text
stringlengths
14
1.76M
2G Second Generation 3G 3${}^{\text{rd}}$ Generation 3GPP 3rd Generation Partnership Project 4G 4${}^{\text{th}}$ Generation 5G 5${}^{\text{th}}$ Generation 5GPPP 5G Infrastructure Public Private Partnership QAM quadrature amplitude modulation ADAS Advanced driver assistance system AD autonomous driving AI artificial intelligence AoA angle of arrival AoD angle of departure API application programming interface AR autoregressive ARQ automatic repeat request BER bit error rate BLER block error rate BPC Binary Power Control BPSK Binary Phase-Shift Keying BRA Balanced Random Allocation BS base station CAM cooperative awareness messages CAP Combinatorial Allocation Problem CAPEX capital expenditure CBF coordinated beamforming CBR congestion busy ratio CDD cyclic delay diversity CDF cumulative distribution function CDL clustered delay line CS Coordinated Scheduling C-ITS cooperative intelligent transportation system CSI channel state information CSIT channel state information at the transmitter D2D device-to-device DCA Dynamic Channel Allocation DCI downlink control information DE Differential Evolution DENM decentralized environmental notification messages DFO Doppler frequency offset DFT Discrete Fourier Transform DIST Distance DL downlink DMA Double Moving Average DMRS Demodulation Reference Signal D2DM D2D Mode DMS D2D Mode Selection DMRS demodulation reference symbol DPC Dirty paper coding DPS Dynamic point switching DRA Dynamic resource assignment DSA Dynamic spectrum access eMBB enhanced mobile broadband eV2X Enhanced vehicle-to-everything EIRP equivalent isotropically radiated power ERTMS European Rail Traffic Management System ETSI European Telecommunications Standards Institute FDD frequency division duplexing FR1 frequency range-1 FR2 frequency range-2 GNSS global navigation satellite system HARQ hybrid automatic repeat request HST high-speed train IAB integrated access and backhaul ITS intelligent transportation system KPI key performance indicator IEEE Institute of Electronics and Electrical Engineers IMT International Mobile Telecommunications IMU inertial measurement unit InC in-coverage IoT Internet of Things ITS intelligent transportation system LDPC low-density parity-check coding LMR land mobile radio LoS line-of-sight LTE Long Term Evolution MAC medium access control mmWave millimeter-wave MBB mobile broadband MCS modulation and coding scheme METIS Mobile Enablers for the Twenty-Twenty Information Society MIMO multiple-input multiple-output MISO multiple-input single-output ML machine learning MRC maximum ratio combining MS mode selection MSE mean square error MTC machine type communications multi-TRP multiple transmission and reception points mMTC massive machine type communications cMTC critical machine type communications NDAF Network Data Analytics Function NF network function NR New Radio NSPS national security and public safety NWC network coding OEM original equipment manufacturer OFDM orthogonal frequency division multiplexing OoC out-of-coverage PSBCH physical sidelink broadcast channel PSFCH physical sidelink feedback channel PSCCH physical sidelink control channel PSSCH physical sidelink shared channel PDCCH physical downlink control channel PDCP packet data convergence protocol PHY physical PLNC physical layer network coding PPPP proximity services per packet priority PPPR proximity services per packet reliability PSD power spectral density RLC radio link control QAM quadrature amplitude modulation QCL quasi co-location QoS quality of service QPSK quadrature phase shift keying PaC partial coverage RAISES Reallocation-based Assignment for Improved Spectral Efficiency and Satisfaction RAN radio access network RA Resource Allocation RAT Radio Access Technology RB resource block RF radio frequency RS reference signal RSRP Reference Signal Received Power SA scheduling assignment SFN Single frequency network SNR signal-to-noise ratio SINR signal-to-interference-plus-noise ratio SC-FDM single carrier frequency division modulation SFBC space-frequency block coding SCI sidelink control information SL sidelink SLAM simultaneous localization and mapping SPS semi-persistent scheduling STC space-time coding SW software TCI transmission configuration indication TBS transmission block size TDD time division duplexing TRP transmission and reception point TTI transmission time interval UAV unmanned aerial vehicle UAM urban air mobility UE user equipment UL uplink URLLC ultra-reliable and low latency communications VUE vehicular user equipment V2I vehicle-to-infrastructure V2N vehicle-to-network V2X vehicle-to-everything V2V vehicle-to-vehicle V2P vehicle-to-pedestrian ZF Zero-Forcing ZMCSCG Zero Mean Circularly Symmetric Complex Gaussian TBS transport block size SCI sidelink control information # 5G New Radio for Automotive, Rail, and Air Transport Gábor Fodor⋆‡ Julia Vinogradova† Peter Hammarberg⋆ Keerthi Kumar Nagalapur⋆ Zhiqiang (Tyler) Qi♭ Hieu Do⋆ Ricardo Blasco† Mirza Uzair Baig⋆ ⋆Ericsson Research Sweden E-mail<EMAIL_ADDRESS> ‡KTH Royal Institute of Technology Sweden. E-mail<EMAIL_ADDRESS> †Ericsson Research Finland Email: <EMAIL_ADDRESS> ♭Ericsson Research China Email<EMAIL_ADDRESS> ###### Abstract The recent and upcoming releases of the 3rd Generation Partnership Project’s 5G New Radio (NR) specifications include features that are motivated by providing connectivity services to a broad set of verticals, including the automotive, rail, and air transport industries. Currently, several radio access network features are being further enhanced or newly introduced in NR to improve 5G’s capability to provide fast, reliable, and non-limiting connectivity for transport applications. In this article, we review the most important characteristics and requirements of a wide range of services that are driven by the desire to help the transport sector to become more sustainable, economically viable, safe, and secure. These requirements will be supported by the evolving and entirely new features of 5G NR systems, including accurate positioning, reference signal design to enable multi- transmission and reception points, service-specific scheduling configuration, and service quality prediction. Keywords: 5G networks, automotive services, high-speed train, urban air mobility, positioning, QoS prediction. ## I Introduction Recent advances in wireless communications, real-time control, sensing, and battery technologies, collaborative spectrum management and sharing, and artificial intelligence are enabling the transport sector to become more cost efficient, secure, and sustainable [1]. Due to new requirements arising in road, railway, air and maritime transport, cellular connectivity, and reliable wireless communications between vehicles and road users are no longer a "nice to have", but are essential parts of cooperative intelligent transportation systems and smart cities [2]. Ericsson predicts that the number of connected cars in operation will rise to more than 500 million in 2025, and the railway sector is making the first steps to digitalize the European Rail Traffic Management System (ERTMS), which includes mission-critical control systems for train operations, including high-speed trains. The unmanned aerial vehicle (UAV) and urban air mobility (UAM) (drone) market is expected to grow from the current estimated USD $4.4$bn to $63.6$bn by 2025 [3]. Apart from smart city applications, there is a growing interest in employing connected UAVs in surface mining, seaports, oil and gas, and other large industrial facilities or in public safety situations in order to improve and optimize industrial processes, enhance operational efficiencies, and create resilience. The digitalization and increasing connectivity in the transport sector are driven by three key factors. First, there are increasing demands imposed virtually by all stakeholders – including passengers, cargo companies, vehicle (car, truck, locomotive, ship) manufacturers, public transport and rail operators, and infrastructure (road, rail, harbor) providers. This broad set of requirements includes being always connected to the Internet and enterprise networks, enjoying safe and secure journeys in urban and rural environments, and minimizing environmental impacts. At the same time, reducing capital and operational expenditures necessitates increasing digitalization, automation, and always-on connectivity, since these technologies make manufacturing, maintaining and operating transport equipment, infrastructure, and services much more efficient. Thirdly, the rapid deployment of 5G networks, and the recent advances in 6G research provide a technology push towards digitalized and connected transport services [4]. In parallel with the above trends in the transport industry, the release 15 (Rel-15) of the 3rd Generation Partnership Project (3GPP) specifications in 2016 marked the birth of the new cellular radio interface for the fifth generation (5G) systems, commonly referred to as New Radio (NR). Although mobile broadband (MBB) services continue to be the main driver for NR, this new radio technology generation inherently has much stronger support for verticals such as the transport industry, as compared to Long Term Evolution (LTE). Additionally, already in Rel-16, new technical features are introduced specifically for supporting critical machine-type communications including ultra-reliable and low latency communications (URLLC) vehicle-to-everything (V2X) services for automotive. Further enhancements targeting special connectivity requirements of the rail operations and remote control of UAVs are also being standardized in the upcoming releases. Compared with 4G systems, 5G NR adopts a new design philosophy and novel technology components, including flexible numerology and waveform design for lower and millimeter-wave frequency bands, minimizing control signaling overhead, multi-hop support by integrated access and backhaul relay, enhanced positioning, and quality of service (QoS) handling mechanisms. Also, further enhanced multiple-input multiple-output (MIMO) techniques enable 5G networks to acquire accurate channel state information (CSI) for both analog and hybrid beamforming and spatial multiplexing applications, which are important for maintaining high spectral efficiency even in high-speed road and rail transport scenarios [5]. Finally, recent 3GPP releases of 5G radio access networks pave the way for advanced radio-based positioning techniques that efficiently complement and make positioning more precise than pure satellite- based positioning techniques [6], which are highly useful for automotive and drone use cases. The present paper serves two purposes. Firstly, we summarize the technical foundations of 5G NR which can fulfill basic requirements imposed by emerging use cases in the transport sector. Secondly, based on an in-depth review of connectivity requirements of transport use cases, we highlight several important new technology enablers which will play a key role in meeting the most stringent requirements. In particular, we focus on the following, * • Positioning techniques that take advantage of combining onboard sensors and cellular network measurements; * • Reference signal design and selecting the appropriate multi- transmission and reception point (TRP) option for spectrum-efficient operations of HSTs and other high-speed user equipments; * • Service-specific scheduling techniques for V2X communications that ensure high resource utilization and service differentiation between low-latency and delay tolerant (lower than best effort) traffic types; * • Novel QoS-prediction techniques that are useful in driverless and driver- assisted road, rail, and drone transport use cases. ## II Technical Foundations of 5G NR and the Initial NR Evolution Targeting the Transport Vertical Figure 1: Key areas and relevant transport use cases in NR Rel-15, Rel-16, and Rel-17. ### II-A Major Features in Rel-15 and Rel-16 As mentioned, Rel-15 and Rel-16 of the 3GPP specifications have been largely driven by requirements of MBB services, including requirements on enhanced data rates, latency, coverage, capacity, and reliability. However, starting already in Rel-15 and continuing in the subsequent 3GPP releases, NR enables new use cases by meeting the requirements imposed by transport use cases, such as connected cars, high-speed trains, and UAVs. While Rel-15 focused on supporting MBB and URLLC applications, Rel-16/17 includes UE power savings, operation in unlicensed spectrum, industrial Internet of Things (IoT) enhancements as well as special radio access network (RAN) features such as physical layer support for unicasting sidelink (device-to-device) for advanced V2X services [7]. A key distinguishing feature of 5G NR from fourth generation (4G) systems is the substantial expansion of the frequency bands, in which NR can be deployed. For transport applications, the following NR-specific features are particularly important (see Figure 1): * • Symmetric physical layer design with orthogonal frequency division multiplexing (OFDM) waveform for all link types, including uplink, downlink, sidelink, and backhaul; * • Wide range of carrier frequencies, bandwidths, and deployment options. 3GPP aims to develop and specify components and physical layer numerology that can operate in frequencies up to 100 GHz. This implies several options for OFDM subcarrier spacing ranging from 15 kHz up to 240 kHz with a proportional change in cyclic prefix duration; * • Native support for dynamic time division duplexing (TDD) as a key technology component. In dynamic TDD, parts of a slot can be adaptively allocated to either uplink or downlink, depending on the prevailing traffic demands; * • Support for massive MIMO, that is a massive number of steerable antenna elements for both transmission and reception, utilizing channel reciprocity in TDD deployments and a flexible CSI acquisition framework. The NR channels and signals, including those used for data transmission, control signaling and synchronization are all designed for optional beamforming. In addition to flexible numerology, native support for dynamic TDD and advanced MIMO features, NR is designed using the principle of ultra-lean design, which aims at minimizing control plane and synchronization signal transmissions when data transmissions are idle. Inherent support for distributed MIMO, also referred to as multi-TRP, is introduced and fully supported in Rel-16. This feature is largely motivated by HST applications, since it allows UEs to receive multiple control and data channels per slot, which enables simultaneous data transmissions from multiple physically separated base stations. ### II-B Major Developments in Rel-17 Looking beyond Rel-16, the NR evolution will be driven by industry verticals, including a variety of transport use cases, such as V2X communications, high- speed trains, UAVs and passenger aircrafts, and maritime communications. These use cases justify new features discussed and planned for Rel-17. MIMO enhancements are expected to support multi-TRP specific tracking reference signals, single frequency network deployments, and non-coherent joint transmissions by multiple base stations, which are particularly useful for providing connectivity to high-speed trains. Furthermore, Rel-17 is studying the integration of non-terrestrial and terrestrial networks in order to support use cases for which terrestrial networks alone cannot provide the required coverage and capacity, including maritime, UAV, and UAM scenarios. ## III Overview of Intelligent Transportation Systems Services and Requirements Figure 2: Use case categories in the automotive and road transport (upper), railway (middle), and UAV and UAM (lower) segments. Figure 2 classifies the automotive, rail, and UAV/UAM use cases in use case categories, together with the key connectivity requirements per category. Regulated C-ITSs provide international or governmental regulated services for road, rail and drone traffic efficiency, sustainability, and safety. Traffic efficiency use cases have relaxed latency requirements, while safety-related data often requires URLLC. A benefit of regulation is to facilitate original equipment manufacturer (OEM) cooperation in standardized information exchange. C-ITS services may also use dedicated spectrum in certain regions; for example, for direct short-range communication using the 3GPP sidelink technology. For rail transport, C-ITS implies station dwell time control and speed/break control to optimize rail network utilization while ensuring safety. Advanced driver assistance systems and autonomous driving (AD) are increasingly employed for both road and rail transport. In Europe, for example, the next generation of the ERTMS will support well-defined levels of automation, including semi-automated (assisted) driving, driverless and unattended train operation. Similarly, advanced pilot assistance systems for UAM and passenger aircrafts are envisioned by various stake-holders of the air transport industry. For this set of applications, URLLC communication and high-accuracy positioning play crucial roles. Fleet management including remote assistance of driverless vehicles is an important application for road, rail, and UAV-based transport. This type of services aim at vehicle fleet owners such as logistics or car-sharing companies. The communication service is primarily used to monitor vehicle locations and the vehicle/driver status. With increasing level of automation in the rail industry and for UAVs, or for a fleet of driverless trucks, fleet management also includes communication support for operations monitoring and remote assistance from a control tower. Convenience and infotainment, based on MBB services for drivers, crew, and passengers are equally important in road, rail, and future UAM transport use cases. Such services deliver content such as traffic news and audio entertainment for car drivers, and gaming and video entertainment for passengers. One specific example is the concept of "Gigabit train" services, which motivate the adoption of HST scenarios in 3GPP. For this set of use cases, the most important requirement is high data rate and low latency connections, which rely heavily on the capability of tracking wireless channels at high vehicle speed. The primary focus in the logistics and connected goods category is on the tracking of transported objects (commodities, merchandise goods, cargo and so on) during the production and transport cycle of the object. Near real time tracking and status monitoring of goods are attractive for cargo companies, customers, and freight train operators. In the vehicle-as-a-sensor use case category, sensors installed in the vehicle sense the environment and can also provide anonymized data to 3rd parties. In road transport, for example, the vehicle-mounted sensors provide information for solutions such as ADAS or AD as well as for monitoring city infrastructure and road status. For rails, railway track monitoring and anomaly detection are supported by various sensors mounted on the train, effectively operating the train as a sensor. Similarly, drones can be equipped with a lot of sensors that help collect data and perform distributed or federated computation for various purposes such as forecasting cloud formation, rain, and other hazardous weather conditions. Just as with the convenience and infotaiment category, the vehicle-as-a-sensor requires high data-rate connections between vehicles or between vehicles and the cellular network at high vehicle speed and dynamic interference conditions. Telematics applications for vehicles include collecting vehicle diagnostics to monitor/adjust the vehicle, while rail telematics rail applications allow continuous status updates from trains to determine state, delay, cargo conditions, software (SW) updates, and geo-fencing. In this category, several applications (e.g. SW updates) tolerate delay, while others are more delay critical. Similarly, for air transport, telematics serve as a tool for collecting air vehicles diagnostics to monitor/adjust the vehicles. To make sense of the vast amount of data collected from vehicles in this group of use cases, the role of artificial intelligence (AI)/ machine learning (ML) is utterly important. In the reverse direction, AI/ML can also play a meaningful role in determining when to perform a certain task to the vehicle in an efficient manner. For example, ML-based spare capacity prediction, which is part of the so-called cellular network QoS prediction, can be used to predict the most economical time for SW update for a set of vehicles [8, 9]. Connected road infrastructure services are operated by cities and road authorities to monitor the state of the traffic and control its flow, such as physical traffic guidance systems, parking management and dynamic traffic signs. For railways, the communication between the rail infrastructure and the locomotive via specific transmission modules and eurobalises is used to send information from line-side electric units to the trains e.g., current speed restrictions for the coming rail segment. For UAV/UAM, an Unmanned Aircraft System Traffic Management (UTM) is used for traffic control, which requires high-accuracy 3D positioning and URLLC communication with the ground control system. ## IV Proposed New Features and Solutions to Support intelligent transportation system (ITS) Requirements Despite the recent and ongoing enhancements to 5G NR, there is still a need to further improve the technology to meet the growing demands of industry verticals. In this section, we summarize the state of the standardization of several specific features and introduce new solutions which can help fulfilling the stringent connectivity requirements of the transport sectors outlined in the preceding section. These components span both radio layers (physical, medium access control) and the service layer of the protocol stack. ### IV-A Advanced Positioning Support and Algorithms With the introduction of NR, 3GPP targets improved positioning capabilities to cater for a number of new use cases, involving indoor, industrial, and automotive applications. Cooperative manoeuvring in the C-ITS category and several ADAS applications rely on accurate positioning, which must remain operational even in global navigation satellite system (GNSS)-problematic areas. Specifically in Rel-16, a set of positioning related features are introduced, which pave the way for enhanced positioning services. These new features include new and improved uplink (UL) and downlink (DL) reference signal design, allowing larger bandwidths and beamforming, assisted by additional measurements and enhanced reporting capabilities. By supporting larger bandwidths than in LTE, higher accuracy of range estimates can be achieved, and with angle of arrival (AoA) and angle of departure (AoD) measurements new positioning schemes exploiting the spatial domain can be supported. Architecture-wise, similarly to LTE, NR positioning is based on the use of a location server. The location server collects and distributes information related to positioning (UE capabilities, assistance data, measurements, position estimates, etc.) to the other entities involved in the positioning procedures (base stations and connected vehicles). A range of positioning schemes, including DL-based and UL-based ones, are used separately or in combination to meet the accuracy requirements in vehicular scenarios. Specifically, in the millimeter-wave (mmWave)-bands, referred to as frequency range-2 (FR2) bands of NR, the AoA and AoD measurements can be enhanced by using large antenna arrays, which facilitate high resolution angular measurements. By unlocking the spatial domain, NR can significantly increase the positioning accuracies for many industrial and automotive use cases [10], [11]. Additionally, to further improve the accuracy and reliability of positioning in GNSS-problematic areas, we propose to fuse acceleration measurements provided by onboard inertial measurement units with measurements on the received NR DL reference signals from multiple NR BSs. Fusing local IMU measurements with measurements on multiple NR DL signals can reach decimeter accuracy in favourable deployments. Furthermore, high accuracy spatial and temporal measurements facilitate the use of advanced positioning schemes such as simultaneous localization and mapping (SLAM), which utilizes consecutive measurements to build a statistical model of the environment, achieving high accuracy even in extreme scenarios by utilizing measurements on DL signals of only a single base station. ### IV-B Further Enhanced MIMO to Support Multiple Transmission and Reception Points HST wireless communication is characterized by a highly time-varying channel and rapid changes of the closest TRPs to the train, resulting from the extreme high velocities. Recognizing these challenges, NR has been designed to support high mobility from day one, and includes features to enable communications with HSTs [12]. Furthermore, several multi-TRP deployment options and features developed under the general MIMO framework can be exploited to support HST communications for the "Gigabit train", while minimizing the need for handovers. We expect this technology component will also play a very important role in UAV/UAM use cases. The multi-TRP options that are the most relevant to HST communications are: * • Dynamic point switching (DPS): Data signals are transmitted from a single TRP at a given time, and the TRP used for transmission is dynamically selected based on the relative quality of channels between the train and a few closest TRPs; * • Single frequency network (SFN): All the TRPs in the SFN area transmit the same data and reference signals to the train; * • SFN with TRP-specific reference signals: The same data signal is transmitted from different TRPs, while some of the reference signals are transmitted in a TRP-specific manner. The first and the third options rely on TRP-specific reference signals, whereas the second option uses common reference signal across the TRPs in the coverage area of the SFN. In addition to supporting TRP-specific reference signals, NR supports associating different reference signals with different channel properties, such as Doppler shift and delay spread, through NR’s quasi co-location (QCL) and transmission configuration indication framework. In Rel-17, different QCL enhancements are investigated to better support advanced channel estimation schemes that can be implemented at the train and to evaluate the need for TRP-side pre-compensation algorithms. In Section V, we evaluate the necessity and performance of the multi-TRP options described above through link level evaluations. Furthermore, beam management enhancements necessary to support HST communications in the higher bands of NR are also investigated in Rel-17. ### IV-C Service-Specific Scheduling To accommodate MBB, URLLC, and machine-type communications services, NR networks employ scheduling algorithms that take into account the current service mix, prevailing channel conditions, traffic load, available carriers, and other factors. Scheduling in multiservice wireless networks has been researched for decades and a vast literature as well as practically deployed scheduling schemes exist. Interestingly, due to requirements imposed by the coexistence of URLLC and delay tolerant services, mmWave communications support in the FR2 bands and serving very high speed UE-s have stimulated renewed research interest in this topic [13]. Some recent works propose scheduling strategies to minimize end-to-end delay for time-critical services [14], or to optimize resource allocation for the coexistence of various services [13]. In NR, service-specific scheduling can also be configured to take into account the specific opportunities that are present in certain deployment scenarios. We propose to customize the scheduler in certain deployment scenarios, which can be illustrated by a scheduling configuration that is suitable for non-time-critical services. This can be applicable for background data transfer for SW updates or uploading sensor measurements in the vehicle-as-a-sensor category. This scheduling mechanism divides vehicles into a high and a low path-gain group. Vehicles belonging to the high path-gain group are eligible for medium access, while scheduling vehicles in the low path-gain group is postponed (dropped) until their path gain improves. The scheme is suitable for highly mobile users (including automotive or urban train use cases) in high-way, urban or suburban scenarios, and can be activated or de-activated based on velocity or other sensor measurements that help to configure the path gain threshold. ### IV-D QoS Prediction The NR QoS framework together with features like URLLC are successful in delivering a minimum guaranteed performance, especially in controlled scenarios. However, highly mobile UEs usually experience time-variant network performance, partly because the actual QoS often exceeds the minimum or guaranteed level, and partly because the system is occasionally not able to fulfill a QoS provision. Interestingly, in many cases, including certain C-ITS, ADAS or telematics applications, these performance fluctuations are not a problem if they can be predicted in advance. Having access to real-time QoS predictions has generated a large interest from the automotive industry [8], as it would allow service providers, mobile network users, and automotive applications to dynamically adapt their behaviors to the prevailing or imminent QoS level. This would allow for enabling services relying on continuous guaranteed performance as well as for exploiting spare capacity for large bulk data transfers in lower than best effort services. Despite the high expectations, QoS prediction is largely an open research topic. The realistically achievable performance and the applicability of this type of algorithms are still unknown. To a large extent, QoS prediction is seen as a ML application with a broad data set consisting of network measurements, device measurements, and application data [9]. In practice, different types of information may be collected with different periodicities, time horizons, resolutions, and accuracies, depending on practical and business-related constraints. Understanding the relevance of each of them is ongoing work and will be instrumental in determining the relative merits and tradeoffs of the different architecture options, which range from network-centric to application-based. As a first step towards supporting predictive QoS in mobile networks, 3GPP enhanced the NR system architecture in Rel-16 to support services providing network data analytics in the 5G core network. To this end, application programming interfaces for exposing network-based predictions were defined, the necessary procedures for collecting the relevant data for the analytics from different network functions as well as from the operations administration and management functionality. In addition, procedures providing analytics (e.g., load, network performance, data congestion, QoS sustainability and UE related analytics) to other network functions were introduced. As usual, the algorithms used to obtain the network analytics are not defined in the specification, and as said, they are a topic of ongoing research. ## V Numerical Examples ### V-A Positioning Figure 3: CDF of the obtained positioning accuracy when using NR only and fused IMU+NR positioning at 5 and 15 dB SNR. We consider a highway scenario with BSs equally spaced and placed at the side of the road. A moving vehicle following a snake-like trajectory is equipped with an onboard IMU sensor and an NR UE. A MIMO system is considered with square antenna arrays at the BS and the UE. A line-of-sight (LoS) downlink propagation is assumed with a grid of Discrete Fourier Transform beams transmitted by the BS. A sensor fusion-based positioning approach is proposed (similar to the one in [11]), for which a Kalman filter is used to fuse the measurements obtained from the IMU and the NR downlink such as the range and the angles-of-arrival. An extension to this is proposed allowing to fuse measurements from multiple BS. Positioning accuracy in terms of positioning cumulative distribution functions are compared in Figure 3 for NR-based method and for the sensor fusion-based method combining IMU with NR. It is assumed that the BSs operate at the millimeter wave frequency of 28 GHz with 256 antennas. There are 4 antennas at the UE, the UE speed is equal to 130 km/h, and signal-to-noise ratio (SNR) is equal to 5 and 15 dB. The BSs are placed at 40 m from the road with the inter- site distance equal to 200 m. The results are averaged over a total distance of 10 km. The number of BSs used to fuse the measurements from is denoted by NbFusedBS. The simulation results show that a large performance gain is obtained for the sensor fusion-based method as compared to the NR only-based method, especially at low SNR. Notice that fusion-based methods allow to achieve a decimeter level accuracy with greater than 90% probability, even at low SNR of 5dB. ### V-B Further Enhanced MIMO to Support Multiple Transmission and Reception Points Figure 4: High-speed train multi-TRP scenario with four base stations (upper) and downlink throughput as a function of train position when using SFN without/with CDD, DPS and SFN with precompensation. (The $x,y$ and $z$ axes are scaled by 10, 10 and 7 respectively for better visualization.) For evaluations, a four-TRP deployment at 2 GHz carrier frequency using 20 MHz bandwidth (50 resource blocks) with TRP hight of 35m, 30 kHz subcarrier spacing, and train speed of 500 km/h is assumed, as illustrated in Figure 4 (upper) is used. The TRP antenna orientation is set to 10 degrees downtilt with an antenna gain of 20.5 dBi. The channel between each TRP and the reception point at the train is modeled using an extended clustered delay line (CDL) channel model. The SNR at train position $D1=0$ m is 16 dB in the SFN deployment, and an hybrid automatic repeat request (HARQ) scheme with a maximum number of 3 retransmissions is employed, using a fixed modulation and coding scheme (MCS) with 64- quadrature amplitude modulation (QAM), low- density parity- check coding with code-rate = 0.428. Figure 4 (lower) shows the throughput as a function of distance in the different deployment options. In the baseline SFN transmission scheme, the throughput for UE locations, close to midpoint of two TRPs does not reach the peak throughput of the modulation and coding scheme used. This is due to the fact that the equivalent channel formed by the combination of the two dominant CDL channels with LoS components having equal and opposite Doppler shifts results in a less frequency selective channel with deep fades across some of the OFDM symbols in a slot. This channel behavior can be modified by adding a TRP-specific cyclic delay diversity which converts the effective channel close to midpoint between TRPs to a frequency selective channel without deep fades across OFDM symbols. The throughput improvement with TRP-specific cyclic delay diversity is shown in the figure. A precompensation scheme where a TRP-specific Doppler frequency offset (DFO) compensation is performed also improves the throughput close to the midpoint, as seen in the figure. However, this scheme requires TRP- specific reference signals in order for the train to estimate the Doppler shifts and additional signaling in the uplink direction to feedback the estimates. The figure also shows the performance of the DPS scheme with genie selection of the TRPs, where a single TRP closest to the train is used for data transmission. In case of DPS, the received SNR at a train position is smaller than in the case of SFN transmissions due to transmission from a single TRP. However, if sufficient SNR can be guaranteed using proper deployment, DPS achieves peak throughput of the MCS used at all train positions. The evaluation results show that a complex scheme, such as SFN with DFO precompensation, does not perform significantly better than the other alternatives. Also, a low-complexity and UE transparent scheme such as SFN with TRP-specific CDD, which can be readily employed and scaled to serve a large number of UEs, performs well by suitably altering the effective channel. ### V-C Service-Specific Scheduling Figure 5: Downlink mean user throughput as a function of the traffic density when employing different drop rates in an automotive highway scenario, in which the base station sites are deployed with 1732m inter-site distance. To illustrate the impact of customizing the scheduler to operate in specific deployment scenarios, we consider an automotive high-way scenario, in which the BS sites are deployed according to the 3GPP recommendation in [15]. Figure 5 shows that dropping low path-gain users improves the mean user throughput, especially in case of high traffic density. Specifically, when all users are scheduled simultaneously (marked as 0% Drop) the mean user throughput drops to close 0 when the traffic is around or higher than 1000 Mbps/km2. When the scheduler is configured to distinguish low and high path- gain vehicles and postpones scheduling vehicles that are momentarily have low path-gain, the average throughput significantly increases. Dropping 50% of the low path-gain users (Drop 50% line), for example, improves the mean user throughput when traffic density lies between 1000-3000 Mbps/ km2. This simple example illustrates that adjusting the drop ratio (by adjusting the threshold between low and high path gain vehicles and setting the drop ratio in the low path-gain group) – according to the deployment parameters and prevailing traffic density – can optimize the system spectral efficiency. The basic rationale for this is that for non-latency-critical services, the user data transmission can wait until the user is moving into good coverage area, while users in poor coverage area can be dropped to improve both system resource utilization and spectral efficiency. As an example, consider the highway scenario, in which vehicles drive at 140km/h. To guarantee 95% MBB-like service coverage (10 Mbps data rate for DL and 2 Mbps for UL), the DL capacity for the non-dropping case is 450 Mbps/km2 (760 Mbps/km2 for UL). It takes 50s (250s for UL) to transmit a 500M file in the DL. For the drop 50% case, for the same traffic density (450 Mbps/km2 for DL and 760 Mbps/km2 for UL), the transmission time is greatly reduced (19s for DL and 31s for UL). Although the Drop 50% case has about 50% coverage hole, this coverage hole lasts until the vehicles drive by a next BS, improve their path loss, and complete the ongoing file transmission. ### V-D QoS Prediction Figure 6: Predictive QoS: Linking network-level QoS level and the application layer (upper), Network-based prediction function utilizing ML (middle) and DL throughput prediction (lower). The most fundamental question of predictive QoS is, perhaps, the accuracy of the predictions. We have studied multiple alternatives to predict DL throughput in different time horizons. Figure 6 summarizes our findings in terms of the CDF for different prediction horizons of the difference between predicted and delivered number bits $B$ over a given time interval $\Delta t$: $\displaystyle e^{\prime}(t,\Delta t)$ $\displaystyle\triangleq\frac{\Big{|}B_{\text{delivered}}(t)-B_{\text{predicted}}(t)\Big{|}}{\Delta t}.$ Our results show that both short-term and long-term predictions are quite accurate in most cases. In the former case, this is due to the ability to predict short term channel quality fluctuations, while other system variables that are harder to predict (e.g., interference level, instantaneous cell load) are relatively stable. In the latter case, the longer interval averages out the instantaneous variations of channel quality and other short-term effects. In contrast, DL throughput prediction in the intermediate regime is much more challenging. In this case, rapid channel fluctuations are often not well predicted, while the averaging effect is still weak. How to bridge the gap between short- and long-term predictions is still an open question. ## VI Concluding Remarks and Outlook 5G NR was designed to enable various use cases, reach a broad range of aggressive performance targets and be deployed in both traditional and mmWave frequency bands. The initial release (Rel-15) of NR included support for flexible numerology, latency-optimized frame structure, massive MIMO, interworking between low and high frequency bands and dynamic TDD. The new features of NR in subsequent releases include enhancements for MIMO, V2X, high-speed UE, and URLLC services, more accurate positioning, and support for non-terrestrial and mission-critical communications. These new standardized features, together with proprietary and algorithmic solutions facilitate connected and intelligent transport services, including automotive, rail, air transport and public safety services. Connected and intelligent transport systems will continue to rely on ubiquitous broadband connectivity as expectations by the automotive, rail and air transport and public safety stakeholders evolve, and new business models emerge. The contours of future 6G systems are already emerging, as the standardization and research communities and end-users in the transport sector define future requirements and solutions. Based on these discussions, 6G technology candidates include both further enhancements of existing features and entirely new features. The former group includes pushing the limits of frequencies towards the lower bands of THz spectrum, while examples for the latter include integrated communication and radar sensing, integrated terrestrial and non-terrestrial networks, and utilizing intelligent reconfigurable surfaces and full-duplex communications. ## References * [1] F. R. Soriano, J. J. Samper-Zapater, J. J. Martinez-Dura, R. V. Cirilo-Gimeno, and J. M. Plume, “Smart mobility trends: Open data and other tools,” _IEEE Intelligent Transportation Systems Magazine_ , vol. 10, no. 2, pp. 6–16, Jul. 2018. * [2] S. Zeadally, M. A. Javed, and E. B. Hamida, “Vehicular communications for ITS: Standardization and challenges,” _IEEE Communications Standards Magazine_ , vol. 4, no. 1, pp. 11–17, Mar. 2020. * [3] M. Alwateer and S. W. Loke, “Emerging drone services: Challenges and societal issues,” _IEEE Technology and Society Magazine_ , vol. 39, no. 3, pp. 47–51, Sep. 2020. * [4] K. Samdanis and T. Taleb, “The road beyond 5G: A vision and insight of the key technologies,” _IEEE Network_ , vol. 34, no. 2, pp. 135–141, Feb. 2020\. * [5] S. Han, C.-L. I, T. Xie, S. Wang, Y. Huang, L. Dai, Q. Sun, and C. Cui, “Achieving high spectrum efficiency on high speed train for 5G New Radio and beyond,” _IEEE Wireless Communications_ , vol. 26, no. 5, pp. 62–69, Oct. 2019. * [6] X. Lin, J. Bergman, F. Gunnarsson, O. Liberg, S. M. Razavi, H. S. Razaghi, H. Rydn, and Y. Sui, “Positioning for the internet of things: A 3GPP perspective,” _IEEE Communications Magazine_ , vol. 55, no. 12, pp. 179–185, Sep. 2017. * [7] S. A. Ashraf, R. Blasco, H. Do, G. Fodor, C. Zhang, and W. Sun, “Supporting vehicle-to-everything services by 5G New Radio Release-16 systems,” _IEEE Communications Standards Magazine_ , vol. 4, no. 1, pp. 26–32, Mar. 2020. * [8] 5GAA Automotive Association, “Making 5G proactive and predictive for the automotive industry,” www.5gaa.org, Tech. Rep., 2019. * [9] D. Raca, A. H. Zahran, C. J. Sreenan, R. K. Sinha, E. Halepovic, R. Jana, and V. Gopalakrishnan, “On leveraging machine and deep learning for throughput prediction in cellular networks: Design, performance, and challenges,” _IEEE Communications Magazine_ , pp. 11–17, Mar. 2020. * [10] H. Wymeersch, G. Seco-Granados, G. Destino, D. Dardari, and F. Tufvesson, “5g mmwave positioning for vehicular networks,” _IEEE Wireless Communications_ , vol. 24, no. 6, pp. 80–86, 2017. * [11] S. S. Mostafavi, S. Sorrentino, M. B. Guldogan, and G.Fodor, “Vehicular positioning using 5G millimeter wave and sensor fusion in highway scenarios,” in _IEEE International Conference on Communications (ICC)_ , Dublin, Ireland, Ireland, Jun. 2020. * [12] G. Noh, B. Hui, and I. Kim, “High speed train communications in 5G: Design elements to mitigate the impact of very high mobility,” _IEEE Wireless Communications_ , pp. 1–9, 2020. * [13] A. A. Esswie and K. I. Pedersen, “Opportunistic spatial preemptive scheduling for URLLC and eMBB coexistence in multi-user 5G networks,” _IEEE Access_ , vol. 6, pp. 38 451 – 38 463, Jul. 2018. * [14] Y. Hu, H. Li, Z. Chang, and Z. Han, “Scheduling strategy for multimedia heterogeneous high-speed train networks,” _IEEE Transactions on Vehicular Technology_ , vol. 66, no. 4, pp. 3265 – 3279, Apr. 2017. * [15] 3GPP, “Study on LTE-based V2X services,” RAN WG TR 36.885, Tech. Rep., 2016\. *[ERTMS]: European Rail Traffic Management System *[UAV]: unmanned aerial vehicle *[UAM]: urban air mobility *[UAVs]: unmanned aerial vehicle *[3GPP]: 3rd Generation Partnership Project *[NR]: New Radio *[MBB]: mobile broadband *[LTE]: Long Term Evolution *[URLLC]: ultra-reliable and low latency communications *[V2X]: vehicle-to-everything *[QoS]: quality of service *[MIMO]: multiple-input multiple-output *[CSI]: channel state information *[TRP]: transmission and reception point *[HSTs]: high-speed train *[IoT]: Internet of Things *[RAN]: radio access network *[OFDM]: orthogonal frequency division multiplexing *[TDD]: time division duplexing *[HST]: high-speed train *[UEs]: user equipment *[C-ITSs]: cooperative intelligent transportation system *[OEM]: original equipment manufacturer *[C-ITS]: cooperative intelligent transportation system *[AD]: autonomous driving *[ADAS]: Advanced driver assistance system *[SW]: software *[AI]: artificial intelligence *[ML]: machine learning *[ITS]: intelligent transportation system *[GNSS]: global navigation satellite system *[UL]: uplink *[DL]: downlink *[AoA]: angle of arrival *[AoD]: angle of departure *[mmWave]: millimeter-wave *[FR2]: frequency range-2 *[BSs]: base station *[IMU]: inertial measurement unit *[SLAM]: simultaneous localization and mapping *[TRPs]: transmission and reception point *[DPS]: Dynamic point switching *[SFN]: Single frequency network *[QCL]: quasi co-location *[UE]: user equipment *[BS]: base station *[LoS]: line-of-sight *[SNR]: signal-to-noise ratio *[CDD]: cyclic delay diversity *[CDL]: clustered delay line *[HARQ]: hybrid automatic repeat request *[MCS]: modulation and coding scheme *[QAM]: quadrature amplitude modulation *[DFO]: Doppler frequency offset *[CDF]: cumulative distribution function
# Efficient MPI-based Communication for GPU-Accelerated Dask Applications ††thanks: This research is supported in part by NSF grants #1818253, #1854828, #1931537, #2007991, #2018627, and XRAC grant #NCR-130002. Aamir Shafi, Jahanzeb Maqbool Hashmi, Hari Subramoni and Dhabaleswar K. (DK) Panda The Ohio State University {shafi.16, hashmi.29, subramoni.1<EMAIL_ADDRESS> ###### Abstract Dask is a popular parallel and distributed computing framework, which rivals Apache Spark to enable task-based scalable processing of big data. The Dask Distributed library forms the basis of this computing engine and provides support for adding new communication devices. It currently has two communication devices: one for TCP and the other for high-speed networks using UCX-Py—a Cython wrapper to UCX. This paper presents the design and implementation of a new communication backend for Dask—called MPI4Dask—that is targeted for modern HPC clusters built with GPUs. MPI4Dask exploits mpi4py over MVAPICH2-GDR, which is a GPU-aware implementation of the Message Passing Interface (MPI) standard. MPI4Dask provides point-to-point asynchronous I/O communication coroutines, which are non-blocking concurrent operations defined using the async/await keywords from the Python’s asyncio framework. Our latency and throughput comparisons suggest that MPI4Dask outperforms UCX by $6\times$ for 1 Byte message and $4\times$ for large messages (2 MBytes and beyond) respectively. We also conduct comparative performance evaluation of MPI4Dask with UCX using two benchmark applications: 1) sum of cuPy array with its transpose, and 2) cuDF merge. MPI4Dask speeds up the overall execution time of the two applications by an average of $3.47\times$ and $3.11\times$ respectively on an in-house cluster built with NVIDIA Tesla V100 GPUs for $1-6$ Dask workers. We also perform scalability analysis of MPI4Dask against UCX for these applications on TACC’s Frontera (GPU) system with upto $32$ Dask workers on $32$ NVIDIA Quadro RTX 5000 GPUs and $256$ CPU cores. MPI4Dask speeds up the execution time for cuPy and cuDF applications by an average of $1.71\times$ and $2.91\times$ respectively for $1-32$ Dask workers on the Frontera (GPU) system. ###### Index Terms: Python, Dask, MPI, MVAPICH2-GDR, Coroutines ## I Introduction With the end of Moore’s Law [1, 2] in sight, the performance advances in the computer industry are likely to be driven by the “Top”—1) hardware architecture, 2) software, and 3) algorithms—as noted by Leiserson recently [3]. This is in stark comparison to the vision “There’s Plenty of Room at the Bottom” laid out by Feynman [4] in 1959 referring to semiconductor physics and fabrication technology. Leiserson [3] argues that the post-Moore generation of software will focus on reducing software engineering bloat as well as exploiting parallelism, locality, and heterogeneous hardware. A representative of the “Top” is the Python programming language, which is a clear winner on the landscape of data science. Python is a classic example of a language benefiting from Moore’s law by traditionally focusing on programmer productivity and reduced development time. In the context of data science applications, Python’s popularity is due to rich set of free and open-source libraries that enable the end-to-end data processing pipeline. These libraries/packages include: core (SciPy), data preparation (NumPy, Pandas), data visualization (Matlibplot), machine learning (Scikit-learn), and deep learning (PyTorch, TensorFlow). However in the post-Moore era, it is vital that Python is able to support parallel and distributed computing as well as exploit emerging architectures especially accelerators. Two popular big data computing frameworks that enable high-performance data science in Python include Dask [5] and Apache Spark [6]. This paper focuses on Dask, which provides support for natively extending popular data processing libraries like numPy and Pandas. This allows incremental development of user applications and lesser modifications to legacy codes when executed with Dask. Traditionally Dask has mainly supported execution on hosts (CPUs) only, which means that it was not able to exploit massive parallelism offered by Graphical Processing Units (GPUs). However this has recently changed as part of the development of the NVIDIA RAPIDS library, which aims to enable parallel and distributed computation on clusters of GPUs. RAPIDS adopted a similar approach—as taken by Dask—of extending already existing Python data processing libraries for the GPU ecosystem. For instance, RAPIDS support processing of distributed data stored in cuPy (numPy-like) and cuDF (Pandas-like) formats. An overarching goal for RAPIDS project is to hide the low-level programming complexities of the CUDA compute environment from Python developers and make it easy to deploy and execute full end-to-end processing pipeline on GPUs. An example of a machine learning library provided by RAPIDS is cuML [7], which is the GPU-counterpart for Scikit-learn. In order to support execution of Dask programs on cluster of GPUs, an efficient communication layer is required. This is extended by the Dask Distributed library that provides essentials for distributed execution of Dask programs on parallel hardware. It is an asynchronous I/O application, which means that it supports non-blocking and concurrent execution of its routines/functions including communication primitives. This mandates the restriction that any communication backend—aiming to be part of Dask Distributed library—must implement coroutines that are non-blocking methods defined and awaited by using the async and await keywords respectively. Coroutines support concurrent mode of execution and hence cannot be invoked like regular Python functions. The main reason behind this non-blocking asynchronous style of programming is to avoid blocking, or delaying, networking applications for completing I/O operations. The Dask Distributed library currently provides two communication devices: one for TCP and the other for UCX-Py that is a Cython-based wrapper library to UCX [8]. The UCX-Py communication device is capable of communicating data to/from GPU device memory directly. However, it fails to deliver high-performance—as revealed by our performance evaluation detailed in Section V—since it delays progressing the communication engine by assigning it to a separate coroutine that executes periodically. We address this performance overhead by advocating an efficient design where communication coroutines also progress the communication engine. The Message Passing Interface (MPI) standard [9] is considered as the defacto programming model for writing parallel applications on modern GPU clusters. The MVAPICH2-GDR [10] library provides high-performance support for communicating data to/from GPU devices via optimized point-to-point and collective routines. As part of this paper, we design, implement, and evaluate a new communication backend, called MPI4Dask, based on the MVAPICH2-GDR library for Dask. MPI4Dask implements point-to-point communication coroutines using mpi4py [11] over MVAPICH2-GDR. To the best of our knowledge, this is the first attempt to use an GPU-enabled MPI library to handle communication requirements of Dask. This is challenging especially because the Dask Distributed library is an asynchronous I/O application. Hence any point-to- point communication operations—implemented via MPI—must be integrated as communication coroutines that typically have performance penalties over regular Python functions. An additional goal here is to avoid making unnecessary modifications to the Dask ecosystem. Another challenge that MPI4Dask addresses is to provide communication isolation and support for dynamic connectivity between Dask components including scheduler, workers, and client. Communication between Dask entities—as provided by TCP and UCX backends—is based on the abstraction of endpoints that are established when processes connect with one another. An endpoint essentially represent a direct point-to-point connection. In MPI4Dask , we devise a strategy, detailed in Section IV, that relies on sub- communicator mechanism provided by MPI—coupled with communicator duplication—to handle this. This allows us to build sub-communicators between processes to mimic a direct endpoint-based connection. In order to motivate the need for MPI4Dask, we present latency comparison between Python communication coroutines implemented using MVAPICH2-GDR and mpi4py with UCX-Py polling and event-based modes in Figure 1. The latency graphs—Figure 1(a) and Figure 1(b) depict that the performance of Python communication coroutines using MVAPICH2-GDR (with mpi4py) are roughly $5\times$ better for small messages ($1$ Byte to $16$ Bytes) and $2-3\times$ better for small-medium messages ($32$ Bytes to $4$ KBytes). For large messages ($2$ MBytes to $128$ MBytes) as shown in Figure 1(c), MVAPICH2-GDR (with mpi4py)-based coroutines are better than UCX-Py point-to-point coroutines by a factor of $3-4\times$. These performance benefits form the basis for our motivation to design and implement MPI-based communication backend for the Dask framework. (a) Latency (Small) (b) Latency (Medium) (c) Throughput (Large) Figure 1: Latency and Bandwidth comparison of MVAPICH2-GDR (using mpi4py) with UCX using Ping-pong Benchmark—based on Python Coroutines—on the RI2 Cluster (V100 GPUs). This demonstrates the performance benefits of of using MVAPICH2-GDR as compared to UCX at the Python layer and forms the major motivation of this paper. We present a detailed performance evaluation of MPI4Dask against TCP and UCX communication backends using a number of micro-benchmarks and application benchmarks—these are presented in Section V. For the micro-benchmark evaluation, MPI4Dask outperforms the UCX communication device in latency and throughput comparison by ping-pong benchmarks by $5\times$ for small messages, $2-3\times$ better for small-medium messages, and $3-4\times$ for large messages. We used these two application benchmarks: 1) sum of cuPy array and its transpose, and 2) cuDF merge on an in-house cluster and the TACC’s Frontera (GPU) cluster. On the in-house cluster, we are witnessing an average speedup of $3.47\times$ for $1-6$ Dask workers for the cuPy application. Communication time for this application has been reduced by $6.92\times$ compared to UCX communication device. For the cuDF application on the in-house cluster, there is an average speedup of $3.11\times$ for $2-6$ Dask workers and $3.22\times$ reduction in communication time. On the Frontera (GPU) cluster, MPI4Dask outperforms UCX by an average factor of $1.71\times$ for the cuPy application with $1-32$ Dask workers. For the cuDF application, MPI4Dask reduces the total execution time by an average factor of $2.91\times$ when compared to UCX for $1-32$ Dask workers. Reasons for better overall performance of MPI4Dask against its counterparts include better point-to-point performance of MVAPICH2-GDR and efficient coroutine implementation. Unlike UCX-Py that implements a separate coroutine to make progress for UCX worker, MPI4Dask ensures cooperative progression where every communication coroutine triggers the communication progression engine. ### I-A Contributions Main contributions of this paper are summarized below: 1. 1. Design and implementation of MPI4Dask that provides high-performance point-to- point communication coroutines for Python-based HPC applications. To the best of our knowledge, MPI4Dask is the first library that implements MPI-based Python common coroutines that work with the asyncio framework on cluster of GPUs. 2. 2. Integration of MPI4Dask with asynchronous Dask Distributed library. This is a pioneering effort that enables MPI-based communication for the Dask ecosystem. 3. 3. Demonstrate the performance benefits of using MPI4Dask compared to TCP and UCX devices using basic micro-benchmarks and two application benchmarks (based on cuPy and cuDF) using an in-house cluster comprising of V100/K80 GPUs. 4. 4. Perform scalability evaluation of MPI4Dask against TCP and UCX communication devices using two application benchmarks (based on cuPy and cuDF) on TACC’s Frontera (GPU) system with upto $32$ Dask workers on $32$ NVIDIA Quadro RTX 5000 GPUs and $256$ CPU cores. Rest of the paper is organized as follows. Section II presented relevant background followed by the design approach of the MPI4Dask library in Section III. Section IV presents implementation details. Later, we evaluate MPI4Dask using a communication micro-benchmark and two application-level benchmarks. Section VII concludes the paper. ## II Background This section covers the background for this paper, which includes the fundamentals of the MPI standard, the Dask framework, and the asyncio package. This section also presents related work. ### II-A Message Passing Interface (MPI) The Message Passing Interface (MPI) API is considered the defacto standard for writing parallel applications. The MPI API defines a set of point-to-point and collective communication routines that are provided as convenience functions to application developers. In the context of the Dask Distributed library, the most relevant set of functions are the non-blocking point-to-point communication functions like MPI_Isend() and MPI_Irecv(). Both of these functions return an MPI_Request object that can be used to invoke MPI_Test() function to check if the non-blocking communication operation has completed or not. In this paper, we make use of a GPU-aware MPI library called MVAPICH2-GDR [10] that provides optimized point-to-point and collective communication support for GPU devices. GPU-awareness here means that the MPI library is capable of directly communication data efficiently to/from GPU memory instead of staging it through the host system. The Dask ecosystem is implemented in the Python programming language. This implies that the MPI communication backend must also be implemented in Python. For this reason, we use the GPU- aware mpi4py library that provides Cython [12] wrappers to native MPI library—MVAPICH2-GDR in this case. For performance reasons, it is important that the Python wrapper library is capable of communicating data to/from GPU memory directly without incurring the overhead of serialization/de- serialization. mpi4py supports efficient exchange of GPU data stored in cuPy and cuDF format. ### II-B Dask Dask is a popular data science framework for Python programmers. It converts user application into a task-graph, which is later executed lazily on distributed hardware. This execution requires data exchange supported through implicit communication by the Dask runtime. The Dask ecosystem is a suite of Python packages. One such package is Dask Distributed that provides various components like scheduler, workers, and client. This library also supports point-to-point functionality between these Dask components. Figure 2 depicts the distributed execution model of a Dask program. There are three types of connections between Dask entities: 1) data connections (solid lines) for exchanging application data, 2) control connections (dotted lines) for exchanging heart-beat messages to maintain status of workers and detect any failures, and 3) dynamic connections (dashed lines) between workers to resolve dependencies, work-stealing, or achieving higher throughput. Currently Dask has two communication devices: 1) the TCP device that exploits the asynchronous Tornado library, 2) the UCX backend that uses UCX-Py [13]—a Cython wrapper library—on top of the native UCX [8] communication library. Figure 2: Dask Execution Model. The scheduler and workers form a Dask cluster. The client executes user program by connecting with the Dask cluster. ### II-C The asyncio Package Since version $3.5$, Python has introduced a new package called asyncio that allows writing concurrent non-blocking I/O applications. The main idea behind this package is to allow networking (or other I/O) applications to efficiently utilize CPU without unnecessarily getting blocked for long-running I/O operations. This is possible because as the Python program executes, it keeps defining tasks that are stored in a task queue and execute concurrently as soon as it is possible to execute them. These tasks are defined through coroutines, which are functions defined using the async keyword and invoked later using the await keyword. ## III Design Overview of MPI4Dask This section presents the design of the MPI4Dask library. We first discuss communication requirements of the Dask framework. This is followed by coverage of layered architecture of Dask with focus on communication devices. We begin with communication requirements of the Dask framework. 1. 1. Scalability: Provide scalability by exploiting low-latency and high-throughput for cluster of GPUs typically deployed in modern data centers. 2. 2. Coroutines: The communication backend for Dask Distributed library is asynchronous and executes within an event loop, as part of an asyncio application. Hence, the communication backend needs to support non-blocking point-to-point send and receive operations through asyncio coroutines defined using the async/await syntax. 3. 3. Elasticity: The Dask cluster (e.g., scheduler and a set of workers) is an elastic entity meaning that the workers dynamically join or leave the cluster. In this context, the communication backend needs to support such dynamic connectivity for clients and workers. Dask, however, also executes with static number of processes using the Dask-MPI library. In this paper we adopt this approach. 4. 4. Serialization/De-serialization: Dask applications support distributed processing of GPU-based buffers including cuPy and cuDF DataFrames. The communication engine should be able to support communication to/from GPU-based Python objects/data-structures supported by Dask. Figure 3 presents a layered architecture for Dask. The top layer represents the Dask framework, which includes support for various data structures and storage formats including Dask Bags, Arrays, and DataFrames. Also there is support for asynchronous execution through Delayed and Future objects. The user application is internally converted to a task graph by Dask. The next layer shows various packages that are part of the Dask ecosystem. These include Dask-MPI, Dask-CUDA, and Dask-Jobqueue. The next layer represents the Dask Distributed library. This package supports distributed computation through scheduler, worker, and client. This library also contains the distributed.comm module, shown here as “Comm Layer”. This layer provides the API that all Dask communication backends must implement. A subset of this API can be seen in Listing 1. As shown, the Dask Distributed library provides TCP and UCX-Py backends represented in this layer by tcp.py and ucx.py. MPI4Dask is implemented as part of the Dask Distributed “Comm Layer” as a multi-layered software sitting over mpi4py, which in turn exploits MVAPICH2-GDR for GPU- aware MPI communication. The bottom layer in Figure 3 represents cluster of GPUs as parallel hardware. Dask is also capable of running on shared memory system like a laptop or a desktop. Figure 3: Dask Layered Architecture with Communication Backends. Yellow boxes are designed and evaluated as a part of this paper. ⬇ 1class Comm(ABC): 2 @abstractmethod 3 def read(self, deserializers=None): 4 @abstractmethod 5 def write(self, msg, serializers=None, 6 on_error=None): 7 8class Listener(ABC): 9 @abstractmethod 10 async def start(self): 11 @abstractmethod 12 def stop(self): 13 14class Connector(ABC): 15 @abstractmethod 16 def connect(self, address, deserialize=True) Listing 1: Subset of the Communication Backend API as Mandated by the Dask Distributed Library ## IV Implementation Details of the MPI4Dask Library This section outlines the implementation details of the MPI4Dask communication library. ### IV-A Bootstrapping the Dask ecosystem and initializing MPI There are many ways to start execution of Dask programs. The manual way is to start the scheduler followed by multiple workers on the command line. After this, the client is ready to connect with the Dask cluster and execute the user program. In order to automate this process, Dask provides a number of utility “Cluster” classes. An example of this is the LocalCUDACluster from the Dask CUDA package. Once an instance of LocalCUDACluster object has been initiated, a client can connect to execute a user program. Other example of utility cluster classes include SLURMCluster, SGECluster, PBSCluster, and others. MPI4Dask currently requires that user initiates the execution of Dask program using the Dask-MPI package. Dask-MPI uses the bootstrapping mechanism provided by MPI libraries to start Dask scheduler, client, and one or more workers. We use the mpirun_rsh utility provided by MVAPICH2-GDR . Figure 2 illustrates this where the mpirun_rsh utility was used to start $4$ MPI processes. Here processes $0$ and $1$ assume the role of scheduler and client, while all the other processes—in this case $2-3$ become worker processes. This bootstrapping is done by the Dask-MPI package and this is the point where mpi4py and MVAPICH2-GDR are also initialized. The CUDA_VISIBLE_DEVICES environment variable is used to map Dask worker processes on a particular GPU in a node. This is particularly important when a node has multiple GPUs since we would want multiple Dask workers to utilize distinct GPUs on the system. Note that Dask-MPI does not provide any communication between Dask components and this is what we address in this paper. ### IV-B Point-to-point Communication Coroutines A challenge that we tackle as part of this work is to implement asynchronous communication coroutines using mpi4py over MVAPICH2-GDR that can be incorporated inside the Dask Distributed layer. To the best of our knowledge, this is first such effort to exploit MPI-based communication inside an asyncio application in Python. mpi4py provides two variants of point-to-point functions. The first option is to use the lowercase methods like Comm.send()/Comm.isend() to communicate data/to from Python objects. This involves picking/unpickling (serialization/de-serialization) of Python objects. The second option is to use methods like Comm.Send()/Comm.Isend() that start with uppercase letter and communicate data to/from directly from user-specified buffer. The second option is efficient and preferred for achieving high-performance in the Dask Distributed library. Since we are implementing MPI4Dask for GPU buffers, the specified buffer to mpi4py communication routines must support the __cuda_array_interface__. As previously mentioned, the Dask Distributed library is an asynchronous application due to which it executes coroutines in a non-blocking and concurrent manner. Also for this reason, using blocking MPI point-to-point methods will result in deadlock for the application. In order to tackle this, our implementation of MPI4Dask makes use of Comm.Isend()/Comm.Irecv() methods that return Request objects. Later, MPI4Dask checks for completion of pending communication using the Request.Test() method. Instead of checking the completion status in a busy-wait loop—that will result in a deadlock situation—MPI4Dask calls the asyncio.sleep() method that allows other coroutines to make progress while waiting for communication to complete. Listings 2 and 3 provide outlines of send and receive communication coroutines in MPI4Dask. ⬇ 1request = comm.Isend([buf, size], dest, tag) 2status = request.Test() 3 4while status is False: 5 await asyncio.sleep(0) 6 status = request.Test() Listing 2: The Comm.Isend()-based Send Communication Coroutine Implemented by MPI4Dask. ⬇ 1request = comm.Irecv([buf, size], src, tag) 2status = request.Test() 3 4while status is False: 5 await asyncio.sleep(0) 6 status = request.Test() Listing 3: The Comm.Irecv()-based Receive Communication Coroutine Implemented by MPI4Dask. ### IV-C Handling Large Messages Dask is a programming framework for writing data science applications and it is common for such libraries to handle and exchange large amounts of data. As a consequence it is possible that the higher layers of the Dask ecosystem attempt to communicate large messages using the underlying communication infrastructure. We have experienced this with MPI4Dask that relies on the point-to-point non-blocking primitives provided by MPI—Comm.Isend() and Comm.Irecv() methods. These functions accept an argument int count that is used to specify the size of the message being sent or received. The maximum value that this parameter can hold is $2^{31}-1$ bytes, which corresponds a message size of $2$ GB$-1$. On the other hand, the Dask Distributed library attempts to communicate messages larger than this value including upto $64$ GB. We have catered for this requirement by dividing the large message into several chunks of $1$ GB for the actual communication using the Comm.Isend() and Comm.Irecv() methods. Typically the buffers specified to these communication functions are Python objects and hence subscriptable using the slice notation array[start:end]. This approach works for numPy and cuPy arrays and hence the buffer argument can be subscripted in a loop to implement chunking. But this strategy does not work for cuDF and RAPIDS Memory Manager (RMM) buffers rmm.DeviceBuffer. For these, we implement chunking by incrementing the offset argument in the communication loop. ### IV-D Providing Communication Isolation After the initial bootstrapping, MPI4Dask has full connectivity between all processes. This is provided by the default communicator MPI_COMM_WORLD initialized by the MPI library. However using this communicator naively might lead to message interference. This is possible when a particular worker might want to use same tag for both data and control message exchanges. There are multiple ways to tackle this. The approach that we choose in this paper is to rely on MPI sub-communicators. Figure 4 outlines this approach. The $4$ MPI processes here engage with one another to form five sub-communicators represented by colored bi-directional edges between a pair of processes. The pseudocode for this is shown in Listing 4. Each process stores a handle to these sub-communicators in a table called comm_table. Creating a new MPI sub- communicator is a costly operation and for this reason all of this is done at the startup. The MPI.Group.Incl() and MPI.COMM_WORLD.Create() functions (from mpi4py) are used for creating sub-communicators that only contain two processes. Complexity of making new sub-communicators is $O(\frac{n^{2}}{2})$ where $n$ is the total number of MPI processes. Our approach has negligible overhead because these sub-communicators are initialized at the startup and re-used later during the program execution. Figure 4: Building new Sub-Communicators from MPI.COMM_WORLD at Startup. The bi-directional edges between processes represent newly built sub- communicators. ⬇ 1for i in range(size): 2 for j in range(i+1, size): 3 incls = [i, j] 4 new_group = MPI.Group.Incl(group, incls) 5 new_comm = MPI.COMM_WORLD.Create(new_group) 6 if rank == i: 7 comm_table.update({j : new_comm}) 8 else if rank == j: 9 comm_table.update({i : new_comm}) Listing 4: Nested Initialization Loop Executed at Each MPI Process during Startup to Build new Sub-communicators from MPI.COMM_WORLD. The newly built sub-communicators encapsulate only two processes and are stored in the comm_table. ### IV-E Handling Dynamic Connectivity The Dask Distributed library support connectivity between Dask components including scheduler, workers, and client. The UCX and TCP communication devices maintain information for remote endpoints since this information is required for the actual communication. In MPI4Dask, we have replaced the abstraction of endpoint with sub-communicator. This enables us to utilize the existing communication infrastructure of the Dask ecosystem. Apart from communications being setup at the startup, Dask allows dynamic connections between workers as well. In this context, new communication channels are only established in the Dask Distributed library through traditional client/server semantics of connect/accept. In MPI4Dask we handle this requirement by starting a server that listens for incoming connections—and invokes a connection handler callback function—using the asyncio.start_server() method. Dask scheduler and worker processes use this function since they act as listeners and other processes are allowed to connect with them. Later any process (scheduler, client, or worker) that attempts connecting to a listener process do so by using the asyncio.open_connection() function. When two processes connect with one another—and this is something that happens frequently in Dask at startup—each process gets the relevant sub-communicator from the comm_table by doing a lookup based on the destination process rank in the MPI.COMM_WORLD communicator. The corresponding sub-communicator from the comm_table is duplicated by using the Comm.Dup() function. This ensures that a new sub-communicator is built for every new dynamic connection. For performance reasons, we are maintaining a configurable cache of these sub- communicators. ## V Performance Evaluation This section presents performance evaluation of MPI4Dask against UCX and TCP (using IPoIB) communication devices using 1) Ping Pong micro-benchmark (Section V-A), 2) Two application benchmarks (cuPy and cuDF) on an in-house cluster (RI2) with two types of GPU nodes with NVIDIA Tesla V100s and NVIDIA Tesla K80s (Section V-B), and 3) Scalability results for the same two application benchmarks on TACC’s Frontera (GPU) cluster with NVIDIA Quadro RTX 5000 GPUs (Section V-C). The hardware specifications for the RI2 cluster and Frontera (GPU) subsystem are shown in Table I. The following versions of the software were used: UCX v1.8.0, UCX-Py v0.17.0, Dask Distributed v2.30, MVAPICH2-GDR v2.3.4, and mpi4py v3.0.3. TABLE I: Hardware specification of the in-house RI2 and TACC’s Frontera (GPU) clusters. Columns 1 and 2 titled RI2-V100 and RI2-K80 provide details for two types of nodes on the RI2 cluster. Column 3 titled RI2-V100 provide details for the nodes used on the Frontera (GPU) cluster. Specification | RI2-V100 | RI2-K80 | Frontra (GPU) ---|---|---|--- Number of Nodes | 16 | 16 | 90 Processor Family | Xeon Broadwell | Xeon Broadwell | Xeon Broadwell Processor Model | E5-2680 v4 | E5-2680 v4 | E5-2620 v4 Clock Speed | 2.4 GHz | 2.4 GHz | 2.1 GHz Sockets | 2 | 2 | 2 Cores Per socket | 14 | 14 | 16 RAM (DDR4) | 128 GB | 128 GB | 128 GPU Family | Tesla V100 | Tesla K80 | Quadro RTX 5000 GPUs | 1 | 2 | 4 GPU Memory | 32 GB | 12 GB | 16 GB Interconnect | IB-EDR (100G) | IB-EDR (100G) | IB-FDR (56G) ### V-A Latency and Throughput Comparison We first perform latency and throughput comparisons between MPI4Dask and other communication backends—in particular UCX-Py—using a Ping Pong benchmark. Also we add UCX (Tag API) and MVAPICH2-GDR to our comparisons for baseline performance. UCX v1.8.0 was used alongwith ucx_perftest [14] to evaluate latency and throughput for UCX Tag API. For MVAPICH2-GDR, we used the osu_latency test that is part of the OSU Micro-Benchmark Suite (OMB) [15]. UCX-Py was also evaluated using its own Ping Pong benchmark [16]. UCX-Py has two modes of execution: 1) polling-based, and 2) event-based. The polling- based mode is latency-bound and hence more efficient than the event-based mode. Comparisons also include mpi4py running over MVAPICH2-GDR—this is labeled as MV2-GDR (mpi4py). The benchmark used for mpi4py is an in-house Python version of OMB. Lastly we plot MPI4Dask that essentially represents the MPI4Dask communication device integrated into the Dask Distributed library. We always execute the TCP device using IP over InfiniBand (IPoIB) protocol that provides best performance for this backend. For this reason, we use TCP and IPoIB interchangeably in this section. (a) Latency (Small) (b) Latency (Medium) (c) Throughput (Large) Figure 5: Latency/Bandwidth comparison of MPI4Dask with UCX-Py (Polling and Event Modes) using Ping Pong Benchmark on RI2 Cluster with V100 GPUs. UCX (Tag API) and MVAPICH2-GDR numbers are also presented for baseline performance. (a) Latency (Small) (b) Latency (Medium) (c) Throughput (Large) Figure 6: Latency/Throughput comparison of MPI4Dask with UCX-Py (Polling and Event Modes) using Ping Pong Benchmark on RI2 Cluster with K80 GPUs. UCX (Tag API) and MVAPICH2-GDR numbers are also presented for baseline performance. Figure 5(a) shows latency for message sizes $1$ Byte to $4$ KByte—here MPI4Dask outperforms UCX-Py (Polling) by $5\times$. Similarly for medium-sized messages presented in Figure 5(b)—$8$ KByte to $512$ KByte—MPI4Dask is better than UCX-Py (Polling) by $2-3\times$. In throughput comparisons for large messages, $1$ MByte and beyond, MPI4Dask outperforms UCX-Py (Polling) by $3-4\times$ in throughput as shown in Figure 5(c). It is important to note that the difference in performance between MV2-GDR (mpi4py) and MV2-GDR is minimal. This means that regular Python functions over native code do not introduce much overhead. However when executing communication coroutines—by employing MPI4Dask or UCX-Py—there is additional overhead due to the asyncio framework. We observe similar performance patterns for Tesla K80 GPUs in latency and throughput comparisons between MPI4Dask and UCX-Py (Polling) as depicted in Figure 6. As noted earlier, we are focusing on performance comparison of MPI4Dask with UCX-Py (Polling) since it is the more efficient mode of the UCX-Py library. ### V-B Application Benchmarks We also evaluated MPI4Dask against UCX-Py for two application benchmarks: 1) sum of cuPy array and its transpose, and 2) cuDF merge. The cuPy benchmark presents strong scaling results as the problem size remains the same as more Dask workers are used in the computation. On the other hand, the cuDF benchmark presents weak scaling results as the problem size increases with increase in the number of Dask workers. This evaluation was done on the in- house RI2 cluster. #### V-B1 Sum of cuPy Array and its Transpose This benchmark [17] creates a cuPy array and distributes its chunks across Dask workers. The benchmark adds these distributed chunks to their transpose, forcing the GPU data to move around over the network. The following operations are performed: y = x + x.T y = y.persist() wait(y) Performance comparison graphs for the first application benchmark—sum of cuPy array with its transpose—are shown in Figure 7. Dask follows the One Process per GPU (OPG) model of execution, which means that a single worker process with multiple threads is initiated for each GPU. On the RI2 cluster, each node has a single GPU. For this reason, we instantiate a single worker process with $28$ worker threads to fully exploit the available CPU cores. Figure 7(a) shows execution time where we are witnessing an average speedup of $3.47\times$ for $2-6$ Dask workers. Figure 7(b) shows the communication time for the benchmark run. This shows that MPI4Dask is better than UCX by $6.92\times$ on average for $2-6$ workers. Average throughput comparison for Dask workers is shown in Figure 7(c), which depicts that MPI4Dask outperform UCX by $5.17\times$ on average for $2-6$ workers. This benchmark application is presenting strong scaling results. As can be seen in Figure 7(a), the cuPy application does not exhibit impressive speedups as the number of workers are increased. This is due to the nature of this benchmark that is designed to stress and evaluate communication performance. This is suitable for this paper because we are primarily interested in comparative performance of communication devices and not demonstrating application-level performance of the Dask ecosystem. #### V-B2 cuDF Merge cuDF DataFrames are table-like data-structure that are stored in the GPU memory. As part of this application [18], a merge operation is carried out for multiple cuDF data frames. Performance comparison graphs for the second application benchmark—cuDF merge operation—are shown in Figure 8. Figure 8(a) shows execution time where we are witnessing an average speedup of $3.11\times$ for $2-6$ Dask workers. Figure 8(b) shows time for communication that is taking place. This shows that MPI4Dask is better than UCX by $3.22\times$ on average for $2-6$ workers. Average throughput comparison for Dask workers is presented in Figure 8(c), which depicts that MPI4Dask outperforms UCX by $3.82\times$ on average for $2-6$ workers. The cuDF merge benchmark is presenting weak scaling results. As can be observed in Figure 8, the execution time with MPI4Dask is increasing slightly with an increase in the number of workers. This is due to efficient communication performance provided by MPI4Dask to the Dask execution. Note that $3.2$, $6.4$, $9.6$, $12.8$, $16$, and $19.2$ GB of data is processed for $1$, $2$, $3$, $4$, $5$, and $6$ Dask workers respectively. Note that UCX-Py is executed in polling-mode that is the efficient mode. Reasons for better overall performance of MPI4Dask against other counterparts include better point-to-point performance of MVAPICH2-GDR , chunking implemented for messages greater than $1$ GB, and efficient coroutine implementation for MPI4Dask as compared to UCX-Py. Unlike UCX-Py that implements a separate coroutine to make communication progress for UCX worker, MPI4Dask ensures cooperative progression where every communication coroutine triggers the communication progression engine. (a) Execution Time Comparison (b) Communication Time Comparison (c) Aggregate Throughput Between Workers Figure 7: Sum of cuPy Array and its Transpose (cuPy Dims: 16K$\times$16K, Chunk size: 4K, Partitions: 16): Performance Comparison between IPoIB, UCX, and MPI4Dask on the RI2 Cluster. This benchmark presents strong scaling results. $28$ threads are started in a single Dask worker. (a) Execution Time Comparison (b) Communication Time Comparison (c) Aggregate Throughput Between Workers Figure 8: cuDF Merge Operation Benchmark (Chunk Size: 1E8, Shuffle: True, Fraction Match: 0.3): Performance Comparison between IPoIB, UCX, and MPI4Dask on the RI2 Cluster. This benchmark presents weak scaling results. $28$ threads are started in a single Dask worker. $3.2$ GB of data is processed per Dask worker (or GPU) ### V-C Scalability Results on the TACC Frontera (GPU) Cluster This section presents the scalability results for the two application benchmarks introduced earlier in Section V-B. This evaluation was done on the Frontera (GPU) system that is equipped with $360$ NVIDIA Quadro RTX 5000 GPUs in $90$ nodes. Following the OPG model of execution, we started $4$ processes on a single node—one process per GPU—with $8$ worker threads in each process. This evaluation is presented in Figure 9. Figure 9(a) shows the execution time comparison between TCP (using IPoIB), UCX, and MPI4Dask devices for the sum of cuPy with its transpose application with $1-32$ Dask workers—here MPI4Dask outperforms UCX by an average factor of $1.71\times$. Figures 9(b) and 9(c) shows the execution time and throughput comparison between all communication devices for the cuDF merge operation. Again, MPI4Dask outperforms UCX by an average factor of $2.91\times$ for overall execution time for $1-32$ Dask workers. The merge operation throughout depicts the rate at which workers merge the cuDF input data and it also follows a similar pattern. The performance results shown in Figure 9(a) present strong scaling results for the cuPy application, while Figures 9(b) and 9(c) present weak scaling results for the cuDF merge operation. The performance shown by $1$ Dask worker in Figures 9(b) and 9(c) is efficient due to small problem size and no communication overhead. However, the problem size becomes larger and more realistic with additional Dask workers. Note that $1.6$ GB of data is processed per GPU for the cuDF application—this means that $51.2$ GB of data is processed on $32$ GPUs. (a) Execution Time Comparison for Sum of cuPy Array and its Transpose Benchmark. (cuPy Dims: 20E3$\times$20E3, Chunk size: 5E2, Partitions: 16E2) (b) Execution Time Comparison for cuDF Merge Benchmark. (Chunk Size: 5E7, Shuffle: True, Fraction Match: 0.3) (c) Average Throughput For Workers for the cuDF Merge Benchmark Figure 9: Execution Time Comparison for Sum of cuPy Array with its Transpose Benchmark, cuDF Merge Benchmark, and Average Throughput for Dask Workers for the cuDF Merge Benchmark. This evaluation was done on the Frontera (GPU) system at TACC. This comparison is between MPI4Dask , UCX, and TCP (using IPoIB) communication devices. There are $4$ GPUs in a single node. Each Dask worker has $8$ threads. ## VI Related Work A framework similar to Dask is the Apache Spark software [6]. It supports programming data science applications using Scala, Java, Python and R. Traditionally Spark has also only supported execution on hosts (CPUs). This meant that it was not able to exploit massive parallelism offered by GPUs. However GPU support has been added by the RAPIDS project recently to the Apache Spark 3.0 through a Spark-RAPIDS plugin [19]. Vanilla versions of Apache Spark only support data communication over Ethernet or other high-speed networks using the TCP backend. There have been efforts [20, 21] to address this shortcoming for Apache Spark. The Dask Distributed library currently support two communication backends. The first device—called TCP—makes use of Python’s Tornado framework [22]. The second device—called UCX—uses a Cython wrapper called UCX-Py [13] over the UCX [8] communication library. In an earlier effort [23], we have developed a new communication device called Blink for Dask. However Blink is limited to CPU- only execution of Dask programs and hence is not relevant here. Also, recently there has been another effort [24] to develop an MPI communication layer on the Dask Distributed github repository. However, this is currently a work in progress and is not addressing GPU-based data science applications, which is the focus of this paper. At this time, it is only possible to use TCP and UCX based devices for Dask on cluster of GPUs. This paper demonstrates that MPI4Dask —developed as part of this work—delivers better performance than the current state of the art communication options for Dask on cluster of GPUs. ## VII Conclusions and Future Work Python is an emerging language on the landscape of scientific computing with support for distributed computing engines like Dask. Rivaling the Apache Spark ecosystem, Dask allows incremental parallelization of Big Data applications. Recently the Dask software has been extended, as part of the NVIDIA RAPIDS framework, to support distributed computation on cluster of GPUs. Support for GPU-based data storage and processing APIs like cuPy and cuDF—GPU-counterparts of numPy and Pandas respectively—has also been added. As part of this work, we have extended the Dask Distributed library with a new communication device called MPI4Dask based on the popular MPI standard. MPI4Dask exploits the mpi4py wrapper software on top of the MVAPICH2-GDR library to offer an efficient alternative to existing backends—TCP and UCX—from within a non- blocking asynchronous I/O framework. This is done by implementing high- performance communication coroutines using MPI that support the async/await syntax. The performance evaluation done, on an in-house cluster built with Tesla V100/K80 GPUs and the Frontera (GPU) cluster equipped with Quadro RTX 5000 GPUs—using the Ping Pong micro-benchmark and two other application benchmarks—suggest that MPI4Dask clearly outperforms other communication devices. This is due to efficient implementation of communication coroutines in MPI4Dask that use the cooperative progress approach with non-blocking point-to-point MPI calls. In the future we plan to extend MPI4Dask with support for dynamic process management, which will allow entities like workers and clients to dynamically join and leave the Dask Cluster. Also we aim to extend MPI4Dask for host-based (CPU only) distributed computation with Dask. ## References * [1] G. E. Moore, “Cramming more components onto integrated circuits,” _Electronics_ , vol. 38, no. 8, April 1965. * [2] G. E. Moore, “Progress in digital integrated electronics [technical literaiture, copyright 1975 ieee. reprinted, with permission. technical digest. international electron devices meeting, ieee, 1975, pp. 11-13.],” _IEEE Solid-State Circuits Society Newsletter_ , vol. 11, no. 3, pp. 36–37, 2006. * [3] C. E. Leiserson, N. C. Thompson, J. S. Emer, B. C. Kuszmaul, B. W. Lampson, D. Sanchez, and T. B. Schardl, “There’s plenty of room at the top: What will drive computer performance after moore’s law?” _Science_ , vol. 368, no. 6495, 2020. [Online]. Available: https://science.sciencemag.org/content/368/6495/eaam9744 * [4] R. P. Feynman, “There’s Plenty of Room at the Bottom,” 1959. * [5] M. Rocklin, “Dask: Parallel computation with blocked algorithms and task scheduling,” in _Proceedings of the 14th Python in Science Conference_ , K. Huff and J. Bergstra, Eds., 2015, pp. 130 – 136. * [6] M. Zaharia, R. S. Xin, P. Wendell, T. Das, M. Armbrust, A. Dave, X. Meng, J. Rosen, S. Venkataraman, M. J. Franklin _et al._ , “Apache Spark: A Unified Engine for Big Data Processing,” _Communications of the ACM_ , vol. 59, no. 11, pp. 56–65, 2016. * [7] S. Raschka, J. T. Patterson, and C. Nolet, “Machine learning in python: Main developments and technology trends in data science, machine learning, and artificial intelligence,” _ArXiv_ , vol. abs/2002.04803, 2020. * [8] P. Shamis, M. G. Venkata, M. G. Lopez, M. B. Baker, O. Hernandez, Y. Itigin, M. Dubman, G. Shainer, R. L. Graham, L. Liss, Y. Shahar, S. Potluri, D. Rossetti, D. Becker, D. Poole, C. Lamb, S. Kumar, C. Stunkel, G. Bosilca, and A. Bouteiller, “UCX: An Open Source Framework for HPC Network APIs and Beyond,” in _2015 IEEE 23rd Annual Symposium on High-Performance Interconnects_ , 2015, pp. 40–43. * [9] “Message Passing Interface (MPI),” http://www.mpi-forum.org. Accessed: . * [10] MVAPICH2-GDR Development Team, “MVAPICH2-GDR User Guide,” http://mvapich.cse.ohio-state.edu/userguide/gdr. * [11] L. D. Dalcin, R. R. Paz, P. A. Kler, and A. Cosimo, “Parallel distributed computing using python,” _Advances in Water Resources_ , vol. 34, no. 9, pp. 1124 – 1139, 2011, new Computational Methods and Software Tools. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0309170811000777 * [12] S. Behnel, R. Bradshaw, C. Citro, L. Dalcin, D. S. Seljebotn, and K. Smith, “Cython: The best of both worlds,” _Computing in Science & Engineering_, vol. 13, no. 2, pp. 31–39, 2011. * [13] UCX, “UCX-Py Github,” https://github.com/rapidsai/ucx-py. * [14] ——, “Performance Measurement,” https://github.com/openucx/ucx/wiki/Performance-measurement. * [15] Network-Based Computing Laboratory, “OSU Micro-Benchmarks 5.6.3,” http://mvapich.cse.ohio-state.edu/benchmarks. * [16] UCX-Py, “Benchmarks,” https://github.com/rapidsai/ucx-py/blob/branch-0.15/benchmarks/local-send-recv.py. * [17] NVIDIA RAPIDS Team, “Sum of cuPy Array and its Transpose,” https://github.com/rapidsai/dask-cuda/blob/branch-0.16/dask_cuda/benchmarks/local_cupy_transpose_sum.py. * [18] ——, “cuDF Merge,” https://github.com/rapidsai/dask-cuda/blob/branch-0.16/dask_cuda/benchmarks/local_cudf_merge.py. * [19] NVIDIA, “RAPIDS Accelerator For Apache Spark,” https://github.com/NVIDIA/spark-rapids. * [20] X. Lu, D. Shankar, S. Gugnani, and D. K. Panda, “High-performance design of apache spark with RDMA and its benefits on various workloads,” in _2016 IEEE International Conference on Big Data (Big Data)_ , 2016, pp. 253–262. * [21] M. Inc., _SparkRDMA_ , 2020 (accessed July 15, 2020). [Online]. Available: https://github.com/Mellanox/SparkRDMA * [22] Tornado Web Server, “Dask TCP Backend,” https://www.tornadoweb.org. * [23] A. Shafi, J. Hashmi, H. Subramoni, and D. Panda, “ Blink: Towards Efficient RDMA-based Communication Coroutines for Parallel Python Applications ,” December 2020. * [24] Ian Thomas, “[WIP]: MPI communication layer,” https://github.com/dask/distributed/pull/4142.
††thanks: This work was supported by the National Natural Science Foundation of China (Nos. 11805294 and 11975021), the China Postdoctoral Science Foundation (2018M631013), the Strategic Priority Research Program of Chinese Academy of Sciences (XDA10010900), the Fundamental Research Funds for the Central Universities, Sun Yat-sen University (19lgpy268), and in part by the CAS Center for Excellence in Particle Physics (CCEPP). # Event vertex and time reconstruction in large-volume liquid scintillator detectors Zi-Yuan Li School of Physics, Sun Yat-Sen University, Guangzhou 510275, China Yu-Mei Zhang<EMAIL_ADDRESS>Sino-French Institute of Nuclear Engineering and Technology, Sun Yat-Sen University, Zhuhai 519082, China Zhen Qian School of Physics, Sun Yat-Sen University, Guangzhou 510275, China Shu Zhang School of Physics, Sun Yat-Sen University, Guangzhou 510275, China Kai-Xuan Huang School of Physics, Sun Yat-Sen University, Guangzhou 510275, China Guo-Fu Cao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China Zi-Yan Deng Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China Gui-Hong Huang Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China Wei-Dong Li Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China Tao Lin Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China Liang-Jian Wen Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China Miao Yu School of Physics and Technology, Wuhan University, Wuhan 430072, China Jia-Heng Zou Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China Wu-Ming Luo<EMAIL_ADDRESS>Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China Zheng-Yun You<EMAIL_ADDRESS>School of Physics, Sun Yat-Sen University, Guangzhou 510275, China ###### Abstract Large-volume liquid scintillator detectors with ultra-low background levels have been widely used to study neutrino physics and search for dark matter. Event vertex and event time are not only useful for event selection but also essential for the reconstruction of event energy. In this study, four event vertex and event time reconstruction algorithms using charge and time information collected by photomultiplier tubes were analyzed comprehensively. The effects of photomultiplier tube properties were also investigated. The results indicate that the transit time spread is the main effect degrading the vertex reconstruction, while the effect of dark noise is limited. In addition, when the event is close to the detector boundary, the charge information provides better performance for vertex reconstruction than the time information. JUNO, Liquid scintillator detector, Neutrino experiment, Vertex reconstruction, Time reconstruction ## I Introduction Liquid scintillators (LSs) have widely been used as detection medium for neutrinos in experiments such as Kamioka Liquid Scintillator Antineutrino Detector (KamLAND) kamland , Borexino borexino , Double Chooz doublechooz , Daya Bay dyb , and Reactor Experiment for Neutrino Oscillation (RENO) reno . KamLAND revealed a large mixing angle (LMA) solution for solar neutrino oscillations. Borexino confirmed the Mikheyev-–Smirnov-–Wolfenstein (MSW) LMA lmamsw model in the sub-MeV region for solar neutrino oscillations. Double Chooz, Daya Bay, and RENO reported nonzero measurements for the mixing angle $\theta_{13}$. The size of such detectors varies from hundreds to thousands of cubic meters. Large-volume liquid scintillator detectors are widely used in the next generation of neutrino experiments, aiming to solve problems such as neutrinoless double-beta decay (SNO+ at the Sudbury Neutrino Observatory sno+ ) and neutrino mass ordering (Jiangmen Underground Neutrino Observatory (JUNO) junophysics ). The sensitivity of these experiments is limited by the energy resolution, detector volume, and detector background. These detectors typically contain a fiducial volume, where the signal-to-noise ratio is maximal. To distinguish between events occurring in the fiducial and non-fiducial regions, the event vertex is reconstructed using the charge and time distribution of photons collected by the photomultiplier tubes (PMTs). Most importantly, to achieve a high energy resolution, an accurate vertex is essential to correct for energy non-uniformity. In addition, the event vertex and event time information can also be used for particle identification, direction reconstruction, event classification, and other purposes. A previous study Liu_2018 investigated vertex reconstruction with time information in JUNO, without discussing the event time reconstruction, dark noise effect, and the improvement based on the charge information when the event is close to the detector boundary. The aim of this study was to analyze vertex and time reconstruction for point-like events in JUNO under more realistic conditions. The main contributions of this study are as follows: 1. $*$ The event time was reconstructed to provide the start time information of the event, which was important for event alignment, event correlation, etc. 2. $*$ The dark noise from PMTs was considered and its effect on the vertex reconstruction was properly controlled. 3. $*$ Two types of large PMTs were considered and handled separately mainly because of the difference in transit time spread (TTS). 4. $*$ An algorithm to provide a more accurate initial vertex value was developed to improve the performance, especially at the detector boundary region. 5. $*$ An algorithm employing charge information to reconstruct the event vertex at the detector boundary was developed. The remainder of this paper is organized as follows. In Sec. II, a brief introduction of the JUNO detector and the configurations of the PMTs used in this study is provided. In Sec. III, the optical processes are described and a simple optical model is introduced. In Sec. IV, two simple algorithms that can quickly provide initial values are compared. In Secs. V and VI, two complex algorithms that provide relatively good performance are introduced in detail. Finally, a performance summary, the discussion, and the conclusions are provided in Secs. VII, VIII, and IX, respectively. ## II The JUNO detector Figure 1: Schematic of the JUNO detector. A schematic of the JUNO detector is shown in Fig. 1. The central detector (CD) is the main part of the JUNO detector assembly, with an acrylic sphere with a diameter of 35.4 m as inner layer. The acrylic sphere is supported by a stainless steel latticed shell with a diameter of 40.1 m and filled with approximately 20,000 tons of LS as target for neutrino detection. The composition of the LS includes linear alkylbenzene (LAB) as solvent, 2,5-diphenyloxazole (PPO) as fluor, and p-bis-(o-methylstyryl)-benzene (bis- MSB) as wavelength shifter. To collect photons, the CD is surrounded by approximately 17,600 20-inch PMTs, including 5000 Hamamatsu R12860 PMTs, 12,600 North Night Vision Technology (NNVT) GDG-6201 PMTs, and approximately 25,600 3-inch XP72B22 PMTs. Around the CD, a 2.5-m-thick water pool is used to shield external radioactivity from the surrounding rocks and is combined with approximately 2,000 20-inch GDG-6201 PMTs to serve as a water Cherenkov detector to veto cosmic ray muons. A top tracker detector, consisting of plastic scintillators, located above the water pool is used for the identification and veto of muon tracks. A more detailed description of the JUNO detector can be found in Refs. junocdr ; junophysics . The main factors affecting the reconstruction of the event vertex and time include the TTS and dark noise of the PMTs. In this study, only the 20-inch PMTs of the CD were used for reconstruction. The number and parameters of the Hamamatsu and NNVT PMTs are summarized in Table 1 pmtsys . In principle, the 3-inch PMTs with $\sigma\simeq 1.6$ ns TTS could also be included to improve the reconstruction performance; however, these PMTs 3inchpmt were not considered in this study because of their small photon detection coverage ($\sim$3%) gdml ; gdml2 . Table 1: Number and parameters of the PMTs used in the reconstruction. The TTS and dark noise rate are the mean values of the distribution measured during the mass testing. However, these are not the final values for JUNO. Company | Number | TTS ($\sigma$) | Dark noise rate ---|---|---|--- Hamamatsu | 5,000 | 1.15 ns | 15 kHz NNVT | 12,600 | 7.65 ns | 32 kHz ## III Optical processes When a charged particle deposits energy in the scintillator, the solvent enters an excited state and transfers energy to the fluor in a non-radiative manner. Scintillation photons are then emitted along the particle track through the radiative de-excitation of the excited fluor within a limited time. The emitted scintillation photons can undergo several different processes while propagating through a large LS detector. At short wavelengths ($<$ 410 nm), photons are mostly absorbed and then re-emitted at longer wavelengths, which maximizes the detection efficiency of the PMTs. At long wavelengths ($>$ 410 nm), photons mainly undergo Rayleigh scattering. A more detailed study of the wavelength-dependent absorption and re-emission can be found in Ref. lsopmdl . Additionally, the refractive indices at 420 nm are 1.50 and 1.34 for the LS and water, respectively. The difference in the refractive indices results in refraction and total reflection at the boundary of the two media, which affects the time-of-flight of the photons. When using the time information in the reconstruction, the time-of-flight is crucial, and it is calculated using the equation $\mathrm{tof}=\sum_{m}\frac{d_{m}}{v_{m}},$ (1) where $\mathrm{tof}$, $d_{m}$, and $v_{m}$ are the time-of-flight, optical path length, and effective light speed, respectively, and $m$ represents different media, in this case the LS and water, in the JUNO experiment. The acrylic sphere (thickness of 12 cm) and acrylic cover (thickness of 1 cm) in front of each PMT were ignored in this study because their refractive indices were similar to that of the LS and their thickness was small compared to that of the LS and water. ### III.1 Optical path length Figure 2: Event display of the optical path from the event vertex to the PMT in the JUNO simulation. The red circle ring is the event vertex and the gray bulbs with blue caps represent the PMTs. The optical path length can be characterized by the start and end positions in the detector, which are the vertex of the event and the position of the PMTs, respectively. The bold cyan curve in Fig. 2 shows a typical example of the optical path of photons detected by the PMT in the JUNO simulation using the event display eventdisplay1 ; eventdisplay2 , and further examples are shown by the thin green curves. There are multiple physically possible paths between these two positions, each of which has a different optical path length, as follows: 1. $*$ owing to absorption and re-emission, the re-emitted photon is not in the same absorption position and the propagation direction also changes; 2. $*$ owing to scattering, the photon changes the original direction of the propagation; and 3. $*$ owing to refraction and total reflection, the photon does not travel in a straight line. As shown in Fig. 2, owing to the various aforementioned optical processes, it is difficult to predict the actual optical path length for each photon. In this paper, a simple optical model is proposed, which uses a straight line connecting the vertex and the PMTs to calculate the optical path length (Fig. 3), and combines with the effective light speed to correct for the time-of- flight. Using this simple optical model reasonable results can be obtained, as discussed in Sec. V. Figure 3: Optical path length from the event vertex to the $i$th PMT. $O$ denotes the center of the detector. In Fig. 3, {$\vec{r}_{0},t_{0}$} represents the event vertex and start time, {$\vec{r}_{i},t_{i}$} is the position of the $i$th PMT and the time of the earliest arriving photon detected by it. The angle between the normal direction of the $i$th PMT and the vector of the position of the $i$th PMT pointing to the event vertex is $\theta_{i}$ and $\alpha_{i}=\arccos(\hat{\vec{r}}_{0}\cdot\hat{\vec{r}}_{i})$. The optical path length of the photon arriving at the $i$th PMT is $d_{pathlength,i}=|\vec{r}_{i}-\vec{r}_{0}|=d_{LS,i}+d_{water,i}$ and the corresponding time-of-flight is $\mathrm{tof}_{i}$. The optical path length in the LS and water can be calculated by simply solving the trigonometric equation. ### III.2 Effective light speed According to Ref. lsopmdl , the emission spectrum of scintillation photons is in the range of approximately 300––600 nm. Typically, the group velocity of the wave packet is used to describe the photon propagation in the medium, which is given by the equation $v_{g}(\lambda)=\frac{c}{n(\lambda)-\lambda\frac{\partial{n(\lambda)}}{\partial\lambda}},$ (2) where $v_{g}$ is the group velocity, $c$ is the speed of light in vacuum, $n$ is the refractive index, and $\lambda$ is the wavelength. By fitting the Sellmeier equation optics , which describes the dispersion of the measurement in Refs. rayleighscat and h2oindex , the refractive index of the LS and water at different wavelengths is shown in the upper panel of Fig. 4. The group velocity of the LS and water can be calculated using Eq. 2 at different wavelengths, as shown in the lower panel of Fig. 4. Figure 4: Dependence of the refractive index (upper panel) and group velocity (lower panel) on the wavelength in the LS and water. The propagation speed of photons in water ($v_{water}$) was determined as the average speed weighted by the probability density function of the photon wavelength, which was obtained from a Monte Carlo (MC) simulation. As the absorption and re-emission change the initial wavelength, determining the propagation speed of photons in the LS ($v_{LS}$) is more complicated. To consider all wavelength-dependent effects that affect the propagation speed of photons, the effective light speed $v_{eff}$ is introduced. In addition, $v_{eff}$ also mitigates the effects by the simplified optical model, which, for example, ignores the refraction at the interface between the LS and water, as well as the change in the optical path length due to Rayleigh scattering. The exact value for $v_{eff}$ can be determined using a data-driven method based on the calibration data as follows: place $\gamma$ sources along the Z-axis, use $v_{LS}$ at 420 nm as the initial value of $v_{eff}$ in the reconstruction algorithm and then, calibrate $v_{eff}$ such that the source positions can be appropriately reconstructed. As no calibration data was available for JUNO, in this study, simulated calibration data were used, and the optimized values for the effective refractive index (c/$v_{eff}$) were 1.546 in the LS and 1.373 in water. In the future, the same method can be applied to the experimental calibration data. ## IV Initial value for vertex and time The TMinuit package tminuit was used for the minimization procedure in the time likelihood and in the charge likelihood algorithm introduced in Secs. V and VI. When there are multiple local minima in the parameter space, an inaccurate initial value results in local instead of global minima, resulting in a lower reconstruction efficiency. For detectors such as JUNO, the initial value needs to be treated carefully because of the total reflection, as discussed in the following subsections. ### IV.1 Charge-based algorithm The charge-based algorithm is essentially based on the charge-weighted average of the positions of the PMTs in an event, and the event vertex can be determined using the equation $\vec{r}_{0}=a\cdot\frac{\sum_{i}{q_{i}\cdot\vec{r}_{i}}}{\sum_{i}{q_{i}}},$ (3) where $q_{i}$ is the charge of the pulses detected by the $i$th PMT and $\vec{r}_{0}$ and $\vec{r}_{i}$ are defined in Fig. 3. A scale factor $a$ is introduced because the charge-based algorithm is inherently biased and an ideal point-like event in a spherical detector is covered by a uniform photocathode. Even if all propagation-related effects, such as absorption and scattering are ignored, the result of a simple integral of the intersections of all photons with the sphere surface shows that the reconstructed position of the event is 2/3 of the true position. The value of $a$ can be tuned based on the calibration data along the $Z$-axis. In this study, $a=1.3$ was used, which was sufficient to provide an initial estimate for the event vertex. Figure 5: Heatmap of $R_{rec}$ (upper panel) and $R_{rec}-R_{true}$ (lower panel) as a function of $R_{true}$ for 4-MeV $e^{+}$ uniformly distributed in space calculated by the charge-based algorithm. As can be seen in Fig. 5, even with the scale factor, owing to total reflection, the reconstructed vertex deviates up to 3 m near the detector boundary. According to Ref. optics , total reflection occurs only when the event vertex is located at an $R$ larger than $R_{LS}\cdot{n_{water}/n_{LS}}\approx 15.9$ m, where $R_{LS}$ is the radius of the acrylic sphere, $n_{LS}$ and $n_{water}$ are the refractive indices in the LS and water, respectively. The total reflection region is defined as $R>15.9$ m while $R<15.9$ m is the central region. If the result from the charge-based algorithm is used as the initial value for the time likelihood algorithm, approximately 18% of events is reconstructed at a local minimum position. In addition, it should be noted that the charge-based algorithm is not able to provide an initial value for the event generation time $t_{0}$. Therefore, a fast time-based algorithm needs to be introduced, which can provide more accurate initial values. ### IV.2 Time-based algorithm The time-based algorithm uses the distribution of the time-of-flight correction time $\Delta{t}$ (defined in Eq. 4) of an event to reconstruct its vertex and $t_{0}$. In practice, the algorithm finds the reconstructed vertex and $t_{0}$ using the following iterations: 1. 1. Apply the charge-based algorithm to obtain the initial vertex. 2. 2. Calculate time-of-flight correction time $\Delta{t}$ for the $i$th PMT as $\Delta{t}_{i}(j)=t_{i}-\mathrm{tof}_{i}(j),$ (4) where $j$ is the iteration step and $t_{i}$, $\mathrm{tof}_{i}$ are defined in Fig. 3. Plot the $\Delta{t}$ distribution for all triggered PMTs, and label the peak position as $\Delta{t}^{peak}$. 3. 3. Calculate the correction vector $\vec{\delta}[\vec{r}(j)]$ as $\vec{\delta}[\vec{r}(j)]=\frac{\sum_{i}(\frac{\Delta{t}_{i}(j)-\Delta{t}^{peak}(j)}{\mathrm{tof}_{i}(j)})\cdot(\vec{r}_{0}(j)-\vec{r}_{i})}{N^{peak}(j)},$ (5) where $\vec{r}_{0}$, and $\vec{r}_{i}$ are defined in Fig. 3. To minimize the effect of scattering, reflection, and dark noise on the bias of the reconstructed vertex, only the pulses appearing in the $(-10\leavevmode\nobreak\ \rm ns,+5\leavevmode\nobreak\ \rm ns)$ window around $\Delta{t}^{peak}$ are included. The time cut also suppresses the effect of the late scintillation photons. The number of triggered PMTs in the window is $N^{peak}$. 4. 4. If $\vec{\delta}[\vec{r}(j)]<1\leavevmode\nobreak\ \rm mm$ or $j=100$, stop the iteration; otherwise, update the vertex with $\vec{r}_{0}(j+1)=\vec{r}_{0}(j)+\vec{\delta}[\vec{r}(j)]$ and go to step 2 to start a new round of iteration. The distribution of $\Delta{t}$ at different iteration steps is shown in Fig. 6. At the beginning of the iteration, the $\Delta{t}$ distribution is wide because the initial vertex is far from the true vertex. As the number of iterations increases, the $\Delta{t}$ distribution becomes more concentrated. Finally, when the requirement in step 4 is satisfied, the iteration stops. In the final step, $\vec{r}_{0}$ is the reconstructed vertex and $\Delta{t}^{peak}$ is the reconstructed time $t_{0}$. Figure 6: $\Delta{t}$ distribution at different iteration steps $j$. After the time-of-flight correction, the $\Delta{t}$ distribution is independent of the event vertex. However, because the earliest arrival time is used, according to the first-order statistic, as discussed in Ref. fos ; fos2 ; fos3 , $t_{i}$ is related to the number of photoelectrons $N^{i}_{pe}$ detected by $i$th PMT. To reduce the bias of the vertex reconstruction, the following form of the time–$N_{pe}$ correction is applied, and in Eq. 4 $t_{i}$ is replaced by $t^{\prime}_{i}$: $t^{\prime}_{i}=t_{i}-p0/\sqrt{N^{i}_{pe}}-p1-p2/N^{i}_{pe}.$ (6) The parameters $(p0,p1,p2)$ with the corresponding values of (9.42, 0.74, $-$4.60) for Hamamatsu PMTs and (41.31, $-$12.04, $-$20.02) for NNVT PMTs were found to minimize the bias and energy dependence of the reconstruction in this study. The difference in the parameters is mainly due to the difference in the TTSs of the PMTs. Following the correction, the times of different PMTs with different values of $N_{pe}$ are aligned. Figure 7: Heatmap of $R_{rec}$ (upper panel) and $R_{rec}-R_{true}$ (lower panel) as a function of $R_{true}$ for 4-MeV $e^{+}$ uniformly distributed in space calculated by the time-based algorithm. As shown in Fig. 7, the time-based algorithm provided a more accurate reconstructed vertex than the charge-based algorithm (Fig. 5). In addition, after the time–$N_{pe}$ correction, the reconstruction shows no obvious bias within the entire detector, even in the total reflection region. The reconstructed result was used as the initial value for the time likelihood algorithm. ## V Time likelihood algorithm ### V.1 Principle of the algorithm The time likelihood algorithm uses the scintillator response function to reconstruct the event vertex. The variable residual time $t_{res}(\vec{r}_{0},t_{0})$ for the $i$th PMT can be described as $t^{i}_{res}(\vec{r}_{0},t_{0})=t_{i}-\mathrm{tof}_{i}-t_{0},$ (7) where $t^{i}_{res}$ is the residual time of the $i$th PMT and $\vec{r}_{0}$, $t_{0}$, $t_{i}$, and $\mathrm{tof}_{i}$ are defined in Fig. 3. The scintillator response function mainly consists of the emission time profile of the scintillation photons and the TTS and the dark noise of PMTs. In principle, the additional delays introduced by the absorption, re-emission, scattering, and total reflection of the photon arriving to the PMT depend on the distance between the emission position and the individual PMTs. However, the differences are only noticeable for the late arrival hits, which are largely suppressed by the requirement for the earliest arriving photons in the time likelihood algorithm. Therefore, in the first-order approximation, the scintillator response function can be considered to be the same for all positions inside the scintillator. The scintillator response function can be described as follows. As described in Sec. III, when a charged particle interacts with a scintillator molecule, the molecule is excited, then de-excites, and emits photons. Typically, the scintillator has more than one component; thus, the emission time profile of the scintillation photons, $f(t_{res})$, can be described as $f(t_{res})=\sum_{k}\frac{\rho_{k}}{\tau_{k}}e^{\frac{-t_{res}}{\tau_{k}}},\sum_{k}\rho_{k}=1,$ (8) where each ${k}$ component is characterized by its decay time $\tau_{k}$ and intensity $\rho_{k}$. The different components result from the different excited states of the scintillator molecules. To consider the spread in the arrival time of photons at the PMTs, a convolution with a Gaussian function is applied, given by $g(t_{res})=\frac{1}{\sqrt{2\pi}\sigma}e^{-\frac{(t_{res}-\nu)^{2}}{2\sigma^{2}}}\cdot{f(t_{res})}.$ (9) where $\sigma$ is the TTS of PMTs and $\nu$ is the average transit time. The dark noise, which occurs without incident photons in the PMTs, is not correlated with any physical event. The fraction of the dark noise in the total number of photoelectrons $\varepsilon_{dn}$ can be calculated based on the data acquisition (DAQ) windows, dark noise rate, and light yield of the LS. The probability of dark noise $\varepsilon(t_{res})$ is constant over time, where $\int_{DAQ}{\varepsilon(t_{res})dt_{res}}=\varepsilon_{dn}$. By adding $\varepsilon(t_{res})$ to $g(t_{res})$ and renormalizing its integral to 1, the probability density function (PDF) of the scintillator response function can be written as $p(t_{res})=(1-\varepsilon_{dn})\cdot{g(t_{res})}+\varepsilon(t_{res}).$ (10) The distribution of the residual time $t_{res}$ of an event for a hypothetical vertex can be compared with $p(t_{res})$. The best-fitting vertex and $t_{0}$ are chosen by minimizing the negative log-likelihood function $\mathcal{L}{(\vec{r}_{0},t_{0})}=-\ln(\prod_{i}p(t^{i}_{res})).$ (11) The parameters in Eq. 10 can be measured experimentally tres1 ; tres2 ; tres3 ; tres4 . In this work, the PDF from the MC simulation for the methodology study was employed. ### V.2 Probability density function Figure 8: PDF of the scintillator response function for PMTs detecting different numbers of photoelectrons. The upper panel shows the response function for Hamamatsu, the lower panel for the NNVT PMTs. Figure 9: Bias of the reconstructed $R$ (left panel), $\theta$ (middle panel), and $\phi$ (right panel) for different energies calculated by the time likelihood algorithm. Figure 10: Resolution of the reconstructed $R$ (left panel), $\theta$ (middle panel), and $\phi$ (right panel) as a function of energy calculated by the time likelihood algorithm. Different colors represent different PMT configurations. The PDF of the scintillator response function for PMTs detecting a single photoelectron was obtained from the MC simulation, using a 4.4-MeV $\gamma$ source located at the center of the detector, such that the distance to all PMTs is the same. For PMTs detecting multiple photoelectrons, the time of the earliest arriving photon is biased toward an earlier time. Therefore, the PDF need to be modified according to the first-order statistic of $p(t_{res})$ or the so-called first photoelectron timing technique fos ; fos2 ; fos3 as $p_{N_{pe}}(t_{res})=N_{pe}p(t_{res})(\int_{t_{res}}^{\infty}p(x)dx)^{N_{pe}-1},$ (12) where $p_{N_{pe}}(t_{res})$ is the PDF of the scintillator response function when the PMTs detect $N_{pe}$ hits. The PDF of two kinds of PMTs is shown in Fig. 8: the upper panel is for Hamamatsu while the lower panel is for NNVT PMTs. As the PDF is affected by the time resolution of the PMTs, the PDF of the NNVT is wider because of its inferior TTS. The inset in the lower panel shows the PDF on a logarithmic scale, and the time constant contribution of the dark noise $\varepsilon(t_{res})$ is clearly visible. ### V.3 Reconstruction performance The reconstructed vertex was compared with the true vertex in spherical coordinates ($R,\theta,\phi$) for the MC $e^{+}$ samples and fitted with a Gaussian function to analyze the bias and resolution. The bias of the reconstruction is shown in Fig. 9, where different colors represent events with different energies. As can be seen in the left panel of Fig. 9, the reconstructed $R$ is consistent with the true value in the central region, while an energy-dependent bias behavior is noticeable near the detector boundary. Given its regular bias behavior, the bias can be corrected with an energy-dependent correction. Moreover, although the reconstructed $R$ is biased, there is no bias in $\theta$ and $\phi$, as shown in the middle and right panels of Fig. 9, respectively. The spatial resolution of the vertex reconstruction as a function of energy is shown in Fig. 10. The $R$ bias was corrected before the analysis of the resolution. To study the individual effect of the TTS and dark noise on the vertex reconstruction, different MC samples were produced with and without these effects. The vertex reconstruction results are shown in Fig. 10. The magenta circles represent the default PMT configuration, as described in Sec. II. The red triangles represent an ideal configuration, which assumes perfect PMTs without the effects of the TTS and dark noise. The black squares represent the configuration of PMTs including only the dark noise effect, while the blue inverted triangles represent the PMT configuration including only the TTS effect. The exact values of the vertex resolution at 1.022 MeV and 10.022 MeV are summarized in Tables 2 and 3, respectively. The energy $E_{true}$ includes the energy of the annihilation gamma rays. The light yield was approximately 1300 detected $N_{pe}$ per 1 MeV of deposited energy in JUNO, and the energy nonlinearities on the light yield were ignored in the approximation. As can be seen in Tables 2 and 3, the dark noise has no effect at high energy and its effect at low energy is also highly limited. The largest effect results from the TTS in the time likelihood algorithm. The energy-dependent vertex resolution is approximately proportional to $1/\sqrt{N_{pe}}$ fos2 . Table 2: Vertex resolution for different PMT configurations at 1.022 MeV (detection of $\sim$1328 $N_{pe}$ in total, corresponding to $\sim$370 $N_{pe}$ detected by Hamamatsu PMTs). PMT configuration | $R$ (mm) | $\theta$ (degrees) | $\phi$ (degrees) ---|---|---|--- Ideal | 60 | 0.25 | 0.31 With dark noise only | 62 | 0.27 | 0.34 With TTS only | 89 | 0.37 | 0.44 With TTS and dark noise | 103 | 0.40 | 0.47 With TTS and dark noise (Hamamatsu PMTs only) | 105 | 0.42 | 0.49 Table 3: Vertex resolution for different PMT configurations at 10.022 MeV (detection of $\sim$13280 $N_{pe}$ in total, corresponding to $\sim$ 3700 $N_{pe}$ detected by Hamamatsu PMTs). PMT configuration | $R$ (mm) | $\theta$ (degrees) | $\phi$ (degrees) ---|---|---|--- Ideal | 19 | 0.08 | 0.11 With dark noise only | 19 | 0.08 | 0.11 With TTS only | 31 | 0.13 | 0.16 With TTS and dark noise | 31 | 0.13 | 0.16 With TTS and dark noise (Hamamatsu PMTs only) | 32 | 0.14 | 0.17 Owing to the low time resolution of the NNVT PMTs, in Fig. 10 only the reconstruction using Hamamatsu PMTs is shown (green circles). In this study, we found that the vertex resolution with Hamamatsu PMTs was similar to that of using all PMTs. The reconstruction speed was 3.5 times faster, because the fraction of the Hamamatsu PMTs was approximately 28% of all PMTs in the CD. Figure 11: Reconstructed event time $t_{0}$ at different energies. The reconstructed event time $t_{0}$ is shown in Fig. 11. The effect of $t_{0}$ is essentially a global shift of an event to match the scintillator response function PDF; in reality, $t_{0}$ is also affected by the trigger time and the time delay from the cable. The absolute value of $t_{0}$ can be neglected; only the relative difference of different events is important for the alignment of events. The small bump near $-$1.6 ns is correlated with the $R$ bias, and the long tail on the right side results from positronium formation. The variation in the reconstructed $t_{0}$ is within a few nanoseconds. ## VI Total reflection region calculated by the charge likelihood algorithm The time likelihood method described in Sec. V introduces a bias in the $R$ direction when the reconstructing events are close to the acrylic sphere. As mentioned in Ref. Smirnov_2003 , using a charge signal with the maximum likelihood method can provide better spatial resolution than the time likelihood algorithm when an event occurs near the detector boundary. In this section, we discuss the charge likelihood algorithm to reconstruct the event vertex in the total reflection region only, while the reconstruction result in the central region is omitted. The charge likelihood algorithm is based on the distribution of the number of photoelectrons in each PMT. With the mean expected number of photoelectrons $\mu(\vec{r_{0}},E)$ detected by each PMT at a given vertex and energy, the probability of observing $N_{pe}$ on a PMT follows a Poisson distribution. Furthermore, 1. $*$ Probability for the $j$th PMT with no hits: $P_{nohit}^{j}(\vec{r_{0}},E)=e^{-\mu_{j}}$, 2. $*$ Probability for the $i$th PMT with $N^{i}_{pe}$ hits: $P_{hit}^{i}(\vec{r_{0}},E)=\frac{\mu_{i}^{N^{i}_{pe}}e^{-\mu_{i}}}{N^{i}_{pe}!}$. Therefore, the probability of observing a hit pattern for an event can be written as $p(\vec{r_{0}},E)=\prod_{j}{P_{nohit}^{j}(\vec{r_{0}},E)}\cdot{\prod_{i}{P_{hit}^{i}(\vec{r_{0}},E)}}.$ (13) The best-fit values of $\vec{r_{0}}$ and $E$ can be obtained by minimizing the negative log-likelihood function $\mathcal{L}{(\vec{r_{0}},E)}=-\ln(p(\vec{r_{0}},E)).$ (14) In principle, $\mu(\vec{r_{0}},E)$ can be expressed by the equation $\mu_{i}(\vec{r_{0}},E)=Y\cdot\frac{\Omega(\vec{r_{0}},r_{i})}{4\pi}\cdot\varepsilon_{i}\cdot{f(\theta_{i})}\cdot{e^{-\sum_{m}{\frac{d_{m}}{\zeta_{m}}}}}\cdot{E}+\delta_{i},$ (15) where $Y$ is the energy scale factor, $\Omega(\vec{r_{0}},r_{i})$ is the solid angle of the $i$th PMT, $\varepsilon_{i}$ is the detection efficiency of the $i$th PMT, $f(\theta_{i})$ is the angular response of the $i$th PMT, $\theta_{i}$ is defined in Fig. 3, $\zeta_{m}$ is the attenuation length attleng in materials, and $\delta_{i}$ is the expected number of dark noise. This equation is based on the assumption that the scintillation light yield is linearly proportional to the energy. Figure 12: Mean expected number of photoelectron distribution as a function of radius $R$ and angle $\alpha$. This map is obtained by placing gamma sources at 29 specific positions along the Z-axis, which can be performed using a calibration procedure Wu_2019 . However, Eq. 15 cannot describe properly the contribution of the indirect light, the effect of light shadows because of the geometric structure, and the effect of the total reflection. Another solution is to use the model- independent method described in Ref. Wu_2019 : the mean expected number of photoelectrons can be obtained by placing gamma sources at 29 specific positions along the Z-axis, which can be performed using a calibration procedure calibsys . In this study, different from Ref. Wu_2019 , we focused on the performance of the vertex reconstruction. The mean expected number of the photoelectron distributions as a function of radius $R$ and angle $\alpha$ is shown in Fig. 12, and the definition of angle $\alpha$ is shown in Fig. 3. The mean expected number of photoelectrons $\mu$ obtained from Fig. 12 was used to calculate the hit probability. Instead of reconstructing ($R,\theta,\phi$) at the same time, $\theta$ and $\phi$ were fixed at the reconstructed values provided by the time likelihood algorithm, and only the event radius $R$ was reconstructed using the charge likelihood algorithm. Therefore, the probability in Eq. 13 can be rewritten as $p(R,E)=\prod_{j}{P_{nohit}^{j}(R,E)}\cdot{\prod_{i}{P_{hit}^{i}(R,E)}}.$ (16) The reconstruction performance, focusing on the total reflection region, is shown in Figs. 13 and 14. In the total reflection region, the mean value of the reconstructed $R$ was consistent with the true $R$, and the resolutions in the $R$ direction were 81 mm at 1.022 MeV and 30 mm at 10.022 MeV. Figure 13: Bias of the reconstructed $R$ in the total reflection region at different energies calculated by the charge likelihood algorithm. Figure 14: Resolution of reconstructed $R$ as a function of energy calculated by the time likelihood and charge likelihood algorithms ($R^{3}\leavevmode\nobreak\ >\leavevmode\nobreak\ 4000\leavevmode\nobreak\ \rm{m}^{3}$) . As the charge distribution provides good radial discrimination ability, this algorithm can provide better resolution and a significantly smaller bias compared with those of the time likelihood algorithm in the total reflection region. ## VII Performance summary The execution time of the reconstruction for each event was tested on a computing cluster with Intel Xeon Gold 6238R CPUs (2.2 GHz), as shown in Fig. 15. The execution time of the charge-based algorithm was in the order of $O(10^{-4})$ s per event, which cannot be presented in the figure. The execution time of the time-based and the time likelihood algorithm was proportional to the event energy and could be reduced by using only the Hamamatsu PMTs for the reconstruction. The execution time of the charge likelihood algorithm was independent of the event energy. Figure 15: Execution time for the reconstruction for different algorithms. The resolutions of the four algorithms in the $R$ direction are shown in Fig. 16. Owing to the large bias of the charge-based algorithm, a correction to remove the position-dependent bias was applied before the analysis of the resolution. Figure 16: Resolution of the reconstructed $R$ as a function of energy for different algorithms. The charged-based algorithm is suitable for online reconstruction tasks that require high speed but do not require high resolution. The time-based algorithm does not rely on MC; it can be used as a data-driven reconstruction method. The time likelihood and charge likelihood algorithms are relatively accurate and each has its own advantages in a specific detector region. ## VIII Discussion The vertex resolutions of KamLAND and Borexino are approximately 12 cm and 10 cm at 1 MeV, respectively, and for JUNO it is approximately 10.5 cm. The diameter of JUNO (35.4 m) is several times larger than that of KamLAND (13 m) and Borexino (8.5 m). Despite its larger size, JUNO is still able to achieve a similar vertex resolution based only on the PMT time information. In this study, various effects on the vertex reconstruction for JUNO were comprehensively analyzed. As expected, the TTS of the PMT is the dominant factor. The vertex reconstruction capability of JUNO is mainly based on the Hamamatsu PMTs. Although the number of NNVT PMTs is more than twice, their time information is not useful in the vertex reconstruction because of their significantly inferior TTS. After considering the light yield of the LS and the PMT coverage, the number of photons detected by the Hamamatsu PMTs is in the same range as those of the three detectors mentioned above. This provides an explanation of their similar vertex resolutions based only on the PMT time information. To fully exploit the large PMT coverage and large number of PMTs in JUNO, the charge information of PMTs also need to be utilized in addition to the time information to constrain the event vertex, especially near the detector boundary. At the same time, the effect of the dark noise of the PMT can be mitigated with appropriate treatment. A more accurate initial value also improves the performance of the vertex reconstruction. In addition to the event vertex, the event time can also be reconstructed simultaneously, which is a useful variable for downstream analyses. In LS detectors, in addition to the scintillation photons there are also Cherenkov photons, whose effects need to be studied in the future. The $R$ bias near the detector edge in the current results also indicates that a more accurate PDF of the scintillator response function is needed to include its dependence on the position as well as the particle type. All vertex reconstruction methods based on PMT time information use the time of the first arrival photons only; in principle, later photons might be useful as well. Therefore, novel methods also need to be explored. Our preliminary studies on algorithms based on machine learning qianz_2021 showed comparable vertex reconstruction performance, and they need to be further investigated ## IX Conclusion In this study, four algorithms for the reconstruction of the event vertex and event time were investigated in detail and verified using MC samples generated by the offline software of JUNO. Considering the TTS and dark noise effects from the PMTs, a vertex resolution of 10 cm or higher can be achieved in the energy range of reactor neutrinos. The TTS has a major effect on the vertex resolution, whereas the effect of dark noise is limited. Near the detector boundary, charge information can constrain the event vertex better compared with time information. The algorithms discussed in this paper are applicable to current and future experiments using similar detection techniques. ## References * (1) KamLAND Collaboration, First Results from KamLAND: Evidence for Reactor Antineutrino Disappearance. Phys. Rev. Lett. 90 (2003) 021802. https://doi.org/10.1103/PhysRevLett.90.021802 * (2) BOREXINO Collaboration, Neutrinos from the primary proton–proton fusion process in the sun. Nature 512 (2014) 383–386. https://doi.org/10.1038/nature13702 * (3) DOUBLE CHOOZ Collaboration, Indication of reactor ${\overline{\nu}}_{e}$ disappearance in the Double CHOOZ experiment. Phys. Rev. Lett. 108 (2012) 131801. https://doi.org/10.1103/PhysRevLett.108.131801 * (4) DAYA BAY Collaboration, Observation of electron-antineutrino disappearance at Daya Bay. Phys. Rev. Lett. 108 (2012) 171803. https://doi.org/10.1103/PhysRevLett.108.171803 * (5) RENO Collaboration, Observation of reactor electron antineutrinos disappearance in the RENO experiment. Phys. Rev. Lett. 108 (2012) 191802. https://doi.org/10.1103/PhysRevLett.108.191802 * (6) S.T. Petcov, M. Piai, The LMA MSW solution of the solar neutrino problem, inverted neutrino mass hierarchy and reactor neutrino experiments. Phys. Lett. B 533 (2002) 94–106. https://doi.org/10.1016/S0370-2693(02)01591-5 * (7) SNO+ Collaboration, Current status and future prospects of the SNO+ experiment. Adv. High Energy Phys. 2016 (2016) 6194250. https://doi.org/10.1155/2016/6194250 * (8) JUNO Collaboration, Neutrino physics with JUNO. J. Phys. G 43 (2016) 030401. https://doi.org/10.1088/0954-3899/43/3/030401 * (9) Q. Liu, M. He, X. Ding et al., A vertex reconstruction algorithm in the central detector of JUNO. JINST 13 (09) (2018) T09005. https://doi.org/10.1088/1748-0221/13/09/t09005 * (10) T. Lin, J.H. Zou, W.D. Li et al., The application of SNiPER to the JUNO simulation. J. Phys. Conf. Ser. 898 (2017) 042029. https://doi.org/10.1088/1742-6596/898/4/042029 * (11) JUNO Collaboration, JUNO Conceptual Design Report. arXiv:1508.07166 [physics.ins-det] * (12) Z.M. Wang, JUNO PMT system and prototyping. J. Phys. Conf. Ser. 888 (2017) 012052. https://doi.org/10.1088/1742-6596/888/1/012052 * (13) G. Wang, Y.K. Heng, X.H. Li et al., Effect of divider current on 7.62cm photomultiplier tube performance. Nucl. Tech. 2018, 41(8): 80402-080402. https://doi.org/10.11889/j.0253-3219.2018.hjs.41.080402 * (14) K.J. Li, Z.Y. You, Y.M. Zhang et al., GDML based geometry management system for offline software in JUNO. Nucl. Instrum. Meth. A 908 (2018) 43–48. https://doi.org/10.1016/j.nima.2018.08.008 * (15) S. Zhang, J.S. Li, Y.J. Su et al., A method for sharing dynamic geometry information in studies on liquid-based detectors. Nucl. Sci. Tech. 32, 21 (2021). https://doi.org/10.1007/s41365-021-00852-8 * (16) Y. Zhang, Z.Y. Yu, X.Y. Li et al., A complete optical model for liquid-scintillator detectors. Nucl. Instrum. Meth. A 967 (2020) 163860. https://doi.org/10.1016/j.nima.2020.163860 * (17) Z.Y. You, K. Li, Y. Zhang et al., A ROOT based event display software for JUNO, JINST 13 (2018) T02002. https://doi.org/10.1088/1748-0221/13/02/t02002 * (18) J. Zhu, Z. You, Y. Zhang et al., A method of detector and event visualization with unity in JUNO. JINST 14 (01) (2019) T01007. https://doi.org/10.1088/1748-0221/14/01/t01007 * (19) M. Born et al., Principles of Optics: Electromagnetic Theory of Propagation, Interference and Diffraction of Light (7th ed.). Cambridge: Cambridge University Press, (1999). * (20) X. Zhou, Q. Liu, M. Wurm et al., Rayleigh scattering of linear alkylbenzene in large liquid scintillator detectors. Rev. Sci. Instrum. 86 (04 2015). https://doi.org/10.1063/1.4927458 * (21) A.N. Bashkatov, E.A. Genina, Water refractive index in dependence on temperature and wavelength: a simple approximation. International Society for Optics and Photonics, SPIE, 2003, pp. 393 – 395. https://doi.org/10.1117/12.518857 * (22) F. James, MINUIT Function Minimization and Error Analysis: Reference Manual Version 94.1 (1994). * (23) G. Ranucci, An analytical approach to the evaluation of the pulse shape discrimination properties of scintillators. Nucl. Instrum. Meth. A 354 (2) (1995) 389 – 399. https://doi.org/10.1016/0168-9002(94)00886-8 * (24) C. Galbiati, K. McCarty, Time and space reconstruction in optical, non-imaging, scintillator-based particle detectors. Nucl. Instrum. Meth. A 568 (2) (2006) 700 – 709. https://doi.org/10.1016/j.nima.2006.07.058 * (25) M. Moszynski, B. Bengtson, Status of timing with plastic scintillation detectors. Nucl. Instrum. Meth. A 158 (1979) 1 – 31. https://doi.org/10.1016/S0029-554X(79)90170-8 * (26) W.L. Zhong, Z.H. Li, C.G. Yang et al., Measurement of decay time of liquid scintillator. Nucl. Instrum. Meth. A 587 (2) (2008) 300 – 303. https://doi.org/10.1016/j.nima.2008.01.077 * (27) T.M. Undagoitia, F. von Feilitzsch, L. Oberauer et al., Fluorescence decay-time constants in organic liquid scintillators. Rev. Sci. Instrum. 80 (4) (2009) 043301. https://doi.org/10.1063/1.3112609 * (28) H.M. O’Keeffe, E. O’Sullivan, M.C. Chen, Scintillation decay time and pulse shape discrimination in oxygenated and deoxygenated solutions of linear alkylbenzene for the SNO+ experiment. Nucl. Instrum. Meth. A 640 (1) (2011) 119 – 122. https://doi.org/10.1016/j.nima.2011.03.027 * (29) H. Takiya, K. Abe, K. Hiraide et al., A measurement of the time profile of scintillation induced by low energy gamma-rays in liquid xenon with theXMASS-I detector. Nucl. Instrum. Meth. A 834 (2016) 192 – 96. https://doi.org/10.1016/j.nima.2016.08.014 * (30) O.J. Smirnov, Energy and Spatial Resolution of a Large-Volume Liquid-Scintillator Detector. Instrum. Exp. Tech. 46 (3) (2003) 327 – 344. https://doi.org/10.1023/a:1024458203966 * (31) R. Zhang, D.W. Cao, C.W. Loh et al. Using monochromatic light to measure attenuation length of liquid scintillator solvent LAB. Nucl. Sci. Tech. 30, 30 (2019). https://doi.org/10.1007/s41365-019-0542-1 * (32) W.J. Wu, M. He, X. Zhou et al., A new method of energy reconstruction for large spherical liquid scintillator detectors. JINST 14 (03) (2019) P03009. https://doi.org/10.1088/1748-0221/14/03/p03009 * (33) G.L. Zhu, J.L. Liu, Q. Wang et al., Ultrasonic positioning system for the calibration of central detector. Nucl. Sci. Tech. 30, 5 (2019). https://doi.org/10.1007/s41365-018-0530-x * (34) Z. Qian, V. Belavin, V. Bokov et al., Vertex and Energy Reconstruction in JUNO with Machine Learning Methods. arXiv:2101.04839 [physics.ins-det]
# Length functions on groups and rigidity Shengkui Ye ###### Abstract Let $G$ be a group. A function $l:G\rightarrow[0,\infty)$ is called a length function if (1) $l(g^{n})=|n|l(g)$ for any $g\in G$ and $n\in\mathbb{Z};$ (2) $l(hgh^{-1})=l(g)$ for any $h,g\in G;$ and (3) $l(ab)\leq l(a)+l(b)$ for commuting elements $a,b.$ Such length functions exist in many branches of mathematics, mainly as stable word lengths, stable norms, smooth measure-theoretic entropy, translation lengths on $\mathrm{CAT}(0)$ spaces and Gromov $\delta$-hyperbolic spaces, stable norms of quasi-cocycles, rotation numbers of circle homeomorphisms, dynamical degrees of birational maps and so on. We study length functions on Lie groups, Gromov hyperbolic groups, arithmetic subgroups, matrix groups over rings and Cremona groups. As applications, we prove that every group homomorphism from an arithmetic subgroup of a simple algebraic $\mathbb{Q}$-group of $\mathbb{Q}$-rank at least $2,$ or a finite-index subgroup of the elementary group $E_{n}(R)$ $(n\geq 3)$ over an associative ring, or the Cremona group $\mathrm{Bir}(P_{\mathbb{C}}^{2})$ to any group $G$ having a purely positive length function must have its image finite. Here $G$ can be outer automorphism group $\mathrm{Out}(F_{n})$ of free groups, mapping classes group $\mathrm{MCG}(\Sigma_{g})$, $\mathrm{CAT}(0)$ groups or Gromov hyperbolic groups, or the group $\mathrm{Diff}(\Sigma,\omega)$ of diffeomorphisms of a hyperbolic closed surface preserving an area form $\omega.$ ### 0.1 Introduction The rigidity phenomena have been studied for many years. The famous Margulis superrigidity implies any group homomorphism between irreducible lattices in semisimple Lie groups of real rank $\mathrm{rk}_{\mathbb{R}}(G)\geq 2$ are virtually induced by group homomorphisms between the Lie groups. Therefore, group homomorphisms from ‘higher’-rank irreducible lattices to ‘lower’-rank irreducible lattices normally have finite images. Farb, Kaimanovich and Masur [26] [39] prove that every homomorphism from an (irreducible) higher rank lattice into the mapping class group $\mathrm{MCG}(\Sigma_{g})$ has a finite image. Bridson and Wade [17] showed that the same superrigidity remains true if the target is replaced with the outer automorphism group $\mathrm{Out}(F_{n})$ of the free group. Mimura [48] proves that every homomorphism from Chevalley group over commutative rings to $\mathrm{MCG}(\Sigma_{g})$ or $\mathrm{Out}(F_{n})$ has a finite image. Many other rigidity results can be found, e.g. [49] [18] [19] [33] [54] and [53]. In this article, we study rigidity phenomena with the notion of length functions. Let $G$ be a group. We call a function $l:G\rightarrow[0,\infty)$ a length function if 1) $l(g^{n})=|n|l(g)$ for any $g\in G$ and $n\in\mathbb{Z};$ 2) $l(aga^{-1})=l(g)$ for any $a,g\in G;$ 3) $l(ab)\leq l(a)+l(b)$ for commuting elements $a,b.$ Such length functions exist in geometric group theory, dynamical system, algebra, algebraic geometry and many other branches of mathematics. For example, the following functions $l$ are length functions (see Section 3 for more examples with details). * • (The stable word lengths) Let $G$ be a group generated by a symmetric (not necessarily finite) set $S.$ For any $g\in G,$ the word length $\phi_{S}(w)=\min\\{n\mid g=s_{1}s_{2}\cdots s_{n},$each $s_{i}\in S\\}$ is the minimal number of elements of $S$ whose product is $g.$ The stable length is defined as $l(g)=\lim_{n\rightarrow\infty}\frac{\phi_{S}(g^{n})}{n}.$ * • (Stable norms) Let $M$ be a compact smooth manifold and $G=\mathrm{Diff}(M)$ the diffeomorphism group consisting of all self-diffeomorphisms. For any diffeomorphism $f:M\rightarrow M,$ let $\|f\|=\sup_{x\in M}\|D_{x}f\|,$ where $D_{x}f$ is the induced linear map between tangent spaces $T_{x}M\rightarrow T_{f(x)}M.$ Define $l(f)=\max\\{\lim_{n\rightarrow+\infty}\frac{\log\|f^{n}\|}{n},\lim_{n\rightarrow+\infty}\frac{\log\|f^{-n}\|}{n}\\}.$ * • (smooth measure-theoretic entropy) Let $M$ be a $C^{\infty}$ closed Riemannian manifold and $G=\mathrm{Diff}_{\mu}^{2}(M)$ consisting of diffeomorphisms of $M$ preserving a Borel probability measure $\mu.$ Let $l(f)=h_{\mu}(f)$ be the measure-theoretic entropy, for any $f\in G=\mathrm{Diff}_{\mu}^{2}(M)$. * • (Translation lengths) Let $(X,d)$ be a metric space and $G=\mathrm{Isom}(X)$ consisting of isometries $\gamma:X\rightarrow X$. Fix $x\in X,$ define $l(\gamma)=\lim_{n\rightarrow\infty}\frac{d(x,\gamma^{n}x)}{n}.$ This contains the translation lengths on $\mathrm{CAT}(0)$ spaces and Gromov $\delta$-hyperbolic spaces as special cases. * • (average norm for quasi-cocycles) Let $G$ be a group and $E$ be a Hilbert space with an $G$-action by linear isometrical action. A function $f:G\rightarrow E$ is a quasi-cocyle if there exists $C>0$ such that $\|f(gh)-f(g)-gf(h)\|<C$ for any $g,h\in G.$ Let $l:G\rightarrow[0,+\infty)$ be defined by $l(g)=\lim_{n\rightarrow\infty}\frac{\|f(g^{n})\|}{n}.$ * • (Rotation numbers of circle homeomorphisms) Let $\mathbb{R}$ be the real line and $G=\mathrm{Home}_{\mathbb{Z}}(\mathbb{R})=\\{f\mid f:R\rightarrow R$ is a monotonically increasing homeomorphism such that $f(x+n)=f(x)$ for any $n\in\mathbb{Z}\\}.$ For any $f\in\mathrm{Home}_{\mathbb{Z}}(\mathbb{R})$ and $x\in[0,1),$ the translation number is defined as $l(f)=\lim_{n\rightarrow\infty}\frac{f^{n}(x)-x}{n}.$ * • (Asymptotic distortions) Let $f$ be a $C^{1+bv}$ diffeomorphism of the closed interval $[0,1]$ or the circle $S^{1}.$ (“bv” means derivative with finite total variation.) The asymptotic distortion of $f$ is defined (by Navas [50]) as $l(f)=\lim_{n\rightarrow\infty}\mathrm{var}(\log Df^{n}).$ This gives a length function $l$ on the group $\mathrm{Diff}^{1+bv}(M)$ of $C^{1+bv}$ diffeomorphisms for $M=[0,1]$ or $S^{1}.$ * • (Dynamical degree) Let $\mathbb{C}P^{n}$ be the complex projective space and $f:\mathbb{C}P^{n}\dashrightarrow\mathbb{C}P^{n}$ be a birational map given by $(x_{0}:x_{1}:\cdots:x_{n})\dashrightarrow(f_{0}:f_{1}:\cdots:f_{n}),$ where the $f_{i}$’s are homogeneous polynomials of the same degree without common factors. The degree of $f$ is $\deg f=\deg f_{i}.$ Define $l(f)=\max\\{\lim_{n\rightarrow\infty}\log\deg(f^{n})^{\frac{1}{n}},\lim_{n\rightarrow\infty}\log\deg(f^{-n})^{\frac{1}{n}}\\}.$ This gives a length function $l:\mathrm{Bir}(\mathbb{C}P^{n})\rightarrow[0,+\infty).$ Here $\mathrm{Bir}(\mathbb{C}P^{n})$ is the group of birational maps, also called Cremona group. The terminologies of length functions are used a lot in the literature (eg. [28], [22]). However, they usually mean different things from ours (in particular, it seems that the condition 3) has not been addressed for commuting elements before). Our first observation is the following result on vanishing of length functions. ###### Theorem 0.1 Let $G_{A}=\mathbb{Z}^{2}\rtimes_{A}\mathbb{Z}$ be an abelian-by-cyclic group, where $A\in\mathrm{SL}_{2}(\mathbb{Z})$. 1. (i) When the absolute value of the trace $|\mathrm{tr}(A)|>2,$ any length function $l:\mathbb{Z}^{2}\rtimes_{A}\mathbb{Z}\rightarrow\mathbb{R}_{\geq 0}$ vanishes on $\mathbb{Z}^{2}.$ 2. (ii) When $|\mathrm{tr}(A)|=2$ and $A\neq I_{2},$ any length function $l:\mathbb{Z}^{2}\rtimes_{A}\mathbb{Z}\rightarrow\mathbb{R}_{\geq 0}$ vanishes on the direct summand of $\mathbb{Z}^{2}$ spanned by eigenvectors of $A$. ###### Corollary 0.2 Suppose that the semi-direct product $G_{A}=\mathbb{Z}^{2}\rtimes_{A}\mathbb{Z}$ acts on a compact manifold by Lipschitz homeomorphisms (or $C^{2}$-diffeomorphisms, resp.). The topological entropy $h_{top}(g)=0$ (or Lyapunov exponents of $g$ are zero, resp.) for any $g\in\mathbb{Z}^{2}$ when $|\mathrm{tr}(A)|>2$ or any eigenvector $g\in\mathbb{Z}^{2}$ when $|\mathrm{tr}(A)|=2.$ It is well-known that the central element in the integral Heisenberg group $G_{A}$ (for $A=\begin{bmatrix}1&1\\\ 0&1\end{bmatrix}$) is distorted in the word metric. When the Heisenberg group $G_{A}$ acts on a $C^{\infty}$ compact Riemannian manifold, Hu-Shi-Wang [36] proves that the topological entropy and all Lyapunov exponents of the central element are zero. These results are special cases of Theorem 0.1 and Corollary 0.2, by choosing special length functions. A length function $l:G\rightarrow[0,\infty)$ is called purely positive if $l(g)>0$ for any infinite-order element $g.$ A group $G$ is called virtually poly-positive, if there is a finite-index subgroup $H<G$ and a subnormal series $1=H_{n}\vartriangleleft H_{n-1}\vartriangleleft\cdots\vartriangleleft H_{0}=H$ such that every finitely generated subgroup of each quotient $H_{i}/H_{i+1}$ $(i=0,...,n-1)$ has a purely positive length function. Our following results are on the rigidity of group homomorphisms. ###### Theorem 0.3 Let $\Gamma$ be an arithmetic subgroup of a simple algebraic $\mathbb{Q}$-group of $\mathbb{Q}$-rank at least $2.$ Suppose that $G$ is virtually poly-positive. Then any group homomorphism $f:\Gamma\rightarrow G$ has its image finite. ###### Theorem 0.4 Let $G$ be a group having a finite-index subgroup $H<G$ and a subnormal series $1=H_{n}\vartriangleleft H_{n-1}\vartriangleleft\cdots\vartriangleleft H_{0}=H$ satisfying that 1. (i) every finitely generated subgroup of each quotient $H_{i}/H_{i+1}$ $(i=0,...,n-1)$ has a purely positive length function, i.e. $G$ is virtually poly-positive; and 2. (ii) any torsion abelian subgroup in every finitely generated subgroup of each quotient $H_{i}/H_{i+1}$ $(i=0,...,n-1)$ is finitely generated. Let $R$ be a finitely generated associative ring with identity and $E_{n}(R)$ the elementary subgroup. Suppose that $\Gamma<E_{n}(R)$ is finite-index subgroup. Then any group homomorphism $f:\Gamma\rightarrow G$ has its image finite when $n\geq 3$. ###### Corollary 0.5 Let $\Gamma$ be an arithmetic subgroup of a simple algebraic $\mathbb{Q}$-group of $\mathbb{Q}$-rank at least $2,$ or a finite-index subgroup of the elementary subgroup $E_{n}(R)$ $(n\geq 3)$ for an associative ring $R.$ Then any group homomorphism $f:E\rightarrow G$ has its image finite. Here $G$ is one of the following groups: * • a Gromov hyperbolic group, * • $\mathrm{CAT(0)}$ group, * • automorphism group $\mathrm{Aut}(F_{k})$ of a free group, * • outer automorphism group $\mathrm{Out}(F_{k})$ of a free group or * • mapping class group $\mathrm{MCG}(\Sigma_{g})$ $(g\geq 2)$. * • the group $\mathrm{Diff}(\Sigma,\omega)$ of diffeomorphisms of a closed surface preserving an area form $\omega.$ ###### Theorem 0.6 Suppose that $G$ is virtually poly-positive. Let $R$ be a finitely generated associative ring of characteristic zero such that any nonzero ideal is of a finite index (eg. the ring of algebraic integers in a number field). Suppose that $S<E_{n}(R)$ is a finite-index subgroup of the elementary group. Then any group homomorphism $f:S\rightarrow G$ has its image finite when $n\geq 3$. ###### Corollary 0.7 Let $R$ be an associative ring of characteristic zero such that any nonzero ideal is of a finite index. Any group homomorphism $f:E\rightarrow G$ has its image finite, where $E<E_{n}(R)$ is finite-index subgroup and $n\geq 3$. Here $G$ is one of the followings: * • $\mathrm{Aut}(F_{k}),\mathrm{Out}(F_{k}),\mathrm{MCG}(\Sigma_{g}),$ * • a hyperbolic group, * • a $\mathrm{CAT}(0)$ group or more generally a semi-hyperbolic group, * • a group acting properly semi-simply on a $\mathrm{CAT}(0)$ space, or * • a group acting properly semi-simply on a $\delta$-hyperbolic space, * • the group $\mathrm{Diff}(\Sigma,\omega)$ of diffeomorphisms of a hyperbolic closed surface preserving an area form $\omega.$ Some relevent cases of Theorem 0.4 and Theorem 0.6 are already established in the literature. Bridson and Wade [17] showed that any group homomorphism from an irreducible lattice in a semisimple Lie group of real rank $\geq 2$ to the mapping class group $\mathrm{MCG}(\Sigma_{g})$ must have its image finite. However, Theorem 0.3 can never hold when $\Gamma$ is a cocompact lattice, since a cocompact lattice has its stable word length purely positive. When the length functions involved in the virtually poly-positive group $G$ are required to be stable word lengths, Theorem 0.3 holds more generally for $\Gamma$ non-uniform irreducible lattices a semisimple Lie group of real rank $\geq 2$ (see Proposition 8.4). When the length functions involved in the virtually poly-positive group $G$ are given by a particular kind of quasi- cocyles, Theorem 0.3 holds more generally for $\Gamma$ with property TT (cf. Py [54], Prop. 2.2). Haettel [33] prove that any action of a high-rank a higher rank lattice on a Gromov-hyperbolic space is elementary (i.e. either elliptic or parabolic). Guirardel and Horbez [33] prove that every group homomorphism from a high-rank lattice to the outer automorphism group of torsion-free hyperbolic group has finite image. Thom [57] (Corollary 4.5) proves that any group homomorphism from a boundedly generated with property (T) to a Gromov hyperbolic group has finite image. Compared with these results, our target group $G$ and the source group $E_{n}(R)$ (can be defined over any non-commutative ring) in Theorem 0.4 are much more general. The inequalities of $n$ in Theorem 0.4, Theorem 0.6 and Corollary 0.5, Corollary 0.7 can not be improved, since $\mathrm{SL}_{2}(\mathbb{Z})$ is hyperbolic. The group $\Gamma$ in Corrolary 0.5 has Kazhdan’s property T (i.e. an arithmetic subgroup of a simple algebraic $\mathbb{Q}$-group of $\mathbb{Q}$-rank at least $2,$ or a finite-index subgroup of the elementary subgroup $E_{n}(R),$ $n\geq 3,$ for an associative ring $R$ has Kazhdan’s property T by [23]). However, there exist hyperbolic groups with Kazhdan’s property $T$ (cf. [32], Section 5.6). This implies that Corrolary 0.5 does not hold generally for groups $\Gamma$ with Kazhdan’s property T. Franks and Handel [27] prove that any group homomorphism from a quasi-simple group containing a subgroup isomorphic to the three-dimensional integer Heisenberg group, to the group $\mathrm{Diff}(\Sigma,\omega)$ of diffeomorphisms of a closed surface preserving an area form $\omega,$ has its image finite (cf. Lemma 8.3). We now study length functions on the Cremona groups. ###### Theorem 0.8 Let $\mathrm{Bir}(\mathbb{P}_{k}^{n})$ $(n\geq 2)$ be the group of birational maps on the projective space $\mathbb{P}_{k}^{n}$ over an algebraic closed field $k$. Any length function $l:\mathrm{Bir}(\mathbb{P}_{k}^{n})\rightarrow[0,+\infty)$ vanishes on the automorphism group $\mathrm{Aut}(\mathbb{P}_{k}^{n})=\mathrm{PGL}_{n+1}(k).$ When $n=2,$ a result of Blanc and Furter [9] (page 7 and Proposition 4.8.10) implies that there are three length functions $l_{1},l_{2},l_{3}$ on $\mathrm{Bir}(\mathbb{P}_{k}^{2})$ such that any element $g\in\mathrm{Bir}(\mathbb{P}_{k}^{n})$ satisfying $l_{1}(g)=l_{2}(g)=l_{3}(g)=0$ is either finite or conjugate to an element in $\mathrm{Aut}(\mathbb{P}_{k}^{2}).$ This implies that the automorphism group $\mathrm{Aut}(\mathbb{P}_{k}^{n})$ (when $k=2$) is one of the ‘largest’ subgroups of $\mathrm{Bir}(\mathbb{P}_{k}^{n})$ on which every length function vanishes. ###### Corollary 0.9 Let $G$ be a virtually poly-positive group. Any group homomorphism $f:\mathrm{Bir}(\mathbb{P}_{k}^{2})\rightarrow G$ is trivial, for an algebraic closed field $k$. In particular, Corrolary 0.9 implies that any quotient group of $\mathrm{Bir}(\mathbb{P}_{k}^{2})$ cannot act properly semisimply neither on a Gromov $\delta$-hyperbolic space nor a $\mathrm{CAT}(0)$ space. This is interesting, considering the following facts. There are (infinite-dimensional) hyperbolic space and cubical complexes, on which $\mathrm{Bir}(P_{k}^{2})$ acts isometrically (see [21], Section 3.1.2 and [45]). The Cremona group $\mathrm{Bir}(\mathbb{P}_{k}^{2})$ is sub-quotient universal: every countable group can be embedded in a quotient group of $\mathrm{Bir}(\mathbb{P}_{k}^{2})$ (see [21], Theorem 4.7). Moreover, Blanc- Lamy-Zimmermann [10] (Theorem E) proves that when $n\geq 3,$ there is a surjection from $\mathrm{Bir}(\mathbb{P}_{k}^{n})$ onto a free product of two- element groups $\mathbb{Z}/2.$ This means that Corrolary 0.9 can never hold for higher dimensional Cremona groups. As byproducts, we give characterizations of length functions on Lie groups. Our next result is that there is essentially only one length function on the special linear group $\mathrm{SL}_{2}(\mathbb{R})$: ###### Theorem 0.10 Let $G=\mathrm{SL}_{2}(\mathbb{R}).$ Any length function $l:G\rightarrow[0,+\infty)$ continuous on the subgroup $SO(2)$ and the diagonal subgroup is proposional to the translation function $\tau(g):=\inf_{x\in X}d(x,gx),$ where $X=\mathrm{SL}_{2}(\mathbb{R})/\mathrm{SO}(2)$ is the upper-half plane. More generally, we study length functions on Lie groups. Let $G$ be a connected semisimple Lie group whose center is finite with an Iwasawa decomposition $G=KAN$. Let $W$ be the Weyl group, i.e. the quotient group of the normalizers $N_{K}(A)$ modulo the centralizers $C_{K}(A)$. Our second result shows that a length function $l$ on $G$ is uniquely determined by its image on $A.$ ###### Theorem 0.11 Let $G$ be a connected semisimple Lie group whose center is finite with an Iwasawa decomposition $G=KAN$. Let $W$ be the Weyl group. 1. (i) Any length function $l$ on $G$ that is continuous on the maximal compact subgroup $K$ is determined by its image on $A.$ 2. (ii) Conversely, any length function $l$ on $A$ that is $W$-invariant (i.e. $l(w\cdot a)=l(a)$) can be extended to be a length function on $G$ that vanishes on the maximal compact subgroup $K.$ The proofs of Theorems 0.10 0.11 are based the Jordan-Chevalley decompositions of algebraic groups and Lie groups. We will prove that any length function on a Heisenberg group vanishes on the central elements (see Lemma 5.2). This is a key step for many other proofs. Based this fact, we prove Theorems 0.3, 0.6, 0.4, 0.8 by looking for Heisenberg subgroups. In Section 1, we give some elementary facts on the length functions. In Section 2, we discuss typical examples of length functions. In later sections, we study length functions on Lie groups, algebraic groups, hyperbolic groups, matrix groups and the Cremona groups. ## 1 Basic properties of length functions ### 1.1 Length functions ###### Definition 1.1 Let $G$ be a group. A function $l:G\rightarrow[0,\infty)$ is called a length function if 1) $l(g^{n})=|n|l(g)$ for any $g\in G$ and $n\in\mathbb{Z}.$ 2) $l(aga^{-1})=l(g)$ for any $a,g\in G.$ 3) $l(ab)\leq l(a)+l(b)$ for commuting elements $a,b.$ ###### Lemma 1.2 Any torsion element $g\in G$ has length $l(g)=0.$ Proof. Note that $l(1)=2l(1)$ and thus $l(1)=0.$ If $g^{n}=1,$ then $l(g)=|n|l(1)=0.$ Recall that a subset $V$ of a real vector space is a convex cone, if $av+bw\in V$ for any $v,w\in V$ and any non-negative real numbers $a,b\geq 0.$ ###### Lemma 1.3 The set $\mathrm{Func}(G)$ of all length functions on a group $G$ is a convex cone. Proof. It is obvious that for two functions $l_{1},l_{2}$ on $G,$ a non- negative linear combination $al_{1}+bl_{2}$ is a new length function. ###### Lemma 1.4 Let $f:G\rightarrow H$ be a group homomorphism between two groups $G$ and $H.$ For any length function $l:H\rightarrow[0,\infty),$ the composite $l\circ f$ is a length function on $G.$ Proof. It is enough to note that a group homomorphism preserves powers of elements, conjugacy classes and commutativity of elements. ###### Corollary 1.5 For a group $G,$ let $\mathrm{Out}(G)=\mathrm{Aut}(G)/\mathrm{Inn}(G)$ be the outer automorphism group. Then $\mathrm{Out}(G)$ acts on the set $\mathrm{Func}(G)$ of all length functions by pre-compositions $l\mapsto l\circ g,$ where $l\in\mathrm{Func}(G),$ $g\in\mathrm{Out}(G)$. This action preserves scalar multiplications and linear combinations (with non-negative coefficients). Proof. For an inner automorphism $I_{g}:G\rightarrow G$ given by $I_{g}(h)=ghg^{-1},$ the length function $l\circ I_{g}=l$ since $l$ is invariant under conjugation. Therefore, the outer automorphism group $\mathrm{Out}(G)$ has an action on $\mathrm{Func}(G).$ It is obvious that the pre-compositions preserve scalar multiplications and linear combinations with non-negative coefficients. ###### Definition 1.6 A length function $l:G\rightarrow[0,\infty)$ is primitive if it is not a composite $l^{\prime}\circ f$ for a non-trivial surjective group homomorphism $f:G\twoheadrightarrow H$ and a length function $l^{\prime}:$ $H\rightarrow[0,\infty).$ ###### Lemma 1.7 Suppose that a length function $l:G\rightarrow[0,\infty)$ vanishes on a central subgroup $H<G.$ Then $l$ factors through the quotient group $G/H.$ In other words, there exists a length function $l^{\prime}:G/H\rightarrow[0,\infty)$ such that $l=l^{\prime}\circ q,$ where $q:G\rightarrow G/H$ is the quotient group homomorphism. Proof. Write $G=\cup gH,$ the union of left cosets. For any $h\in H,$ we have $l(gh)\leq l(g)+l(h)=l(g)$ and $l(g)=l(ghh^{-1})\leq l(gh).$ Therefore, $l(gh)=l(g)$ for any $h\in H.$ Define $l^{\prime}(gH)=l(g).$ Then $l^{\prime}$ is a length function on the quotient group $G/H.$ The required property follows the definition easily. ###### Corollary 1.8 Suppose that a group $G$ has non-trivial finite central subgroup $Z(G).$ Any length function $l$ on $G$ factors through $G/Z(G).$ Proof. This follows Lemma 1.7 and Lemma 1.2. ###### Lemma 1.9 Let $G$ be a group. Suppose that any non-trivial normal subgroup $H\vartriangleleft G$ is of finite index. Then any non-vanishing length function $l:G\rightarrow[0,\infty)$ is primitive. Proof. Suppose that $l$ is a composite $l^{\prime}\circ f$ for a non-trivial surjective group homomorphism $f:G\twoheadrightarrow H$ and a length function $l^{\prime}:H\rightarrow[0,\infty).$ By the assumption of $G,$ the quotient group $H$ is finite. This implies that $l^{\prime}$ and thus $l$ vanishes, which is a contradiction. ###### Theorem 1.10 Let $\Gamma$ be an irreducible lattice in a connected irreducible semisimple Lie group of real rank $\geq 2.$ Then any non-vanishing length function $l:\Gamma\rightarrow[0,\infty)$ factors through a primitive function on $\Gamma/Z(\Gamma)$. Proof. By the Margulis-Kazhdan theorem (see [62], Theorem 8.1.2), any normal subgroup $N$ of $\Gamma$ either lies in the center of $\Gamma$ (and hence it is finite) or the quotient group $\Gamma/N$ is finite. Corollary 1.8 implies that $l$ factors through a length function $l^{\prime}$ on $\Gamma/Z(\Gamma).$ The previous lemma 1.9 implies that $l^{\prime}$ is primitive. ## 2 Examples of length functions Let’s see a general example first. Let $G$ be a goup and $f:G\rightarrow[0,+\infty)$ be a function satisfying $f(gh)\leq f(g)+f(h)$ and $f(g)=f(g^{-1})$ for any elements $g,h\in G.$ Define $l:G\rightarrow[0,+\infty)$ by $l(g)=\lim_{n\rightarrow\infty}\frac{f(g^{n})}{n}$ for any $g\in G.$ ###### Lemma 2.1 The function $l$ is a length function in the sense of Definition 1.1. Proof. For any $g\in G,$and natural numbers $n,m,$ we have $f(g^{n+m})\leq f(g^{n})+f(g^{m}).$ This means that $\\{f(g^{n})\\}_{n=1}^{\infty}$ is a subadditive sequence and thus the limit $\lim_{n\rightarrow\infty}\frac{f(g^{n})}{n}$ exists. This shows that $l$ is well-defined. From the definition of $l,$ it is clear that $l(g^{n})=|n|l(g)$ for any integer $n.$ Let $h\in G.$ We have $l(hgh^{-1})=\lim_{n\rightarrow\infty}\frac{f(hg^{n}h^{-1})}{n}\leq\lim_{n\rightarrow\infty}\frac{f(h)+f(g^{n})+f(h^{-1})}{n}=\lim_{n\rightarrow\infty}\frac{f(g^{n})}{n}=l(g).$ Similarly, we have $l(g)=l(h^{-1}(hgh^{-1})h)\leq l(hgh^{-1})$ and thus $l(g)=l(hgh^{-1}).$ For commuting elements $a,b,$ we have $(ab)^{n}=a^{n}b^{n}.$ Therefore, $\displaystyle l(ab)$ $\displaystyle=$ $\displaystyle\lim_{n\rightarrow\infty}\frac{f((ab)^{n})}{n}=\lim_{n\rightarrow\infty}\frac{f(a^{n}b^{n})}{n}$ $\displaystyle\leq$ $\displaystyle\lim_{n\rightarrow\infty}\frac{f(a^{n})+f(b^{n})}{n}\leq l(a)+l(b).$ Many (but not all) length functions $l$ come from subadditive functions $f.$ ### 2.1 Stable word lengths Let $G$ be a group generated by a (not necessarily finite) set $S$ satisfying $s^{-1}\in S$ for each $s\in S.$ For any $g\in G,$ the word length $\phi_{S}(w)=\min\\{n\mid g=s_{1}s_{2}\cdots s_{n},$each $s_{i}\in S\\}$ is the minimal number of elements of $S$ whose product is $g.$ The stable length $l(g)=\lim_{n\rightarrow\infty}\frac{\phi_{S}(g^{n})}{n}.$ Since $\phi_{S}(g^{n})$ is subadditive, the limit always exists. ###### Lemma 2.2 The stable length $l:G\rightarrow[0,+\infty)$ is a length function in the sense of Definition 1.1. Proof. From the definition of the word length $\phi_{S},$ it is clear that $\phi_{S}(gh)\leq\phi_{S}(g)+\phi_{S}(h)$ and $\phi_{S}(g)=\phi_{S}(g^{-1})$ for any $g,h\in G.$ The claim is proved by Lemma 2.1. When $S$ is the set of commutators, the $l(g)$ is called the stable commutator length, which is related to lots of topics in low-dimensional topology (see Calegari [20]). ### 2.2 Growth rate Let $G$ be a group generated by a finite set $S$ satisfying $s^{-1}\in S$ for each $s\in S.$ Suppose $|\cdot|_{S}$ is the word length of $(G,S).$ For any automorphism $\alpha:G\rightarrow G,$ define $l^{\prime}(\alpha)=\max\\{|\alpha(s_{i})|_{S}:s_{i}\in S\\}.$ Let $l(\alpha)=\lim_{n\rightarrow\infty}\frac{\log l^{\prime}(\alpha^{n})}{n}.$ This number $l(\alpha)$ is called the algebraic entropy of $\alpha$ (cf. [40], Definition 3.1.9, page 114). ###### Lemma 2.3 Let $\mathrm{Aut}(G)$ be the group of automorphisms of $G.$ The function $l:\mathrm{Aut}(G)\rightarrow[0,+\infty)$ is a length function in the sense of Definition 1.1. Proof. Since $\alpha(s_{i})^{-1}=\alpha^{-1}(s_{i})$ for any $s_{i}\in S,$ we know that $l^{\prime}(\alpha)=l^{\prime}(\alpha^{-1}).$ For another automorphism $\beta:G\rightarrow G,$ let $l^{\prime}(\beta)=|\beta(s_{i})|_{S}$ for some $s_{i}\in S.$ Suppose that $\beta(s_{i})=s_{i_{1}}s_{i_{2}}\cdots s_{i_{k}}$ with $k=l^{\prime}(\beta).$ Then $|(\alpha\beta)(s_{i})|_{S}=|\alpha(s_{i_{1}})\alpha(s_{i_{2}})\cdots\alpha(s_{i_{k}})|_{S}\leq l^{\prime}(\alpha)k.$ This proves that $l^{\prime}(\alpha\beta)\leq l^{\prime}(\alpha)l^{\prime}(\beta).$ The claim is proved by Lemma 2.1. Fix $g\in G.$ For any automorphism $\alpha:G\rightarrow G,$ define $b_{n}=|\alpha^{n}(g)|_{S}.$ Suppose that $g=s_{1}s_{2}\cdots s_{k}$ with $k=|g|_{S}.$ Note that $b_{n}=|\alpha^{n}(g)|_{S}=|\alpha^{n}(s_{1})\alpha^{n}(s_{2})\cdots\alpha^{n}(s_{k})|_{S}\leq l^{\prime}(\alpha^{n})|g|_{S}.$ Therefore, we have $\lim\sup_{n\rightarrow\infty}\frac{\log b_{n}}{n}\leq l(\alpha).$ This implies that $l(\alpha)$ is an upper bounded for growth rate of $\\{|\alpha^{n}(g)|_{S}\\}.$ The growth rate is studied a lot in geometric group theory (for example, see [43] for growth of automorphisms of free groups). ### 2.3 Matrix norms and group acting on smooth manifolds For a square matrix $A,$ the matrix norm $\|A\|=\sup_{\|x\|=1}\|Ax\|.$ Define the stable norm $s(A)=\lim_{n\rightarrow+\infty}\frac{\log\|A^{n}\|}{n}.$ Since $\|AB\|\leq\|A\|\|B\|$ for any two matrices $A,B,$ the sequence $\\{\log\|A^{n}\|\\}_{n=1}^{\infty}$ is subadditive and thus the limit exists. ###### Lemma 2.4 Let $G=\mathrm{GL}_{n}(\mathbb{R})$ be the general linear group. The function $l:G\rightarrow[0,+\infty)$ defined by $l(g)=\max\\{s(g),s(g^{-1})\\}$ is a length function in the sense of Definition 1.1. Proof. From the definition of the matrix norm, it is clear that $\log\|gh\|\leq\log\|g\|+\log\|h\|$ for any $g,h\in G.$ Then $l(g)=\max\\{s(g),s(g^{-1})\\}$ is a length function by Lemma 2.1. Let $M$ be a compact smooth manifold and $\mathrm{Diff}(M)$ the diffeomorphism group consisting of all self-diffeomorphisms. For any diffeomorphism $f:M\rightarrow M,$ let $\|f\|=\sup_{x\in M}\|D_{x}f\|,$ where $D_{x}f$ is the induced linear map between tangent spaces $T_{x}M\rightarrow T_{f(x)}M.$ Define $l(f)=\max\\{\lim_{n\rightarrow+\infty}\frac{\log\|f^{n}\|}{n},\lim_{n\rightarrow+\infty}\frac{\log\|f^{-n}\|}{n}\\}.$ A similar argument as the proof of the previous lemma proves the following. ###### Lemma 2.5 Let $G$ be a group acting on a smooth manifold $M$ by diffeomorphisms. The function $l:G\rightarrow[0,+\infty)$ is a length function in the sense of Definition 1.1. For an $f$-invariant Borel probability measure $\mu$ on $M,$ it is well known (see [O]) that there exists a measurable subset $\Gamma_{f}\subset M$ with $\mu(\Gamma_{f})=1$ such that for all $x\in\Gamma_{f}$ and $u\in T_{x}M,$ the limit $\chi(x,u,f)=\lim\frac{1}{n}\log\|D_{x}f^{n}(u)\|$ exists and is called Lyapunov exponent of $u$ at $x.$ From the definitions, we know that $\chi(x,u,f)\leq l(f)$ for any $x\in\Gamma_{f}$ and $u\in T_{x}M.$ ### 2.4 Smooth measure-theoretic entropy Let $T:X\rightarrow X$ be a measure-preserving map of the probability space $(X,\mathfrak{B},m).$ For a finite-sub-$\sigma$-algebra $A=\\{A_{1},A_{2},...,A_{k}\\}$ of $\mathfrak{B},$ denote by $\displaystyle H(A)$ $\displaystyle=$ $\displaystyle-\sum m(A_{i})\log m(A_{i}),$ $\displaystyle h(T,A)$ $\displaystyle=$ $\displaystyle\lim\frac{1}{n}H(\vee_{i=0}^{n-1}T^{-i}A),$ where $\vee_{i=0}^{n-1}T^{-i}A$ is a set consisting of sets of the form $\cap_{i=0}^{n-1}T^{-i}A_{j_{i}}.$ The entropy of $T$ is defined as $h_{m}(T)=\sup h(T,A),$ where the supremum is taken over all finite sub- algebra $A$ of $\mathfrak{B}.$ For more details, see Walters [58] (Section 4.4). ###### Lemma 2.6 Let $M$ be a $C^{\infty}$ closed Riemannian manifold and $G=\mathrm{Diff}_{\mu}^{2}(M)$ consisting of diffeomorphisms of $M$ preserving a Borel probability measure $\mu.$ The entropy $h_{\mu}$ is a length function on $\mathrm{Diff}_{\mu}^{2}(M)$ in the sense of Definition 1.1. Proof. For any $f,g\in\mathrm{Diff}_{\mu}^{2}(M)$ and integer $n,$ it is well- known that $h_{\mu}(f^{n})=|n|h_{\mu}(f)$ and $h_{\mu}(f)=h_{\mu}(gfg^{-1})$ (cf. [58], Theorem 4.11 and Theorem 4.13). Hu [35] proves that $h_{\mu}(fg)\leq h_{\mu}(f)+h_{\mu}(g)$ when $fg=gf.$ ### 2.5 Stable translation length on metric spaces Let $(X,d)$ be a metric space and $\gamma:X\rightarrow X$ an isometry. Fix $x\in X.$ Note that $d(x,\gamma_{1}\gamma_{2}x)\leq d(x,\gamma_{1}x)+d(\gamma_{1}x,\gamma_{1}\gamma_{2}x)=d(x,\gamma_{1}x)+d(x,\gamma_{2}x)$ and $d(x,\gamma_{1}x)=d(x,\gamma_{1}^{-1}x)$ for any isometries $\gamma_{1},\gamma_{2}.$ Define $l(\gamma)=\lim_{n\rightarrow\infty}\frac{d(x,\gamma^{n}x)}{n}.$ For any $y\in X,$ we have $\displaystyle d(x,\gamma^{n}x)$ $\displaystyle\leq$ $\displaystyle d(x,y)+d(y,\gamma^{n}y)+d(\gamma^{n}y,\gamma^{n}x)$ $\displaystyle=$ $\displaystyle 2d(x,y)+d(y,\gamma^{n}y)$ and thus $\lim_{n\rightarrow\infty}\frac{d(x,\gamma^{n}x)}{n}\leq\lim_{n\rightarrow\infty}\frac{d(y,\gamma^{n}y)}{n}.$ Similarly, we have the other direction $\lim_{n\rightarrow\infty}\frac{d(y,\gamma^{n}y)}{n}\leq\lim_{n\rightarrow\infty}\frac{d(x,\gamma^{n}x)}{n}.$ This shows that the definition of $l(\gamma)$ does not depend on the choice of $x.$ ###### Lemma 2.7 Let $G$ be a group acting isometrically on a metric space $X.$ Then the function $l:G\rightarrow[0,+\infty)$ defined by $g\longmapsto l(g)$ as above is a length function in the sense of Definition 1.1. Proof. This follows Lemma 2.1. ### 2.6 Translation lengths of isometries of CAT(0) spaces In this subsection, we will prove that the translation length on a CAT(0) space defines a length function. First, let us introduce some notations. Let $(X,d_{X})$ be a geodesic metric space, i.e. any two points $x,y\in X$ can be connected by a path $[x,y]$ of length $d_{X}(x,y)$. For three points $x,y,z\in X,$ the geodesic triangle $\Delta(x,y,z)$ consists of the three vertices $x,y,z$ and the three geodesics $[x,y],[y,z]$ and $[z,x].$ Let $\mathbb{R}^{2}$ be the Euclidean plane with the standard distance $d_{\mathbb{R}^{2}}$ and $\bar{\Delta}$ a triangle in $\mathbb{R}^{2}$ with the same edge lengths as $\Delta$. Denote by $\varphi:\Delta\rightarrow\bar{\Delta}$ the map sending each edge of $\Delta$ to the corresponding edge of $\bar{\Delta}.$ The space $X$ is called a CAT(0) space if for any triangle $\Delta$ and two elements $a,b\in\Delta,$ we have the inequality $d_{X}(a,b)\leq d_{\mathbb{R}^{2}}(\varphi(a),\varphi(b)).$ The typical examples of CAT(0) spaces include simplicial trees, hyperbolic spaces, products of CAT(0) spaces and so on. From now on, we assume that $X$ is a complete CAT(0) space. Denote by Isom$(X)$ the isometry group of $X.$ For any $g\in$ Isom$(X)$, let $\mathrm{Minset}(g)=\\{x\in X:d(x,gx)\leq d(y,gy)\text{ for any }y\in X\\}$ and let $\tau(g)=\inf\nolimits_{x\in X}d(x,gx)$ be the translation length of $g.$ When the fixed-point set $\mathrm{Fix}(g)\neq\emptyset,$ we call $g$ elliptic. When $\mathrm{Minset}(g)\neq\emptyset$ and $d_{X}(x,gx)=\tau(g)>0$ for any $x\in\mathrm{Minset}(g),$ we call $g$ hyperbolic. The group element $g$ is called semisimple if the minimal set $\mathrm{Minset}(g)$ is not empty, i.e. it is either elliptic or hyperbolic. A subset $C$ of a CAT(0) space if convex, if any two points $x,y\in C$ can connected by the geodesic segment $[x,y]\subset C.$ A group $G$ is called CAT(0) if $G$ acts properly discontinuously and cocompactly on a CAT(0) space $X$. In such a case, any infinite-order element in $G$ acts hyperbolically on $X.$ For more details on CAT(0) spaces, see the book of Bridson and Haefliger [16]. The following was proved by Ballmann-Gromov-Schroeder [2] (Lemma 6.6, page 83). The original proof was for Hardmard manifolds, which also holds for general cases. For completeness, we give details here. ###### Lemma 2.8 Let $\gamma:X\rightarrow X$ be an isometry of a complete CAT(0) space $X.$ For any $x_{0}\in X,$ we have $\tau(\gamma):=\inf_{x\in X}d(\gamma x,x)=\lim_{k\rightarrow\infty}\frac{d(\gamma^{k}x_{0},x_{0})}{k}.$ Proof. For any $p=x_{0}\in X,$ let $m$ be the middle point of $[p,\gamma p].$ We have that $d(m,\gamma m)\leq\frac{1}{2}d(p,\gamma^{2}p)$ by the convexity of length functions. Therefore, $d(p,\gamma^{2}p)\geq 2\tau(\gamma)$ and $\tau(\gamma^{2})\geq 2\tau(\gamma).$ Note that $d(p,\gamma^{2}p)\leq d(p,\gamma p)+d(\gamma p,\gamma^{2}p)=2d(p,\gamma p)$ and thus $\tau(\gamma^{2})\leq 2\tau(\gamma).$ Inductively, we have $2^{n}\tau(\gamma)\leq d(p,\gamma^{2^{n}}p)\leq 2^{n}d(p,\gamma p).$ Note that the limit $\lim_{k\rightarrow\infty}\frac{d(\gamma^{k}p,p)}{k}$ exists and is independent of $p$ (see the previous subsection). Therefore, the limit $\lim_{k\rightarrow\infty}\frac{d(\gamma^{k}p,p)}{k}$ equals to $\tau(\gamma).$ ###### Corollary 2.9 Let $X$ be a complete CAT(0) space and $G$ a group acting on $X$ by isometries. For any $g\in G,$ define $\tau(g)=\inf_{x\in X}d(x,gx)$ as the translation length. Then $\tau:G\rightarrow[0,+\infty)$ is a length function in the sense of Definition 1.1. Proof. This follows Lemma 2.8 and Lemma 2.7. ### 2.7 Translation lengths of Gromov $\delta$-hyperbolic spaces Let $\delta>0.$ A geodesic metric space $X$ is called Gromov $\delta$-hyperbolic if for any geodesic triangle $\Delta xyz$ one side $[x,y]$ is contained a $\delta$-neighborhood of the other two edges $[x,z]\cup[y,z].$ Fix $x_{0}\in X.$ Any isometry $\gamma:X\rightarrow X$ is called elliptic if $\\{\gamma^{n}x\\}_{ne\mathbb{Z}}$ is bounded. If the orbit map $\mathbb{Z}\rightarrow X$ given by $n\mapsto\gamma^{n}x_{0}$ is quasi- isometric (i.e. there exists $A\geq 1$ and $B\geq 0$ such that $\frac{1}{A}|n-m|-B\leq d_{X}(\gamma^{n}x_{0},\gamma^{m}x_{0})\leq A|n-m|+B$ for any integers $n,m$), we call $\gamma$ is hyperbolic. Otherwise, we call $\gamma$ is parabolic. Define $l(\gamma)=\lim_{n\rightarrow\infty}\frac{d(\gamma^{n}x_{0},x_{0})}{n}.$ For any group $G$ acts isometrically on a $\delta$-hyperbolic space, the function $l:G\rightarrow[0,\infty)$ is a length function by Lemma 2.7. A finitely generated group $G$ is Gromov $\delta$-hyperbolic if for some finite generating set $S,$ the Caley graph $\Gamma(G,S)$ is Gromov $\delta$-hyperbolic. Any infinite-order element $g$ in a Gromov $\delta$-hyperbolic group is hyperbolic and thus has positive length $l(g)>0$ (cf. [32], 8.1.D). For more details on hyperbolic spaces and hyperbolic groups, see the book [32] of Gromov. ### 2.8 Quasi-cocycles Let $G$ be a group and $(E,\|\cdot\|)$ be a normed vector space with an $G$-action by linear isometries. A function $f:G\rightarrow E$ is a quasi- cocyle if there exists $C>0$ such that $\|f(gh)-f(g)-gf(h)\|<C$ for any $g,h\in G.$ Let $l:G\rightarrow[0,+\infty)$ be defined by $l(g)=\lim_{n\rightarrow\infty}\frac{\|f(g^{n})\|}{n}.$ Note that $\|f(g^{n+m})\|\leq\|f(g^{n})\|+\|f(g^{m})\|+C$ for any integers $n,m\geq 0.$ This general subadditive property implies that the limit $\lim_{n\rightarrow\infty}\frac{\|f(g^{n})\|}{n}$ exists (see [56], Theorem 1.9.2, page 22). We call $l$ the average norm. Many applications of quasi- cocycles can be found in [49]. ###### Lemma 2.10 For any quasi-cocyle $f:G\rightarrow E,$ the average norm $l$ is a length function. Proof. For any natural number $n,$ we have $\|f(1)-f(g^{-n})-g^{-n}f(g^{n})\|<C$ and thus $\|\frac{f(1)-f(g^{-n})-g^{-n}f(g^{n})}{n}\|<\frac{C}{n}.$ Taking the limit, we have $\lim_{n\rightarrow\infty}\frac{\|f(g^{-n})\|}{n}=\lim_{n\rightarrow\infty}\frac{\|f(g^{n})\|}{n}.$ Therefore, for any $k\in\mathbb{Z},$ we have $l(g^{k})=\lim_{n\rightarrow\infty}\frac{\|f(g^{kn})\|}{n}=|k|l(g).$ For any $h\in G,$ we have $\|f(hg^{n}h^{-1})\|\leq\|f(h)\|+\|f(h^{-1})\|+\|f(g^{n})\|+2C.$ Therefore, we have $l(hgh^{-1})=\lim_{n\rightarrow\infty}\frac{\|f(hg^{n}h^{-1})\|}{n}\leq\lim_{n\rightarrow\infty}\frac{\|f(g^{n})\|}{n}=l(g).$ Similarly, we have $l(g)=l(h^{-1}(hgh^{-1})h)\leq l(hgh^{-1}).$ When $g,h$ commutes, we have $\displaystyle l(gh)$ $\displaystyle=$ $\displaystyle\lim_{n\rightarrow\infty}\frac{\|f((gh)^{n})\|}{n}=\lim_{n\rightarrow\infty}\frac{\|f(g^{n}h^{n})\|}{n}$ $\displaystyle\leq$ $\displaystyle\lim_{n\rightarrow\infty}\frac{\|f(g^{n})\|+\|f(h^{n})\|+C}{n}=l(g)+l(h).$ ### 2.9 Rotation number Let $\mathbb{R}$ be the real line and $\mathrm{Home}_{\mathbb{Z}}(\mathbb{R})=\\{f\mid f:\mathbb{R}\rightarrow\mathbb{R}$ is a monotonically increasing homeomorphism such that $f(x+n)=f(x)$ for any $n\in\mathbb{Z}\\}.$ For any $f\in\mathrm{Home}_{\mathbb{Z}}(\mathbb{R})$ and $x\in[0,1),$ the translation number is defined as $l(f)=\lim_{n\rightarrow\infty}\frac{f^{n}(x)-x}{n}.$ It is well-known that $l(f)$ exists and is independent of $x$ (see [51], Prop. 2.22, p.31). Note that every $f\in\mathrm{Home}_{\mathbb{Z}}(\mathbb{R})$ induces an orientation-preserving homeomorphism of the circle $S^{1}.$ ###### Proposition 2.11 The absolute value of the translation number $|l|:\mathrm{Home}_{\mathbb{Z}}(\mathbb{R})\rightarrow[0,\infty)$ is a length function in the sense of Definition 1.1. Proof. For any $f\in\mathrm{Home}_{\mathbb{Z}}(\mathbb{R})$ and $k\in\mathbb{Z}\backslash\\{0\\},$ we have that $l(f^{k})=\lim_{n\rightarrow\infty}\frac{f^{kn}(x)-x}{n}=k\lim_{n\rightarrow\infty}\frac{f^{kn}(x)-x}{nk}=kl(f).$ For any $a\in\mathrm{Home}_{\mathbb{Z}}(\mathbb{R}),$ we have that $\displaystyle\mid$ $\displaystyle l(afa^{-1})-l(f)\mid=\lim_{n\rightarrow\infty}\mid\frac{af^{n}(a^{-1}x)-x-f^{n}(x)+x}{n}\mid$ $\displaystyle=$ $\displaystyle\lim_{n\rightarrow\infty}\mid\frac{af^{n}(a^{-1}x)-f^{n}(a^{-1}x)+f^{n}(a^{-1}x)-f^{n}(x)}{n}\mid$ $\displaystyle=$ $\displaystyle 0,$ since $a$ is bounded on $[0,1]$ and $\mid f^{n}(a^{-1}x)-f^{n}(x)\mid\leq 2+|a^{-1}x-x|.$ For commuting elements $f,g\in\mathrm{Home}_{\mathbb{Z}}(\mathbb{R}),$ we have that $l(fg)=\lim_{n\rightarrow\infty}\frac{f^{n}(g^{n}(x))-x}{n}=\lim_{n\rightarrow\infty}\frac{f^{n}(g^{n}(x))-g^{n}(x)+g^{n}(x)-x}{n}.$ Suppose that $g^{n}(x)=k_{n}+x_{n}$ for $k_{n}\in\mathbb{Z}$ and $x_{n}\in[0,1).$ Then $\displaystyle\lim_{n\rightarrow\infty}\frac{f^{n}(g^{n}x)-g^{n}(x)+g^{n}(x)-x}{n}$ $\displaystyle=$ $\displaystyle\lim_{n\rightarrow\infty}\frac{f^{n}(0)-0+g^{n}(x)-x}{n}=l(f)+l(g).$ Therefore, we get $|l(fg)|\leq|l(f)|+|l(g)|.$ ###### Remark 2.12 It is actually true that the rotation number $l$ is multiplicative on any amenable group (see [51], Prop. 2.2.11 and the proof of Prop. 2.2.10, page 36). This implies that the absolute rotation number $|l|$ is subadditive on any amenable group. In other words, for any amenable group $G<\mathrm{Home}_{\mathbb{Z}}(\mathbb{R})$ and any $g,h\in G$ we have $|l(gh)|\leq|l(g)|+|l(h)|.$ ### 2.10 Asymptotic distortions Let $f$ be a $C^{1+bv}$ diffeomorphism of the closed interval $[0,1]$ or the circle $S^{1}.$ (“bv” means derivative with finite total variation.) The asymptotic distortion of $f$ is defined as $l(f)=\mathrm{dist}_{\infty}(f)=\lim_{n\rightarrow\infty}\mathrm{var}(\log Df^{n}).$ It’s proved by Eynard-Bontemps amd Navas ([24], pages 7-8) that 1. (1) $\mathrm{dist}_{\infty}(f^{n})=|n|\mathrm{dist}_{\infty}(f)$ for all $n\in\mathbb{Z}$; 2. (2) $\mathrm{dist}_{\infty}(hfh^{-1})=\mathrm{dist}_{\infty}(f)$ for every $C^{1+bv}$ diffeomorphism $h;$ 3. (3) $\mathrm{dist}_{\infty}(f\circ g)\leq\mathrm{dist}_{\infty}(f)+\mathrm{dist}_{\infty}(g)$ for commuting $f,g.$ Therefore, the asymptotic distortion is a length function $l$ on the group $\mathrm{Diff}^{1+bv}(M)$ of $C^{1+bv}$ diffeomorphisms for $M=[0,1]$ or $S^{1}.$ ### 2.11 Dynamical degrees of Cremona groups Let $k$ be a field and $\mathbb{P}_{k}^{n}=k^{n+1}\backslash\\{0\\}/\\{\lambda\sim\lambda x:\lambda\neq 0\\}$ be the projective space. A rational map from $\mathbb{P}_{k}^{n}$ to itself is a map of the following type $(x_{0}:x_{1}:\cdots:x_{n})\dashrightarrow(f_{0}:f_{1}:\cdots:f_{n})$ where the $f_{i}$’s are homogeneous polynomials of the same degree without common factor. The degree of $f$ is $\deg f=\deg f_{i}.$ A birational map from $\mathbb{P}_{k}^{n}$ to itself is a rational map $f:\mathbb{P}_{k}^{n}\dashrightarrow\mathbb{P}_{k}^{n}$ such that there exists a rational map $g:\mathbb{P}_{k}^{n}\dashrightarrow\mathbb{P}_{k}^{n}$ such that $f\circ g=g\circ f=\mathrm{id}.$ The group $\mathrm{Bir}(\mathbb{P}_{k}^{n})$ of birational maps is called the Cremona group (also denoted as $\mathrm{Cr}_{n}(k)$). It is well-known that $\mathrm{Bir}(\mathbb{P}_{k}^{n})$ is isomorphic to the group $\mathrm{Aut}_{k}(k(x_{1},x_{2},\cdots,x_{n}))$ of self-isomorphisms of the field $k(x_{1},x_{2},\cdots,x_{n})$ of the rational functions in $n$ indeterminates over $k.$ The (first) dynamical degree $\lambda(f)$ of $f\in\mathrm{Bir}(\mathbb{P}_{k}^{n})$ is defined as $\lambda(f)=\max\\{\lim_{n\rightarrow\infty}\deg(f^{n})^{\frac{1}{n}},\lim_{n\rightarrow\infty}\deg(f^{-n})^{\frac{1}{n}}\\}.$ Since $\deg(f^{n})^{\frac{1}{n}}$ is sub-multiplicative, the limit exists. ###### Lemma 2.13 Let $l(f)=\log\lambda(f).$ Then $l:\mathrm{Bir}(\mathbb{P}_{k}^{n})\rightarrow[0,+\infty)$ is a length function. Proof. Without loss of generality, we assume that $\lambda(f)=\lim_{n\rightarrow\infty}\deg(f^{n})^{\frac{1}{n}},$ while the other case can be considered similarly. For any $k\in\mathbb{N},$ it is easy that $l(f^{k})=\lim_{n\rightarrow\infty}\frac{\log\deg f^{nk}}{n}=kl(f).$ For any $h\in\mathrm{Bir}(\mathbb{P}_{k}^{n}),$ we have $l(hfh^{-1})=\lim_{n\rightarrow\infty}\frac{\log\deg hf^{n}h^{-1}}{n}=\lim_{n\rightarrow\infty}\frac{\log\deg f^{nk}}{n}=l(f).$ For commuting maps $f,g,$ we have $(fg)^{n}=f^{n}g^{n}.$ Therefore, $\displaystyle l(fg)$ $\displaystyle=$ $\displaystyle\lim_{n\rightarrow\infty}\frac{\log\deg f^{n}g^{n}}{n}$ $\displaystyle\leq$ $\displaystyle\lim_{n\rightarrow\infty}\frac{\log\deg f^{n}}{n}+\lim_{n\rightarrow\infty}\frac{\log\deg g^{n}}{n}=l(f)+l(g).$ This checks the three conditions of the length function. It is surprising that when $n=2$ and $k$ is an algebraically closed field, the length function $l(f)$ is given by the translation length $\tau(f)$ on an (infinite-dimensional) Gromov $\delta$-hyperbolic space (see Blanc-Cantat [8], Theorem 4.4). Some other length functions are studied by Blanc and Furter [9] for groups of birational maps, eg. dynamical number of base-points and dynamical length. ## 3 Groups with purely positive length functions ###### Definition 3.1 A length function $l$ on a group $G$ is said to be purely positive if $l(g)>0$ for any infinite-order element $g\in G.$ In this section, we show that the (Gromov) hyperbolic group, mapping class group and outer automorphism groups of free groups have purely positive length functions. First, let us recall the relevant definitions. A geodesic metric space $X$ is $\delta$-hyperbolic (for some real number $\delta>0$) if for any geodesic triangle $\Delta xyz$ in $X,$ one side is contained the $\delta$-neighborhood of the other two sides. A group $G$ is (Gromov) hyperbolic if $G$ acts properly discontinuously and cocompactly on a $\delta$-hyperbolic space $X$. ###### Definition 3.2 1. (i) An element $g$ in a group $G$ is called primitive if it cannot be writen as a proper power $\alpha^{n},$ where $\alpha\in G$ and $|n|\geq 2;$ 2. (ii) A group $G$ has unique-root property if every infinite-order element $g$ is a proper power of a unique (up to sign) primitive element, i.e. $g=\gamma^{n}=\gamma_{1}^{m}$ for primitive elements $\gamma,\gamma_{1}$ will imply $\gamma=\gamma^{\pm}.$ The following fact is well-known. ###### Lemma 3.3 A torsion-free hyperbolic group has unique-root property. Proof. Let $G$ be a torsion-free hyperbolic group and $1\neq g\in G.$ Suppose that $g=\gamma^{n}=\gamma_{1}^{m}$ for primitive elements $\gamma$ and $\gamma_{1}.$ The set $C_{G}(g)$ of centralizers is virtually cyclic (cf. [16], Corollary 3.10, page 462). By a result of Serre, a torsion-free virtually free group is free. Since $G$ is torsion-free, the group $C_{G}(g)$ is thus free and thus cyclic, say generated by $t$. Since $\gamma$ and $\gamma^{\prime}$ are primitive, they are $t^{\pm}.$ ###### Remark 3.4 The previous lemma does not hold for general hyperbolic groups with torsions. For example, let $G=\mathbb{Z}/2\times\mathbb{Z}.$ We have $(0,2)=(0,1)^{2}=(1,1)^{2}$ and $(0,1),$ $(1,1)$ are both primitive. For a group $G,$ let $P(G)$ be the set of all primitive elements. We call two primitive elements $\gamma,\gamma^{\prime}$ are general conjugate if there exists $g\in G$ such that $g\gamma g^{-1}=\gamma^{\prime}$ or $g\gamma^{-1}g=\gamma^{\prime}$. Let $\mathrm{CP}(G)$ be the general conjugacy classes of primitive elements. For a set $S,$ let $S_{\mathbb{R}}$ be the set of all real functions on $S$. The convex polyhedral cone spanned by $S$ is the subset $\\{\mathop{\textstyle\sum}\nolimits_{s\in S}a_{s}s\mid a_{s}\geq 0\ \\}\subset S_{\mathbb{R}}.$ ###### Lemma 3.5 Let $G$ be a torsion-free hyperbolic group. The set of all length functions on $G$ is the convex polyhedral cone spanned by the general conjugacy classes $\mathrm{CP}(G).$ Proof. Let $l$ be a length function on $G.$ Then $l$ gives an element $\mathop{\textstyle\sum}\nolimits_{s\in S}a_{s}s$ in the convex polyhedral cone by $a_{s}=l(s).$ Conversely, for any general conjugacy classes $[s]\in\mathrm{CP}(G)$ with $s$ a primitive element, let $l_{s}$ be the function defined by $l_{s}(s^{\pm})=1$ and $l_{s}(\gamma)=0$ for element $\gamma$ in any other general conjugacy classes. For any $1\neq g\in G,$ there is a unique (up to sign) primitive element $\gamma$ such that $g=\gamma^{n}.$ Define $l_{s}(g)=|n|l_{s}(\gamma).$ Then $l_{s}$ satisfies conditions (1) and (2) in Definition 1.1. The condition (3) is satisfied automatically, since any commuting pair of elements $a,b$ generate a cyclic group in a torsion-free hyperbolic group. Any element $\mathop{\textstyle\sum}\nolimits_{s\in S}a_{s}s$ gives a length function on $G$ as a combination of $a_{s}l_{s}$. ###### Lemma 3.6 Let $G\ $be one of the following groups: * • automorphism group $\mathrm{Aut}(F_{k})$ of a free group; * • outer automorphism group $\mathrm{Out}(F_{k})$ of a free group or * • mapping class group $\mathrm{MCG}(\Sigma_{g,m})$ (where $\Sigma_{g,m}$ is an oriented surface of genus $g$ and $m$ punctures); * • a hyperbolic group, * • a $\mathrm{CAT}(0)$ group or more generally * • a semi-hyperbolic group, * • a group acting properly semi-simply on a $\mathrm{CAT}(0)$ space, * • a group acting properly semi-simply on a $\delta$-hyperbolic space. Then $G$ has a purely positive length function. Proof. Note that hyperbolic groups and $\mathrm{CAT}(0)$ groups are semihyperbolic (see [16], Prop. 4.6 and Cor. 4.8, Chapter III.$\Gamma$). For a semihyperbolic group $G$ acting a metric space $X$ (actually $X=G$), the translation $\tau$ is a length function by Lemma 2.7. Moreover, for any infinite-order element $g\in G,$ the length $\tau(g)>0$ (cf. [16], Lemma 4.18, page 479). For group acting properly semisimply on a $\mathrm{CAT}(0)$ space (or a $\delta$-hyperbolic space), the translation $l(\gamma)=\lim_{n\rightarrow\infty}\frac{d(x,\gamma^{n}x)}{n}$ is a length function (cf. Lemma 2.7). For any hyperbolic $\gamma$, we get $l(\gamma)>0.$ For any elliptic $\gamma,$ it is finite-order since the action is proper. Alibegovic [1] proves that the stable word length of $\mathrm{Aut}(F_{n}),\mathrm{Out}(F_{n})$ are purely positive. Farb, Lubotzky and Minsky [25] prove that Dehn twists and more generally all elements of infinite order in $\mathrm{MCG}(\Sigma_{g,m})$ have positive translation length. ###### Definition 3.7 A group $G$ is called poly-positive (or has a poly-positive length), if there is a subnormal series $1=H_{n}\vartriangleleft H_{n-1}\vartriangleleft\cdots\vartriangleleft H_{0}=G$ such that every finitely generated subgroup of the quotient $H_{i}/H_{i+1}$ $(i=0,...,n-1)$ has a purely positive length function. Recall that a group $G$ is poly-free, if there is a subnormal series $1=H_{n}\vartriangleleft H_{n-1}\vartriangleleft\cdots\vartriangleleft H_{0}=G$ such that the successive quotient $H_{i}/H_{i+1}$ is free $(i=0,...,n-1).$ Since a free group is hyperbolic, it has a purely positive length function. This implies that a poly-free group is poly-positive. A group is said to have a virtual property if there is a finite-index subgroup has the property. Let $\Sigma$ be a closed oriented surfaec endowed with an area form $\omega.$ Denote by $\mathrm{Diff}(\Sigma,\omega)$ the group of diffeomorphisms preserving $\omega$ and $\mathrm{Diff}_{0}(\Sigma,\omega)$ the subgroup consisting of diffeomorphisms isotopic to the identity. ###### Lemma 3.8 When the genuse of $\Sigma$ is greater than $1,$ the group $\mathrm{Diff}_{0}(\Sigma,\omega)$ and $\mathrm{Diff}(\Sigma,\omega)$ is poly- positive. Proof. This is eseentially proved by Py [54] (Section 1). There is a group homomorphism $\alpha:\mathrm{Diff}_{0}(\Sigma,\omega)\rightarrow H_{1}(\Sigma,\mathbb{R})$ with $\ker\alpha=\mathrm{Ham}(\Sigma,\omega)$ the group of Hamiltonian diffeomorphisms of $\Sigma.$ Polterovich [53] (1.6.C.) proves that any finitely generated group of $\mathrm{Ham}(\Sigma,\omega)$ has a purely positive stable word length. Since the quotient group $\mathrm{Diff}(\Sigma,\omega)/\mathrm{Diff}_{0}(\Sigma,\omega)$ is a subgroup of the mapping class group $\mathrm{MCG}(\Sigma),$ which has a purely positive stable word length by Farb-Lubotzky-Minsky [25], the group $\mathrm{Diff}(\Sigma,\omega)$ is poly-positive. ## 4 Vanishing of length functions on abelian-by-cyclic groups We will need the following result proved in [28]. ###### Lemma 4.1 Given a group $G$, let $l:G\rightarrow[0,+\infty)$ be function such that 1) $l(e)=0;$ 2) $l(x^{n})=|n|l(x)$ for any $x\in G,$ any $n\in\mathbb{Z};$ 3) $l(xy)\leq l(x)+l(y)$ for any $x,y\in G.$ Then there exist a real Banach space $(\mathbb{B},\|\|)$ and a group homomorphism $\varphi:G\rightarrow\mathbb{B}$ such that $l(x)=\|\varphi(x)\|$ for all $x\in G.$ Further more, if $l(x)>0$ for any $x\neq e$, one can take $\varphi$ to be injective, i.e., an isometric embedding. Let $\mathbb{Z}^{2}\rtimes_{A}\mathbb{Z}$ be an abelian by cyclic group, where $A=\begin{bmatrix}a&b\\\ c&d\end{bmatrix}\in\mathrm{GL}_{2}(\mathbb{Z})$. We prove Theorem 0.1 by proving the following two theorems. ###### Theorem 4.2 When the absolute value of the trace $|\mathrm{tr}(A)|>2,$ any length function $l:\mathbb{Z}^{2}\rtimes_{A}\mathbb{Z}\rightarrow\mathbb{R}_{\geq 0}$ vanishes on $\mathbb{Z}^{2}.$ Proof. Let $A=\begin{bmatrix}a&b\\\ c&d\end{bmatrix}\in\mathrm{GL}_{2}(\mathbb{Z}).$ Suppose that $t$ is a generator of $\mathbb{Z}$ and $t\begin{bmatrix}x\\\ y\end{bmatrix}t^{-1}=A\begin{bmatrix}x\\\ y\end{bmatrix}$ for any $x,y\in\mathbb{Z}.$ Note that $(0,t^{k})(v,0)(0,t^{k})^{-1}=(A^{k}v,0)$ for any $v\in\mathbb{Z}^{2}$ and $k\in\mathbb{Z}$. Therefore, an element $v\in\mathbb{Z}^{2}$ is conjugate to $A^{k}v$ for any integer $k.$ Note that $A\begin{bmatrix}1\\\ 0\end{bmatrix}=\begin{bmatrix}a\\\ c\end{bmatrix},A^{2}\begin{bmatrix}1\\\ 0\end{bmatrix}=\begin{bmatrix}a^{2}+bc\\\ ac+dc\end{bmatrix}$ and $\begin{bmatrix}a^{2}+bc\\\ ac+dc\end{bmatrix}=(a+d)\begin{bmatrix}a\\\ c\end{bmatrix}-(ad-bc)\begin{bmatrix}1\\\ 0\end{bmatrix}.$ Therefore, we have $\displaystyle|a+d|l(\begin{bmatrix}1\\\ 0\end{bmatrix})$ $\displaystyle=$ $\displaystyle l((a+d)\begin{bmatrix}a\\\ c\end{bmatrix})$ $\displaystyle=$ $\displaystyle l(\begin{bmatrix}a^{2}+bc\\\ ac+dc\end{bmatrix}+(ad- bc)\begin{bmatrix}1\\\ 0\end{bmatrix})$ $\displaystyle\leq$ $\displaystyle(1+|ad-bc|)l(\begin{bmatrix}1\\\ 0\end{bmatrix}).$ When $ad-bc=\pm 1$ and $|a+d|>2,$ we must have $l(\begin{bmatrix}1\\\ 0\end{bmatrix})=0.$ Similarly, we can prove that $l(\begin{bmatrix}0\\\ 1\end{bmatrix})=0.$ Since $l$ is subadditive on $\mathbb{Z}^{2},$ we get that $l$ vanishes on $\mathbb{Z}^{2}.$ ###### Theorem 4.3 When the absolute value $|\mathrm{tr}(A)|=|a+d|=2,I_{2}\neq A\in\mathrm{SL}_{2}(\mathbb{Z}),$ any length function $l:\mathbb{Z}^{2}\rtimes_{A}\mathbb{Z}\rightarrow\mathbb{R}_{\geq 0}$ vanishes on the direct summand of $\mathbb{Z}^{2}$ spanned by eigenvectors of $A$. Proof. We may assume that $A=\begin{bmatrix}1&n\\\ 0&1\end{bmatrix},n\neq 0.$ For any integer $k\geq 0$ and $v\in\mathbb{Z}^{2},$ we have $t^{k}vt^{-k}=A^{k}v.$ Take $v=\begin{bmatrix}0\\\ 1\end{bmatrix}$ to get that $t^{k}\begin{bmatrix}0\\\ 1\end{bmatrix}t^{-k}=\begin{bmatrix}kn\\\ 0\end{bmatrix}+\begin{bmatrix}0\\\ 1\end{bmatrix}.$ Since the function $l|_{\mathbb{Z}^{2}}$ is given by the norm of a Banach space according to Lemma 4.1, we get that $\displaystyle k|n|l(\begin{bmatrix}1\\\ 0\end{bmatrix})$ $\displaystyle\leq$ $\displaystyle l(t^{k}\begin{bmatrix}0\\\ 1\end{bmatrix}t^{-k})+l(\begin{bmatrix}0\\\ 1\end{bmatrix})$ $\displaystyle=$ $\displaystyle 2l(\begin{bmatrix}0\\\ 1\end{bmatrix}).$ Since $k$ is arbitrary, we get that $l(\begin{bmatrix}1\\\ 0\end{bmatrix})=0.$ ###### Remark 4.4 When $A=\begin{bmatrix}1&1\\\ 0&1\end{bmatrix},$ the semidirect product $G=\mathbb{Z}^{2}\rtimes_{A}\mathbb{Z}$ is a Heisenberg group. A length function on $G/Z(G)\cong\mathbb{Z}^{2}$ gives a length function on $G.$ In particular, a length function of $G$ may not vanish on the second component $\begin{bmatrix}0\\\ 1\end{bmatrix}\in\mathbb{Z}^{2}<G.$ ###### Remark 4.5 When $A\in\mathrm{SL}_{2}(\mathbb{Z})$ has $|\mathrm{tr}(A)|<2,$ the matrix $A$ is of finite order and the semi-direct product $\mathbb{Z}^{2}\rtimes_{A}\mathbb{Z}$ contains $\mathbb{Z}^{3}$ as a finite- index normal subgroup. Actually, in this case the group $\mathbb{Z}^{2}\rtimes_{A}\mathbb{Z}$ is the fundamental group of a flat $3$-manifold $M$ (see [61], Theorem 3.5.5). Therefore, the group $\mathbb{Z}^{2}\rtimes_{A}\mathbb{Z}$ acts freely properly discontinuously isometrically and cocompactly on the universal cover $\tilde{M}=\mathbb{R}^{3}.$ This means the translation length gives a purely positive length function on $\mathbb{Z}^{2}\rtimes_{A}\mathbb{Z}$. ###### Lemma 4.6 Let $A\in\mathrm{GL}_{n}(\mathbb{Z})$ be a matrix and $G=\mathbb{Z}^{n}\rtimes_{A}\mathbb{Z}$ the semi-direct product. Let $\sum_{i=0}^{n}a_{i}x^{i}$ be the characterisitic polynomial of some power $A^{k}$. Suppose that for some $k,$ there is a coefficient $a_{i}$ such that $|a_{i}|>\sum_{j\neq i}|a_{j}|.$ Any length function $l$ of $G$ vanishes on $\mathbb{Z}^{n}.$ Proof. Let $t$ be a generator of $Z$ and $tat^{-1}=Aa$ for any $a\in Z^{n}.$ Note that for any integer $m,$ we have $t^{m}at^{-m}=A^{m}a$ and $l(a)=l(A^{m}a).$ Note that $\sum_{i=0}^{n}a_{i}A^{ki}=0$ and thus $\displaystyle\sum_{i=0}^{n}a_{i}A^{ki}a$ $\displaystyle=$ $\displaystyle 0,$ $\displaystyle l(\sum_{i=0}^{n}a_{i}A^{ki}a)$ $\displaystyle=$ $\displaystyle 0$ for any $a\in\mathbb{Z}^{n}.$ Therefore, $|a_{i}|l(a)=|a_{i}|l(A^{ki}a)=l(\sum_{j\neq i}a_{j}A^{kj}a)\leq\sum_{j\neq i}|a_{j}|l(a).$ This implies that $l(a)=0.$ Proof of Corollary 0.2. When the group action is $C^{2},$ define $l(f)=\max\\{\lim_{n\rightarrow+\infty}\frac{\log\sup_{x\in M}\|D_{x}f^{n}\|}{n},\lim_{n\rightarrow+\infty}\frac{\log\sup_{x\in M}\|D_{x}f^{-n}\|}{n}\\}$ for any diffeomorphism $f:M\rightarrow M.$ Lemma 2.5 shows that $l$ is a length function, which is an upper bound of the Lyapunov exponents. When the group action is Lipschitz, define $L(f)=\sup_{x\neq y}\frac{d(fx,fy)}{d(x,y)}$ for a Lipschitz-homeomorphism $f:M\rightarrow M.$ Since $L(fg)\leq L(f)L(g)$ for two Lipschitz-homeomorphisms $f,g:M\rightarrow M,$ we have that $l(f):=\lim_{n\rightarrow\infty}\frac{\max\\{\log(L(f^{n})),\log(L(f^{-n}))\\}}{n}$ gives a length function by Lemma 2.1. Note that $l(f)\geq h_{top}(f)$ (see [40], Theorem 3.2.9, page 124). The vanishings of the topological entropy $h_{top}$ and the Lyapunov exponents in Corolary 0.2 are proved by Theorem 0.1 considering these length functions. ## 5 Classification of length functions on nilpotent groups The following lemma is a key step for our proof of the vanishing of length functions on Heisenberg groups. ###### Lemma 5.1 Let $G=\langle a,b,c\mid aba^{-1}b^{-1}=c,ac=ca,bc=cb\rangle$ be the Heisenberg group. Suppose that $f:G\rightarrow\mathbb{R}$ is a conjugation-invariant function, i.e. $f(xgx^{-1})=f(g)$ for any $x,g\in G.$ For any coprime integers (not-all-zero) $m,n$ and any integer $k,$ we have $f(a^{m}b^{n}c^{k})=f(a^{m}b^{n}).$ Proof. It is well-known that for any integers $n,m,$ we have $[a^{n},b^{m}]=c^{nm}.$ Actually, since $aba^{-1}b^{-1}=c,$ we have $ba^{-1}b^{-1}=a^{-1}c$ and thus $ba^{-n}b^{-1}=a^{-n}c^{n}$ for any integer $n.$ Therefore, $a^{n}ba^{-n}b^{-1}=c^{n}$ and $a^{n}ba^{-n}=c^{n}b,$ $a^{n}b^{m}a^{-n}=a^{nm}b^{m}$ for any integer $m.$ This means $[a^{n},b^{m}]=c^{nm}.$ For any coprime $m,n$, and any integer $k,$ let $s,t\in\mathbb{Z}$ such that $ms+nt=k.$ We have $a^{-m}b^{-s}a^{m}b^{s}=c^{ms},b^{-s}a^{m}b^{s}=a^{m}c^{ms},b^{-s}a^{m}b^{n}b^{s}=a^{m}b^{n}c^{ms}$ and $a^{t}b^{n}a^{-t}b^{-n}=c^{nt},a^{t}b^{n}a^{-t}=b^{n}c^{nt},a^{t}a^{m}b^{n}a^{-t}=a^{m}b^{n}c^{nt}.$ Therefore, $a^{t}(b^{-s}a^{m}b^{n}b^{s})a^{-t}=a^{t}(a^{m}b^{n}c^{ms})a^{-t}=a^{m}b^{n}c^{nt+ms}=a^{m}b^{n}c^{k}.$ When $f$ is conjugation-invariant, we get $f(a^{m}b^{n}c^{k})=f(a^{m}b^{n})$ for any coprime $m,n$, and any integer $k.$ ###### Lemma 5.2 Let $G=\langle a,b,c\mid aba^{-1}b^{-1}=c,ac=ca,bc=cb\rangle$ be the Heisenberg group. Any length function $l:G\rightarrow[0,\infty)$ (in the sense of Definition 1.1) factors through the abelization $G_{\mathrm{ab}}:=G/[G,G]\cong\mathbb{Z}^{2}.$ In other words, there is a function $l^{\prime}:G_{\mathrm{ab}}\rightarrow[0,\infty)$ such that $l^{\prime}(x^{n})=|n|l^{\prime}(x)$ for any $x\in G_{\mathrm{ab}},$ any integer $n$ and $l=l^{\prime}\circ q$, where $q:G\rightarrow G_{\mathrm{ab}}$ is the natural quotient group homomorphism. Proof. Let $H=\langle c\rangle\cong\mathbb{Z}$ and write $G=\cup gH$ the union of left cosets. We choose the representative $g_{ij}=a^{i}b^{j}$ with $(i,j)\in\mathbb{Z}^{2}.$ Note that the subgroup $\langle g_{ij},c\rangle$ generated by $g_{ij},c$ is isomorphic to $\mathbb{Z}^{2}$ for coprime $i,j.$ The length function $l$ is subadditive on $\langle g_{ij},c\rangle.$ By Lemma 4.1, there is a Banach space $\mathbb{B}$ and a group homomorphism $\varphi:$ $\langle g_{ij},c\rangle\rightarrow\mathbb{B}$ such that $l(g)=\|\varphi(g)\|$ for any $g\in$ $\langle g_{ij},c\rangle.$ Lemma 5.1 implies that $\|\varphi(g_{ij})+\varphi(c^{k})\|=\|\varphi(g_{ij})\|$ for any integer $k.$ Since $\|\varphi(g_{ij})\|=\|\varphi(g_{ij})+\varphi(c^{k})\|\geq|k|\|\varphi(c)\|-\|\varphi(g_{ij})\|$ for any $k,$ we have that $\|\varphi(c)\|=l(c)=0.$ This implies that $l(g_{ij}^{n}c^{m})=\|\varphi(g_{ij}^{n})+\varphi(c^{m})\|=l(g_{ij}^{n})$ for any integers $m,n.$ Moreover, for any integers $m,n$ and coprime $i,j,$ we have $a^{ni}b^{nj}c^{m}=(a^{i}b^{j})^{n}c^{k}$ for some integer $k.$ This implies that $l(a^{ni}b^{nj}c^{m})=l((a^{i}b^{j})^{n})=|n|l(a^{i}b^{j}).$ Therefore, the function $l$ is constant on each coset $gH.$ Define $l^{\prime}(gH)=l(g).$ Since $l(g^{k})=|l|l(g),$ we have that $l^{\prime}(g^{k}H)=|k|l^{\prime}(gH).$ The proof is finished. Denote by $S^{\prime}=\\{(m,n)\mid m,n$ are coprime integers$\\}$ be the set of coprime integer pairs and define an equivalence relation by $(m,n)\sim(m^{\prime},n^{\prime})$ if $(m,n)=\pm(m^{\prime},n^{\prime})$. Let $S=S^{\prime}/\sim$ be the equivalence classes. ###### Theorem 5.3 Let $G=\langle a,b,c\mid aba^{-1}b^{-1}=c,ac=ca,bc=cb\rangle$ be the Heisenberg group. The set of all length functions $l:G\rightarrow[0,\infty)$ (in the sense of Definition 1.1) is the convex polyhedral cone $\mathbb{R}_{\geq 0}[S]=\\{\sum_{s\in S}a_{s}s\mid a_{s}\in\mathbb{R}_{\geq 0},s\in S\\}.$ Proof. Similar to the proof of the previous lemma, we let $H=\langle c\rangle\cong\mathbb{Z}$ and write $G=\cup gH$ the union of left cosets. We choose the representative $g_{ij}=a^{i}b^{j}$ with $(i,j)\in\mathbb{Z}^{2}.$ Let $T$ be the set of length function $l:G\rightarrow[0,\infty).$ For any length functions $l,$ let $\varphi(l)=\sum_{s\in S}a_{s}s\in\mathbb{R}_{\geq 0}[S],$ where $a_{s}=l(a^{i}b^{j})$ with $(i,j)$ a representative of $s.$ Note that $l(a^{-i}b^{-j})=l(b^{-j}a^{-i}c^{ij})=l(b^{-j}a^{-i})=l(a^{i}b^{j}),$ which implies that $a_{s}$ is well-defined. We have defined a function $\varphi:T\rightarrow\mathbb{R}_{\geq 0}[S].$ If $\varphi(l_{1})=\varphi(l_{2})$ for two functions $l_{1},l_{2},$ then $l_{1}(a^{i}b^{j})=l_{2}(a^{i}b^{j})$ for coprime integers $i,j.$ Since both $l_{1},l_{2}$ are conjugation-invariant, Lemma 5.1 implies that $l_{1},l_{2}$ coincide on any coset $a^{i}b^{j}H$ and thus on the whole group $G.$ This proves the injectivity of $\varphi.$ For any $\sum_{s\in S}a_{s}s,$ we define a function $l:G=\cup a^{i}b^{j}H\rightarrow[0,\infty).$ For any coprime integers $i,j,$ define $l(a^{i}b^{j}z)=a_{s}$ for any representative $(i,j)$ of $s$ and any $z\in H.$ For any general integers $m,n$ and $z\in H,$ define $l(a^{m}b^{n}z)=l(a^{m}b^{n})=|\gcd(m,n)|l(a^{m/\gcd(m,n)}b^{n/\gcd(m,n)})$ and $l(z)=0.$ From the definition, it is obvious that $l$ is homogeous. Note that any element of $G$ is of the form $a^{k}b^{s}c^{t}$ for integers $k,s,t\in\mathbb{Z}.$ For any two elements $a^{k}b^{s}c^{t},a^{k^{\prime}}b^{s^{\prime}}c^{t^{\prime}}$ we have the conjugation $a^{k^{\prime}}b^{s^{\prime}}c^{t^{\prime}}a^{k}b^{s}c^{t}(a^{k^{\prime}}b^{s^{\prime}}c^{t^{\prime}})^{-1}=a^{k}b^{s}c^{t^{\prime\prime}}$ for some $t^{\prime\prime}\in\mathbb{Z}.$ Therefore, we see that $l$ is conjugation-invariant. The previous equality also shows that two elements $g,h$ are commuting if and only if they lies simutanously in $\langle a^{i}b^{j},c\rangle$ for a pair of coprime integers $i,j.$ By construction, we have $l(g)=l(h).$ This proves the surjectivity of $\varphi.$ ## 6 Length functions on matrix groups In this section, we study length functions on matrix groups $\mathrm{SL}_{n}(\mathbb{R})$. As the proofs are elementary, we present here in a separated section, without using profound results on Lie groups and algebraic groups. The following lemma is obvious. ###### Lemma 6.1 Let $G_{p,q}=\langle x,t:tx^{p}t^{-1}=x^{q}\rangle$ be a Baumslag-Solitar group. When $|p|\neq|q|,$ any length function $l$ on $G$ has $l(x)=0.$ Proof. Note that $|p|l(x)=l(x^{p})=l(x^{q})=|q|l(x),$ which implies $l(x)=0.$ Let $V^{n}$ be a finite-dimensional vector space over a field $K$ and $A:V\rightarrow V$ a unipotent linear transformation (i.e. $A^{k}=0$ for some positive integer $k).$ The following fact is from linear algebra (see the Lemma of page 313 in [4]. Since the reference is in Chinese, we repeat the proof here). ###### Lemma 6.2 $I+A$ is conjugate to a direct sum of Jordan blocks with $1s$ along the diagonal. Proof. We prove that $V$ has a basis $\\{a_{1},Aa_{1},\cdots,A^{k_{1}-1}a_{1},a_{2},Aa_{2},\cdots,A^{k_{2}-1}a_{2},\cdots,a_{s},\cdots,Aa_{s},\cdots,A^{k_{s}-1}a_{s}\\}$ satisfying $A^{k_{i}}a_{i}=0$ for each $i,$ which implies that the representation matrix of $I+A$ is similar to a direct sum of Jordan blocks with $1$ along the diagonal. The proof is based on the induction of $\dim V.$ When $\dim V=1,$ choose $0\neq v\in V.$ Suppose that $Av=\lambda v.$ Then $A^{k}v=\lambda^{k}v=0$ and thus $\lambda=0.$ Suppose that the case is proved for vector spaces of dimension $k<n.$ Note that the invariant subspace $AV$ is a proper subspace of $V$ (otherwise, $AV=V$ implies $A^{k}V=A^{k-1}V=V=0$). By induction, the subspace $AV$ has a basis $\\{a_{1},Aa_{1},\cdots,A^{k_{1}-1}a_{1},a_{2},Aa_{2},\cdots,A^{k_{2}-1}a_{2},\cdots,a_{s},\cdots,Aa_{s},\cdots,A^{k_{s}-1}a_{s}\\}.$ Choose $b_{i}\in V$ satisfying $A(b_{i})=a_{i}.$ Then $A$ maps the set $\\{b_{1},Ab_{1},\cdots,A^{k_{1}}b_{1},b_{2},Ab_{2},\cdots,A^{k_{2}}b_{2},b_{s},\cdots,Ab_{s},\cdots,A^{k_{s}}b_{s}\\}$ to the basis $\\{a_{1},Aa_{1},\cdots,A^{k_{1}-1}a_{1},a_{2},Aa_{2},\cdots,A^{k_{2}-1}a_{2},\cdots,a_{s},\cdots,Aa_{s},\cdots,A^{k_{s}-1}a_{s}\\}.$ This implies that the former set is linearly independent (noting that $A(A^{k_{i}}b_{i})=0$). Extend this set to be a $V^{\prime}s$ basis $\\{b_{1},Ab_{1},\cdots,A^{k_{1}-1}b_{1},b_{2},Ab_{2},\cdots,A^{k_{2}-1}b_{2},b_{s},\cdots,Ab_{s},\cdots,A^{k_{s}-1}b_{s},b_{s+1},\cdots,b_{s^{\prime}}\\}.$ Note that $Ab_{i}=0$ for $i\geq s+1$ and $A^{k_{i}+1}b_{i}=A^{k_{i}}a_{i}=0$ for each $i\leq s.$ This finishes the proof. ###### Corollary 6.3 Let $A_{n\times n}$ be a strictly upper triangular matrix over a field $K$ of characteristic $\mathrm{ch}(K)\neq 2$. Then $A^{2}$ is conjugate to $A.$ Proof. Suppose that $A=I+u$ for a nilpotent matrix $u.$ Lemma 6.2 implies that $A^{2}$ is conjugate to a direct sum of Jordan blocks. Without loss of generality, we assume $A$ is a Jordan block. Then $A^{2}=I+2u+u^{2}.$ By Lemma 6.2 again, $A^{2}$ is conjugate to a direct sum of Jordan blocks with $1s$ along the diagonal. The minimal polynomial of $A^{2}$ is $(x-1)^{n},$ which shows that there is only one block in the direct sum and thus $A^{2}$ is conjugate to $A.$ Recall that a matrix $A\in\mathrm{GL}_{n}(\mathbb{R})$ is called semisimple if as a complex matrix $A$ is conjugate to a diagonal matrix. A semisimple matrix $A$ is elliptic (respectively, hyperbolic) if all its (complex) eigenvalues have modulus $1$ (respectively, are $>0$). The following lemma is the complete multiplicative Jordan (or Jordan-Chevalley) decomposition (cf. [34], Lemma 7.1, page 430). ###### Lemma 6.4 Each $A\in$ $\mathrm{GL}_{n}(\mathbb{R})$ can be uniquely writen as $A=ehu,$ where $e,h,u\in\mathrm{GL}_{n}(\mathbb{R})$ are elliptic, hyperbolic and unipotent, respectively, and all three commute. The following result characterizes the continuous length functions on compact Lie groups. ###### Lemma 6.5 Let $G$ be a compact connected Lie group and $l$ a continuous length function on $G$. Then $l=0.$ Proof. For any element $g\in G,$ there is a maximal torus $T\backepsilon g.$ For finite order element $h\in T,$ we have $l(h)=0.$ Note that the set of finite-order elements is dense in $T.$ Since $l$ is continuous, $l$ vanishes on $T$ and thus $l(g)=0$ for any $g.$ ###### Theorem 6.6 Let $G=\mathrm{SL}_{n}(\mathbb{R})$ $(n\geq 2).$ Let $l:G\rightarrow[0,+\infty)$ be a length function, which is continuous on compact subgroups and the subgroup of diagonal matrices with positive diagonal entries. Then $l$ is uniquely determined by its images on the subgroup $D$ of diagonal matrices with positive diagonal entries. Proof. For any $g\in\mathrm{SL}_{n}(\mathbb{R}),$ let $g=ehu$ be the Jordan decomposition for commuting elements $e,h,u$, where $e$ is elliptic, $h$ is hyperbolic and $u$ is unipotent (see Lemma 6.4) after multiplications by suitable powers of derterminants. Then $l(g)\leq l(e)+l(h)+l(u).$ For any unipotent matrix $u,$ there is an invertible matrix $a$ such that $aua^{-1}$ is strictly upper triangular (see [34], Theorem 7.2, page 431). Lemma 6.3 implies that $u^{2}$ is conjugate to $u$. Therefore, $l(u)=0$ by Lemma 6.1. Since $l$ vainishes on a compact Lie group (cf. Lemma 6.5), we have that $l(e)=0$ for any elliptic matrix $e.$ Therefore, $l(g)\leq l(h).$ Similarly, $l(h)=l(e^{-1}gu^{-1})\leq l(g)$ which implies $l(g)=l(h).$ Note that a hyperbolic matrix is conjugate to a real diagonal matrix with positive diagonal entries. Proof of Theorem 0.10. By Theorem 6.6, the length function $l$ is determined by its image on the subgroup $D$ generated by $h_{12}(x),x\in\mathbb{R}_{>0}.$ Take $x=e^{t},t\in\mathbb{R}$. We have $l(h_{12}(e^{\frac{k}{l}}))=\frac{|k|}{|l|}l(h_{12}(e))$ for any rational number $\frac{k}{l}.$ Since $l$ satisfies the condition 2) of the definition and is continuous on $D$, we see that $l|_{D}$ is determined by the image $l(h_{12}(e))$ (actually, any real number $t$ is a limit of a rational sequence). Note that the translation function $\tau$ vainishes on compact subgroups and is continuous on the subgroup of diagonal subgroups with positive diagonal entries (cf. [16], Cor. 10.42 and Ex. 10.43, page 320). Therefore, $l$ is proposional to $\tau.$ Actually, $\tau$ can be determined explicitly by the formula $\mathrm{tr}(A)=\pm 2$cosh$\frac{\tau(A)}{2}$ (for nonzero $\tau(A)$), where $\mathrm{tr}$ is the trace and cosh is the hyperbolic cosine function (see [5], Section 7.34, page 173). This implies that $l(A)$ is determined by the spectrum radius of $A$ (which could also be seen clearly by the matrix norm). Let $h_{1i}(x)$ ($i=2,\cdots,n$) be an $n\times n$ diagonal matrix whose $(1,1)$-entry is $x,$ $(i,i)$-entry is $x^{-1},$ while other diagonal entries are $1$s and non-diagonal entries are $0$s. The subgroup $D<\mathrm{SL}_{n}(\mathbb{R})$ of diagonal matrices with positive diagonal entries is isomorphic to $(\mathbb{R}_{>0})^{n-1}$ and $D$ is generated by the matrices $h_{1i}(x)$ ($i=2,\cdots,n$) whose $(1,1)$-entry is $x,$ $(i,i)$-entry is $x^{-1}.$ Since $h_{1i}(x)$ $(i\neq 1)$ is conjugate to $h_{12}(x),$ a length function $l:\mathrm{SL}_{n}(\mathbb{R})\rightarrow[0,+\infty)$ is completely determined by its image on the convex hull spanned by $h_{12}(e),h_{13}(e),\cdots,h_{1n}(e)$ (see Theorem 7.10 for a more general result on Lie groups). Here $e$ is the Euler’s number in the natural exponential function. ###### Corollary 6.7 Let $l:\mathrm{SL}_{2}(\mathbb{R})\rightarrow[0,+\infty)$ be a non-trivial length function that is continuous on the subgroup $SO(2)$ and the diagonal subgroup. Then $l(g)>0$ if and only if $g$ is hyperbolic. Proof. It is well-known that the elements in $\mathrm{SL}_{2}(\mathbb{R})$ are classified as elliptic, hyperbolic and parabolic elements. Moreover, the translation length $\tau$ vanishes on the compact subgroup $SO(2)$ and the parabolic elements. The corollary follows Theorem 0.10. When the length function $l$ is the asymptotic distortion function $\mathrm{dist}_{\infty}$, Corollary 6.7 is known to Navas [24] (Proposition 4). ## 7 Length functions on algebraic and Lie groups For an algebraic group $G,$ let $k[G]$ be the regular ring. For any $g\in G,$ let $\rho_{g}:k[G]\rightarrow k[G]$ be the right translation by $x.$ The following is the famous Jordan (or Jordan-Chevalley) decomposition. ###### Lemma 7.1 ([37] p.99) Let $G$ be an algebraic group and $g\in G.$ There exists unique elements $g_{s},g_{u}$ such that $g_{s}g_{u}=g_{u}g_{s},$ and $\rho_{g_{s}}$ is semisimple, $\rho_{g_{u}}$ is unipotent. ###### Lemma 7.2 [46] Let $G$ be a reductive connected algebraic group over an algebraically closed field $k.$ The conjugacy classes of unipotent elements in $G$ is finite. ###### Lemma 7.3 ([29], Theorem 3.4) If $G$ is a reductive linear algebraic group defined over a field $k$ and $g\in G(k)$ then the set of conjugacy classes in $G(k)$ which when base changed to the algebraic closed field $\bar{k}$ are equal to the conjugacy class of $g$ in $G(\bar{k})$ is in bijection with the subset of $H^{1}(\bar{k}/k,Z(g)(k)),$ the Galois cohomology group. ###### Definition 7.4 A field $k$ is of type (F) if for any integer $n$ there exist only finitely many extensions of $k$ of degree $n$ (in a fixed algebraic closure $\bar{k}$ of $k$). Examples of fields of type (F) include: the field $\mathbb{R}$ of reals, a finite field, the field of formal power series over an algebraically closed field. ###### Lemma 7.5 [Borel-Serre [12], Theorem 6.2] Let $k$ be a field of type (F) and let $H$ be a linear algebraic group defined over $k$. The set $H^{1}(\bar{k}/k,H(k))$ is finite. ###### Lemma 7.6 Let $G(k)$ be a reductive linear algebraic group over a field of type (F) and $l$ a length function on $G.$ Then $l(g)=l(g_{s}),$ where $g_{s}$ is the semisimple part of $g.$ Proof. By the Jordan decomposition $g=g_{s}g_{u},$ we have $l(g)\leq l(g_{s})+l(g_{u})$ and $l(g_{s})\leq l(g)+l(g_{u}^{-1}).$ Note that for any integer $n,$ $g_{u}^{n}$ is also unipotent. By the Lemma 7.2, Lemma 7.3 and Lemma 7.5, there are only finitely many conjugacy classes of unipotent elements. This implies that $g_{u}^{n_{1}}=g_{u}^{n_{2}}$ for distinct positive integers $n_{1},n_{2}.$ Therefore, we have $n_{1}l(g_{u})=n_{2}l(g_{u}),$ which implies that $l(g_{u})=0$ and thus $l(g)=l(g_{s}).$ A Lie group $G$ is semisimple if its maximal connected solvable normal subgroup is trivial. Let $\mathfrak{g}$ be its Lie algebra and let $\exp:\mathfrak{g}\rightarrow G$ denote the exponential map. An element $x\in\mathfrak{g}$ is real semi-simple if $Ad(x)$ is diagonalizable over $\mathbb{R}$. An element $g\in G$ is called hyperbolic (resp. unipotent) if $g$ is of the form $g=\exp(x)$ where $x$ is real semi-simple (resp. nilpotent). In either case the element $x$ is easily seen to be unique and we write $x=\log g$. The following is the Jordan decomposition in Lie groups. An element $e\in G$ is elliptic if $Ad(e)$ is diagonalizable over $\mathbb{C}$ with eigenvalues $1$. ###### Lemma 7.7 ([42], Prop. 2.1 and Remark 2.1) 1. 1. Let $g\in G$ be arbitrary. Then g may be uniquely written $g=e(g)h(g)u(g)$ where $e(g)$ is elliptic, $h(g)$ is hyperbolic and $u(g)$ is unipotent and where the three elements $e(g),h(g),u(g)$ commute. 2. 2. An element $f\in G$ commutes with $g$ if and only if $f$ commutes with the three components. Moreover, if $f,g$ commutes, then $e(fg)=e(f)e(g),h(fg)=h(f)h(g),u(fg)=u(f)u(g).$ ###### Lemma 7.8 [Eberlein, Prop. 1.14.6, page 63]Let $G$ be a connected semisimple Lie group whose center is trivial. Then there exists an integer $n\geq 2$ and an algebraic group $G^{\ast}<\mathrm{GL}_{n}(\mathbb{C})$ defined over $\mathbb{Q}$ such that $G$ is isomorphic to $G_{\mathbb{R}}^{\ast 0}$ (the connected component of $G_{\mathbb{R}}^{\ast 0}$ containing the identity) as a Lie group. Let $G=KAN$ be an Iwasawa decomposition. The Weyl group $W$ is the finite group defined as the quotient of the normalizer of $A$ in $K$ modulo the centralizer of $A$ in $K.$ For an element $x\in A,$ let $W(h)$ be the set of all elements in $A$ which are conjugate to $x$ in $G.$ ###### Lemma 7.9 ([42], Prop. 2.4) An element $h\in G$ is hyperbolic if and only if it is conjugate to an element in $A.$ In such a case, $W(h)$ is a single $W$-orbit in $A.$ ###### Theorem 7.10 Let $G$ be a connected semisimple Lie group whose center is finite with an Iwasawa decomposition $G=KAN$. Let $W$ be the Weyl group. 1. (i) Any length function $l$ on $G$ that is continuous on the maximal compact subgroup $K$ is determined by its image on $A.$ 2. (ii) Conversely, any length function $l$ on $A$ that is $W$-invariant (i.e. $l(w\cdot a)=l(a)$) can be extended to be a length function on $G$ that vanishes on the maximal compact subgroup $K.$ Proof. (i) Let $Z$ be the center of $G.$ Then $G/Z$ is connected with trivial center. For any $z\in Z,g\in G,$ we have $l(z)=0$ and $l(gz)=l(g).$ The length function $l$ factors through a length function on $G/Z.$ We may assume that $G$ has the trivial center. For any $g\in G,$ the Jordan decomposition gives $g=ehu,$ where $e$ is elliptic, $h$ is hyperbolic and $u$ is unipotent and where the three elements $e,h,u$ commute (cf. Lemma 7.7). By Lemma 7.8, the Lie group $G$ is an algebraic group. Lemma 7.6 implies that $l$ vanies on unipotent elements and $l(g)=l(eh).$ Since $l$ vanishes on $e$ (cf. Lemma 6.5), we have $l(g)=l(h).$ Therefore, the function $l$ is determined by its image on $A.$ (ii) Let $l$ be a length function $l$ on $A$ that is $W$-invariant. We first extend $l$ to the set $H$ of all the conjugates of $A.$ For any $g\in G,a\in A,$ define $l^{\prime}(gag^{-1})=l(a).$ If $g_{1}a_{1}g_{1}^{-1}=g_{2}a_{2}g_{2}^{-1}$ for $g_{1},g_{2}\in G,a_{1},a_{2}\in A,$ then $g_{1}^{-1}g_{2}a_{2}g_{2}^{-1}g_{1}=a_{1}.$ By Lemma 7.9, there exists an element $w\in W$ such that $w\cdot a_{2}=a_{1}.$ Therefore, we have $l(a_{1})=l(a_{2})$ and thus $l^{\prime}$ is well-defined on the set $H$ of conjugates of elements in $A.$ Such a set $H$ is the set consisting of hyperbolic elements by Lemma 7.9. We then extend $l^{\prime}$ on the set of all conjugates of elements in $K.$ For any $g\in G,k\in K,$ define $l^{\prime}(gkg^{-1})=0.$ If $g_{1}kg_{1}^{-1}=g_{2}ag_{2}^{-1}$ for $g_{1},g_{2}\in G,k\in K,a\in A,$ then $g_{1}^{-1}g_{2}ag_{2}^{-1}g_{1}=k.$ Then $k$ is both hyperbolic and elliptic. The only element which is both elliptic and hyperbolic is the identity element. Therefore, we have $k=a=1$ and $l^{\prime}(g_{1}kg_{1}^{-1})=l(g_{2}ag_{2}^{-1})=1.$ This shows that $l^{\prime}$ is well-defined on the set of hyperbolic elements and elliptic elements. For any unipotent element $u\in G,$ define $l^{\prime}(u)=0.$ For any element $g\in G,$ let $g=ehu$ be the Jordan decomposition. Define $l^{\prime}(g)=l^{\prime}(h).$ We check the function $l^{\prime}$ is a length function on $G.$ The definition shows that $l^{\prime}$ is conjugate invariant. For any positive integer $n$ and any $g\in G$ with Jordan decomposition $g=ehu,$ we have $g^{n}=e^{n}h^{n}u^{n}$ and thus $l^{\prime}(g^{n})=l^{\prime}(h^{n}).$ But $h^{n}$ is hyperbolic and conjugate to an element in $A$ (see Lemma 7.9). Therefore, we have $l^{\prime}(h^{n})=|n|l^{\prime}(h)$ and thus $l^{\prime}(g^{n})=|n|l^{\prime}(g).$ If $g_{1}=e_{1}h_{1}u_{1}$ commutes with $g_{2}=e_{2}h_{2}u_{2},$ then $g_{1}g_{2}=e_{1}e_{2}h_{1}h_{2}u_{1}u_{2}$ (cf. Lemma 7.7) and $l^{\prime}(g_{1}g_{2})=l^{\prime}(h_{1}h_{2}).$ Since $h_{1},h_{2}$ are commuting hyperbolic elements, they are conjuate simutaniously to elements in $A.$ Therefore, we have $l^{\prime}(h_{1}h_{2})\leq l^{\prime}(h_{1})+l^{\prime}(h_{2})$ and $l^{\prime}(g_{1}g_{2})\leq l^{\prime}(g_{1})+l^{\prime}(g_{2}).$ ###### Remark 7.11 A length function $l$ on $A$ is determined by a group homomorphism $f:A\rightarrow\mathbb{B},$ for a real Banach space $(\mathbb{B},\|\|),$ satisfying $l(a)=\|f(a)\|$ (see Lemma 4.1). The previous theorem implies that a length function $l$ on the Lie group $G$ (that is continuous on compact subgroup) is uniqely determined by such a group homomorphism $f:A\rightarrow\mathbb{B}$ such that $\|f(a)\|=\|f(wa)\|$ for any $a\in A$ and $w\in W,$ the Weyl group. Let $G$ be a connected semisimple Lie group whose center is finite with an Iwasawa decomposition $G=KAN$. Let $\exp:\mathfrak{g}\rightarrow G$ be the exponent map from the Lie algebra $\mathfrak{g}$ with subalgebra $\mathfrak{h}$ corresponding to $A$. ###### Theorem 7.12 Suppose that $l$ is a length function on $G$ that is continuous on $K$ and $A.$ Then $l$ is determined by its image on $\exp(v)$ $($unit vector $v\in\mathfrak{h})$ in a fixed closed Weyl chamber of $A.$ Proof. Let $Z$ be the center of $G.$ Then $G/Z$ is connected with trivial center. The length function $l$ factors through an length function on $G/Z$ (cf. Corrolary 1.8). For any $g\in G$ we have $g=ehu,$ where $e$ is elliptic, $h$ is hyperbolic and $u$ is unipotent and where the three elements $e,h,u$ commute (cf. Lemma 7.7). By Lemma 7.8 and Lemma 7.6, we have $l(g)=l(eh).$ Since $l$ vanishes on $e$ (cf. Lemma 6.5), we have $l(g)=l(h).$ Any element $h\in A$ is conjugate to an element in a fixed Weyl chamber $C$ (cf. [38], Theorem 8.20, page 254). For any element $\exp(x)\in C,$ with unit vector $x\in\mathfrak{h},$ the one-parameter subgroup $\exp(\mathbb{R}x)$ lies in $A.$ Since $l$ is continuous on $A,$ the function $l$ is determined by its image on $\exp(\mathbb{Q}x).$ Note that $l(e^{\frac{m}{n}x})=\frac{1}{n}l(e^{mx})=\frac{m}{n}l(e^{x})$ for any rational number $\frac{m}{n}.$ The function $l$ is determined by $l(e^{x}),$ for all unit vectors $x$ in the fixed closed Weyl chamber. ###### Corollary 7.13 Let $G$ be a connected semisimple Lie group whose center is finite of real rank $1.$ There is essentially only one length function on $G.$ In order words, any continuous length function is proportional to the translation function on the symmetric space $G/K.$ Proof. When the real rank of $G$ is $1,$ a closed Weyl chamber is of dimension $1$. Therefore, the previous theorem implies that any continuous length function is determined by its image on a unit vector in a split torus. ## 8 Rigidity of group homomorphisms on arithmetic groups Let $V$ denote a finite-dimensional vector space over $\mathbb{C}$, endowed with a $\mathbb{Q}$-structure. Recall that the arithmetic subgroup is defined as the following (cf. Borel [11], page 37). ###### Definition 8.1 Let $G$ be a $\mathbb{Q}$-subgroup of $\mathrm{GL}(V)$. A subgroup $\Gamma$ of $G_{\mathbb{Q}}$ is said to be arithmetic if there exists a lattice $L$ of $V_{\mathbb{Q}}$ such that $\Gamma$ is commensurable with $G_{L}=\\{g\in G_{\mathbb{Q}}:gL=L\\}.$ ###### Theorem 8.2 Let $\Gamma$ be an arithmetic subgroup of a simple algebraic $\mathbb{Q}$-group of $\mathbb{Q}$-rank at least $2.$ Suppose that $H$ is a group with a purely positive length function. Then any group homomorphism $f:\Gamma\rightarrow H$ has its image finite. Recall that a group $G$ is quasi-simple, if any non-trivial normal subgroup is either finite or of finite index. The Margulis-Kazhdan theorem (see [62], Theorem 8.1.2) implies that an irreducible lattice (and hence) in a semisimple Lie group of real rank $\geq 2$ is quasi-simple. ###### Lemma 8.3 Let $\Gamma$ be a finitely generated quasi-simple group with contains a Heisenberg subgroup, i.e. there are elements torsion-free elements $a,b,c\in\Gamma$ satisfying $[a,b]=c,[a,c]=[c,b]=1$. Suppose that $G$ has a virtually poly-positive length. Then any group homomorphism $f:\Gamma\rightarrow G$ has its image finite. Proof. Suppose that $G$ has a finite-index subgroup $H$ and a subnormal series $1=H_{n}\vartriangleleft H_{n-1}\vartriangleleft\cdots\vartriangleleft H_{0}=H$ such that every finitely generated subgroup of $H_{i}/H_{i+1}$ has a purely positive length function. Without loss of generality, we assume that $H$ is normal. Let $f:\Gamma\rightarrow G$ be a homomorphism. The kernel of the composite $f_{0}:\Gamma\overset{f}{\rightarrow}G\rightarrow G/H$ is finitely generated. Suppose that the image of the composite $f_{1}:\ker f_{0}\overset{f}{\rightarrow}H\rightarrow H/H_{1}$ has a purely positive length function $l.$ After passing to finite-index subgroups, we may still suppose that $\ker f_{0}$ contains a Heisenberg subgroup $\langle a,b,c\rangle.$ By Lemma 5.2, the length function $l$ vanishes on $f_{1}(c).$ Therefore $f_{1}(c^{k})=1\in H/H_{1}$ for some integer $k>0.$ The normal subgroup $\ker f_{1}$ containing $c^{k}$ is of finite index. Now we have map $\ker f_{1}\rightarrow H_{1}$ induced by $f.$ An induction argument shows that $f$ maps some power $c^{d}$ of the central element of the Heisenberg subgroup into the identity $1\in G.$ Therefore, the image of $f$ is finite. Proof of Theorem 8.2. It is well-known that that $G$ contains a $\mathbb{Q}$-split simple subgroup whose root system is the reduced subsystem of the $\mathbb{Q}$-root system of $G$ (see [13], Theorem 7.2, page 117). Replacing $G$ with this $\mathbb{Q}$-subgroup, we may assume $G$ is $\mathbb{Q}$-split and the root system of G is reduced. Because $G$ is simple and $\mathbb{Q}$-rank($G$)$\geq 2$, we know that the $\mathbb{Q}$-root system of G is irreducible and has rank at least two. Therefore, the $\mathbb{Q}$-root system of $G$ contains an irreducible subsystem of rank two, that is, a root subsystem of type $A_{2},B_{2},G_{2}$ (see [60], page 338). For $A_{2},$ choose $\\{\alpha_{1},\alpha_{3}\\}$ as a set of simple roots (see Figure 1). Then the root element $x_{\alpha_{1}+\alpha_{3}}(rs)=x_{\alpha_{2}}(rs)$ is a commutator $[x_{\alpha_{1}}(r),x_{\alpha_{3}}(s)],$ with $x_{\alpha_{2}}(rs)$ commutes with $x_{\alpha_{1}}(r),x_{\alpha_{3}}(s).$ For $G_{2}$, the long roots form a subsystem of $A_{2}.$ For $B_{2},$ choose $\\{\alpha_{1},\alpha_{4}\\}$ as a set of simple roots (see Figure 2). The long root element $x_{\alpha_{3}}(2rs)$ is a commutator $[x_{\alpha_{2}}(r),x_{\alpha_{4}}(s)]$ of the two short root elements, and $x_{\alpha_{3}}(2rs)$ commutes with $x_{\alpha_{2}}(r),x_{\alpha_{4}}(s)$ (cf. [37], Proposition of page 211). This shows that the arithmetic subgroup $\Gamma$ contains a Heisenberg subgroup. The theorem is then proved by Lemma 8.3. If we consider special length functions, general results can be proved. When we consider the stable word lengths, the following is essentially already known (cf. Polterovich [53], Corollary 1.1.D and its proof). ###### Proposition 8.4 Let $\Gamma$ be an irreducible non-uniform lattice in a semisimple connected, Lie group without compact factors and with finite center of real rank $\geq 2.$ Assume that a group $G$ has a virtually poly-positive stable word length. In other words, the group $G$ has a finite-index subgroup $H$ and a subnormal series $1=H_{n}\vartriangleleft H_{n-1}\vartriangleleft\cdots\vartriangleleft H_{0}=H$ such that every finitely generated subgroup of $H_{i}/H_{i+1}$ ($i=0,1,\cdots,n-1$) has a purely positive stable word length. Then any group homomorphism $f:\Gamma\rightarrow G$ has its image finite. Proof. Without loss of generality, we assume that $f$ takes image in $H.$ Since a lattice is finitely generated, $\Gamma$ has its image in $H_{0}/H_{1}$ finitely generated. When the image has a purely positive word length, any distorted element in $\Gamma$ must have trivial image in $H_{0}/H_{1}$ (see ). Lubotzky, Mozes and Raghunathan [44] prove that irreducible non-uniform lattices in higher rank Lie groups have non-trivial distortion elements (They actually prove the stronger result that there are elements in the group whose word length has logarithmic growth). Then a finite-index subgroup $\Gamma_{0}<\Gamma$ will have image in $H_{1}$, since high-rank irreducible lattices are quasi-simple. An induction argument finishes the proof. When we consider the length given by quasi-cocyles, the following is also essentially already known ( cf. Py [54], Prop. 2.2, following Burger-Monod [18] [19]). Recall that a locally compact group has property (TT) if any continuous rough action on a Hilbert space has bounded orbits (see [49], page 172). Burger-Monod proves that an irreducible lattice $\Gamma$ in a high-rank semisimple Lie group has property (TT). ###### Proposition 8.5 Let $\Gamma$ be an irreducible lattice in a semisimple connected, Lie group without compact factors and with finite center of real rank $\geq 2.$. Assume that a group $G$ has a virtually poly-positive average norm for quasi- cocycles. In other words, the group $G$ has a finite-index subgroup $H$ and a subnormal series $1=H_{n}\vartriangleleft H_{n-1}\vartriangleleft\cdots\vartriangleleft H_{0}=H$ such that every finitely generated subgroup of $H_{i}/H_{i+1}$ ($i=0,1,\cdots,n-1$) has a purely positive length given by a quasi-cocycle with values in Hilbert spaces. Then any group homomorphism $f:\Gamma\rightarrow G$ has its image finite. Proof. Note that a group $\Gamma$ has property (TT) if and only if $H^{1}(\Gamma;E)=0$ and $\ker(H_{b}^{2}(\Gamma;E)\rightarrow H^{2}(\Gamma;E))=0$ for any linear isometric action of $\Gamma$ on a Hilbert space $E.$ Here $H_{b}^{2}(\Gamma;E)$ is the second bounded cohomology group. Suppose that $u:\Gamma\rightarrow E$ is a quasi-cocyle. There is a bounded map $v:\Gamma\rightarrow E$ and a $1$-cocycle $w:\Gamma\rightarrow E$ such that $u=v+w,$ by Proposition 2.1 of Py [54]. Since $\Gamma$ has property T, there exists $x_{0}\in E$ such that $w(\gamma)=\gamma x_{0}-x_{0}.$ Therefore, we have $\displaystyle\frac{\|u(\gamma^{n})\|}{n}$ $\displaystyle=$ $\displaystyle\frac{\|v(\gamma^{n})+w(\gamma^{n})\|}{n}$ $\displaystyle=$ $\displaystyle\frac{\|v(\gamma^{n})+\gamma^{n}x_{0}-x_{0}\|}{n}$ $\displaystyle\leq$ $\displaystyle\frac{\|v(\gamma^{n})\|+2\|x_{0}\|}{n}\rightarrow 0.$ Without loss of generality, we assume that $G=H.$ Suppose that any finitely generated subgroup of $H/H_{1}$ has a purely positive average norm $l$ given by a cocycle. The composite $\Gamma\overset{f}{\rightarrow}H\rightarrow H/H_{1}$ has a finite-index kernel $\Gamma_{0}$, since $l$ vanishes on infinite-order elements of the image. This implies that $f(\Gamma_{0})$ lies in $H_{1}.$ A similar argument proves that $\ker f$ is of finite index in the general case. ## 9 Rigidity of group homomorphisms on matrix groups ### 9.1 Steinberg groups over finite rings Recall that a ring $R$ is right Artinian if any non-empty family of right ideals contains minimal elements. A ring $R$ is semi-local if $R/\mathrm{rad}(R)$ is right Artinian (see Bass’ K-theory book [3] page 79 and page 86), where $\mathrm{rad}(R)$ is the Jacobson radical. Let $n$ be a positive integer and $R^{n}$ the free $R$-module of rank $n$ with standard basis. A vector $(a_{1},\ldots,a_{n})$ in $R^{n}$ is called _right unimodular_ if there are elements $b_{1},\ldots,b_{n}\in R$ such that $a_{1}b_{1}+\cdots+a_{n}b_{n}=1$. The _stable range condition_ $\mathrm{sr}_{m}$ says that if $(a_{1},\ldots,a_{m+1})$ is a right unimodular vector then there exist elements $b_{1},\ldots,b_{m}\in R$ such that $(a_{1}+a_{m+1}b_{1},\ldots,a_{m}+a_{m+1}b_{m})$ is right unimodular. It follows easily that $\mathrm{sr}_{m}\Rightarrow\mathrm{sr}_{n}$ for any $n\geq m$. A semi-local ring has the stable range $\mathrm{sr}_{2}$ ( [3], page 267, the proof of Theorem 9.1). A finite ring $R$ is right Artinian and thus has $\mathrm{sr}_{2}.$ The stable range $\mathrm{sr}(R)=\min\\{m:R\text{ has }\mathrm{sr}_{m+1}\\}.$ Thus $\mathrm{sr}(R)=1$ for a finite ring $R$. We briefly recall the definitions of the elementary subgroups $E_{n}(R)$ of the general linear group $\mathrm{GL}_{n}(R)$, and the Steinberg groups $\mathrm{St}_{n}(R)$. Let $R$ be an associative ring with identity and $n\geq 2$ be an integer. The general linear group $\mathrm{GL}_{n}(R)$ is the group of all $n\times n$ invertible matrices with entries in $R$. For an element $r\in R$ and any integers $i,j$ such that $1\leq i\neq j\leq n,$ denote by $e_{ij}(r)$ the elementary $n\times n$ matrix with $1s$ in the diagonal positions and $r$ in the $(i,j)$-th position and zeros elsewhere. The group $E_{n}(R)$ is generated by all such $e_{ij}(r),$ i.e. $E_{n}(R)=\langle e_{ij}(r)|1\leq i\neq j\leq n,r\in R\rangle.$ Denote by $I_{n}$ the identity matrix and by $[a,b]$ the commutator $aba^{-1}b^{-1}.$ The following lemma displays the commutator formulas for $E_{n}(R)$ (cf. Lemma 9.4 in [47]). ###### Lemma 9.1 Let $R$ be a ring and $r,s\in R.$ Then for distinct integers $i,j,k,l$ with $1\leq i,j,k,l\leq n,$ the following hold: 1. (1) $e_{ij}(r+s)=e_{ij}(r)e_{ij}(s);$ 2. (2) $[e_{ij}(r),e_{jk}(s)]=e_{ik}(rs);$ 3. (3) $[e_{ij}(r),e_{kl}(s)]=I_{n}.$ By Lemma 9.1, the group $E_{n}(R)$ $(n\geq 3)$ is finitely generated when the ring $R$ is finitely generated. Moreover, when $n\geq 3,$ the group $E_{n}(R)$ is normally generated by any elementary matrix $e_{ij}(1).$ The commutator formulas can be used to define Steinberg groups as follows. For $n\geq 3,$ the Steinberg group $\mathrm{St}_{n}(R)$ is the group generated by the symbols $\\{x_{ij}(r):1\leq i\neq j\leq n,r\in R\\}$ subject to the following relations: 1. (St$1$) $x_{ij}(r+s)=x_{ij}(r)x_{ij}(s);$ 2. (St$2$) $[x_{ij}(r),x_{jk}(s)]=x_{ik}(rs)$ for $i\neq k;$ 3. (St$3)$ $[x_{ij}(r),x_{kl}(s)]=1$ for $i\neq l,j\neq k.$ There is an obvious surjection $\mathrm{St}_{n}(R)\rightarrow E_{n}(R)$ defined by $x_{ij}(r)\longmapsto e_{ij}(r).$ For any ideal $I\vartriangleleft R,$ let $p:R\rightarrow R/I$ be the quotient map. Then the map $p$ induces a group homomorphism $p_{\ast}:\mathrm{St}_{n}(R)\rightarrow\mathrm{St}_{n}(R/I).$ Denote by $\mathrm{St}_{n}(R,I)$ (resp., $E_{n}(R,I)$) the subgroup of $\mathrm{St}_{n}(R)$ (resp., $E_{n}(R)$) normally generated by elements of the form $x_{ij}(r)$ (resp., $e_{ij}(r)$) for $r\in I$ and $1\leq i\neq j\leq n.$ In fact, $\mathrm{St}_{n}(R,I)$ is the kernel of $p_{\ast}$ (cf. Lemma 13.18 in Magurn [47] and its proof). However, $E_{n}(R,I)$ may not be the kernel of $E_{n}(R)\rightarrow E_{n}(R/I)$ induced by $p.$ ###### Lemma 9.2 When $n\geq\mathrm{sr}(R)+2,$ the natural map $\mathrm{St}_{n}(R)\rightarrow\mathrm{St}_{n+1}(R)$ is injective. In particular, when $R$ is finite, the Steinberg group $\mathrm{St}_{n}(R)$ is finite for any $n\geq 3.$ Proof. Let $W(n,R)$ be the kernel of the natural map $\mathrm{St}_{n}(R)\rightarrow\mathrm{St}_{n+1}(R).$ When $n\geq\mathrm{sr}(R)+2,$ the kernel $W(n,R)$ is trivial (cf. Kolster [41], Theorem 3.1 and Cor. 2.10). When $n$ is sufficient large, the Steinberg group $\mathrm{St}_{n}(R)$ is the universal central extension of $E_{n}(R)$ (cf. [59], Proposition 5.5.1. page 240). Therefore, the kernel $\mathrm{St}_{n}(R)\rightarrow E_{n}(R)$ is the second homology group $H_{2}(E_{n}(R);\mathbb{Z}).$ When $R$ is finite, both $E_{n}(R)$ and $H_{2}(E_{n}(R);\mathbb{Z})$ are finite. Therefore, the group $\mathrm{St}_{n}(R)$ is finite for any $n\geq 3.$ ### 9.2 Rigidity of group homomorphisms on matrix groups ###### Theorem 9.3 Suppose that $G$ is a group satisfying that 1) $G$ has a purely positive length function, i.e. there is a length function $l:G\rightarrow[0,\infty)$ such that $l(g)>0$ for any infinite-order element $g;$ and 2) any torsion abelian subgroup of $G$ is finitely generated. Let $R$ be an associative ring with identity and $\mathrm{St}_{n}(R)$ the Steinberg group. Suppose that $S<\mathrm{St}_{n}(R)$ is a finite-index subgroup. Then any group homomorphism $f:\mathrm{St}_{n}(R)\rightarrow G$ has its image finite when $n\geq 3$. Proof. Since any ring $R$ is a quotient of a free (non-commutative) ring $\mathbb{Z}\langle X\rangle$ for some set $X$ and $\mathrm{St}_{n}(R)$ is functorial with respect to the ring $R,$ we assume without loss of generality that $R=$ $\mathbb{Z}\langle X\rangle$. We prove the case $S=\mathrm{St}_{n}(R)\ $first. Let $x_{ij}=\langle x_{ij}(r):r\in R\rangle,$ which is isomorphic to the abelian group $R.$ Note that $[x_{12}(1),x_{23}(1)]=x_{13}(1)$ and $x_{13}(1)$ commutes with $x_{12}(1)$ and $x_{23}(1).$ Lemma 5.2 implies any length function vanishes on $x_{13}(1).$ By Lemma 1.4, the length $l(f(x_{13}(1)))=0.$ Note that $x_{ij}(r)$ is conjugate to $x_{13}(r)$ for any $r\in R$ and $i,j$ satisfying $1\leq i\neq j\leq n.$ Since $l$ is purely postive, we get that $f(x_{12}(1))$ is of finite order. Let $I=\ker f|_{x_{12}}.$ Then $I\neq\varnothing,$ as $f(x_{12}(1))$ is of finite order. For any $x\in I,$ and $y\in R,$ we have $x_{12}(xy)=[x_{13}(x),x_{32}(y)].$ Therefore, $f(x_{12}(xy))=[f(x_{13}(x),f(x_{32}(y)))]=1$ and thus $xy\in I.$ Similarly, we have $f(x_{12}(yx))=f([x_{13}(y),x_{32}(x)])=1.$ This proves that $I$ is a (two- sided) ideal. Note that $f(e_{12})=R/I$ is a torsion abelian group. By the assumption 2), the quotient ring $R/I$ is finite. Let $\mathrm{St}_{n}(R,I)$ be the normal subgroup of $\mathrm{St}_{n}(R)$ generated by $x_{ij}(r),r\in I.$ There is a short exact sequence $\begin{array}[]{ccc}1\rightarrow\mathrm{St}_{n}(R,I)\rightarrow&\mathrm{St}_{n}(R)\rightarrow&\mathrm{St}_{n}(R/I)\rightarrow 1.\end{array}$ Since $R/I$ is finite, we know that $\mathrm{St}_{n}(R/I)$ is finite by Lemma 9.2. This proves that $\mathop{\mathrm{I}m}f$ is finite since $f$ factors through $\mathrm{St}_{n}(R/I)$. For general finite-index subgroup $S,$ we assume $S$ is normal in $\mathrm{St}_{n}(R)$ after passing to a finite-index subgroup of $S.$ A similar proof shows that $S$ contains $\mathrm{St}_{n}(R,I)$ for some ideal $I$ with the quotient ring $R/I$ finite. Therefore, the image $\mathop{\mathrm{I}m}f$ is finite. ###### Theorem 9.4 Suppose that $G$ is a group having a purely positive length function $l$. Let $R$ be an associative ring of characteristic zero such that any nonzero ideal is of a finite index (eg. the ring of algebraic integers in a number field). Suppose that $S<\mathrm{St}_{n}(R)$ is a finite-index subgroup of the Steinberg group. Then any group homomorphism $f:S\rightarrow G$ has its image finite when $n\geq 3$. Proof. The proof is similar to that of Theorem 9.3. Let $I=\ker f|_{x_{12}},$ where $x_{12}=S\cap\langle x_{12}(r):r\in R\rangle.$ Since $R$ is of characteristic zero and the length $l(f(x_{12}(k)))=0$ for some integer $k,$ we have $f(x_{12}(k))$ is of finite order. Therefore, $f(x_{12}(k^{\prime}))=1$ for some integer $k^{\prime},$ which proves that the ideal $I$ is nozero. Since $I$ is of finite index in $R,$ we get that $\mathrm{St}_{n}(R,I)$ is of finite index in $S.$ This finishes the proof. Since the natural map $\mathrm{St}_{n}(R)\rightarrow E_{n}(R)$ is surjective, any group homomorphism $f:E_{n}(R)\rightarrow G$ can be lifted to be a group homomorphism $\mathrm{St}_{n}(R)\rightarrow G.$ Moreover, a finite-index subgroup $E$ of $E_{n}(R)$ is lifted to be a finite-index subgroup $S$ of $\mathrm{St}_{n}(R).$ Theorem 0.4 and Theorem 0.6 follows Theorem 9.3 and Theorem 9.4, by inductive arugments on the subnormal series as those of the proofs of Theorem 0.3. Proof of Corollary 0.5 and Corollary 0.7. For Corollary 0.5, it is enough to check the two conditions for $G$ in Theorem 0.4. Lemma 3.6 proves that $G$ has a purely positive length function. When $G$ is a $\mathrm{CAT}(0)$ group, (i.e. $G$ acts properly and cocompactly on a $\mathrm{CAT}(0)$ space), then any solvable subgroup of $G$ is finitely generated (and actually virtually abelian, see the Solvable Subgroup Theorem of [16], Theorem 7.8, page 249). When $G$ is hyperbolic, it’s well-known that $G$ contains finitely many conjugacy classes of finite subgroups and thus a torsion abelian subgroup is finite (see [16], Theorem 3.2, page 459). Birman-Lubotzky-McCarthy [7] proves that any abelian subgroup of the mapping class groups for orientable surfaces is finitely generated. Bestvina-Handel [6] proves that every solvable subgroup of $\mathrm{Out}(F_{k})$ has a finite index subgroup that is finitely generated and free abelian. When $G$ is the diffeomorphism group $\mathrm{Diff}(\Sigma,\omega),$ there is a subnormal series (see the proof of Lemma 3.8) $1\vartriangleleft\mathrm{Ham}(\Sigma,\omega)\vartriangleleft\mathrm{Diff}_{0}(\Sigma,\omega)\vartriangleleft\mathrm{Diff}(\Sigma,\omega),$ with subquoitents in $\mathrm{Ham}(\Sigma,\omega),$ $H_{1}(\Sigma,\mathbb{R})$ and the mapping class group $\mathrm{MCG}(\Sigma).$ Any abelian subgroup of a finitely generated subgroup of these groups is finitely generated. Corrolary 0.7 follows Theorem 9.4 and Lemma 3.6. ###### Remark 9.5 An infinite torsion abelian group may act properly on a simplicial tree (see [16], Example 7.11, page 250). Therefore, condition 2) in Theorem 0.4 does not hold for every group $G$ acting properly on a $\mathrm{CAT}(0)$ (or a Gromov hyperbolic) space. We don’t know whether the condition 2) can be dropped. ## 10 Length functions on Cremona groups Let $k$ be a field and $k(x_{1},x_{2},\cdots,x_{n})$ be the field of rational functions in $n$ indeterminates over $k.$ It is well-known that the Cremona group $\mathrm{Cr}_{n}(k)$ is isomorphic to the automorphism group $\mathrm{Aut}_{k}(k(x_{1},x_{2},\cdots,x_{n}))$ of the field $k(x_{1},x_{2},\cdots,x_{n})$. ###### Lemma 10.1 Let $f:k(x_{1},x_{2},\cdots,x_{n})\rightarrow k(x_{1},x_{2},\cdots,x_{n})$ be given by $f(x_{1})=\alpha x_{1},f(x_{i})=x_{i}$ for some $0\neq\alpha\in k$ and any $i=2,\cdots,n.$ Then $f$ lies in the center of a Heisenberg subgroup. In other words, there exists $g,h\in\mathrm{Cr}_{n}(k)$ such that $[g,h]=ghg^{-1}h^{-1}=f,[g,f]=1$ and $[h,f]=1.$ Proof. Let $g,h:k(x_{1},x_{2},\cdots,x_{n})\rightarrow k(x_{1},x_{2},\cdots,x_{n})$ be given by $g(x_{1})=x_{1}x_{2},g(x_{i})=x_{i}(i=2,\cdots,n)$ and $h(x_{1})=x_{1},h(x_{2})=\alpha^{-1}x_{2},h(x_{j})=x_{j}(j=3,\cdots,n).$ It can be directly checked that $[g,h]=f,[g,f]=1$ and $[h,f]=1.$ ###### Lemma 10.2 Let $l:\mathrm{Bir}(\mathbb{P}_{k}^{n})\rightarrow[0,\infty)$ be a length function $(n\geq 2)$. Then $l$ vanishes on diagonal elements and unipotent elements of $\mathrm{Aut}(\mathbb{P}_{k}^{n})=\mathrm{PGL}_{n+1}(k).$ Proof. Let $g=\mathrm{diag}(a_{0},a_{1},\cdots,a_{n})\in\mathrm{PGL}_{n+1}(k)$ be a diagonal element. Note that $l$ is subadditive on the diagonal subgroups. In order to prove $l(g)=0,$ it is enough to prove that $l(\mathrm{diag}(1,\cdots,1,a_{i},1,\cdots,1))=0,$ where $\mathrm{diag}(1,\cdots,1,a_{i},1,\cdots,1)$ is the diagonal matrix with $a_{i}$ in the $(i,i)$-th position and all other diagonal entries are $1.$ But $\mathrm{diag}(1,\cdots,1,a_{i},1,\cdots,1)$ is conjugate to $\mathrm{diag}(1,\alpha,1,\cdots,1)$ for $\alpha=a_{i}.$ Lemma 10.1 implies that $\mathrm{diag}(1,\alpha,1,\cdots,1)$ lies in the center of a Heisenberg group. Therefore, $l(\mathrm{diag}(1,\alpha,1,\cdots,1))=0$ by Lemma 5.2. This proves $l(g)=0.$ The vanishing of $l$ on unipotent elements follows Corollary 6.3 when the characteristic of $k$ is not $2$. When the characteristic of $k$ is $2,$ any unipotent element $A=I+u$ (where $u$ is nilpotent) is of finite order. This means $l(A)=0$. Proof of Theorem 0.8. When $k$ is algebraically closed, the Jordan normal form implies that any element $g\in\mathrm{PGL}_{n}(k)$ is conjugate to the form $sn$ with $s$ diagonal and $n$ the strictly upper triangular matrix. Moreover, $sn=ns.$ Therefore, $l(f)\leq l(s)+l(n).$ By Lemma 10.2, $l(s)=l(n)=0$ and thus $l(g)=0$. Proof of Corollary 0.9. Let $f:\mathrm{Bir}(\mathbb{P}_{k}^{2})\rightarrow G$ be a group homomorphism. Suppose that $G$ has a purely positive length function $l.$ By Theorem 0.8, the purely positive length function $l$ on $G$ will vanish on $f(\mathrm{PGL}_{3}(k)).$ Since $k$ is infinite and $\mathrm{PGL}_{3}(k)$ is a simple group, we get that $\mathrm{PGL}_{3}(k)$ lies in the $\ker f.$ By Noether-Castelnuovo Theorem, $\mathrm{Bir}(\mathbb{P}_{k}^{2})$ is generated by $\mathrm{PGL}_{3}(k)$ and an involution. Moreover, the $\mathrm{Bir}(\mathbb{P}_{k}^{2})$ is normally generated by $\mathrm{PGL}_{3}(k).$ Therefore, the group homomorphism $f$ is trivial. The general case is proved by an inductive argument on the subnormal series of a finite-index subgroup of $G.$ ###### Lemma 10.3 Let $\mathrm{Bir}(\mathbb{P}_{\mathbb{R}}^{n})$ $(n\geq 2)$ be the real Cremona group. Any length function $l:\mathrm{Bir}(\mathbb{P}_{k}^{n})\rightarrow[0,\infty),$ which is continuous on $\mathrm{PSO}(n+1)<\mathrm{Aut}(\mathbb{P}_{\mathbb{R}}^{n}),$ vanishes on $\mathrm{PGL}_{n+1}(\mathbb{R}).$ Proof. By Lemma 10.2, the length function $l$ vanishes on diagonal matrices of $\mathrm{PGL}_{n+1}(\mathbb{R}).$ Theorem 0.11 implies that $l$ vanishes on the whole group $\mathrm{PGL}_{n+1}(\mathbb{R}).$ Acknowledgements The author wants to thank many people for helpful discussions, including Wenyuan Yang on a discussion of hyperbolic groups, C. Weibel on a discussion on Steinberg groups of finite rings, Feng Su on a discussion of Lie groups, Ying Zhang on a discussion of translation lengths of hyperbolic spaces, Enhui Shi for a discussion on smooth measure-theoretic entropy. ## References * [1] E. Alibegović, Translation lengths in $\mathrm{Out}(F_{n})$, Geom. Dedicata 92 (2002), 87–93. * [2] W. Ballmann, M. Gromov, and V. Schroeder, Manifolds of nonpositive curvature, Progress in Mathematics, vol. 61, Birkhäuser Boston Inc., Boston, MA, 1985. * [3] H. Bass, Algebraic K-Theory, Benjamin, New York, 1968. * [4] Algebra Group of Beida, Gao Deng Dai Shu (Advanced algebra), High education press, 2013. * [5] A. Beardon, The geometry of discrete groups, Graduate Texts in Math., Vol. 91, Springer-Verlag, New York, 1983. * [6] M. Bestvina, M. Handel, Solvable Subgroups of $\mathrm{Out}(F_{n})$ are Virtually Abelian, Geometriae Dedicata volume 104, pages 71–96 (2004). * [7] J. S. Birman, A. Lubotzky, and J. McCarthy, Abelian and solvable subgroups of the mapping class groups, Duke Math. J. 50 (1983), no. 4, 1107–1120. * [8] J. Blanc and S. Cantat, Dynamical degrees of birational transformations of projective surfaces, J. Amer. Math. Soc. 29 (2016), 415-471. * [9] J. Blanc and J.-P. Furter, Length in the Cremona group, Ann. H. Lebesgue 2 (2019), p.187-257. * [10] J. Blanc, S. Lamy and S. Zimmermann, Quotients of higher dimensional Cremona groups, Acta Math. (to appear), arXiv:1901.04145. * [11] A. Borel, Introduction to Arithmetic groups, (translated by L. Pham and translation edited by D. Morris) American Math. Soc., 2019. * [12] A. Borel, J.-P. Serre, Théorèmes de finitude en cohomologie galoisienne, Comment. Math. Helv. 39 (1964) 111–164 (in French). * [13] A. Borel and J. Tits, Groupes réductifs, Inst. Hautes Études. Sci. Publ. Math. 27 (1965), 55-150. * [14] M.R. Bridson, The rhombic dodecahedron and semisimple actions of $Aut(F_{n})$ on $CAT(0)$ spaces, Fund. Math. 214 (2011). * [15] M.R. Bridson, Length functions, curvature and the dimension of discrete groups, Mathematical Research Letters 8, 557–567 (2001). * [16] M.R. Bridson and A. Haefliger, Metric spaces of nonpositive curvature, Grundlehren der Math. Wiss. 319, Springer-Verlag, Berlin, 1999. * [17] M. Bridson, R. Wade, Actions of higher-rank lattices on free groups. Compos. Math. 147, 1573–1580 (2011). * [18] M. Burger and N. Monod, Bounded cohomology of lattices in higher rank Lie groups, J. Eur. Math. Soc. (JEMS) 1, No. 2 (1999), 199–235. * [19] M. Burger and N. Monod, Continuous bounded cohomology and applications to rigidity theory, Geom. Funct. Anal. 12, No. 2 (2002), 219–280. * [20] D. Calegari, scl, MSJ Memoirs, 20. Mathematical Society of Japan, Tokyo, 2009. * [21] S. Cantat, The Cremona group in two variables. Proceedings of the sixth European Congress of Math., 187:211–225, 2013. * [22] D. Calegari, M. Freedman, Distortion in transformation groups, Geometry & Topology 10 (2006) 267–293. * [23] M. Ershov, A. Jaikin-Zapirain, Property (T) for noncommutative universal lattices, Invent. Math. 179 (2010) 303–347. * [24] H. Eynard-Bontemps and A. Navas, Mather invariant, distortion, and conjugates for diffeomorphisms of the interval, arXiv:1912.09305. * [25] B. Farb, A. Lubotzky and Y. N Minsky, Rank-1 phenomena for mapping class groups, Duke Mathematical Journal, Volume 106, Number 3, 581–597, 2001. * [26] B. Farb, H. Masur, Superrigidity and mapping class groups, Topology 37, 1169–1176, 1998. * [27] J. Franks and M. Handel, Area preserving group actions on surfaces, Geom. Topol. 7, (2003), 757–771. * [28] T. Fritz, S. Gadgil, A. Khare, P. Nielsen, L. Silberman and T. Tao, Homogeneous length functions on groups, Algebra and Number Theory, Vol. 12 (2018), No. 7, 1773–1786. * [29] S. Garge and A. Singh, Finiteness of z-classes in reductive groups, Journal of Algebra 554 (2020) 41–53. * [30] S.M. Gersten, A presentation for the special automorphism group of a free group, J. Pure Appl. Algebra, 33 (1984) 269-279. * [31] S.M. Gersten, The automorphism group of a free group is not a CAT(0) group, Proc. Amer. Math. Soc. 121 (1994), 999-1002. * [32] M. Gromov, Hyperbolic groups, in “Essays in Groups Theory” (S. Gersten, ed.), MSRI Publications 5 (1989) 75–263. * [33] T. Haettel, Hyperbolic rigidity of higher rank lattices (with appendix by Vincent Guirardel and Camille Horbez), arXiv:1607.02004. * [34] S. Helgason, Differential geometry, Lie groups and symmetric spaces, Academic Press, New York, 1978. * [35] H. Hu, Some ergodic properties of commuting diffeomorphisms. Ergodic Theory Dynam. Systems 13 (1993), no. 1, 73–100. * [36] H. Hu, E. Shi, Z. Wang, Some ergodic and rigidity properties of discrete Heisenberg group actions, Israel Journal of Mathematics 228 (2018), pages 933–972. * [37] J. Humphreys, Linear algebraic groups, Springer, 1975\. * [38] J. Humphreys, Introduction to Lie algebras and representation theory, Springer-Verlag, New York, 1972. * [39] A. Kaimanovich, H. Masur, The Poisson boundary of the mapping class group. Invent. Math. 125, 221–264 (1996). * [40] A. Katok and B. Hasselblatt, Introduction to the Modern Theory of Dynamical Systems, Encyclopedia of Mathematics and Its Applications 54, Cambridge University Press, 1995. * [41] M. Kolster, On injective stability for $K_{2}$, Proceedings, Oberwolfach, 1980, in “Lecture Notes in Mathematics No. 966,” pp. 128-168, Springer-Verlag, Berlin/Heidelberg/New York, 1982. * [42] B. Kostant, On convexity, the Weyl group and the Iwasawa decomposition, Annales Scientifiques de l’Éole Normale Supérieure, Série 4, 6(1973): 413–455. * [43] G. Levitt, Counting growth types of automorphisms of free groups. Geometric and Functional Analysis, 19(4):1119–1146, 2009. * [44] A. Lubotzky, S. Mozes and M.S. Raghunathan, The word and riemannian metrics on lattices in semisimple Lie groups, IHES Publ. Math. 91 (2000), 5-53. * [45] A. Lonjou and C. Urech, Actions of Cremona groups on CAT(0) cube complexes, arXiv:2001.00783v1. * [46] G. Lusztig, On the finiteness of the number of unipotent classes, Invent. Math. 34 (3) (1976) 201–213. * [47] B.A. Magurn. An algebraic introduction to K-theory, Cambridge University Press, 2002. * [48] M. Mimura, Superrigidity from Chevalley groups into acylindrically hyperbolic groups via quasi-cocycles, Journal of the European Mathematical Society, Volume 20, Issue 1, 2018, pp. 103-117. * [49] N. Monod, Continuous bounded cohomology of locally compact groups, Lecture Notes in Mathematics 1758, Springer-Verlag, Berlin, (2001). * [50] A. Navas, On conjugates and the asymptotic distortion of 1-dimensional $C^{1+bv}$ diffeomorphisms, arXiv:1811.06077. * [51] A. Navas, Groups of Circle Diffeomorphisms, Univ. Chicago Press, Chicago, 2011. * [52] V. I. Oseledec, A multiplicative ergodic theorem. Lyapunov characteristic numbers for dynamical systems. Trans. Moscow Math. Soc. 19 (1968), 197-221. * [53] L. Polterovich, Growth of maps, distortion in groups and symplectic geometry, Invent. Math. 150, No. 3 (2002), 655–686. * [54] P. Py, Some remarks on area-preserving actions of lattices, in Geometry, Rigidity and Group Actions, The University of Chicago Press, Chicago and London (2011). * [55] J-P Serre, Trees, Springer Monographs in Mathematics, Springer, Berlin (2003). Translated from the French original by John Stillwell, Corrected 2nd printing of the 1980 English translation. * [56] M. Steele, Probability theory and combinatorial optimization, SIAM, Philadelphia (1997). * [57] A. Thom, Low degree bounded cohomology and L2-invariants for negatively curved groups, Groups Geom. Dyn. 3 (2009), 343–358. * [58] P. Walters, An Introduction to Ergodic Theory, Springer-Verlag, New York, 1981. * [59] C. Weibel, The K-book: an introduction to algebraic K-theory. Graduate studies in mathematics, 145, American mathematical society, Providence, 2013. * [60] D. Witte, Arithmetic Groups of Higher $\mathbb{Q}$-Rank Cannot Act on 1-Manifolds, Proceedings of the American Mathematical Society, Vol. 122, No. 2 (Oct. 1994), pp. 333-340. * [61] J. Wolf, Spaces of Constant Curvature, Mc Graw-Hill Book Company, New York, 1972. * [62] R. Zimmer, Ergodic theory and semisimple groups, Birkhäuser Verlag, Basel, 1984. NYU-ECNU Institute of Mathematical Sciences at NYU Shanghai, 3663 Zhongshan Road North, Shanghai, 200062, China E-mail<EMAIL_ADDRESS>
# The convergence rate of of multivariate operators on simplex in Orlicz space111This work was supported by the Ningxia Science and Technology Department [grant numbers 2019BEB04003]; and Ningxia Education Department [grant numbers NXYLXK2019A8] Wan Ma Lihong Chang Yongxia Qiang<EMAIL_ADDRESS>School of Mathematics and Computer Science, Ningxia Normal University, Guyuan City, Ningxia 756000, People’s Republic of China ###### Abstract The approximation of functions in Orlicz space by multivariate operators on simplex is considered. The convergence rate is given by using modulus of smoothness. ###### keywords: Stancu-Kantorovič operator, Meyer-König-Zeller operator, Convergence rate , Orlicz space ††journal: J. Math. Anal. Appl. ## 1 Introduction Let $\Phi(u)$ be a N-function, $\Psi$ be the complementary function of $\Phi$. We will say that $\Phi$ satisfies the $\Delta_{2}$-condition if $\Phi(2u)\leq c\Phi(u)$ for any $u\geq u_{0}\geq 0$ with some constant $c$ independent of $u.$ For $\triangle=\\{x=(x_{1},x_{2})\in\mathrm{R^{2}}:x_{1}+x_{2}\leq 1,x_{1},x_{2}\geq 0\\},$ the Orlicz space $L_{\Phi}^{\ast}(\triangle)$ corresponding to the function $\Phi$ consists of all Lebesgue-measurable functions $f(x)$ on $\triangle$ such that integral $\int_{\triangle}f(x)g(x)\mbox{d}x$ is finite for any measurable functions $g(x)$ with $\int_{\triangle}\Psi(g(x))\mbox{d}x<\infty.$ It is well-known that the space $L_{\Phi}^{\ast}(\triangle)$ becomes a complete normed space with Orlicz norm $\|f\|_{\Phi}=\|f\|_{[\Phi,\triangle]}=\sup\left\\{\left|\int_{\triangle}f(x)g(x)\mbox{d}x\right|:\int_{\triangle}\Psi(g(x))\mbox{d}x\leq 1\right\\}.$ It can be proved that $\|f\|_{\Phi}=\inf_{\alpha>0}\frac{1}{\alpha}\left\\{1+\int_{\triangle}\Phi(\alpha f(x))\mbox{d}x\right\\}.$ See [1] for the above. For $f\in L_{\Phi}^{\ast}(\triangle),$ we first extend $f(x)$ from $\triangle$ to $D=[0,1]\times[0,1]$ according to $f(x)=f(1-x)$, and then extend $f(x)$ to $\mathrm{R^{2}}$ with period 1. The nonnegative function $\Omega_{\mathrm{R^{2}}}^{2}(f,r)_{\Phi}=\sup\\{\omega_{h}^{2}(f,r)_{\Phi}:h=(h_{1},h_{2})\in\mathrm{R^{2}},|h|=1\\}$ of the variable $r\geq 0$ will be called the 2-th order modulus of continuity of the function $f\in L_{\Phi}^{\ast}(\triangle)$ in the Orlicz norm $\Phi.$ Here, $|h|=\sqrt{h_{1}^{2}+h_{2}^{2}},$ and $\omega_{h}^{2}(f,r)_{\Phi}=\sup_{|t|\leq r}\|f(x+th)+f(x-th)-2f(x)\|_{\Phi}$ is the 2-th order modulus of continuity in the direction $h$ of the function $f.$ For any Lebesgue-measurable function $f(x)$ on $\triangle,$ the functional $K_{n}(f;x)=\sum_{k_{1}=0}^{\infty}\sum_{k_{2}=0}^{\infty}{\tilde{p}_{n,k_{1},k_{2}}(x)}c_{n,k_{1},k_{2}}\int_{\triangle_{k_{1},k_{2}}}f(u)\mbox{d}u$ (1.1) is called Meyer-König-Zeller-$\mathrm{Kantorovi\check{c}}$ operator on $\triangle$ [2]; the functional $K_{n,s}(f;x)=\sum_{k+l\leq n}{b_{n,k,l,s}(x)}(n+2)^{2}\int_{I_{n,k,l}}f(u)\mbox{d}u$ (1.2) is called Stancu-$\mathrm{Kantorovi\check{c}}$ operator on $\triangle$[3], where $x\in\triangle,$ $n\in\mathrm{Z^{+}},$ $k,s,l$ are nonnegative integers, $0\leq s<\frac{n}{2},$ and ${\tilde{p}_{n,k_{1},k_{2}}(x)}=\frac{(n+k_{1}+k_{2})!}{n!k_{1}!k_{2}!}x_{1}^{k_{1}}x_{2}^{k_{2}}(1-x_{1}-x_{2})^{n+1},$ $c_{n,k_{1},k_{2}}=\frac{(n+k_{1}+k_{2})^{2}(n+k_{1}+k_{2}+1)^{2}}{(n+k_{1})(n+k_{2})},$ $\triangle_{k_{1},k_{2}}=\left[\frac{k_{1}}{n+k_{1}+k_{2}},\frac{k_{1}+1}{n+k_{1}+k_{2}+1}\right]\times\left[\frac{k_{2}}{n+k_{1}+k_{2}},\frac{k_{2}+1}{n+k_{1}+k_{2}+1}\right],$ $I_{n,k,l}=\left[\frac{k}{n+2},\frac{k+1}{n+2}\right]\times\left[\frac{l}{n+2},\frac{l+1}{n+2}\right],$ ${p_{n,k,l}(x)}=\frac{n!}{k!l!(n-k-l)!}x_{1}^{k}x_{2}^{l}(1-x_{1}-x_{2})^{n-k-l},$ ${b_{n,k,l,s}(x)}=\begin{cases}(1-x_{1}-x_{2})p_{n-s,k,l}(x),\quad k+l\leq n-s,0\leq k,l<s;\\\ (1-x_{1}-x_{2})p_{n-s,k,l}(x)+x_{1}p_{n-s,k-s,l}(x),\\\ \qquad\qquad\qquad\qquad\qquad\quad\;\;\;k+l\leq n-s,s\leq k,0\leq l<s;\\\ (1-x_{1}-x_{2})p_{n-s,k,l}(x)+x_{2}p_{n-s,k,l-s}(x),\\\ \qquad\qquad\qquad\qquad\qquad\quad\;\;\;k+l\leq n-s,s\leq l,0\leq k<s;\\\ (1-x_{1}-x_{2})p_{n-s,k,l}(x)+x_{1}p_{n-s,k-s,l}(x)+x_{2}p_{n-s,k,l-s}(x,y),\\\ \qquad\qquad\qquad\qquad\qquad\qquad k+l\leq n-s,s\leq k,s\leq l;\\\ x_{1}p_{n-s,k-s,l}(x),\quad\qquad\qquad n-s<k+l\leq n,s\leq k,0\leq l<s;\\\ x_{2}p_{n-s,k,l-s}(x),\quad\qquad\qquad n-s<k+l\leq n,0\leq k<s,s\leq l;\\\ x_{1}p_{n-s,k-s,l}(x)+x_{2}p_{n-s,k,l-s}(x),\\\ \qquad\qquad\qquad\qquad\qquad\qquad\;n-s<k+l\leq n,s\leq k,s\leq l.\end{cases}$ Denote $C$ a constant independent of $f,n,s,$ and its value can be different in different positions. The convergence rate of the operators (1.1) and (1.2) in space $L_{p}$ has been studied (see [2, 3]). This paper intends to investigate their convergence in space $L_{\Phi}^{\ast}(\triangle),$ and the main results are as follows. ###### Theorem 1.1. For $f\in L_{\Phi}^{\ast}(\triangle),$ if a N-function $\Phi(u)$ satisfies the $\Delta_{2}$-condition, then $\|K_{n}(f)-f\|_{\Phi}\leq C\left(\frac{1}{n}\|f\|_{\Phi}+\Omega_{\mathrm{R^{2}}}^{2}\left(f,\sqrt{\frac{1}{n}}\right)_{\Phi}\right).$ ###### Theorem 1.2. For $f\in L_{\Phi}^{\ast}(\triangle),$ if a N-function $\Phi(u)$ satisfies the $\Delta_{2}$-condition, then $\|K_{n,s}(f)-f\|_{\Phi}\leq C\left(\frac{1}{n}\|f\|_{\Phi}+\Omega_{\mathrm{R^{2}}}^{2}\left(f,\sqrt{\frac{1}{n}}\right)_{\Phi}\right).$ ## 2 Lemmas ###### Lemma 2.1. $K_{n}$ is a bounded linear operator, and $\|K_{n}\|_{\Phi}\leq 2.$ ###### Proof. The linearity of $K_{n}$ is obvious. The following proves $\|K_{n}\|_{\Phi}\leq 2.$ After calculation, we can get $mes\triangle_{k_{1},k_{2}}=\frac{(n+k_{1})(n+k_{2})}{(n+k_{1}+k_{2})^{2}(n+k_{1}+k_{2}+1)^{2}},$ $\int_{\triangle}\tilde{p}_{n,k_{1},k_{2}}(x)\mbox{d}x=\frac{n+1}{(n+k_{1}+k_{2}+3)(n+k_{1}+k_{2}+2)(n+k_{1}+k_{2}+1)}.$ By using the _Lemma 1_ in [2], Jensen inequality of convex function and the _Theorem 1.4_ in [1], we obtain $\displaystyle\|K_{n}(f)\|_{\Phi}$ $\displaystyle=\inf_{\alpha>0}\frac{1}{\alpha}\left\\{1+\int_{\triangle}\Phi\left(\alpha K_{n}(f;x)\right)\mbox{d}x\right\\}$ $\displaystyle=\inf_{\alpha>0}\frac{1}{\alpha}\left\\{1+\int_{\triangle}\Phi\left(\alpha\sum_{k_{1}=0}^{\infty}\sum_{k_{2}=0}^{\infty}\tilde{p}_{n,k_{1},k_{2}}(x)c_{n,k_{1},k_{2}}\int_{\triangle_{k_{1},k_{2}}}f(u)\mbox{d}u\right)\mbox{d}x\right\\}$ $\displaystyle\leq\inf_{\alpha>0}\frac{1}{\alpha}\left\\{1+\sum_{k_{1}=0}^{\infty}\sum_{k_{2}=0}^{\infty}\int_{\triangle}\tilde{p}_{n,k_{1},k_{2}}(x)\mbox{d}xc_{n,k_{1},k_{2}}\int_{\triangle_{k_{1},k_{2}}}\Phi\left(\alpha f(u)\right)\mbox{d}u\right\\}$ $\displaystyle\leq\inf_{\alpha>0}\frac{1}{\alpha}\\{1+2\sum_{k_{1}=0}^{\infty}\sum_{k_{2}=0}^{\infty}\int_{\triangle_{k_{1},k_{2}}}\Phi\left(\alpha f(u)\right)\mbox{d}u\\}$ $\displaystyle\leq\inf_{\alpha>0}\frac{1}{\alpha}\left\\{1+\int_{\triangle}\Phi\left(2\alpha f(u)\right)\mbox{d}u_{1}\mbox{d}u_{2}\right\\}$ $\displaystyle=\|2f\|_{\Phi}$ $\displaystyle=2\|f\|_{\Phi}.$ Namely $\|K_{n}\|_{\Phi}\leq 2.\qed$ ###### Lemma 2.2. $K_{n,s}$ is a bounded linear operator, and $\|K_{n,s}\|_{\Phi}\leq 12.$ ###### Proof. The linearity of $K_{n,s}$ is obvious. The following proves $\|K_{n,s}\|_{\Phi}\leq 12.$ After calculation, we can get $mesI_{n,k,l}=mesI_{n,k+s,l}=mesI_{n,k,l+s}=\frac{1}{(n+2)^{2}},$ $\int_{\triangle}p_{n-s,k,l}(x)\mbox{d}x=\frac{1}{(n-s+2)(n-s+1)}.$ By using the _Lemma 2.1_ in [3], Jensen inequality of convex function and the _Theorem 1.4_ in [1], we obtain $\displaystyle\|K_{n,s}(f)\|_{\Phi}$ $\displaystyle\leq\inf_{\alpha>0}\frac{1}{\alpha}\left\\{1+\sum_{k+l\leq n-s}\int_{\triangle}p_{n-s,k,l}(x)\mbox{d}x\left(\int_{I_{n,k,l}}+\int_{I_{n,k+s,l}}\int_{I_{n,k,l+s}}\right)\Phi\left(\alpha f(u)\right)\mbox{d}u\right\\}$ $\displaystyle\leq\inf_{\alpha>0}\frac{1}{\alpha}\left\\{1+\frac{3(n+2)^{2}}{(n-s+2)(n-s+1)}\int_{\triangle}\Phi(\alpha f(u))\mbox{d}u\right\\}.$ $\displaystyle\leq\inf_{\alpha>0}\frac{1}{\alpha}\left\\{1+\int_{\triangle}\Phi\left(\alpha 12f(u)\right)\mbox{d}u\right\\}$ $\displaystyle=\|12f\|_{\Phi}$ $\displaystyle=12\|f\|_{\Phi}.$ Namely $\|K_{n,s}\|_{\Phi}\leq 12.\qed$ ###### Lemma 2.3. The following holds for (1.1). $K_{n}(1;x)=1,\quad K_{n}\left((u_{i}-x_{i})^{i};x\right)\leq\frac{C}{n},\quad i=1,2.$ ###### Lemma 2.4. The following holds for (1.2). $K_{n,s}(1;x)=1,\quad K_{n,s}\left((u_{i}-x_{i})^{i};x\right)\leq\frac{C}{n},\quad i=1,2.$ The proof of the Lemma 2.3 and Lemma 2.4 can be obtained from the _Lemma 3_ in [2] and the _Lemma 2.1_ in [3]. ###### Lemma 2.5. If we denote $f_{r}$ the Steklov mean function for $f\in L_{\Phi}^{\ast}(\triangle),$ i.e. $f_{r}(x)=\frac{1}{r^{4}}\int_{[-r/2,r/2]^{4}}f(x+u+v)\mbox{d}s\mbox{d}t,$ then $\left\|{f_{r}}\right\|_{\Phi}\leq C\left\|f\right\|_{\Phi},$ (2.1) $\left\|{f-f_{r}}\right\|_{\Phi}\leq C\Omega_{R^{2}}^{2}\left({f,r}\right)_{\Phi},$ (2.2) $\left\|{\frac{{\partial f_{r}}}{{\partial x_{1}}}}\right\|_{\Phi}\leq C\left({\left\|{f_{r}}\right\|_{\Phi}+\left\|{\frac{{\partial^{2}f_{r}}}{{\partial x_{1}^{2}}}}\right\|_{\Phi}}\right),$ (2.3) $\left\|{\frac{{\partial f_{r}}}{{\partial x_{2}}}}\right\|_{\Phi}\leq C\left({\left\|{f_{r}}\right\|_{\Phi}+\left\|{\frac{{\partial^{2}f_{r}}}{{\partial x_{2}^{2}}}}\right\|_{\Phi}}\right),$ (2.4) $\left\|{\frac{{\partial^{2}f_{r}}}{{\partial x_{1}^{2}}}}\right\|_{\Phi}+\left\|{\frac{{\partial^{2}f_{r}}}{{\partial x_{2}^{2}}}}\right\|_{\Phi}+\left\|{\frac{{\partial^{2}f_{r}}}{{\partial x_{1}\partial x_{2}}}}\right\|_{\Phi}\leq\frac{C}{{r^{2}}}\Omega_{R^{2}}^{2}\left({f,r}\right)_{\Phi}.$ (2.5) (2.1), (2.2), (2.5) can be directly verified, and the proof of (2.3),(2.4) is similar to that of the _Lemma 1a_ in [4]. If N-function $\Phi$ satisfies the $\Delta_{2}$-condition, then $L_{\Phi}^{\ast}$ is separable. This leads to the following conclusion[5]. ###### Lemma 2.6. If N-function $\Phi$ satisfies the $\Delta_{2}$-condition, then $\left\|\sup_{u_{1}\neq x_{1}}\frac{1}{u_{1}-x_{1}}\int_{x_{1}}^{u_{1}}\left|\frac{\partial^{2}f_{r}(\xi,x_{2})}{\partial\xi^{2}}\right|\mbox{d}\xi\right\|_{\Phi}\leq C\left\|{\frac{{\partial^{2}f_{r}}}{{\partial x_{1}^{2}}}}\right\|_{\Phi},$ $\left\|\sup_{u_{2}\neq x_{2}}\frac{1}{u_{2}-x_{2}}\int_{x_{2}}^{u_{2}}\left|\frac{\partial^{2}f_{r}(x_{1},\eta)}{\partial\eta^{2}}\right|\mbox{d}\eta\right\|_{\Phi}\leq C\left\|{\frac{{\partial^{2}f_{r}}}{{\partial x_{2}^{2}}}}\right\|_{\Phi},$ $\left\|\sup_{u_{2}\neq x_{2}}\frac{1}{u_{2}-x_{2}}\int_{x_{2}}^{u_{2}}\left(\sup_{u_{1}\neq x_{1}}\int_{x_{1}}^{u_{1}}\left|\frac{\partial^{2}f_{r}(\xi,\eta)}{\partial\xi\partial\eta}\right|\mbox{d}\xi\right)\mbox{d}\eta\right\|_{\Phi}\leq C\left\|{\frac{{\partial^{2}f_{r}}}{{\partial x_{1}\partial x_{2}}}}\right\|_{\Phi}.$ ## 3 Proof of the main results The proof of the Theorem 1.1 and Theorem 1.2 is similar, so only the Theorem 1.1 is proved below. ###### Proof. Because $\displaystyle f_{r}(u)-f_{r}(x)=(u_{1}-x_{1})\frac{\partial f_{r}(x)}{\partial x_{1}}$ $\displaystyle+(u_{2}-x_{2})\frac{\partial f_{r}(x)}{\partial x_{2}}+\int_{x_{1}}^{u_{1}}(u_{1}-\xi)\frac{\partial^{2}f_{r}(\xi,x_{2})}{\partial\xi^{2}}\mbox{d}\xi$ $\displaystyle+\int_{x_{2}}^{u_{2}}(u_{2}-\eta)\frac{\partial^{2}f_{r}(x_{1},\eta)}{\partial\eta^{2}}\mbox{d}\eta+\int_{x_{1}}^{u_{1}}\int_{x_{2}}^{u_{2}}\frac{\partial^{2}f_{r}(\xi,\eta)}{\partial\xi\partial\eta}\mbox{d}\eta\mbox{d}\xi,$ so $\displaystyle\left|K_{n}(f_{r};x)-f_{r}(x)\right|\leq$ $\displaystyle\left|K_{n}((u_{1}-x_{1});x)\right|\left|\frac{\partial f_{r}(x)}{\partial x_{1}}\right|+\left|K_{n}((u_{2}-x_{2});x)\right|\left|\frac{\partial f_{r}(x)}{\partial x_{2}}\right|+$ $\displaystyle\left|K_{n}((u_{1}-x_{1})^{2};x)\right|\left(\sup_{u_{1}\neq x_{1}}\frac{1}{u_{1}-x_{1}}\int_{x_{1}}^{u_{1}}\left|\frac{\partial^{2}f_{r}(\xi,x_{2})}{\partial\xi^{2}}\right|\mbox{d}\xi\right)+$ $\displaystyle\left|K_{n}((u_{2}-x_{2})^{2};x)\right|\left(\sup_{u_{2}\neq x_{2}}\frac{1}{u_{2}-x_{2}}\int_{x_{2}}^{u_{2}}\left|\frac{\partial^{2}f_{r}(x_{1},\eta)}{\partial\eta^{2}}\right|\mbox{d}\eta\right)+$ $\displaystyle\left|K_{n}(|u_{1}-x_{1}||u_{2}-x_{2}|;x)\right|\left(\sup_{u_{2}\neq x_{2}}\frac{1}{u_{2}-x_{2}}\int_{x_{2}}^{u_{2}}\left(\sup_{u_{1}\neq x_{1}}\int_{x_{1}}^{u_{1}}\left|\frac{\partial^{2}f_{r}(\xi,\eta)}{\partial\xi\partial\eta}\right|\mbox{d}\xi\right)\mbox{d}\eta\right).$ Noticing $|K_{n}(|u_{1}-x_{1}||u_{2}-x_{2}|;x)|\leq\frac{1}{2}|K_{n}((u_{1}-x_{1})^{2};x)|+|K_{n}((u_{2}-x_{2})^{2};x)|,$ we continue the above estimation using the Lemma 2.3, Lemma 2.5 and Lemma 2.6. $\displaystyle\|K_{n}(f_{r})-f_{r}\|_{\Phi}$ $\displaystyle\leq\frac{C}{n}\left(\left\|\frac{\partial f_{r}}{\partial x_{1}}\right\|_{\Phi}+\left\|\frac{\partial f_{r}}{\partial x_{2}}\right\|_{\Phi}+\left\|\frac{\partial^{2}f_{r}}{\partial x_{1}^{2}}\right\|_{\Phi}+\left\|\frac{\partial^{2}f_{r}}{\partial x_{2}^{2}}\right\|_{\Phi}+\left\|\frac{\partial^{2}f_{r}}{\partial x_{1}\partial x_{2}}\right\|_{\Phi}\right)$ $\displaystyle\leq\frac{C}{n}\left(\left\|f\right\|_{\Phi}+\frac{1}{r^{2}}\Omega_{\mathrm{R}^{2}}^{2}(f,r)_{\Phi}\right).$ If $r=\sqrt{\frac{1}{n}},$ then $\|K_{n}(f_{r})-f_{r}\|_{\Phi}\leq\frac{C}{n}\left(\|f\|_{\Phi}+n\Omega_{\mathrm{R}^{2}}^{2}\left(f,\sqrt{\frac{1}{n}}\right)_{\Phi}\right).$ For $f\in L_{\Phi}^{\ast}(\triangle),$ using the Lemma 2.1 and Lemma 2.5 we get $\displaystyle\|K_{n}(f)-f\|_{\Phi}$ $\displaystyle\leq\|K_{n}(f)-K_{n}(f_{r})\|_{\Phi}+\|K_{n}(f_{r})-f_{r}\|_{\Phi}+\|f_{r}-f\|_{\Phi}$ $\displaystyle\leq 3\|f_{r}-f\|_{\Phi}+\|K_{n}(f_{r})-f_{r}\|_{\Phi}$ $\displaystyle\leq C\Omega_{\mathrm{R}^{2}}^{2}(f,r)_{\Phi}+C\left(\frac{1}{n}\|f\|_{\Phi}+\Omega_{\mathrm{R}^{2}}^{2}\left(f,\sqrt{\frac{1}{n}}\right)_{\Phi}\right)$ $\displaystyle\leq C\left(\frac{1}{n}\|f\|_{\Phi}+\Omega_{\mathrm{R}^{2}}^{2}(f,\sqrt{\frac{1}{n}})_{\Phi}\right).\hskip 99.58464pt\qed$ ## 4 Remark If N-function $\Phi(u)=u^{p}$ $(1<p<\infty),$ then $L_{\Phi}^{\ast}(\triangle)=L_{p}.$ Thus the corresponding conclusions in [2] and [3] can be obtained from the Theorem 1.1 and Theorem 1.2. ## References ## References * [1] C. X. Wu, T. F. Wang, Orlicz space and its applications, Hei Long Jiang Science and Technology Press, Beijing, 1983. * [2] J. Y. Xiong, Approximation by the multivariate meyer-konig-zeller-kantorovic operator, Journal of Beijing Normal University (Natural Science) 30 (4) (1994) 439–447. * [3] R. Y. Y. Jing Yi Xiong, F. L. Cao, Approximation theorems of the stancu-kantorovic polynomials on a simplex, Journal of Qufu Normal University 19 (4) (1993) 29–34. * [4] A. R. K. Ramazanov, On approximation by polynomials and rational functions in orlicz spaces, Analysis Mathematica 19 (1984) 117–132. * [5] D. L. Xie, The order of approximation by positive continuous operator in orlicz space, Journal of Hangzhou University 8 (2) (1981) 142–146.
# High efficient multipartite entanglement purification using hyperentanglement Lan Zhou,1 Pei-Shun Yan,2 Wei Zhong,2 Yu-Bo<EMAIL_ADDRESS>1 School of Science, Nanjing University of Posts and Telecommunications, Nanjing, 210003, China 2Institute of Quantum Information and Technology, Nanjing University of Posts and Telecommunications, Nanjing, 210003, China ###### Abstract Multipartite entanglement plays an important role in controlled quantum teleportation, quantum secret sharing, quantum metrology and some other important quantum information branches. However, the maximally multipartite entangled state will degrade into the mixed state because of the noise. We present an efficient multipartite entanglement purification protocol (EPP) which can distill the high quality entangled states from low quality entangled states for $N$-photon systems in a Greenberger-Horne-Zeilinger (GHZ) state in only linear optics. After performing the protocol, the spatial-mode entanglement is used to purify the polarization entanglement and one pair of high quality polarization entangled state will be obtained. This EPP has several advantages. Firstly, with the same purification success probability, this EPP only requires one pair of multipartite GHZ state, while existing EPPs usually require two pairs of multipartite GHZ state. Secondly, if consider the practical transmission and detector efficiency, this EPP may be extremely useful for the ratio of purification efficiency is increased rapidly with both the number of photons and the transmission distance. Thirdly, this protocol requires linear optics and does not add additional measurement operations, so that it is feasible for experiment. All these advantages will make this protocol have potential application for future quantum information processing. ###### pacs: 03.67.Lx ## I Introduction Entanglement plays an important role in quantum communication and computation. Quantum teleportation teleportation , quantum key distributionQKD , dense codingdensecoding , quantum secure direct communicationQSDC1 ; QSDC2 ; QSDC3 , distributed quantum computing computation , distributed secure quantum machine learning DSQML , and other important branches all require the parties to share the entanglement. Besides the bipartite entanglement, multipartite entanglement also plays an important role in controlled quantum teleportation cteleportation1 ; cteleportation2 , quantum secret sharingQSS1 ; QSS2 ; QSS3 , quantum state sharingQSTS1 ; QSTS2 ; QSTS3 , quantum metrologymetrology1 ; metrology2 , and so on. Recently, the multipartite entangled state named Greenberger-Horne-Zeilinger (GHZ) states was been used in some important quantum communication experiment, such as long-distance measurement-device- independent multiparty quantum communication chenzb , equitable multiparty quantum communication without a trusted third party jeong1 and quantum teleportation of shared quantum secret jeong2 . The GHZ state also have been realized with superconducting system superconduct1 ; superconduct2 , trapped ions ion , and photonic system photon . The detection of the multipartite entanglement structure has also been reported detection . In a practical application, the quantum system should inevitably interact with its environment, and the environment noise will degrade the entanglement. In general, the decoherence will make the maximally entangled state become a mixed state. The degraded entanglement will decrease the efficiency of the quantum communication and it also will make the quantum communication become insecure repeater . Entanglement purification is a powerful tool to distill the high quality entangled states from the low quality entangled states purification1 ; purification2 ; purification3 ; purification4 ; purification5 ; addpurification1 ; addpurification3 ; experiment2 ; purification6 ; purification7 ; purification8 ; purification9 ; purification11 ; purification12 ; purification13 ; purification14 ; purification15 ; purification16 ; purification17 ; purification18 ; purification19 ; purification20 ; purification21 ; purification22 ; purification23 ; purification24 ; purification25 ; purification26 ; addpurification2 ; purification27 ; addpurification4 ; shengprl . Entanglement purification has been widely discussed since Bennett et al. proposed the first entanglement purification protocol (EPP) purification1 . EPPs for bipartite system were proposed in photonic system purification3 ; purification4 ; purification5 ; purification7 ; purification8 ; purification9 , electionsaddpurification1 , quantum dotspurification11 ; purification12 , atoms experiment2 ; purification13 and so on. For example, in 2001, Pan et al. proposed the polarization EPP with only feasible linear optics elements purification3 . In 2008, the EPP with spontaneous parametric down conversion (SPDC) source based on cross-Kerr nonlinearity was proposed purification7 . In this EPP, the purified high quality entangled state can be remained for further application and the remained entangled states can also be repeated to perform the purification to obtain the higher entangled states. The experiments of entanglement purification in optical system were also reported purification4 . In 2017, Chen et al. realized the nested entanglement purification for quantum repeaters. In this experiment, the entanglement purification and entanglement swapping can be realized simultaneously purification17 . On the other hand, the double-pair noise components from the SPDC source can be eliminated automatically. This work was extended to the multi-copy cases addpurification2 . The optimal entanglement purification was also investigated purification26 . Recently, the first high efficient and long-distance entanglement purification using hyperentanglement was demonstrated shengprl . The hyperentanglement was first distributed to 11 km and the spatial entanglement was used to purification polarization entanglement. The authors also demonstrated its powerful application in entanglement-based QKD. For multipartite system, Murao et al. described the first EPP with controlled- not (CNOT) gate multipurification1 . In 2003, Dur et al. described the EPP for Graph state multipurification2 . In 2007, this protocol was extended to high- dimension multipartite system with the generalized CNOT gate multipurification3 . In 2008, the multipartite EPP for polarization entangled states with cross-Kerr nonlinearity was proposed multipurification4 . In 2011, Deng proposed the multipartite EPP using entanglement link from subspace multipurification5 . In his protocol, the discussed items in conventional EPPs still have entanglement in a subspace and they can be reused with entanglement link. There are another kind of EPP for multipartite entanglement system, named deterministic EPP multipurification6 ; multipurification7 . In these EPPs, they exploit the hyperentanglement to perform the purification. Such EPPs are based on the condition that the spatial mode entanglement is robust and it does not suffer from the noise. Therefore, the spatial mode entanglement or the frequency entanglement can be completely transformed to the polarization entanglement. In conventional EPPs, they all require two copies of low quality entangled states. After performing the CNOT or similar operations, one pair of high quality entangled state is remained, if the purification is successful. On the other hand, if the purification is a failure, both pairs should be discarded. In this paper, we will describe an efficient EPP for multipartite polarization entangled systems in a GHZ state, inspired the idea of Ref.shengprl . Different from existing EPPs for multipartite system, this protocol only requires one pair of hyperentangled state. By performing the CNOT gate between two degrees of freedom, the spatial entanglement is consumed. Therefore, if the protocol is successful, one can obtain a high quality polarization entangled state. This EPP is based on linear optics and it is also feasible for current experiment condition. This protocol is organized as follows. In Sec.II, we describe this EPP for bit-flip error. In Sec.III, we describe this protocol for phase-flip error. In Sec.IV, we extend this EPP to a general case for arbitrary N-photon GHZ state. In Sec.V, we present a discussion. Finally, in Sec. VI, we will provide a conclusion. ## II Multipartite entanglement purification for bit-flip error In this section, we describe this EPP with a simple example. The three-photon GHZ states can be written as follows. $\displaystyle|\Phi_{0}^{\pm}\rangle_{ABC}=\frac{1}{\sqrt{2}}(|H\rangle_{A}|H\rangle_{B}|H\rangle_{C}\pm|V\rangle_{A}|V\rangle_{B}|V\rangle_{C}),$ $\displaystyle|\Phi_{1}^{\pm}\rangle_{ABC}=\frac{1}{\sqrt{2}}(|H\rangle_{A}|H\rangle_{B}|V\rangle_{C}\pm|V\rangle_{A}|V\rangle_{B}|H\rangle_{C}),$ $\displaystyle|\Phi_{2}^{\pm}\rangle_{ABC}=\frac{1}{\sqrt{2}}(|H\rangle_{A}|V\rangle_{B}|H\rangle_{C}\pm|V\rangle_{A}|H\rangle_{B}|V\rangle_{C}),$ $\displaystyle|\Phi_{3}^{\pm}\rangle_{ABC}=\frac{1}{\sqrt{2}}(|V\rangle_{A}|H\rangle_{B}|H\rangle_{C}\pm|H\rangle_{A}|V\rangle_{B}|V\rangle_{C}).$ (1) Here $|H\rangle$ denotes the horizonal polarization and $|V\rangle$ denotes the vertical polarization of the photon, respectively. Figure 1: Schematic drawing showing the principle of entanglement purification for bit-flip error. HWP45 is the half-wave plate setting as 45∘. The PBS is the polarization beam splitter which can transmit the $|H\rangle$ polarization and reflect the $|V\rangle$ polarization photon. The polarizing beam displacers (BD) can couple $|H\rangle$ and $|V\rangle$ polarization components from different spatial modes. The spatial mode GHZ states can be written as follows. $\displaystyle|\phi_{0}^{\pm}\rangle_{ABC}=\frac{1}{\sqrt{2}}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{1}\rangle_{C}\pm|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{2}\rangle_{C}),$ $\displaystyle|\phi_{1}^{\pm}\rangle_{ABC}=\frac{1}{\sqrt{2}}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{2}\rangle_{C}\pm|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{1}\rangle_{C}),$ $\displaystyle|\phi_{2}^{\pm}\rangle_{ABC}=\frac{1}{\sqrt{2}}(|a_{1}\rangle_{A}|b_{2}\rangle_{B}|c_{1}\rangle_{C}\pm|a_{2}\rangle_{A}|b_{1}\rangle_{B}|c_{2}\rangle_{C}),$ $\displaystyle|\phi_{3}^{\pm}\rangle_{ABC}=\frac{1}{\sqrt{2}}(|a_{2}\rangle_{A}|b_{1}\rangle_{B}|c_{1}\rangle_{C}\pm|a_{1}\rangle_{A}|b_{2}\rangle_{B}|c_{2}\rangle_{C}).$ (2) Here a1, a2, b1, b2, c1 and c2 are the spatial modes as shown in Fig. 1. The PBS is the polarization beam splitter which can transmit the $|H\rangle$ polarization and reflect the $|V\rangle$ polarization photon. The polarizing beam displacers (BD) can couple $|H\rangle$ and $|V\rangle$ polarization components from different spatial modes. HWP45 is the half-wave plate setting as 45∘. It can convert $|H\rangle$ polarization to $|V\rangle$ and $|V\rangle$ to $|H\rangle$ , respectively. The entanglement source $S$ emits a pair of hyperentangled GHZ state of the form $\displaystyle|\Psi\rangle_{ABC}=|\Phi_{0}^{+}\rangle_{ABC}\otimes|\phi_{0}^{+}\rangle_{ABC}.$ (3) The hyperentangled GHZ state is distributed to Alice, Bob and Charlie, respectively. During distribution, if a bit-flip error occurs on both the polarization and spatial modes entangled state, the original state $|\Psi\rangle_{ABC}$ will become a mixed state as $\displaystyle\rho_{ABC}=\rho^{P}_{ABC}\otimes\rho^{S}_{ABC}.$ (4) Here $\rho^{P}_{ABC}$ and $\rho^{S}_{ABC}$ are the mixed state in polarization and spatial modes. They can be written as $\displaystyle\rho^{P}_{ABC}=F_{1}|\Phi_{0}^{+}\rangle_{ABC}\langle\Phi_{0}^{+}|+(1-F_{{}_{1}})|\Phi_{1}^{+}\rangle_{ABC}\langle\Phi_{1}^{+}|,$ (5) and $\displaystyle\rho^{S}_{ABC}=F_{2}|\phi_{0}^{+}\rangle_{ABC}\langle\phi_{0}^{+}|+(1-F_{{}_{2}})|\phi_{1}^{+}\rangle_{ABC}\langle\phi_{1}^{+}|.$ (6) From Eq.(4), the mixed state $\rho_{ABC}$ can be described as follows. With the probability of $F_{1}\otimes F_{2}$, it is in the state $|\Phi_{0}^{+}\rangle_{ABC}\otimes|\phi_{0}^{+}\rangle_{ABC}$. With the probability of $(1-F_{1})(1-F_{2})$, it is in the state $|\Phi_{1}^{+}\rangle_{ABC}\otimes|\phi_{1}^{+}\rangle_{ABC}$. With the probability of $F_{1}(1-F_{2})$ and $(1-F_{{}_{1}})F_{2}$, they are in the states $|\Phi_{0}^{+}\rangle_{ABC}\otimes|\phi_{1}^{+}\rangle_{ABC}$ and $|\Phi_{1}^{+}\rangle_{ABC}\otimes|\phi_{0}^{+}\rangle_{ABC}$, respectively. The first case $|\Phi_{0}^{+}\rangle_{ABC}\otimes|\phi_{0}^{+}\rangle_{ABC}$ can be described as $\displaystyle|\Phi_{0}^{+}\rangle_{ABC}\otimes|\phi_{0}^{+}\rangle_{ABC}$ (7) $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|H\rangle_{A}|H\rangle_{B}|H\rangle_{C}+|V\rangle_{A}|V\rangle_{B}|V\rangle_{C})$ $\displaystyle\otimes$ $\displaystyle\frac{1}{\sqrt{2}}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{1}\rangle_{C}+|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{2}\rangle_{C})$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|H\rangle_{a_{1}}|H\rangle_{b_{1}}|H\rangle_{c_{1}}+|H\rangle_{a_{2}}|H\rangle_{b_{2}}|H\rangle_{c_{2}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{1}}|V\rangle_{b_{1}}|V\rangle_{c_{1}}+|V\rangle_{a_{2}}|V\rangle_{b_{2}}|V\rangle_{c_{2}})$ $\displaystyle\rightarrow$ $\displaystyle\frac{1}{2}(|V\rangle_{a_{3}}|V\rangle_{b_{3}}|V\rangle_{c_{3}}+|H\rangle_{a_{6}}|H\rangle_{b_{6}}|H\rangle_{c_{6}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{5}}|V\rangle_{b_{5}}|V\rangle_{c_{5}}+|H\rangle_{a_{4}}|H\rangle_{b_{4}}|H\rangle_{c_{4}}).$ From Fig. 1, items $|V\rangle_{a_{3}}|V\rangle_{b_{3}}|V\rangle_{c_{3}}$ and $|H\rangle_{a_{4}}|H\rangle_{b_{4}}|H\rangle_{c_{4}}$ will couple in the BD1, BD2, and BD3, respectively, and finally become the polarization entangled state $|\Phi_{0}^{+}\rangle$ in the output modes D1D2D3. On the other hand, items $|H\rangle_{a_{6}}|H\rangle_{b_{6}}|H\rangle_{c_{6}}$ and $|V\rangle_{a_{5}}|V\rangle_{b_{5}}|V\rangle_{c_{5}}$ will also couple in the BD4, BD5 and BD6 and become the polarization entangled state $|\Phi_{0}^{+}\rangle$ in the output modes D4D5D6. The second case $|\Phi_{1}^{+}\rangle_{ABC}\otimes|\phi_{1}^{+}\rangle_{ABC}$ can evolve as $\displaystyle|\Phi_{1}^{+}\rangle_{ABC}\otimes|\phi_{1}^{+}\rangle_{ABC}$ (8) $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|H\rangle_{A}|H\rangle_{B}|V\rangle_{C}+|V\rangle_{A}|V\rangle_{B}|H\rangle_{C})$ $\displaystyle\otimes$ $\displaystyle\frac{1}{\sqrt{2}}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{2}\rangle_{C}+|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{1}\rangle_{C})$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|H\rangle_{a_{1}}|H\rangle_{b_{1}}|V\rangle_{c_{2}}+|H\rangle_{a_{2}}|H\rangle_{b_{2}}|V\rangle_{c_{1}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{1}}|V\rangle_{b_{1}}|H\rangle_{c_{2}}+|V\rangle_{a_{2}}|V\rangle_{b_{2}}|H\rangle_{c_{1}})$ $\displaystyle\rightarrow$ $\displaystyle\frac{1}{2}(|V\rangle_{a_{3}}|V\rangle_{b_{3}}|H\rangle_{c_{4}}+|H\rangle_{a_{6}}|H\rangle_{b_{6}}|V\rangle_{c_{5}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{5}}|V\rangle_{b_{5}}|H\rangle_{c_{6}}+|H\rangle_{a_{4}}|H\rangle_{b_{4}}|V\rangle_{c_{3}}).$ Items $|V\rangle_{a_{3}}|V\rangle_{b_{3}}|H\rangle_{c_{4}}$ and $|H\rangle_{a_{4}}|H\rangle_{b_{4}}|V\rangle_{c_{3}}$ will become $|\Phi_{1}^{+}\rangle$ in the output modes D1D2D3. Items $|H\rangle_{a_{6}}|H\rangle_{b_{6}}|V\rangle_{c_{5}}$ and $|V\rangle_{a_{5}}|V\rangle_{b_{5}}|H\rangle_{c_{6}}$ will also become $|\Phi_{1}^{+}\rangle$ in the output modes D4D5D6. On the other hand, the cases $|\Phi_{0}^{+}\rangle_{ABC}\otimes|\phi_{1}^{+}\rangle_{ABC}$ and $|\Phi_{1}^{+}\rangle_{ABC}\otimes|\phi_{0}^{+}\rangle_{ABC}$ cannot make the three photons in the output modes D1D2D3 or D4D5D6. For example, $|\Phi_{0}^{+}\rangle_{ABC}\otimes|\phi_{1}^{+}\rangle_{ABC}$ will lead the three photons become $|\Phi_{1}^{+}\rangle$ in output modes D1D2D6 or D4D5D3. Case $|\Phi_{1}^{+}\rangle_{ABC}\otimes|\phi_{0}^{+}\rangle_{ABC}$ will become $|\Phi_{0}^{+}\rangle$ in the output modes D1D2D6 or D4D5D3. In this way, by selecting the output modes D1D2D3 or D4D5D6 each having a photon, they can ultimately obtain a new mixed state $\displaystyle\rho^{\prime}_{ABC}=F^{\prime}|\Phi_{0}^{+}\rangle_{ABC}\langle\Phi_{0}^{+}|+(1-F^{\prime})|\Phi_{1}^{+}\rangle_{ABC}\langle\Phi_{1}^{+}|.$ (9) Here $F^{\prime}$ is $\displaystyle F^{\prime}=\frac{F_{1}F_{2}}{F_{1}F_{2}+(1-F_{1})(1-F_{2})}.$ (10) Obviously, if $F_{1}>\frac{1}{2}$ and $F_{2}>\frac{1}{2}$, we can obtain $F^{\prime}>F_{1}$ and $F^{\prime}>F_{2}$. In this way, we complete the purification of bit-flip error and the success probability is $F_{1}F_{2}+(1-F_{1})(1-F_{2})$. ## III Multipartite entanglement purification for phase-flip error In this section, we will describe the purification of phase-flip error. If a phase-flip error occurs in polarization part and spatial mode part, the mixed state can be written as $\displaystyle\varrho_{ABC}=\varrho^{P}_{ABC}\otimes\varrho^{S}_{ABC}.$ (11) Here $\displaystyle\varrho^{P}_{ABC}=F_{3}|\Phi_{0}^{+}\rangle_{ABC}\langle\Phi_{0}^{+}|+(1-F_{3})|\Phi_{0}^{-}\rangle_{ABC}\langle\Phi_{0}^{-}|,$ (12) and $\displaystyle\varrho^{S}_{ABC}=F_{4}|\phi_{0}^{+}\rangle_{ABC}\langle\phi_{0}^{+}|+(1-F_{4})|\phi_{0}^{-}\rangle_{ABC}\langle\phi_{0}^{-}|.$ (13) Figure 2: Schematic drawing showing the principle of entanglement purification for phase-flip error. The HWP22.5 can transform the $|H\rangle$ polarization to $\frac{1}{\sqrt{2}}(|H\rangle+|V\rangle)$ and $|V\rangle$ polarization to $\frac{1}{\sqrt{2}}(|H\rangle-|V\rangle)$. The BS is the 50:50 beam splitter. $|a_{1}\rangle\rightarrow\frac{1}{\sqrt{2}}(|a_{1}\rangle+|a_{2}\rangle)$ and $|a_{2}\rangle\rightarrow\frac{1}{\sqrt{2}}(|a_{1}\rangle-|a_{2}\rangle)$. Therefore, the HWP22.5 and BS both act as the role of Hadamard operation for polarization and spatial mode qubits, respectively. The principle of phase-flip error is shown in Fig.2. Before purification, they should transform the phase-flip error to bit-flip error using the setup $T_{i},i=1,2,3$. Here HWP22.5 can perform the Hadamard operation and make $|H\rangle\rightarrow\frac{1}{\sqrt{2}}(|H\rangle+|V\rangle)$ and $|V\rangle\rightarrow\frac{1}{\sqrt{2}}(|H\rangle-|V\rangle)$. The beam splitter (BS) can also act as the role of Hadamard operation for spatial mode qubit. It can make $|a_{1}\rangle\rightarrow\frac{1}{\sqrt{2}}(|a_{1}\rangle+|a_{2}\rangle)$ and $|a_{2}\rangle\rightarrow\frac{1}{\sqrt{2}}(|a_{1}\rangle-|a_{2}\rangle)$. After performing the Hadamard operation, the GHZ states in polarization and spatial mode as shown in Eq. (1) and (2) can be rewritten as $\displaystyle|\Psi_{0}^{+}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|H\rangle_{A}|H\rangle_{B}|H\rangle_{C}+|H\rangle_{A}|V\rangle_{B}|V\rangle_{C}$ $\displaystyle+$ $\displaystyle|V\rangle_{A}|H\rangle_{B}|V\rangle_{C}+|V\rangle_{A}|V\rangle_{B}|H\rangle_{C}),$ $\displaystyle|\Psi_{0}^{-}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|H\rangle_{A}|H\rangle_{B}|V\rangle_{C}+|H\rangle_{A}|V\rangle_{B}|H\rangle_{C}$ $\displaystyle+$ $\displaystyle|V\rangle_{A}|H\rangle_{B}|H\rangle_{C}+|V\rangle_{A}|V\rangle_{B}|V\rangle_{C}),$ $\displaystyle|\Psi_{1}^{+}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|H\rangle_{A}|H\rangle_{B}|H\rangle_{C}+|H\rangle_{A}|V\rangle_{B}|V\rangle_{C}$ $\displaystyle-$ $\displaystyle|V\rangle_{A}|H\rangle_{B}|V\rangle_{C}-|V\rangle_{A}|V\rangle_{B}|H\rangle_{C}),$ $\displaystyle|\Psi_{1}^{-}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|H\rangle_{A}|H\rangle_{B}|V\rangle_{C}+|H\rangle_{A}|V\rangle_{B}|H\rangle_{C}$ $\displaystyle-$ $\displaystyle|V\rangle_{A}|H\rangle_{B}|H\rangle_{C}-|V\rangle_{A}|V\rangle_{B}|V\rangle_{C}),$ $\displaystyle|\Psi_{2}^{+}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|H\rangle_{A}|H\rangle_{B}|H\rangle_{C}-|H\rangle_{A}|V\rangle_{B}|V\rangle_{C}$ $\displaystyle+$ $\displaystyle|V\rangle_{A}|H\rangle_{B}|V\rangle_{C}-|V\rangle_{A}|V\rangle_{B}|H\rangle_{C}),$ $\displaystyle|\Psi_{2}^{-}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|H\rangle_{A}|H\rangle_{B}|V\rangle_{C}-|H\rangle_{A}|V\rangle_{B}|H\rangle_{C}$ $\displaystyle+$ $\displaystyle|V\rangle_{A}|H\rangle_{B}|H\rangle_{C}-|V\rangle_{A}|V\rangle_{B}|V\rangle_{C}),$ $\displaystyle|\Psi_{3}^{+}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|H\rangle_{A}|H\rangle_{B}|H\rangle_{C}-|H\rangle_{A}|V\rangle_{B}|V\rangle_{C}$ $\displaystyle-$ $\displaystyle|V\rangle_{A}|H\rangle_{B}|V\rangle_{C}+|V\rangle_{A}|V\rangle_{B}|H\rangle_{C}),$ $\displaystyle|\Psi_{3}^{-}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|H\rangle_{A}|H\rangle_{B}|V\rangle_{C}-|H\rangle_{A}|V\rangle_{B}|H\rangle_{C}$ $\displaystyle-$ $\displaystyle|V\rangle_{A}|H\rangle_{B}|H\rangle_{C}+|V\rangle_{A}|V\rangle_{B}|V\rangle_{C}),$ $\displaystyle|\psi_{0}^{+}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{1}\rangle_{C}+|a_{1}\rangle_{A}|b_{2}\rangle_{B}|c_{2}\rangle_{C}$ $\displaystyle+$ $\displaystyle|a_{2}\rangle_{A}|b_{1}\rangle_{B}|c_{2}\rangle_{C}+|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{1}\rangle_{C}),$ $\displaystyle|\psi_{0}^{-}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{2}\rangle_{C}+|a_{1}\rangle_{A}|b_{2}\rangle_{B}|c_{1}\rangle_{C}$ $\displaystyle+$ $\displaystyle|a_{2}\rangle_{A}|b_{1}\rangle_{B}|c_{1}\rangle_{C}+|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{2}\rangle_{C}),$ $\displaystyle|\psi_{1}^{+}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{1}\rangle_{C}+|a_{1}\rangle_{A}|b_{2}\rangle_{B}|c_{2}\rangle_{C}$ $\displaystyle-$ $\displaystyle|a_{2}\rangle_{A}|b_{1}\rangle_{B}|c_{2}\rangle_{C}-|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{1}\rangle_{C}),$ $\displaystyle|\psi_{1}^{-}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{2}\rangle_{C}+|a_{1}\rangle_{A}|b_{2}\rangle_{B}|c_{1}\rangle_{C}$ $\displaystyle-$ $\displaystyle|a_{2}\rangle_{A}|b_{1}\rangle_{B}|c_{1}\rangle_{C}-|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{2}\rangle_{C}),$ $\displaystyle|\psi_{2}^{+}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{1}\rangle_{C}-|a_{1}\rangle_{A}|b_{2}\rangle_{B}|c_{2}\rangle_{C}$ $\displaystyle+$ $\displaystyle|a_{2}\rangle_{A}|b_{1}\rangle_{B}|c_{2}\rangle_{C}-|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{1}\rangle_{C}),$ $\displaystyle|\psi_{2}^{-}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{2}\rangle_{C}-|a_{1}\rangle_{A}|b_{2}\rangle_{B}|c_{1}\rangle_{C}$ $\displaystyle+$ $\displaystyle|a_{2}\rangle_{A}|b_{1}\rangle_{B}|c_{1}\rangle_{C}-|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{2}\rangle_{C}),$ $\displaystyle|\psi_{3}^{+}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{1}\rangle_{C}-|a_{1}\rangle_{A}|b_{2}\rangle_{B}|c_{2}\rangle_{C}$ $\displaystyle-$ $\displaystyle|a_{2}\rangle_{A}|b_{1}\rangle_{B}|c_{2}\rangle_{C}+|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{1}\rangle_{C}),$ $\displaystyle|\psi_{3}^{-}\rangle_{ABC}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{2}\rangle_{C}-|a_{1}\rangle_{A}|b_{2}\rangle_{B}|c_{1}\rangle_{C}$ $\displaystyle-$ $\displaystyle|a_{2}\rangle_{A}|b_{1}\rangle_{B}|c_{1}\rangle_{C}+|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{2}\rangle_{C}).$ After performing the Hadamard operations, the mixed state in Eq.(11) can be rewritten as $\displaystyle\sigma_{ABC}=\sigma^{P}_{ABC}\otimes\sigma^{S}_{ABC}.$ (16) Here $\displaystyle\sigma^{P}_{ABC}=F_{3}|\Psi_{0}^{+}\rangle_{ABC}\langle\Psi_{0}^{+}|+(1-F_{3})|\Psi_{0}^{-}\rangle_{ABC}\langle\Psi_{0}^{-}|,$ (17) and $\displaystyle\sigma^{S}_{ABC}=F_{4}|\psi_{0}^{+}\rangle_{ABC}\langle\psi_{0}^{+}|+(1-F_{4})|\psi_{0}^{-}\rangle_{ABC}\langle\psi_{0}^{-}|.$ (18) Therefore, $\sigma_{ABC}$ can be described as follows. With the probability of $F_{3}F_{4}$, it is in the state $|\Psi_{0}^{+}\rangle_{ABC}\otimes|\psi_{0}^{+}\rangle_{ABC}$. With the probability of $F_{3}(1-F_{4})$, it is in the state $|\Psi_{0}^{+}\rangle_{ABC}\otimes|\psi_{0}^{-}\rangle_{ABC}$. With the probability of $(1-F_{3})F_{4}$, it is in the state $|\Psi_{0}^{-}\rangle_{ABC}\otimes|\psi_{0}^{+}\rangle_{ABC}$. With the probability of $(1-F_{3})(1-F_{4})$, it is in the state $|\Psi_{0}^{-}\rangle_{ABC}\otimes|\psi_{0}^{-}\rangle_{ABC}$. We first discuss the case $|\Psi_{0}^{+}\rangle_{ABC}\otimes|\psi_{0}^{+}\rangle_{ABC}$. It will evolve as $\displaystyle|\Psi_{0}^{+}\rangle_{ABC}\otimes|\psi_{0}^{+}\rangle_{ABC}$ (19) $\displaystyle=$ $\displaystyle\frac{1}{2}(|H\rangle_{A}|H\rangle_{B}|H\rangle_{C}+|H\rangle_{A}|V\rangle_{B}|V\rangle_{C}$ $\displaystyle+$ $\displaystyle|V\rangle_{A}|H\rangle_{B}|V\rangle_{C}+|V\rangle_{A}|V\rangle_{B}|H\rangle_{C})$ $\displaystyle\otimes$ $\displaystyle\frac{1}{2}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{1}\rangle_{C}+|a_{1}\rangle_{A}|b_{2}\rangle_{B}|c_{2}\rangle_{C}$ $\displaystyle+$ $\displaystyle|a_{2}\rangle_{A}|b_{1}\rangle_{B}|c_{2}\rangle_{C}+|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{1}\rangle_{C})$ $\displaystyle=$ $\displaystyle\frac{1}{4}(|H\rangle_{a_{1}}|H\rangle_{b_{1}}|H\rangle_{c_{1}}+|H\rangle_{a_{1}}|H\rangle_{b_{2}}|H\rangle_{c_{2}}$ $\displaystyle+$ $\displaystyle|H\rangle_{a_{2}}|H\rangle_{b_{1}}|H\rangle_{c_{2}}+|H\rangle_{a_{2}}|H\rangle_{b_{2}}|H\rangle_{c_{1}}$ $\displaystyle+$ $\displaystyle|H\rangle_{a_{1}}|V\rangle_{b_{1}}|V\rangle_{c_{1}}+|H\rangle_{a_{1}}|V\rangle_{b_{2}}|V\rangle_{c_{2}}$ $\displaystyle+$ $\displaystyle|H\rangle_{a_{2}}|V\rangle_{b_{1}}|V\rangle_{c_{2}}+|H\rangle_{a_{2}}|V\rangle_{b_{2}}|V\rangle_{c_{1}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{1}}|H\rangle_{b_{1}}|V\rangle_{c_{1}}+|V\rangle_{a_{1}}|H\rangle_{b_{2}}|V\rangle_{c_{2}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{2}}|H\rangle_{b_{1}}|V\rangle_{c_{2}}+|V\rangle_{a_{2}}|H\rangle_{b_{2}}|V\rangle_{c_{1}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{1}}|V\rangle_{b_{1}}|H\rangle_{c_{1}}+|V\rangle_{a_{1}}|V\rangle_{b_{2}}|H\rangle_{c_{2}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{2}}|V\rangle_{b_{1}}|H\rangle_{c_{2}}+|V\rangle_{a_{2}}|V\rangle_{b_{2}}|H\rangle_{c_{1}})$ $\displaystyle\rightarrow$ $\displaystyle\frac{1}{4}(|V\rangle_{a_{3}}|V\rangle_{b_{3}}|V\rangle_{c_{3}}+|V\rangle_{a_{3}}|H\rangle_{b_{6}}|H\rangle_{c_{6}}$ $\displaystyle+$ $\displaystyle|H\rangle_{a_{6}}|V\rangle_{b_{3}}|H\rangle_{c_{6}}+|H\rangle_{a_{6}}|H\rangle_{b_{6}}|V\rangle_{c_{3}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{3}}|V\rangle_{b_{5}}|V\rangle_{c_{5}}+|V\rangle_{a_{3}}|H\rangle_{b_{4}}|H\rangle_{c_{4}}$ $\displaystyle+$ $\displaystyle|H\rangle_{a_{6}}|V\rangle_{b_{5}}|H\rangle_{c_{4}}+|H\rangle_{a_{6}}|H\rangle_{b_{4}}|V\rangle_{c_{5}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{5}}|V\rangle_{b_{3}}|V\rangle_{c_{5}}+|V\rangle_{a_{5}}|H\rangle_{b_{6}}|H\rangle_{c_{4}}$ $\displaystyle+$ $\displaystyle|H\rangle_{a_{4}}|V\rangle_{b_{3}}|H\rangle_{c_{4}}+|H\rangle_{a_{4}}|H\rangle_{b_{6}}|V\rangle_{c_{5}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{5}}|V\rangle_{b_{5}}|V\rangle_{c_{3}}+|V\rangle_{a_{5}}|H\rangle_{b_{4}}|H\rangle_{c_{6}}$ $\displaystyle+$ $\displaystyle|H\rangle_{a_{4}}|V\rangle_{b_{5}}|H\rangle_{c_{6}}+|H\rangle_{a_{4}}|H\rangle_{b_{4}}|V\rangle_{c_{3}}).$ From Eq. (19), item $|V\rangle_{a_{3}}|V\rangle_{b_{3}}|V\rangle_{c_{3}}$, $|V\rangle_{a_{3}}|H\rangle_{b_{4}}|H\rangle_{c_{4}}$, $|H\rangle_{a_{4}}|V\rangle_{b_{3}}|H\rangle_{c_{4}}$ and $|H\rangle_{a_{4}}|H\rangle_{b_{4}}|V\rangle_{c_{3}}$ will be in the output modes D1D2D3 and become $\displaystyle\frac{1}{2}(|V\rangle_{D_{1}}|V\rangle_{D_{2}}|V\rangle_{D_{3}}+|V\rangle_{D_{1}}|H\rangle_{D_{2}}|H\rangle_{D_{3}}$ (20) $\displaystyle+$ $\displaystyle|H\rangle_{D_{1}}|V\rangle_{D_{2}}|H\rangle_{D_{3}}+|H\rangle_{D_{1}}|H\rangle_{D_{2}}|V\rangle_{D_{3}}).$ By performing bit-flip operation on each photon, state in Eq.(20) can be changed to $|\Psi_{0}^{+}\rangle_{ABC}$. Finally, they can change $|\Psi_{0}^{+}\rangle_{ABC}$ to $|\Phi_{0}^{+}\rangle_{ABC}$ by adding another Hadamard operation on each photon. On the other hand, from Eq.(19), by selecting the output modes D1D5D6, D4D5D3, or D4D2D6, they can obtain the same state in Eq.(20). In this way, they can also obtain $|\Phi_{0}^{+}\rangle_{ABC}$. The case $|\Psi_{0}^{-}\rangle_{ABC}\otimes|\psi_{0}^{-}\rangle_{ABC}$ can be evolve as $\displaystyle|\Psi_{0}^{-}\rangle_{ABC}\otimes|\psi_{0}^{-}\rangle_{ABC}$ (21) $\displaystyle=$ $\displaystyle\frac{1}{2}(|H\rangle_{A}|H\rangle_{B}|V\rangle_{C}+|H\rangle_{A}|V\rangle_{B}|H\rangle_{C}$ $\displaystyle+$ $\displaystyle|V\rangle_{A}|H\rangle_{B}|H\rangle_{C}+|V\rangle_{A}|V\rangle_{B}|V\rangle_{C})$ $\displaystyle\otimes\frac{1}{2}(|a_{1}\rangle_{A}|b_{1}\rangle_{B}|c_{2}\rangle_{C}+|a_{1}\rangle_{A}|b_{2}\rangle_{B}|c_{1}\rangle_{C}$ $\displaystyle+$ $\displaystyle|a_{2}\rangle_{A}|b_{1}\rangle_{B}|c_{1}\rangle_{C}+|a_{2}\rangle_{A}|b_{2}\rangle_{B}|c_{2}\rangle_{C})$ $\displaystyle\rightarrow\frac{1}{4}(|V\rangle_{a_{3}}|V\rangle_{b_{3}}|H\rangle_{c_{4}}+|V\rangle_{a_{3}}|H\rangle_{b_{6}}|V\rangle_{c_{5}}$ $\displaystyle+$ $\displaystyle|H\rangle_{a_{6}}|V\rangle_{b_{3}}|V\rangle_{c_{5}}+|H\rangle_{a_{6}}|H\rangle_{b_{6}}|H\rangle_{c_{4}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{3}}|V\rangle_{b_{5}}|H\rangle_{c_{6}}+|V\rangle_{a_{3}}|H\rangle_{b_{4}}|V\rangle_{c_{3}}$ $\displaystyle+$ $\displaystyle|H\rangle_{a_{6}}|V\rangle_{b_{5}}|V\rangle_{c_{3}}+|H\rangle_{a_{6}}|H\rangle_{b_{6}}|H\rangle_{c_{6}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{5}}|V\rangle_{b_{3}}|H\rangle_{c_{6}}+|V\rangle_{a_{5}}|H\rangle_{b_{6}}|V\rangle_{c_{3}}$ $\displaystyle+$ $\displaystyle|H\rangle_{a_{4}}|V\rangle_{b_{3}}|V\rangle_{c_{3}}+|H\rangle_{a_{4}}|H\rangle_{b_{6}}|H\rangle_{c_{6}}$ $\displaystyle+$ $\displaystyle|V\rangle_{a_{5}}|V\rangle_{b_{5}}|H\rangle_{c_{4}}+|V\rangle_{a_{5}}|H\rangle_{b_{4}}|H\rangle_{c_{5}}$ $\displaystyle+$ $\displaystyle|H\rangle_{a_{4}}|V\rangle_{b_{5}}|V\rangle_{c_{5}}+|H\rangle_{a_{4}}|H\rangle_{b_{4}}|H\rangle_{c_{4}}).$ From Eq.(21), items $|V\rangle_{a_{3}}|V\rangle_{b_{3}}|H\rangle_{c_{4}}$, $|V\rangle_{a_{3}}|H\rangle_{b_{4}}|V\rangle_{c_{3}}$, $|H\rangle_{a_{4}}|V\rangle_{b_{3}}|V\rangle_{c_{3}}$ and $|H\rangle_{a_{4}}|H\rangle_{b_{4}}|H\rangle_{c_{4}}$ will be in the output modes D1D2D3 and become $\displaystyle\frac{1}{2}(|V\rangle_{D_{1}}|V\rangle_{D_{2}}|H\rangle_{D_{3}}+|V\rangle_{D_{1}}|H\rangle_{D_{2}}|V\rangle_{D_{3}}$ (22) $\displaystyle+$ $\displaystyle|H\rangle_{D_{1}}|V\rangle_{D_{2}}|V\rangle_{D_{3}}+|H\rangle_{D_{1}}|H\rangle_{D_{2}}|H\rangle_{D_{3}}).$ State in Eq.(22) can be changed to $|\Psi_{0}^{-}\rangle_{ABC}$ by adding bit- flip operation on each photon. By adding another Hadamard operation on each photon, $|\Psi_{0}^{-}\rangle_{ABC}$ can be converted to $|\Phi_{0}^{-}\rangle_{ABC}$. On the other hand, from Eq.(21), by selecting the output modes D1D5D6, D4D5D3, or D4D2D6, they can obtain the same state in Eq.(22). In this way, they can also obtain $|\Phi_{0}^{-}\rangle_{ABC}$. The other cases $|\Psi_{0}^{+}\rangle_{ABC}\otimes|\psi_{0}^{-}\rangle_{ABC}$ and $|\Psi_{0}^{-}\rangle_{ABC}\otimes|\psi_{0}^{-}\rangle_{ABC}$ will lead the photons in the output modes D1D2D6, D1D5D3, D4D2D3, or D4D5D6. Therefore, by selecting the cases that the output modes D1D2D3, D1D5D6, D4D5D3, or D4D2D6 contain one photon, cases $|\Psi_{0}^{+}\rangle_{ABC}\otimes|\psi_{0}^{-}\rangle_{ABC}$ and $|\Psi_{0}^{-}\rangle_{ABC}\otimes|\psi_{0}^{-}\rangle_{ABC}$ can be eliminated automatically. Finally, with the probability of $F_{3}F_{4}$, they will obtain $|\Phi_{0}^{+}\rangle_{ABC}$. With the probability of $(1-F_{3})(1-F_{4})$, they will obtain $|\Phi_{0}^{-}\rangle_{ABC}$. The new mixed state can be rewritten as $\displaystyle\sigma^{\prime}_{ABC}=F^{\prime\prime}|\Psi_{0}^{+}\rangle_{ABC}\langle\Psi_{0}^{+}|+(1-F^{\prime\prime})|\Psi_{0}^{-}\rangle_{ABC}\langle\Psi_{0}^{-}|.$ Here $F^{\prime\prime}$ is $\displaystyle F^{\prime\prime}=\frac{F_{3}F_{4}}{F_{3}F_{4}+(1-F_{3})(1-F_{4})}.$ (24) Similar to Eq.(10), $F^{\prime\prime}>F_{3}$ and $F^{\prime\prime}>F_{4}$ if $F_{3}>\frac{1}{2}$ and $F_{4}>\frac{1}{2}$. ## IV Arbitrary multipartite entanglement purification It is easy to extend the EPP to the arbitrary GHZ state. The $m$-photon GHZ state in polarization can be described as $\displaystyle|\Phi_{0}^{\pm}\rangle_{m}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|H\rangle_{1}|H\rangle_{2}\cdots|H\rangle_{m}\pm|V\rangle_{1}|V\rangle_{2}\cdots|V\rangle_{m}),$ $\displaystyle|\Phi_{1}^{\pm}\rangle_{m}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|H\rangle_{1}|H\rangle_{2}\cdots|V\rangle_{m}\pm|V\rangle_{1}|V\rangle_{2}\cdots|H\rangle_{m}),$ $\displaystyle\cdots$ , $\displaystyle|\Phi_{2^{m-1}}^{\pm}\rangle_{m}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|V\rangle_{1}|H\rangle_{2}\cdots|H\rangle_{m}\pm|H\rangle_{1}|V\rangle_{2}\cdots|V\rangle_{m}),$ Figure 3: Schematic drawing showing the principle of entanglement purification for arbitrary GHZ state. On the other hand, the $m$-photon GHZ state in spatial mode can be described as $\displaystyle|\phi_{0}^{\pm}\rangle_{m}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|a_{1}\rangle_{1}|b_{1}\rangle_{2}$ $\displaystyle\cdots$ $\displaystyle|m_{1}\rangle_{m}\pm|a_{2}\rangle_{1}|b_{2}\rangle_{2}\cdots|m_{2}\rangle_{m}),$ $\displaystyle|\phi_{1}^{\pm}\rangle_{m}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|a_{1}\rangle_{1}|b_{1}\rangle_{2}$ $\displaystyle\cdots$ $\displaystyle|m_{2}\rangle_{m}\pm|a_{2}\rangle_{1}|b_{2}\rangle_{2}\cdots|m_{1}\rangle_{m}),$ $\displaystyle\cdots$ , $\displaystyle|\phi_{2^{m-1}}^{\pm}\rangle_{m}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|a_{2}\rangle_{1}|b_{1}\rangle_{2}$ (26) $\displaystyle\cdots$ $\displaystyle|m_{1}\rangle_{m}\pm|a_{1}\rangle_{1}|b_{1}\rangle_{2}\cdots|m_{2}\rangle_{m}).$ As shown in Fig. 3, the entanglement source prepares the $m$-photon hyperentangled state $|\Psi\rangle_{m}$ of the form $\displaystyle|\Psi\rangle_{m}=|\Phi_{0}^{+}\rangle_{m}\otimes|\phi_{0}^{+}\rangle_{m}.$ (27) Such hyperentangled state is distributed to $m$ parties, named Bob1, Bob2, $\cdots$, and Bobm. After distribution, the initial state becomes a mixed state as $\displaystyle\rho_{m}=\rho^{P}_{m}\otimes\rho^{P}_{m}.$ (28) Here $\rho^{P}_{m}$ can be written as $\displaystyle\rho^{P}_{m}=F_{1}|\Phi_{0}^{+}\rangle_{m}\langle\Phi_{0}^{+}|+(1-F_{{}_{1}})|\Phi_{1}^{+}\rangle_{m}\langle\Phi_{1}^{+}|.$ (29) $\rho^{P}_{m}$ can be written as $\displaystyle\rho^{S}_{m}=F_{2}|\phi_{0}^{+}\rangle_{m}\langle\phi_{0}^{+}|+(1-F_{{}_{2}})|\phi_{1}^{+}\rangle_{m}\langle\phi_{1}^{+}|.$ (30) The purification is similar as described in above section. By selecting the output modes D1,D2, $\cdots$, Dm exactly contain one photon, they can ultimately obtain a high fidelity mixed state in polarization. The fidelity $F^{\prime}$ is same as it is shown in Eq. (10). On the other hand, if the phase-flip error occurs, one can also convert it to the bit-flip error, and perform the purification in a next step. In this way, one can purify the arbitrary $m$-photon GHZ state. ## V Discussion So far, we have completely described this EPP. We first described the EPP for three-photon GHZ state with a bit-flip error. Then we explained the EPP with a phase-flip error. In this way, all the errors can be purified. Finally, we extend this EPP for arbitrary GHZ state and which can be purified in the same way. In above, we suppose that the bit-flip error occurs on the first qubit. In a practical transmission, the hyperentanglement in polarization and spatial mode will suffer from different errors. For example, the bit-flip error occurs on the first qubit in polarization and which makes the polarization part become the mixed state in Eq. (5), while the bit-flip error occurs on the second qubit in spatial mode and makes the spatial part become $\displaystyle\rho^{\prime S}_{ABC}=F_{3}|\phi_{0}^{+}\rangle_{ABC}\langle\phi_{0}^{+}|+(1-F_{{}_{3}})|\phi_{2}^{+}\rangle_{ABC}\langle\phi_{2}^{+}|.$ (31) Therefore, the hyperentangled mixed state can be written as $\displaystyle\rho^{\prime}_{ABC}=\rho^{P}_{ABC}\otimes\rho^{\prime S}_{ABC}.$ (32) With the probability of F1F3, it is in the state $|\Phi_{0}^{+}\rangle_{ABC}\otimes|\phi_{0}^{+}\rangle_{ABC}$. Such state will make the three photons in the output modes D1D2D3 or D4D5D6. With the probability of (1-F${}_{1})$F3, it is in the state $|\Phi_{1}^{+}\rangle_{ABC}\otimes|\phi_{0}^{+}\rangle_{ABC}$. Such state will make the three photons in the output modes D1D2D6 or D4D5D3. With the probability of F1(1-F${}_{3})$, it is in the state $|\Phi_{0}^{+}\rangle_{ABC}\otimes|\phi_{2}^{+}\rangle_{ABC}$. Such state will make the three photons in the output modes D1D5D3 or D4D2D6. Finally, with the probability of (1-F1)(1-F3), it is in the state $|\Phi_{1}^{+}\rangle_{ABC}\otimes|\phi_{2}^{+}\rangle_{ABC}$. Such state will make the three photons in the output modes D1D5D6 or D4D2D3. In this way, by selecting the cases D1D2D3 or D4D5D6, they can ultimately obtain the polarization state $|\Phi_{0}^{+}\rangle_{ABC}$. Interestingly, if the mixed state is described as shown in Eq.(32), the bit-flip error can be completely purified. The second case $|\Phi_{1}^{+}\rangle_{ABC}\otimes|\phi_{0}^{+}\rangle_{ABC}$ will lead the photons in D1D2D6 or D4D5D3 and such state will become $|\Phi_{1}^{+}\rangle_{ABC}$. They can add a bit-flip operation on the first photon and convert it to $|\Phi_{0}^{+}\rangle_{ABC}$ deterministically. On the other hand, the second case $|\Phi_{0}^{+}\rangle_{ABC}\otimes|\phi_{2}^{+}\rangle_{ABC}$ will lead the photons in the spatial modes D1D5D3 or D4D2D6 and become $|\Phi_{2}^{+}\rangle_{ABC}$. They can also add the bit-flip operation on the second photon and convert it to $|\Phi_{0}^{+}\rangle_{ABC}$ deterministically. The final case $|\Phi_{1}^{+}\rangle_{ABC}\otimes|\phi_{2}^{+}\rangle_{ABC}$ will also become $|\Phi_{1}^{+}\rangle_{ABC}$ and can be converted to $|\Phi_{0}^{+}\rangle_{ABC}$ deterministically. In this way, they can obtain the maximally pure entangled state $|\Phi_{0}^{+}\rangle_{ABC}$ with the probability of 100$\%$. For a general mixed state with bit-flip error, the polarization part and spatial-mode part can be written as $\displaystyle\rho^{\prime\prime P}_{ABC}$ $\displaystyle=$ $\displaystyle F_{1}|\Phi_{0}^{+}\rangle_{ABC}\langle\Phi_{0}^{+}|+F_{2}|\Phi_{1}^{+}\rangle_{ABC}\langle\Phi_{1}^{+}|$ (33) $\displaystyle+$ $\displaystyle F_{3}|\Phi_{2}^{+}\rangle_{ABC}\langle\Phi_{2}^{+}|+F_{4}|\Phi_{3}^{+}\rangle_{ABC}\langle\Phi_{3}^{+}|,$ and $\displaystyle\rho^{\prime\prime S}_{ABC}$ $\displaystyle=$ $\displaystyle F_{4}|\phi_{0}^{+}\rangle_{ABC}\langle\phi_{0}^{+}|+F_{5}|\phi_{1}^{+}\rangle_{ABC}\langle\phi_{1}^{+}|$ (34) $\displaystyle+$ $\displaystyle F_{6}|\phi_{2}^{+}\rangle_{ABC}\langle\phi_{2}^{+}|+F_{7}|\phi_{3}^{+}\rangle_{ABC}\langle\phi_{3}^{+}|.$ Here $F_{1}+F_{2}+F_{3}+F_{4}=1$ and $F_{5}+F_{6}+F_{7}+F_{8}=1$. Similarly, by selecting the output modes D1D2D3 or D4D5D6, they can obtain a new mixed state as $\displaystyle\rho_{ABC}^{\prime\prime}$ $\displaystyle=$ $\displaystyle F^{\prime}_{1}|\Phi_{0}^{+}\rangle_{ABC}\langle\Phi_{0}^{+}|+F^{\prime}_{2}|\Phi_{1}^{+}\rangle_{ABC}\langle\Phi_{1}^{+}|$ (35) $\displaystyle+$ $\displaystyle F^{\prime}_{3}|\Phi_{2}^{+}\rangle_{ABC}\langle\Phi_{2}^{+}|+F^{\prime}_{4}|\Phi_{3}^{+}\rangle_{ABC}\langle\Phi_{3}^{+}|.$ Here $\displaystyle F^{\prime}_{1}=\frac{F_{1}F_{5}}{F_{1}F_{5}+F_{2}F_{6}+F_{3}F_{7}+F_{4}F_{8}},$ $\displaystyle F^{\prime}_{2}=\frac{F_{2}F_{6}}{F_{1}F_{5}+F_{2}F_{6}+F_{3}F_{7}+F_{4}F_{8}},$ $\displaystyle F^{\prime}_{3}=\frac{F_{3}F_{7}}{F_{1}F_{5}+F_{2}F_{6}+F_{3}F_{7}+F_{4}F_{8}},$ $\displaystyle F^{\prime}_{4}=\frac{F_{4}F_{8}}{F_{1}F_{5}+F_{2}F_{6}+F_{3}F_{7}+F_{4}F_{8}}.$ (36) If $F_{1}>\frac{1}{2}$ and $F_{5}>\frac{1}{2}$, we can also obtain $F^{\prime}_{1}>F_{1}$ and $F^{\prime}_{1}>F_{5}$. In this way, we can realize the general purification. It is interesting to calculate the purification efficiency in a practical environment. As shown in Fig. 1, the N-photon GHZ state was distributed to $N$ parties. The transmission efficiency is $\eta_{t}=e^{-\frac{L}{L_{0}}}$. The detector efficiency is $\eta_{d}$. The $\eta_{c}$ is the probability of coupling a photon to the single-photon detector. $L_{0}$ is the attenuation length of the channel (25 km for commercial fibre) munro . $L$ is transmission distance. The success probability is $p_{1}=F_{1}F_{2}+(1-F_{1})(1-F_{2})$. For N-photon purification, the total purification efficiency can be calculated as $\displaystyle P^{N}_{one}=p_{1}\eta^{N}_{t}\eta^{N}_{d}\eta^{N}_{c}.$ (37) In existing multipartite EPPs multipurification1 ; multipurification3 ; multipurification4 ; multipurification7 , they exploit two pairs of N-photon GHZ states to perform the purification. Therefore, for linear optical system, the total purification efficiency can be calculated as $\displaystyle P^{N}_{two}=\frac{1}{4}p_{1}\eta^{2N}_{t}\eta^{2N}_{d}\eta^{2N}_{c}.$ (38) The ratio of $P^{N}_{one}$ and $P^{N}_{two}$ can be calculated as $\displaystyle R=\frac{P^{N}_{one}}{P^{N}_{two}}=\frac{4}{\eta^{N}_{t}\eta^{N}_{d}\eta^{N}_{c}}=\frac{4}{(e^{-\frac{L}{L_{0}}})^{N}\eta^{N}_{d}\eta^{N}_{c}}.$ (39) If we let $\eta_{d}=0.9$, $\eta_{c}=0.95$ munro . Figure 4: The ratio $R$ of entanglement purification efficiency plotted against length of entanglement distribution. We let the photon number of GHZ state as $N=3$ and $N=6$, respectively. Figure 5: The ratio $R$ of entanglement purification efficiency plotted against photon number of GHZ state. We let $L=L_{0}=25km$. Fig. 4 shows the relationship between the coefficient $L$ and $R$. Here we let $N=3$ and $N=6$ respectively. We change the distance $L$ from 20km to 100km. The ratio $R$ increases rapidly. The $R$ can reach more than $10^{10}$ when $N=6$ and $L=100$km. In Fig. 5, we also calculated the $R$ altered with $N$. We let $L=L_{0}=25$km. We also showed that the $R$ increases rapidly with the photon number $N$. On the other hand, in existing multipartite EPPs multipurification1 ; multipurification3 ; multipurification4 ; multipurification7 , after each party performing the CNOT or similar operation, they should measure the target particles to judge that the purification is successful or not. In this EPP, the parties are not required to measure the particles and they can judge whether the purification is successful or not according to the output modes of the photons. In this way, this EPP is more economical and practical in future application. Finally, let us briefly discuss the possible realization. This protocol mainly exploits the common linear optics, such as PBS, BS, BD, HWP. Meanwhile, this protocol require the multi-partite hyperentanglement. Such hyperentanglement was also realized in experiment source , which show that this protocol is feasible in current experiment condition. ## VI Conclusion In conclusion, we have proposed the EPP for multipartite entanglement purification using hyperentanglement. After performing the EPP, the spatial entanglement can be used to purify the polarization entanglement. Different from the previous works, this EPP has several advantages. Firstly, with the same purification success probability, this EPP only requires one pair of multipartite GHZ states, while existing EPPs usually require two pairs of multipartite GHZ state. Secondly, if consider the practical transmission and detector efficiency, this EPP may be extremely useful for the ratio of purification efficiency increases rapidly with both the number of photons and the transmission distance. Thirdly, this protocol requires linear optics and does not add additional measurement operations, so that it is feasible for experiment. All these advantages will make this protocol have potential application for future quantum information processing. ## ACKNOWLEDGEMENTS This work was supported by the National Natural Science Foundation of China (No. 11974189). ## References * (1) C. H. Bennett, et al. Phys. Rev. Lett. 83, 3081(1993). * (2) A. Ekert, Phys. Rev. Lett. 67, 661 (1991). * (3) C. H. Bennett, and S. J.Wiesner, Phys. Rev. Lett. 69, 2881 (1992). * (4) G. L. Long, and X.S. Liu, Phys. Rev. A 65, 032302 (2002). * (5) F. G. Deng, G. L. Long, and X.S. Liu, Phys. Rev. A 68, 042317 (2003). * (6) W. Zhang, D.S. Ding, Y. B. Sheng, et al. Phys. Rev. Lett. 118, 220501 (2017). * (7) D. Gottesman, and I. L. Chuang, Nature 402, 390 (1999). * (8) Y. B. Sheng, and L. Zhou, Sci. Bull. 62, 1025 (2017). * (9) A. Karlsson, and M. Bourennane, Phys. Rev. A 58, 4394 (1998). * (10) F. G. Deng, C. Y. Li, Y. S. Li, H. Y. Zhou, and Y. Wang, Phys. Rev. A 72, 022338 (2005) * (11) M. Hillery, V. Buz̆ek, and A. Berthiaume, Phys. Rev. A 59, 1829 (1999). * (12) A. Karlsson, M. Koashi, and N. Imoto, Phys. Rev. A 59, 162 (1999). * (13) L. Xiao, G. L. Long, F. G. Deng, and J. W. Pan, Phys. Rev. A 69, 052307 (2004). * (14) R. Cleve, D. Gottesman, and H. K. Lo, Phys. Rev. Lett. 83, 648 (1999). * (15) A. M. Lance, T. Symul, W. P. Bowen, B. C. Sanders, and P. K. Lam, Phys. Rev. Lett. 92, 177903 (2004). * (16) F. G. Deng, X. H. Li, C. Y. Li, P. Zhou, and H. Y. Zhou, Phys. Rev. A 72, 044301 (2005) * (17) D. J. Wineland, J. J. Bollinger, W. M. Itano, F. L. Moore, and D. J. Heinzen, Phys. Rev. A 46, R6797 (1992). * (18) V. Giovannetti, S. Lloyd, and L. Maccone, Phys. Rev. Lett. 96, 010401 (2006). * (19) Y. Fu, H. L. Yin, T. Y. Chen, and Z. B. Chen, Phys. Rev. Lett. 114, 090501 (2015). * (20) T. Pramanik, D. H. Lee, Y. W. Cho, H. T. Lim, S. W. Han, H. Jung, S. Moon, K. J. Lee, and Y. S. Kim, Phys. Rev. Appl. 14, 064074 (2020). * (21) S. M. Lee, S. W. Lee, H. Jeong, and H. S. Park, Phys. Rev. Lett. 124, 060501 (2020). * (22) C. Song, K. Xu, W. Liu, C.-p. Yang, S.-B. Zheng, H. Deng, Q. Xie, K. Huang, Q. Guo, L. Zhang et al., Phys. Rev. Lett. 119, 180511 (2017). * (23) M. Gong, et al., Phys. Rev. Lett. 122, 110501 (2019) * (24) T. Monz, P. Schindler, J. T. Barreiro, M. Chwalla, D. Nigg, W. A. Coish, M. Harlander, W. Häsel, M. Hennrich, and R. Blatt, Phys. Rev. Lett. 106, 130506 (2011). * (25) X.-L.Wang, Y.-H. Luo, H.-L. Huang, M.-C. Chen, Z.-E. Su, C. Liu, C. Chen, W. Li, Y.-Q. Fang, X. Jiang et al., Phys. Rev. Lett. 120, 260502 (2018). * (26) Y. Zhou, Q. Zhao, X. Yuan, and X. F. Ma, npj Quant. Inf. 5, 83 (2019). * (27) Briegel, H.-J., Duer, W., Cirac, J. I. & Zoller, P. Phys. Rev. Lett. 81, 5932-5935 (1998). * (28) C. H. Bennett, G. Brassard, S. Popescu, B. Schumacher, J.A. Smolin, W.K. Wootters, Phys. Rev. Lett. 76, 722 (1996). * (29) D. Deutsch, A. Ekert, R. Jozsa, C. Macchiavello, S. Popescu, and A. Sanpera, Phys. Rev. Lett. 77, 2818 (1996). * (30) J. W. Pan, C. Simon, A. Zeilinger, Nature 410, 1067(2001). * (31) J. W. Pan, S. Gasparonl, R. Ursin, G. Weihs, A. Zeilinger, Nature 423, 417 (2003). * (32) C. Simon, J. W. Pan, Phys. Rev. Lett. 89, 257901(2002) . * (33) X. L. Feng, L. C. Kwek, C.H. Oh, Phys. Rev. A 71 064301 (2005). * (34) M. Yang, W. Song, Z. L. Cao, Phys. Rev. A 71, 012308 (2005). * (35) Reichl, R. et al. Nature 443, 838-841 (2006). * (36) Duer, W., and Briegel, H. J. Rep. Pro. Phys. 70, 1381-1424 (2007). * (37) Y.B. Sheng, F. G. Deng, H.Y. Zhou, Phys. Rev. A 77, 042308 (2008). * (38) Y.B. Sheng, F. G. Deng, Phys. Rev. A 81, 032307(2010). * (39) Y.B. Sheng, F. G. Deng, Phys. Rev. A 82, 044305(2010). * (40) C. Wang, Y. Zhang, G. S. Jin, Phys. Rev. A 84, 032307 (2011). * (41) C. Wang, Y. Zhang, G. S. Jin, Quantum Inf. Comput. 11, 988 (2011). * (42) D. Gonta and P. van Loock, Phys. Rev. A 86, 052312 (2012). * (43) Y.B. Sheng, L. Zhou, G. L. Long, Phys. Rev. A 88, 022302(2013). * (44) B. C. Ren, F. F. Du, and F. G. Deng, Phys. Rev. A 90, 052309 (2014). * (45) Y.B. Sheng, L. Zhou, Sci. Rep. 5, 7815(2015). * (46) G.Y. Wang, Q. Liu, F.G. Deng, Phys. Rev. A 94, 032319 (2016). * (47) T.J. Wang, S.C. Mi, C. Wang, Opt. Express 25, 2969(2017). * (48) F. F. Du, Y. T. Liu, Z. R. Shi, Y. X. Liang, J. Tang, and J. Liu, Opt. Express 27, 27046 (2019). * (49) L. Zhou, Y.B. Sheng, Ann. Physics 10, 385 (2017). * (50) H. Zhang, Q. Liu, X. S. Xu, J. Xiong, A. Alsaedi, T. Hayat, F. G. Deng, Phys. Rev. A 96, 052330 (2017). * (51) L. Zhou, Y.B. Sheng, Ann. Physics 385, 10 (2017). * (52) L. K. Chen, H. L. Yong, P. Xu, X. C. Yao, T. Xiang, Z. D. Li, C. Liu, H. Lu, N. L. Liu, L. Li, T. Yang, C. Z. Peng, B. Zhao, Y.A. Chen, J. W. Pan, Nature Photon. 11, 695 (2017). * (53) Kalb, N. et al. Science 356, 928 (2017). * (54) J. Miguel-Ramiro and W. Dür, Phys. Rev. A 98, 042309 (2018). * (55) S. Krastanov, V. V. Albert, and L. Jiang, Quantum 3, 123123 (2019). * (56) L. Zhou, S. S. Zhang, W. Zhong, Y. B. Sheng, Ann. Phys. 412, 168042 (2020) * (57) L. Zhou, W. Zhong, Y. B. Sheng, Opt. Express 28, 2291 (2020). * (58) D. Y. Chen, Z. Lin, M. Yang, Q. Yang, X. P. Zang, and Z. L. Cao, Phys. Rev. A 102, 022425 (2020). * (59) X. M. Hu, C. X. Huang, Y. B. Sheng, L. Zhou, B. H. Liu, Y. Guo, C. Zhang, W. B. Xing, Y. F. Huang, C. F. Li, and G. C. Guo, Phys. Rev. Lett. 126, 010503 (2021). * (60) M. Murao, M.B. Plenio, S. Popescu, V. Vedral, P.L. Knight, Phys. Rev. A 57, R4075 (1998). * (61) W. Dür, H. Aschauer, and H. J. Briegel, Phys. Rev. Lett. 91, 107903 (2003). * (62) Y. W. Cheong, S. W. Lee, J. Lee, H. W. Lee, Phys. Rev. A 76, 042314 (2007). * (63) Y. B. Sheng, F. G. Deng, B. K. Zhao, T. J. Wang, H. Y. Zhou, Eur. Phys. J. D 55, 235 (2009). * (64) F.G. Deng, Phys. Rev. A 83, 062316 (2011). * (65) Y. B. Sheng, G. L. Long, F. G. Deng, Phys. Lett. A 376, 314 (2011). * (66) F. G. Deng, Phys. Rev. A 84, 052312 (2011). * (67) X. L. Wang, Y. H. Luo, H. L. Huang, et al., Phys. Rev. Lett. 120, 260502 (2018). * (68) W. J. Munro, A. M. Stephens, S. J. Devitt, K. A. Harrison, and K. Nemoto, Nat. Photon. 6, 777 (2012).
# Snapshot Hyperspectral Imaging Based on Weighted High-order Singular Value Regularization Niankai Cheng1, Hua Huang2, Lei Zhang1, and Lizhi Wang1 Corresponding author: Hua<EMAIL_ADDRESS>1School of Computer Science and Technology, Beijing Institute of Technology, Beijing 100081, China 2School of Artifical Intelligence, Beijing Normal University, Beijing 100875, China ###### Abstract Snapshot hyperspectral imaging can capture the 3D hyperspectral image (HSI) with a single 2D measurement and has attracted increasing attention recently. Recovering the underlying HSI from the compressive measurement is an ill-posed problem and exploiting the image prior is essential for solving this ill-posed problem. However, existing reconstruction methods always start from modeling image prior with the 1D vector or 2D matrix and cannot fully exploit the structurally spectral-spatial nature in 3D HSI, thus leading to a poor fidelity. In this paper, we propose an effective high-order tensor optimization based method to boost the reconstruction fidelity for snapshot hyperspectral imaging. We first build high-order tensors by exploiting the spatial-spectral correlation in HSI. Then, we propose a weight high-order singular value regularization (WHOSVR) based low-rank tensor recovery model to characterize the structure prior of HSI. By integrating the structure prior in WHOSVR with the system imaging process, we develop an optimization framework for HSI reconstruction, which is finally solved via the alternating minimization algorithm. Extensive experiments implemented on two representative systems demonstrate that our method outperforms state-of-the- art methods. ## I Introduction Hyperspectral imaging techniques, which are able to capture the 3D hyperspetral image (HSI) with multiple discrete bands at specific frequencies, have attracted increasing interests in recent years. By providing abundant spatial and spectral information simultaneously, HSI can be used for many visual tasks that traditional gray image or color image cannot accomplish. In recent years, HSI has been applied in various vision tasks, such as classification [1], segmentation [2], recognition [3] and tracking [4]. HSI consists of 2D spatial information and 1D spectral information. To obtain the 3D HSI, conventional hyperspectral imaging techniques scan the scene along 1D or 2D coordinate [5, 6, 7, 8]. This imaging is time-consuming and can not be used in dynamic scenes. To address the limitations of conventional spectral imaging techniques, many snapshot hyperspectral imaging systems based on compressive sensing theory [9, 10] has been developed. Among those numerous imaging systems, the coded aperture snapshot spectral imager (CASSI) [11, 12], stands out in the dynamic scene due to its advantage in collecting the 3D data from different wavelengths with one shot. As an advancement system of CASSI, the latest proposed design of dual-camera compressive hyperspectral imaging (DCCHI) incorporates a common panchromatic camera to collect more information simultaneously with CASSI [13, 14]. With the complementary details from the uncoded panchromatic branch, DCCHI can improve the reconstruction fidelity significantly, which obtains the potential ability to apply snapshot hyperspectral imaging into practice [15]. Since the number of snapshot hyperspectral imaging systems is far less than what is required by the Nyquist sampling frequency, the reconstruction is severely under-determined. Such a difficult issue can be addressed by solving a convex optimization problem regularized with a HSI prior. So far considerable reconstruction methods have been proposed in this domain [16, 17, 18, 19, 20, 21]. However, most of them start from expressing HSI as 1D vector or 2D matrix and cannot fully exploit the structurally spectral-spatial prior in 3D HSI. Such a compromise treatment results in poor reconstruction quality and hinder the application of snapshot hyperspectral imaging. Figure 1: Overview of the proposed method. We reconstruct HSI from the compressive measurement. Our reconstruction method, including matching, weighted high-order singular value regularization (shortened as WHOSVR in the figure) and projection, is iteratively performed. To overcome the aforementioned drawbacks, we propose a novel tensor based reconstruction method with weight high-order singular value regularization (WHOSVR). Our key observation is that, compared with the 1D vector and 2D matrix, the high-order tensor is more faithful to delivery the structure information of the 3D HSI. Such advantage of tensor motivates us to exploit the high-order representation for snapshot hyperspectral imaging reconstruction and promote the reconstruction fidelity. Specifically, we first build a high-order tensor by exploiting the high correlation in the spatial and spectral domains for each exemplar cubic patch. Then, to characterize the high-order structure prior of HSI, we propose a WHOSVR model, where the singular values of tensor are treated adaptively, to finish an accurate low- rank tensor recovery. By integrating the structure prior in WHOSVR with the system imaging process, we develop an optimization algorithm for HSI reconstruction, which is finally solved via an alternating minimization algorithm. To the best of our knowledge, it is the first time to exploiting the prior of HSI with WHOSVR for snapshot hyperspectral imaging. Extensive experiments implemented on both CASSI and DCCHI demonstrate that our method outperforms state-of-the-art methods. ## II Related Works ### II-A Hyperspectral Imaging System Conventional scanning-based hyperspectral imaging systems directly trade the temporal resolution for the spatial/ spectral resolution and thus lose the ability to record dynamic scenes [5, 6, 7, 8]. To capture the dynamic scenes, several snapshot hyperspectral imaging systems have been developed in the last decades [22, 23, 24]. However, these systems still suffer from the trade-off between the spatial and temporal resolution. Leveraging the compressive sensing theory, CASSI stands out as a promising solution for the trade-off between the spatial and temporal resolution recently. CASSI employs one disperser or two dispersers to capture a 2D encoded image of the target [11, 12]. Then the underlying HSI can be reconstructed from the compressive measurement by solving an ill-posed problem. Several hardware advancements of CASSI have been developed to improve the performance, e.g. multiple snapshot imaging system [25, 26] and spatial- spectral compressive spectral image [18]. The latest dual-camera design, i.e., DCCHI [13, 14], incorporates a co-located panchromatic camera to collect more information simultaneously with CASSI. With the complementary details from the uncoded image, DCCHI obtains a significant improvement on reconstruction quality and the potential ability to be applied into practice. In this paper, we propose a high-order tensor optimization based method to boost the reconstruction accuracy for snapshot hyperspectral imaging. Meanwhile, our method is a general approach and can be easily extends different systems. ### II-B Hyperspectral Image Reconstruction Recovering the underlying HSI from the compressive measurement plays an essential role for snapshot hyperspectral imaging. Based on compressive sensing, HSI can be reconstructed by solving optimization problems with prior knowledge based regularization. With the hypothesis that nature images owns piecewise smooth property, TV prior based methods have been widely used in snapshot hyperspectral imaging [27, 28, 29]. By conducting wavelet transform or over-completed learned dictionary as sparsity basis, various sparse reconstruction methods have been developed [12, 14, 16, 19, 17]. Matrix rank minimization regularization based on spectral-spatial correlation are also adopted [20, 21]. However, all these methods always start their modeling with vector or matrix and ignore the high-dimensionality nature of HSI, thus leading to poor reconstruction quality. In recent years, several methods on deep learning, including ADMM-Net [30] and ISTA-Net [31] have been developed for compressive sensing of nature image. However, the heterogeneity of HSI makes those methods difficult to be extended for snapshot hyperspectral imaging. A convolutional autoencoder (AE) based method was first designed for CASSI to learn a non-linear sparse representation [32]. A recent work named HSCNN [33], which treats the reconstruction task as an image enhancement task, was proposed for CASSI. HRNet utilized two separate networks to explore the spatial and spectral similarity [34]. The latest design in [35] combined the neural network and the optimization framework to learn the spectral regularization prior (SRP) and achieved suboptimal results. However, these methods all try to learn a single image prior information with network and still ignore the high-dimensionality intrinsic structure of HSI. ## III Approach ### III-A Notations and Preliminaries We first introduce notations and preliminaries as follows. Matrices are denoted as boldface capital letters, e.g., $\bm{X}$, vectors are represented with blodface lowercase letters, e.g. $\bm{x}$, and scalars are indicated as lowercase letters, e.g., $x$. A tensor of order $N$ as boldface Euler script ${\bm{\mathcal{X}}}\in{\mathbb{R}^{{I_{1}}\times{I_{2}}\times\cdot\cdot\cdot{I_{N}}}}$. By varying index $i_{n}$ while keeping the others fixed, the mode-$n$ fiber of ${\mathcal{X}}$ can be obtained. By arranging the mode-$n$ fibers of $\bm{\mathcal{X}}$ as column vectors, we get the mode-$n$ unfolding matrix $\bm{X}_{(n)}\in{\mathbb{R}^{{I_{n}}\times({I_{1}}\cdot\cdot\cdot{I_{n-1}}{I_{n+1}}\cdot\cdot\cdot{I_{N}})}}$. $\text{fold}_{n}(\cdot)$ is the operator that converts the matrix back to the tensor format along the mode-$n$. The _Frobenius_ norm of tensor is defined as the square root of the sum of the squares of all its elements, i.e., ${\left\|{\bm{\mathcal{X}}}\right\|_{F}}=\sqrt{\sum\nolimits_{{i_{1}}=1}^{{I_{1}}}{\sum\nolimits_{{i_{2}}=1}^{{I_{2}}}{\cdot\cdot\cdot\sum\nolimits_{{i_{N}}=1}^{{I_{N}}}{x({{i_{1}},{i_{2}},\cdot\cdot\cdot,{i_{N}}})^{2}}}}}.$ (1) The tensor $n$-mode product is defined as multiplying a tensor by a matrix in mode-$n$. For example, the $n$-mode product of ${\bm{\mathcal{X}}}\in{\mathbb{R}^{{I_{1}}\times{I_{2}}\times\cdot\cdot\cdot{I_{N}}}}$ by a matrix ${\bm{A}}\in{\mathbb{R}^{{J}\times{I_{n}}}}$, which denoted as ${\bm{\mathcal{X}}}~{}{\times}_{n}~{}{\bm{A}}$, is an $N$-order tensor ${\bm{\mathcal{B}}}\in{\mathbb{R}^{{I_{1}}\times\cdot\cdot\cdot{J}\times\cdot\cdot\cdot{I_{N}}}}$, with entries $b({i_{1}},...,{i_{n-1}},j,{i_{n-1}},...,{i_{N}})=\sum\limits_{{i_{n}}=1}^{{i_{N}}}{x\left({{i_{1}},{i_{2}},...,{i_{N}}}\right)a\left({j,{i_{n}}}\right)}.$ (2) With the Tucker decomposition [36], an $N$-order tensor ${\bm{\mathcal{X}}}$ can be decomposed into the following form: ${\bm{\mathcal{X}}}=\bm{\mathcal{G}}~{}{\times}_{1}~{}{\bm{U}_{1}}~{}{\times}_{2}~{}{\bm{U}_{2}}~{}{\times}_{3}\cdot\cdot\cdot{\times}_{N}{\bm{U}_{N}},$ (3) where ${\bm{\mathcal{G}}}\in{\mathbb{R}^{{R_{1}}\times{R_{2}}\times\cdot\cdot\cdot{R_{N}}}}$ is called the core tensor, which is similar to the singular values in matrix SVD, and ${\bm{U}_{i}}\in{\mathbb{R}^{{I_{i}}\times{R_{i}}}}$ is the orthogonal base matrix, which is also similar to the principal components matrix in matrix SVD. Therefore, the Tucker decomposition can be regarded as high-order SVD (HOSVD). Figure 2: Diagram of two representative snapshot hyperspectral imaging systems. ### III-B Observation Model We then give a brief introduction to the observation model of the two representative systems, i.e., CASSI and DCCHI. It is worth mention that our method is general in this filed and suited for other systems, such as the multiple snapshot imaging system and the spatial-spectral encoded imaging system. As shown in Fig. 2, the incident light in CASSI is first projected on the plane of a coded aperture through a objective lens. After spatial coding by the coded aperture, the modulated light goes through a relay lens and is spectrally dispersed in the vertical direction by Amici prism. Finally, the modulated and dispersed spectral information is captured by a panchromatic camera. Let $\bm{\mathcal{F}}\in{\mathbb{R}^{{I}\times{J}\times{\Lambda}}}$ denote the original HSI and $f({i,j,\lambda})$ is its element, where $1{\leq}i{\leq}I$, $1{\leq}j{\leq}J$ index the spatial coordinate and $1{\leq}\lambda{\leq}\Lambda$ indexes the spectral coordinate. The compressive measurement at position $(i,j)$ on the focal plane of CASSI can be represented as: ${y^{c}}(i,j)=\sum\nolimits_{\lambda=1}^{\Lambda}{\rho(\lambda)}\varphi(i-\phi(\lambda),j)f(i-\phi(\lambda),j,\lambda),$ (4) where $\varphi(i,j)$ denotes the modulation pattern of the coded aperture, $\phi(\lambda)$ denotes the dispersion introduced by Amici prism and $\rho(\lambda)$ is the spectral response of the detector. For brevity, let ${\bm{Y}}^{c}\in{\mathbb{R}^{{(I+\Lambda-1)}\times J}}$ denote the matrix representation of ${y^{c}}(i,j)$, and $\bm{\Phi}^{c}$ denote the forward imaging function of CASSI, which is jointly determined by $\rho(\lambda)$, $\varphi(i,j)$ and $\phi(\lambda)$. Then the matrix form of CASSI imaging can be expressed as: ${\bm{Y}}^{c}=\bm{\Phi}^{c}(\bm{\mathcal{F}}).$ (5) As shown in Fig. 2, DCCHI consists of a CASSI branch and a panchromatic camera branch. The incident light first is divided into two directions by the beam splitter equivalently. The light in one direction is captured by the CASSI system according to the above imaging principle, while the light on the other direction is captured directly by a panchromatic camera. The compressive measurement ${y^{p}}(i,j)$ on the panchromatic detector can be represented as: ${y^{p}}(i,j)=\sum\nolimits_{\lambda=1}^{\Lambda}{\rho(\lambda)}f(i,j,\lambda).$ (6) Let ${\bm{Y}}^{p}\in{\mathbb{R}^{{I\times J}}}$ denote the matrix representation of ${y^{p}}(i,j)$, and $\bm{\Phi}^{p}$ denote the forward imaging function of the panchromatic camera, which is determined by $\rho(\lambda)$. Then the matrix form of the panchromatic branch can be expressed as: ${\bm{Y}}^{p}=\bm{\Phi}^{p}(\bm{\mathcal{F}}),$ (7) A general imaging representation of snapshot hyperspectral imaging can be formulated as: ${\bm{Y}}=\bm{\Phi}(\bm{\mathcal{F}}).$ (8) Figure 3: Low-rank property analysis. We exploit the nonlocal similarity across spatial and spectral dimensions to reformulate a low-rank tensor. Then we implement HOSVD on the tensor and show the distribution of singular values in the core tensor. For CASSI, $\bm{Y}=\bm{Y}^{c}$ and $\bm{\Phi}=\bm{\Phi}^{c}$. For DCCHI, $\bm{Y}=[\bm{Y}^{c};\bm{Y}^{p}]$ and $\bm{\Phi}=[\bm{\Phi}^{c};\bm{\Phi}^{p}]$. The goal of HSI reconstruction is to estimate $\bm{\mathcal{F}}$ from the compressive measurement $\bm{Y}$. ### III-C Weighted High-order Singular Value Regularization The key of reconstruction algorithm is to fully exploit the prior information of HSI and build a suitable regularization model. So far most of existing reconstuction methods are on account of two important properties of HSI, i.e., the spatial self-similarity and the spectral correlation. The spatial self- similarity states a nature that there are many image patches around each exemplar patch with the same texture structure. While the spectral correlation indicates that HSI contains a small amount of basis materials and thus exhibits rich redundancy in spectra. In this paper, we utilize tensor low-rank regularization to take such two properties into consideration simultaneously and promote the reconstruction accuracy. TABLE I: Average reconstruction results (PSNR(dB)/SSIM/ERGAS/RMSE) of different methods on CASSI. Indexes | TV | AMP | 3DSR | NSR | LRMA | AE | ISTA | HSCNN | HRNet | SPR | Ours ---|---|---|---|---|---|---|---|---|---|---|--- PSNR | 23.16 | 23.18 | 23.636 | 26.13 | 25.94 | 25.72 | 20.60 | 25.09 | 22.83 | 24.48 | 28.05 SSIM | 0.7130 | 0.6600 | 0.7311 | 0.7610 | 0.7930 | 0.7720 | 0.5499 | 0.7334 | 0.6648 | 0.7395 | 0.8302 ERGAS | 258.32 | 256.76 | 245.153 | 189.19 | 195.63 | 197.32 | 344.57 | 206.97 | 268.65 | 224.19 | 153.06 RMSE | 0.0469 | 0.0474 | 0.0457 | 0.0333 | 0.0315 | 0.0333 | 0.0653 | 0.0373 | 0.0496 | 0.0451 | 0.0236 For one cubic patch with the size of ${s}\times{s}\times\Lambda$ across full bands of HSI $\bm{\mathcal{F}}\in{\mathbb{R}^{{I}\times{J}\times{\Lambda}}}$, we search for its $k-1$ nearest neighbors patches in a local window. By reordering the spatial block of each band into a 1D column vector, the constructed 3-order tensor $\bm{\mathcal{S}}$ with the size of $s^{2}\times\Lambda\times k$ is formed. The constructed tensor simultaneously exhibits the spatial self-similarity (mode-1), the spectral correlation (mode-2) and the joint correlation (mode-3). Here we take the multi- dimensionality property of tensor and introduce low-rank tensor recovery model to preserve the structure information of HSI. Consequently, we build a basic regularization towards low-rank tensor recovery: ${\Gamma(\bm{\mathcal{S}})}=\tau{\left\|{{{\bf{R}}(\bm{\mathcal{F}})}-\bm{\mathcal{S}}}\right\|_{F}^{2}}+\text{rank}(\bm{\mathcal{S}}),$ (9) where ${\bf{R}}(\bm{\mathcal{F}})$ represents extracting the 3D tensor from $\bm{\mathcal{F}}$ and $\tau$ denotes the penalty factor. With the Tucker decomposition [36], the rank of a 3-order tensor can be defined as the sum of ranks of unfolding matrices along three modes. However, considering the rank of different modes separately still ignore the correlation between different modes. Meanwhile, estimating the rank of a matrix is still a NP-hard problem. We return to the definition of Tucker decomposition and analyze the low rank property of the constructed tensor. As shown in Fig. 3, we implement Tucker decomposition on a nonlocal tensor extracted from a clean HSI and show the distribution of singular values in the core tensor. We can see that the singular values tend to be dropping to zero fleetly, which indicates that the core tensor exhibits significant sparsity. Therefore, we can introduce a weighted $\ell_{1}$ norm to pursue the tensor rank minimization, i.e.,: $\text{rank}(\bm{\mathcal{S}})=\left\|\bf{w}\circ\bm{\mathcal{G}}\right\|_{1}~{}~{}s.t.\bm{\mathcal{S}}=\bm{\mathcal{G}}~{}{\times}_{1}~{}{\bm{U}_{1}}~{}{\times}_{2}~{}{\bm{U}_{2}}~{}{\times}_{3}~{}{\bm{U}_{3}},$ (10) where $\left\|\bf{w}\circ\bm{\mathcal{G}}\right\|_{1}=\sum\nolimits_{n}{{w_{n}}\left|{{g_{n}}}\right|}$ and ${g_{n}}$ is the element of $\bm{\mathcal{G}}$. The weight $w_{n}$ is set as: $w_{n}^{t+1}={c\mathord{\left/{\vphantom{1{\left({\left|{w_{i}^{t}}\right|+\varepsilon}\right)}}}\right.\kern-1.2pt}{\left({\left|{w_{i}^{t}}\right|+\varepsilon}\right)}},$ (11) where $t$ denotes the $t$-th iteration, $c$ is a positive constant number and $\varepsilon\leq 10^{-6}$. Such formulation derives a meaningful outcome that the singular values can be penalized adaptively and structure information can be better preserved. Specifically, those greater singular values in the $t$-th iteration, which deliver more important structure information, will get a smaller weight and be shrunk less at $(t+1)$-th iteration. | | | | | ---|---|---|---|---|--- TV | AMP | 3DSR | NSR | LRMA | AE (20.13 / 0.6750) | (19.61 / 0.5699) | (20.81 / 0.7030) | (23.46 / 0.7559) | (22.73 / 0.8063) | (22.84 / 0.7593) | | | | | ISTA | HSCNN | HRNet | SRP | Ours | GT (18.20 / 0.5173) | 22.66 / 0.8107) | (19.76 / 0.6314) | (22.16 / 0.7421) | (26.88 / 0.8756) | (PSNR /SSIM) | | | | | TV | AMP | 3DSR | NSR | LRMA | AE (20.42 / 0.7696) | (22.96 / 0.7202) | (22.02 / 0.7796) | (26.90 / 0.8402) | (24.93 / 0.8633) | (26.40 / 0.8496) | | | | | ISTA | HSCNN | HRNet | SRP | Ours | GT (19.66 / 0.5135) | (25.80 / 0.8335) | (21.51 / 0.6998) | (23.83 / 0.7675) | (30.39 / 0.9159) | (PSNR / SSIM) Figure 4: Reconstructed quality comparison of CASSI. The PSNR and SSIM for the result images of _chart and stuffed toy_ and _stuffed toys_ are shown in the parenthesis. By comparing the reconstructed results and ground truth (shortened as GT in the figure), our method obtains better spatial contents and textures. By combining Equation 9 and 10, we obtain the WHOSVR model as: ${\Gamma(\bm{\mathcal{G}})}=\tau{\left\|{{{\bf{R}}(\bm{\mathcal{F}})}-\bm{\mathcal{G}}~{}{\times}_{1}~{}{\bm{U}_{1}}~{}{\times}_{2}~{}{\bm{U}_{2}}~{}{\times}_{3}~{}{\bm{U}_{3}}}\right\|_{F}^{2}}+\left\|\bf{w}\circ\bm{\mathcal{G}}\right\|_{1}.$ (12) In the following, we will illustrate the WHOSVR based reconstruction method for snapshot hyperspectral imaging. ### III-D Reconstruction Method Based on the analysis above, a general reconstruction formulation for snapshot hyperspectral imaging is proposed: $\begin{split}\mathop{\min}\limits_{\bm{\mathcal{F}},\bm{\mathcal{G}}_{l}}\;&\frac{1}{2}\left\|{{\bm{Y}}-\bm{\Phi}(\bm{\mathcal{F}})}\right\|_{F}^{2}+\sum\nolimits_{l=1}^{L}\big{(}\\\ &\tau\left\|{{{\bf{R}}_{l}(\bm{\mathcal{F}})}-\bm{\mathcal{G}}_{l}~{}{\times}_{1}~{}{\bm{U}_{l,1}}~{}{\times}_{2}~{}{\bm{U}_{l,2}}~{}{\times}_{3}~{}{\bm{U}_{l,3}}}\right\|_{F}^{2}+\left\|{\bf{w}}_{l}\circ{\bm{\mathcal{G}}}_{l}\right\|_{1}\big{)},\end{split}$ (13) where $L$ denotes the total number of constructed tensors. To optimize Equation 13, we adopt an alternating minimization scheme to split it into two finer subproblems: updating core tensor $\bm{\mathcal{G}}_{l}$ and updating the whole HSI $\bm{\mathcal{F}}$. #### III-D1 Updating Core Tensor $\bm{\mathcal{G}}_{l}$ By fixing HSI $\bm{\mathcal{F}}$, we can estimate each low-rank tensor $\bm{\mathcal{G}}_{l}$ independently by solving the following reformulated equation: $\mathop{\min}\limits_{\bm{\mathcal{G}}_{l}}\tau\left\|{{{\bf{R}}_{l}(\bm{\mathcal{F}})}-\bm{\mathcal{G}}_{l}~{}{\times}_{1}~{}{\bm{U}_{l,1}}~{}{\times}_{2}~{}{\bm{U}_{l,2}}~{}{\times}_{3}~{}{\bm{U}_{l,3}}}\right\|_{F}^{2}+\left\|{\bf{w}}_{l}\circ\bm{\mathcal{G}}_{l}\right\|_{1},$ (14) Given ${{\bf{R}}_{l}(\bm{\mathcal{F}})}={\hat{\bm{\mathcal{G}}}_{l}}~{}{\times}_{1}~{}{\bm{U}_{l,1}}~{}{\times}_{2}~{}{\bm{U}_{l,2}}~{}{\times}_{3}~{}{\bm{U}_{l,3}}$ be the Tucker decomposition of ${{\bf{R}}_{l}(\bm{\mathcal{F}})}$, the solution of $\bm{\mathcal{G}}_{l}$ in Equation 14 is: ${g_{n}}=\text{max}({\hat{g_{n}}}-\frac{w_{n}}{2\tau},0),$ (15) where $\hat{g_{n}}$ is the element of $\hat{\bm{\mathcal{G}}}_{l}$. TV | AMP | 3DSR | NSR | LRMA | Ours ---|---|---|---|---|--- | | | | | | | | | | Figure 5: The error maps comparison of two typical scenes on DCCHI. It shows that our method can produce relatively higher spatial fidelity. #### III-D2 Updating Whole HSI $\bm{\mathcal{F}}$ Once we obtain the core tensor $\bm{\mathcal{G}}_{l}$, the whole HSI $\bm{\mathcal{F}}$ can be updated by solving the following problem: $\begin{split}&\qquad\mathop{\min}\limits_{\bm{\mathcal{F}}}\;\frac{1}{2}\left\|{{\bm{Y}}-\bm{\Phi}(\bm{\mathcal{F}})}\right\|_{F}^{2}+\\\ &\sum\nolimits_{l=1}^{L}\tau\left\|{{{\bf{R}}_{l}(\bm{\mathcal{F}})}-\bm{\mathcal{G}}_{l}{\times}_{1}{\bm{U}_{l,1}}{\times}_{2}{\bm{U}_{l,2}}{\times}_{3}{\bm{U}_{l,3}}}\right\|_{F}^{2}.\end{split}$ (16) Equation 16 is a quadratic optimization problem, so $\bm{\mathcal{F}}$ admits a straightforward least-square solution: $\displaystyle\bm{\mathcal{F}}=\big{(}{{\bm{\Phi}^{T}}\bm{\Phi}+2\tau\sum\nolimits_{l}{{{{{{\bf{R}}_{l}^{T}}}}}{{\bf{R}}_{l}}}}\big{)}^{{\rm{-}}1}}\big{(}{{\bm{\Phi}^{T}}(\bm{Y})$ (17) $\displaystyle+2\tau\sum\nolimits_{l=1}^{L}{{{{{{\bf{R}}_{l}^{T}}}}}(\bm{\mathcal{G}}_{l}~{}{\times}_{1}~{}{\bm{U}_{l,1}}~{}{\times}_{2}~{}{\bm{U}_{l,2}}~{}{\times}_{3}~{}{\bm{U}_{l,3}})}\big{)}.$ In practice, Equation 17 can be solved by the conjugate gradient algorithm [37]. | | | ---|---|---|--- Figure 6: The absolute spectral error on DCCHI between the ground truth and reconstructed results of the white labels in Fig. 4. It shows that our method obtains the highest spectral accuracy. TABLE II: Average reconstruction results of different methods on DCCHI. Indexes | TV | AMP | 3DSR | NSR | LRMA | Ours ---|---|---|---|---|---|--- PSNR | 28.51 | 28.52 | 28.32 | 32.58 | 37.45 | 37.81 SSIM | 0.8938 | 0.8526 | 0.9037 | 0.9377 | 0.9730 | 0.9733 ERGAS | 167.14 | 140.75 | 163.67 | 107.71 | 57.30 | 51.38 RMSE | 0.0525 | 0.0263 | 0.0337 | 0.0285 | 0.0132 | 0.0069 ## IV Experiments In this section, we conduct experiments on two representative snapshot hyperspectral imaging systems, i.e., CASSI and DCCHI, to verify the performance of our method. ### IV-A Implementation Details We generate the mask of coded aperture in CASSI as a random Bernoulli matrix with $p=0.5$. The dispersion of the Amici prism obeys a linear distribution across the wavelength dimension. The Columbia dataset, which contains 32 various real-world objects from 400nm to 700nm (31 bands with 10nm interval), are used as synthetic data. In our experiment, the resolution of all tested images is cropped into $256\times 256$ at the center region cross full bands. Meanwhile, we use 22 HSIs for training and 10 HSIs for testing. Our algorithm is compared with 5 prior knowledge based methods, i.e., TV regularization solved by TwIST [38], 3D sparse reconstruction (3DSR) [18], approximate message passing (AMP) [17], nonlocal sparse representation (NSR) [19] and low-rank matrix approximation (LRMA) [20] and 5 deep learning based methods, i.e., AE [32], HSCNN [33], ISTA-Net (shorten as ISTA) [31], HRNet [34] and SRP[35]. The parameters for our method are set as following. The penalty factor $\tau=1$, the constant $c=0.0055$, and the cubic spatial size $s=5$ with step length of 4. We search $k=45$ nonlocal similar patches within a [-20, 20] window. We set the max iteration number of the alternating minimization scheme as 600 for the stop criterion. For the competitive methods we make great efforts to achieve the best results for all the competing methods according to their publication and released codes. We execute our experiments on a platform of the Windows 10 64-bit system with I7 6700 and 64GB RAM. For quantitative evaluation, four image quality indexes are employed, including peek signal-to-noise ratio (PSNR), structure similarity (SSIM) [39], erreur relative globale adimensionnelle de synthèse (ERGAS)[40] and root mean square error (RMSE). PSNR and SSIM measure the visual quality and the structure similarity, respectively. ERGAS and RMSE measure the spectral fidelity. Generally, a bigger PSNR and SSIM and a smaller ERGAS and RMSE suggest a better reconstruction fidelity. ### IV-B Evaluation Results The average quantitative results of all methods on CASSI are shown in Table I. The best values for each index are highlighted in bold. We can see that our method obtains remarkable promotion in PSNR, SSIM, ERGAS and RMSE compared with other methods. Specifically, our method produces noticeable quality promotion compared with TV, AMP, 3DSR and NSR. It indicates that low rank prior can recover more structure information than sparse prior. The promotion upon LRMA demonstrates that the high-dimensional tensor is more powerful than matrix to exploit the intrinsic nature of HSI. Meanwhile, promotion upon the deep learning based methods indicates that the tensor based optimization can more faithfully account for the high-dimensionality structure. For a visual comparison, we convert HSI into RGB image using the CIE color mapping function. The synthetic RGB images of two representative scenes, i.e., _chart and stuffed toy_ and _stuffed toys_ , are shown in Fig. 4. We can see that our method can produce not only clearer spatial textures but also higher spectral accuracy. Next, we evaluate the reconstruction performance on DCCHI. It should be noted that the mismatch in the two branches of DCCHI results in that block layout based learning methods can not be implemented. So here we compare our method with the prior regularization based methods. The average quantitative results are presented in Table II. It shows that our method can produce the best quantitative performance. We then show the average absolute error maps between the ground truth and restored results across spectra in Fig. 5. It can be seen that the results produced by our method are closer to the ground truth compared with other methods, which verifies that our method obtains higher spatial accuracy. Futher, we show the average absolute error curves between the ground truth and reconstructed results across spectra in Fig. 6. It demonstrates that our method obtains higher spectral accuracy. ## V Conclusion In this paper, we proposed a novel and general reconstruction method with low- rank tensor recovery based on WHOSVR for snapshot hyperspectral imaging. We introduce 3D tensors to exploit the high-order nature of HSI, including spatial self-similarity, spectral correlation and spatial-spectral joint correlation. Specifically, we first construct a 3D tensor for each exemplar cubic patch of HSI. We then proposed to utilize WHOSVR to characterize the high correlation in each mode of the formulated tensors. Finally, an iterative optimization algorithm based on WHOSVR was developed to finish high-accuracy HSI reconstruction. Through experiments implemented on two representative systems verified that our method outperforms state-of-the-art methods. ## Acknowledgment This work is supported by the National Natural Science Foundation of China under Grant No. 62072038 and No. 61922014. ## References * [1] H. Kim Min, Todd Alan Harvey, David S. Kittle, Holly Rushmeier, Julie Dorsey, Richard O. Prum, and David J. Brady. 3d imaging spectroscopy for measuring hyperspectral patterns on solid objects. ACM Transactions on Graphics, 31(4):1–11, 2012. * [2] Yuliya Tarabalka, Jocelyn Chanussot, and Jón Atli Benediktsson. Segmentation and classification of hyperspectral images using minimum spanning forest grown from automatically selected markers. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 40(5):1267–1279, 2009. * [3] Zhihong Pan, Glenn Healey, Manish Prasad, and Bruce Tromberg. Face recognition in hyperspectral images. IEEE Transactions on Pattern Analysis and Machine Intelligence, 25(12):1552–1560, 2003. * [4] Hien Van Nguyen, Amit Banerjee, and Rama Chellappa. Tracking via object reflectance using a hyperspectral video camera. In IEEE Conference on Computer Vision and Pattern Recognition Workshops, pages 44–51, 2010. * [5] Gregg Vane, Robert O Green, Thomas G Chrien, Harry T Enmark, Earl G Hansen, and Wallace M Porter. A system overview of the airborne visible/infrared imaging spectrometer (aviris). In Imaging Spectroscopy II, volume 834, pages 22–32, 1987. * [6] Robert W Basedow, Dwayne C Carmer, and Mark E Anderson. Hydice system: Implementation and performance. In Imaging Spectrometry, volume 2480, pages 258–268, 1995. * [7] Masahiro Yamaguchi, Hideaki Haneishi, Hiroyuki Fukuda, Junko Kishimoto, Hiroshi Kanazawa, Masaru Tsuchida, Ryo Iwama, and Nagaaki Ohyama. High-fidelity video and still-image communication based on spectral information: Natural vision system and its applications. In Spectral Imaging: Eighth International Symposium on Multispectral Color Science, volume 6062, page 60620G, 2006. * [8] Yoav Y. Schechner and Shree K. Nayar. Generalized mosaicing: Wide field of view multispectral imaging. IEEE Transactions on Pattern Analysis and Machine Intelligence, 24(10):1334–1348, 2002. * [9] Richard G Baraniuk. Compressive sensing [lecture notes]. IEEE Signal Processing Magazine, 24(4):118–121, 2007. * [10] Emmanuel J Candes. Compressive sampling. In Proceedings of the international congress of mathematicians, volume 3, pages 1433–1452, 2006. * [11] Gonzalo R Arce, David J Brady, Lawrence Carin, Henry Arguello, and David S Kittle. Compressive coded aperture spectral imaging: An introduction. IEEE Signal Processing Magazine, 31(1):105–115, 2014. * [12] Ashwin Wagadarikar, Renu John, Rebecca Willett, and David Brady. Single disperser design for coded aperture snapshot spectral imaging. Applied Optics, 47(10):B44–B51, 2008. * [13] Lizhi Wang, Zhiwei Xiong, Dahua Gao, Guangming Shi, and Feng Wu. Dual-camera design for coded aperture snapshot spectral imaging. Applied Optics, 54(4):848–58, 2015. * [14] Lizhi Wang, Zhiwei Xiong, Dahua Gao, Guangming Shi, Wenjun Zeng, and Feng Wu. High-speed hyperspectral video acquisition with a dual-camera architecture. In IEEE Conference on Computer Vision and Pattern Recognition, pages 4942–4950, 2015. * [15] Shipeng Zhang, Huang Hua, and Ying Fu. Fast parallel implementation of dual-camera compressive hyperspectral imaging system. IEEE Transactions on Circuits and Systems for Video Technology, 2018\. * [16] Mário AT Figueiredo, Robert D Nowak, and Stephen J Wright. Gradient projection for sparse reconstruction: Application to compressed sensing and other inverse problems. IEEE Journal of selected topics in signal processing, 1(4):586–597, 2007. * [17] Jin Tan, Yanting Ma, Hoover Rueda, Dror Baron, and Gonzalo R Arce. Compressive hyperspectral imaging via approximate message passing. IEEE Journal of Selected Topics in Signal Processing, 10(2):389–401, 2016. * [18] Xing Lin, Yebin Liu, Jiamin Wu, and Qionghai Dai. Spatial-spectral encoded compressive hyperspectral imaging. ACM Transactions on Graphics, 33(6):1–11, 2014. * [19] Lizhi Wang, Zhiwei Xiong, Guangming Shi, Feng Wu, and Wenjun Zeng. Adaptive nonlocal sparse representation for dual-camera compressive hyperspectral imaging. IEEE Transactions on Pattern Analysis and Machine Intelligence, 39(10):2104–2111, 2017. * [20] Ying Fu, Yinqiang Zheng, Imari Sato, and Yoichi Sato. Exploiting spectral-spatial correlation for coded hyperspectral image restoration. In IEEE Conference on Computer Vision and Pattern Recognition, pages 3727–3736, 2016. * [21] Yang Liu, Xin Yuan, Jinli Suo, David Brady, and Qionghai Dai. Rank minimization for snapshot compressive imaging. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2018\. * [22] Xun Cao, Hao Du, Xin Tong, Qionghai Dai, and Stephen Lin. A prism-mask system for multispectral video acquisition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 33(12):2423–2435, 2011. * [23] Xun Cao, Tao Yue, Xing Lin, Stephen Lin, Xin Yuan, Qionghai Dai, Lawrence Carin, and David J Brady. Computational snapshot multispectral cameras: toward dynamic capture of the spectral world. IEEE Signal Processing Magazine, 33(5):95–108, 2016. * [24] Liang Gao and Lihong V Wang. A review of snapshot multidimensional optical imaging: Measuring photon tags in parallel. Physics reports, 616:1–37, 2016. * [25] David Kittle, Kerkil Choi, Ashwin Wagadarikar, and David J Brady. Multiframe image estimation for coded aperture snapshot spectral imagers. Applied Optics, 49(36):6824, 2010. * [26] Yuehao Wu, Iftekhar O Mirza, Gonzalo R Arce, and Dennis W Prather. Development of a digital-micromirror-device-based multishot snapshot spectral imaging system. Optics Letters, 36(14):2692–2694, 2011. * [27] Ashwin A Wagadarikar, Nikos P Pitsianis, Xiaobai Sun, and David J Brady. Spectral image estimation for coded aperture snapshot spectral imagers. In Image Reconstruction from Incomplete Data V, volume 7076, page 707602, 2008. * [28] Shipeng Zhang, Huang Hua, and Ying Fu. Fast parallel implementation of dual-camera compressive hyperspectral imaging system. IEEE Transactions on Circuits and Systems for Video Technology, pages 1–1, 2018. * [29] Shipeng Zhang, Lizhi Wang, Ying Fu, and Hua Huang. Gpu assisted towards real-time reconstruction for dual-camera compressive hyperspectral imaging. In Pacific Rim Conference on Multimedia, pages 711–720. Springer, 2018. * [30] Jian Sun, Huibin Li, and Zongben Xu. Deep admm-net for compressive sensing mri. In Advances in Neural Information Processing Systems, pages 10–18, 2016. * [31] Jian Zhang and Bernard Ghanem. Ista-net: Interpretable optimization-inspired deep network for image compressive sensing. In IEEE Conference on Computer Vision and Pattern Recognition, pages 1828–1837, 2018. * [32] Inchang Choi, Daniel S Jeon, Giljoo Nam, Diego Gutierrez, and Min H Kim. High-quality hyperspectral reconstruction using a spectral prior. ACM Transactions on Graphics, 36(6):218, 2017. * [33] Zhiwei Xiong, Zhan Shi, Huiqun Li, Lizhi Wang, Dong Liu, and Feng Wu. Hscnn: Cnn-based hyperspectral image recovery from spectrally undersampled projections. In IEEE International Conference on Computer Vision Workshops, pages 518–525, 2017. * [34] Lizhi Wang, Tao Zhang, Ying Fu, and Hua Huang. Hyperreconnet: Joint coded aperture optimization and image reconstruction for compressive hyperspectral imaging. IEEE Transactions on Imaging Processing, 28(5):2257–2270, 2018\. * [35] Lizhi Wang, Chen Sun, Ying Fu, Min H Kim, and Hua Huang. Hyperspectral image reconstruction using a deep spatial-spectral prior. In IEEE Conference on Computer Vision and Pattern Recognition, pages 8032–8041, 2019. * [36] Ledyard R Tucker. Some mathematical notes on three-mode factor analysis. Psychometrika, 31(3):279–311, 1966. * [37] Magnus R. Hestenes and Eduard Stiefel. Methods of conjugate gradients for solving linear systems. J. Research Nat. Bur. Standards, 49:409–436 (1953), 1952. * [38] José M Bioucas-Dias and Mário AT Figueiredo. A new twist: Two-step iterative shrinkage/thresholding algorithms for image restoration. IEEE Transactions on Image Processing, 16(12):2992–3004, 2007. * [39] Zhou Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli. Image quality assessment: from error visibility to structural similarity. IEEE Transactions on Image Processing, 13(4):600–612, 2004. * [40] Lucien Wald. Data fusion: definitions and architectures: fusion of images of different spatial resolutions. Presses des MINES, 2002.
# Numerical analysis of a deep learning formulation of elastic full waveform inversion with high order total variation regularization in different parameterization Tianze Zhang11footnotemark: 1 Jian Sun11footnotemark: 122footnotemark: 2 Kristopher A. Innanen11footnotemark: 1 Daniel Trad11footnotemark: 1 ###### Abstract We have formulated elastic seismic full waveform inversion (FWI) within a deep learning environment. Training this network with a single seismic data set is equivalent to carrying out elastic FWI. There are three main motivations for this effort. The first is an interest in developing an inversion algorithm which is more or less equivalent to standard elastic FWI but which is ready- made for a range of cloud computational architectures. The second is an interest in algorithms which can later (i.e., not in this paper) be coupled with a more involved training component, involving multiple datasets. The third is a general interest in developing the idea of theory-guiding within machine learning solutions for large geophysical problems, wherein the number of degrees of freedom within a network, and the reliance on exhaustive training data, are both reduced by constraining the network with physical rules. In our formulation, a recurrent neural network is set up with rules enforcing elastic wave propagation, with the wavefield projected onto a measurement surface acting as the synthetic data to be compared with observed seismic data. Gradients for iterative updating of an elastic model, with a variety of parameterizations and misfit functionals, can be efficiently constructed within the network through the automatic differential method. With this method, the inversion based on complex misfits can be calculated. We survey the impact of different complex misfits (based on the $l_{2}$, $l_{1}$) with high order total variation (TV) regulations on multiparameter elastic FWI recovery of models within velocity/density, modulus/density, and stiffness parameter/density parameterizations. We analyze parameter cross-talk. Inversion results on simple and complex models show that the RNN elastic FWI with high order TV regulation using $l_{1}$ norm can help mitigate cross-talk issues with gradient-based optimization methods. ## 1 Introduction It has recently been shown (Sun et al.,, 2020) that seismic wave propagation can be simulated with a specialized recurrent neural network (RNN), and that the process of training such a network with a single seismic data set is equivalent to carrying out seismic full waveform inversion (FWI). We are motivated to extend and expand on these results, because of (1) the apparent potential for wider training of such a network to combat common FWI issues such as modelling error; (2) the opportunities for efficient computation (e.g., cloud) offered by FWI realized on platforms like TensorFlow, and (3) an interest in the behavior of more complex RNN-FWI formulations than previously analyzed, e.g., multi-parameter elastic FWI. In this paper we report on progress on the third of these. The application of machine learning methods to seismic problems has been underway for decades; for example, Röth and Tarantola, (1994) presented a neural network-based inversion of time-domain seismic data for acoustic velocity depth-profiles. However, the evolution of these network approaches into deep learning methods, and the results which have subsequently become possible, make many aspects of the discipline quite new, and explain the major recent surge in development and interest. Now, novel seismic applications are being reported in fault detection, denoising, reservoir prediction and velocity inversion (e.g., Jin et al.,, 2019; Zheng et al.,, 2019; Peters et al.,, 2019; Chen et al.,, 2019; Li et al., 2019a, ; Smith et al.,, 2019; Shustak and Landa,, 2018). Specifically in seismic velocity inversion, Sun and Demanet, (2018) applied a deep learning method to the problem of bandwidth extension; Zhang et al., (2019) designed an end-to-end framework, the velocity-GAN, which generates velocity images directly from the raw seismic waveform data; Wu and Lin, (2018) trained a network with an encoder-decoder structure to model the correspondence between seismic data and subsurface velocity structures; and Yang and Ma, (2019) investigated a supervised, deep convolutional neural network for velocity-model building directly from raw seismograms. The above examples are purely data-driven, in the sense that they involve no assumed theoretical/physical relationships between the input layer (e.g., velocity model) and output layer (e.g., seismic data). We believe that the advantages of the purely data-driven methods are that once the training for the network to perform inversion is finished, take the data-driven training for a network that can perform FWI as an example, the raw seismograms can be directly mapped to the velocity models. This would become a faster inversion method compared with the conventional inversion method that requires iterations. However, seismic inversion is a sophisticated issue, so how we choose the sufficient amount of training data sets that can represent such complex wave physics features and their corresponding velocity models is a hard problem. Here we distinguish between such methods and those belonging to the theory- guided AI paradigm (e.g., Wagner and Rondinelli,, 2016; Wu et al.,, 2018; Karpatne et al.,, 2017). Theory-guided deep learning networks are, broadly, those which enforce physical rules on the relationships between variables and/or layers. This may occur through the training of a standard network with simulated or synthetic data, which are created using physical rules, or by holding some weights in a network, chosen to force the network to mimic a physical process, fixed and un-trainable. Theory-guiding was explicitly used in the former sense by Downton and Hampson, (2018), in which a network designed to predict well log properties from seismic amplitudes was trained not only with real data but with synthetics derived from the Zoeppritz equations. Bar-Sinai et al., (2019) and Raissi, (2018) built deep convolutional neural networks (or CNNs) to solve partial differential equations, i.e., explicitly using theoretical models within the design, which is an example of theory guiding in the latter sense. Sun et al., (2020), in the work we extend in this paper, similarly set up a deep recurrent neural network to simulate the propagation of a seismic wave through a general acoustic medium. The network is set up in such a way that the trainable weights correspond to the medium property unknowns (i.e., wave velocity model), and the non-trainable weights correspond to the mathematical rules (differencing, etc.) enforcing wave propagation. The output layer was the wave field projected onto a measurement surface, i.e., a simulation of measured seismic data. The training of the Sun et al., (2020) network with a single data set was shown to be an instance of full waveform inversion. Conventional (i.e., not network-based) seismic full waveform inversion, or FWI, is a complex data fitting procedure aimed at extracting important information from seismic records. Key ingredients are an efficient forward modeling method to simulate synthetic data, and a local differential approach, in which the gradient and direction are calculated to update the model. When updating several parameters simultaneously, which is a multiparameter full waveform inversion, error in one parameter tends to produce updates in others, a phenomenon referred to as inter-parameter trade-off or cross-talk (e.g., Kamei and Pratt,, 2013; Alkhalifah and Plessix,, 2014; Innanen,, 2014; Pan et al.,, 2018; Keating and Innanen,, 2019). The degree of trade-off or cross-talk between parameters can depend sensitively on the specific parameterization; even though, for instance, $\lambda$, $\mu$ and $\rho$ have the same information content as $V_{P}$, $V_{S}$, and $\rho$, updating in one can produce markedly different convergence properties and final models that doing so in the other. The coupling between different physical parameters is controlled by the physical relationships amongst these parameters, and the relationships between the parameters and the wave which interacts with them. Tarantola, (1986), by using the scattering or radiation pattern, systematically analyzed the effect of different model parameterizations on isotropic FWI. He suggested that the greater the difference in the scattering pattern between each parameter, the better the parameters can be resolved. Köhn et al., (2012) showed that across all geophysical parameterizations, within isotropic-elastic FWI, density is always the most challenging to resolve (e.g., Tarantola,, 1984; Plessix,, 2006; Kamath et al.,, 2018; Lailly,, 1983; Operto et al.,, 2013; Oh and Alkhalifah,, 2016; Pan et al.,, 2016; Keating and Innanen,, 2017). In RNN-based FWI, then, a natural step is the extension of the Sun et al., (2020) result, which involves a scalar-acoustic formulation of FWI, to a multi-parameter elastic version. Cells within a deep recurrent neural network are designed such that the propagation of information through the network is equivalent to the propagation of an elastic wavefield through a 2D heterogeneous and isotropic-elastic medium; the network is equipped to explore a range of parameterizations and misfit functionals for training based on seismic data. As with the acoustic network, the output layer is a projection of the wave field onto some defined measurement surface, simulating measured data. In addition to providing a framework for inversion methods which mixes the features of FWI with the training capacity of a deep learning network, this approach also allows for efficient calculation of the derivatives of the residual through automatic differential (AD) methods (e.g., Li et al., 2019b, ; Sambridge et al.,, 2007), an engine for which is provided by the open-source machine learning library Pytorch (Paszke et al.,, 2017). Our recurrent neural network is designed using this library. The forward simulation of wave propagation is represented by a Dynamic Computational Graph (DCG), which records how each internal parameter is calculated from all previous ones. In the inversion, the partial derivatives of the residual with respect to any parameter (within a trainable class) is computed by (1) backpropagating within the computational graph to that parameter, and (2) calculating the partial derivative along the path using the chain rule. This paper is organized as follows. First, we introduce the basic structure of the recurrent neural network and how the gradients can be calculated using the backpropagation method. Second, we demonstrate how the elastic FWI RNN cell is constructed in this paper and how the waveflieds propagate in the RNN cells. Third, we explain the $l_{2}$ and $l_{1}$ misfits with high order TV regulation and the mathematical expression for the gradients based on these misfits. Fourth, we use simple layers models and complex over-thrust models to perform inversions with various parameterizations using different misfits. Finally, we discuss the conclusions of this work. ## 2 Recurrent Neural Network A recurrent neural network (RNN) is a machine learning network suitable for dealing with data which have some sequential meaning. Within such a network, the information generated in one cell layer can be stored and used by the next layer. This design has natural applicability in the processing and interpretation of the time evolution of physical processes (Sun et al.,, 2020), and time-series data in general; examples include language modeling (Mikolov,, 2012), speech recognition Graves et al., (2013), and machine translation (Kalchbrenner and Blunsom,, 2013; Zaremba et al.,, 2014). Figure 1: Forward propagation through an example RNN. Cells are identical in form, with the output of one being the input of the next. $\textbf{O}=[O_{1},O_{2},O_{3},{\cdots}]$ are the internal variables in each cell; $\textbf{S}=[S_{1},S_{2}]$ are the inputs; $\textbf{P}=[P_{1},P_{2}]$ are the outputs; $\textbf{L}=[L_{1},L_{2}]$ are the labeled data; $\textbf{W}=[W_{1},W_{2}]$ are the trainable weights. Black dashed line indicates forward propagation. Figure 2: Residual backward propagation through an example RNN. $\textbf{R}=[R_{1},R_{2}]$ are the residuals at each RNN cell, which are calculated using absolute error($l_{1}$ norm). Gray solid line shows how gradients are calculated using back propagation from the residual to the trainable weights in RNN cell along the computational graph. Figure 1 illustrates the forward propagation of information through an example RNN, in which each RNN cell represents an instant in time: the $\textbf{O}=[O_{1},O_{2},O_{3},{\cdots}]$ are the internal variables in each RNN cell; $\textbf{S}=[S_{1},S_{2}]$ are the input at each time step; $\textbf{P}=[P_{1},P_{2}]$ are the output; $\textbf{L}=[L_{1},L_{2}]$ are the labeled data; and $\textbf{W}=[W_{1},W_{2}]$ are the trainable parameters at each time step. Mathematical operations relating the internal variables within a cell, and those relating internal variables across adjacent cells, are represented as arrows. To train this network is to select trainable weights such that labeled data L and RNN output P are as close as possible. Specifically, in training we determine the parameters W, through a gradient-based optimization involving the partial derivatives of the residual with respect to each $W_{i}$. These derivatives are determined through backpropagation, which is a repeated application the chain rule, organized to resemble flow in the reverse direction along with the arrows in the network. For the example RNN in Figure 1, this takes the form illustrated in Figure 2, in which the sequence $\textbf{R}=[R_{1},R_{2}]$ represents the residuals at each time step. Within this example, to calculate the partial derivative of $R_{1}$ with respect to $W_{1}$, we back-propagate from node $R_{1}$ to $W_{1}$: $\frac{{\partial}R_{1}}{{\partial}W_{1}}=\frac{{\partial}R_{1}}{{\partial}P_{1}}\frac{{\partial}P_{1}}{{\partial}W_{1}}=\frac{{\partial}R_{1}}{{\partial}P_{1}}\frac{{\partial}P_{1}}{{\partial}O_{3}}\frac{{\partial}O_{3}}{{\partial}W_{1}}=-2$ (1) To calculate the partial derivative of $R_{2}$ with respect to $W_{1}$, we backpropagate from $R_{2}$ to $W_{1}$: $\frac{{\partial}R_{2}}{{\partial}W_{1}}=\frac{{\partial}R_{2}}{{\partial}P_{2}}\frac{{\partial}P_{2}}{{\partial}W_{1}}=\frac{{\partial}R_{2}}{{\partial}P_{2}}\frac{{\partial}P_{2}}{{\partial}O_{5}}\frac{{\partial}O_{5}}{{\partial}O_{4}}\frac{{\partial}O_{4}}{{\partial}O_{3}}\frac{{\partial}O_{3}}{{\partial}W_{1}}=-4O_{4}$ (2) If the RNN was set up to propagate through two-time steps, the gradient for $W_{1}$ is $-4O_{4}+2$. Real RNNs are more complex and involve propagation through larger numbers of time steps, but all are optimized through a process similar to this. These derivatives are the basis for gradients in the optimization misfit function; using them the W are updated and iterations continue. ## 3 A Recurrent Neural Network formulation of EFWI Wave propagation can be simulated using suitably-designed RNNs (Sun et al.,, 2020; Richardson,, 2018; Hughes et al.,, 2019). Here we take the acoustic wave propagation approach of Sun et al., (2020) as a starting point, and formulate an RNN which simulates the propagation of an elastic wave through isotropic elastic medium. The underlying equations are the 2D velocity-stress form of the elastodynamic equations (Virieux,, 1986; Liu and Sen,, 2009): $\begin{aligned} &{\frac{\partial{{v}}_{x}}{\partial t}=\frac{1}{\rho}\left(\frac{\partial{\sigma}_{xx}}{\partial x}+\frac{\partial{\sigma}_{xz}}{\partial z}\right)}\\\ &{\frac{\partial{{v}}_{z}}{\partial t}=\frac{1}{\rho}\left(\frac{\partial{\sigma}_{xz}}{\partial x}+\frac{\partial{\sigma}_{zz}}{\partial z}\right)}\\\ &{\frac{\partial{\sigma}_{xx}}{\partial t}=({\lambda}+2{\mu})\frac{\partial{{v}}_{x}}{\partial x}+{\lambda}\frac{\partial{{v}}_{z}}{\partial z}}\\\ &{\frac{\partial{\sigma}_{zz}}{\partial t}=({\lambda}+2{\mu})\frac{\partial{{v}}_{z}}{\partial z}+{\lambda}\frac{\partial{{v}}_{x}}{\partial x}}\\\ &{\frac{\partial{\sigma}_{xz}}{\partial t}={\mu}\left(\frac{\partial{{v}}_{x}}{\partial z}+\frac{\partial{{v}}_{z}}{\partial x}\right)}\\\ \end{aligned}.,$ (3) where ${{v}}_{x}$ and ${{v}}_{z}$ are the $x$ and $z$ components of the particle velocity, ${{\sigma}_{xx}}$, ${{\sigma}_{zz}}$ and ${{\sigma}_{xz}}$ are three 2D components of the stress tensor. Discretized spatial distributions of the Lamé parameters $\lambda$ and $\mu$, and the density ${\rho}$, form the elastic model. Figure 3: The structure of each RNN. In this figure, $\partial_{x}{\sigma}_{xx}^{t}$, $\partial_{z}{\sigma}_{zz}^{t}$, $\partial_{x}{\sigma}_{xz}^{t}$, $\partial_{z}{\sigma}_{xz}^{t}$, $\partial_{x}{v}_{x}^{t+\frac{1}{2}}$, $\partial_{z}{v}_{x}^{t+\frac{1}{2}}$, $\partial_{x}{v}_{z}^{t+\frac{1}{2}}$, $\partial_{z}{v}_{z}^{t+\frac{1}{2}}$ are the internal variables. ${v}^{t-\frac{1}{2}}_{x}$, ${v}^{t-\frac{1}{2}}_{z}$, ${\sigma}_{xx}^{t}$, ${\sigma}_{zz}^{t}$, ${\sigma}_{xz}^{t}$, is communicated between the RNN cells. ${\lambda}$, ${\mu}$,$\rho$ are trainable parameters. Algorithm 1 Sequence of calculations in the RNN cell with PML boundary 1:Source: $s_{x}$, $s_{z}$; velocity and stress fields at the previous time step, parameters :$\lambda$, $\mu$, $\rho$; time step: $dt$. PML damping coefficients $d_{x}$. $d_{z}$. 2:Update velocity field at $t+\frac{1}{2}$ and stress fields at $t+1$ 3:${\sigma_{xx}^{t}}\leftarrow{\sigma_{xx}^{t}}+{s}_{x}$ 4:${\sigma_{zz}^{t}}\leftarrow{\sigma_{zz}^{t}}+{s}_{z}$ 5:$\partial_{x}{\sigma_{xx}^{t}}\leftarrow({\sigma_{xx}^{t}}*{\mathbf{k}}_{x_{1}})/\rho$ 6:$\partial_{z}{\sigma_{xz}^{t}}\leftarrow({\sigma_{xz}^{t}}*{\mathbf{k}}_{z_{2}})/\rho$ 7:$\partial_{x}{\sigma_{xz}^{t}}\leftarrow({\sigma_{xz}^{t}}*{\mathbf{k}}_{x_{2}})/\rho$ 8:$\partial_{z}{\sigma_{zz}^{t}}\leftarrow({\sigma_{zz}^{t}}*{\mathbf{k}}_{z_{1}})/\rho$ 9:${{v}_{x}^{t+\frac{1}{2}}}_{x}\leftarrow(1-dtd_{x}){{v}_{x}^{t-\frac{1}{2}}}_{x}+dt(\partial_{x}{\sigma_{xx}^{t}})$ 10:${{v}_{x}^{t+\frac{1}{2}}}_{z}\leftarrow(1-dtd_{z}){{v}_{x}^{t-\frac{1}{2}}}_{z}+dt(\partial_{z}{\sigma_{xz}^{t}})$ 11:${v}_{x}^{t+\frac{1}{2}}\leftarrow{{v}_{x}^{t+\frac{1}{2}}}_{x}+{{v}_{x}^{t+\frac{1}{2}}}_{z}$ 12:${{v}_{z}^{t+\frac{1}{2}}}_{x}\leftarrow(1-dtd_{x}){{v}_{z}^{t-\frac{1}{2}}}_{x}+dt(\partial_{x}{\sigma_{xz}^{t}})$ 13:${{v}_{z}^{t+\frac{1}{2}}}_{z}\leftarrow(1-dtd_{z}){{v}_{z}^{t-\frac{1}{2}}}_{z}+dt(\partial_{z}{\sigma_{zz}^{t}})$ 14:${v}_{z}^{t+\frac{1}{2}}\leftarrow{{v}_{x}^{t+\frac{1}{2}}}_{x}+{{v}_{x}^{t+\frac{1}{2}}}_{z}$ 15:$\partial_{x}{v}_{x}^{t+\frac{1}{2}}\leftarrow{v}_{x}^{t+\frac{1}{2}}*{\mathbf{k}}_{x_{2}}$ 16:$\partial_{z}{v}_{x}^{t+\frac{1}{2}}\leftarrow{v}_{x}^{t+\frac{1}{2}}*{\mathbf{k}}_{z_{1}}$ 17:$\partial_{x}{v}_{z}^{t+\frac{1}{2}}\leftarrow{v}_{z}^{t+\frac{1}{2}}*{\mathbf{k}}_{x_{1}}$ 18:$\partial_{z}{v}_{z}^{t+\frac{1}{2}}\leftarrow{v}_{z}^{t+\frac{1}{2}}*{\mathbf{k}}_{z_{2}}$ 19:${{\sigma_{xx}}^{t+1}}_{x}\leftarrow(1-dtdx){{\sigma_{xx}}^{t}}_{x}+dt({\lambda+2\mu})\partial_{x}{v}_{x}^{t+\frac{1}{2}}$ 20:${{\sigma_{zz}}^{t+1}}_{x}\leftarrow(1-dtdz){{\sigma_{zz}}^{t}}_{x}+dt(\lambda)\partial_{z}{v}_{z}^{t+\frac{1}{2}}$ 21:${{\sigma_{zz}}^{t+1}}\leftarrow{{\sigma_{xx}}^{t+1}}_{x}+{{\sigma_{xx}}^{t+1}}_{z}$ 22:${{\sigma_{zz}}^{t+1}}_{x}\leftarrow(1-dtdx){{\sigma_{xx}}^{t}}_{x}+dt({\lambda})\partial_{x}{v}_{x}^{t+\frac{1}{2}}$ 23:${{\sigma_{zz}}^{t+1}}_{z}\leftarrow(1-dtdz){{\sigma_{zz}}^{t}}_{z}+dt(\lambda+2\mu)\partial_{z}{v}_{z}^{t+\frac{1}{2}}$ 24:${{\sigma_{zz}}^{t+1}}\leftarrow{{\sigma_{zz}}^{t+1}}_{x}+{{\sigma_{zz}}^{t+1}}_{z}$ 25:${{\sigma_{xz}}^{t+1}}_{x}\leftarrow(1-dtdx){{\sigma_{xz}}^{t}}_{x}+dt({\mu})\partial_{z}{v}_{x}^{t+\frac{1}{2}}$ 26:${{\sigma_{xz}}^{t+1}}_{z}\leftarrow(1-dtdz){{\sigma_{xz}}^{t}}_{z}+dt(\mu)\partial_{x}{v}_{z}^{t+\frac{1}{2}}$ 27:${{\sigma_{xz}}^{t+1}}\leftarrow{{\sigma_{xz}}^{t+1}}_{x}+{{\sigma_{xz}}^{t+1}}_{z}$ In Figure 3 the structure of an RNN cell which produces a staggered-grid finite difference solution for the velocity and stress fields is illustrated. At each time step the discrete sources $s_{x}$ and $s_{z}$ act as inputs; the velocity and stress information, ${v}^{t-\frac{1}{2}}_{x}$, ${v}^{t-\frac{1}{2}}_{z}$, ${\sigma}_{xx}^{t}$, ${\sigma}_{zz}^{t}$, and ${\sigma}_{xz}^{t}$, is communicated between the RNN cells ; the partial derivative fields, $\partial_{x}{\sigma}_{xx}^{t}$, $\partial_{z}{\sigma}_{zz}^{t}$, $\partial_{x}{\sigma}_{xz}^{t}$, $\partial_{z}{\sigma}_{xz}^{t}$, $\partial_{x}{v}_{x}^{t+\frac{1}{2}}$, $\partial_{z}{v}_{x}^{t+\frac{1}{2}}$, $\partial_{x}{v}_{z}^{t+\frac{1}{2}}$, $\partial_{z}{v}_{z}^{t+\frac{1}{2}}$ are the internal variables in each RNN cell, which correspond to O in Figure 2; and, $\lambda$, $\mu$ and $\rho$ are included as trainable weights, which correspond to W in Figure 2. In Algorithm 1 pseudocode detailing these calculations within the RNN cell is provided. The $*$ symbol represents the machine learning image convolution operator. This image convolution is the process of adding each element of the image to its local neighbors, weighted by the image convolution kernel. We find that this image convolution operator is also capable of calculating space partial derivatives if the convolution kernel is designed according to the finite difference coefficients. Details about the image convolution operation can be found in Podlozhnyuk, (2007). $dx$, $dz$ are the grid intervals, and the image convolution kernels are: ${\mathbf{k}}_{x_{1}}={\mathbf{a}}/dx$, ${\mathbf{k}}_{x_{2}}={\mathbf{b}}/dx$, ${\mathbf{k}}_{z_{1}}={\mathbf{a}}^{T}/dz$, and ${\mathbf{k}}_{z_{2}}={\mathbf{b}}^{T}/dz$, where ${\mathbf{a}}=[0,1/24,-9/8,9/8,-1/24]$ and ${\mathbf{b}}=[1/24,-9/8,9/8,-1/24,0]$. ${\mathbf{a}}$ and ${\mathbf{b}}$ are 1$\times$5 dimension arrays. ${\mathbf{k}}_{x_{1}}$ and ${\mathbf{k}}_{x_{2}}$ are kernels, for the image convolution process, responsible for calculating the staggered grid space partial derivative in x direction. ${\mathbf{k}}_{z_{1}}$ and ${\mathbf{k}}_{z_{2}}$ are kernels, for the image convolution process, responsible for calculating the staggered grid space partial derivative in z direction, and that is also why the arrays, ${\mathbf{a}}$ and ${\mathbf{b}}$, are transposed in ${\mathbf{k}}_{z_{1}}$ and ${\mathbf{k}}_{z_{2}}$. Space partial derivative calculated in this way is, mathematically, the same with conventional staggered grid method (e.g.Virieux, (1986)). In this study, we achieve this process in an image convolution way. In algorithm 1, in order to implement the PML boundary, all the stress fields and the velocity fields need to be split into their x and z components. In algorithm 1, $d_{x}$ and $d_{z}$ are the PML damping coefficients in x direction and z direction. $d_{x}$ can be expressed as: $d_{x}(i)=d_{0x}(\frac{i}{n_{pmlx}})^{p}$ (4) ,where * represents either x or z direction. i is PML layer number starting from the effective calculation boundary. $n_{pmlx}$ is the PML layer number in $x$ direction. p is an integer and the value is from 1-4. $d_{0x}$ can be expressed as: $d_{0x}=log(\frac{1}{R})\frac{r{V_{s}}}{n_{pmlx}{\delta}_{x}}$ (5) ,where $R$ is a theoretically reflection coefficient, $r$ is a value ranging from 3-4. ${\delta}_{x}$ is the grid length in $x$ direction. $d_{z}$ can also be calculated in the same way. Figure 4: Velocity field forward and residual backward propagation under formatting of the RNN. The shot records formed at each time step correspond to the output of RNN cell P in Figure 2. Observed shot records correspond to labeled data (L in Figure 2). Residual shot record correspond to residual information (R in Figure 2). Back propagation starts from the residual shot record. The activity of the RNN network is illustrated in “unfolded” form in Figure 4. Above the unfolded network, horizontal velocity fields associated with a point source at the top left of a model are plotted at three times during propagation of the wave information through the network, the third being at $t_{max}$, the maximum receiving time. The wave field values are stored at positions selected to match multicomponent receivers; these form shot records as time evolves, which becomes the output data at each time. These shot records correspond to variables P in Figures 1 and 2, and the observed data. For FWI problem, the observed data is obtained from the true model. For real seismic data inversion, the observed data is obtained from field survey. In this RNN based elastic FWI the observed data is considered as the labeled data. The residuals are calculated at the last computational time as Figure 4 illustrated and this along with a selected norm defines the misfit function used to train the network. Algorithm 2 describes the training of the network, i.e., the process of elastic RNN FWI. Partial derivatives of the residual with respect to the trainable parameters (in this case $\lambda$, $\mu$ and $\rho$) are calculated through backpropagation using the automatic differential method, as set out in the previous section. After we have the gradients then we can use an optimization method and step length to update the trainable parameters and reduce the misfit and start another iteration. In step 4, RNN() is the network discussed above, whose output is the synthetic data; costFunc(), in step 5, is the misfit or loss function chosen to measure the difference between the synthetic data and observed data; loss.backward() begins the backpropagation within the computational graph and produces the gradients for each parameter, with which the current parameter model can be updated and another iteration started. Algorithm 2 Loop for elastic RNN FWI 1:Set trainable parameters: $\lambda$, $\mu$, ${\rho}$ in this test. 2:Set optimizers for parameters: $Optimizer_{1}$, $Optimizer_{2}$ and $Optimizer_{3}$ for ${\lambda}$, $\mu$ and ${\rho}$ respectively. 3:for iter $\in[1,maxiter]$ or not converge do 4: $D_{syn}$ = RNN($\lambda$, $\mu$, ${\rho}$): generate synthetic data 5: loss = costFunc($D_{syn}$, $D_{obs}$): calculate misfits 6: loss.backward(): Backpropagation and give gradients for the parameters 7: optimizers.step(): update parameters 8:end for The gradient calculated within the training process above essentially reproduces the adjoint-state calculations within FWI, as discussed by Sun et al., (2020). Formulated as a problem of RNN training, the gradient calculation occurs rapidly and in a manner suitable for cloud computational architectures; also, it allows the researcher to efficiently alter misfit function choices and parameterization in order to do high-level optimization. However, it involves the storage of the whole wavefield, and thus should be expected to have significant memory requirements. ## 4 Misfits with high order TV regulation Here we fist introduce the elastic RNN misfits based on $l_{2}$ norm with high order TV regularization: $\displaystyle{\mathbf{\Phi}}_{l2}^{TV}({\mathbf{m_{\lambda},{m_{\mu},m_{\rho}}}},{\alpha_{1}^{\lambda}},{\alpha_{1}^{\mu}},{\alpha_{1}^{\rho}},{\alpha_{2}^{\lambda}},{\alpha_{2}^{\mu}},{\alpha_{2}^{\rho}})=\frac{1}{2}\lVert\mathbf{D_{syn}}\mathbf{(m_{\lambda},m_{\mu},m_{\rho})}-\mathbf{D_{obs}}\rVert^{2}_{2}+$ (6) $\displaystyle{\alpha_{1}^{\lambda}}\mathbf{{\Theta}_{TV}}\mathbf{(m_{\lambda})}+{\alpha_{1}^{\mu}}\mathbf{{\Theta}_{TV}}\mathbf{(m_{\mu})}+{\alpha_{1}^{\rho}}\mathbf{{\Theta}_{{TV}}}\mathbf{(m_{\rho})}+$ $\displaystyle{\alpha_{2}^{\lambda}}\mathbf{\Upsilon_{TV}}\mathbf{(m_{\lambda})}+{\alpha_{2}^{\mu}}\mathbf{\Upsilon_{TV}}\mathbf{(m_{\mu})}+{\alpha_{2}^{\rho}}\mathbf{\Upsilon_{TV}}\mathbf{(m_{\rho})}$ ,where ${\alpha_{1}^{\lambda}}$, ${\alpha_{1}^{\mu}}$, ${\alpha_{1}^{\rho}}$, ${\alpha_{1}^{\lambda}}$, ${\alpha_{1}^{\mu}}$, ${\alpha_{1}^{\rho}}$, are vector of Lagrange multipliers, and $\mathbf{{\Theta}_{TV}}$, $\mathbf{{\Upsilon}_{TV}}$ represents first and second order TV regularization functions respectively. $\mathbf{D_{syn}}\mathbf{(m_{\lambda},m_{\mu},m_{\rho})}$ represents the synthetic data, which is the function of the model parameters, and in this equation they are $V_{P}$, $V_{S}$ and $\rho$. $\mathbf{{\Theta}_{TV}}$ and $\mathbf{{\Upsilon}_{TV}}$ represent functions for calculating the first and second order TV regulations for the models. The first order TV regulation term can be expressed as: $\displaystyle TV_{1}(\mathbf{(m)})=$ $\displaystyle\sum_{i=1}^{n-1}\sum_{j=1}^{m-1}|M_{i+1,j}-M_{i,j}|+\sum_{i=1}^{n-1}\sum_{j=1}^{m-1}|M_{i,j+1}-M_{i,j}|=$ (7) $\displaystyle\begin{pmatrix}{\nabla}_{x},{\nabla}_{z}\end{pmatrix}\begin{pmatrix}\mathbf{m},\\\ \mathbf{m}\end{pmatrix}=\begin{pmatrix}\mathbf{\mathcal{L}_{x}},\mathbf{\mathcal{L}_{z}},\end{pmatrix}\begin{pmatrix}\mathbf{m},\\\ \mathbf{m}\end{pmatrix}=\mathbf{{\Theta}_{TV}}\mathbf{(m)}$ The second order TV regulation term can be expressed as: $\displaystyle TV_{2}(\mathbf{(m)})=$ $\displaystyle\sum_{i=1}^{n-1}\sum_{j=1}^{m-1}|M_{i+1,j}-2M_{i,j}+M_{i-1,j}|+\sum_{i=1}^{n-1}\sum_{j=1}^{m-1}|M_{i,j+1}-2M_{i,j}+M_{i,j-1}|$ (8) $\displaystyle=$ $\displaystyle\begin{pmatrix}{\nabla}_{xx},{\nabla}_{zz}\end{pmatrix}\begin{pmatrix}\mathbf{m},\\\ \mathbf{m}\end{pmatrix}=\begin{pmatrix}\mathbf{\mathcal{K}_{xx}},\mathbf{\mathcal{K}_{zz}},\end{pmatrix}\begin{pmatrix}\mathbf{m},\\\ \mathbf{m}\end{pmatrix}=\mathbf{{\Upsilon}_{TV}}\mathbf{(m)}$ The derivative of ${\mathbf{\Phi}}_{l2}^{TV}$ for each parameter ,which is the gradient for $V_{P}$, $V_{S}$ and $\rho$ based on the $l_{2}^{TV}$ norm can be expressed as: $\displaystyle\begin{pmatrix}\frac{{\partial}{\mathbf{\Phi}}_{l2}^{TV}}{{\partial}{\mathbf{m_{\lambda}}}}\\\ \frac{{\partial}{\mathbf{\Phi}}_{l2}^{TV}}{{\partial}{\mathbf{m_{\mu}}}}\\\ \frac{{\partial}{\mathbf{\Phi}}_{l2}^{TV}}{{\partial}{\mathbf{m_{\rho}}}}\\\ \end{pmatrix}=\begin{pmatrix}\mathbf{G_{{l2}_{\lambda}}}\\\ \mathbf{G_{{l2}_{\mu}}}\\\ \mathbf{G_{{l2}_{\rho}}}\\\ \end{pmatrix}+\begin{pmatrix}\mathbf{R_{\lambda}}\\\ \mathbf{R_{\mu}}\\\ \mathbf{R_{\rho}}\\\ \end{pmatrix}$ (9) , where $\mathbf{G_{{l2}_{\lambda}}}$,$\mathbf{G_{{l2}_{\mu}}}$,$\mathbf{G_{{l2}_{\rho}}}$ are the gradient for $\lambda$, $\mu$, ${\rho}$. $\mathbf{R_{\lambda}}$,$\mathbf{R_{\mu}}$,$\mathbf{R_{\rho}}$ are the regulation terms and the mathematical expressions for these regulation terms are: $\displaystyle\begin{pmatrix}\mathbf{R_{\lambda}}\\\ \mathbf{R_{\mu}}\\\ \mathbf{R_{\rho}}\\\ \end{pmatrix}=\begin{pmatrix}&{\alpha_{1}^{\lambda}}\mathbf{\mathcal{L}^{T}_{x}}\mathbf{Q_{x_{\lambda}}}\hskip 7.22743pt&{\alpha_{1}^{\lambda}}\mathbf{\mathcal{L}^{T}_{z}}\mathbf{Q_{z_{\lambda}}},\hskip 7.22743pt&{\alpha_{2}^{\lambda}}\mathbf{\mathcal{K}^{T}_{xx}}\mathbf{Q_{{xx}_{\lambda}}}\hskip 7.22743pt&{\alpha_{2}^{\lambda}}\mathbf{\mathcal{K}^{T}_{zz}}\mathbf{Q_{{zz}_{\lambda}}}\\\ &{\alpha_{1}^{\mu}}\mathbf{\mathcal{L}^{T}_{x}}\mathbf{Q_{{x}_{\mu}}}\hskip 7.22743pt&{\alpha_{1}^{\mu}}\mathbf{\mathcal{L}^{T}_{z}}\mathbf{Q_{{z}_{\mu}}}\hskip 7.22743pt&{\alpha_{2}^{\mu}}\mathbf{\mathcal{K}^{T}_{xx}}\mathbf{Q_{{xx}_{\mu}}}\hskip 7.22743pt&{\alpha_{2}^{\mu}}\mathbf{\mathcal{K}^{T}_{zz}}\mathbf{Q_{{zz}_{\mu}}}\\\ &{\alpha_{1}^{\rho}}\mathbf{\mathcal{L}^{T}_{x}}\mathbf{Q_{{x}_{\rho}}}\hskip 7.22743pt&{\alpha_{1}^{\rho}}\mathbf{\mathcal{L}^{T}_{z}}\mathbf{Q_{{z}_{\rho}}}\hskip 7.22743pt&{\alpha_{2}^{\rho}}\mathbf{\mathcal{K}^{T}_{xx}}\mathbf{Q_{{xx}_{\rho}}}\hskip 7.22743pt&{\alpha_{2}^{\rho}}\mathbf{\mathcal{K}^{T}_{zz}}\mathbf{Q_{{zz}_{\rho}}}\end{pmatrix}\begin{pmatrix}\mathbf{\mathcal{L}_{x}}\\\ \mathbf{\mathcal{L}_{z}}\\\ \mathbf{\mathcal{K}_{xx}}\\\ \mathbf{\mathcal{K}_{zz}}\end{pmatrix}\cdot\begin{pmatrix}\mathbf{m_{\lambda}}\\\ \mathbf{m_{\mu}}\\\ \mathbf{m_{\rho}}\end{pmatrix}$ (10) $\begin{pmatrix}\mathbf{q_{{x}_{\lambda}}}&\mathbf{q_{{x}_{\mu}}}&\mathbf{q_{{x}_{\rho}}}\\\ \mathbf{q_{{z}_{\lambda}}}&\mathbf{q_{{z}_{\mu}}}&\mathbf{q_{{z}_{\rho}}}\\\ \mathbf{q_{{xx}_{\lambda}}}&\mathbf{q_{{xx}_{\mu}}}&\mathbf{q_{{xx}_{\rho}}}\\\ \mathbf{q_{{zz}_{\lambda}}}&\mathbf{q_{{zz}_{\mu}}}&\mathbf{q_{{zz}_{\rho}}}\\\ \end{pmatrix}=\begin{pmatrix}\mathbf{\mathcal{L}_{x}}\\\ \mathbf{\mathcal{L}_{z}}\\\ \mathbf{\mathcal{K}_{xx}}\\\ \mathbf{\mathcal{K}_{zz}}\\\ \end{pmatrix}\begin{pmatrix}\mathbf{m_{\lambda}},\mathbf{m_{\mu}},\mathbf{m_{\rho}}\end{pmatrix}$ (11) $\begin{pmatrix}\mathrm{Q_{{x}_{\lambda}}}&\mathrm{Q_{{x}_{\mu}}}&\mathrm{Q_{{x}_{\rho}}}\\\ \mathrm{Q_{{z}_{\lambda}}}&\mathrm{Q_{{z}_{\mu}}}&\mathrm{Q_{{z}_{\rho}}}\\\ \mathrm{Q_{{xx}_{\lambda}}}&\mathrm{Q_{{xx}_{\mu}}}&\mathrm{Q_{{xx}_{\rho}}}\\\ \mathrm{Q_{{zz}_{\lambda}}}&\mathrm{Q_{{zz}_{\mu}}}&\mathrm{Q_{{zz}_{\rho}}}\\\ \end{pmatrix}=\begin{pmatrix}\mathrm{\frac{1}{|q_{{x}_{\lambda}}|}}&\mathrm{\frac{1}{|q_{{x}_{\mu}}|}}&\mathrm{\frac{1}{|q_{{x}_{\rho}}|}}\\\ \mathrm{\frac{1}{|q_{{z}_{\lambda}}|}}&\mathrm{\frac{1}{|q_{{z}_{\mu}}|}}&\mathrm{\frac{1}{|q_{{z}_{\rho}}|}}\\\ \mathrm{\frac{1}{|q_{{xx}_{\lambda}}|}}&\mathrm{\frac{1}{|q_{{xx}_{\mu}}|}}&\mathrm{\frac{1}{|q_{{xx}_{\rho}}|}}\\\ \mathrm{\frac{1}{|q_{{zz}_{\lambda}}|}}&\mathrm{\frac{1}{|q_{{zz}_{\mu}}|}}&\mathrm{\frac{1}{|q_{{zz}_{\rho}}|}}\end{pmatrix}$ (12) $\mathbf{T}$ means the transpose of the matrix, $\cdot$ meas dot product. $\mathbf{q_{{x}_{\mathbf{\lambda}}}}$ represent the first order TV regularization vector in x direction for parameter ${\lambda}$. $\mathrm{q_{{x}_{{\lambda}}}}$ represent the values in vector $\mathbf{q_{{x}_{\lambda}}}$. $\mathrm{Q_{{x}_{\lambda}}}$ is the absolute inverse of $\mathrm{q_{{x}_{\lambda}}}$. $\mathrm{Q_{{x}_{\lambda}}}$ are elements in vector $\mathbf{Q_{{x}_{\lambda}}}$. $\mathbf{q_{{xx}_{\mathbf{\lambda}}}}$ represent the second order TV regularization vector in x direction for parameter ${\lambda}$. $\mathrm{q_{{xx}_{{\lambda}}}}$ represent the values in vector $\mathbf{q_{{xx}_{\lambda}}}$. $\mathrm{Q_{{xx}_{\lambda}}}$ is the absolute inverse of $\mathrm{q_{{xx}_{\lambda}}}$. $\mathrm{Q_{{xx}_{\lambda}}}$ are elements in vector $\mathbf{Q_{{xx}_{\lambda}}}$. Other values in equations (9) and (10) can be also deduced like this. $\mathbf{\mathcal{L}_{x}}$, $\mathbf{\mathcal{L}_{z}}$ are the first order differential vector to give the first order total variations in x and z directions respectively. $\mathbf{\mathcal{K}_{xx}}$, $\mathbf{\mathcal{K}_{zz}}$ are the second order differential vector to give the second order total variations in x and z directions respectively. If we were to use $l1$ norm objective function with TV regulation. The objective function can be written as: $\displaystyle{\mathbf{\Phi}}_{l1}^{TV}({\mathbf{m_{\lambda},{m_{\mu},m_{\rho}}}},{\alpha_{1}^{\lambda}},{\alpha_{1}^{\mu}},{\alpha_{1}^{\rho}},{\alpha_{2}^{\lambda}},{\alpha_{2}^{\mu}},{\alpha_{2}^{\rho}})=\lVert\mathbf{D_{syn}}\mathbf{(m_{\lambda},m_{\mu},m_{\rho})}-\mathbf{D_{obs}}\rVert+$ (13) $\displaystyle{\alpha_{1}^{\lambda}}\mathbf{{\Theta}_{TV}}\mathbf{(m_{\lambda})}+{\alpha_{1}^{\mu}}\mathbf{{\Theta}_{TV}}\mathbf{(m_{\mu})}+{\alpha_{1}^{\rho}}\mathbf{{\Theta}_{{TV}}}\mathbf{(m_{\rho})}+$ $\displaystyle{\alpha_{2}^{\lambda}}\mathbf{\Upsilon_{TV}}\mathbf{(m_{\lambda})}+{\alpha_{2}^{\mu}}\mathbf{\Upsilon_{TV}}\mathbf{(m_{\mu})}+{\alpha_{2}^{\rho}}\mathbf{\Upsilon_{TV}}\mathbf{(m_{\rho})}$ The gradient for each parameter based on l1 norm: $\displaystyle\begin{pmatrix}\frac{{\partial}{\mathbf{\Phi}}_{l1}^{TV}}{{\partial}{\mathbf{m_{\lambda}}}}\\\ \frac{{\partial}{\mathbf{\Phi}}_{l1}^{TV}}{{\partial}{\mathbf{m_{\mu}}}}\\\ \frac{{\partial}{\mathbf{\Phi}}_{l1}^{TV}}{{\partial}{\mathbf{m_{\rho}}}}\\\ \end{pmatrix}=\begin{pmatrix}\mathbf{G_{{l1}_{\lambda}}}\\\ \mathbf{G_{{l1}_{\mu}}}\\\ \mathbf{G_{{l1}_{\rho}}}\\\ \end{pmatrix}+\begin{pmatrix}\mathbf{R_{\lambda}}\\\ \mathbf{R_{\mu}}\\\ \mathbf{R_{\rho}}\\\ \end{pmatrix}$ (14) , where $\mathbf{G_{{l1}_{\lambda}}}$,$\mathbf{G_{{l1}_{\mu}}}$,$\mathbf{G_{{l1}_{\rho}}}$ are the gradient for $\lambda$, $\mu$, ${\rho}$ using $l_{1}$ norm as misfit function. The gradient calculation in $l_{1}$ norm is using a differenta adjoint source (Pyun et al., (2009) Brossier et al., (2010)). The adjoint source for the adjoint fields for $l_{1}$ norm is , In the case of real arithmetic numbers, the term $\frac{{\Delta\mathbf{d}}}{|\Delta\mathbf{d}|}$ corresponds to the function $sign$. In this study, we did not meet conditions when $\Delta\mathbf{d}=0$. The detail gradient expression using the adjoint state method for parameters $\lambda$, $\mu$ and $\rho$ based on $l_{2}$ and $l_{1}$ norm can be expressed as: $\begin{array}[]{l}\mathbf{G_{{l2}_{{\lambda}}}}=-\sum_{\mathbf{x_{s}}}\sum_{\mathbf{x_{g}}}\int_{0}^{T}\\\ \left(\left(\partial_{x}\tilde{u}_{x}\left(\mathbf{r},\mathbf{r}_{s},t\right)+\partial_{z}\tilde{u}_{z}\left(\mathbf{r},\mathbf{r}_{s},t\right)\right)\left(\partial_{x}\tilde{u}_{x}^{*_{l2}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)+\partial_{z}\tilde{u}_{z}^{*_{l2}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)\right)\right)\end{array}$ (15) $\begin{array}[]{l}\mathbf{G_{{l2}_{\mu}}}=-\sum_{\mathbf{x_{s}}}\sum_{\mathbf{x_{g}}}\int_{0}^{T}\\\ \left(\left(\partial_{z}\tilde{u}_{x}\left(\mathbf{r},\mathbf{r}_{s},t\right)+\partial_{x}\tilde{u}_{z}\left(\mathbf{r},\mathbf{r}_{s},t\right)\right)\left(\partial_{z}\tilde{u}_{x}^{*_{l2}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)+\partial_{x}\tilde{u}_{z}^{*_{l2}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)\right)\right)\\\ -2\left(\left(\partial_{x}\tilde{u}_{x}\left(\mathbf{r},\mathbf{r}_{s},t\right)\partial_{x}\tilde{u}_{x}^{*_{l2}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)+\partial_{z}\tilde{u}_{z}\left(\mathbf{r},\mathbf{r}_{s},t\right)\partial_{z}\tilde{u}_{z}^{*_{l2}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)\right)\right)\end{array}$ (16) $\mathbf{G_{{l2}_{\rho}}}=\sum_{\mathbf{x_{s}}}\sum_{\mathbf{x_{g}}}\int_{0}^{T}\\\ \left(\left(\partial_{t}\tilde{u}_{x}\left(\mathbf{r},\mathbf{r}_{s},t\right)\partial_{t}\tilde{u}_{x}^{*_{l2}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)+\partial_{t}\tilde{u}_{z}\left(\mathbf{r},\mathbf{r}_{s},t\right)\partial_{t}\tilde{u}_{z}^{*_{l2}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)\right)\right)$ (17) $\begin{array}[]{l}\mathbf{G_{{l1}_{{\lambda}}}}=-\sum_{\mathbf{x_{s}}}\sum_{\mathbf{x_{g}}}\int_{0}^{T}\\\ \left(\left(\partial_{x}\tilde{u}_{x}\left(\mathbf{r},\mathbf{r}_{s},t\right)+\partial_{z}\tilde{u}_{z}\left(\mathbf{r},\mathbf{r}_{s},t\right)\right)\left(\partial_{x}\tilde{u}_{x}^{*_{l1}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)+\partial_{z}\tilde{u}_{z}^{*_{l1}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)\right)\right)\end{array}$ (18) $\begin{array}[]{l}\mathbf{G_{{l1}_{\mu}}}=-\sum_{\mathbf{x_{s}}}\sum_{\mathbf{x_{g}}}\int_{0}^{T}\\\ \left(\left(\partial_{z}\tilde{u}_{x}\left(\mathbf{r},\mathbf{r}_{s},t\right)+\partial_{x}\tilde{u}_{z}\left(\mathbf{r},\mathbf{r}_{s},t\right)\right)\left(\partial_{z}\tilde{u}_{x}^{*_{l1}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)+\partial_{x}\tilde{u}_{z}^{*_{l1}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)\right)\right)\\\ -2\left(\left(\partial_{x}\tilde{u}_{x}\left(\mathbf{r},\mathbf{r}_{s},t\right)\partial_{x}\tilde{u}_{x}^{*_{l1}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)+\partial_{z}\tilde{u}_{z}\left(\mathbf{r},\mathbf{r}_{s},t\right)\partial_{z}\tilde{u}_{z}^{*_{l1}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)\right)\right)\end{array}$ (19) $\mathbf{G_{{l1}_{\rho}}}=\\\ \sum_{\mathbf{x_{s}}}\sum_{\mathbf{x_{g}}}\int_{0}^{T}\left(\left(\partial_{t}\tilde{u}_{x}\left(\mathbf{r},\mathbf{r}_{s},t\right)\partial_{t}\tilde{u}_{x}^{*_{l1}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)+\partial_{t}\tilde{u}_{z}\left(\mathbf{r},\mathbf{r}_{s},t\right)\partial_{t}\tilde{u}_{z}^{*_{l1}}\left(\mathbf{r},\mathbf{r}_{g},T-t\right)\right)\right)$ (20) $\mathbf{G_{{l2}_{\lambda}}}$, $\mathbf{G_{{l2}_{\mu}}}$, $\mathbf{G_{{l2}_{\rho}}}$, are gradients for $\lambda$, $\mu$ and $\rho$ using $l_{2}$ norm as misfit respectively. $\mathbf{G_{{l1}_{\lambda}}}$, $\mathbf{G_{{l1}_{\mu}}}$, $\mathbf{G_{{l1}_{\rho}}}$, are gradients for $\lambda$, $\mu$ and $\rho$ using $l_{1}$ norm as misfit respectively. $\tilde{u_{x}}^{*_{l1}}$ and $\tilde{u_{z}}^{*_{l1}}$ are the adjoint wavefields generated by the l1 norm adjoint source, $\tilde{u_{x}}^{*_{l2}}$ and $\tilde{u_{z}}^{*_{l2}}$ are the adjoint wavefields generated by the l2 norm adjoint source. $T$ is the total receiving time for the shot records. $\mathbf{r_{s}}$, $\mathbf{r_{g}}$ represent the source and receivers locations respectively. $\mathbf{r}$ represent the model perturbation locations for $\lambda$, $\mu$ and $\rho$ model. Figure 5 shows the gradient calculated using the adjoint state method and the Automatic Difference method. Figure 5 (a), (b), (c) are the normalized ${\lambda}$, ${\mu}$ and $\rho$ gradients calculated by using the adjont state method. Figure 5 (d), (e), (f) are the normalized ${\lambda}$, ${\mu}$ and $\rho$ gradients calculated by using the Automatic Differential method. The gradients calculated by using the Automatic Difference method, contains more information about the model, for instance the lower part of the model, indicating that they can better reconstruct the mode. Now we rewrite the misfit function as: $\displaystyle{\mathbf{\Phi}}^{TV}=\mathbf{J_{D}}+\mathbf{J_{r1}}+\mathbf{J_{r2}}$ (21) ,where $\mathbf{J_{D}}$ represents the any kind of norm misfit between observed data and synthetic data. $\mathbf{J_{r1}}={\alpha_{1}^{\lambda}}\mathbf{{\Theta}_{TV}}\mathbf{(m_{\lambda})}+{\alpha_{1}^{\mu}}\mathbf{{\Theta}_{TV}}\mathbf{(m_{\mu})}+{\alpha_{1}^{\rho}}\mathbf{{\Theta}_{{TV}}}\mathbf{(m_{\rho})}$. $\mathbf{J_{r2}}={\alpha_{2}^{\lambda}}\mathbf{\Upsilon_{TV}}\mathbf{(m_{\lambda})}+{\alpha_{2}^{\mu}}\mathbf{\Upsilon_{TV}}\mathbf{(m_{\mu})}+{\alpha_{2}^{\rho}}\mathbf{\Upsilon_{TV}}\mathbf{(m_{\rho})}$. The value $\alpha_{1}^{\lambda}$, ${\alpha_{1}^{\mu}}$, ${\alpha_{1}^{\rho}}$, $\alpha_{2}^{\lambda}$, ${\alpha_{2}^{\mu}}$, ${\alpha_{2}^{\rho}}$. are chosen according to the following formula (Guitton et al., (2012)). $T=\frac{\mathbf{J_{D}}}{\mathbf{J_{r1}}+\mathbf{J_{r2}}}$ (22) We should control the values for $\alpha_{1}^{\lambda}$, ${\alpha_{1}^{\mu}}$, ${\alpha_{1}^{\rho}}$, $\alpha_{2}^{\lambda}$, ${\alpha_{2}^{\mu}}$, ${\alpha_{2}^{\rho}}$ and keep value T between 1 and 10. T should be relatively large when, noise occurs in the data (Xiang and Zhang, (2016)). Figure 5: (a) ${\lambda}$ gradient given by the adjoint state method. (b) $\mu$ gradient based by the adjoint state method. (c) $\rho$ gradient based by the adjoint state method. (d) ${\lambda}$ gradient given by the AD method. (e) $\mu$ gradient based by the AD method. (f) $\rho$ gradient based by the AD method. ### 4.1 Parameterization testing By modifying the RNN cells, and changing the trainable parameters, we can examine the influence of parameterization on waveform inversion within the deep learning formulation. Three sets of parameter classes are considered: the velocity parameterization (D-V model), involving P-wave velocity, S-wave velocity, and density; the modulus parameterization (D-M model), involving the Lamé parameters ${\lambda}$ and ${\mu}$, and density; and, the stiffness matrix model (D-S model), involving ${C_{11}}$, $C_{44}$ and density ${\rho}$. In these tests, the size of each model is 40${\times}90$. 7 source points are evenly distributed across the surface of the model; the source is a Ricker wavelet with a dominant frequency of $30Hz$. The grid length of the model is $dx=dz=4$m. In Figure 6 (a)-(b) are true and initial $V_{P}$ model, (c)-(d) are the true and initial $V_{S}$ model, (e)-(f) are the true and initial $\rho$ model we use in this test. In Figures 7 are the inversion results using the D-V parameterization. Figure 7 (a)-(d) are the inversion results for $V_{P}$ generated by $l_{2}$, $l_{2}^{TV}$, $l_{1}$ and $l_{1}^{TV}$ norm. Figure 7 (e)-(h) are the inversion results for $V_{S}$ generated by $l_{2}$, $l_{2}^{TV}$, $l_{1}$ and $l_{1}^{TV}$ norm. Figure 7 (i)-(l) are the inversion results for $\rho$ generated by $l_{2}$, $l_{2}^{TV}$, $l_{1}$ and $l_{1}^{TV}$ norm. In Figure 7 (a), (e) and (i) we can see the cross talk between different parameters as the back arrows indicate, In Figure 7 (b), (c), (d) we can see that by changing the misfits and adding high order TV regulations, the cross talk between $V_{P}$ and $\rho$ has been reduced. In Figure 7 (h) and (l) we can see that by using $l_{1}^{TV}$ norm, the cross talk between density and $V_{S}$ has been mitigated, while in (j) and (k) we still see the cross talk between $V_{S}$ and density. From Figures 7 we can see that $l_{1}$ norm with high order TV regulation can help to mitigate the cross talk problem. Figure 6: (a) true Vp model, (b) initial Vs model, (c) true Vs model, (d) initial vs model, (e) true ${\rho}$ mode, (f) initial ${\rho}$ model. Figure 7: D-V parameterization inversion results. (a)-(d), $V_{P}$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}$ norm, $l_{1}^{TV}$ norm inversion results respectively, (e)-(h), $V_{S}$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l1$ norm $l_{1}^{TV}$ norm, inversion results respectively, (i)-(l), $\rho$ $l_{2}$ norm $l_{2}^{TV}$ norm, $l_{1}$ norm, $l_{1}^{TV}$ norm inversion results respectively. Next the modulus parameterization is examined, in which we seek to recover ${\lambda}$, ${\mu}$ and ${\rho}$ models. This occurs through a straightforward modification of the RNN cell, and again a change in the trainable parameters from velocities to moduli. Figure 8 (a)-(d) are the inversion results for $\lambda$ generated by $l_{2}$, $l_{2}^{TV}$, $l_{1}$ and $l_{1}^{TV}$ norm. Figure 8 (e)-(h) are the inversion results for $\mu$ generated by $l_{2}$, $l_{2}^{TV}$, $l_{1}$ and $l_{1}^{TV}$ norm. Figure 8 (i)-(l) are the inversion results for $\rho$ generated by $l_{2}$, $l_{2}^{TV}$, $l_{1}$ and $l_{1}^{TV}$ norm. In Figure 8 (b) and (d) we can see that by using the high order TV regulations in $l_{2}$ and $l_{1}$ norm the cross talk between density and $\lambda$ has been mitigated. Figure (j) shows that In this parameterization by using the high order TV regulation on $l_{2}$ norm can provide better inversion results for density as well. Figure 8: D-M parameterization inversion results. (a)-(d), ${\lambda}$ $l_{2}$, norm, $l_{2}^{TV}$ norm, $l_{1}$, norm $l_{1}^{TV}$ norm inversion results respectively, (e)-(h), $\mu$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l1$ norm, $l_{1}^{TV}$ norm inversion results respectively, (i)-(l), $\rho$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}$ norm, $l_{1}^{TV}$ norm inversion results respectively. Inversion results for models in the S-D parameterization are plotted in Figure 9, generated using the change of variables ${C_{11}}=V_{P}^{2}{\rho}$ and ${C_{44}}=V_{S}^{2}{\rho}$. Figure 9 (a)-(d) are the inversion results for $\lambda$ generated by $l_{2}$, $l_{2}^{TV}$, $l_{1}$ and $l_{1}^{TV}$ norm. Figure 9 (e)-(h) are the inversion results for $\mu$ generated by $l_{2}$, $l_{2}^{TV}$, $l_{1}$ and $l_{1}^{TV}$ norm. Figure 9 (i)-(l) are the inversion results for $\rho$ generated by $l_{2}$, $l_{2}^{TV}$, $l_{1}$ and $l_{1}^{TV}$ norm. In Figure 9 we can still see the cross talk between $c44$ and $c11$ as the black arrows pointing out. However this cross talk has been mitigated by in figure (d), by using the $l_{1}^{TV}$ norm misfit. The inversion results above shows that the RNN based high order TV regulation FWi based on the $l_{2}$ and $l_{1}$ norm has the ability to mitigate cross talk problem with only gradient based methods. The RNN based $l_{2}^{TV}$ RNN FWI in D-M parameterization and $l_{1}^{TV}$ RNN FWI in D-S parameterization provide better inversion results than other inversion tests. Figure 9: D-S parameterization inversion results. (a)-(d), $c11$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}$ norm, $l_{1}^{TV}$ norm inversion results respectively, (e)-(h), $c44$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l1$ norm $l_{1}^{TV}$ norm, inversion results respectively, (i)-(l), $\rho$ $l_{2}$ norm $l_{2}^{TV}$ norm, $l_{1}$ norm, $l_{1}^{TV}$ norm inversion results respectively. Figure 10: (a) true $V_{P}$ model, (b) initial $V_{P}$ model, (c) true $V_{S}$ model (d) initial Vs model, (e) ${\rho}$ true (f) initial $\rho$ Figure 11: (a) true $V_{P}$ model, (b) initial $V_{P}$ model, (c) true $V_{S}$ model (d) initial $V_{S}$ model, (e) ${\rho}$ true (f) initial $\rho$ Figure 12: M-D parameterization inversion results. (a)-(c), $\lambda$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm inversion results respectively, (c)-(f), $\mu$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm, inversion results respectively, (g)-(i), $\rho$ $l_{2}$ norm $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm inversion results respectively Figure 13: Profiles through the recovered elastic models. (a) Vertical $\lambda$ profiles; (b) vertical $\mu$ profiles; (c) vertical ${\rho}$ profiles Next, we will verify the proposed methods on the over-thrust model. Figure 10 (a), (c), (e) demonstrate the true models for $V_{P}$, $V_{S}$, and density $\rho$ model, and (b), (d) and (h) are the initial models for $V_{P}$, $V_{S}$, and density $\rho$ respectively. The size of the model is 121 $\times$ 240\. The grid length of the model is 10m. 12 shots are evenly distributed on the surface of the model and every grid point has a receiver. The source of the wavelet is Ricker’s wavelet with main frequency 20Hz. Figure Figure 11 shows the inversion results by using the conventional FWI. Figure Figure 11 (a) is the inversion for $V_{P}$, Figure Figure 11 (b) is the inversion for $V_{S}$, Figure Figure 11 (c) is the inversion for $\rho$. From figure 11 we can see that the overall inversion resolution by using the conventional FWI is poor. Figure 12 shows the inversion results by using D-M parameterization. Figure 12 (a)-(c) are ${\lambda}$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm inversion results respectively, (d)-(f) are $\mu$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm, inversion results respectively, (g)-(i) are $\rho$ $l_{2}$ norm $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm inversion results respectively. In this parameterization we get unstable inversion results for parameter $\lambda$. However, in D-M parameterization Figure 13 (f), the small half arc structure at around 1000m of the model has been recovered in , as the black arrows indicate. Figure 13 shows the profiles through the recovered elastic modules at 1000m of the models based on D-M parameterization. In Figure 13, the black lines are the true values, the yellow lines are the initial values, red lines are the inversion results for $l_{2}$ norm, green lines are the inversion results for $l_{2}^{TV}$ norm and blue lines are the inversion results for $l_{1}^{TV}$ norm. Compared with the true lines we can also see that , Figure 13 (b) and Figure 13 (c), $l_{1}^{TV}$ norm inversion results are more close to the true values. Figure 14: D-V parameterization inversion results. (a)-(d), $V_{P}$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm inversion results respectively, (e)-(h), $V_{S}$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm, inversion results respectively, (i)-(l), $\rho$ $l_{2}$ norm $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm inversion results respectively. Figure 15: Profiles through the recovered elastic models. (a) Vertical $V_{P}$ profiles; (b) vertical $V_{S}$ profiles; (c) vertical ${\rho}$ profiles. Figure 16: D-V parameterization inversion results . (a) Vp model misfit using $l_{2}$,$l_{2}^{TV}$ and $l_{1}^{TV}$ norm.(b) Vs model misfit using $l_{2}$,$l_{2}^{TV}$ and $l_{1}^{TV}$ norm.(c) $\rho$ model misfit using $l_{2}$,$l_{2}^{TV}$ and $l_{1}^{TV}$ norm Figure 14 shows the inversion results by using D-V parameterization. In this parameterization all the three parameters are stable. Figure 14 (a)-(c) are, $V_{P}$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm inversion results respectively, (d)-(f) are $V_{S}$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm, inversion results respectively, (g)-(i) are $\rho$ $l_{2}$ norm $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm inversion results respectively. Compared with the true models ,we can see that the inversion results generated by using the $l_{2}$ norm are less robust compared with other misfits. $l_{1}$ norm with high order TV regulation can provide more accurate inversion results. This can also be seen from Figure 14, which is the profiles through the recovered elastic modules at 1000m of the models. In Figure 14, the black lines are the true velocities and density, the yellow lines are the initial values, red lines are the inversion results for $l_{2}$ norm, green lines are the inversion results for $l_{2}^{TV}$ norm and blue lines are the inversion results for $l_{1}^{TV}$ norm. Figure 15 (a) shows the results for $V_{P}$. Figure 15 (b) shows the results for $V_{S}$. Figure 15 (c) shows the results for $\rho$. In all the three figures we can see that blue lines are closer to the true values compared with other lines, which means that the $l_{1}^{TV}$ norm can provide us with more accurate inversion results. Figure 16 shows the D-V parameterization inversion model misfits in each iteration. The red lines are the inversion using $l_{2}$ norm. The green lines are the inversion using $l_{2}^{TV}$ norm. The red lines are the inversion using $l_{1}^{TV}$ norm. Figure 16 shows that we can get higher accuracy inversion results by using the $l_{1}^{TV}$ norm with fewer iterations. Figure 17 shows the inversion results by using D-S parameterization. Figure 17 (a)-(c), are ${c11}$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm inversion results respectively, (d)-(f) are $c44$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm, inversion results respectively, (g)-(i) are $\rho$ $l_{2}$ norm $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm inversion results respectively. In this parameterization the small half arc structure at around 1000m of the model is clearly resolved in this parameterization by using the $l_{1}^{TV}$ norm. Compared with other misfits , again the $l_{1}^{TV}$ norm provide us the best resolution for the model. Figure 13 shows the profiles through the recovered elastic modules at 1000m of the models based on D-M parameterization.. Red lines, green lines and blue lines are $l_{2}$, $l_{2}^{TV}$, and $l_{1}^{TV}$ norm inversion results respectively. We can also see that in D-S parameteriation. $l_{1}^{TV}$ provide us with inversion results that is more close the to true values. Figure 18 shows how model misfits in D-M parameterization changes in each iteration. The blue line, the $l_{1}^{TV}$ norm inversion, has the fastest model misfit decline rate. We can conclude in this parameterization $l_{1}^{TV}$ can generate more accuracy inversion results with fewer iterations. Figure 17: S-D parameterization inversion results. (a)-(c), $c11$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm inversion results respectively, (c)-(f), $c44$ $l_{2}$ norm, $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm, inversion results respectively, (g)-(i), $\rho$ $l_{2}$ norm $l_{2}^{TV}$ norm, $l_{1}^{TV}$ norm inversion results respectively. Figure 18: Profiles through the recovered elastic models based on . (a) Vertical $c11$ profiles; (b) vertical $c44$ profiles; (c) vertical ${\rho}$ profiles Figure 19: D-S parameterization inversion results . (a) $c11$ model misfit using $l_{2}$,$l_{2}^{TV}$ and $l_{1}^{TV}$ norm.(b) $c44$ model misfit using $l_{2}$,$l_{2}^{TV}$ and $l_{1}^{TV}$ norm.(c) $\rho$ model misfit using $l_{2}$,$l_{2}^{TV}$ and $l_{1}^{TV}$ norm ## 5 Random noise testing In this section, we will test the sensitivity of this deep learning method to data contaminated with random noise drawn from a Gaussian distribution. In Figure 20, we plot a trace from an example shot profile with different ratios of noise. Figure 20: Noise free and noise shotrecords. (a) Noise free record and records with Gaussian noise $std$ = $0.3$. (b) Noise free record and records with Gaussian noise $std$ = $0.5$. (c) Noise free record and records with Gaussian noise $std$ = $1$. In (a), the red signal is the record without noise, and the blue line is the record with Gaussian random noise. The mean value of the noise is zero, and the standard deviation is 0.3 ($std$ = $0.3$); in (b) the noise-free data and data with noise at $std=0.5$ are plotted; in (c) the noise-free data and the data with noise at $std=1$ are plotted. Figure 21: Random noise testing inversion results. (a),(e),(i), noise free inversion results for $V_{P}$, $V_{S}$, and density (b),(f),(j) Gaussian noise $std=0.3$ inversion results for $V_{P}$, $V_{S}$, and density, (c),(g),(k) Gaussian noise $std=0.5$ inversion results for $V_{P}$, $V_{S}$, and density, (d),(h),(l) Gaussian noise $std=1.0$ inversion results for $V_{P}$, $V_{S}$, and density. In Figure 21a, e and i the inversion results from noise-free inversion based on the RNN are plotted. In Figure 21b, f and j, the inversion results with noise at $std=0.3$ for $V_{P}$, $V_{S}$, and ${\rho}$ are plotted, respectively; in Figure 21c, g, and k, the inversion results with noise at $std=0.5$ for $V_{P}$, $V_{S}$, and ${\rho}$ are plotted; in Figure 21d, h, and l are likewise for $std=1.0$. We conclude that a moderate amount of random error in the data used for the RNN training leads to acceptable results, though some blurring is introduced and detail in the structure is lost. The $V_{S}$ recovery appears to be much more sensitive to noise than are those of $V_{P}$ or ${\rho}$. ## 6 Conclusions Elastic multi-parameter full waveform inversion can be formulated as a strongly constrained, theory-based deep-learning network. Specifically, a recurrent neural network, set up with rules enforcing elastic wave propagation, with the wavefield projected onto a measurement surface acting as the labeled data to be compared with observed seismic data, recapitulates elastic FWI but with both (1) the opportunity for data-driven learning to be incorporated, and (2) a design supported by powerful cloud computing architectures. Each cell of the recurrent neural network is designed according to the isotropic-elastic wave equation. The partial derivatives of the data residual with respect to the trainable parameters which act to represent the elastic media are calculated by using the intelligent automatic differential method. With the automatic differential method, gradients can be automatically calculated via the chain rule, guided by backpropagation along the paths within the computational graph. The automatic differential method produces a high level of computational efficiency and scalability for the calculation of gradients for different parameters in elastic media. The formulation is suitable for exploring numerical features of different misfits and different parameterizations, with an aim of improving the resolution of the recovered elastic models, and mitigate cross-talk. We compared RNN waveform inversions based on $l_{2}$, $l_{1}$ with high order total variations. We used this RNN synthetic environment to compare density- velocity (P-wave velocity, S-wave velocity, and density), modulus-density (Lamé parameters and density) and S-D ($C_{11}$, $C_{44}$ and density) parameterizations, and their exposure to cross-talk for the varying misfit functions. Our results suggest generally that this approach to full waveform inversion is consistent and stable. S-D $l1$ norm with high order TV regulation can better resolve the elastic synthetic models. $l1$ norm with high order TV regulation can mitigate the cross-talk between parameters with gradient-based method. ## References * Alkhalifah and Plessix, (2014) Alkhalifah, T., and R.-É. Plessix, 2014, A recipe for practical full-waveform inversion in anisotropic media: An analytical parameter resolution study: Geophysics, 79, R91–R101. * Bar-Sinai et al., (2019) Bar-Sinai, Y., S. Hoyer, J. Hickey, and M. P. Brenner, 2019, Learning data-driven discretizations for partial differential equations: Proceedings of the National Academy of Sciences, 116, 15344–15349. * Brossier et al., (2010) Brossier, R., S. Operto, and J. Virieux, 2010, Which data residual norm for robust elastic frequency-domain full waveform inversion?: Geophysics, 75, R37–R46. * Chen et al., (2019) Chen, Y., M. Zhang, M. Bai, and W. Chen, 2019, Improving the signal-to-noise ratio of seismological datasets by unsupervised machine learning: Seismological Research Letters. * Downton and Hampson, (2018) Downton, J. E., and D. P. Hampson, 2018, Deep neural networks to predict reservoir properties from seismic: CSEG Geoconvention. * Graves et al., (2013) Graves, A., A.-r. Mohamed, and G. Hinton, 2013, Speech recognition with deep recurrent neural networks: 2013 IEEE international conference on acoustics, speech and signal processing, IEEE, 6645–6649. * Guitton et al., (2012) Guitton, A., G. Ayeni, and E. Díaz, 2012, Constrained full-waveform inversion by model reparameterizationgeologically constrained fwi: Geophysics, 77, R117–R127. * Hughes et al., (2019) Hughes, T. W., I. A. Williamson, M. Minkov, and S. Fan, 2019, Wave physics as an analog recurrent neural network: Science Advances, 5, eaay6946. * Innanen, (2014) Innanen, K. A., 2014, Seismic avo and the inverse hessian in precritical reflection full waveform inversion: Geophysical Journal International, 199, 717–734. * Jin et al., (2019) Jin, G., K. Mendoza, B. Roy, and D. G. Buswell, 2019, Machine learning-based fracture-hit detection algorithm using lfdas signal: The Leading Edge, 38, 520–524. * Kalchbrenner and Blunsom, (2013) Kalchbrenner, N., and P. Blunsom, 2013, Recurrent continuous translation models: 1700–1709. * Kamath et al., (2018) Kamath, N., R. Brossier, L. Métivier, and P. Yang, 2018, 3d acoustic/visco-acoustic time-domain fwi of obc data from the valhall field, in SEG Technical Program Expanded Abstracts 2018: Society of Exploration Geophysicists, 1093–1097. * Kamei and Pratt, (2013) Kamei, R., and R. Pratt, 2013, Inversion strategies for visco-acoustic waveform inversion: Geophysical Journal International, 194, 859–884. * Karpatne et al., (2017) Karpatne, A., G. Atluri, J. H. Faghmous, M. Steinbach, A. Banerjee, A. Ganguly, S. Shekhar, N. Samatova, and V. Kumar, 2017, Theory-guided data science: A new paradigm for scientific discovery from data: IEEE Transactions on Knowledge and Data Engineering, 29, 2318–2331. * Keating and Innanen, (2017) Keating, S., and K. A. Innanen, 2017, Crosstalk and frequency bands in truncated newton an-acoustic full-waveform inversion: 1416–1421. * Keating and Innanen, (2019) ——–, 2019, Parameter crosstalk and modeling errors in viscoacoustic seismic full-waveform inversion: Geophysics, 84, R641–R653. * Köhn et al., (2012) Köhn, D., D. De Nil, A. Kurzmann, A. Przebindowska, and T. Bohlen, 2012, On the influence of model parametrization in elastic full waveform tomography: Geophysical Journal International, 191, 325–345. * Lailly, (1983) Lailly, P., 1983, as a sequence of before stack migrations: Conference on Inverse Scattering–Theory and Application, Siam, 206. * (19) Li, C., Y. Zhang, and C. C. Mosher, 2019a, A hybrid learning-based framework for seismic denoising: The Leading Edge, 38, 542–549. * (20) Li, D., K. Xu, J. M. Harris, and E. Darve, 2019b, Time-lapse full waveform inversion for subsurface flow problems with intelligent automatic differentiation: arXiv preprint arXiv:1912.07552. * Liu and Sen, (2009) Liu, Y., and M. K. Sen, 2009, An implicit staggered-grid finite-difference method for seismic modelling: Geophysical Journal International, 179, 459–474. * Mikolov, (2012) Mikolov, T., 2012, Statistical language models based on neural networks: Master’s thesis. * Oh and Alkhalifah, (2016) Oh, J.-W., and T. Alkhalifah, 2016, Elastic orthorhombic anisotropic parameter inversion: An analysis of parameterizationelastic orthorhombic anisotropic fwi: Geophysics, 81, C279–C293. * Operto et al., (2013) Operto, S., Y. Gholami, V. Prieux, A. Ribodetti, R. Brossier, L. Metivier, and J. Virieux, 2013, A guided tour of multiparameter full-waveform inversion with multicomponent data: From theory to practice: The leading edge, 32, 1040–1054. * Pan et al., (2018) Pan, W., K. A. Innanen, and Y. Geng, 2018, Elastic full-waveform inversion and parametrization analysis applied to walk-away vertical seismic profile data for unconventional (heavy oil) reservoir characterization: Geophysical Journal International, 213, 1934–1968. * Pan et al., (2016) Pan, W., K. A. Innanen, G. F. Margrave, M. C. Fehler, X. Fang, and J. Li, 2016, Estimation of elastic constants for hti media using gauss-newton and full-newton multiparameter full-waveform inversion: Geophysics, 81, R275–R291. * Paszke et al., (2017) Paszke, A., S. Gross, S. Chintala, G. Chanan, E. Yang, Z. DeVito, Z. Lin, A. Desmaison, L. Antiga, and A. Lerer, 2017, Automatic differentiation in pytorch. * Peters et al., (2019) Peters, B., E. Haber, and J. Granek, 2019, Neural-networks for geophysicists and their application to seismic data interpretation: arXiv preprint arXiv:1903.11215. * Plessix, (2006) Plessix, R.-E., 2006, A review of the adjoint-state method for computing the gradient of a functional with geophysical applications: Geophysical Journal International, 167, 495–503. * Podlozhnyuk, (2007) Podlozhnyuk, V., 2007, Image convolution with cuda: NVIDIA Corporation white paper, June, 2097. * Pyun et al., (2009) Pyun, S., W. Son, and C. Shin, 2009, Frequency-domain waveform inversion using an l 1-norm objective function: Exploration Geophysics, 40, 227–232. * Raissi, (2018) Raissi, M., 2018, Deep hidden physics models: Deep learning of nonlinear partial differential equations: The Journal of Machine Learning Research, 19, 932–955. * Richardson, (2018) Richardson, A., 2018, Seismic full-waveform inversion using deep learning tools and techniques: arXiv preprint arXiv:1801.07232. * Röth and Tarantola, (1994) Röth, G., and A. Tarantola, 1994, Neural networks and inversion of seismic data: Journal of Geophysical Research: Solid Earth, 99, 6753–6768. * Sambridge et al., (2007) Sambridge, M., P. Rickwood, N. Rawlinson, and S. Sommacal, 2007, Automatic differentiation in geophysical inverse problems: Geophysical Journal International, 170, 1–8. * Shustak and Landa, (2018) Shustak, M., and E. Landa, 2018, Time reversal for wave refocusing and scatterer detection using machine learning: Geophysics, 83, T257–T263. * Smith et al., (2019) Smith, R., T. Mukerji, and T. Lupo, 2019, Correlating geologic and seismic data with unconventional resource production curves using machine learning: Geophysics, 84, O39–O47. * Sun and Demanet, (2018) Sun, H., and L. Demanet, 2018, Low frequency extrapolation with deep learning, in SEG Technical Program Expanded Abstracts 2018: Society of Exploration Geophysicists, 2011–2015. * Sun et al., (2020) Sun, J., Z. Niu, K. A. Innanen, J. Li, and D. O. Trad, 2020, A theory-guided deep-learning formulation and optimization of seismic waveform inversion: Geophysics, 85, R87–R99. * Tarantola, (1984) Tarantola, A., 1984, Inversion of seismic reflection data in the acoustic approximation: Geophysics, 49, 1259–1266. * Tarantola, (1986) ——–, 1986, A strategy for nonlinear elastic inversion of seismic reflection data: Geophysics, 51, 1893–1903. * Virieux, (1986) Virieux, J., 1986, P-sv wave propagation in heterogeneous media: Velocity-stress finite-difference method: Geophysics, 51, 889–901. * Wagner and Rondinelli, (2016) Wagner, N., and J. M. Rondinelli, 2016, Theory-guided machine learning in materials science: Frontiers in Materials, 3, 28. * Wu et al., (2018) Wu, J.-L., H. Xiao, and E. Paterson, 2018, Physics-informed machine learning approach for augmenting turbulence models: A comprehensive framework: Physical Review Fluids, 3, 074602. * Wu and Lin, (2018) Wu, Y., and Y. Lin, 2018, Inversionnet: A real-time and accurate full waveform inversion with cnns and continuous crfs: arXiv preprint arXiv:1811.07875. * Xiang and Zhang, (2016) Xiang, S., and H. Zhang, 2016, Efficient edge-guided full-waveform inversion by canny edge detection and bilateral filtering algorithms: Geophysical Supplements to the Monthly Notices of the Royal Astronomical Society, 207, 1049–1061. * Yang and Ma, (2019) Yang, F., and J. Ma, 2019, Deep-learning inversion: a next generation seismic velocity-model building method: Geophysics, 84, 1–133. * Zaremba et al., (2014) Zaremba, W., I. Sutskever, and O. Vinyals, 2014, Recurrent neural network regularization: arXiv preprint arXiv:1409.2329. * Zhang et al., (2019) Zhang, Z., Y. Wu, Z. Zhou, and Y. Lin, 2019, Velocitygan: Subsurface velocity image estimation using conditional adversarial networks: 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), IEEE, 705–714. * Zheng et al., (2019) Zheng, Y., Q. Zhang, A. Yusifov, and Y. Shi, 2019, Applications of supervised deep learning for seismic interpretation and inversion: The Leading Edge, 38, 526–533.
# Differentially Private SGD with Non-Smooth Losses††thanks: Corresponding author: Yiming Ying. Email<EMAIL_ADDRESS> Puyu Wang†, Yunwen Lei‡, Yiming Ying∗ and Hai Zhang† † School of Mathematics, Northwest University, Xi’an, 710127, China ‡ School of Computer Science, University of Birmingham, Birmingham B15 2TT, UK ∗Department of Mathematics and Statistics, State University of New York at Albany, Albany, NY, 12222, USA ###### Abstract In this paper, we are concerned with differentially private stochastic gradient descent (SGD) algorithms in the setting of stochastic convex optimization (SCO). Most of the existing work requires the loss to be Lipschitz continuous and strongly smooth, and the model parameter to be uniformly bounded. However, these assumptions are restrictive as many popular losses violate these conditions including the hinge loss for SVM, the absolute loss in robust regression, and even the least square loss in an unbounded domain. We significantly relax these restrictive assumptions and establish privacy and generalization (utility) guarantees for private SGD algorithms using output and gradient perturbations associated with non-smooth convex losses. Specifically, the loss function is relaxed to have an $\alpha$-Hölder continuous gradient (referred to as $\alpha$-Hölder smoothness) which instantiates the Lipschitz continuity ($\alpha=0$) and the strong smoothness ($\alpha=1$). We prove that noisy SGD with $\alpha$-Hölder smooth losses using gradient perturbation can guarantee $(\epsilon,\delta)$-differential privacy (DP) and attain optimal excess population risk $\mathcal{O}\Big{(}\frac{\sqrt{d\log(1/\delta)}}{n\epsilon}+\frac{1}{\sqrt{n}}\Big{)}$, up to logarithmic terms, with the gradient complexity $\mathcal{O}(n^{2-\alpha\over 1+\alpha}+n).$ This shows an important trade-off between $\alpha$-Hölder smoothness of the loss and the computational complexity for private SGD with statistically optimal performance. In particular, our results indicate that $\alpha$-Hölder smoothness with $\alpha\geq{1/2}$ is sufficient to guarantee $(\epsilon,\delta)$-DP of noisy SGD algorithms while achieving optimal excess risk with the linear gradient complexity $\mathcal{O}(n).$ Keywords: Stochastic Gradient Descent , Algorithmic Stability, Differential Privacy, Generalization ## 1 Introduction Stochastic gradient descent (SGD) algorithms are widely employed to train a wide range of machine learning (ML) models such as SVM, logistic regression, and deep neural networks. It is an iterative algorithm which replaces the true gradient on the entire training data by a randomized gradient estimated from a random subset (mini-batch) of the available data. As opposed to gradient descent algorithms, this reduces the computational burden at each iteration trading for a lower convergence rate [5]. There is a large amount of work considering the optimization error (convergence analysis) of SGD and its variants in the linear case [2, 19, 20, 29, 30] as well as the general setting of reproducing kernel Hilbert spaces [10, 23, 28, 36, 37, 32]. At the same time, data collected often contain sensitive information such as individual records from schools and hospitals, financial records for fraud detection, online behavior from social media and genomic data from cancer diagnosis. Modern ML algorithms can explore the fine-grained information about data in order to make a perfect prediction which, however, can lead to privacy leakage [8, 31]. To a large extent, SGD algorithms have become the workhorse behind the remarkable progress of ML and AI. Therefore, it is of pivotal importance for developing privacy-preserving SGD algorithms to protect the privacy of the data. Differential privacy (DP) [12, 14] has emerged as a well- accepted mathematical definition of privacy which ensures that an attacker gets roughly the same information from the dataset regardless of whether an individual is present or not. Its related technologies have been adopted by Google [15], Apple [25], Microsoft [11] and the US Census Bureau [1]. In this paper, we are concerned with differentially private SGD algorithms in the setting of stochastic convex optimization (SCO). Specifically, let the input space $\mathcal{X}$ be a domain in some Euclidean space, the output space $\mathcal{Y}\subseteq\mathbb{R}$, and $\mathcal{Z}=\mathcal{X}\times\mathcal{Y}.$ Denote the loss function by $\ell:\mathbb{R}^{d}\times\mathcal{Z}\mapsto[0,\infty)$ and assume, for any $z\in\mathcal{Z}$, that $\ell(\cdot,z)$ is convex with respect to (w.r.t.) the first argument. SCO aims to minimize the expected (population) risk, i.e. $\mathcal{R}(\mathbf{w}):=\mathbb{E}_{z}[\ell(\mathbf{w},z)]$, where the model parameter $\mathbf{w}$ belongs to a (not necessarily bounded) domain $\mathcal{W}\subseteq\mathbb{R}^{d}$, and the expectation is taken w.r.t. $z$ according to a population distribution $\mathcal{D}.$ While the population distribution is usually unknown, we have access to a finite set of $n$ training data points denoted by $S=\\{z_{i}\in\mathcal{Z}:i=1,2,\ldots,n\\}.$ It is assumed to be independently and identically distributed (i.i.d.) according to the distribution $\mathcal{D}$ on $\mathcal{Z}.$ In this context, one often considers SGD algorithms to solve the Empirical Risk Minimization (ERM) problem defined by $\min_{\mathbf{w}\in\mathcal{W}}\Bigl{\\{}\mathcal{R}_{S}(\mathbf{w}):=\frac{1}{n}\sum_{i=1}^{n}\ell(\mathbf{w},z_{i})\Bigr{\\}}.$ For a randomized algorithm (e.g., SGD) $\mathcal{A}$ to solve the above ERM problem, let $\mathcal{A}(S)$ be the output of algorithm $\mathcal{A}$ based on the dataset $S$. Then, its statistical generalization performance is measured by the excess (population) risk, i.e., the discrepancy between the expected risk $\mathcal{R}(\mathcal{A}(S))$ and the least possible one in $\mathcal{W}$, which is defined by $\epsilon_{\text{risk}}(\mathcal{A}(S))=\mathcal{R}(\mathcal{A}(S))-\min_{\mathbf{w}\in\mathcal{W}}\mathcal{R}(\mathbf{w}).$ Along this line, there are a considerable amount of work [35, 4, 16] on analyzing the excess risk of private SGD algorithms in the setting of SCO. However, most of such approaches often require two assumptions: 1) the loss $\ell$ is $L$-Lipschitz and $\beta$-smooth; 2) the domain $\mathcal{W}$ is uniformly bounded. These assumptions are very restrictive as many popular losses violate these conditions including the hinge loss $(1-y\mathbf{w}^{T}x)^{q}_{+}$ for $q$-norm soft margin SVM and the $q$-norm loss $|y-\mathbf{w}^{T}x|^{q}$ in regression with $1\leq q\leq 2.$ More specifically, the work [35] assumed the loss to be Lipschitz continuous and strongly smooth and showed that the private SGD algorithm with output perturbation can achieve $(\epsilon,\delta)$-DP and an excess risk rate $\mathcal{O}(\frac{(d\log(1/\delta))^{1/4}}{\sqrt{n\epsilon}})$ when the gradient complexity (i.e. the number of computing gradients) $T=n$. The study [4] proved, under the same assumptions, that the private SGD algorithm with gradient perturbation can achieve an optimal excess risk rate $\mathcal{O}\Big{(}\frac{\sqrt{d\log(1/\delta)}}{n\epsilon}+\frac{1}{\sqrt{n}}\Big{)}$ while guaranteeing its $(\epsilon,\delta)$-DP. To deal with the non- smoothness, it used the Moreau envelope technique to smooth the loss function and got the optimal rate. However, the algorithm is computationally inefficient with a gradient complexity $\mathcal{O}\Big{(}n^{4.5}\sqrt{\epsilon}+\frac{n^{6.5}\epsilon^{4.5}}{(d\log(1/\delta))^{2}}\Big{)}.$ The work [16] improved the gradient complexity of the algorithm to $\mathcal{O}(n^{2}\log(\frac{1}{\delta}))$ by localizing the approximate minimizer of the population loss on each phase. Recently, [3] showed that a simple variant of noisy projected SGD yields the optimal rate with gradient complexity $\mathcal{O}(n^{2})$. However, it only focused on the Lipschitz continuous losses and assumed that the parameter domain $\mathcal{W}$ is bounded. Our main contribution is to significantly relax these restrictive assumptions and to prove both privacy and generalization (utility) guarantees for private SGD algorithms with non-smooth convex losses in both bounded and unbounded domains. Specifically, the loss function $\ell(\mathbf{w},z)$ is relaxed to have an $\alpha$-Hölder continuous gradient w.r.t. the first argument, i.e., there exists $L>0$ such that, for any $\mathbf{w},\mathbf{w}^{\prime}\in\mathcal{W}$ and any $z\in\mathcal{Z}$, $\|\partial\ell(\mathbf{w},z)-\partial\ell(\mathbf{w}^{\prime},z)\|_{2}\leq L\|\mathbf{w}-\mathbf{w}^{\prime}\|_{2}^{\alpha},$ where $\|\cdot\|_{2}$ denotes the Euclidean norm, $\partial\ell(\mathbf{w},z)$ denotes a subgradient of $\ell$ w.r.t. the first argument. For the sake of notional simplicity, we refer to this condition as $\alpha$-Hölder smoothness with parameter $L$. The smoothness parameter $\alpha\in[0,1]$ characterizes the smoothness of the loss function $\ell(\cdot,z)$. The case of $\alpha=0$ corresponds to the Lipschitz continuity of the loss $\ell$ while $\alpha=1$ means its strong smoothness. This definition instantiates many non-smooth loss functions mentioned above. For instance, the hinge loss for $q$-norm soft- margin SVM and $q$-norm loss for regression mentioned above with $q\in[1,2]$ are $(q-1)$-Hölder smooth. In particular, we prove that noisy SGD with $\alpha$-Hölder smooth losses using gradient perturbation can guarantee $(\epsilon,\delta)$-DP and attain the optimal excess population risk $\mathcal{O}\Big{(}\frac{\sqrt{d\log(1/\delta)}}{n\epsilon}+\frac{1}{\sqrt{n}}\Big{)}$, up to logarithmic terms, with gradient complexity $\mathcal{O}(n^{2-\alpha\over 1+\alpha}+n).$ This shows an important trade-off between $\alpha$-Hölder smoothness of the loss and the computational complexity for private SGD in order to achieve statistically optimal performance. In particular, our results indicate that $\alpha$-Hölder smoothness with $\alpha\geq{1/2}$ is sufficient to guarantee $(\epsilon,\delta)$-DP of noisy SGD algorithms while achieving the optimal excess risk with linear gradient complexity $\mathcal{O}(n).$ Table 1 summarizes the upper bound of the excess population risk, gradient complexity of the aforementioned algorithms in comparison to our methods. Our key idea to handle general Hölder smooth losses is to establish the approximate non-expansiveness of the gradient mapping, and the refined boundedness of the iterates of SGD algorithms when domain $\mathcal{W}$ is unbounded. This allows us to show the uniform argument stability [24] of the iterates of SGD algorithms with high probability w.r.t. the internal randomness of the algorithm (not w.r.t. the data $S$), and consequently estimate the generalization error of differentially private SGD with non- smooth losses. Reference | Loss | Method | Utility bounds | Gradient Complexity | Domain ---|---|---|---|---|--- [35] | Lipschitz | Output | $\mathcal{O}\Big{(}\frac{(d\log(\frac{1}{\delta}))^{\frac{1}{4}}}{\sqrt{n\epsilon}}\Big{)}$ | $\mathcal{O}\big{(}n\big{)}$ | bounded & smooth [4] | Lipschitz | Gradient | $\mathcal{O}\Big{(}\frac{\sqrt{d\log(\frac{1}{\delta})}}{n\epsilon}+\frac{1}{\sqrt{n}}\Big{)}$ | $\mathcal{O}\Big{(}n^{1.5}\sqrt{\epsilon}+\frac{(n\epsilon)^{2.5}}{d\log(\frac{1}{\delta})}\Big{)}$ | bounded & smooth Lipschitz | Gradient | $\mathcal{O}\Big{(}\frac{\sqrt{d\log(\frac{1}{\delta})}}{n\epsilon}+\frac{1}{\sqrt{n}}\Big{)}$ | $\mathcal{O}\Big{(}n^{4.5}\sqrt{\epsilon}+\frac{n^{6.5}\epsilon^{4.5}}{(d\log(\frac{1}{\delta}))^{2}}\Big{)}$ | bounded [16] | Lipschitz | Phased Output | $\mathcal{O}\Big{(}\frac{\sqrt{d\log(\frac{1}{\delta})}}{n\epsilon}+\frac{1}{\sqrt{n}}\Big{)}$ | $\mathcal{O}\big{(}n\big{)}$ | bounded & smooth Lipschitz | Phased ERM | $\mathcal{O}\Big{(}\frac{\sqrt{d\log(\frac{1}{\delta})}}{n\epsilon}+\frac{1}{\sqrt{n}}\Big{)}$ | $\mathcal{O}\big{(}n^{2}\log(\frac{1}{\delta})\big{)}$ | bounded [3] | Lipschitz | Gradient | $\mathcal{O}\Big{(}\frac{\sqrt{d\log(\frac{1}{\delta})}}{n\epsilon}+\frac{1}{\sqrt{n}}\Big{)}$ | $\mathcal{O}\big{(}n^{2}\big{)}$ | bounded Ours | $\alpha$-Hölder | Output | $\mathcal{O}\Big{(}\frac{(d\log(\frac{1}{\delta}))^{\frac{1}{4}}\sqrt{\log(\frac{n}{\delta})}}{\sqrt{n\epsilon}}\Big{)}$ | $\mathcal{O}\big{(}n^{\frac{2-\alpha}{1+\alpha}}+n\big{)}$ | bounded smooth $\alpha$-Hölder | Output | $\mathcal{O}\Big{(}\frac{\sqrt{d\log(\frac{1}{\delta})}\log(\frac{n}{\delta})}{n^{\frac{2}{3+\alpha}}\epsilon}+\frac{\log(\frac{n}{\delta})}{n^{\frac{1}{3+\alpha}}}\Big{)}$ | $\mathcal{O}\big{(}n^{\frac{-\alpha^{2}-3\alpha+6}{(1+\alpha)(3+\alpha)}}+n\big{)}$ | unbounded smooth $\alpha$-Hölder | Gradient | $\mathcal{O}\Big{(}\frac{\sqrt{d\log(\frac{1}{\delta})}}{n\epsilon}+\frac{1}{\sqrt{n}}\Big{)}$ | $\mathcal{O}\big{(}n^{\frac{2-\alpha}{1+\alpha}}+n\big{)}$ | bounded smooth Table 1: Comparison of different $(\epsilon,\delta)$-DP algorithms. We report the method, utility (generalization) bound, gradient complexity and parameter domain for three types of convex losses, i.e. Lipschitz, Lipschitz and smooth, and $\alpha$-Hölder smooth. Here Output, Gradient, Phased Output and Phased ERM denote output perturbation which adds Gaussian noise to the output of non- private SGD, gradient perturbation which adds Gaussian noise at each SGD update, phased output perturbation and phased ERM output perturbation [16], respectively. The gradient complexity is the total number of computing the gradient on one datum in the algorithm. Organization of the Paper. The rest of the paper is organized as follows. The formulation of SGD algorithms and the main results are given in Section 2. We provide the proofs in Section 3 and conclude the paper in Section 4. ## 2 Problem Formulation and Main Results ### 2.1 Preliminaries Throughout the paper, we assume that the loss function $\ell:\mathcal{W}\times\mathcal{Z}\rightarrow\mathbb{R}$ is convex w.r.t. the first argument, i.e., for any $z\in\mathcal{Z}$ and $\mathbf{w},\mathbf{w}^{\prime}\in\mathcal{W}$, there holds $\ell(\mathbf{w},z)\geq\ell(\mathbf{w}^{\prime},z)+\langle\partial\ell(\mathbf{w}^{\prime},z),\mathbf{w}-\mathbf{w}^{\prime}\rangle$ where $\partial\ell(\mathbf{w}^{\prime},z)$ denotes a subgradient of $\ell(\cdot,z)$ in the first argument. We restrict our attention to the (projected) stochastic gradient descent algorithm which is defined as below. ###### Definition 1 (Stochastic Gradient Descent). Let $\mathcal{W}\subseteq\mathbb{R}^{d}$ be convex, $T$ denote the number of iterations, and $\text{Proj}_{\mathcal{W}}$ denote the projection to $\mathcal{W}$. Let $\mathbf{w}_{1}=\mathbf{0}\in\mathbb{R}^{d}$ be an initial point, and $\\{\eta_{t}\\}_{t=1}^{T-1}$ be a sequence of positive step sizes. At step $t\in\\{1,\ldots,T-1\\}$, the update rule of (projected) stochastic gradient decent is given by $\mathbf{w}_{t+1}=\text{Proj}_{\mathcal{W}}\big{(}\mathbf{w}_{t}-\eta_{t}\partial\ell(\mathbf{w}_{t},z_{i_{t}})\big{)},$ (1) where $\\{i_{t}\\}$ is uniformly drawn from $[n]:=\\{1,2,\ldots,n\\}$. When $\mathcal{W}=\mathbb{R}^{d}$, then (1) is reduced to $\mathbf{w}_{t+1}=\mathbf{w}_{t}-\eta_{t}\partial\ell(\mathbf{w}_{t},z_{i_{t}}).$ For a randomized learning algorithm $\mathcal{A}:\mathcal{Z}^{n}\rightarrow\mathcal{W}$, let $\mathcal{A}(S)$ denote the model produced by running $\mathcal{A}$ over the training dataset $S$. We say two datasets $S$ and $S^{\prime}$ are neighboring datasets, denoted by $S\simeq S^{\prime}$, if they differ by a single datum. We consider the following high-probabilistic version of the uniform argument stability (UAS), which is an extension of the UAS in expectation [24]. ###### Definition 2 (Uniform argument stability). We say an algorithm $\mathcal{A}$ has $\Delta_{\mathcal{A}}$-UAS with probability at least $1-\gamma$ ($\gamma\in(0,1)$) if $\mathbb{P}_{\mathcal{A}}(\sup_{S\simeq S^{\prime}}\delta_{\mathcal{A}}(S,S^{\prime})\geq\Delta_{\mathcal{A}})\leq\gamma,$ where $\delta_{\mathcal{\mathcal{A}}}(S,S^{\prime}):=\|\mathcal{A}(S)-\mathcal{A}(S^{\prime})\|_{2}.$ We will use UAS to study generalization bounds with high probability. In particular, the following lemma as a straightforward extension of Corollary 8 in [7] establishes the relationship between UAS and generalization errors. The proof is given in the Appendix for completeness. ###### Lemma 1. Suppose $\ell$ is nonnegative, convex and $\alpha$-Hölder smooth with parameter $L$. Let $M_{0}=\sup_{z\in\mathcal{Z}}\ell(0,z)$ and $M=\sup_{z\in\mathcal{Z}}\|\partial\ell(0,z)\|_{2}$. Let $\mathcal{A}$ be a randomized algorithm with the output of $\mathcal{A}$ bounded by $G$ and $\mathbb{P}_{\mathcal{A}}(\sup_{S\simeq S^{\prime}}\delta_{\mathcal{A}}(S,S^{\prime})\geq\Delta_{\mathcal{A}})\leq\gamma_{0}.$ Then there exists a constant $c>0$ such that for any distribution $\mathcal{D}$ over $\mathcal{Z}$ and any $\gamma\in(0,1)$, there holds $\mathbb{P}_{\mathcal{S}\sim\mathcal{D}^{n},\mathcal{A}}\biggl{[}|\mathcal{R}(\mathcal{A(S)})-\mathcal{R}_{S}(\mathcal{A(S)})|\geq c\bigg{(}(M+LG^{\alpha})\Delta_{\mathcal{A}}\log(n)\log(1/{\gamma})+\big{(}M_{0}+(M+LG^{\alpha})G\big{)}\sqrt{n^{-1}\log(1/\gamma)}\bigg{)}\biggr{]}\leq\gamma_{0}+\gamma.$ Differential privacy [13] is a de facto standard privacy measure for a randomized algorithm $\mathcal{A}.$ ###### Definition 3 (Differential Privacy). We say a randomized algorithm $\mathcal{A}$ satisfies $(\epsilon,\delta)$-DP if, for any two neighboring datasets $S$ and $S^{\prime}$ and any event $E$ in the output space of $\mathcal{A}$, there holds $\mathbb{P}(\mathcal{A}(S)\in E)\leq e^{\epsilon}\mathbb{P}(\mathcal{A}(S^{\prime})\in E)+\delta.$ In particular, we call it satisfies $\epsilon$-DP if $\delta=0$. We also need the following concept called $\ell_{2}$-sensitivity. ###### Definition 4 ($\ell_{2}$-sensitivity). The $\ell_{2}$-sensitivity of a function (mechanism) $\mathcal{M}:\mathcal{Z}^{n}\rightarrow\mathcal{W}$ is defined as $\Delta=\sup_{S\simeq S^{\prime}}\|\mathcal{M}(S)-\mathcal{M}(S^{\prime})\|_{2},$ where $S$ and $S^{\prime}$ are neighboring datasets. A basic mechanism to obtain $(\epsilon,\delta)$-DP from a given function $\mathcal{M}:\mathcal{Z}^{n}\rightarrow\mathcal{W}$ is to add a random noise from a Gaussian distribution $\mathcal{N}(0,\sigma^{2}\mathbf{I}_{d})$ where $\sigma$ is proportional to its $\ell_{2}$-sensitivity. This mechanism is often referred to as Gaussian mechanism as stated in the following lemma. ###### Lemma 2 ([14]). Given a function $\mathcal{M}:\mathcal{Z}^{n}\rightarrow\mathcal{W}$ with the $\ell_{2}$-sensitivity $\Delta$ and a dataset $S\subset\mathcal{Z}^{n}$, and assume that $\sigma\geq\frac{\sqrt{2\log(1.25/\delta)}\Delta}{\epsilon}$. The following Gaussian mechanism yields $(\epsilon,\delta)$-DP: $\mathcal{G}(S,\sigma):=\mathcal{M}(S)+\mathbf{b},~{}~{}\mathbf{b}\sim\mathcal{N}(0,\sigma^{2}\mathbf{I}_{d}),$ where $\mathbf{I}_{d}$ is the identity matrix in $\mathbb{R}^{d\times d}$. Although the concept of $(\epsilon,\delta)$-DP is widely used in privacy- preserving methods, its composition and subsampling amplification results are relatively loose, which are not suitable for iterative SGD algorithms. Based on the Rényi divergence, the work [26] proposed Rényi differential privacy (RDP) as a relaxation of DP to achieve tighter analysis of composition and amplification mechanisms. ###### Definition 5 (RDP [26]). For $\lambda>1$, $\rho>0$, a randomized mechanism $\mathcal{A}$ satisfies $(\lambda,\rho)$-RDP, if, for all neighboring datasets $S$ and $S^{\prime}$, we have $D_{\lambda}\big{(}\mathcal{A}(S)\parallel\mathcal{A}(S^{\prime})\big{)}:=\frac{1}{\lambda-1}\log\int\Big{(}\frac{P_{\mathcal{A}(S)}(\theta)}{P_{\mathcal{A}(S^{\prime})}(\theta)}\Big{)}^{\lambda}dP_{\mathcal{A}(S^{\prime})}(\theta)\leq\rho,$ where $P_{\mathcal{A}(S)}(\theta)$ and $P_{\mathcal{A}(S^{\prime})}(\theta)$ are the density of $\mathcal{A}(S)$ and $\mathcal{A}(S^{\prime})$, respectively. As $\lambda\rightarrow\infty$, RDP reduces to $\epsilon$-DP, i.e., $\mathcal{A}$ satisfies $\epsilon$-DP if and only if $D_{\infty}\big{(}\mathcal{A}(S)||\mathcal{A}(S^{\prime})\big{)}\leq\epsilon$ for any neighboring datasets $S$ and $S^{\prime}$. Our analysis requires the introduction of several lemmas on useful properties of RDP listed below. First, we introduce the privacy amplification of RDP by uniform subsampling, which is fundamental to establish privacy guarantees of noisy SGD algorithms. In general, a uniform subsampling scheme first draws a subset with size $pn$ uniformly at random with a subsampling rate $p\leq 1$, and then applies a known randomized mechanism to the subset. ###### Lemma 3 ([22]). Consider a function $\mathcal{M}:\mathcal{Z}^{n}\rightarrow\mathcal{W}$ with the $\ell_{2}$-sensitivity $\Delta$, and a dataset $S\subset\mathcal{Z}^{n}$. The Gaussian mechanism $\mathcal{G}(S,\sigma)=\mathcal{M}(S)+\mathbf{b}$, where $\mathbf{b}\sim\mathcal{N}(0,\sigma^{2}\mathbf{I}_{d})$, applied to a subset of samples that are drawn uniformly without replacement with subsampling rate $p$ satisfies $(\lambda,3.5p^{2}\lambda\Delta^{2}/\sigma^{2})$-RDP if $\sigma^{2}\geq 0.67\Delta^{2}$ and $\lambda-1\leq\frac{2\sigma^{2}}{3\Delta^{2}}\log\big{(}\frac{1}{\lambda p(1+\sigma^{2}/\Delta^{2})}\big{)}$. The following adaptive composition theorem of RDP establishes the privacy of a composition of several adaptive mechanisms in terms of that of individual mechanisms. We say a sequence of mechanisms $(\mathcal{A}_{1},\ldots,\mathcal{A}_{k})$ are chosen adaptively if $\mathcal{A}_{i}$ can be chosen based on the outputs of the previous mechanisms $\mathcal{A}_{1}(S),\ldots,\mathcal{A}_{i-1}(S)$ for any $i\in[k]$. ###### Lemma 4 (Adaptive Composition of RDP [26]). If a mechanism $\mathcal{A}$ consists of a sequence of adaptive mechanisms $(\mathcal{A}_{1},\ldots,\mathcal{A}_{k})$ with $\mathcal{A}_{i}$ satisfying $(\lambda,\rho_{i})$-RDP, $i\in[k]$, then $\mathcal{A}$ satisfies $(\lambda,\sum_{i=1}^{k}\rho_{i})$-RDP. Lemme 4 tells us that the derivation of the privacy guarantee for a composition mechanism is simple and direct. This is the underlying reason that we adopt RDP in our subsequent privacy analysis. The following lemma allows us to further convert RDP back to $(\epsilon,\delta)$-DP. ###### Lemma 5 (From RDP to $(\epsilon,\delta)$-DP [26]). If a randomized mechanism $\mathcal{A}$ satisfies $(\lambda,\rho)$-RDP, then $\mathcal{A}$ satisfies $(\rho+\log(1/\delta)/(\lambda-1),\delta)$-DP for all $\delta\in(0,1)$. The following lemma shows that a post-processing procedure always preserves privacy. ###### Lemma 6 (Post-processing [26]). Let $\mathcal{A}:\mathcal{Z}^{n}\rightarrow\mathcal{W}_{1}$ satisfy $(\lambda,\rho)$-RDP and $f:\mathcal{W}_{1}\rightarrow\mathcal{W}_{2}$ be an arbitrary function. Then $f\circ\mathcal{A}:\mathcal{Z}^{n}\rightarrow\mathcal{W}_{2}$ satisfies $(\lambda,\rho)$-RDP. ### 2.2 Main Results We present our main results here. First, we state a key bound of UAS for SGD when $\mathcal{W}\subseteq\mathbb{R}^{d}$ and the loss function is $\alpha$-Hölder smooth. Then, we propose two privacy-preserving SGD-type algorithms using output and gradient perturbations, and present the corresponding privacy and generalization (utility) guarantees. The utility guarantees in terms of the excess risk typically rely on two main errors: optimization errors and generalization errors, as shown soon in (3) and (4) for the algorithms with output and gradient perturbations, respectively. We will apply techniques in optimization theory to handle the optimization errors [27], and the concept of UAS [6, 17, 24], which was given in Definition 2 in Subsection 2.1, to estimate the generalization errors. #### 2.2.1 UAS bound of SGD with Non-Smooth Losses We begin by stating the key result on the distance between two iterate trajectories produced by SGD on neighboring datasets. Let $c_{\alpha,1}=\begin{cases}(1+1/\alpha)^{\frac{\alpha}{1+\alpha}}L^{\frac{1}{1+\alpha}},&\mbox{if }\alpha\in(0,1]\\\ M+L,&\mbox{if }\alpha=0.\end{cases}$ (2) and $c_{\alpha,2}=\sqrt{\frac{1-\alpha}{1+\alpha}}(2^{-\alpha}L)^{\frac{1}{1-\alpha}}$, where $M=\sup_{z\in\mathcal{Z}}\|\partial\ell(0,z)\|_{2}$. In addition, define $C_{\alpha}=\frac{1-\alpha}{1+\alpha}c^{\frac{2(1+\alpha)}{1-\alpha}}_{\alpha,1}\big{(}\frac{\alpha}{1+\alpha}\big{)}^{\frac{2\alpha}{1-\alpha}}+2\sup_{z\in\mathcal{Z}}\ell(0;z)$. Furthermore, let $\mathcal{B}(0,r)$ denote the Euclidean ball of radius $r>0$ centered at $0\in\mathbb{R}^{d}$. Without loss of generality, we assume $\eta>1/T$. ###### Theorem 7. Suppose that the loss function $\ell$ is convex and $\alpha$-Hölder smooth with parameter $L$. Let $\mathcal{A}$ be the SGD with $T$ iterations and $\eta_{t}=\eta<\min\\{1,1/L\\}$, and $\bar{\mathbf{w}}=\frac{1}{T}\sum_{t=1}^{T}\mathbf{w}_{t}$ be the output produced by $\mathcal{A}$. Further, let $c_{\gamma,T}=\max\Big{\\{}\big{(}3n\log(n/\gamma)/T\big{)}^{\frac{1}{2}},3n\log(n/\gamma)/T\Big{\\}}$. 1. (a) If $\ell$ is nonnegative and $\mathcal{W}=\mathbb{R}^{d}$, then, for any $\gamma\in(0,1)$, there holds $\mathbb{P}_{\mathcal{A}}\Big{(}\sup_{S\simeq S^{\prime}}\delta_{\mathcal{A}}(S,S^{\prime})\geq\Delta_{SGD}(\gamma)\Big{)}\leq\gamma,$ where $\Delta_{SGD}(\gamma)=\Big{(}e\big{(}c^{2}_{\alpha,2}T\eta^{\frac{2}{1-\alpha}}+4\big{(}M+L(C_{\alpha}T\eta)^{\frac{\alpha}{2}}\big{)}^{2}\eta^{2}\Big{(}1+\frac{T}{n}(1+c_{\gamma,T})\Big{)}\frac{T}{n}(1+c_{\gamma,T})\big{)}\Big{)}^{1/2}$. 2. (b) If $\mathcal{W}\subseteq\mathcal{B}(0,R)$ with $R>0$, then, for any $\gamma\in(0,1)$, there holds $\mathbb{P}_{\mathcal{A}}\Big{(}\sup_{S\simeq S^{\prime}}\delta_{\mathcal{A}}(S,S^{\prime})\geq\tilde{\Delta}_{\text{SGD}}(\gamma)\Big{)}\leq\gamma,$ where $\tilde{\Delta}_{\text{SGD}}(\gamma)=\Big{(}e\big{(}c^{2}_{\alpha,2}T\eta^{\frac{2}{1-\alpha}}+4\big{(}M+LR^{\alpha}\big{)}^{2}\eta^{2}\Big{(}1+\frac{T}{n}(1+c_{\gamma,T})\Big{)}\frac{T}{n}(1+c_{\gamma,T})\big{)}\Big{)}^{1/2}$. ###### Remark 1. Under the reasonable assumption of $T\geq n$, we have $c_{\gamma,T}=\mathcal{O}(\log(n/\gamma))$. Then $\Delta_{\text{SGD}}(\gamma)=\mathcal{O}\Big{(}\sqrt{T}\eta^{\frac{1}{1-\alpha}}+\frac{(T\eta)^{1+\alpha/2}\log(n/\gamma)}{n}\Big{)}$ and $\tilde{\Delta}_{\text{SGD}}(\gamma)=\mathcal{O}\Big{(}\sqrt{T}\eta^{\frac{1}{1-\alpha}}+\frac{T\eta\log(n/\gamma)}{n}\Big{)}$. In addition, if $\ell$ is strongly smooth, i.e., $\alpha=1$, the first term in the UAS bounds tends to $0$ under the typical assumption of $\eta<1$. In this case we have $\Delta_{\text{SGD}}(\gamma)=\mathcal{O}\Big{(}\frac{\big{(}T\eta\big{)}^{3/2}\log(n/\gamma)}{n}\Big{)}$ and $\tilde{\Delta}_{\text{SGD}}(\gamma)=\mathcal{O}\Big{(}\frac{T\eta\log(n/\gamma)}{n}\Big{)}$. The work [3] established the high probability upper bound of the random variable of the argument stability $\delta_{SGD}$ in the order of $\mathcal{O}(\sqrt{T}\eta+\frac{T\eta}{n})$ for Lipschitz continuous losses under an additional assumption $\gamma\geq\exp(-n/2)$. Our result gives the upper bound of $\sup_{S\simeq S^{\prime}}\delta_{SGD}(S,S^{\prime})$ in the order of $\mathcal{O}(\sqrt{T}\eta+\frac{T\eta\log(n/\gamma)}{n})$ for any $\gamma\in(0,1)$ for the case of $\alpha=0$. The work [17] gave the bound of $\mathcal{O}({T\eta}/{n})$ in expectation for Lipschitz continuous and smooth loss functions. As a comparison, our stability bounds are stated with high probability and do not require the Lipschitz condition. Under a further Lipschitz condition, our stability bounds actually recover the bound $\mathcal{O}({T\eta}/{n})$ in [17] in the smooth case. Indeed, both the term $\big{(}M+(C_{\alpha}T\eta)^{\frac{\alpha}{2}}\big{)}^{2}$ and the term $\big{(}M+LR^{\alpha}\big{)}^{2}$ are due to controlling the magnitude of gradients, and can be replaced by $L^{2}$ for $L$-Lipschitz losses. #### 2.2.2 Differentially Private SGD with Output Perturbation 1: Inputs: Data $S=\\{z_{i}\in\mathcal{Z}:i=1,\ldots,n\\}$, $\alpha$-Hölder smooth loss $\ell(\mathbf{w},z)$ with parameter $L$, the convex set $\mathcal{W}$, step size $\eta$, number of iterations $T$, and privacy parameters $\epsilon$, $\delta$ 2: Set: $\mathbf{w}_{1}=\mathbf{0}$ 3: for $t=1$ to $T$ do 4: Sample $i_{t}\sim\text{Unif}([n])$ 5: $\mathbf{w}_{t+1}=\text{Proj}_{\mathcal{W}}(\mathbf{w}_{t}-\eta\partial\ell(\mathbf{w}_{t};z_{i_{t}}))$ 6: end for 7: if $\mathcal{W}=\mathbb{R}^{d}$ then 8: let $\Delta=\Delta_{\text{SGD}}(\delta/2)$ 9: else if $\mathcal{W}\subseteq\mathcal{B}(0,R)$ then 10: let $\Delta=\tilde{\Delta}_{\text{SGD}}(\delta/2)$ 11: end if 12: Compute: $\sigma^{2}=\frac{2\log(2.5/\delta)\Delta^{2}}{\epsilon^{2}}$ 13: return: ${\mathbf{w}}_{\text{priv}}=\frac{1}{T}\sum_{t=1}^{T}\mathbf{w}_{t}+\mathbf{b}$ where $\mathbf{b}\sim\mathcal{N}(0,\sigma^{2}\mathbf{I}_{d})$ Algorithm 1 Differentially Private SGD with Output perturbation (DP-SGD- Output) Output perturbation [9, 13] is a common approach to achieve $(\epsilon,\delta)$-DP. The main idea is to add a random noise $\mathbf{b}$ to the output of the SGD algorithm, where $\mathbf{b}$ is randomly sampled from the Gaussian distribution with mean $0$ and variance proportional to the $\ell_{2}$-sensitivity of SGD. In Algorithm 1, we propose the private SGD algorithm with output perturbation for non-smooth losses in both bounded domain $\mathcal{W}\subseteq\mathcal{B}(0,R)$ and unbounded domain $\mathcal{W}=\mathbb{R}^{d}$. The difference in these two cases is that we add random noise with different variances according to the sensitivity analysis of SGD stated in Theorem 7. In the sequel, we present the privacy and utility guarantees for Algorithm 1. ###### Theorem 8 (Privacy guarantee). Suppose that the loss function $\ell$ is convex, nonnegative and $\alpha$-Hölder smooth with parameter $L$. Then Algorithm 1 (DP-SGD-Output) satisfies $(\epsilon,\delta)$-DP. According to the definitions, the $\ell_{2}$-sensitivity of SGD is identical to the UAS of SGD: $\sup_{S\simeq S^{\prime}}\delta_{SGD}(S,S^{\prime})$. In this sense, the proof of Theorem 8 directly follows from Theorem 7 and Lemma 2. For completeness, we include the detailed proof in Subsection 3.2. Recall that the empirical risk is defined by $\mathcal{R}_{S}(\mathbf{w})=\frac{1}{n}\sum_{i=1}^{n}\ell(\mathbf{w},z_{i})$, and the population risk is $\mathcal{R}(\mathbf{w})=\mathbb{E}_{z}[\ell(\mathbf{w},z)]$. Let $\mathbf{w}^{*}\in\arg\min_{\mathbf{w}\in\mathcal{W}}\mathcal{R}(\mathbf{w})$ be the one with the best prediction performance over $\mathcal{W}$. We use the notation $B\asymp\tilde{B}$ if there exist constants $c_{1},c_{2}>0$ such that $c_{1}\tilde{B}<B\leq c_{2}\tilde{B}$. Without loss of generality, we always assume $\|\mathbf{w}^{*}\|_{2}\geq 1$. ###### Theorem 9 (Utility guarantee for unbounded domain). Suppose the loss function $\ell$ is nonnegative, convex and $\alpha$-Hölder smooth with parameter $L$. Let $\mathbf{w}_{\text{priv}}$ be the output produced by Algorithm 1 with $\mathcal{W}=\mathbb{R}^{d}$ and $\eta=n^{\frac{1}{3+\alpha}}/\big{(}T(\log(\frac{1}{\gamma}))^{\frac{1}{3+\alpha}}\big{)}$. Let $T\asymp n^{\frac{-\alpha^{2}-3\alpha+6}{(1+\alpha)(3+\alpha)}}$ if $0\leq\alpha<\frac{\sqrt{73}-7}{4}$, and $T\asymp n$ else. Then, for any $\gamma\in(4\max\\{\exp(-d/8),\delta\\},1)$, with probability at least $1-\gamma$ over the randomness in both the sample and the algorithm, there holds $\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})=\|\mathbf{w}^{*}\|_{2}^{2}\cdot\mathcal{O}\bigg{(}\frac{\sqrt{d\log(1/\delta)}{\log(n/\delta)}}{(\log(1/\gamma))^{\frac{1+\alpha}{4(3+\alpha)}}n^{\frac{2}{3+\alpha}}\epsilon}+\frac{\log(n)\big{(}\log(1/\gamma)\big{)}^{\frac{2}{3+\alpha}}{\log(n/\delta)}}{n^{\frac{1}{3+\alpha}}}\bigg{)}.$ To examine the excess population risk $\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})$, we use the following error decomposition: $\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})=[\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\bar{\mathbf{w}})]+[\mathcal{R}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\bar{\mathbf{w}})]+[\mathcal{R}_{S}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\mathbf{w}^{*})]+[\mathcal{R}_{S}(\mathbf{w}^{*})-\mathcal{R}(\mathbf{w}^{*})],$ (3) where $\bar{\mathbf{w}}=\frac{1}{T}\sum_{t=1}^{T}\mathbf{w}_{t}$ is the output of non-private SGD. The first term is due to the added noise $\mathbf{b}$, which can be estimated by the Chernoff bound for Gaussian random vectors. The second term is the generalization error of SGD, which can be handled by the stability analysis. The third term is an optimization error and can be controlled by standard techniques in optimization theory. Finally, the last term can be bounded by $\mathcal{O}(1/\sqrt{n})$ by Hoeffding inequality. The proof of Theorem 9 is given in Subsection 3.2. Now, we turn our attention to the utility guarantee for the case with a bounded domain. ###### Theorem 10 (Utility guarantees for bounded domain). If the loss function $\ell$ is nonnegative, convex and $\alpha$-Hölder smooth with parameter $L$. Let $\mathbf{w}_{\text{priv}}$ be the output produced by Algorithm 1 with $\mathcal{W}\subseteq\mathcal{B}(0,R)$. Let $T\asymp n^{\frac{2-\alpha}{1+\alpha}}$ if $\alpha<\frac{1}{2}$, $T\asymp n$ else, and choose $\eta=1/\Big{(}T\max\Big{\\{}\frac{\sqrt{\log(n/\delta)\log(n)\log(1/\gamma)}}{\sqrt{n}},\frac{\big{(}d\log(1/\delta)\big{)}^{1/4}\sqrt{\log(n/\delta)}(\log(1/\gamma))^{1/8}}{\sqrt{n\epsilon}}\Big{\\}}\Big{)}$. Then for any $\gamma\in(4\max\\{\exp(-d/8),\delta\\},1)$, with probability at least $1-\gamma$ over the randomness in both the sample and the algorithm, there holds $\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})=\|\mathbf{w}^{*}\|_{2}^{2}\cdot\mathcal{O}\bigg{(}\frac{\big{(}d\log(1/\delta)\big{)}^{\frac{1}{4}}(\log(1/\gamma))^{\frac{1}{8}}{\sqrt{\log(n/\delta)}}}{\sqrt{n\epsilon}}+\frac{\sqrt{\log(n)\log(1/\gamma){\log(n/\delta)}}}{\sqrt{n}}\bigg{)}.$ The definition of $\alpha$-Hölder smoothness and the convexity of $\ell$ imply the following inequalities $\|\partial\ell(\mathbf{w};z)\|_{2}\leq M+LR^{\alpha}\text{ and }\ell(\mathbf{w};z)\leq\ell(0;z)+MR+LR^{1+\alpha},\quad\forall z\in\mathcal{Z},\mathbf{w}\in\mathcal{W}.$ These together with Theorem 8 and Theorem 9 imply the privacy and utility guarantees in the above theorem. The detailed proof is given in Subsection 3.2. ###### Remark 2. The private SGD algorithm with output perturbation was studied in [35] under both the Lipschitz continuity and the strong smoothness assumption, where the excess population risk for one-pass private SGD (i.e. the total iteration number $T=n$) with a bounded parameter domain was bounded by $\mathcal{O}\big{(}(n\epsilon)^{-\frac{1}{2}}(d\log(1/\delta)^{\frac{1}{4}}\big{)}$. As a comparison, we show that the same rate (up to a logarithmic factor) $\mathcal{O}\big{(}(n\epsilon)^{-\frac{1}{2}}(d\log(1/\delta))^{\frac{1}{4}}\log^{\frac{1}{2}}(n/\delta)\big{)}$ can be achieved for general $\alpha$-Hölder smooth losses by taking $T=\mathcal{O}(n^{2-\alpha\over 1+\alpha}+n).$ Our results extend the output perturbation for private SGD algorithms to a more general class of non-smooth losses. #### 2.2.3 Differentially Private SGD with Gradient Perturbation 1: Inputs: Data $S=\\{z_{i}\in\mathcal{Z}:i=1,\ldots,n\\}$, loss function $\ell(\mathbf{w},z)$ with Hölder parameters $\alpha$ and $L$, the convex set $\mathcal{W}\subseteq\mathcal{B}(0,R)$, step size $\eta$, number of iterations $T$, privacy parameters $\epsilon$, $\delta$, and constant $\beta$. 2: Set: $\mathbf{w}_{1}=\mathbf{0}$ 3: Compute $\sigma^{2}=\frac{14(M+LR^{\alpha})^{2}T}{\beta n^{2}\epsilon}\Big{(}\frac{\log(1/\delta)}{(1-\beta)\epsilon}+1\Big{)}$ 4: for $t=1$ to $T$ do 5: Sample $i_{t}\sim\text{Unif}([n])$ 6: $\mathbf{w}_{t+1}=\text{Proj}_{\mathcal{W}}\big{(}\mathbf{w}_{t}-\eta(\partial\ell(\mathbf{w}_{t};z_{i_{t}})+\mathbf{b}_{t})\big{)}$, where $\mathbf{b}_{t}\sim\mathcal{N}(0,\sigma^{2}\mathbf{I}_{d})$ 7: end for 8: return: ${\mathbf{w}}_{\text{priv}}=\frac{1}{T}\sum_{t=1}^{T}\mathbf{w}_{t}$ Algorithm 2 Differentially Private SGD with Gradient perturbation (DP-SGD- Gradient) An alternative approach to achieve $(\epsilon,\delta)$-DP is gradient perturbation, i.e., adding Gaussian noise to the stochastic gradient at each update. The detailed algorithm is described in Algorithm 2, whose privacy guarantee is established in the following theorem. ###### Theorem 11 (Privacy guarantee). Suppose the loss function $\ell$ is nonnegative, convex and $\alpha$-Hölder smooth with parameter $L$. Then Algorithm 2 (DP-SGD-Gradient) satisfies $(\epsilon,\delta)$-DP if there exists $\beta\in(0,1)$ such that $\frac{\sigma^{2}}{4(M+LR^{\alpha})^{2}}\geq 0.67$ and $\lambda-1\leq\frac{\sigma^{2}}{6(M+LR^{\alpha})^{2}}\log\Big{(}\frac{n}{\lambda(1+\frac{\sigma^{2}}{4(M+LR^{\alpha})^{2}})}\Big{)}$ hold with $\lambda=\frac{\log(1/\delta)}{(1-\beta)\epsilon}+1$. Since $\mathcal{W}\subseteq\mathcal{B}(0,R)$, the Hölder smoothness of $\ell$ implies that $\|\partial\ell(\mathbf{w}_{t},z)\|_{2}\leq M+LR^{\alpha}$ for any $t\in[T]$ and any $z\in\mathcal{Z}$, from which we know that the $\ell_{2}$-sensitivity of the function $\mathcal{M}_{t}=\partial\ell(\mathbf{w}_{t},z)$ can be bounded by $2(M+LR^{\alpha})$. By Lemma 3 and the post-processing property of DP, it is easy to show that the update of $\mathbf{w}_{t}$ satisfies $(\frac{\log(1/\delta)}{(1-\beta)\epsilon}+1,\frac{\beta\epsilon}{T})$-RDP for any $t\in[T]$. Furthermore, by the composition theorem of RDP and the relationship between $(\epsilon,\delta)$-DP and RDP, we can show that the proposed algorithm satisfies $(\epsilon,\delta)$-DP. The detailed proof can be found in Subsection 3.3. Other than the privacy guarantees, the DP-SGD-Gradient algorithm also enjoys utility guarantees as stated in the following theorem. ###### Theorem 12 (Utility guarantee). Suppose the loss function $\ell$ is nonnegative, convex and $\alpha$-Hölder smooth with parameter $L$. Let $\mathbf{w}_{\text{priv}}$ be the output produced by Algorithm 2 with $\eta=\frac{1}{T}\max\big{\\{}\frac{\sqrt{\log(n)\log(n/\gamma)\log(1/\gamma)}}{\sqrt{n}},\frac{\sqrt{d\log(1/\delta)}(\log(1/\gamma))^{\frac{1}{4}}}{n\epsilon}\big{\\}}$. Furthermore, let $T\asymp n^{\frac{2-\alpha}{1+\alpha}}$ if $\alpha<\frac{1}{2}$, and $T\asymp n$ else. Then, for any $\gamma\in(18\exp(-Td/8),1)$, with probability at least $1-\gamma$ over the randomness in both the sample and the algorithm, there holds $\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})=\|\mathbf{w}^{*}\|_{2}^{2}\cdot\mathcal{O}\bigg{(}\frac{\sqrt{d\log(1/\delta)\log(1/\gamma)}}{n\epsilon}+\frac{\sqrt{\log(n){\log(n/\gamma)}\log(1/\gamma)}}{\sqrt{n}}\bigg{)}.$ Our basic idea to prove Theorem 12 is to use the following error decomposition: $\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})=[\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}_{S}(\mathbf{w}_{\text{priv}})]+[\mathcal{R}_{S}(\mathbf{w}_{\text{priv}})-\mathcal{R}_{S}(\mathbf{w}^{*})]+[\mathcal{R}_{S}(\mathbf{w}^{*})-\mathcal{R}(\mathbf{w}^{*})].$ (4) Similar to the proof of Theorem 9, the generalization error $\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}_{S}(\mathbf{w}_{\text{priv}})$ can be handled by the UAS bound, the optimization error $\mathcal{R}_{S}(\mathbf{w}_{\text{priv}})-\mathcal{R}_{S}(\mathbf{w}^{*})$ can be estimated by standard techniques in optimization [[, e.g.]]Nem, and the last term $\mathcal{R}_{S}(\mathbf{w}^{*})-\mathcal{R}(\mathbf{w}^{*})$ can be bounded by the Hoeffding inequality. The detailed proof can be found in Subsection 3.3. ###### Remark 3. We now compare our results with the related work under a bounded domain assumption. The work [4] established the optimal rate $\mathcal{O}(\frac{1}{n\epsilon}{\sqrt{d\log(1/\delta)}}+\frac{1}{\sqrt{n}})$ for the excess population risk of private SCO algorithm in either smooth case ($\alpha=1$) or non-smooth case ($\alpha=0$). However, their algorithm has a large gradient complexity $\mathcal{O}\Big{(}n^{4.5}\sqrt{\epsilon}+\frac{n^{6.5}\epsilon^{4.5}}{(d\log(\frac{1}{\delta}))^{2}}\Big{)}$. The work [16] proposed a private phased ERM algorithm for SCO, which can achieve the optimal excess population risk for non-smooth losses with a better gradient complexity of the order $\mathcal{O}(n^{2}\log(1/{\delta}))$. The very recent work [3] improved the gradient complexity to $\mathcal{O}(n^{2})$. As a comparison, we show that SGD with gradient complexity $\mathcal{O}(n^{2-\alpha\over 1+\alpha}+n)$ is able to achieve the optimal (up to logarithmic terms) excess population risk $\mathcal{O}(\frac{1}{n\epsilon}{\sqrt{d\log(1/\delta)}}+\frac{1}{\sqrt{n}})$ for general $\alpha$-Hölder smooth losses. Our results match the existing gradient complexity for both the smooth case in [4] and the Lipschitz continuity case [3]. An interesting observation is that our algorithm can achieve the optimal utility guarantee with the linear gradient complexity $\mathcal{O}(n)$ for $\alpha\geq 1/2$, which shows that a relaxation of the strong smoothness from $\alpha=1$ to $\alpha\geq 1/2$ does not bring any harm in both the generalization and computation complexity. Now, we give a sufficient condition for the existence of $\beta$ in Theorem 11 under a specific parameter setting. ###### Lemma 13. Let $n\geq 18$, $T=n$ and $\delta=1/{n^{2}}$. If $\epsilon\geq\frac{7(n^{\frac{1}{3}}-1)+4\log(n)n+7}{2n(n^{\frac{1}{3}}-1)},$ then there exists $\beta\in(0,1)$ such that Algorithm 2 satisfies $(\epsilon,\delta)$-DP. Figure 1: The sufficient condition for the existence of $\beta$ in Lemma 13. The shaded area is the area where the sufficient condition in Lemma 13 holds true, i.e., $\epsilon\geq\big{(}7(n^{\frac{1}{3}}-1)+4\log(n)n+7\big{)}/\big{(}2n(n^{\frac{1}{3}}-1)\big{)}$. ###### Remark 4. Privacy parameters $\epsilon$ and $\delta$ together quantify the privacy risk. $\epsilon$ is often called the privacy budget controlling the degree of privacy leakage. A larger value of $\epsilon$ implies higher privacy risk. Therefore, the value of $\epsilon$ depends on how much privacy the user needs to protect. Theoretically, the value of $\epsilon$ is less than 1. However, in practice, to obtain the desired utility, a larger privacy budget, i.e., $\epsilon\geq 1$, is always acceptable [35, 33]. For instance, Apple uses a privacy budget $\epsilon=8$ for Safari Auto-play intent detection, and $\epsilon=2$ for Health types111https://www.apple.com/privacy/docs/Differential_Privacy_Overview.pdf. Parameter $\delta$ is the probability with which $e^{\epsilon}$ fails to bound the ratio between the two probabilities in the definition of differential privacy, i.e., the probability of privacy protection failure. For meaningful privacy guarantees, according to [14] the value of $\delta$ should be much smaller than $1/n$. In particular, we always choose $\delta=1/n^{2}$. For DP- SGD-Gradient algorithm, another constant we should discuss is $\beta$ which depends on the choice of the number of iterations $T$, size of training data $n$, privacy parameters $\epsilon$ and $\delta$. The appearance of this parameter is due to the use of subsampling result for RDP (see Lemma 3). The condition in Lemma 13 ensures the existence of $\beta\in(0,1)$ such that Algorithm 2 satisfies DP. In practical applications, we search in $(0,1)$ for all $\beta$ that satisfy the RDP conditions in Theorem 11. Note that the closer the $\beta$ is to $1/2$, the smaller the variance of the noise added to the algorithm in each iteration. Therefore, we choose the value that is closest to $1/2$ of all $\beta$ that meets the RDP conditions as the value of $\beta$. We end this section with a final remark on the challenges of proving DP for Algorithm 2 when $\mathcal{W}$ is unbounded. ###### Remark 5. To make Algorithm 2 satisfy DP when $\mathcal{W}=\mathbb{R}^{d}$, the variance $\sigma_{t}$ of the noise $\mathbf{b}_{t}$ added in the $t$-th iteration should be proportional to the $\ell_{2}$-sensitivity $\Delta_{t}=\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})-\partial\ell(\mathbf{w}_{t},z^{\prime}_{i_{t}})\|_{2}$. The definition of Hölder smoothness implies that $\Delta_{t}\leq 2(M+L\|\mathbf{w}_{t}\|^{\alpha}_{2})$. When $\alpha=0$, we have $\Delta_{t}\leq 2(M+L)$ and the privacy guarantee can be established in a way similar to Theorem 11. When $\alpha\in(0,1]$, we have to establish an upper bound of $\|\mathbf{w}_{t}\|_{2}$. Since $\mathbf{w}_{t}=\mathbf{w}_{t-1}-\eta(\partial\ell(\mathbf{w}_{t-1},z_{i_{t-1}})+\mathbf{b}_{t-1})$ ($\mathbf{b}_{t-1}\sim\mathcal{N}(0,\sigma_{t-1}^{2}\mathbf{I}_{d})$), we can only give a bound of $\|\mathbf{w}_{t}\|_{2}$ with high probability. Thus, the sensitivity $\Delta_{t}$ can not be uniformly bounded in this case. Therefore, the first challenge is how to analyze the privacy guarantee when the sensitivity changes at each iteration and all of them can not be uniformly bounded. Furthermore, by using the property of the Gaussian vector, we can prove that $\|\mathbf{w}_{t}\|_{2}=\mathcal{O}(\sqrt{t\eta}+\eta\sum_{j=1}^{t-1}\sigma_{j}+\eta\sqrt{d\sum_{j=1}^{t-1}\sigma_{j}^{2}})$ with high probability. However, as mentioned above, the variance $\sigma_{t}$ should be proportional to $\Delta_{t}$ whose upper bound involves $\|\mathbf{w}_{t}\|^{\alpha}_{2}$. Thus, $\sigma_{t}$ is proportional to $(t\eta)^{\alpha/2}+\eta^{\alpha}(\sum_{j=1}^{t-1}\sigma_{j})^{\alpha}+\eta^{\alpha}(d\sum_{j=1}^{t-1}\sigma_{j}^{2})^{\alpha/2}.$ For this reason, it seems difficult to give a clear expression for an upper bound of $\|\mathbf{w}_{t}\|_{2}.$ ## 3 Proofs of Main Results Before presenting the detailed proof, we first introduce some useful lemmas on the concentration behavior of random variables. ###### Lemma 14 (Chernoff bound for Bernoulli variable [34]). Let $X_{1},\ldots,X_{k}$ be independent random variables taking values in $\\{0,1\\}$. Let $X=\sum_{i=1}^{k}X_{i}$ and $\mu=\mathbb{E}[X]$. The following statements hold. 1. (a) For any $\tilde{\gamma}\in(0,1)$, with probability at least $1-\exp\big{(}-\mu\tilde{\gamma}^{2}/3\big{)}$, there holds $X\leq(1+\tilde{\gamma})\mu$. 2. (b) For any $\tilde{\gamma}\geq 1$, with probability at least $1-\exp\big{(}-\mu\tilde{\gamma}/3\big{)}$, there holds $X\leq(1+\tilde{\gamma})\mu$. ###### Lemma 15 (Chernoff bound for the $\ell_{2}$-norm of Gaussian vector [34]). Let $X_{1},\ldots,X_{k}$ be i.i.d. standard Gaussian random variables, and $\mathbf{X}=[X_{1},\ldots,X_{k}]\in\mathbb{R}^{k}$. Then for any $t\in(0,1)$, with probability at least $1-\exp(-kt^{2}/8)$, there holds $\|\mathbf{X}\|_{2}^{2}\leq k(1+t).$ ###### Lemma 16 (Hoeffding inequality [18]). Let $X_{1},\ldots,X_{k}$ be independent random variables such that $a_{i}\leq X_{i}\leq b_{i}$ with probability 1 for all $i\in[k]$. Let ${X}=\frac{1}{k}\sum_{i=1}^{k}X_{i}$. Then for any $t>0$, with probability at least $1-\exp(-2t^{2}/\sum_{i}(b_{i}-a_{i})^{2})$, there holds ${X}-\mathbb{E}[{X}]\leq t.$ ###### Lemma 17 (Azuma-Hoeffding inequality [18]). Let $X_{1},\ldots,X_{k}$ be a sequence of random variables where $X_{i}$ may depend on the previous random variables $X_{1},\ldots,X_{i-1}$ for all $i=1,\ldots,k$. Consider a sequence of functionals $\xi_{i}(X_{1},\ldots,X_{i})$, $i\in[k]$. If $|\xi_{i}-\mathbb{E}_{X_{i}}[\xi_{i}]|\leq b_{i}$ for each $i$. Then for all $t>0$, with probability at least $1-\exp(-t^{2}/(2\sum_{i}b_{i}^{2}))$, there holds $\sum_{i=1}^{k}\xi_{i}-\sum_{i=1}^{k}\mathbb{E}_{X_{i}}[\xi_{i}]\leq t$. ###### Lemma 18 (Tail bound of sub-Gaussian variable [34]). Let $X$ be a sub-Gaussian random variable with mean $\mu$ and sub-Gaussian parameter $v^{2}$. Then, for any $t\geq 0$, we have, with probability at least $1-\exp\big{(}-t^{2}/(2v^{2})\big{)}$, that $X-\mu\leq t$. ### 3.1 Proofs on UAS bound of SGD on Non-smooth Losses Our stability analysis for unbounded domain requires the following lemma on the self-bounding property for Hölder smooth losses. ###### Lemma 19. ([21, 37]) Suppose the loss function $\ell$ is nonnegative, convex and $\alpha$-Hölder smooth with parameter $L$. Then for $c_{\alpha,1}$ defined as (2) we have $\|\partial\ell(\mathbf{w},z)\|_{2}\leq c_{\alpha,1}\ell^{\frac{\alpha}{1+\alpha}}(\mathbf{w},z),\quad\forall\mathbf{w}\in\mathbb{R}^{d},z\in\mathcal{Z}.$ Based on Lemma 19, we develop the following bound on the iterates produced by the SGD update (1) which is critical to analyze the privacy and utility guarantees in the case of unbounded domain. Recall that $M=\sup_{z\in\mathcal{Z}}\|\partial\ell(0,z)\|_{2}$. ###### Lemma 20. Suppose the loss function $\ell$ is nonnegative, convex and $\alpha$-Hölder smooth with parameter $L$. Let $\\{\mathbf{w}_{t}\\}_{t=1}^{T}$ be the sequence produced by SGD with $T$ iterations when $\mathcal{W}=\mathbb{R}^{d}$ and $\eta_{t}<\min\\{1,1/L\\}$. Then, for any $t\in[T]$, there holds $\|\mathbf{w}_{t+1}\|_{2}^{2}\leq C_{\alpha}\sum_{j=1}^{t}\eta_{j},$ where $C_{\alpha}=\frac{1-\alpha}{1+\alpha}c^{\frac{2(1+\alpha)}{1-\alpha}}_{\alpha,1}\big{(}\frac{\alpha}{1+\alpha}\big{)}^{\frac{2\alpha}{1-\alpha}}+2\sup_{z\in\mathcal{Z}}\ell(0;z)$. ###### Proof. The update rule $\mathbf{w}_{t+1}=\mathbf{w}_{t}-\eta_{t}\partial\ell(\mathbf{w}_{t},z_{i_{t}})$ implies that $\displaystyle\|\mathbf{w}_{t+1}\|_{2}^{2}$ $\displaystyle=\|\mathbf{w}_{t}-\eta_{t}\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}=\|\mathbf{w}_{t}\|_{2}^{2}+\eta_{t}^{2}\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}-2\eta_{t}\langle\mathbf{w}_{t},\partial\ell(\mathbf{w}_{t},z_{i_{t}})\rangle.$ (5) First, we consider the case $\alpha=0$. By the definition of Hölder smoothness, we know $\ell$ is $(M+L)$-Lipschitz continuous. Furthermore, by the convexity of $\ell$, we have $\displaystyle\eta_{t}\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}-2\langle\mathbf{w}_{t},\partial\ell(\mathbf{w}_{t},z_{i_{t}})$ $\displaystyle\leq\eta_{t}\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}+2\big{(}\ell(0,z_{i_{t}})-\ell(\mathbf{w}_{t},z_{i_{t}})\big{)}$ $\displaystyle\leq(M+L)^{2}+2\sup_{z\in\mathcal{Z}}\ell(0,z),$ where in the last inequality we have used $\eta_{t}<1$ and the nonnegativity of $\ell$. Now, putting the above inequality back into (5) and taking the summation gives $\|\mathbf{w}_{t+1}\|_{2}^{2}\leq\big{(}(M+L)^{2}+2\sup_{z\in\mathcal{Z}}\ell(0;z)\big{)}\sum_{j=1}^{t}\eta_{j}.$ (6) Then, we consider the case $\alpha=1$. In this case, Lemma 19 implies $\|\partial\ell(\mathbf{w};z)\|_{2}^{2}\leq 2L\ell(\mathbf{w};z)$. Therefore, $\displaystyle\eta_{t}\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}-2\langle\mathbf{w}_{t},\partial\ell(\mathbf{w}_{t},z_{i_{t}})\rangle\leq 2\eta_{t}L\ell(\mathbf{w}_{t},z_{i_{t}})+2\ell(0,z_{i_{t}})-2\ell(\mathbf{w}_{t},z_{i_{t}})\leq 2\ell(0,z_{i_{t}}),$ where we have used the convexity of $\ell$ and $\eta_{t}<1/L$. Plugging the above inequality back into (5) and taking the summation yield that $\|\mathbf{w}_{t+1}\|_{2}^{2}\leq 2\sup_{z\in\mathcal{Z}}\ell(0,z)\sum_{j=1}^{t}\eta_{j}.$ (7) Finally, we consider the case $\alpha\in(0,1)$. According to the self-bounding property and the convexity, we know $\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}\leq c_{\alpha,1}\ell^{\frac{\alpha}{1+\alpha}}(\mathbf{w}_{t},z_{i_{t}})\leq c_{\alpha,1}\big{(}\langle\mathbf{w}_{t},\partial\ell(\mathbf{w}_{t},z_{i_{t}})\rangle+\ell(0,z_{i_{t}})\big{)}^{\frac{\alpha}{1+\alpha}}.$ Therefore, for $\alpha\in(0,1)$ there holds $\displaystyle\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}$ $\displaystyle\leq c^{2}_{\alpha,1}\big{(}\langle\mathbf{w}_{t},\partial\ell(\mathbf{w}_{t},z_{i_{t}})\rangle+\ell(0,z_{i_{t}})\big{)}^{\frac{2\alpha}{1+\alpha}}$ $\displaystyle=\Big{(}\frac{1+\alpha}{\alpha\eta_{t}}\big{(}\langle\mathbf{w}_{t},\partial\ell(\mathbf{w}_{t},z_{i_{t}})\rangle+\ell(0,z_{i_{t}})\big{)}\Big{)}^{\frac{2\alpha}{1+\alpha}}\cdot\Big{(}c^{2}_{\alpha,1}\big{(}\frac{1+\alpha}{\alpha\eta_{t}}\big{)}^{-\frac{2\alpha}{1+\alpha}}\Big{)}$ $\displaystyle\leq\frac{2\alpha}{1+\alpha}\Big{(}\frac{1+\alpha}{\alpha\eta_{t}}\big{(}\langle\mathbf{w}_{t},\partial\ell(\mathbf{w}_{t},z_{i_{t}})\rangle+\ell(0,z_{i_{t}})\big{)}\Big{)}+\frac{1-\alpha}{1+\alpha}\Big{(}c^{2}_{\alpha,1}\big{(}\frac{1+\alpha}{\alpha\eta_{t}}\big{)}^{-\frac{2\alpha}{1+\alpha}}\Big{)}^{\frac{1+\alpha}{1-\alpha}}$ $\displaystyle=2\eta_{t}^{-1}\big{(}\langle\mathbf{w}_{t},\partial\ell(\mathbf{w}_{t},z_{i_{t}})\rangle+\ell(0,z_{i_{t}})\big{)}+\frac{1-\alpha}{1+\alpha}c^{\frac{2(1+\alpha)}{1-\alpha}}_{\alpha,1}\big{(}\frac{\alpha}{1+\alpha}\big{)}^{\frac{2\alpha}{1-\alpha}}\eta_{t}^{\frac{2\alpha}{1-\alpha}},$ where the last inequality used Young’s inequality $ab\leq\frac{1}{p}a^{p}+\frac{1}{q}b^{q}$ with $\frac{1}{p}+\frac{1}{q}=1.$ Putting the above inequality into (5), we have $\|\mathbf{w}_{t+1}\|_{2}^{2}\leq\|\mathbf{w}_{t}\|_{2}^{2}+\frac{1-\alpha}{1+\alpha}c^{\frac{2(1+\alpha)}{1-\alpha}}_{\alpha,1}\big{(}\frac{\alpha}{1+\alpha}\big{)}^{\frac{2\alpha}{1-\alpha}}\eta_{t}^{\frac{2}{1-\alpha}}+2\ell(0,z_{i_{t}})\eta_{t},$ If the step size $\eta_{t}<1$, then $\|\mathbf{w}_{t+1}\|_{2}^{2}\leq\|\mathbf{w}_{t}\|_{2}^{2}+\bigg{(}\frac{1-\alpha}{1+\alpha}c^{\frac{2(1+\alpha)}{1-\alpha}}_{\alpha,1}\big{(}\frac{\alpha}{1+\alpha}\big{)}^{\frac{2\alpha}{1-\alpha}}+2\sup_{z\in\mathcal{Z}}\ell(0;z)\bigg{)}\eta_{t}.$ Taking a summation of the above inequality, we get $\|\mathbf{w}_{t+1}\|_{2}^{2}\leq\Big{(}\frac{1-\alpha}{1+\alpha}c^{\frac{2(1+\alpha)}{1-\alpha}}_{\alpha,1}\big{(}\frac{\alpha}{1+\alpha}\big{)}^{\frac{2\alpha}{1-\alpha}}+2\sup_{z\in\mathcal{Z}}\ell(0;z)\Big{)}\sum_{j=1}^{t}\eta_{j}.$ (8) The desired result follows directly from (6), (7) and (8) for different values of $\alpha.$ ∎ The following lemma shows the approximately non-expensive behavior of the gradient mapping $\mathbf{w}\mapsto\mathbf{w}-\eta\partial\ell(\mathbf{w},z)$. The case $\alpha\in[0,1)$ can be found in Lei and Ying [21], and the case $\alpha=1$ can be found in Hardt [17]. ###### Lemma 21. Suppose the loss function $\ell$ is convex and $\alpha$-Hölder smooth with parameter $L$. Then for all $\mathbf{w},\mathbf{w}^{\prime}\in\mathbb{R}^{d}$ and $\eta\leq 2/L$ there holds $\|\mathbf{w}-\eta\partial\ell(\mathbf{w},z)-\mathbf{w}^{\prime}+\eta\partial\ell(\mathbf{w}^{\prime},z)\|_{2}^{2}\leq\|\mathbf{w}-\mathbf{w}^{\prime}\|_{2}^{2}+\frac{1-\alpha}{1+\alpha}(2^{-\alpha}L)^{\frac{2}{1-\alpha}}\eta^{\frac{2}{1-\alpha}}.$ With the above preparation, we are now ready to prove Theorem 7. ###### Proof of Theorem 7. (a) Assume that $S$ and $S^{\prime}$ differ by the $i$-th datum, i.e., $z_{i}\neq z^{\prime}_{i}.$ Let $\\{\mathbf{w}_{t}\\}_{t=1}^{T}$ and $\\{\mathbf{w}^{\prime}_{t}\\}_{t=1}^{T}$ be the sequence produced by SGD update (1) based on $S$ and $S^{\prime}$, respectively. For simplicity, let $c^{2}_{\alpha,2}=\frac{1-\alpha}{1+\alpha}(2^{-\alpha}L)^{\frac{2}{1-\alpha}}$. Note that when $\mathcal{W}=\mathbb{R}^{d}$, Eq. (1) reduces to $\mathbf{w}_{t+1}=\mathbf{w}_{t}-\eta\partial\ell(\mathbf{w}_{t},z_{i_{t}})$. For any $t\in[T]$, we consider the following two cases. Case 1: If $i_{t}\neq i$, Lemma 21 implies that $\|\mathbf{w}_{t+1}-\mathbf{w}^{\prime}_{t+1}\|_{2}^{2}=\|\mathbf{w}_{t}-\eta_{t}\partial\ell(\mathbf{w}_{t},z_{i_{t}})-\mathbf{w}^{\prime}_{t}+\eta_{t}\partial\ell(\mathbf{w}^{\prime}_{t},z_{i_{t}})\|_{2}^{2}\leq\|\mathbf{w}_{t}-\mathbf{w}^{\prime}_{t}\|_{2}^{2}+c^{2}_{\alpha,2}\eta_{t}^{\frac{2}{1-\alpha}}.$ Case 2: If $i_{t}=i$, it follows from the elementary inequality $(a+b)^{2}\leq(1+p)a^{2}+(1+1/p)b^{2}$ that $\displaystyle\|\mathbf{w}_{t+1}-\mathbf{w}^{\prime}_{t+1}\|_{2}^{2}$ $\displaystyle=\|\mathbf{w}_{t}-\eta_{t}\partial\ell(\mathbf{w}_{t},z_{i})-\mathbf{w}^{\prime}_{t}+\eta_{t}\partial\ell(\mathbf{w}^{\prime}_{t},z^{\prime}_{i})\|_{2}^{2}$ $\displaystyle\leq(1+p)\|\mathbf{w}_{t}-\mathbf{w}^{\prime}_{t}\|_{2}^{2}+(1+1/p)\eta_{t}^{2}\|\partial\ell(\mathbf{w}^{\prime}_{t},z^{\prime}_{i})-\partial\ell(\mathbf{w}_{t},z_{i})\|_{2}^{2}.$ According to the definition of Hölder smoothness and Lemma 20, we know $\|\partial\ell(\mathbf{w}_{t},z)\|_{2}\leq M+L\Big{(}C_{\alpha}\sum_{j=1}^{t-1}\eta_{j}\Big{)}^{\frac{\alpha}{2}}:=c_{\alpha,t}.$ (9) Combining the above two cases and (9) together, we have $\|\mathbf{w}_{t+1}-\mathbf{w}^{\prime}_{t+1}\|_{2}^{2}\leq(1+p)^{\mathbb{I}_{[i_{t}=i]}}\|\mathbf{w}_{t}-\mathbf{w}^{\prime}_{t}\|_{2}^{2}+c^{2}_{\alpha,2}\eta_{t}^{\frac{2}{1-\alpha}}+4(1+1/p)\mathbb{I}_{[i_{t}=i]}c^{2}_{\alpha,t}\eta_{t}^{2},$ where $\mathbb{I}_{[i_{t}=i]}$ is the indicator function, i.e., $\mathbb{I}_{[i_{t}=i]}=1$ if $i_{t}=i$ and $0$ otherwise. Applying the above inequality recursively, we get $\|\mathbf{w}_{t+1}-\mathbf{w}^{\prime}_{t+1}\|_{2}^{2}\leq\prod_{k=1}^{t}(1+p)^{\mathbb{I}_{[i_{k}=i]}}\|\mathbf{w}_{1}-\mathbf{w}^{\prime}_{1}\|_{2}^{2}+\Big{(}c^{2}_{\alpha,2}\sum_{k=1}^{t}\eta_{k}^{\frac{2}{1-\alpha}}+4\sum_{k=1}^{t}c_{\alpha,k}^{2}\eta_{k}^{2}(1+1/p){\mathbb{I}_{[i_{k}=i]}}\Big{)}\prod_{j=k+1}^{t}(1+p)^{\mathbb{I}_{[i_{j}=i]}}.$ Since $\mathbf{w}_{1}=\mathbf{w}_{1}^{\prime}$ and $\eta_{t}=\eta$, we further get $\displaystyle\|\mathbf{w}_{t+1}-\mathbf{w}^{\prime}_{t+1}\|_{2}^{2}$ $\displaystyle\leq\prod_{j=2}^{t}(1+p)^{\mathbb{I}_{[i_{j}=i]}}\Big{(}c^{2}_{\alpha,2}t\eta^{\frac{2}{1-\alpha}}+4\eta^{2}\sum_{k=1}^{t}c^{2}_{\alpha,k}(1+1/p){\mathbb{I}_{[i_{k}=i]}}\Big{)}$ $\displaystyle\leq(1+p)^{\sum_{j=2}^{t}\mathbb{I}_{[i_{j}=i]}}\Big{(}c^{2}_{\alpha,2}t\eta^{\frac{2}{1-\alpha}}+4c^{2}_{\alpha,t}\eta^{2}(1+1/p)\sum_{k=1}^{t}\mathbb{I}_{[i_{k}=i]}\Big{)}.$ (10) Applying Lemma 14 with $X_{j}=\mathbb{I}_{[i_{j}=i]}$ and $X=\sum_{j=1}^{t}X_{j}$, for any $\exp(-t/3n)\leq\gamma\leq 1$, with probability at least $1-\frac{\gamma}{n}$, there holds $\sum_{j=1}^{t}\mathbb{I}_{[i_{j}=i]}\leq\frac{t}{n}\Big{(}1+\frac{\sqrt{3\log(1/\gamma)}}{\sqrt{t/n}}\Big{)}.$ For any $0<\gamma<\exp(-t/3n)$, with probability at least $1-\frac{\gamma}{n}$, there holds $\sum_{j=1}^{t}\mathbb{I}_{[i_{j}=i]}\leq\frac{t}{n}\Big{(}1+\frac{3\log(1/\gamma)}{t/n}\Big{)}.$ Plug the above two inequalities back into (3.1), and let $c_{\gamma,t}=\max\Big{\\{}\sqrt{\frac{3\log(n/{\gamma})}{t/n}},\frac{3\log(n/{\gamma})}{t/n}\Big{\\}}$. Then, for any $\gamma\in(0,1)$, with probability at least $1-\frac{\gamma}{n}$, we have $\|\mathbf{w}_{t+1}-\mathbf{w}^{\prime}_{t+1}\|_{2}^{2}\leq(1+p)^{\frac{t}{n}(1+c_{\gamma,t})}\Big{(}c^{2}_{\alpha,2}t\eta^{\frac{2}{1-\alpha}}+4c^{2}_{\alpha,t}\eta^{2}(1+1/p)\frac{t}{n}(1+c_{\gamma,t})\Big{)}.$ Let $p=\frac{1}{\frac{t}{n}(1+c_{\gamma,t})}$. Then we know $(1+p)^{\frac{t}{n}(1+c_{\gamma,t})}\leq e$ and therefore $\displaystyle\|\mathbf{w}_{t+1}-\mathbf{w}^{\prime}_{t+1}\|_{2}^{2}\leq e\Big{(}c^{2}_{\alpha,2}t\eta^{\frac{2}{1-\alpha}}+4c^{2}_{\alpha,t}\eta^{2}\Big{(}1+\frac{t}{n}(1+c_{\gamma,t})\Big{)}\frac{t}{n}(1+c_{\gamma,t})\Big{)}.$ (11) This together with the inequality $c^{2}_{\alpha,t}\leq\big{(}M+L(C_{\alpha}t\eta)^{\frac{\alpha}{2}}\big{)}^{2}$ due to Lemma 20, we have, with probability at least $1-\frac{\gamma}{n}$, that $\|\mathbf{w}_{t+1}-\mathbf{w}^{\prime}_{t+1}\|_{2}^{2}\leq e\Big{(}c^{2}_{\alpha,2}t\eta^{\frac{2}{1-\alpha}}+4\big{(}M+L(C_{\alpha}t\eta)^{\frac{\alpha}{2}}\big{)}^{2}\eta^{2}\Big{(}1+\frac{t}{n}(1+c_{\gamma,t})\Big{)}\frac{t}{n}(1+c_{\gamma,t})\Big{)}.$ By taking a union bound of probabilities over $i=1,\ldots,n$, with probability at least $1-\gamma$, there holds $\sup_{S\simeq S^{\prime}}\|\mathbf{w}_{t+1}-\mathbf{w}^{\prime}_{t+1}\|_{2}^{2}\leq e\Big{(}c^{2}_{\alpha,2}t\eta^{\frac{2}{1-\alpha}}+4\big{(}M+L(C_{\alpha}t\eta)^{\frac{\alpha}{2}}\big{)}^{2}\eta^{2}\Big{(}1+\frac{t}{n}(1+c_{\gamma,t})\Big{)}\frac{t}{n}(1+c_{\gamma,t})\Big{)}.$ Let $\Delta_{SGD}(\gamma)=\Big{(}e\big{(}c^{2}_{\alpha,2}T\eta^{\frac{2}{1-\alpha}}+4\big{(}M+L(C_{\alpha}T\eta)^{\frac{\alpha}{2}}\big{)}^{2}\eta^{2}\Big{(}1+\frac{T}{n}(1+c_{\gamma,T})\Big{)}\frac{T}{n}(1+c_{\gamma,T})\big{)}\Big{)}^{1/2}$. Recall that $\mathcal{A}$ is the SGD with $T$ iterations, and $\bar{\mathbf{w}}=\frac{1}{T}\sum_{t=1}^{T}\mathbf{w}_{t}$ is the output produced by $\mathcal{A}$. Hence, $\sup_{S\simeq S^{\prime}}\delta_{\mathcal{A}}(S,S^{\prime})=\sup_{S\simeq S^{\prime}}\|\bar{\mathbf{w}}-\bar{\mathbf{w}}^{\prime}\|_{2}$. By the convexity of the $\ell_{2}$-norm, with probability at least $1-\gamma$, we have $\sup_{S\simeq S^{\prime}}\delta_{\mathcal{A}}(S,S^{\prime})\leq\frac{1}{T}\sum_{t=1}^{T}\sup_{S\simeq S^{\prime}}\|\mathbf{w}_{t}-\mathbf{w}_{t}^{\prime}\|_{2}\leq\Delta_{SGD}(\gamma).$ This completes the proof of part (a). (b) For the case $\mathcal{W}\subseteq\mathcal{B}(0,R)$, the analysis is similar to the case $\mathcal{W}=\mathbb{R}^{d}$ except using a different estimate for the term $\|\partial\ell(\mathbf{w}_{t},z)\|_{2}$. Indeed, in this case we have $\|\mathbf{w}_{t}\|_{2}\leq R$, which together with the Hölder smoothness, implies $\|\partial\ell(\mathbf{w}_{t},z)\|_{2}\leq M+LR^{\alpha}$ for any $t\in[T]$ and $z\in\mathcal{Z}$. Now, replacing $c_{\alpha,t}=M+LR^{\alpha}$ in (9) and putting $c_{\alpha,t}$ back into (11), with probability at least $1-\frac{\gamma}{n}$, we obtain $\sup_{S\simeq S^{\prime}}\|\mathbf{w}_{t+1}-\mathbf{w}^{\prime}_{t+1}\|_{2}^{2}\leq e\Big{(}c^{2}_{\alpha,2}t\eta^{\frac{2}{1-\alpha}}+4\big{(}M+LR^{\alpha}\big{)}^{2}\eta^{2}\Big{(}1+\frac{t}{n}(1+c_{\gamma,t})\Big{)}\frac{t}{n}(1+c_{\gamma,t})\Big{)}.$ Now, let $\tilde{\Delta}_{SGD}(\gamma)=\Big{(}e\big{(}c^{2}_{\alpha,2}T\eta^{\frac{2}{1-\alpha}}+4\big{(}M+LR^{\alpha}\big{)}^{2}\eta^{2}\Big{(}1+\frac{T}{n}(1+c_{\gamma,T})\Big{)}\frac{T}{n}(1+c_{\gamma,T})\big{)}\Big{)}^{1/2}.$ The convexity of a norm implies, with probability at least $1-\gamma$, that $\sup_{S\simeq S^{\prime}}\delta_{\mathcal{A}}(S,S^{\prime})\leq\frac{1}{T}\sum_{t=1}^{T}\sup_{S\simeq S^{\prime}}\|\mathbf{w}_{t}-\mathbf{w}_{t}^{\prime}\|_{2}\leq\tilde{\Delta}_{SGD}(\gamma).$ The proof of the theorem is completed. ∎ ### 3.2 Proofs on Differentially Private SGD with Output Perturbation In this subsection, we prove the privacy and utility guarantees for output perturbation (i.e. Algorithm 1). We consider both the unbounded domain $\mathcal{W}=\mathbb{R}^{d}$ and bounded domain $\mathcal{W}\subseteq\mathcal{B}(0,R)$. We first prove Theorem 8 on the privacy guarantee of Algorithm 1. ###### Proof of Theorem 8. Let $\mathcal{A}$ be the SGD with $T$ iterations, $\bar{\mathbf{w}}=\frac{1}{T}\sum_{t=1}^{T}\mathbf{w}_{t}$ be the output of $\mathcal{A}$. First, consider the unbounded domain case, i.e., $\mathcal{W}=\mathbb{R}^{d}$. Let $I=\\{i_{1},\ldots,i_{T}\\}$ be the sequence of sampling after $T$ iterations in $\mathcal{A}$. Define $\mathcal{B}=\big{\\{}I:\sup_{S\simeq S^{\prime}}\delta_{\mathcal{A}}(S,S^{\prime})\leq\Delta_{\text{SGD}}(\delta/2)\big{\\}}.$ Part (a) in Theorem 7 implies that $\mathbb{P}(I\in\mathcal{B})\geq 1-\delta/2$. Further, according to the definitions, we know the $\ell_{2}$-sensitivity of $\mathcal{A}$ is identical to the UAS of $\mathcal{A}$. Thus, if $I\in\mathcal{B}$, then Lemma 2 with $\delta^{\prime}=\delta/2$ implies Algorithm 1 satisfies $(\epsilon,\delta/2)$-DP. For any neighboring datasets $S$ and $S^{\prime}$, let $\mathbf{w}_{\text{priv}}$ and $\mathbf{w}^{\prime}_{\text{priv}}$ be the output produced by Algorithm 1 based on $S$ and $S^{\prime}$, respectively. Hence, for any $E\subseteq\mathbb{R}^{d}$ we have $\displaystyle\mathbb{P}(\mathbf{w}_{\text{priv}}\in E)$ $\displaystyle=\mathbb{P}(\mathbf{w}_{\text{priv}}\in E\cap I\in\mathcal{B})+\mathbb{P}(\mathbf{w}_{\text{priv}}\in E\cap I\in\mathcal{B}^{c})$ $\displaystyle\leq\mathbb{P}(\mathbf{w}_{\text{priv}}\in E|I\in\mathcal{B})\mathbb{P}(I\in\mathcal{B})+\frac{\delta}{2}\leq\Big{(}e^{\epsilon}\mathbb{P}(\mathbf{w}_{\text{priv}}^{\prime}\in E|I\in\mathcal{B})+\frac{\delta}{2}\Big{)}\mathbb{P}(I\in\mathcal{B})+\frac{\delta}{2}$ $\displaystyle\leq e^{\epsilon}\mathbb{P}(\mathbf{w}_{\text{priv}}^{\prime}\in E\cap I\in\mathcal{B})+\delta\leq e^{\epsilon}\mathbb{P}(\mathbf{w}_{\text{priv}}^{\prime}\in E)+\delta,$ where in the second inequality we have used the definition of DP. Therefore, Algorithm 1 satisfies $(\epsilon,\delta)$-DP when $\mathcal{W}=\mathbb{R}^{d}$. The bounded domain case can be proved in a similar way by using part (b) of Theorem 7. The proof is completed. ∎ Now, we turn to the utility guarantees of Algorithm 1. Recall that the excess population risk $\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})$ can be decomposed as follows ($\bar{\mathbf{w}}=\frac{1}{T}\sum_{t=1}^{T}\mathbf{w}_{t}$) $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})$ $\displaystyle=[\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\bar{\mathbf{w}})]+[\mathcal{R}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\bar{\mathbf{w}})]+[\mathcal{R}_{S}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\mathbf{w}^{*})]+[\mathcal{R}_{S}(\mathbf{w}^{*})-\mathcal{R}(\mathbf{w}^{*})].$ (12) We now introduce three lemmas to control the first three terms on the right hand side of (12). The following lemma controls the error resulting from the added noise. ###### Lemma 22. Suppose the loss function $\ell$ is nonnegative, convex and $\alpha$-Hölder smooth with parameter $L$. Let $\mathbf{w}_{\text{priv}}$ be the output produced by Algorithm 1 based on the dataset $S=\\{z_{1},\cdots,z_{n}\\}$ with $\eta_{t}=\eta<\min\\{1,1/L\\}$. Then for any $\gamma\in(4\exp(-d/8),1)$, the following statements hold true. 1. (a) If $\mathcal{W}=\mathbb{R}^{d}$, then, with probability at least $1-\frac{\gamma}{4}$, there holds $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\bar{\mathbf{w}})=\mathcal{O}\Big{(}(T\eta)^{\frac{\alpha}{2}}\sigma\sqrt{d}(\log(1/\gamma))^{\frac{1}{4}}+\sigma^{1+\alpha}d^{\frac{1+\alpha}{2}}(\log(1/\gamma))^{\frac{1+\alpha}{4}}\Big{)}.$ 2. (b) If $\mathcal{W}\subseteq\mathcal{B}(0,R)$ with $R>0$, then, with probability at least $1-\frac{\gamma}{4}$, we have $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\bar{\mathbf{w}})=\mathcal{O}\Big{(}\sigma\sqrt{d}(\log(1/\gamma))^{\frac{1}{4}}+\sigma^{1+\alpha}d^{\frac{1+\alpha}{2}}(\log(1/\gamma))^{\frac{1+\alpha}{4}}\Big{)}.$ ###### Proof. (a) First, we consider the case $\mathcal{W}=\mathbb{R}^{d}$. Note that $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\bar{\mathbf{w}})$ $\displaystyle=\mathbb{E}_{z}[\ell(\mathbf{w}_{\text{priv}},z)-\ell(\bar{\mathbf{w}},z)]\leq\mathbb{E}_{z}[\langle\partial\ell(\mathbf{w}_{\text{priv}},z),\mathbf{w}_{\text{priv}}-\bar{\mathbf{w}}\rangle]$ $\displaystyle\leq\mathbb{E}_{z}[\|\partial\ell(\mathbf{w}_{\text{priv}},z)\|_{2}\|\mathbf{b}\|_{2}]\leq(M+L\|\mathbf{w}_{\text{priv}}\|_{2}^{\alpha})\|\mathbf{b}\|_{2}$ $\displaystyle\leq(M+L\|\bar{\mathbf{w}}\|^{\alpha}_{2})\|\mathbf{b}\|_{2}+L\|\mathbf{b}\|_{2}^{1+\alpha},$ (13) where the first inequality is due to the convexity of $\ell$, the second inequality follows from the Cauchy-Schwartz inequality, the third inequality is due to the definition of Hölder smoothness, and the last inequality uses $\mathbf{w}_{\text{priv}}=\bar{\mathbf{w}}+\mathbf{b}$. Hence, to estimate $\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\bar{\mathbf{w}})$, it suffices to bound $\|\mathbf{b}\|_{2}$ and $\|\bar{\mathbf{w}}\|_{2}$. Since $\mathbf{b}\sim\mathcal{N}(0,\sigma^{2}\mathbf{I})$, then for any $\gamma\in(\exp(-d/8),1)$, Lemma 15 implies, with probability at least $1-\frac{\gamma}{4}$, that $\|\mathbf{b}\|_{2}\leq\sigma\sqrt{d}\Big{(}1+\Big{(}\frac{8}{d}\log\big{(}4/\gamma\big{)}\Big{)}^{\frac{1}{4}}\Big{)}.$ (14) Further, by the convexity of a norm and Lemma 20, we know $\|\bar{\mathbf{w}}\|_{2}\leq\frac{1}{T}\sum_{t=1}^{T}\|\mathbf{w}_{t}\|_{2}\leq\big{(}C_{\alpha}T\eta\big{)}^{\frac{1}{2}}.$ (15) Putting the above inequality and (14) back into (3.2) yields $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\bar{\mathbf{w}})$ $\displaystyle\leq\big{(}M+L(C_{\alpha}T\eta)^{\frac{\alpha}{2}}\big{)}\sigma\sqrt{d}\Big{(}1+\Big{(}\frac{8}{d}\log\big{(}4/\gamma\big{)}\Big{)}^{\frac{1}{4}}\Big{)}+L\sigma^{1+\alpha}d^{\frac{1+\alpha}{2}}\Big{(}1+\Big{(}\frac{8}{d}\log\big{(}4/\gamma\big{)}\Big{)}^{\frac{1}{4}}\Big{)}^{1+\alpha}$ $\displaystyle=\mathcal{O}\Big{(}(T\eta)^{\frac{\alpha}{2}}\sigma\sqrt{d}\big{(}\log(1/\gamma)\big{)}^{\frac{1}{4}}+\sigma^{1+\alpha}d^{\frac{1+\alpha}{2}}\big{(}\log(1/\gamma)\big{)}^{\frac{1+\alpha}{4}}\Big{)}.$ This completes the proof of part (a). (b) The proof for the unbounded domain case is similar to that of the bounded domain. Since $\|\mathbf{w}_{t}\|_{2}\leq R$ for $t\in[T]$ in this case, then $\|\bar{\mathbf{w}}\|_{2}\leq\frac{1}{T}\sum_{t=1}^{T}\|\mathbf{w}_{t}\|_{2}\leq R.$ (16) Plugging (16) and (14) back into (3.2) yield the result in part (b). ∎ In the following lemma, we use the stability of SGD to control the generalization error $\mathcal{R}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\bar{\mathbf{w}})$. ###### Lemma 23. Suppose the loss function $\ell$ is nonnegative, convex, and $\alpha$-Hölder smooth with parameter $L$. Let $\mathcal{A}$ be the SGD with $T$ iterations and $\eta_{t}=\eta<\min\\{1,1/L\\}$ based on the dataset $S=\\{z_{1},\cdots,z_{n}\\}$, and $\bar{\mathbf{w}}=\frac{1}{T}\sum_{t=1}^{T}\mathbf{w}_{t}$ be the output produced by $\mathcal{A}$. Then for any $\gamma\in(4\delta,1)$, the following statements hold true. 1. (a) If $\mathcal{W}=\mathbb{R}^{d}$, then, with probability at least $1-\frac{\gamma}{4}$, there holds $\displaystyle\mathcal{R}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\bar{\mathbf{w}})=\mathcal{O}\Big{(}(T\eta)^{\frac{\alpha}{2}}\Delta_{\text{SGD}}(\delta/2)\log(n)\log(1/\gamma)+(T\eta)^{\frac{1+\alpha}{2}}\sqrt{n^{-\frac{1}{2}}\log(1/\gamma)}\Big{)}.$ 2. (b) If $\mathcal{W}\subseteq\mathcal{B}(0,R)$ with $R>0$, then, with probability at least $1-\frac{\gamma}{4}$, we have $\displaystyle\mathcal{R}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\bar{\mathbf{w}})=\mathcal{O}\Big{(}\tilde{\Delta}_{\text{SGD}}(\delta/2)\log(n)\log(1/\gamma)+\sqrt{n^{-\frac{1}{2}}\log(1/\gamma)}\Big{)}.$ ###### Proof. (a) Consider the unbounded domain case. Part (a) in Theorem 7 implies, with probability at least $1-\frac{\delta}{2}$, that $\sup_{S\simeq S^{\prime}}\delta_{\mathcal{A}}(S,S^{\prime})\leq\Delta_{\text{SGD}}(\delta/2).$ (17) Since $\gamma\geq 4\delta$, then we know (17) holds with probability at least $1-\frac{\gamma}{8}$. According to the result $\|\bar{\mathbf{w}}\|_{2}\leq\sqrt{C_{\alpha}T\eta}$ by (15) and Lemma 1 with $G=\sqrt{C_{\alpha}T\eta}$ together, we derive the following inequality with probability at least $1-\frac{\gamma}{8}-\frac{\gamma}{8}=1-\frac{\gamma}{4}$ $\displaystyle\mathcal{R}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\bar{\mathbf{w}})$ $\displaystyle\leq c\bigg{(}(M+L(C_{\alpha}T\eta)^{\frac{\alpha}{2}})\Delta_{\text{SGD}}(\delta/2)\log(n)\log({8}/{\gamma})+\big{(}\sup_{z\in\mathcal{Z}}\ell(0,z)+(M+L(T\eta)^{\frac{\alpha}{2}})\sqrt{T\eta}\big{)}\sqrt{\frac{\log({8}/{\gamma})}{n}}\bigg{)}$ $\displaystyle=\mathcal{O}\bigg{(}(T\eta)^{\frac{\alpha}{2}}\Delta_{\text{SGD}}(\delta/2)\log(n)\log(1/\gamma)+(T\eta)^{\frac{1+\alpha}{2}}\sqrt{\frac{\log(1/\gamma)}{n}}\bigg{)},$ where $c>0$ is a constant. The proof of part (a) is completed. (b) For the case $\mathcal{W}\subseteq\mathcal{B}(0,R)$, the proof follows a similar argument as part (a). Indeed, part (b) in Theorem 7 implies, with probability at least $1-\frac{\gamma}{8}$, that $\sup_{S\simeq S^{\prime}}\delta_{\mathcal{A}}(S,S^{\prime})\leq\tilde{\Delta}_{\text{SGD}}(\delta/2).$ (18) Note that $\|\bar{\mathbf{w}}\|_{2}\leq R$ in this case, then combining (18) and Lemma 1 with $G=R$ together, with probability at least $1-\frac{\gamma}{4}$, we have $\displaystyle\mathcal{R}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\bar{\mathbf{w}})$ $\displaystyle\leq c\bigg{(}(M+LR^{\alpha})\tilde{\Delta}_{\text{SGD}}(\delta/2)\log(n)\log({8}/{\gamma})+\big{(}\sup_{z\in\mathcal{Z}}\ell(0,z)+(M+LR^{\alpha})R\big{)}\sqrt{\frac{\log({8}/{\gamma})}{n}}\bigg{)}$ $\displaystyle=\mathcal{O}\bigg{(}\tilde{\Delta}_{\text{SGD}}(\delta/2)\log(n)\log(1/\gamma)+\sqrt{\frac{\log(1/\gamma)}{n}}\bigg{)},$ where $c>0$ is a constant. This completes the proof of part (b). ∎ In the following lemma, we use techniques in optimization theory to control the optimization error $\mathcal{R}_{S}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\mathbf{w}^{*})$. ###### Lemma 24. Suppose the loss function $\ell$ is nonnegative, convex and $\alpha$-Hölder smooth with parameter $L$. Let $\mathcal{A}$ be the SGD with $T$ iterations and $\eta_{t}=\eta<\min\\{1,1/L\\}$ based on the dataset $S=\\{z_{1},\cdots,z_{n}\\}$, and $\bar{\mathbf{w}}=\frac{1}{T}\sum_{t=1}^{T}\mathbf{w}_{t}$ be the output produced by $\mathcal{A}$. Then, for any $\gamma\in(0,1)$, the following statements hold true. 1. (a) If $\mathcal{W}=\mathbb{R}^{d}$, then, with probability at least $1-\frac{\gamma}{4}$, there holds $\mathcal{R}_{S}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\mathbf{w}^{*})=\mathcal{O}\bigg{(}\eta^{\frac{1+\alpha}{2}}T^{\frac{\alpha}{2}}\sqrt{\log(1/\gamma)}+\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{T}}+\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{\eta T}+\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\eta\bigg{)}.$ 2. (b) If $\mathcal{W}\subseteq\mathcal{B}(0,R)$ with $R>0$, then, with probability at least $1-\frac{\gamma}{4}$, we have $\mathcal{R}_{S}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\mathbf{w}^{*})=\mathcal{O}\bigg{(}\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{T}}+\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{\eta T}+\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\eta\bigg{)}.$ ###### Proof. (a) We first consider the case $\mathcal{W}=\mathbb{R}^{d}$. From the convexity of $\ell$, we have $\displaystyle\mathcal{R}_{S}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\mathbf{w}^{*})$ $\displaystyle\leq\frac{1}{T}\sum_{t=1}^{T}\mathcal{R}_{S}(\mathbf{w}_{t})-\mathcal{R}_{S}(\mathbf{w}^{*})$ $\displaystyle=\frac{1}{T}\sum_{t=1}^{T}[\mathcal{R}_{S}(\mathbf{w}_{t})-\ell(\mathbf{w}_{t},z_{i_{t}})]+\frac{1}{T}\sum_{t=1}^{T}[\ell(\mathbf{w}^{*},z_{i_{t}})-\mathcal{R}_{S}(\mathbf{w}^{*})]+\frac{1}{T}\sum_{t=1}^{T}[\ell(\mathbf{w}_{t},z_{i_{t}})-\ell(\mathbf{w}^{*},z_{i_{t}})].$ (19) First, we consider the upper bound of $\frac{1}{T}\sum_{t=1}^{T}[\mathcal{R}_{S}(\mathbf{w}_{t})-\ell(\mathbf{w}_{t};z_{i_{t}})]$. Since $\\{z_{i_{t}}\\}$ is uniformly sampled from the dataset $S$, then for all $t=1,\ldots,T$ we obtain $\mathbb{E}_{z_{i_{t}}}[\ell(\mathbf{w}_{t},z_{i_{t}})|\mathbf{w}_{1},...,\mathbf{w}_{t-1}]=\mathcal{R}_{S}(\mathbf{w}_{t}).$ By the convexity of $\ell$, the definition of Hölder smoothness and Lemma 20, for any $z\in\mathcal{Z}$ and all $t\in[T]$, there holds $\displaystyle\ell(\mathbf{w}_{t},z)$ $\displaystyle\leq\sup_{z}\ell(0,z)+\langle\partial\ell(\mathbf{w}_{t},z),\mathbf{w}_{t}\rangle\leq\sup_{z}\ell(0,z)+\|\partial\ell(\mathbf{w}_{t},z)\|_{2}\|\mathbf{w}_{t}\|_{2}$ $\displaystyle\leq\sup_{z}\ell(0,z)+(M+L\|\mathbf{w}_{t}\|_{2}^{\alpha})\|\mathbf{w}_{t}\|_{2}\leq\sup_{z}\ell(0,z)+M(C_{\alpha}T\eta)^{\frac{1}{2}}+L(C_{\alpha}T\eta)^{\frac{1+\alpha}{2}}.$ (20) Similarly, for any $z\in\mathcal{Z}$, we have $\displaystyle\ell(\mathbf{w}^{*},z)\leq\sup_{z}\ell(0;z)+M\|\mathbf{w}^{*}\|_{2}+L\|\mathbf{w}^{*}\|_{2}^{1+\alpha}.$ (21) Now, combining Lemma 17 with (3.2) and noting $\eta>1/T$, we get the following inequality with probability at least $1-\frac{\gamma}{8}$ $\displaystyle\frac{1}{T}\sum_{t=1}^{T}[\mathcal{R}_{S}(\mathbf{w}_{t})-\ell(\mathbf{w}_{t},z_{i_{t}})]$ $\displaystyle\leq\big{(}\sup_{z}\ell(0,z)+M(C_{\alpha}T\eta)^{\frac{1}{2}}+L(C_{\alpha}T\eta)^{\frac{1+\alpha}{2}}\big{)}\sqrt{\frac{2\log(\frac{8}{\gamma})}{T}}=\mathcal{O}\Big{(}\eta^{\frac{1+\alpha}{2}}T^{\frac{\alpha}{2}}\sqrt{\log(1/\gamma)}\Big{)}.$ (22) According to Lemma 16, with probability at least $1-\frac{\gamma}{8}$, there holds $\displaystyle\frac{1}{T}\sum_{t=1}^{T}[\ell(\mathbf{w}^{*};z_{i_{t}})-\mathcal{R}_{S}(\mathbf{w}^{*})]$ $\displaystyle\leq\big{(}\sup_{z}\ell(0,z)+M\|\mathbf{w}^{*}\|_{2}+L\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\big{)}\sqrt{\frac{\log({8}/{\gamma})}{2T}}=\mathcal{O}\Big{(}\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{T}}\Big{)}.$ (23) Finally, we consider the term $\frac{1}{T}\sum_{t=1}^{T}[\ell(\mathbf{w}_{t},z_{i_{t}})-\ell(\mathbf{w}^{*},z_{t})]$. The update rule implies $\mathbf{w}_{t+1}-\mathbf{w}^{*}=\big{(}\mathbf{w}_{t}-\mathbf{w}^{*}\big{)}-\eta\partial\ell(\mathbf{w}_{t},z_{i_{t}})$, from which we know $\displaystyle\|\mathbf{w}_{t+1}-\mathbf{w}^{*}\|_{2}^{2}$ $\displaystyle=\|\big{(}\mathbf{w}_{t}-\mathbf{w}^{*}\big{)}-\eta\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}$ $\displaystyle=\|\mathbf{w}_{t}-\mathbf{w}^{*}\|_{2}^{2}+\eta^{2}\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}-2\eta\langle\partial\ell(\mathbf{w}_{t},z_{i_{t}}),\mathbf{w}_{t}-\mathbf{w}^{*}\rangle.$ It then follows that $\langle\partial\ell(\mathbf{w}_{t},z_{i_{t}}),\mathbf{w}_{t}-\mathbf{w}^{*}\rangle=\frac{1}{2\eta}\big{(}\|\mathbf{w}_{t}-\mathbf{w}^{*}\|_{2}^{2}-\|\mathbf{w}_{t+1}-\mathbf{w}^{*}\|_{2}^{2}\big{)}+\frac{\eta}{2}\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}.$ Combining the above inequality and the convexity of $\ell$ together, we derive $\displaystyle\frac{1}{T}\sum_{t=1}^{T}[\ell(\mathbf{w}_{t},z_{i_{t}})-\ell(\mathbf{w}^{*},z_{i_{t}})]$ $\displaystyle\leq\frac{1}{T}\sum_{t=1}^{T}\Bigl{[}\frac{1}{2\eta}\big{(}\|\mathbf{w}_{t}-\mathbf{w}^{*}\|_{2}^{2}-\|\mathbf{w}_{t+1}-\mathbf{w}^{*}\|_{2}^{2}\big{)}+\frac{\eta}{2}\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}\Bigr{]}$ $\displaystyle\leq\frac{1}{2T\eta}\|\mathbf{w}_{1}-\mathbf{w}^{*}\|_{2}^{2}+\frac{\eta}{2T}\sum_{t=1}^{T}\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}.$ (24) Since $0\leq\frac{2\alpha}{1+\alpha}\leq 1$, Lemma 19 implies the following inequality for any $t=1,\ldots,T$ $\|\partial\ell(\mathbf{w}_{t};z_{i_{t}})\|_{2}^{2}\leq c_{\alpha,1}\ell^{\frac{2\alpha}{1+\alpha}}(\mathbf{w}_{t};z_{i_{t}})\leq c_{\alpha,1}\max\\{\ell(\mathbf{w}_{t};z_{i_{t}}),1\\}\leq c_{\alpha,1}\ell(\mathbf{w}_{t};z_{i_{t}})+c_{\alpha,1}.$ Putting $\|\partial\ell(\mathbf{w}_{t};z_{i_{t}})\|^{2}_{2}\leq c_{\alpha,1}\ell(\mathbf{w}_{t};z_{i_{t}})+c_{\alpha,1}$ back into (3.2) and noting $\|\mathbf{w}_{1}\|_{2}=0$, we have $\displaystyle\frac{1}{T}\sum_{t=1}^{T}[\ell(\mathbf{w}_{t},z_{i_{t}})-\ell(\mathbf{w}^{*},z_{i_{t}})]$ $\displaystyle\leq\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{2\eta T}+\frac{c_{\alpha,1}\eta}{2T}\sum_{t=1}^{T}\ell(\mathbf{w}_{t},z_{i_{t}})+\frac{c_{\alpha,1}\eta}{2}.$ Rearranging the above inequality and using (21), we derive $\displaystyle\frac{1}{T}\sum_{t=1}^{T}[\ell(\mathbf{w}_{t},z_{i_{t}})-\ell(\mathbf{w}^{*},z_{i_{t}})]$ $\displaystyle\leq\frac{1}{1-\frac{c_{\alpha,1}\eta}{2}}\Bigl{(}\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{2\eta T}+\frac{c_{\alpha,1}\eta}{2T}\sum_{t=1}^{T}\ell(\mathbf{w}^{*},z_{i_{t}})+\frac{c_{\alpha,1}\eta}{2}\Bigr{)}$ $\displaystyle=\mathcal{O}\Big{(}\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{\eta T}+\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\eta\Big{)}.$ (25) Now, plugging (22), (23) and (3.2) back into (3.2), we derive $\mathcal{R}_{S}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\mathbf{w}^{*})=\mathcal{O}\Big{(}\eta^{\frac{1+\alpha}{2}}T^{\frac{\alpha}{2}}\sqrt{\log(1/\gamma)}+\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{T}}+\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{\eta T}+\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\eta\Big{)}$ with probability at least $1-\frac{\gamma}{4}$, which completes the proof of part (a). (b) Consider the bounded domain case. Since $\|\mathbf{w}_{t}\|_{2}\leq R$ for any $t\in[T]$, then by the convexity of $\ell$ and the definition of Hölder smoothness, for any $z\in\mathcal{Z}$, there holds $\ell(\mathbf{w}_{t},z)\leq\sup_{z}\ell(0,z)+(M+LR^{\alpha})R.$ Combining the above inequality and Lemma 17 together, with probability at least $1-\frac{\gamma}{8}$, we obtain $\displaystyle\frac{1}{T}\sum_{t=1}^{T}[\mathcal{R}_{S}(\mathbf{w}_{t})-\ell(\mathbf{w}_{t},z_{i_{t}})]$ $\displaystyle\leq\big{(}\sup_{z}\ell(0,z)+(M+LR^{\alpha})R\big{)}\sqrt{\frac{2\log(\frac{8}{\gamma})}{T}}=\mathcal{O}\Big{(}\sqrt{\frac{\log(1/\gamma)}{T}}\Big{)}.$ (26) Since $\|\mathbf{w}_{t+1}-\mathbf{w}^{*}\|^{2}_{2}=\|\text{Proj}_{\mathcal{W}}\big{(}\mathbf{w}_{t}-\eta\partial\ell(\mathbf{w}_{t},z_{i_{t}})\big{)}-\mathbf{w}^{*}\|_{2}^{2}\leq\|(\mathbf{w}_{t}-\mathbf{w}^{*})-\eta\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}$, then (3.2) also holds true in this case. Putting (26), (23) and (3.2) back into (3.2), with probability at least $1-\frac{\gamma}{4}$, we have $\mathcal{R}_{S}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\mathbf{w}^{*})=\mathcal{O}\bigg{(}\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{T}}+\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{\eta T}+\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\eta\bigg{)}.$ The proof is completed. ∎ Now, we are in a position to prove the utility guarantee for DP-SGD-Output algorithm. First, we give the proof for the unbounded domain case (i.e. Theorem 9). ###### Proof of Theorem 9. Note that $\mathcal{R}_{S}(\mathbf{w}^{*})-\mathcal{R}(\mathbf{w}^{*})=\mathcal{R}_{S}(\mathbf{w}^{*})-\mathbb{E}_{S}[\mathcal{R}_{S}(\mathbf{w}^{*})]$. By Hoeffding inequality and (21), with probability at least $1-\frac{\gamma}{4}$, there holds $\displaystyle\mathcal{R}_{S}(\mathbf{w}^{*})-\mathcal{R}(\mathbf{w}^{*})\leq\Big{(}\sup_{z\in\mathcal{Z}}\ell(0,z)+M\|\mathbf{w}^{*}\|_{2}+L\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\Big{)}\sqrt{\frac{\log({4}/{\gamma})}{2n}}=\mathcal{O}\Big{(}\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{n}}\Big{)}.$ (27) Combining part (a) in Lemmas 22, 23, 24 and (27) together, with probability at least $1-\gamma$, the population excess risk can be bounded as follows $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})$ $\displaystyle=[\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\bar{\mathbf{w}})]+[\mathcal{R}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\bar{\mathbf{w}})]+[\mathcal{R}_{S}(\bar{\mathbf{w}})-\mathcal{R}_{S}(\mathbf{w}^{*})]+[\mathcal{R}_{S}(\mathbf{w}^{*})-\mathcal{R}(\mathbf{w}^{*})]$ $\displaystyle=\mathcal{O}\bigg{(}(T\eta)^{\frac{\alpha}{2}}\sigma\sqrt{d}\big{(}\log(1/\gamma)\big{)}^{\frac{1}{4}}+\sigma^{1+\alpha}d^{\frac{1+\alpha}{2}}(\log(1/\gamma))^{\frac{1+\alpha}{4}}+(T\eta)^{\frac{\alpha}{2}}\Delta_{\text{SGD}}(\delta/2)\log(n)\log(1/\gamma)+\eta^{\frac{1+\alpha}{2}}\Big{(}T^{\frac{1+\alpha}{2}}\sqrt{\frac{\log(1/\gamma)}{n}}$ $\displaystyle\qquad+T^{\frac{\alpha}{2}}\sqrt{\log(1/\gamma)}\Big{)}+\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{\eta T}+\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\eta+\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{n}}\bigg{)}.$ (28) Plugging $\Delta_{\text{SGD}}(\delta/2)=\mathcal{O}\Big{(}\sqrt{T}\eta^{\frac{1}{1-\alpha}}+\frac{(T\eta)^{1+\frac{\alpha}{2}}\log(n/\delta)}{n}\Big{)}$ and $\sigma=\mathcal{O}(\frac{\sqrt{\log(1/\delta)}\Delta_{\text{SGD}}(\delta/2)}{\epsilon})$ back into (3.2), we have $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})$ $\displaystyle=\mathcal{O}\bigg{(}T^{\frac{1+\alpha}{2}}\sqrt{\frac{\log(1/\gamma)}{n}}\eta^{\frac{1+\alpha}{2}}+\frac{T^{1+\alpha}\sqrt{d\log(1/\delta)}\big{(}\log(1/\gamma)\big{)}^{\frac{1}{4}}\log(n/\delta)}{n\epsilon}\eta^{1+\alpha}+\frac{d^{\frac{1+\alpha}{2}}(\log(1/\gamma))^{\frac{1+\alpha}{4}}\big{(}T\log(\frac{1}{\delta})\big{)}^{\frac{1+\alpha}{2}}}{\epsilon^{1+\alpha}}\eta^{\frac{1+\alpha}{1-\alpha}}$ $\displaystyle\qquad+\frac{\big{(}d\log(1/\delta)\big{)}^{\frac{1+\alpha}{2}}(\log(1/\gamma))^{\frac{1+\alpha}{4}}T^{(1+\frac{\alpha}{2})(1+\alpha)}\big{(}\log(n/\delta)\big{)}^{1+\alpha}}{(n\epsilon)^{1+\alpha}}\eta^{(1+\frac{\alpha}{2})(1+\alpha)}+\frac{T^{\frac{1+\alpha}{2}}\sqrt{d\log(\frac{1}{\delta})}\big{(}\log(1/\gamma)\big{)}^{\frac{1}{4}}}{\epsilon}\eta^{\frac{2+\alpha-\alpha^{2}}{2(1-\alpha)}}$ $\displaystyle\qquad+T^{\frac{1+\alpha}{2}}\log(n)\log(1/\gamma)\eta^{\frac{2+\alpha-\alpha^{2}}{2(1-\alpha)}}+\frac{T^{1+\alpha}\log(n/\delta)\log(n)\log(1/\gamma)}{n}\eta^{1+\alpha}+\frac{1}{\eta T}+\eta+\sqrt{\frac{\log(1/\gamma)}{n}}\bigg{)}\cdot\|\mathbf{w}^{*}\|_{2}^{2}.$ (29) Taking the derivative of $\frac{1}{T\eta}+T^{\frac{1+\alpha}{2}}\sqrt{\frac{\log(1/\gamma)}{n}}\eta^{\frac{1+\alpha}{2}}$ w.r.t $\eta$ and setting it to $0$, then we have $\eta=n^{\frac{1}{3+\alpha}}/\big{(}T(\log(1/\gamma))^{\frac{1}{3+\alpha}}\big{)}$. Putting this $\eta$ back into (3.2), we obtain $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})$ $\displaystyle=\mathcal{O}\bigg{(}\frac{n^{\frac{(2-\alpha)(1+\alpha)}{2(1-\alpha)(3+\alpha)}}\sqrt{d\log(1/\delta)}}{T^{\frac{1+\alpha}{2(1-\alpha)}}\epsilon\big{(}\log(1/\gamma)\big{)}^{\frac{1+4\alpha-\alpha^{2}}{4(1-\alpha)(3+\alpha)}}}+\frac{\sqrt{d\log(1/\delta)}\log(n/\delta)}{n^{\frac{2}{3+\alpha}}\epsilon\big{(}\log(1/\gamma)\big{)}^{\frac{1+\alpha}{4(3+\alpha)}}}+\Big{(}\frac{\sqrt{d\log(1/\delta)}\log(n/\delta)}{n^{\frac{4+\alpha}{2(3+\alpha)}}\epsilon\big{(}\log(1/\gamma)\big{)}^{\frac{1+\alpha}{4(3+\alpha)}}}\Big{)}^{1+\alpha}$ $\displaystyle\quad+\Big{(}\frac{n^{\frac{1}{(1-\alpha)(3+\alpha)}}\sqrt{d\log(1/\delta)}}{T^{\frac{1+\alpha}{2(1-\alpha)}}\epsilon\big{(}\log(1/\gamma)\big{)}^{\frac{(1+\alpha)^{2}}{4(1-\alpha)(3+\alpha)}}}\Big{)}^{1+\alpha}+\log(n)\log(n/\delta)\big{(}\log(1/\gamma)\big{)}^{\frac{2}{3+\alpha}}\Big{(}\frac{n^{\frac{2+\alpha-\alpha^{2}}{2(3+\alpha)(1-\alpha)}}}{T^{\frac{1+\alpha}{2(1-\alpha)}}}+\frac{1}{n^{\frac{2}{3+\alpha}}}+\frac{1}{n^{\frac{1}{3+\alpha}}}+\frac{n^{\frac{1}{3+\alpha}}}{T}\Big{)}\bigg{)}\cdot\|\mathbf{w}^{*}\|_{2}^{2}.$ (30) To achieve the best rate with a minimal computational cost, we choose the smallest $T$ such that $\frac{n^{\frac{(2-\alpha)(1+\alpha)}{2(1-\alpha)(3+\alpha)}}}{T^{\frac{1+\alpha}{2(1-\alpha)}}}=\mathcal{O}(\frac{1}{n^{\frac{2}{3+\alpha}}})$, $\frac{n^{\frac{1+\alpha}{(1-\alpha)(3+\alpha)}}}{T^{\frac{(1+\alpha)^{2}}{2(1-\alpha)}}}=\mathcal{O}(\frac{1}{n^{\frac{(4+\alpha)(1+\alpha)}{2(3+\alpha)}}})$ and $\frac{n^{\frac{2+\alpha-\alpha^{2}}{2(3+\alpha)(1-\alpha)}}}{T^{\frac{1+\alpha}{2(1-\alpha)}}}+\frac{1}{n^{\frac{2}{3+\alpha}}}+\frac{n^{\frac{1}{3+\alpha}}}{T}=\mathcal{O}(\frac{1}{n^{\frac{1}{3+\alpha}}})$. Hence, we set $T\asymp n^{\frac{-\alpha^{2}-3\alpha+6}{(1+\alpha)(3+\alpha)}}$ if $0\leq\alpha\leq\frac{\sqrt{73}-7}{4}$, and $T\asymp n$ else. Now, putting the choice of $T$ back into (3.2), we derive $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})=$ $\displaystyle\mathcal{O}\bigg{(}\frac{\sqrt{d\log(1/\delta)}\log(n/\delta)}{(\log(1/\gamma))^{\frac{1+\alpha}{4(3+\alpha)}}n^{\frac{2}{3+\alpha}}\epsilon}+\Bigl{(}\frac{\sqrt{d\log(1/\delta)}\log(n/\delta)}{(\log(1/\gamma))^{\frac{1+\alpha}{4(3+\alpha)}}n^{\frac{4+\alpha}{2(3+\alpha)}}\epsilon}\Bigr{)}^{1+\alpha}$ $\displaystyle\quad+\frac{\log(n)\big{(}\log(1/\gamma)\big{)}^{\frac{2}{3+\alpha}}\log(n/\delta)}{n^{\frac{1}{3+\alpha}}}\bigg{)}\cdot\|\mathbf{w}^{*}\|_{2}^{2}.$ Without loss of generality, we assume the first term of the above utility bound is less than 1. Therefore, with probability at least $1-\gamma$, there holds $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})=\|\mathbf{w}^{*}\|_{2}^{2}\cdot\mathcal{O}\bigg{(}\frac{\sqrt{d\log(1/\delta)}\log(n/\delta)}{(\log(1/\gamma))^{\frac{1+\alpha}{4(3+\alpha)}}n^{\frac{2}{3+\alpha}}\epsilon}+\frac{\log(n)\big{(}\log(1/\gamma)\big{)}^{\frac{2}{3+\alpha}}\log(n/\delta)}{n^{\frac{1}{3+\alpha}}}\bigg{)}.$ The proof is completed. ∎ Finally, we provide the proof of utility guarantee for the DP-SGD-Output algorithm when $\mathcal{W}\subseteq\mathcal{B}(0,R)$ (i.e. Theorem 10). ###### Proof of Theorem 10. The proof is similar to that of Theorem 9. Indeed, plugging part (b) in Lemmas 22, 23, 24 and (27) back into (12), with probability at least $1-\gamma$, the population excess risk can be bounded as follows $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})$ $\displaystyle=\mathcal{O}\Big{(}\sigma\sqrt{d}(\log(1/\gamma))^{\frac{1}{4}}+\sigma^{1+\alpha}d^{\frac{1+\alpha}{2}}(\log(1/\gamma))^{\frac{1+\alpha}{4}}+\tilde{\Delta}_{\text{SGD}}(\delta/2)\log(n)\log(1/\gamma)+\sqrt{\frac{\log(1/\gamma)}{n}}$ $\displaystyle\qquad+\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{T\eta}+\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\eta+\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{n}}\Big{)}.$ Note that $\tilde{\Delta}_{\text{SGD}}(\delta/2)=\mathcal{O}(\sqrt{T}\eta^{\frac{1}{1-\alpha}}+\frac{T\eta\log(n/\delta)}{n})$ and $\sigma=\frac{\sqrt{2\log(2.5/\delta)}\tilde{\Delta}_{\text{SGD}}(\delta/2)}{\epsilon}$. Then we have $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})$ $\displaystyle=\mathcal{O}\biggl{(}\big{(}\frac{T\log(n/\delta))\log(n)\log(1/\gamma)}{n}+\frac{T\sqrt{d\log(1/\delta)}\log(n/\delta)(\log(1/\gamma))^{\frac{1}{4}}}{n\epsilon}\big{)}\eta$ $\displaystyle\qquad+\big{(}\frac{\sqrt{\log(1/\delta)Td}(\log(1/\gamma))^{\frac{1}{4}}}{\epsilon}+\sqrt{T}\log(n)\log(1/\gamma)\big{)}\eta^{\frac{1}{1-\alpha}}+\frac{(Td\log(1/\delta))^{\frac{1+\alpha}{2}}(\log(1/\gamma))^{\frac{1+\alpha}{4}}}{\epsilon^{1+\alpha}}\eta^{\frac{1+\alpha}{1-\alpha}}$ $\displaystyle\qquad+\big{(}\frac{T\sqrt{d\log(1/\delta)}\log(n/\delta)(\log(1/\gamma))^{\frac{1}{4}}}{n\epsilon}\big{)}^{1+\alpha}\eta^{1+\alpha}+\frac{1}{T\eta}+\sqrt{\frac{\log(1/\gamma)}{n}}\biggr{)}\cdot\|\mathbf{w}^{*}\|_{2}^{2}.$ (31) Consider the tradeoff between $1/\eta$ and $\eta$. Taking the derivative of $\big{(}\frac{T\log(n/\delta)\log(n)\log(1/\gamma)}{n}+\frac{T\sqrt{d\log(1/\delta)}\log(n/\delta)(\log(1/\gamma))^{1/4}}{n\epsilon}\big{)}\eta$ $+\frac{1}{T\eta}$ w.r.t $\eta$ and setting it to $0$, we have $\eta=1/\Big{(}T\max\Big{\\{}\frac{\sqrt{\log(n/\delta)\log(n)\log(1/\gamma)}}{\sqrt{n}},\frac{\big{(}d\log(1/\delta)\big{)}^{1/4}\sqrt{\log(n/\delta)}(\log(1/\gamma))^{1/8}}{\sqrt{n\epsilon}}\Big{\\}}\Big{)}$. Then putting the value of $\eta$ back into (3.2), we obtain $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})$ $\displaystyle=\mathcal{O}\bigg{(}\frac{\big{(}d\log(1/\delta)\big{)}^{\frac{1}{4}}(\log(1/\gamma))^{\frac{1}{8}}\sqrt{\log(n/\delta)}}{\sqrt{n\epsilon}}+\Bigl{(}\frac{\big{(}d\log(1/\delta)\big{)}^{\frac{1}{4}}(\log(1/\gamma))^{\frac{1}{8}}\sqrt{\log(n/\delta)}}{\sqrt{n\epsilon}}\Bigr{)}^{1+\alpha}$ $\displaystyle\quad+\frac{\big{(}d\log(1/\delta)\big{)}^{\frac{1-2\alpha}{4(1-\alpha)}}(\log(1/\gamma))^{\frac{1-2\alpha}{8(1-\alpha)}}n^{\frac{1}{2(1-\alpha)}}\epsilon^{\frac{2\alpha-1}{2(1-\alpha)}}}{T^{\frac{1+\alpha}{2(1-\alpha)}}(\log(n/\delta))^{\frac{1}{2(1-\alpha)}}}+\Big{(}\frac{\big{(}d\log(1/\delta)\big{)}^{\frac{1-2\alpha}{4(1-\alpha)}}(\log(1/\gamma))^{\frac{1-2\alpha}{8(1-\alpha)}}n^{\frac{1}{2(1-\alpha)}}\epsilon^{\frac{2\alpha-1}{2(1-\alpha)}}}{T^{\frac{1+\alpha}{2(1-\alpha)}}(\log(n/\delta))^{\frac{1}{2(1-\alpha)}}}\Big{)}^{1+\alpha}$ $\displaystyle\quad+\sqrt{\log(n)\log(1/\gamma)\log(1/\delta)}\big{(}\frac{1}{\sqrt{n}}+\frac{n^{\frac{1}{2(1-\alpha)}}}{T^{\frac{1+\alpha}{2(1-\alpha)}}}\big{)}\bigg{)}\cdot\|\mathbf{w}^{*}\|_{2}^{2}.$ Similarly, we choose the smallest $T$ such that $\frac{n^{\frac{1}{2(1-\alpha)}}}{T^{\frac{1+\alpha}{2(1-\alpha)}}}=\mathcal{O}(\frac{1}{\sqrt{n}})$. Hence, we set $T\asymp n^{\frac{2-\alpha}{1+\alpha}}$ if $\alpha<\frac{1}{2}$, and $T\asymp n$ else. Since $\frac{1}{4}\geq\frac{1-2\alpha}{2(1-\alpha)}$, we have $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})=\mathcal{O}\bigg{(}$ $\displaystyle\frac{\big{(}d\log(1/\delta)\big{)}^{\frac{1}{4}}(\log(1/\gamma))^{\frac{1}{8}}\sqrt{\log(n/\delta)}}{\sqrt{n\epsilon}}+\bigl{(}\frac{\big{(}d\log(1/\delta)\big{)}^{\frac{1}{4}}(\log(1/\gamma))^{\frac{1}{8}}\sqrt{\log(n/\delta)}}{\sqrt{n\epsilon}}\bigr{)}^{1+\alpha}$ $\displaystyle+\frac{\sqrt{\log(n)\log(1/\gamma)\log(n/\delta)}}{\sqrt{n}}\bigg{)}\cdot\|\mathbf{w}^{*}\|_{2}^{2}.$ It is reasonable to assume the first term is less than $1$ here. Therefore, with probability at least $1-\gamma$, there holds $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})=\|\mathbf{w}^{*}\|_{2}^{2}\cdot\mathcal{O}\Big{(}\frac{\big{(}d\log(1/\delta)\big{)}^{\frac{1}{4}}(\log(1/\gamma))^{\frac{1}{8}}\sqrt{\log(n/\delta)}}{\sqrt{n\epsilon}}+\frac{\sqrt{\log(n)\log(1/\gamma)\log(n/\delta)}}{\sqrt{n}}\Big{)}.$ The proof is completed. ∎ ### 3.3 Proofs on Differential Privacy of SGD with Gradient Perturbation We now turn to the analysis for DP-SGD-Gradient algorithm (i.e. Algorithm 2) and provide the proofs for Theorems 11 and 12. We start with the proof of Theorem 11 on the privacy guarantee for Algorithm 2. ###### Proof of Theorem 11. Consider the mechanism $\mathcal{G}_{t}=\mathcal{M}_{t}+\mathbf{b}_{t}$, where $\mathcal{M}_{t}=\partial\ell(\mathbf{w}_{t},z_{i_{t}})$. For any $\mathbf{w}_{t}\in\mathcal{W}$ and any $z_{i_{t}},z^{\prime}_{i_{t}}\in\mathcal{Z}$, the definition of $\alpha$-Hölder smoothness implies that $\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})-\partial\ell(\mathbf{w}_{t},z^{\prime}_{i_{t}})\|_{2}\leq 2\big{(}M+L\|\mathbf{w}_{t}\|^{\alpha}_{2}\big{)}\leq 2(M+LR^{\alpha}).$ Therefore, the $\ell_{2}$-sensitivity of $\mathcal{M}_{t}$ is $2(M+LR^{\alpha})$. Let $\sigma^{2}=\frac{14(M+LR^{\alpha})^{2}T}{\beta n^{2}\epsilon}\Big{(}\frac{\log(1/\delta)}{(1-\beta)\epsilon}+1\Big{)}.$ Lemma 3 with $p=\frac{1}{n}$ implies that $\mathcal{G}_{t}$ satisfies $\Big{(}\lambda,\frac{\lambda\beta\epsilon}{T\big{(}\frac{\log(1/\delta)}{(1-\beta)\epsilon}+1\big{)}}\Big{)}$-RDP if the following conditions hold $\displaystyle\frac{\sigma^{2}}{4(M+LR^{\alpha})^{2}}\geq 0.67$ (32) and $\displaystyle\lambda-1\leq\frac{\sigma^{2}}{6(M+LR^{\alpha})^{2}}\log\Big{(}\frac{n}{\lambda(1+\frac{\sigma^{2}}{4(M+LR^{\alpha})^{2}})}\Big{)}.$ (33) Let $\lambda=\frac{\log(1/\delta)}{(1-\beta)\epsilon}+1$. We obtain that $\mathcal{G}_{t}$ satisfies $(\frac{\log(1/\delta)}{(1-\beta)\epsilon}+1,\frac{\beta\epsilon}{T})$-RDP. Then by the post-processing property of DP (see Lemma 6), we know $\mathbf{w}_{t+1}$ also satisfies $(\frac{\log(1/\delta)}{(1-\beta)\epsilon}+1,\frac{\beta\epsilon}{T})$-RDP for any $t=0,...,T-1$. Furthermore, according to the adaptive composition theorem of RDP (see Lemma 4), Algorithm 2 satisfies $(\frac{\log(1/\delta)}{(1-\beta)\epsilon}+1,\beta\epsilon)$-RDP. Finally, by Lemma 5, the output of Algorithm 2 satisfies $(\epsilon,\delta)$-DP as long as (32) and (33) hold. ∎ Now, we turn to the generalization analysis of Algorithm 2. First, we estimate the generalization error $\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}_{S}(\mathbf{w}_{\text{priv}})$ in (4). ###### Lemma 25. Suppose the loss function $\ell$ is nonnegative, convex and $\alpha$-Hölder smooth with parameter $L$. Let $\mathbf{w}_{\text{priv}}$ be the output produced by Algorithm 2 based on $S=\\{z_{1},\cdots,z_{n}\\}$ with $\eta_{t}=\eta<\min\\{1,1/L\\}$. Then for any $\gamma\in(0,1)$, with probability at least $1-\frac{\gamma}{3}$, there holds $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}_{S}(\mathbf{w}_{\text{priv}})$ $\displaystyle=\mathcal{O}\Big{(}\tilde{\Delta}_{\text{SGD}}(\gamma/6)\log(n)\log(1/\gamma)+\sqrt{\frac{\log(1/\gamma)}{n}}\Big{)}.$ ###### Proof. Part (b) in Theorem 7 implies that $\tilde{\Delta}_{\text{SGD}}(\gamma/6)=\mathcal{O}\Big{(}\sqrt{T}\eta^{\frac{1}{1-\alpha}}+\frac{T\eta\log(n/\gamma)}{n}\Big{)}$ with probability at least $1-\frac{\gamma}{6}$. Since the noise added to the gradient in each iteration is the same for the neighboring datasets $S$ and $S^{\prime}$, the noise addition does not impact the stability analysis. Therefore, the UAS bound of the noisy SGD is equivalent to the SGD. According to Lemma 1 and $\|\mathbf{w}_{\text{priv}}\|_{2}\leq R$, we derive the following inequality with probability at least $1-(\frac{\gamma}{6}+\frac{\gamma}{6})$ $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}_{S}(\mathbf{w}_{\text{priv}})$ $\displaystyle\leq c\Big{(}(M+LR^{\alpha})\tilde{\Delta}_{\text{SGD}}(\gamma/6)\log(n)\log(6/{\gamma})+\big{(}M_{0}+(M+LR^{\alpha}\big{)}R\sqrt{\frac{\log(6/{\gamma})}{n}}\Big{)}$ $\displaystyle=\mathcal{O}\Big{(}\tilde{\Delta}_{\text{SGD}}(\gamma/6)\log(n)\log(1/\gamma)+\sqrt{\frac{\log(1/\gamma)}{n}}\Big{)},$ where $c>0$ is a constant. The proof is completed. ∎ The following lemma gives an upper bound for the second term $\mathcal{R}_{S}(\mathbf{w}_{\text{priv}})-\mathcal{R}_{S}(\mathbf{w}^{*})$ in (4). ###### Lemma 26. Suppose the loss function $\ell$ is nonnegative, convex and $\alpha$-Hölder smooth with parameter $L$. Let $\mathbf{w}_{\text{priv}}$ be the output produced by Algorithm 2 based on $S=\\{z_{1},\cdots,z_{n}\\}$ with $\eta_{t}=\eta<\min\\{1,1/L\\}$. Then, for any $\gamma\in(18\exp(-dT/8),1)$, with probability at least $1-\frac{\gamma}{3}$, there holds $\displaystyle\mathcal{R}_{S}(\mathbf{w}_{\text{priv}})-\mathcal{R}_{S}(\mathbf{w}^{*})=\mathcal{O}\Big{(}$ $\displaystyle\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{T}}+\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{T\eta}+\eta+\frac{\sqrt{\log(1/\delta)\log(1/\gamma)}(\|\mathbf{w}^{*}\|_{2}+\eta)}{n\epsilon}$ $\displaystyle+\frac{\eta Td\log(\frac{1}{\delta})\sqrt{\log(\frac{1}{\gamma})}}{n^{2}\epsilon^{2}}\Big{)}.$ ###### Proof. To estimate the term $\mathcal{R}_{S}(\mathbf{w}_{\text{priv}})-\mathcal{R}_{S}(\mathbf{w}^{*})$, we decompose it as $\displaystyle\mathcal{R}_{S}(\mathbf{w}_{\text{priv}})-\mathcal{R}_{S}(\mathbf{w}^{*})$ $\displaystyle\leq\frac{1}{T}\sum_{t=1}^{T}[\mathcal{R}_{S}(\mathbf{w}_{t})-\ell(\mathbf{w}_{t},z_{i_{t}})]+\frac{1}{T}\sum_{t=1}^{T}[\ell(\mathbf{w}^{*},z_{i_{t}})-\mathcal{R}_{S}(\mathbf{w}^{*})]+\frac{1}{T}\sum_{t=1}^{T}[\ell(\mathbf{w}_{t},z_{i_{t}})-\ell(\mathbf{w}^{*},z_{i_{t}})].$ (34) Similar to the analysis in (3.2) and (21), we have $\ell(\mathbf{w}^{*},z)=\mathcal{O}(\|\mathbf{w}^{*}\|_{2}^{1+\alpha})$ for all $z\in\mathcal{Z}$ and $\ell(\mathbf{w}_{t},z)=\mathcal{O}(R+R^{1+\alpha})$ for all $t=1,\ldots,T$ and $z\in\mathcal{Z}$. Therefore, Azuma-Hoeffding inequality (see Lemma 17) yields, with probability at least $1-\frac{\gamma}{9}$, that $\displaystyle\frac{1}{T}\sum_{t=1}^{T}[\mathcal{R}_{S}(\mathbf{w}_{t})-\ell(\mathbf{w}_{t},z_{t})]\leq\big{(}\sup_{z\in{\mathcal{Z}}}\ell(0,z)+\sup_{t=1,\ldots,T;z\in\mathcal{Z}}\ell(\mathbf{w}_{t},z)\big{)}\sqrt{\frac{\log({9}/{\gamma})}{2T}}=\mathcal{O}\Big{(}(R+R^{1+\alpha})\sqrt{\frac{\log(1/\gamma)}{T}}\Big{)}.$ (35) In addition, Hoeffding inequality (see Lemma 16) implies, with probability at least $1-\frac{\gamma}{9}$, that $\displaystyle\frac{1}{T}\sum_{t=1}^{T}[\ell(\mathbf{w}^{*},z_{i_{t}})-\mathcal{R}_{S}(\mathbf{w}^{*})]\leq(\sup_{z\in{\mathcal{Z}}}\ell(0,z)+\sup_{z\in{\mathcal{Z}}}\ell(\mathbf{w}^{*},z))\sqrt{\frac{\log({9}/{\gamma})}{2T}}=\mathcal{O}\Big{(}\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{T}}\Big{)}.$ (36) Finally, we try to bound $\frac{1}{T}\sum_{t=1}^{T}[\ell(\mathbf{w}_{t},z_{i_{t}})-\ell(\mathbf{w}^{*},z_{i_{t}})]$. The SGD update rule implies that $\|\mathbf{w}_{t+1}-\mathbf{w}^{*}\|_{2}^{2}=\|\text{Proj}_{\mathcal{W}}\big{(}\mathbf{w}_{t}-\eta(\partial\ell(\mathbf{w}_{t},z_{i_{t}})+\mathbf{b}_{t})\big{)}-\mathbf{w}^{*}\|_{2}^{2}\leq\|(\mathbf{w}_{t}-\mathbf{w}^{*})-\eta(\partial\ell(\mathbf{w}_{t},z_{i_{t}})+\mathbf{b}_{t})\|_{2}^{2}$, then we have $\langle\mathbf{w}_{t}-\mathbf{w}^{*},\partial\ell(\mathbf{w}_{t},z_{i_{t}})\rangle\leq\frac{1}{2\eta}\big{(}\|\mathbf{w}_{t}-\mathbf{w}^{*}\|_{2}^{2}-\|\mathbf{w}_{t+1}-\mathbf{w}^{*}\|_{2}^{2}\big{)}+\frac{\eta}{2}\big{(}\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}+\|\mathbf{b}_{t}\|_{2}^{2}\big{)}-\langle\mathbf{b}_{t},\mathbf{w}_{t}-\mathbf{w}^{*}-\eta\partial\ell(\mathbf{w}_{t},z_{i_{t}})\rangle$. Further, noting $\|\mathbf{w}_{1}\|_{2}=0$, then by the convexity of $\ell$ we have $\displaystyle\frac{1}{T}\sum_{t=1}^{T}[\ell(\mathbf{w}_{t},z_{i_{t}})-\ell(\mathbf{w}^{*},z_{i_{t}})]\leq\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{2T\eta}+\frac{\eta}{2T}\sum_{t=1}^{T}\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}-\frac{1}{T}\sum_{t=1}^{T}\langle\mathbf{b}_{t},\mathbf{w}_{t}-\mathbf{w}^{*}-\eta\partial\ell(\mathbf{w}_{t},z_{i_{t}})\rangle+\frac{\eta}{2T}\sum_{t=1}^{T}\|\mathbf{b}_{t}\|_{2}^{2}.$ The definition of $\alpha$-Hölder smoothness implies that $\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}\leq M+L\|\mathbf{w}_{t}\|^{\alpha}_{2}\leq M+LR^{\alpha}$ for any $t$. Then, there hold $\displaystyle\frac{\eta}{2T}\sum_{t=1}^{T}\|\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}^{2}$ $\displaystyle\leq\frac{\eta}{2T}\sum_{t=1}^{T}(M+L\|\mathbf{w}_{t}\|^{\alpha}_{2})^{2}=\mathcal{O}(\eta),$ and $\|\mathbf{w}_{t}-\mathbf{w}^{*}-\eta\partial\ell(\mathbf{w}_{t},z_{i_{t}})\|_{2}\leq\|\mathbf{w}^{*}\|_{2}+R+\eta(M+LR^{\alpha}).$ Since $\mathbf{b}_{t}$ is an $\sigma^{2}$-sub-Gaussian random vector, $\frac{1}{T}\langle\mathbf{b}_{t},\mathbf{w}_{t}-\mathbf{w}^{*}-\eta\partial\ell(\mathbf{w}_{t},z_{i_{t}})\rangle$ is an $\frac{\sigma^{2}}{T^{2}}\big{(}\|\mathbf{w}^{*}\|_{2}+R+\eta(M+LR^{\alpha})\big{)}^{2}$-sub- Gaussian random vector. Note that the sub-Gaussian parameter $\frac{\sigma^{2}}{T^{2}}\big{(}\|\mathbf{w}^{*}\|_{2}+R+\eta(M+LR^{\alpha})\big{)}^{2}$ is independent of $\mathbf{w}_{t-1}$ and $\mathbf{b}_{t-1}$. Hence, $\frac{1}{T}\sum_{t=1}^{T}\langle\mathbf{b}_{t},\mathbf{w}_{t}-\mathbf{w}^{*}-\eta\partial\ell(\mathbf{w}_{t},z_{i_{t}})\rangle$ is an $\frac{\sigma^{2}\sum_{t=1}^{T}(\|\mathbf{w}^{*}\|_{2}+R+\eta(M+LR^{\alpha}))^{2}}{T^{2}}$-sub- Gaussian random vector. Since $\sigma^{2}=\mathcal{O}(\frac{T\log(1/\delta)}{n^{2}\epsilon^{2}})$, the tail bound of Sub-Gaussian variables (see Lemma 18) implies, with probability at least $1-\frac{\gamma}{18}$, that $\displaystyle\frac{1}{T}\sum_{t=1}^{T}\langle\mathbf{b}_{t},\mathbf{w}_{t}-\mathbf{w}^{*}-\eta\partial\ell(\mathbf{w}_{t},z_{i_{t}})\rangle$ $\displaystyle\leq\frac{\Big{(}\sigma^{2}\big{(}\|\mathbf{w}^{*}\|_{2}+R+\eta(M+LR^{\alpha})\big{)}^{2}\Big{)}^{\frac{1}{2}}}{\sqrt{T}}\sqrt{2\log({18}/{\gamma})}$ $\displaystyle=\mathcal{O}\Big{(}\sigma(\|\mathbf{w}^{*}\|_{2}+\eta)\sqrt{\frac{\log(1/\gamma)}{T}}\Big{)}=\mathcal{O}\Big{(}\frac{\sqrt{\log(1/\delta)\log(1/\gamma)}(\|\mathbf{w}^{*}\|_{2}+\eta)}{n\epsilon}\Big{)}.$ According to the Chernoff bound for the $\ell_{2}$-norm of Gaussian vector with $\mathbf{X}=[\mathbf{b}_{11},...,\mathbf{b}_{1d},\mathbf{b}_{21}...,\mathbf{b}_{Td}]\in\mathbb{R}^{Td}$(see Lemma 15), for any $\gamma\in(18\exp(-dT/8),1)$, with probability at least $1-\frac{\gamma}{18}$, there holds $\frac{\eta}{2T}\sum_{t=1}^{T}\|\mathbf{b}_{t}\|_{2}^{2}\leq\frac{\eta d}{2T}\Big{(}1+(\frac{1}{d}\log({18}/{\gamma}))^{\frac{1}{2}}\Big{)}T\sigma^{2}=\mathcal{O}\Big{(}\frac{\eta Td\log(\frac{1}{\delta})\sqrt{\log(\frac{1}{\gamma})}}{n^{2}\epsilon^{2}}\Big{)}.$ Therefore, with probability at least $1-\frac{\gamma}{9}$, there holds $\displaystyle\frac{1}{T}\sum_{t=1}^{T}[\ell(\mathbf{w}_{t},z_{i_{t}})-\ell(\mathbf{w}^{*},z_{i_{t}})]\leq\mathcal{O}\Big{(}\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{T\eta}+\eta+\frac{\sqrt{\log(1/\delta)\log(1/\gamma)}(\|\mathbf{w}^{*}\|_{2}+\eta)}{n\epsilon}+\frac{\eta Td\log(1/\delta)\sqrt{\log(1/\gamma)}}{n^{2}\epsilon^{2}}\Big{)}.$ (37) Putting (35), (36) and (37) back into (34), we obtain, with probability at least $1-\frac{\gamma}{3}$, that $\displaystyle\mathcal{R}_{S}(\mathbf{w}_{\text{priv}})-\mathcal{R}_{S}(\mathbf{w}^{*})$ $\displaystyle=\mathcal{O}\Big{(}\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{T}}+\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{T\eta}+\eta+\frac{\sqrt{\log({1}/{\delta})\log(1/\gamma)}(\|\mathbf{w}^{*}\|_{2}+\eta)}{n\epsilon}+\frac{\eta Td\log({1}/{\delta})\sqrt{\log({1}/{\gamma})}}{n^{2}\epsilon^{2}}\Big{)}.$ The proof is completed. ∎ Now, we are ready to prove the utility theorem for DP-SGD-Gradient algorithm. ###### Proof of Theorem 12. The Hoeffding inequality implies, with probability at least $1-\frac{\gamma}{3}$, that $\displaystyle\mathcal{R}_{S}(\mathbf{w}^{*})-\mathcal{R}(\mathbf{w}^{*})\leq\big{(}\sup_{z\in\mathcal{Z}}\ell(0,z)+\sup_{z\in{\mathcal{Z}}}\ell(\mathbf{w}^{*},z)\big{)}\sqrt{\frac{\log({3}/{\gamma})}{2n}}=\mathcal{O}\Big{(}\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{n}}\Big{)}.$ Combining Lemma 25, Lemma 26 and the above inequality together, with probability at least $1-\gamma$, we obtain $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})=\mathcal{O}\Big{(}$ $\displaystyle\tilde{\Delta}_{\text{SGD}}(\gamma/6)\log(n)\log(1/\gamma)+\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{T\eta}+\eta+\frac{\sqrt{\log(1/\delta)\log(1/\gamma)}(\|\mathbf{w}^{*}\|_{2}+\eta)}{n\epsilon}$ $\displaystyle+\frac{\eta Td\log(\frac{1}{\delta})\sqrt{\log(\frac{1}{\gamma})}}{n^{2}\epsilon^{2}}+\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{n}}\Big{)}.$ Now, putting $\tilde{\Delta}_{\text{SGD}}(\gamma/6)=\mathcal{O}(\sqrt{T}\eta^{\frac{1}{1-\alpha}}+\frac{T\eta\log(n/\gamma)}{n})$ back into the above estimate, we have $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})$ $\displaystyle=\mathcal{O}\Big{(}\sqrt{T}\log(n)\log(1/\gamma)\eta^{\frac{1}{1-\alpha}}+\frac{\|\mathbf{w}^{*}\|_{2}^{2}}{T\eta}+\eta\Big{(}\frac{Td\log(1/\delta)\sqrt{\log(1/\gamma)}}{n^{2}\epsilon^{2}}+\frac{T\log(n)\log(n/\gamma)\log(1/\gamma)}{n}\Big{)}$ $\displaystyle\qquad+\|\mathbf{w}^{*}\|_{2}^{1+\alpha}\sqrt{\frac{\log(1/\gamma)}{n}}+\frac{\|\mathbf{w}^{*}\|_{2}\sqrt{\log(1/\delta)\log(1/\gamma)}}{n\epsilon}\Big{)}.$ (38) To choose a suitable $\eta$ and $T$ such that the algorithm achieves the optimal rate, we consider the trade-off between $1/\eta$ and $\eta$. We take the derivative of $\frac{1}{T\eta}+\eta\big{(}\frac{Td\log(1/\delta)\sqrt{\log(1/\gamma)}}{n^{2}\epsilon^{2}}+\frac{T\log(n)\log(n/\gamma)\log(1/\gamma)}{n}\big{)}$ w.r.t $\eta$ and set it to $0$, then we have $\eta=1/T\cdot\max\big{\\{}\frac{\sqrt{\log(n)\log(n/\gamma)\log(1/\gamma)}}{\sqrt{n}},\frac{\sqrt{d\log(1/\delta)}(\log(1/\gamma))^{\frac{1}{4}}}{n\epsilon}\big{\\}}$. Putting the value of $\eta$ back into (3.3), we obtain $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})$ $\displaystyle=\mathcal{O}\bigg{(}\frac{\big{(}\log(n)\log(1/\gamma)\big{)}^{\frac{1-2\alpha}{2(1-\alpha)}}n^{\frac{1}{2(1-\alpha)}}}{T^{\frac{1+\alpha}{2(1-\alpha)}}(\log(n/\gamma))^{\frac{1}{2(1-\alpha)}}}+\frac{\sqrt{d\log(1/\delta)\log(1/\gamma)}}{n\epsilon}+\frac{\sqrt{\log(n)\log(n/\gamma)\log(1/\gamma)}}{\sqrt{n}}\bigg{)}\cdot\|\mathbf{w}^{*}\|_{2}^{2}.$ In addition, if $n=\mathcal{O}(T^{\frac{1+\alpha}{2-\alpha}})$, then there holds $\displaystyle\mathcal{R}(\mathbf{w}_{\text{priv}})-\mathcal{R}(\mathbf{w}^{*})=$ $\displaystyle\|\mathbf{w}^{*}\|_{2}^{2}\cdot\mathcal{O}\Big{(}\frac{\sqrt{d\log(1/\delta)\log(1/\gamma)}}{n\epsilon}+\frac{\sqrt{\log(n)\log(n/\gamma)\log(1/\gamma)}}{\sqrt{n}}\Big{)}.$ The above bound matches the optimal rate $\mathcal{O}\big{(}\frac{\sqrt{d\log(1/\delta)}}{n\epsilon}+\frac{1}{\sqrt{n}}\big{)}$. Furthermore, we want the algorithm to achieve the optimal rate with a low computational cost. Therefore, we set $T\asymp n^{\frac{2-\alpha}{1+\alpha}}$ if $\alpha<\frac{1}{2}$, and $T\asymp n$ else. The proof is completed. ∎ Finally, we give the proof of Lemma 13 on the existence of $\beta$ for Algorithm 2 to be $(\epsilon,\delta)$-DP. ###### Proof of Lemma 13. We give sufficient conditions for the existence of $\beta\in(0,1)$ such that RDP conditions (32) and (33) hold with $\sigma^{2}=\frac{14(M+LR^{\alpha})^{2}\lambda}{\beta n\epsilon}$ and $\lambda=\frac{2\log(n)}{(1-\beta)\epsilon}+1$ in Theorem 11. Condition (32) with $T=n$ and $\delta=\frac{1}{n^{2}}$ is equivalent to $\displaystyle f(\beta):=\beta^{2}-\Big{(}1+\frac{7}{1.34n\epsilon}\Big{)}\beta+\frac{7(2\log(n)+\epsilon)}{1.34n\epsilon^{2}}\geq 0.$ (39) If $\big{(}1+\frac{7}{1.34n\epsilon}\big{)}^{2}<\frac{28(2\log(n)+\epsilon)}{1.34n\epsilon^{2}}$, then $f(\beta)\geq 0$ for all $\beta$. Then (32) holds for any $\beta\in(0,1)$. If $\big{(}1+\frac{7}{1.34n\epsilon}\big{)}^{2}\geq\frac{28(2\log(n)+\epsilon)}{1.34n\epsilon^{2}}$, then $\beta\in(0,\beta_{1}]\cup[\beta_{2},+\infty)$ such that the above condition holds, where $\beta_{1,2}=\frac{1}{2}\Big{(}\big{(}1+\frac{7}{1.34n\epsilon}\big{)}\mp\sqrt{\big{(}1+\frac{7}{1.34n\epsilon}\big{)}^{2}-\frac{28(2\log(n)+\epsilon)}{1.34n\epsilon^{2}}}\Big{)}$ are two roots of $f(\beta)=0$. Now, we consider the second RDP condition. Plugging $\sigma^{2}=\frac{14(M+LR^{\alpha})^{2}\lambda}{\beta n\epsilon}$ back into (33), we derive $\displaystyle\frac{3\beta n\epsilon(\lambda-1)}{7\lambda}+\log(\lambda)+\log(1+\frac{7\lambda}{2\beta n\epsilon})\leq\log(n).$ (40) To guarantee (40), it suffices that the following three inequalities hold $\frac{3\beta n\epsilon(\lambda-1)}{7\lambda}\leq\frac{\log(n)}{3},$ (41) $\log(\lambda)\leq\frac{\log(n)}{3},$ (42) $\log\big{(}1+\frac{7\lambda}{2\beta n\epsilon}\big{)}\leq\frac{\log(n)}{3}.$ (43) We set $\lambda=\frac{2\log(n)}{(1-\beta)\epsilon}+1$ in the above three inequalities. Since $\lambda>1$, then (41) holds if $\beta\leq 7\log(n)/9n\epsilon$. Eq. (42) reduces to $\beta\leq 1-\frac{2\log(n)}{(n^{1/3}-1)\epsilon}$. Moreover, (43) is equivalent to the following inequality $\displaystyle g(\beta):=\beta^{2}-(1+\frac{7}{2n(n^{\frac{1}{3}}-1)\epsilon})\beta+\frac{7(2\log(n)+\epsilon)}{2n(n^{\frac{1}{3}}-1)\epsilon^{2}}\leq 0.$ (44) There exists at least one $\beta$ such that $g(\beta)\leq 0$ if $(1+\frac{7}{2n(n^{1/3}-1)\epsilon})^{2}-\frac{14(2\log(n)+\epsilon)}{n(n^{1/3}-1)\epsilon^{2}}\geq 0$, which can be ensured by the condition $\epsilon\geq\frac{7}{2n(n^{1/3}-1)}+2\sqrt{\frac{7\log(n)}{n(n^{1/3}-1)}}$. Furthermore, $g(\beta)\leq 0$ for all $\beta\in[\beta_{3},\beta_{4}]$, where $\beta_{3,4}=\frac{1}{2}\Big{(}\big{(}1+\frac{7}{2n(n^{1/3}-1)\epsilon}\big{)}\mp\sqrt{\big{(}1+\frac{7}{2n(n^{1/3}-1)\epsilon}\big{)}^{2}-\frac{14(2\log(n)+\epsilon)}{n\big{(}n^{1/3}-1\big{)}\epsilon^{2}}}\Big{)}$ are two roots of $g(\beta)=0$. Finally, note that $\max\bigg{\\{}\frac{7}{2n(n^{\frac{1}{3}}-1)}+2\sqrt{\frac{7\log(n)}{n(n^{\frac{1}{3}}-1)}},\frac{\log(n)\big{(}14\log(n)(n^{\frac{1}{3}}-1)+162n-63\big{)}}{9n\big{(}2\log(n)(n^{\frac{1}{3}}-1)-9\big{)}}\bigg{\\}}\leq\frac{7(n^{\frac{1}{3}}-1)+4\log(n)n+7}{2n(n^{\frac{1}{3}}-1)}.$ Then if $n\geq 18$ and $\epsilon\geq\frac{7(n^{\frac{1}{3}}-1)+4\log(n)n+7}{2n(n^{\frac{1}{3}}-1)},$ there hold $\displaystyle\beta_{3}\leq\min\bigg{\\{}\frac{7\log(n)}{9n\epsilon},1-\frac{2\log(n)}{(n^{\frac{1}{3}}-1)\epsilon}\bigg{\\}}$ (45) and $\displaystyle\beta_{3}\leq\beta_{1}\;\text{\ if\ }\big{(}1+\frac{7}{1.34n\epsilon^{2}}\big{)}^{2}\geq\frac{28(2\log(n)+\epsilon)}{1.34n\epsilon^{2}}.$ (46) Conditions (45) and (46) ensure the existence of at least one consistent $\beta\in(0,1)$ such that (39), (41), (42), (43) and (44) hold, which imply that (32) and (33) hold. The proof is completed. ∎ ## 4 Conclusion In this paper, we are concerned with differentially private SGD algorithms with non-smooth losses in the setting of stochastic convex optimization. In particular, we assume that the loss function is $\alpha$-Hölder smooth (i.e., the gradient is $\alpha$-Hölder continuous). We systematically studied the output and gradient perturbations for SGD and established their privacy as well as utility guarantees. For the output perturbation, we proved that our private SGD with $\alpha$-Hölder smooth losses in a bounded $\mathcal{W}$ can achieve $(\epsilon,\delta)$-DP with the excess risk rate $\mathcal{O}\Big{(}\frac{(d\log(1/\delta))^{1/4}\sqrt{\log(n/\delta)}}{\sqrt{n\epsilon}}\Big{)}$, up to some logarithmic terms, and gradient complexity $T=\mathcal{O}(n^{2-\alpha\over 1+\alpha}+n)$, which extends the results of [35] in the strongly-smooth case. We also established similar results for SGD algorithms with output perturbation in an unbounded domain $\mathcal{W}=\mathbb{R}^{d}$ with excess risk $\mathcal{O}\Big{(}\frac{\sqrt{d\log(1/\delta)}\log(n/\delta)}{n^{\frac{2}{3+\alpha}}\epsilon}+\frac{\log(n/\delta)}{n^{\frac{1}{3+\alpha}}}\Big{)}$, up to some logarithmic terms, which are the first-ever known results of this kind for unbounded domains. For the gradient perturbation, we show that private SGD with $\alpha$-Hölder smooth losses in a bounded domain $\mathcal{W}$ can achieve optimal excess risk $\mathcal{O}\Big{(}\frac{\sqrt{d\log(1/\delta)}}{n\epsilon}+\frac{1}{\sqrt{n}}\Big{)}$ with gradient complexity $T=\mathcal{O}(n^{2-\alpha\over 1+\alpha}+n).$ Whether one can derive privacy and utility guarantees for gradient perturbation in an unbounded domain still remains a challenging open question to us. Acknowledgement. This work was done while Puyu Wang was a visiting student at SUNY Albany. The corresponding author is Yiming Ying, whose work is supported by NSF grants IIS-1816227 and IIS-2008532. The work of Hai Zhang is supported by NSFC grant U1811461. ## References * [1] John M Abowd. The challenge of scientific reproducibility and privacy protection for statistical agencies. Census Scientific Advisory Committee, 2016. * [2] Francis Bach and Eric Moulines. Non-strongly-convex smooth stochastic approximation with convergence rate o (1/n). In Advances in neural information processing systems, pages 773–781, 2013. * [3] Raef Bassily, Vitaly Feldman, Cristóbal Guzmán, and Kunal Talwar. Stability of stochastic gradient descent on nonsmooth convex losses. Advances in Neural Information Processing Systems, 33, 2020. * [4] Raef Bassily, Vitaly Feldman, Kunal Talwar, and Abhradeep Guha Thakurta. Private stochastic convex optimization with optimal rates. In Advances in Neural Information Processing Systems, pages 11279–11288, 2019. * [5] Léon Bottou and Olivier Bousquet. The tradeoffs of large scale learning. In Advances in neural information processing systems, pages 161–168, 2008. * [6] Olivier Bousquet and André Elisseeff. Stability and generalization. Journal of machine learning research, 2(Mar):499–526, 2002. * [7] Olivier Bousquet, Yegor Klochkov, and Nikita Zhivotovskiy. Sharper bounds for uniformly stable algorithms. arXiv preprint arXiv:1910.07833, 2019. * [8] Nicholas Carlini, Chang Liu, Úlfar Erlingsson, Jernej Kos, and Dawn Song. The secret sharer: Evaluating and testing unintended memorization in neural networks. In 28th $\\{$USENIX$\\}$ Security Symposium ($\\{$USENIX$\\}$ Security 19), pages 267–284, 2019. * [9] Kamalika Chaudhuri, Claire Monteleoni, and Anand D Sarwate. Differentially private empirical risk minimization. Journal of Machine Learning Research, 12(Mar):1069–1109, 2011. * [10] Aymeric Dieuleveut and Francis Bach. Nonparametric stochastic approximation with large step-sizes. The Annals of Statistics, 44(4):1363–1399, 2016. * [11] Bolin Ding, Janardhan Kulkarni, and Sergey Yekhanin. Collecting telemetry data privately. In Advances in Neural Information Processing Systems, pages 3571–3580, 2017. * [12] Cynthia Dwork and Jing Lei. Differential privacy and robust statistics. In Proceedings of the forty-first annual ACM symposium on Theory of computing, pages 371–380, 2009. * [13] Cynthia Dwork, Frank McSherry, Kobbi Nissim, and Adam Smith. Calibrating noise to sensitivity in private data analysis. In Theory of cryptography conference, pages 265–284. Springer, 2006\. * [14] Cynthia Dwork, Aaron Roth, et al. The algorithmic foundations of differential privacy. Foundations and Trends® in Theoretical Computer Science, 9(3–4):211–407, 2014. * [15] Úlfar Erlingsson, Vasyl Pihur, and Aleksandra Korolova. Rappor: Randomized aggregatable privacy-preserving ordinal response. In Proceedings of the 2014 ACM SIGSAC conference on computer and communications security, pages 1054–1067, 2014. * [16] Vitaly Feldman, Tomer Koren, and Kunal Talwar. Private stochastic convex optimization: optimal rates in linear time. In Proceedings of the 52nd Annual ACM SIGACT Symposium on Theory of Computing, pages 439–449, 2020. * [17] Moritz Hardt, Ben Recht, and Yoram Singer. Train faster, generalize better: Stability of stochastic gradient descent. In International Conference on Machine Learning, pages 1225–1234, 2016. * [18] Wassily Hoeffding. Probability inequalities for sums of bounded random variables. In The Collected Works of Wassily Hoeffding, pages 409–426. Springer, 1994. * [19] Rie Johnson and Tong Zhang. Accelerating stochastic gradient descent using predictive variance reduction. In Advances in neural information processing systems, pages 315–323, 2013. * [20] Simon Lacoste-Julien, Mark Schmidt, and Francis Bach. A simpler approach to obtaining an o (1/t) convergence rate for the projected stochastic subgradient method. arXiv preprint arXiv:1212.2002, 2012. * [21] Yunwen Lei and Yiming Ying. Fine-grained analysis of stability and generalization for stochastic gradient descent. In International Conference on Machine Learning, pages 5809–5819. PMLR, 2020. * [22] Zhicong Liang, Bao Wang, Quanquan Gu, Stanley Osher, and Yuan Yao. Exploring private federated learning with laplacian smoothing. arXiv preprint arXiv:2005.00218, 2020. * [23] Junhong Lin and Lorenzo Rosasco. Optimal learning for multi-pass stochastic gradient methods. In Advances in Neural Information Processing Systems, pages 4556–4564, 2016. * [24] Tongliang Liu, Gábor Lugosi, Gergely Neu, and Dacheng Tao. Algorithmic stability and hypothesis complexity. arXiv preprint arXiv:1702.08712, 2017. * [25] Robert McMillan. Apple tries to peek at user habits without violating privacy. The Wall Street Journal, 2016. * [26] Ilya Mironov. Rényi differential privacy. In 2017 IEEE 30th Computer Security Foundations Symposium (CSF), pages 263–275. IEEE, 2017. * [27] Arkadi Nemirovski, Anatoli Juditsky, Guanghui Lan, and Alexander Shapiro. Robust stochastic approximation approach to stochastic programming. SIAM Journal on optimization, 19(4):1574–1609, 2009. * [28] Francesco Orabona. Simultaneous model selection and optimization through parameter-free stochastic learning. In Advances in Neural Information Processing Systems, pages 1116–1124, 2014. * [29] Alexander Rakhlin, Ohad Shamir, and Karthik Sridharan. Making gradient descent optimal for strongly convex stochastic optimization. In Proceedings of the 29th International Conference on Machine Learning, pages 449–456, 2012. * [30] Ohad Shamir and Tong Zhang. Stochastic gradient descent for non-smooth optimization: Convergence results and optimal averaging schemes. In International Conference on Machine Learning, pages 71–79, 2013\. * [31] Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov. Membership inference attacks against machine learning models. In 2017 IEEE Symposium on Security and Privacy (SP), pages 3–18. IEEE, 2017. * [32] Steve Smale and Yuan Yao. Online learning algorithms. Foundations of Computational Mathematics, 6(2):145–170, 2006. * [33] Shuang Song, Kamalika Chaudhuri, and Anand Sarwate. Learning from data with heterogeneous noise using sgd. In Artificial Intelligence and Statistics, pages 894–902, 2015\. * [34] Martin J Wainwright. High-dimensional statistics: A non-asymptotic viewpoint, volume 48. Cambridge University Press, 2019. * [35] Xi Wu, Fengan Li, Arun Kumar, Kamalika Chaudhuri, Somesh Jha, and Jeffrey Naughton. Bolt-on differential privacy for scalable stochastic gradient descent-based analytics. In Proceedings of the 2017 ACM International Conference on Management of Data, pages 1307–1322, 2017. * [36] Yiming Ying and Massimiliano Pontil. Online gradient descent learning algorithms. Foundations of Computational Mathematics, 8(5):561–596, 2008. * [37] Yiming Ying and Ding-Xuan Zhou. Unregularized online learning algorithms with general loss functions. Applied and Computational Harmonic Analysis, 42(2):224–244, 2017\. Appendix: Proof of Lemma 1 In the appendix, we present the proof of Lemma 1. To this aim, we introduce the following lemma. ###### Lemma 27. Suppose $\ell$ is nonnegative, convex and $\alpha$-Hölder smooth. Let $\mathcal{A}$ be a randomized algorithm with $\sup_{S\simeq S^{\prime}}\delta_{\mathcal{A}}(S,S^{\prime})\leq\Delta_{\mathcal{A}}$. Suppose the output of $\mathcal{A}$ is bounded by $G>0$ and let $M_{0}=\sup_{z\in\mathcal{Z}}\ell(0,z)$, $M=\sup_{z\in\mathcal{Z}}\|\partial\ell(0,z)\|_{2}$. Then for any $\gamma\in(0,1)$, there holds $\mathbb{P}_{\mathcal{S}\sim\mathcal{D}^{n},\mathcal{A}}\left[|\mathcal{R}(\mathcal{A(S)})-\mathcal{R}_{S}(\mathcal{A(S)})|\geq c\bigg{(}(M+LG^{\alpha})\Delta_{\mathcal{A}}\log(n)\log(1/{\gamma})+\big{(}M_{0}+(M+LG^{\alpha})G\big{)}\sqrt{n^{-1}\log(1/\gamma)}\bigg{)}\right]\leq\gamma.$ ###### Proof. By the convexity of $\ell$ and the definition of $\alpha$-Hölder smoothness, we have for any $S$ and $S^{\prime}$, $\displaystyle\ell(\mathcal{A}(S),z)$ $\displaystyle\leq\sup_{z\in\mathcal{Z}}\ell(0,z)+\langle\partial\ell(\mathcal{A}(S),z),\mathcal{A}(S)\rangle\leq M_{0}+\|\partial\ell(\mathcal{A}(S),z)\|_{2}\|\mathcal{A}(S)\|_{2}$ $\displaystyle\leq M_{0}+(M+L\|\mathcal{A}(S)\|^{\alpha}_{2})\|\mathcal{A}(S)\|_{2}\leq M_{0}+(M+LG^{\alpha})G$ (47) and $\displaystyle\sup_{z\in\mathcal{Z}}|\ell(\mathcal{A}(S),z)-\ell(\mathcal{A}(S^{\prime}),z)|$ $\displaystyle\leq\max\big{\\{}\|\partial\ell(\mathcal{A}(S),z)\|_{2},\|\partial\ell(\mathcal{A}(S^{\prime}),z)\|_{2}\big{\\}}\|\mathcal{A}(S)-\mathcal{A}(S^{\prime})\|_{2}$ $\displaystyle\leq(M+LG^{\alpha})\|\mathcal{A}(S)-\mathcal{A}(S^{\prime})\|_{2}.$ Note $\sup_{S\simeq S^{\prime}}\delta_{\mathcal{A}}(S,S^{\prime})\leq\Delta_{\mathcal{A}}$ and $\delta_{\mathcal{A}}(S,S^{\prime})=\|\mathcal{A}(S)-\mathcal{A}(S^{\prime})\|_{2}$. Then for any neighboring datasets $S\simeq S^{\prime}$, we have $\displaystyle\sup_{z\in\mathcal{Z}}|\ell(\mathcal{A}(S),z)-\ell(\mathcal{A}(S^{\prime}),z)|$ $\displaystyle\leq(M+LG^{\alpha})\Delta_{\mathcal{A}}.$ (48) Combining Eq. (Proof.), Eq. (48) and Corollary 8 in [7] together, we derive the following probabilistic inequality $\mathbb{P}_{\mathcal{S}\sim\mathcal{D}^{n},\mathcal{A}}\left[|\mathcal{R}(\mathcal{A(S)})-\mathcal{R}_{S}(\mathcal{A(S)})|\geq c\bigg{(}(M+LG^{\alpha})\Delta_{\mathcal{A}}\log(n)\log(1/{\gamma})+\big{(}M_{0}+(M+LG^{\alpha})G\big{)}\sqrt{n^{-1}\log(1/\gamma)}\bigg{)}\right]\leq\gamma.$ The proof is completed. ∎ ###### Proof of Lemma 1. Let $E_{1}=\\{\mathcal{A}:\sup_{S\simeq S^{\prime}}\|\mathcal{A}(S)-\mathcal{A}(S^{\prime})\|_{2}\geq\Delta_{\mathcal{A}}\\}$ and $E_{2}=\Big{\\{}(S,\mathcal{A}):|\mathcal{R}(\mathcal{A(S)})-\mathcal{R}_{S}(\mathcal{A(S)})|\geq c\Big{(}(M+LG^{\alpha})\Delta_{\mathcal{A}}\log(n)\log(1/{\gamma})+\big{(}M_{0}+(M+LG^{\alpha})G\big{)}\sqrt{n^{-1}\log(1/\gamma)}\Big{)}\Big{\\}}$. Then by the assumption we have $\mathbb{P}_{\mathcal{A}}[\mathcal{A}\in E_{1}]\leq\gamma_{0}$. Further, according to Lemma 27, for any $\gamma\in(0,1)$, we have $\mathbb{P}_{S,\mathcal{A}}[(S,\mathcal{A})\in E_{2}\cap\mathcal{A}\notin E_{1}]\leq\gamma$. Therefore, $\displaystyle\mathbb{P}_{S,\mathcal{A}}[(S,\mathcal{A})\in E_{2}]$ $\displaystyle=\mathbb{P}_{S,\mathcal{A}}[(S,\mathcal{A})\in E_{2}\cap\mathcal{A}\in E_{1}]+\mathbb{P}_{S,\mathcal{A}}[(S,\mathcal{A})\in E_{2}\cap\mathcal{A}\notin E_{1}]$ $\displaystyle\leq\mathbb{P}[\mathcal{A}\in E_{1}]+\mathbb{P}_{S,\mathcal{A}}[(S,\mathcal{A})\in E_{2}\cap\mathcal{A}\notin E_{1}]\leq\gamma_{0}+\gamma.$ The proof is completed. ∎
# A Two-stream Neural Network for Pose-based Hand Gesture Recognition Chuankun Li, Shuai Li, Yanbo Gao, Xiang Zhang, Wanqing Li Chuankun Li is with the North University of China, Taiyuan, China (email: chuankun@nuc.edu.cn).Shuai Li and Yanbo Gao are with Shandong University. (e-mail: {shuaili<EMAIL_ADDRESS>Xiang Zhang is with University of Electronic Science and Technology of China, Chengdu, China. (e-mail: <EMAIL_ADDRESS>Wanqing Li is with the Advanced Multimedia Research Lab, University of Wollongong, Wollongong, Australia. (wanqing@uow.edu.au). ###### Abstract Pose based hand gesture recognition has been widely studied in the recent years. Compared with full body action recognition, hand gesture involves joints that are more spatially closely distributed with stronger collaboration. This nature requires a different approach from action recognition to capturing the complex spatial features. Many gesture categories, such as “Grab” and “Pinch”, have very similar motion or temporal patterns posing a challenge on temporal processing. To address these challenges, this paper proposes a two-stream neural network with one stream being a self-attention based graph convolutional network (SAGCN) extracting the short-term temporal information and hierarchical spatial information, and the other being a residual-connection enhanced bidirectional Independently Recurrent Neural Network (RBi-IndRNN) for extracting long-term temporal information. The self-attention based graph convolutional network has a dynamic self-attention mechanism to adaptively exploit the relationships of all hand joints in addition to the fixed topology and local feature extraction in the GCN. On the other hand, the residual-connection enhanced Bi-IndRNN extends an IndRNN with the capability of bidirectional processing for temporal modelling. The two streams are fused together for recognition. The Dynamic Hand Gesture dataset and First-Person Hand Action dataset are used to validate its effectiveness, and our method achieves state-of-the-art performance. ###### Index Terms: Hand Gesture Recognition, Graph Convolutional Network, Bidirectional Independently Recurrent Neural Network. ## I Introduction Hand gesture recognition (HGR) is a hot research topic in artificial intelligence and computer vision, and has a large number of applications in the human-machine systems [1, 2, 3, 4]. Nowadays, depth sensors, like Intel RealSense and Microsoft Kinect, have become easily accessible, so have the body and hand skeletons [5, 6, 7]. Accordingly, hand gesture recognition based on pose/skeleton has attracted more and more interests[8, 9, 10, 11, 12, 13]. In the past few years, hand gesture recognition based on handcrafted features has been widely reported [8, 14, 15, 16]. Generally, shape of connected joints, hand orientation, surface normal orientation are often used as spatial features, and dynamic time warping or hidden Markov model are then used to process temporal information for classification. Recently, deep learning has also been explored for hand gestures recognition. [11, 17, 12, 18, 19]. One approach is to encode joint sequences into texture images and feed into Convolutional Neural Networks (CNNs) in order to extract discriminative features for gesture recognition. Several methods [12, 19, 20, 21] have been proposed along this approach. However, these methods cannot effectively and efficiently express the dependency between joints, since hand joints are not distributed in a regular grid but in a non-Euclidean domain. To address this problem, graph convolutional networks (GCN) [22, 21, 23] expressing the dependency among joints with a graph have been proposed. For instance, the method in [22] sets four types of edges to capture relationship between non-adjacent joints. However, the topology of the graph is fixed which is ineffective to deal with varying joint relationships for different hand gestures. A typical example is that, the connection between tip of thumb and tip of forefinger in gesture “Write” is likely to be strong, but it is not the case for gestures “Prick” and “Tap”. Modelling gesture dependent collaboration among joints is especially important for robust hand gesture recognition. Furthermore, conventional ST-GCN has limited receptive field in temporal domain, hence, long-term temporal information cannot be effectively learned. Another approach is to apply Recurrent Neural Networks (RNNs) to a hand skeleton sequence for classification [11, 18, 17]. However, due to the problem of gradient exploding and vanishing in the temporal direction or gradient decay along layers in training a RNN, shallow RNNs are often adopted. This paper adopts the Independently Recurrent Neural Network (IndRNN) [24, 25] as the basic component to develop a deep residual bidirectional IndRNN (RBi- IndRNN) for effective extraction of long-term temporal features. The contributions are summarized in the following. * • A two-stream neural network is proposed. One stream is a self-attention based graph convolutional network (SAGCN) and the other is a deep residual- connection enhanced bidirectional Independently Recurrent Neural Network (RBi- IndRNN). * • The SAGCN is specifically designed to model the strong collaboration among joints in hand gestures. In particular, a global correlation map is first adaptively constructed using a self-attention mechanism to characterize pairwise relationships among all the joints and it works together with a static adjacency map of the hand skeleton to capture varying spatial patterns for different gestures. * • The deep residual bidirectional IndRNN (RBi-IndRNN) is developed, which extends the residual connection enhanced IndRNN with bidirectional processing to exploit the bidirectional temporal context and long-term temporal information for challenging gestures having similar motion patterns such as “Grab” vs “Pinch”, “Swipe Left” vs “Swipe Right”. The proposed RBi-IndRNN compensate the inability of the SAGCN in learning long-range temporal patterns. * • Extensive experiments have been performed on two datasets including Dynamic Hand Gesture dataset (DHG) [8] and First-Person Hand Action datasets (FPHA) [26] and the proposed method performs the best. Detailed analysis on the experimental results is also presented to provide insight on the proposed two- stream networks. Part of this work has been presented in ISCAS 2020 [27] where the method only consists of a plain Bi-IndRNN. This paper extends the method in [27] to a two- stream network. One stream is a self-attention based graph convolutional network for capturing spatial and short-term temporal information, and the other stream is the deep residual-connection enhanced Bi-IndRNN to exploit the long-term temporal pattern. More experiments are also conducted with thorough analysis. The rest of this article is structured as follows. Section II provides a review of the relevant literature. The proposed two-stream network is described in Section III. Extensive experiment and detailed analysis are given in section IV. Section V concludes the paper. ## II Related Work Some related and representative works on hand gesture recognition are introduced in this section. They can be categorized into two approaches: handcrafted features-based [14, 15, 28, 16, 8] and deep learning-based [11, 12, 19, 17, 20, 29]. ### II-A Handcrafted Feature based Hand Gesture Recognition Under the category of handcrafted features based methods, Ren et al. [14] modelled hand skeleton as a signature and proposed FEMD (finger-earth mover distance) metric treating each finger as a cluster and penalizing the empty finger-hole to recognize the hand gesture. Wang et al. [15] proposed a mapping score-disparity cost map as the hand representation and used a SVM classifier for recognition. Kuznetsova et al. [28] used some geometric features to characterize the hand skeleton such as point distances and angle between the two lines created by two points. The multi-layered random forest is used for classification. Marin et al. [16] constructed a feature set containing the position and direction of fingertips, and performed classification using multi-class SVM classifier for recognition of hand gestures. Smedt et al. [8] designed features based on hand shape and combined featues of wrist rotations and hand directions, and utilized temporal pyramid to model these features. ### II-B Deep Learning based Hand Gesture Recognition In recent years, deep learning based methods have been widely studied for recognizing hand gesture. Depending on the types of deep neural networks, they can be usually divided into two ways: RNNs-based and CNNs-based ones. Figure 1: Framework of the proposed two-stream network. SAGCN module focuses on hierarchical spatial information and short-term temporal information, and RBi-IndRNN focuses on long-term temporal information. For the CNNs-based approach, it takes the advantage of CNNs to process spatial hand joints or each hand joint over time as grid data and extract local features. Devineau et al. [12] employed a parallel CNN to conduct temporal convolution processing for the world coordinate sequence of the skeleton joint separately, and fused the time-varying characteristics of each joint into a full connection layer (FC) and Softmax function for hand gesture classification. Liu et al. [29] firstly used RGB and depth map based faster region proposal networks to detect hand region. Then the video was segmented by using hand positions. From segmented videos, hand-oriented spatio-temporal features were extracted using the 3D convolutional networks for recognizing the hand gestures. Narayana et al. [19] utilized a spatial focus of attention to construct 12-stream networks called a FOA-Nets. Each network extracted local features from different modality. Then a sparse network architecture was designed to fuse the 12 channels. Pavlo et al. [20] used a recurrent 3DCNN for hand gesture recognition. In the recurrent 3DCNN, connectionist temporal classification (CTC) was utilized to synchronously segment video and classify dynamic hand gestures from multi-modal data. However, these methods based on CNNs cannot utilize the graphical structure of hand joints explicitly. To exploit this, Li et al. [22] employed a hand gesture based graph convolutional network (HGCN) to capture the linkage and motion information of different joints. However, the topology of the graph is fixed which limits capability of the network. In addition, the GCN is not designed for learning long-term temporal relationship. Nguyen et al. [21] employed a neural network to exploit gesture representation from hand joints by using SPD matrix. This method focuses on relationship of local joints, but ignores two joints that are far away from each other in the matrix. Motivated by the concept of attention based graph neural network (GCN) [30, 31], we propose a self-attention based GCN (SAGCN) to effectively explore the often strong collaboration among hand joints in gesture. In the RNNs-based approach, handcrafted features or features extracted from CNN of joint sequences are used as input to RNNs to explore temporal information. Chen et al. [11] utilized angle features of joints to characterize the finger movement and global features of rotation and translation. Three LSTM networks were used to process these features, respectively. Shin et al. [18] used two local features (finger and palm features) and a global feature (Pose feature) to capture the spatial information, and used a GRU-RNN network to process each feature part. Zhang et al. [17] employed a deep framework including 3DCNN layer, ConvLSTM layer, 2DCNN layer and temporal pooling layer to capture short-term spatio-temporal information. However, due to the difficulty in constructing a deep RNN, these methods usually employ a relatively shallow RNN network. Consequently, they often fail to recognize hand gesture having complex and long-term temporal patterns. In this paper, a residual connection enhanced bidirectional Independently Recurrent Neural Network (RBi-IndRNN) is employed to address this issue. ## III Proposed Method Fig. 1 illustrates the proposed two-stream network, one stream is a self- attention based Graph Convolutional Network learning the spatial and short- term temporal features, and the other stream is a residual Bi-IndRNN learning long-term temporal features. The two streams are then fused at score level for final classification. Figure 2: Sample of hand skeleton in the DHG 14/28 dataset [8]. Referencing to the gravity center of the hand joints (purple cross), the neighbouring connected joints of each joint, called root (blue circle), are grouped into two: 1) centripetal group (yellow circle): adjacent nodes closer to the gravity center of the hand joint; 2)centrifugal group: the rest connected neighbors (green circle). ### III-A Self-attention based Graph Convolutional Network Graph Convolutional Networks (GCNs) have been proposed for structured but non- grid data and extended for recognizing human action based on skeleton. A typical example is the spatial temporal GCN (ST-GCN) [32]. However, ST-GCN mainly explores the interaction or collaboration between local (e.g. first and second order neighbouring) joints and the topology of the graph representing the human body is fixed or static. Such representation is ineffective in exploring the collaboration among non-neighbouring joints and, hence, in differentiating actions with collaboration of different body parts. This problem becomes severe in hand gesture recognition where hand joints, regardless of being connected or not, are more closely and strongly collaborated than body joints in action recognition. To tackle this problem, a self-attention based ST-GCN method is developed. As illustrated in the left image of Fig. 2, a hand skeleton graph is usually defined with each hand joint being a node and anatomical connectivity between joints being edges. The graph is expressed as an adjacency matrix whose elements representing connected neighbouring joints are one and, otherwise, zero. Convolutional operations are performed at each node, the node itself, referred to as a root node to differentiate it from its connected nodes in the following, and its neighbouring connected nodes defined by the adjacency matrix. In order to characterize the local structure, different weights are learned for the root node and its connected nodes. Since the number of its connected nodes varies from node to node, a spatial configuration partitioning strategy is developed for hand joints similarly as in [32] to generalize the operations by defining two groups of neighbouring connected nodes and assuming the same weight for nodes in the same group. Specifically, a gravity center is first generated by calculating the average coordinate of all hand joints in a frame. Using this gravity center as a reference point, the neighbouring adjacent nodes/joints of each root node are grouped into two: 1) centripetal group: adjacent nodes closer to the gravity center of the hand joint; 2) the centrifugal group: the other neighbouring nodes locating farther away from the gravity center than the root node/joint, as illustrated the right image in Fig. 2. With the above adjacency matrix and partitioning strategy, a convolution operation on each node can be expressed as follows. $\mathbf{f_{out}}=\sigma(\sum_{k}^{K_{v}}\mathbf{W_{k}f_{in}}\mathbf{A_{k}})$ (1) where $K_{v}$ is the kernel size of the spatial dimension, i.e., the number of groups of connected neighbors, and is set to 3 (the two groups plus the root). For brevity, $\mathbf{A_{k}}$ denotes the normalized adjacency matrix $\mathbf{\Lambda_{k}^{-\frac{1}{2}}(A_{k})\Lambda_{k}^{-\frac{1}{2}}}$ similarly as in[31], where $\Lambda_{k}^{ii}=\sum_{j}(A_{k}^{(ij)})$. $\mathbf{W_{k}}$ is the weight vector of the $1\times 1$ convolution operation. The nonlinear activation function ReLU is used for $\sigma$. The above adjacency matrix is predefined anatomically and static. It represents each joint and its connected neighbouring joints, hence, the convolution operation extracts local features. However, such a local process is not able to capture directly the collaboration among non-connected joints that is often dynamic within a gesture and varies from gesture to gesture. Therefore, we propose in this paper a dynamic attention matrix or map to adaptively characterize the collaboration among all nodes/joints. The attention matrix is obtained via a self-attention mechanism and is calculated in a feature space by projecting the node features with a weight $\mathbf{W_{a}}$. The attention matrix is especially useful for the spatially closely-located, connected or non-connected hand joints that have strong collaboration. The detailed process is expressed as follows: $\displaystyle\mathbf{f_{a}}$ $\displaystyle=\mathbf{W_{a}{f_{in}}}$ (2) $\displaystyle\mathbf{A_{g}}$ $\displaystyle=\frac{exp(\mathbf{f_{a}}\otimes\mathbf{f_{a}^{T}})}{\sum exp(\mathbf{f_{a}}\otimes\mathbf{f_{a}^{T}})}$ (3) where $\mathbf{f_{in}}$ is input and $\mathbf{W}$ maps the input to a feature space (done via a convolutional operation in the experiments). $\mathbf{f_{a}^{T}}$ is the transpose of $\mathbf{f_{a}}$, and $\otimes$ is matrix multiplication. $\mathbf{A_{g}}$ is an attention matrix to indicate the relationship of the pair-wise nodes/joints and is normalized to $(0,1)$ via softmax. In order to capture local structure and global collaboration among the joins together, $\mathbf{A_{g}}$ is combined with the adjacency matrix as follows. $\mathbf{f_{out}}=\sigma(\sum_{k}^{K_{v}}\mathbf{W_{k}f_{in}}\mathbf{A_{k}}+\mathbf{W_{g}f_{in}}\mathbf{A_{g}})$ (4) where $\sum_{k}^{K_{v}}\mathbf{W_{k}f_{in}}\mathbf{A_{k}}$ captures local structure of joints and their connected neighbors and $\mathbf{W_{g}f_{in}}\mathbf{A_{g}}$ captures the global collaboration among all the joints. In this way, the proposed self-attention based GCN, termed as SAGCN, can process both local and global features together. Finally, the spatial features at each time step are processed over time with convolution in the same way as TCN in [32]. Figure 3: Illustration of an SAGCN unit. Figure 4: The recognition network built upon the proposed SAGCN. Fig. 3 illustrates the proposed SAGCN. An input feature is first processed by a convolution operation to a feature space, and then multiplied and normalized with a Softmax function to obtain the attention map $A_{g}$. Together with the input adjacency matrix, the GCN and TCN process the input feature spatially and temporally with batch normalization and ReLU activation functions. $N$, $C$ and $T$ in the figure represent the total number of the vertexes, number of convolutional channels and the length of hand sequences, respectively. The structure of the proposed SAGCN used in the experiments is illustrated in Fig. 4. The network contains six layers of SAGCN units and in each SAGCN layer the numbers of kernels of the attention component, the GCN component and the TCN component are set the same. The numbers of kernels for the six SAGCN layers are 64, 64, 128, 128, 256, and 256, respectively. The stride for the convolution in the fourth TCN is set to $2$ as a pooling operation over time. After six layers of SAGCN units, global average pooling (GPA) is used to pool the spatial-temporal features, and feed into FC and Softmax function for gesture classification. Figure 5: Structure of the proposed bidirectional IndRNN. ### III-B Residual Bidirectional Independently Recurrent Neural Network (RBi- IndRNN) As seen from the architecture of an SAGCN unit, it focuses on hierarchical spatial features and relatively short-term temporal features since it has a relatively small receptive field of the temporal convolution operations. It would be difficult for an SAGCN to capture long-term temporal features, hence, to distinguish gestures that are differentiated in long-term motion patterns, such as “Swipe up” and “Swipe down”. To address this shortcoming of the SAGCN, a second stream using the recent IndRNN [24, 25] is proposed in this paper. $\mathbf{h}_{t}=\sigma(\mathbf{Wx}_{t}+\mathbf{u}\odot\mathbf{h}_{t-1}+\mathbf{b})$ (5) where $\mathbf{x}_{t}\in\mathbb{R}^{M}$ is input of IndRNN network and $\mathbf{h}_{t}\in\mathbb{R}^{N}$ represents hidden states at time-step $t$. $\mathbf{W}\in\mathbb{R}^{N\times M}$, $\mathbf{u}\in\mathbb{R}^{N}$ and $\mathbf{b}\in\mathbb{R}^{N}$ are the weights need to be learned. $\odot$ is dot product, and $N$ is the number of neurons. One of the key advantages of IndRNN, comparing with the conventional RNN, is that IndRNN can capture longer temporal information by regulating the recurrent weights to avoid gradient vanishing and exploding in training. Moreover, multiple IndRNN layers are able to be efficiently stacked to build a deeper network with low complexity. Considering the success of the Bidirectional Recurrent Neural Networks (Bi- RNN) in action recognition [33] and language modelling [34], a Bidirectional IndRNN (Bi-IndRNN) is constructed whose architecture is shown in Fig. 5. The features are extracted by two directions of temporal processing using IndRNNs, and features of two directions are concatenated and fed into next layer. The Bi-IndRNN is able to capture the temporal relationship in two directions. Considering that IndRNN can effectively work with ReLU, we further develop a residual connection enhanced Bi-IndRNN (RBi-IndRNN) as shown in Fig. 6a, by adding an identity shortcut (skip-connection) to bypass the non-linear transformation of the input feature in order to facilitate the gradient backpropagation. This skip-connection does not affect the temporal processing, but makes the deeper features a summation of the shallower features. This paper adopts a pre-activation type of residual function with batch normalization, Bi-IndRNN and then weight processing as shown in Fig. 6a. (a) (b) Figure 6: Illustration of the (a): RBi-IndRNN module and (b): proposed 6-layer RBi-IndRNN for hand gesture recognition. In the experiments, a deep RBi-IndRNN of 6-layers and the framework is shown in Fig. 6b. After six layers of RBi-IndRNN, the features of last time-step are fed into a FC layer for gesture classification. Moreover, the temporal displacement of each joint describing the movement between adjacent frames similarly as the optical flow is extracted and concatenated with original skeletal joints as input to the RBi-IndRNN. Figure 7: A sample attention matrix of gesture “Grab” obtained by the proposed SAGCN. (a) (b) Figure 8: Confusion matrices on the DHG-14 dataset using (a): RBi-IndRNN and (b): SAGCN. In order to reduce the complexity of the two-stream network training, the SAGCN and the RBi-IndRNN are trained separately. The resulted models are then used to test and the probability vectors produced by the SAGCN and RBi-IndRNN are fused by multiplication. The class with the max probability is recognized as the gesture class and can be expressed as follows. $label=argmax({v_{SAGCN}}\odot{v_{RBi-IndRNN}})$ (6) where $v$ represents a probability vector, $\odot$ is dot product, and $argmax(\cdot)$ is to find the index position with the maximum probability. ## IV Experiments ### IV-A Datasets Two widely used datasets for HGR, namely the Dynamic Hand Gesture (DHG) 14/28 dataset [8] and the First-Person Hand Action (FPHA) dataset [26], are used for experiments. The DHG 14/28 dataset [8] is captured by Intel RealSense depth cameras containing hand joint data (3-dimension world coordinate ($x$, $y$, $z$)) and depth map sequences as shown in Fig. 2. This DHG 14/28 dataset is constructed with 2800 gesture sequences containing 14 gestures performed by 20 subjects. Each sequence ranging from 20 to 50 frames is assigned with a class label. The categories of this dataset include “Swipe x (SX)”, “Swipe down (SD)”, “Rotation counter-clockwise (RCC)”, “Tap (T)”, “Rotation clockwise (RC)”, “Swipe right (SR)”, “Pinch (P)”, “Swipe up (SU)”, “Shake (SH)”, “Grab (G)”, “Swipe left (SL)”, “Swipe v (SV)”,“Expand (E)”, “Swipe + (S+)”. Depending on the number of fingers used, gestures are classified with either 14 labels or 28 labels. The evaluation protocols in [12] are adopted, namely 1960 video as training samples and other sequences for testing. And 5% of the training data is randomly selected for validation. 20 frames are sampled from each sequence and fed into the proposed networks described in Section III for training and classification. The FPHA dataset [26] contains 1175 sequences from 45 different gesture classes with high viewpoint, speed, intra-subject variability and inter- subject variability of style, viewpoint and scale. This dataset is captured in 3 different scenarios (kitchen, office and social) and performed by 6 subjects. Compare with DHG 14/28 dataset, FPHA dataset has 21 hand joints and the palm joint is missed. This is a challenging dataset due to the similar motion patterns and involvement of many different objects. The same evaluation strategy in [26] are used. ### IV-B Training Setup The experiments are conducted on the Pytorch platform using a 1070Ti GPU card. Adaptive Moment Estimation (Adam) [35] is used as optimization function of training network. The batch size is set to 64 and 32 for DHG and FPHA datasets, respectively. Dropout [36] is set to 0.2 and 0.5 and the initial learning rate is set to $2*10^{-4}$ and $2*10^{-3}$ for RBi-IndRNN and SAGCN respectively. When the validation accuracy is improved, learning rate is decayed by 10. The number of neurons in each RBi-IndRNN layer is $512$. And 1024 neurons are used in FC of the RBi-IndRNN due to the bidirectional processing. TABLE I: Comparison of w/o self-attention for the GCN on two datasets Method | DHG-14 | DHG-28 | FPHA ---|---|---|--- GCN | 90.83% | 87.74% | 83.48% SAGCN | 93.33% | 91.54% | 87.83% TABLE II: Results on the two datasets under different settings of RBi-IndRNN Method | DHG-14 | DHG-28 | FPHA ---|---|---|--- IndRNN(joint coordinate) | 92.07% | 85.82% | 84.52% IndRNN(joint coordinate + displacement) | 92.19% | 88.87% | 86.78% Bi-IndRNN(joint coordinate + displacement) | 93.15% | 91.13% | 88.35% RBi-IndRNN(joint coordinate + displacement) | 94.05% | 91.90% | 88.87% TABLE III: Performance of the proposed method on the DHG dataset with comparisons to the existing methods in terms of accuracy Method | modality | 14 gestures | 28 gestures | average ---|---|---|---|--- HO4D Normals [37] | Depth | 78.53% | 74.03% | 76.28% Motion Trajectories [13] | Pose | 79.61% | 62.00% | 70.80% CNN for key frames [9] | Pose | 82.90% | 71.90% | 77.40% JAS and HOG2 [10] | Pose | 83.85% | 76.53% | 80.19% RNN+Motion feature [11] | Pose | 84.68% | 80.32% | 82.50% HoHD+HoWR+SoCJ [8] | Pose | 88.24% | 81.90% | 85.07% Parallel CNN [12] | Pose | 91.28% | 84.35% | 87.82% HG-GCN [22] | Pose | 92.80% | 88.30% | 90.55% leap motion controller [38] | Pose | 97.62% | 91.43% | 94.53% PB-GRU-RNN [18] | Pose | 95.21% | 93.23% | 94.22% proposed method | Pose | 96.31% | 94.05% | 95.18% Figure 9: Confusion matrix of the proposed network on DHG-14. Figure 10: Confusion matrix of the proposed network on DHG-28. TABLE IV: Results on the FPHA dataset and comparisons to the existing methods in terms of the accuracy Method | modality | 14 gestures ---|---|--- Two stream [39] | RGB | 75.30% Joint angles similarities and HOG2 [10] | Depth+Pose | 66.78% Histogram of Oriented 4D Normals [37] | Depth | 70.61% JOULE [40] | RGB+Depth+Pose | 78.78% Novel View [41] | Depth | 69.21% 2-layer LSTM [42] | Pose | 80.14% Moving Pose [43] | Pose | 56.34% Lie Group [44] | Pose | 82.69% HBRNN [45] | Pose | 77.40% Gram Matrix [46] | Pose | 85.39% TF [47] | Pose | 80.69% SPD Matrix Learning [48] | Pose | 84.35% Grassmann Manifolds [49] | Pose | 77.57% proposed method | Pose | 90.26% ### IV-C Ablation Study on Some Key Factors #### IV-C1 Contribution of the self-attention in the SAGCN The self-attention model used in the proposed SAGCN is evaluated against the GCN. The results on the two datasets by using GCN network and SAGCN are tabulated in Table I. It shows that the self-attention model used in the proposed SAGCN significantly improves the performance, compared with original GCN network. Taking the FPHA dataset as an example, the performance is improved from $83.48\%$ to $87.83\%$ in terms of accuracy. It indicates that the collaboration among all hand joints is important and is well captured by the proposed attention model via the attention matrix. A visual illustration, an attention matrix of gesture “Grab”, is shown in Fig. 7. This matrix demonstrates and also verifies the intuition that collaboration between thumb and forefinger joints is very important for recognizing gesture “Grab”. #### IV-C2 Contribution of the bidirectional processing and residual connection in the RBi-IndRNN The bidirectional processing and residual connection in the RBi-IndRNN is evaluated against the plain IndRNN. Table II shows the recognition results under different settings. The classification accuracies of the plain IndRNN only using joint world coordinates are 92.07%, 85.82% and 84.52% on DHG14, DHG-28 and FPHA datasets, respectively, which outperforms some methods (as illustrated in the following Table IV and Table III). When using the temporal displacement and joint world coordinates, the performance is further improved to 92.19%, 88.87% and 86.78%, respectively. It improves by $3.05$ and $2.26$ percentage points for DHG-28 and FPHA datasets respectively, compared with only using joint coordinates. Meanwhile, the RBi-IndRNN with the temporal displacement performs better than Bi-IndRNN and gets the best performance as shown in Table II. #### IV-C3 Contribution of the long-term temporal feature processing using RBi-IndRNN against SAGCN As discussed in Section III, SAGCN is not capable of learning long-term temporal features, but the RBi-IndRNN is. By comparing the performance of SAGCN and RBi-IndRNN as illustrated in Table I and Table II, respectively, RBi-IndRNN outperforms the SAGCN. To further understand what types of hand gestures that SAGCN and RBi-IndRNN are better at in the proposed framework, the confusion matrices of the SAGCN and RBi-IndRNN recognition results on the DHG-14 dataset are presented in Fig. 8a and 8b, respectively. By comparing the two confusion matrices, it can be seen that SAGCN performs better for hand gestures having much spatial variation and less temporal movement such as “ Swipe left ”, while RBi-IndRNN performs better for hand gestures with complex temporal motion such as “Grab ”. This indicates that SAGCN and RBi-IndRNN are complementary to each other. Figure 11: Confusion matrix of the proposed network on the FPHA dataset. ### IV-D Results on the DHG Dataset Results of the proposed method comparing with the existing hand-crafted feature based and deep learning based method [37, 13, 10, 8, 9, 12, 21, 38, 18] are shown in Table III. It can be seen that approaches based on deep learning generally outperform handcrafted feature based methods. Compared with methods based on CNNs [9, 12], graph convolutional networks based method [22] can capture spatial relationship of hand joints better and perform better. The proposed method outperforms these approaches, and achieves the best performance. Fig. 9 and Fig. 10 show the confusion matrices on the DHG-14 and DHG-28 datasets, respectively, obtained with the proposed method. It can be seen that most hand gestures can be recognized effectively except hand gesture “Grab (G)” and “Pinch (P)”. This is mostly because some Grab gestures only using the thumb and forefinger is of similar movement with “Pinch”. Therefore, networks to explore subtle differences are still highly desired and need to be investigated in the future. However, compared with results as illustrated in 8a of the previous work [27], our proposed method improves performance on all hand gestures. ### IV-E Results on the FPHA Dataset The comparison between the existing methods and the proposed method on the FPHA dataset is shown in Table IV. It also shows that our method achieves better performance than existing methods [39, 40, 47, 46, 49]. The confusion matrix of FPHA dataset is illustrated in Fig. 11. Most of the hand gestures are able to be accurately recognized. However, it is still difficult to recognize some gestures effectively such as “open wallet”, “unfold glasses” and “take letter from envelope” as shown in the confusion matrix. This is mostly because these hand gestures involve hand-object interaction, which cannot be well captured by the skeleton/pose alone. ## V Conclusion In this paper, a two-stream neural network is presented for recognizing the pose-based hand gesture. One is an SAGCN network having a self-attention mechanism to adaptively explore the collaboration among all joints in the spatial and short-term temporal domains. The other is a RBi-IndRNN to explore the long-term temporal dependency, compensating the weakness of SAGCN in processing the temporal features. The bidirectional processing and residual connections used in the RBi-IndRNN have proven to be effective in learning temporal patterns. State-of-the-art results are achieved by our two-stream neural network on two representative hand gesture datasets. Thorough analysis with ablation studies have also been conducted, validating the effectiveness of the proposed method. ## References * [1] L. Cheng, Y. Liu, Z. Hou, M. Tan, D. Du, and M. Fei, “A rapid spiking neural network approach with an application on hand gesture recognition,” _IEEE Transactions on Cognitive and Developmental Systems_ , pp. 1–10, 2019\. * [2] Y. Xue, Z. Ju, K. Xiang, J. Chen, and H. Liu, “Multimodal human hand motion sensing and analysis: A review,” _IEEE Transactions on Cognitive and Developmental Systems_ , vol. 11, pp. 162–175, 2019. * [3] Y. Yang, F. Duan, J. Ren, J. Xue, Y. Lv, C. Zhu, and H. Yokoi, “Performance comparison of gestures recognition system based on different classifiers,” _IEEE Transactions on Cognitive and Developmental Systems_ , pp. 1–10, 2020\. * [4] H. Cheng, L. Yang, and Z. Liu, “Survey on 3d hand gesture recognition,” _IEEE Transactions on Circuits and Systems for Video Technology_ , vol. 26, no. 9, pp. 1659–1673, 2016. * [5] C. Li, B. Zhang, C. Chen, Q. Ye, J. Han, G. Guo, and R. Ji, “Deep manifold structure transfer for action recognition,” _IEEE Transactions on Image Processing_ , vol. 28, pp. 4646–4658, 2019. * [6] C. Cao, C. Lan, Y. Zhang, W. Zeng, H. Lu, and Y. Zhang, “Skeleton-based action recognition with gated convolutional neural networks,” _IEEE Transactions on Circuits and Systems for Video Technology_ , vol. 29, pp. 3247–3257, 2019. * [7] S. Song, C. Lan, J. Xing, W. Zeng, and J. Liu, “An end-to-end spatio-temporal attention model for human action recognition from skeleton data,” in _AAAI_ , 2017. * [8] Q. De Smedt, H. Wannous, and J.-P. Vandeborre, “Skeleton-based dynamic hand gesture recognition,” in _IEEE Conference on Computer Vision and Pattern Recognition Workshops_ , 2016, pp. 1–9. * [9] Q. De Smedt, H. Wannous, J.-P. Vandeborre, J. Guerry, B. Le Saux, and D. Filliat, “Shrec’17 track: 3d hand gesture recognition using a depth and skeletal dataset,” in _Eurographics Workshop on 3D Object Retrieval_ , 2017\. * [10] E. Ohn-Bar and M. Trivedi, “Joint angles similarities and hog2 for action recognition,” in _IEEE conference on computer vision and pattern recognition workshops_ , 2013, pp. 465–470. * [11] X. Chen, H. Guo, G. Wang, and L. Zhang, “Motion feature augmented recurrent neural network for skeleton-based dynamic hand gesture recognition,” in _IEEE International Conference on Image Processing_ , 2017, pp. 2881–2885. * [12] G. Devineau, F. Moutarde, W. Xi, and J. Yang, “Deep learning for hand gesture recognition on skeletal data,” in _IEEE International Conference on Automatic Face & Gesture Recognition_, 2018, pp. 106–113. * [13] M. Devanne, H. Wannous, S. Berretti, P. Pala, M. Daoudi, and A. Del Bimbo, “3-d human action recognition by shape analysis of motion trajectories on riemannian manifold,” _IEEE transactions on cybernetics_ , vol. 45, no. 7, pp. 1340–1352, 2015. * [14] Z. Ren, J. Yuan, and Z. Zhang, “Robust hand gesture recognition based on finger-earth mover’s distance with a commodity depth camera,” in _ACM international conference on Multimedia_ , 2011, pp. 1093–1096. * [15] H. Wang, Q. Wang, and X. Chen, “Hand posture recognition from disparity cost map,” in _Asian Conference on Computer Vision_ , 2012. * [16] G. Marin, F. Dominio, and P. Zanuttigh, “Hand gesture recognition with leap motion and kinect devices,” in _IEEE International Conference on Image Processing_ , 2014, pp. 1565–1569. * [17] L. Zhang, G. Zhu, P. Shen, J. Song, S. A. Shah, and M. Bennamoun, “Learning spatiotemporal features using 3dcnn and convolutional lstm for gesture recognition,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2017, pp. 3120–3128. * [18] S. Shin and W. Kim, “Skeleton-based dynamic hand gesture recognition using a part-based gru-rnn for gesture-based interface,” _IEEE Access_ , vol. 8, pp. 50 236–50 243, 2020. * [19] P. Narayana, J. R. Beveridge, and B. A. Draper, “Gesture recognition: Focus on the hands,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2018, pp. 5235–5244. * [20] P. Molchanov, X. Yang, S. Gupta, K. Kim, S. Tyree, and J. Kautz, “Online detection and classification of dynamic hand gestures with recurrent 3d convolutional neural network,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2016, pp. 4207–4215. * [21] X. S. Nguyen, L. Brun, O. Lezoray, and S. Bougleux, “A neural network based on spd manifold learning for skeleton-based hand gesture recognition,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2019, pp. 12 036–12 045. * [22] Y. Li, Z. He, X. Ye, Z. He, and K. Han, “Spatial temporal graph convolutional networks for skeleton-based dynamic hand gesture recognition,” _Eurasip Journal on Image and Video Processing_ , vol. 2019, no. 1, pp. 1–7, 2019. * [23] G. Hu, B. Cui, Y. He, and S. Yu, “Progressive relation learning for group activity recognition,” _2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)_ , pp. 977–986, 2020. * [24] S. Li, W. Li, C. Cook, C. Zhu, and Y. Gao, “Independently recurrent neural network (indrnn): Building a longer and deeper rnn,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2018, pp. 5457–5466. * [25] S. Li, W. Li, C. Cook, and Y. Gao, “Deep independently recurrent neural network (indrnn),” _ArXiv_ , vol. abs/1910.06251, 2019. * [26] G. Garcia-Hernando, S. Yuan, S. Baek, and T.-K. Kim, “First-person hand action benchmark with rgb-d videos and 3d hand pose annotations,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2018. * [27] S. Li, L. Zheng, C. Zhu, and Y. Gao, “Bidirectional independently recurrent neural network for skeleton-based hand gesture recognition,” in _Proceedings of IEEE International Symposium on Circuits and Systems_ , 2020\. * [28] A. Kuznetsova, L. Leal-Taixé, and B. Rosenhahn, “Real-time sign language recognition using a consumer depth camera,” in _2013 IEEE International Conference on Computer Vision Workshops_ , 2013, pp. 83–90. * [29] Z. Liu, X. Chai, Z. Liu, and X. Chen, “Continuous gesture recognition with hand-oriented spatiotemporal feature,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2017, pp. 3056–3064. * [30] S. Guo, Y. Lin, N. Feng, C. Song, and H. Wan, “Attention based spatial-temporal graph convolutional networks for traffic flow forecasting,” in _AAAI Conference on Artificial Intelligence_ , 2019. * [31] L. Shi, Y. Zhang, J. Cheng, and H. Lu, “Two-stream adaptive graph convolutional networks for skeleton-based action recognition,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2019, pp. 12 018–12 027. * [32] S. Yan, Y. Xiong, D. Lin, and X. Tang, “Spatial temporal graph convolutional networks for skeleton-based action recognition,” in _AAAI Conference on Artificial Intelligence_ , 2018, pp. 7444–7452. * [33] X. Liu, Y. Li, and Q. Wang, “Multi-view hierarchical bidirectional recurrent neural network for depth video sequence based action recognition,” _International Journal of Pattern Recognition and Artificial Intelligence_ , p. 1850033, 2018. * [34] E. Arisoy, A. Sethy, B. Ramabhadran, and S. Chen, “Bidirectional recurrent neural network language models for automatic speech recognition,” in _IEEE International Conference on Acoustics, Speech and Signal Processing_ , 2015, pp. 5421–5425. * [35] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” _arXiv preprint arXiv:1412.6980_ , 2014. * [36] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Dropout: a simple way to prevent neural networks from overfitting,” _The Journal of Machine Learning Research_ , vol. 15, no. 1, pp. 1929–1958, 2014. * [37] O. Oreifej and Z. Liu, “Hon4d: Histogram of oriented 4d normals for activity recognition from depth sequences,” in _IEEE conference on computer vision and pattern recognition_ , 2013, pp. 716–723. * [38] D. Avola, M. Bernardi, L. Cinque, G. L. Foresti, and C. Massaroni, “Exploiting recurrent neural networks and leap motion controller for the recognition of sign language and semaphoric hand gestures,” _IEEE Transactions on Multimedia_ , vol. 21, no. 1, pp. 234–245, 2019. * [39] C. Feichtenhofer, A. Pinz, and A. Zisserman, “Convolutional two-stream network fusion for video action recognition,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2016, pp. 1933–1941. * [40] J. Hu, W.-S. Zheng, J.-H. Lai, and J. Zhang, “Jointly learning heterogeneous features for rgb-d activity recognition,” _IEEE Transactions on Pattern Analysis and Machine Intelligence_ , vol. 39, pp. 2186–2200, 2017. * [41] H. Rahmani and A. Mian, “3d action recognition from novel viewpoints,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2016, pp. 1506–1515. * [42] W. Zhu, C. Lan, J. Xing, W. Zeng, Y. Li, L. Shen, and X. Xie, “Co-occurrence feature learning for skeleton based action recognition using regularized deep lstm networks,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2016, pp. 3697–3703. * [43] M. Zanfir, M. Leordeanu, and C. Sminchisescu, “The moving pose: An efficient 3d kinematics descriptor for low-latency action recognition and detection,” in _IEEE International Conference on Computer Vision_ , 2013, pp. 2752–2759. * [44] R. Vemulapalli, F. Arrate, and R. Chellappa, “Human action recognition by representing 3d skeletons as points in a lie group,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2014, pp. 588–595. * [45] Y. Du, W. Wang, and L. Wang, “Hierarchical recurrent neural network for skeleton based action recognition,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2015, pp. 1110–1118. * [46] X. Zhang, Y. Wang, M. Gou, M. Sznaier, and O. Camps, “Efficient temporal sequence comparison and classification using gram matrix embeddings on a riemannian manifold,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2016, pp. 4498–4507. * [47] G. Garciahernando and T. Kim, “Transition forests: Learning discriminative temporal transitions for action recognition and detection,” in _IEEE Conference on Computer Vision and Pattern Recognition_ , 2017, pp. 407–415. * [48] Z. Huang and L. Van Gool, “A riemannian network for spd matrix learning,” in _AAAI Conference on Artificial Intelligence_ , 2016, pp. 2036–2042. * [49] Z. Huang, J. Wu, and L. Van Gool, “Building deep networks on grassmann manifolds,” in _AAAI Conference on Artificial Intelligence_ , 2018, pp. 3279–3286. | Chuankun Li received the BE degree in electronic information engineering from North University of China, Taiyuan, China, in 2012 and received the MS degree in communication and information system from North University of China, Taiyuan, China, in 2015. He received the Ph.D degree with School of electronic information engineering , Tianjin University, China in 2020. His current research interests include computer vision and machine learning. ---|--- | Shuai Li is currently with the School of Control Science and Engineering, ShanDong University (SDU), China, as a Professor and QiLu Young Scholar. He was with the School of Information and Communication Engineering, University of Electronic Science and Technology of China, China, as an Associate Professor from 2018-2020. He received his Ph.D. degree from the University of Wollongong, Australia, in 2018. His research interests include image/video coding, 3D video processing and computer vision. He was a co-recipient of two best paper awards at the IEEE BMSB 2014 and IIH-MSP 2013, respectively. ---|--- | Yanbo Gao is currently with the School of Software, Shandong University (SDU), Jinan, China, as an Associate Professor. She was with the School of Information and Communication Engineering, University of Electronic Science and Technology of China (UESTC), Chengdu, China, as a Post-doctor from 2018-2020. She received her Ph.D. degree from UESTC in 2018. Her research interests include video coding, 3D video processing and light field image coding. She was a co-recipient of the best student paper awards at the IEEE BMSB 2018. ---|--- | Xiang Zhang received the B.S. and M.S. degrees from University of Electronic Science and Technology of China, Chengdu, China, and the Ph.D. degree from Shanghai Jiaotong University, Shanghai, China, in 2003, 2006, and 2009, respectively. He is an Associate Professor with the School of Information and Communication Engineering, University of Electronic Science and Technology of China. His research interests include video analysis and machine learning. ---|--- | Wanqing Li (M’97-SM’05) received his Ph.D. in electronic engineering from The University of Western Australia. He was a Principal Researcher (98-03) at Motorola Research Lab and a visiting researcher (08, 10 and 13) at Microsoft Research US. He is currently an Associate Professor and Co-Director of Advanced Multimedia Research Lab (AMRL) of UOW, Australia. His research areas are machine learning, 3D computer vision, 3D multimedia signal processing and medical image analysis. Dr. Li currently serves as an Associate Editor for IEEE Transactions on Circuits and Systems for Video Technology and IEEE Transactions on Multimedia. He was an Associator for Journal of Visual Communication and Image Representation. ---|---
# Machine Learning Percolation Model Shu Cheng1 Huai Zhang<EMAIL_ADDRESS>Yaolin Shi Key Laboratory of Computational Geodynamics, College of Earth and Planetary Sciences, University of Chinese Academy of Sciences, No.19(A) Yuquan Road, Shijingshan District, Beijing 100049, China Fei He1 Ka-Di Zhu<EMAIL_ADDRESS>Key Laboratory of Artificial Structures and Quantum Control (Ministry of Education), School of Physics and Astronomy, Shanghai Jiao Tong University, 800 Dong Chuan Road, Shanghai 200240, China ###### Abstract Recent advances in machine learning have become increasingly popular in the applications of phase transitions and critical phenomena. By machine learning approaches, we try to identify the physical characteristics in the two- dimensional percolation model. To achieve this, we adopt Monte Carlo simulation to generate dataset at first, and then we employ several approaches to analyze the dataset. Four kinds of convolutional neural networks (CNNs), one variational autoencoder (VAE), one convolutional VAE (cVAE), one principal component analysis (PCA), and one $k$-means are used for identifying order parameter, the permeability, and the critical transition point. The former three kinds of CNNs can simulate the two order parameters and the permeability with high accuracy, and good extrapolating performance. The former two kinds of CNNs have high anti-noise ability. To validate the robustness of the former three kinds of CNNs, we also use the VAE and the cVAE to generate new percolating configurations to add perturbations into the raw configurations. We find that there is no difference by using the raw or the perturbed configurations to identify the physical characteristics, under the prerequisite of corresponding labels. In the case of lacking labels, we use unsupervised learning to detect the physical characteristics. The PCA, a classical unsupervised learning, performs well when identifying the permeability but fails to deduce order parameter. Hence, we apply the fourth kinds of CNNs with different preset thresholds, and identify a new order parameter and the critical transition point. Our findings indicate that the effectiveness of machine learning still needs to be evaluated in the applications of phase transitions and critical phenomena. ††preprint: APS/123-QED ## I Introduction Machine learning methods have rapidly become pervasive instruments due to better fitting quality and predictive quality in comparison with traditional models in terms of phase transitions and critical phenomena. Usually, machine learning can be divided into supervised and unsupervised learning. In the former, the machine receives a set of inputs and labels. Supervised learning models are trained with high accuracy to predict labels. The effectiveness of supervised learning has been examined by many predecessors on Ising models [1, 2, 3, 4, 5, 6, 7, 8], Kitaev chain models [3], disordered quantum spin chain models [3], Bose-Hubbard models [6], SSH models [6], SU(2) lattice gauge theory [7], topological states models [9], $q$-state Potts models [10, 11], uncorrelated configuration models [12], Hubbard models [13, 12], and $XY$ models [14], ect. On the other hand, in unsupervised learning models, there are no labels. Unsupervised learning can be used as meaningful analysis tools, such as sample generation, feature extraction, cluster analysis. Principal component analysis (PCA) is one of the unsupervised learning techniques. Recently investigators have examined the PCA’s effectiveness for exploring physical features without labels in the applications of phase transitions and critical phenomena [15, 14, 8, 13, 16, 17]. Variational autoencoder (VAE) and convolutional VAE (cVAE), another two classical unsupervised learning technique, incorporated into generative neural networks, are used for data reconstruction and dimensional reduction in respect of phase transitions and critical phenomena [8, 18]. Although machine learning approaches have been applied successfully in phase transitions and critical phenomena, there is only one study on the percolation model [17]. Motivated by predecessors, we conduct much more comprehensive studies, which combine supervised learning with unsupervised learning, to detect the physical characteristics in the percolation model. Our work is considered from the following several aspects. First, we use the former three kinds of deep convolutional neural networks (CNNs) to deduce the two order parameters and the permeabilities in the two-dimensional percolation model. our inspiration and method come from [2, 3], whose study both focus on Ising model. Nevertheless, the above CNNs are trained on the known configurations from the dataset obtained by Monte Carlo simulation. [8, 18] find that VAE and cVAE can reconstruct samples in Ising model. Hence, we use the VAE and the cVAE to generate new configurations that are out of the dataset. After generating the new configurations, we pour them into the former three kinds of CNNs, respectively. Having explored supervised learning, we now move on to unsupervised learning. Here we try to identify physical characteristics without labels in the two- dimensional percolation model. [15] takes the first principal component obtained by PCA as the order parameter in Ising model. In contrast to [15], by using preprocessing on the unpercolating clusters, [17] also successfully finds the order parameter in percolation model by PCA. In this study, we try to use the PCA to extract relevant low-dimensional representations to discover physical characteristics. In an actual situation, we may not know the labels when identifying order parameter. To overcome the difficulty associated with missing labels, [12] changes the preset threshold between the labels zero and one so as to make incorrect labels between the preset and the true thresholds. Hence, we deliberately change the preset thresholds, determined by $k$-means, between the labels zero and one. Here we use the fourth kinds of CNNs which receives the raw configurations as input and the labels determined by the preset thresholds. This paper is organized as follows. In Sec. II, we describe the two- dimensional percolation model and the dataset from Monte Carlo simulation. In Sec. III, we give a brief introduction to CNNs, VAE and cVAE, and PCA. Next, we provide dozens of machine learning models to capture the physical characteristics and discuss the results in Sec. IV. Finally, we conclude with a summary in Sec. V. ## II The two-dimensional Percolation model For percolation models, what we need to do is to capture the physical characteristics. A suitable dataset should be constructed to fulfill this objective. Various models in physical dynamics can be simulated mathematically by the Monte Carlo method, and it has been proved to be valid for using the Monte Carlo simulation to capture different physical features in phase transitions and critical phenomena [15, 14, 8]. In this study, the Monte Carlo simulation for the two-dimensional percolation model is carried out as follows. First, 40 values of permeability range from 0.41 to 0.80 with an interval of 0.01. For each permeability, the initial samples consist of 1000 percolating configurations. To train the machine learning models, the matrix $\bm{X}$ with the size of $M\times N$ (see Eq. 6) is used for storing 40,000 raw percolating configurations. $\displaystyle\bm{X}=\left(\begin{array}[]{ccccc}a_{1,1}&a_{1,2}&\ldots&a_{1,N-1}&a_{1,N}\\\ a_{2,1}&a_{2,2}&\ldots&a_{2,N-1}&a_{2,N}\\\ \vdots&\vdots&\ddots&\vdots&\vdots\\\ a_{M-1,1}&a_{M-1,2}&\ldots&a_{M-1,N-1}&a_{M-1,N}\\\ a_{M,1}&a_{M,2}&\ldots&a_{M,N-1}&a_{M,N}\\\ \end{array}\right)_{M\times N}.$ (6) In Eq. (6), $M=40,000$, $N=L\times L$, and $L=28$. $M$ and $N$ represent the number of the configurations and the lattices, respectively. Each row $\bm{R}_{i}$ ($i=1,2,\ldots,M$) in the matrix $\bm{X}$ is a configuration with one dimension, can be reshaped as the matrix $\bm{X}_{i}$ ($i=1,2,\ldots,M$) with the size of $L\times L$ (see Eq. 10). Furthermore, each column $\bm{C}_{j}$ ($j=1,2,\ldots,N$) in the matrix $\bm{X}$ represents one lattice with different configurations. Moreover, the element $a_{ij}$ ($i=1,2,\ldots,M;j=1,2,\ldots,N$) in matrix $\bm{X}$ and the element $b_{kl}$ ($k,l=1,2,\ldots,L$) in matrix $\bm{X}_{i}$ take 0 when the corresponding lattice is occupied and take 1 otherwise. $\displaystyle\bm{X}_{i}=\left(\begin{array}[]{ccccc}b_{11}&b_{12}&\ldots&b_{1L-1}&b_{1L}\\\ &&\vdots&&\\\ b_{L1}&b_{L2}&\ldots&b_{LL-1}&b_{LL}\\\ \end{array}\right)_{L\times L}.$ (10) Figure 1: (a) The relationship between the permeabilities $\\{0.41,0.42,\ldots,0.80\\}$ and the raw $\varPi(\bm{p},L)$. (b) The relationship between the permeabilities $\\{0.41,0.42,\ldots,0.80\\}$ and the raw $P(\bm{p},L)$. Except for the raw configurations, the dataset also encompasses order parameter. In the two-dimensional percolation model, order parameter includes the percolation probability $\varPi(\bm{p},L)$ and the density of the spanning cluster $P(\bm{p},L)$. $\varPi(\bm{p},L)$ refers to the probability that there is one connected path from one side to another in $\bm{X}_{i}$. That is to say, $\varPi(\bm{p},L)$ is a function of the permeability $\bm{p}$ in the system with the size of $L\times L$ (see Fig. 1(a)). With a connectivity of 4, we can identify the cluster for each lattice $b_{kl}$. The clusters are marked sequentially with an unique index. Note that the two lattices having the same index belong to the same cluster. If there are more than one cluster, the greatest cluster is chosen as the result. In this way, we can count up how many times the configurations $\\{\bm{X}_{1},\bm{X}_{2},\ldots,\bm{X}_{M}\\}$ are percolated for given $\bm{p}$ and $L$. For each permeability $p$, $\varPi(\bm{p},L)_{p}$ is expressed in Eq. (11). $\displaystyle\varPi(\bm{p},L)_{p}=\frac{1}{1000}\sum_{r=1}^{1000}{S_{r}},$ (11) Where $S_{r}$ refers that the two-dimensional configurations $\\{\bm{X}_{e},\bm{X}_{2\times e},\ldots,\bm{X}_{1000\times e}\\}$ for the permeability $p$ is percolated or not. Clearly, $S_{r}$ takes 0 if the corresponding lattice is occupied and takes 1 otherwise. $r=\\{1,2,\ldots,1000\\}$, $p\in\\{0.41,0.42,\ldots,0.80\\}$, $e=(p-0.40)\times 100$, and $\varPi(\bm{p},L)_{p}\in[0,1]$. Another representation of order parameter is the density of the spanning cluster $P(\bm{p},L)$. In contrast to the $\varPi(\bm{p},L)$, $P(\bm{p},L)$ is associated with spanning cluster. Therefore, for all configurations $\\{\bm{X}_{1},\bm{X}_{2},\ldots,\bm{X}_{M}\\}$, $P(\bm{p},L)$ is characterized by that whether or not each lattice $b_{kl}$ belongs to the total spanning cluster. Similarly, $P(\bm{p},L)$ is a function of the permeability $\bm{p}$ in the system with the size of $L\times L$ (see Fig. 1(b)). For each permeability $p$, $P(\bm{p},L)_{p}$ is expressed in Eq. (12). $\displaystyle P(\bm{p},L)_{p}=\frac{1}{1000\times L\times L}\sum_{r=1}^{1000}{S_{r}^{{}^{\prime}}}.$ (12) Where $S_{r}^{{}^{\prime}}$ counts up the total number of lattices that belong to the spanning cluster for each configuration in $\\{\bm{X}_{e},\bm{X}_{2\times e},\ldots,\bm{X}_{1000\times e}\\}$ for the permeability $p$. Obviously, $r=\\{1,2,\ldots,1000\\}$, $p\in\\{0.41,0.42,\ldots,0.80\\}$, $e=(p-0.40)\times 100$, $0\leq S_{r}^{{}^{\prime}}<L\times L$, and $P(\bm{p},L)_{p}\in[0,1)$. ## III Machine learning methods ### III.1 CNNs In this section, we will focus on the two-dimensional percolation model and the dataset obtained by the Monte Carlo simulation. This section will discuss several machine learning approaches, including CNNs, VAE and cVAE, and PCA, to deduce physical characteristics. Let us first introduce CNNs. CNNs, supervised learning methods, are particularly useful in solving realistic problem for many disciplines, such as physics[3], chemistry [19], medicine [20], economics [21], biology [22], and geophysics [23, 24], ect. In the applications of phase transitions and critical phenomena, many predecessors utilize CNNs to detect physical features, especially order parameter [1, 25, 26, 27, 10, 5, 2, 6]. In this study, the four kinds of CNNs are not only used to detect the two order parameters ($\varPi(\bm{p},L)$ and $P(\bm{p},L)$), but also to detect the permeability $\bm{p}$. Figure 2: The structure of the CNNs with four layers, including “Conv1”, “Conv2”, “FC”, and “Output”. The square with black and white lattices is percolating configurations $\\{\bm{X}_{1},\bm{X}_{2},\ldots,\bm{x}_{M}\\}$. The light yellow cuboids (“Conv1” and “Conv2”) stand for convolution layers. The bright orange cuboids stand for max-pooling layers. The light purple cuboid “FC” refer to a fully connected layer. The input layer “Input” has the size of $28\times 28$. The first layer “Conv1” with “filter1” filters has the size of “size1”$\times$“size1”. The second layer “Conv2” with “filter2” filters has the size of “size2”$\times$“size2”. The layer “Flatten” owns the size of “size3”. The third layer “FC” owns the size of “size4”. And the last layer “Output” represents for a fully connected layer with one neuron. Next, we demonstrate the architecture of the CNNs (see Fig. 2). The structure of the CNNs has four layers, including two convolution layers and two fully connected layers. The percolating configurations $\\{\bm{X}_{1},\bm{X}_{2},\ldots,\bm{X}_{M}\\}$ are taken as inputs. Consequently, the CNNs receive the corresponding order parameters ($\varPi(\bm{p},L)$ and $P(\bm{p},L)$) or the permeability $\bm{p}$ as outputs. ### III.2 VAE and cVAE In the former section (see Sec. III.1), the raw configuration $\bm{X}$ at different permeability $\bm{p}$ is generated by Monte Carlo simulation. However, what if configurations are not in the raw configurations, can we still identify the physical features as well? Here we consider to use the VAE and the cVAE to generate new configuration $\hat{\bm{X}}_{\text{VAE}}$ and $\hat{\bm{X}}_{\text{cVAE}}$, respectively. Figure 3: The structure of the VAE with an encoder and a decoder. The left large light purple cuboid refers to the encoder with two fully connected layers, i.e., “FC1”, and “$\bm{\mu}$” or “$\bm{\sigma}$”, with the size of “size1”, and “size2”. And the right large light red cuboid is the decoder with two fully connected layers, including “FC2”, and “Output” with the size of “size3”, and 784. The outputs of the encoder are the mean value “$\bm{\mu}$” and the standard deviation “$\bm{\sigma}$”. The input “$\bm{Z}$” of the decoder is sampled from the normal distribution with “$\bm{\mu}$” and “$\bm{\sigma}$”. The green cuboid consists of “$\bm{\mu}$”, “$\bm{\sigma}$”, and “$\bm{Z}$”. The rectangles with 784 black and white lattices represent percolating configuration $\bm{X}$ on the left and its reconstruction $\hat{\bm{X}}$ on the right, respectively. VAE (see Fig. 3), a generative network, bases on the variational Bayes inference proposed by [28]. Contracted with traditional AE (see Fig. S. 1), the VAE describes latent variables with probability. From that point, the VAE shows great values in data generation. Just like AE, the VAE is composed of an encoder and a decoder. The VAE uses two different CNNs as two probability density distributions. The encoder in the VAE, called the inference network $p_{\text{encoder}}(\bm{Z}|\bm{X})$, can generate the latent variables $\bm{Z}$. And the decoder in the VAE, called the generating network $p_{\text{decoder}}(\hat{\bm{X}}|\bm{Z})$, reconstructs the raw configuration $\bm{X}$. Unlike AE, the encoder and the decoder in VAE are constrained by the two probability density distributions. Figure 4: The structure of the cVAE with an encoder and a decoder. The left large light purple cuboid refers to the encoder with three layers, including the input layer “Input” with the size of “28$\times$28”, two convolution layers (“Conv1” and “Conv2”) with the size of “filter1$\times$size1$\times$size1” and “filter2$\times$size2$\times$size2”, a flatten layer “Flatten” with the size of “filter2$\times$size2$\times$size2”, and a fully connected layer (“$\bm{\mu}$” or “$\bm{\sigma}$”) with the size of “size3”. And the right large light red cuboid is the decoder with four layers, including the layer “$\bm{Z}$” with the size of “size3”, a fully connected layer “FC” with the size of “filter1$\times$size4$\times$size4”, a reshape layer “Reshape” with the size of “filter1$\times$size4$\times$size4”, two transposed convolution layers (“Conv3” and “Conv4”) with the size of “filter2$\times$size5$\times$size5” and “filter1$\times$28$\times$28”, and a output layer “Output” with the size of “28$\times$28”. The outputs of the encoder are the mean value “$\bm{\mu}$” and the standard deviation “$\bm{\sigma}$”. The input of the decoder “$\bm{Z}$” is sampled from the normal distribution with “$\bm{\mu}$” and “$\bm{\sigma}$”. The green cuboid is consist of “$\bm{\mu}$”, “$\bm{\sigma}$”, and “$\bm{Z}$”. The squares with 784 black and white lattices represent percolating configurations $\\{\bm{X}_{1},\bm{X}_{2},\ldots,\bm{X}_{M}\\}$ on the left and their reconstructions $\\{\hat{\bm{X}}_{1},\hat{\bm{X}}_{2},\ldots,\hat{\bm{X}}_{M}\\}$ on the right, respectively. To better reconstruct the raw two-dimensional configuration $\bm{X}$, the fully connected layer in the VAE is replaced by the convolution layer. Now we have got the cVAE. The architecture of the cVAE is shown in Fig. 3. Usually, the performance of the cVAE is better than the VAE due to the configuration $\bm{X}$ with spatial attribute. In our work, the VAE and the cVAE are both used for generating THE new configuration $\hat{\bm{X}}$. ### III.3 PCA The Sec. III.1 and Sec. III.2 focus on supervised learning, which hypothesize that the labels exist for the raw configuration $\bm{X}$ on the percolation model. However, though we can detect the permeability values $\\{0.41,0.42,\ldots,0.80\\}$ and the two order parameters ($\varPi(\bm{p},L)$ and $P(\bm{p},L)$) by supervised learning, label dearth often occurs. Thus, it is imperative to identify the labels, such as $\varPi(\bm{p},L)$, $P(\bm{p},L)$, and $\bm{p}$. Some recent studies have shown that the first principal component obtained by PCA can be regarded as typical physical quantities [15, 8, 14, 17]. Base on these studies, we explore the meaning of the first principal component on the percolation model. As it is well-known, PCA can reduce the dimension of the matrix $\bm{X}$. First, we compute the mean value $\bm{X}_{\text{mean}}=1/M\sum_{i=1}^{M}a_{ij}(i=1,2,\ldots,M;j=1,2,\ldots,N)$ for each column $\bm{C}_{j}(j=1,2,\ldots,N)$ in the matrix $\bm{X}$. Then we get the centered matrix $\bm{X}_{\text{centered}}$ that is expressed as $\bm{X}_{\text{centered}}=\bm{X}-\bm{X}_{\text{mean}}$. After obtaining $\bm{X}_{\text{centered}}$, by an orthogonal linear transformation expressed as $\bm{X}_{\text{centered}}^{T}\bm{X}_{\text{centered}}\bm{W}=\bm{\lambda}\bm{W}$, we extract the eigenvectors $\bm{W}$ and the eigenvalues $\bm{\lambda}$. The eigenvectors $\bm{W}$ are composed with $\bm{w}_{1},\bm{w}_{2},\ldots,\bm{w}_{N}$. The eigenvalues are sorted in the descending order, i.e., $\bm{\lambda}_{1}\geq\bm{\lambda}_{2}\geq\ldots\geq\bm{\lambda}_{N}\geq 0$. The normalized eigenvalues $\tilde{\bm{\lambda}}_{j}(j=1,2,\ldots,N)$ are expressed as $\bm{\lambda}_{j}/\sum_{j=1}^{N}{\bm{\lambda}_{j}}$. The row $\bm{R}_{i}$ in matrix $\bm{X}$ can be transformed into $\bm{X}^{{}^{\prime}}_{i}=\bm{R}_{i}\bm{W}$. Eq. 13 represents the statistic average every 40 intervals for each permeability $p$. This process is quite similar to the process of calculating the two order parameters, i.e., $\varPi(\bm{p},L)$ and $P(\bm{p},L)$. Table 1 shows the procedure of PCA algorithm. $\langle\bm{X}^{{}^{\prime}}\rangle=\frac{1}{1000}\sum_{i=1}^{1000}|\bm{X}^{{}^{\prime}}_{i\times 40}|$ (13) Table 1: The procedure of PCA Algorithm Require: the raw configuration $\bm{X}$ --- 1\. Compute the mean value $\bm{X}_{\text{mean}}$ for the column $\bm{C}_{j}$ in the matrix $\bm{X}$; 2\. Get the centered matrix $\bm{X}_{\text{centered}}$; 3\. Compute the eigenvectors $\bm{W}$ and the eigenvalues $\bm{\lambda}$ by an orthogonal linear transformation; 4\. Transform $\bm{R}_{i}$ into $\bm{X}^{{}^{\prime}}_{i}$; 5\. Get the statistic average $\langle\bm{X}^{{}^{\prime}}\rangle$ with every 40 intervals for each permeability $p$. ## IV Results and Discussion ### IV.1 Simulate the two order parameters by two CNNs In this section, we consider to use the approches in Sec. III to capture the physic features. First we make use of TensorFlow 2.2 library to perform the CNNs with four layers. To predict the two order parameters ($\varPi(\bm{p},L)$ and $P(\bm{p},L)$), two kinds of CNNs (CNNs-I and CNNs-II) are constructed. The first two layers of the former two kinds of CNNs are composed of two convolution layers (“Con1” and “Con2”), both of which possessed “filter1”=32 and “filter2”=64 filters with the size of $3\times 3$, and a stride of 1. Each convolution layer is followed by a max-pooling layer with the size of $2\times 2$. The final convolution layer “Con2” is strongly interlinked to a fully connected layer “FC” with 128 variables. The output layer “Output”, following by “FC”, is a fully connected layer. For the two convolution layers and the fully connected layer “FC”, a rectified linear unit (ReLU) $\bm{a}=\text{max}(0,\bm{x})$ [29] is chosen as activation function due to its reliability and validity. However, the output layer has no activation function. After determining the framework of CNNs-I and CNNs-II, here we mention how to train CNNs-I and CNNs-II for deducing $\varPi(\bm{p},L)$ and $P(\bm{p},L)$. First, we carry out an Adam algorithm [30] as an optimizer to update parameters, i.e., weights and biases. Then, a mini batch size of 256 and a learning rate of $10^{-4}$ are selected for its timesaving. Following this treatment, CNNs-I and CNNs-II are trained on 1000 epochs for 40,000 uncorrelated and shuffled configurations, respectively. Before training, we split $\\{\bm{X}_{1},\bm{X}_{2},\ldots,\bm{X}_{M}\\}$, $\varPi(\bm{p},L)$ and $P(\bm{p},L)$ into 32,000 training set and 8,000 testing set. While training, we monitor three indicators, including the loss function (i.e., mean squared error (MSE, see Eq. 14)), mean average error (MAE, see Eq. 15), and root mean squared error (RMSE, see Eq. 16), for training and testing set [24]. In Eq. 14-16, ${y_{i}}^{\text{raw}}$ and ${y_{i}}^{\text{pred}}$ refer to $\varPi(\bm{p},L)$/$P(\bm{p},L)$ and its predictions. If the loss function in testing set reaches the minimum, then the optimal CNNs-I and CNNs-II will be obtained. As can be seen from Fig. S. 2, these indicators gradually decrease. In Table. S. 3, the errors of the optimal CNNs-I and CNNs-II are very small. What stands out in Fig. S. 2 is that CNNs-I and CNNs-II have high stability, consistency, and faster convergence rate. $\mathrm{MSE}=\frac{1}{M}\sum_{i=1}^{M}({y_{i}}^{\text{pred}}-{y_{i}}^{\text{raw}})^{2}\ $ (14) $\mathrm{MAE}=\frac{1}{M}\sum_{i=1}^{M}|{y_{i}}^{\text{pred}}-{y_{i}}^{\text{raw}}|$ (15) $\mathrm{RMSE}=\sqrt{\frac{1}{M}\sum_{i=1}^{M}({y_{i}}^{\text{pred}}-{y_{i}}^{\text{raw}})^{2}}$ (16) Figure 5: (a) The relationship between the permeability values $\\{0.41,0.42,\ldots,0.80\\}$ and the raw $\varPi(\bm{p},L)$ or the statistic average from the outputs of CNNs-I. (b) The relationship between the permeability values $\\{0.41,0.42,\ldots,0.80\\}$ and the raw $P(\bm{p},L)$ or the statistic average from the outputs of CNNs-II. Before assess CNNs-I and CNNs-II, we have to explain what is meant by statistic average. Statistic average can be defined as the averages of CNNs-I’s or CNNs-II’s outputs for each permeability $p$. As shown in Fig. 5, there is a clear trend of phase transition between the permeability values $\\{0.41,0.42,\ldots,0.80\\}$ and $\varPi(\bm{p},L)$/$P(\bm{p},L)$. The two grey lines in Fig. 5, which are the same as the two blue lines in Fig. 1, represent the relationship between the permeability values $\\{0.41,0.42,\ldots,0.80\\}$ and the raw $\varPi(\bm{p},L)$ or $P(\bm{p},L)$. Likewise, the two blue lines in Fig. 5, represent the relationship between the permeability values $\\{0.41,0.42,\ldots,0.80\\}$ and the statistic average from the outputs of CNNs-I and CNNs-II. The overlapping of the two kinds of lines shows that CNNs-I and CNNs-II can be used to deduce the two order parameters and the process of phase transition. To overcome the difficulty associated with the percolation model being near the critical transition point, we truncate the dataset. Specifically, we remove the data near the critical transition point, and only retain the data far away from the critical transition point. Here we take the simulation of $\varPi(\bm{p},L)$ as an example. The retained data with the raw $\varPi(\bm{p},L)$ rangs from 0 to 0.1, and 0.9 to 1. As shown in Fig. S. 3 and the middle red points in Fig. 6, we find that CNNs-I can extrapolate $\varPi(\bm{p},L)$ to missing data by learning the retained data. Figure 6: Identification of phase transition with truncated dataset by CNNs-I. The permeability values $\\{0.41,0.42,\ldots,0.80\\}$ and $\varPi(\bm{p},L)$ connected with red points are artificially removed from the dataset. And CNNs-I makes the judgment only by learning the data associated with blue points. The grey curve show that $\varPi(\bm{p},L)$ shifts with the permeability values $\\{0.41,0.42,\ldots,0.80\\}$. This section demonstrate that CNNs-I and CNNs-II can be two effective tools for detecting $\varPi(\bm{p},L)$ and $P(\bm{p},L)$, respectively. Additional test should be made to verify that whether or not CNNs-I and CNNs-II are robust against noise. To address this issue, we deliberately invert a proportion, i.e., 5%, 10%, and 20%, of the labels for the raw $\varPi(\bm{p},L)$ and $P(\bm{p},L)$ and verify that whether or not the “artificial” noises can affect the predicted $\varPi(\bm{p},L)$ and $P(\bm{p},L)$. Fig. 7 and Fig. S. 4 demonstrate that CNNs-I and CNNs-II are robust against noise. As the labeling error rates increase, the same trend is evident in the outputs of CNNs-I and CNNs-II within a relatively small difference (see Fig. 7). Therefore, we draw the conclusion that noises have little effect on detecting $\varPi(\bm{p},L)$ and $P(\bm{p},L)$. Figure 7: Robustness of the two order parameters ($\varPi(\bm{p},L)$ and $P(\bm{p},L)$) with noises for CNNs-I and CNNs-II. (a) The relationship between the permeability values $\\{0.41,0.42,\ldots,0.80\\}$ and the raw $\varPi(\bm{p},L)$ or the statistic average from the outputs of CNNs-I under different noisy inputs. (b) The relationship between the permeability values $\\{0.41,0.42,\ldots,0.80\\}$ and the raw $P(\bm{p},L)$ or the statistic average from the outputs of CNNs-II under different noisy inputs. ### IV.2 Simulate the permeability by one CNNs Just like we use CNNs-I and CNNs-II in Sec. IV.1, here we use the same structure for CNNs-III and strategies for training the permeability $\bm{p}$. The only distinction between CNNs-III and CNNs-I/CNNs-II is that the outputs for CNNs-III are the permeability $\bm{p}$ instead of the two order parameters. Fig. S. 5 shows the performance of CNNs-III. With successive increases in epochs, the MSE, MAE, and RMSE continue to decrease until no longer dropping. Another measure of CNNs-III’s performance is concerned with the difference between the raw permeability $\bm{p}$ and its prediction $\hat{\bm{p}}$. The blue circles in Fig. 8 show that there is a strong positive correlation between the raw permeability $\bm{p}$ and its prediction $\hat{\bm{p}}$. Further statistical tests reveal that most of the gap between the raw permeability $\bm{p}$ and its prediction $\hat{\bm{p}}$ is less than 0.1. The result proves that CNNs-III has the advantage of convenient use and high precision. Figure 8: The correlation between the raw permeability $\bm{p}$ and its predictions. The blue circles refer to the relationship between the raw permeability $\bm{p}$ in dataset and its predictions $\hat{\bm{p}}$ by CNNs- III. The red circles refer to the relationship between the raw permeability $\bm{p}_{\text{ex}}$ out of dataset and its prediction $\hat{\bm{p}}_{\text{extrapolated}}$ by CNNs-III. The raw permeability $\bm{p}_{\text{ex}}$ out of dataset not only range from 0.01 to 0.4, and from 0.81 to 1.0, with an interval of 0.01. The black line refers to the correlation between the raw permeabilities $\\{\bm{p},\bm{p}_{\text{extrapolated}}\\}$ ranges from 0.01 to 1.0 with an interval of 0.01 and their average predictions. On the other hand, extrapolation ability can also reflect the performance of CNNs-III. To do this, we use Monte Carlo simulation to generate new dataset. The new permeability $\bm{p}$ not only ranges from 0.01 to 0.40, but also from 0.81 to 1.00, with an interval of 0.01. Just like Sec. II, we perform 1050 Monte Carlo steps and keep the last 1000 steps. As a consequence, 60,000 configurations are generated. The red circles in Fig. 8 exhibit the extrapolation ability of CNNs-III. Our results show that no significant correlation between the extrapolated permeability $\bm{p}_{\text{ex}}$ ranging from 0.81 to 1.00 and their prediction $\hat{\bm{p}}_{\text{extrapolated}}$. However, in Fig. 8, the results indicate that CNNs-III has good extrapolation ability for the extrapolated permeability $\bm{p}_{\text{ex}}$ from 0.01 to 0.40. ### IV.3 Generate new configurations by one VAE and one cVAE Though CNNs-I, CNNs-II, and CNNs-III are valid when detecting the two order parameters ($\varPi(\bm{p},L)$ and $P(\bm{p},L)$) and the permeability $\bm{p}$, the validity of these three CNNs is unkown for percolating configurations outside of the dataset. As is shown in Fig. 3 and Fig. 4, we use the same network structures for the VAE and the cVAE to generate new configurations ($\hat{\bm{X}}_{\text{vae}}$ and $\hat{\bm{X}}_{\text{cvae}}$). Actually, $\hat{\bm{X}}_{\text{vae}}$ and $\hat{\bm{X}}_{\text{cvae}}$ can be regarded as adding some noise into the raw configuration $\bm{X}$. Let us first consider VAE. Just like AE (see Fig. S. 1), the VAE is also composed of an encoder and a decoder. The encoder of the VAE owns two fully connected layers, both of which follows with a ReLU activation function. The first layer of the encoder possess “size1”=512 neurons. Another 512 neurons, including 256 mean “$\bm{\mu}$” and 256 variance “$\bm{\sigma}$”, are taken into account for the second layer of the encoder. By resampling from the Gaussian distribution with the mean “$\bm{\mu}$” and the variance “$\bm{\sigma}$”, we obtain 256 latent variables “$\bm{Z}$” which are the inputs of the decoder. For the decoder of the VAE, two fully connected layers follow with the outputs of the encoder. For symmetry, the first layer of the decoder also contains “size1”=512 neurons and follows with a ReLU activation function. And the output layer of the decoder contains 784 neurons which are used to reconstruct the raw configuration $\bm{X}$. Thus, the neurons in the output layer are the same as that in the input layer. Moving on now to consider cVAE. The encoder of the cVAE is composed of one input layer, two hidden convolution layers with “filter1”=32 and “filter2”=64 filters with the size of 3 and a stride of 2, and a ReLU activation function. The output layer in the encoder is a fully connected flatten layer with 800 neurons (400 mean “$\bm{\mu}$” and 400 variance “$\bm{\sigma}$”) without activation function. By resampling from the Gaussian distribution with the mean “$\bm{\mu}$” and the variance “$\bm{\sigma}$”, we obtain 400 latent variables “$\bm{Z}$”. The reason why latent variables in the cVAE is more than the VAE is that the cVAE needs to consider more complex spatial characteristic. The decoder of the cVAE is composed of an input layer with 400 latent variables “$\bm{Z}$”. A fully connected layer “FC” with 1,568 neurons is followed by “$\bm{Z}$”. After reshaping the outputs of “FC” into three dimension, we feed the data into two transposed convolution layers (“Conv3” and “Conv4”) and one output layer “Output”. The filters in these deconvolution layers are 64, 32 and 1 with the size of 3 and the stride of 2. After excluding the output layer “Output” without activation functions, there exist two ReLU activation functions followed by “Conv3” and “Conv4”, respectively. We train the VAE and the cVAE over $10^{3}$ epochs using the Adam optimizer, a learning rate of $10^{-3}$, and a mini batch size of 256. To train the VAE and the cVAE, we use the sum of binary cross-entropy (see Eq. 17) and the Kullback-Leibler (KL) divergence (see Eq. 18) as the loss function [18]. In Eq. 17-18, $\bm{x}_{i}^{\text{raw}}$ and $\bm{x}_{i}^{\text{pred}}$ represent each raw configuration with one/two dimension and its prediction. As shown in Fig. S. 6, the loss function, the binary cross-entropy, and the KL divergence vary with epochs for the VAE and the cVAE. Here we focus on the minimum value of loss function. From Fig. S. 6, the optimal cVAE performs better than the optimal VAE. $\displaystyle\text{BinaryCrossEntropy}=-\sum_{i=1}^{M}((\bm{x}_{i}^{\text{pred}}\times\text{log}(\bm{x}_{i}^{\text{raw}})$ $\displaystyle+(1-\bm{x}_{i}^{\text{pred}})\times\text{log}(1-\bm{x}_{i}^{\text{raw}})).$ (17) $\text{KL}_{\text{divergence}}=-\sum_{i=1}^{M}\Biggl{(}\bm{x}_{i}^{\text{raw}}\times\text{log}\left(\frac{\bm{x}_{i}^{\text{raw}}}{\bm{x}_{i}^{\text{pred}}}\right)\Biggr{)}.$ (18) For a more visual comparison, we show the snapshots of the raw configurations $\\{\bm{X}_{1},\bm{X}_{2},\ldots,\bm{X}_{M}\\}$, and compare them to the VAE- generated and cVAE-generated configurations in Fig. 9. As we can see from Fig. 9, the configurations from the Monte Carlo simulation, the VAE and the cVAE are very close to each other. Figure 9: Snapshots of percolating configurations for $p\in\\{0.46,0.60,0.75\\}$. The configurations in the top, middle, and bottom panels are sampled from the Monte Carlo simulation, the VAE, and the cVAE, respectively. Figure 10: The relationship between the permeability values $\\{0.41,0.42,\ldots,0.80\\}$ and the $\varPi(\bm{p},L)$ or the statistic averages from the outputs of CNNs-I that originate from the outputs of the VAE and the cVAE. (b) The relationship between the permeability values $\\{0.41,0.42,\ldots,0.80\\}$ and $P(\bm{p},L)$ or the statistic averages from the outputs of CNNs-II that originate from the outputs of the VAE and the cVAE. After reconstructing the configurations ($\hat{\bm{X}}_{\text{vae}}$ and $\hat{\bm{X}}_{\text{cvae}}$) through the VAE and the cVAE and pouring $\hat{\bm{X}}_{\text{vae}}$ and $\hat{\bm{X}}_{\text{cvae}}$ into CNNs-I, CNNs-II, and CNNs-III, we can detect $\varPi(\bm{p},L)$, $P(\bm{p},L)$, and the permeability $\bm{p}$. Fig. 10 shows the relationships between the permeability values $\\{0.41,0.42,\ldots,0.80\\}$ and the statistic average from the outputs in CNNs-I and CNNs-II, respectively. From the red and purple lines in Fig. 10, by using $\hat{\bm{X}}_{\text{vae}}$ and $\hat{\bm{X}}_{\text{cvae}}$, we can obtain the two order parameters as well. From the Fig. 11, the raw permeability $\bm{p}$ is remarkably correlated linearly with its prediction $\hat{\bm{p}}$ through the VAE/cVAE and CNNs-III. Thus, subtle change in the raw configurations does not effect the catch of physical features. Figure 11: The correlation between the raw permeability $\bm{p}$ and its predictions through VAE/cVAE and CNNs-III. The blue circles refers to the relationship between the raw permeability $\bm{p}$ and its predictions by VAE/cVAE and CNNs-III. And the black line refers to the correlation between the raw permeability $\bm{p}$ and its average predictions. ### IV.4 Identify the characteristic of the first principal component by one PCA This section will discuss how to capture the physic characteristics without labels. Various studies suggest to use PCA to identify order parameters [15, 8, 14, 17]. Therefore, we try to verify the feasibility of PCA in capturing order parameter on the percolation model. First, we perform the PCA to the raw configuration $\bm{X}$. Fig. 12 exhibit the $N$ normalized eigenvalues $\tilde{\bm{\lambda}}_{n}=\bm{\lambda}_{n}/\sum_{n=1}^{N}\bm{\lambda}_{n}$. $\tilde{\bm{\lambda}}_{n}$ is also called as the explained variance ratios. The most noteworthy information in Fig. 12 is that there is one dominant principal component $\tilde{\bm{\lambda}}_{1}$, whcih is the largest one among $\tilde{\bm{\lambda}}_{n}$ and much larger than other explained variance ratios. Thus, $\tilde{\bm{\lambda}}_{1}$ plays a key role when dealing with dimension reduction. Based on $\tilde{\bm{\lambda}}_{1}$, the raw configuration $\bm{X}$ are mapped to another matrix $\bm{Y}=\bm{X}\tilde{\bm{\lambda}}_{1}$. Figure 12: The explained variance ratios obtained from the raw configuration $\bm{X}$ by the PCA, with the horizontal axis indicating corresponding component labels. The largest value of the explained variance ratios locating at the top-left corner means that there exists one dominant principle component. In Fig 13, we construct the matrix $\bm{Y}^{{}^{\prime}}=\\{{\bm{X}\tilde{\bm{\lambda}}_{1},\bm{X}\tilde{\bm{\lambda}}_{2}}\\}$ by the first two eigenvalues and their eigenvectors. We use 40,000 blue scatter points to plot the the relationship between $\bm{X}\tilde{\bm{\lambda}}_{1}$ and $\bm{X}\tilde{\bm{\lambda}}_{2}$ on 40 permeability values ranging from 0.41 to 0.8. Just like in Fig. 12, there is only one dominant representation on the percolation model due to the first principal component is much more important than the second principal component. Figure 13: Projection of the raw configuration $\bm{X}$ onto the plane of the first two dominant principal components, i.e., $\bm{X}\tilde{\bm{\lambda}}_{1}$ and $\bm{X}\tilde{\bm{\lambda}}_{2}$. The color bar on the right indicates the permeability values $\\{0.41,0.42,\ldots,0.80\\}$. Having analysed the importance of the first principal component, we now move on to discuss the meaning of the first principal component. In Fig. 14, we focus on the quantified first principal component as a function of the permeability $\bm{p}$ and the two order parameters, i.e., $\varPi(\bm{p},L)$ and $P(\bm{p},L)$. From Fig. 14, we can see that there is a strong linear correlation between the quantified first principal component and the permeability $\bm{p}$. And the relationships between the quantified first principal component and two order parameters ($\varPi(\bm{p},L)$ and $P(\bm{p},L)$) are similar to the relationships between the permeability values $\\{0.41,0.42,\ldots,0.80\\}$ and the two order parameters. Our results are significant different from former study which demonstrates that the quantified first principal component can be taken as order parameter by data preprocessing [17]. A possible explanation may be that [17] evaluates the first principal component by removing the raw dataset with certain attributes on the percolation model. Therefore, we assume that the quantified first principal component obtained by PCA may not be taken as order parameter for various physical models. Another possible explanation is that, under certain conditions, the quantified first principal component can be regarded as order parameter. Figure 14: Taking the normalized quantified first principal component $\bm{X}\tilde{\bm{\lambda}}_{1}$ as a function of the permeability $\bm{X}$ and the two order parameters, i.e., $\varPi(\bm{p},L)$ and $P(\bm{p},L)$. ### IV.5 Identify physic characteristics by one $k$-means and one CNNs From Sec. IV.4, no significant corresponding is found between the normalized quantified first principal component and the two order parameters. Here we eagerly wonder how to identify order parameter. And another physical characteristics we desire to explore is the critical transition point. So we try to find a way to capture order parameter from the raw configurations $\\{\bm{X}_{1},\bm{X}_{2},\ldots,\bm{X}_{M}\\}$ and their permeability $\bm{p}$ on the percolation model. As it is well-known, for the two-dimensional percolating configurations, the critical transition point is equal to 0.593 in theoretical calculation. Though the critical transition point is already known, we wonder that whether or not the critical transition point can be found by machine. To do this, we first use a cluster analysis algorithm named $k$-means [31] to separate the raw configurations $\\{\bm{X}_{1},\bm{X}_{2},\ldots,\bm{X}_{M}\\}$ into two categories. The minimum and maximum value of their permeability $\bm{p}$ are 0.55 and 0.80 in the first cluster, and 0.41 and 0.65 in the second cluster. Note that there is an overlapping interval between 0.55 and 0.65 for the two categories. According to the overlapping interval, we hypothesize that the critical threshold $p_{c}$ is set to be 11 values, i.e., 0.55, 0.56, $\ldots$, and 0.65. For each raw configuration, if the permeability is smaller than $p_{c}$, there will exist no percolating cluster and its label will be marked as 0; otherwise, there will exist at least one percolating cluster and its label will be marked as 1. To detect physic characteristics, we use the fourth kinds of CNNs (CNNs-IV) with the structure in Fig. 2 for 11 preset critical thresholds from 0.55 to 0.65. Note that the output layer use a sigmoid activation function expressed as $\bm{a}=1/(1+e^{-\bm{x}})$, to make sure the outputs are between 0 and 1. Another critical thing to pay attention is that the He normal distribution initializer [32] and L2 regularization [33] are used in the layer of “Conv1”, “Conv2”, and “FC” on CNNs-IV. To avoid overfiting, in addition to L2 regularization, we also use a dropout layer with a dropout rate of 0.5 on “FC”. A Mini batch size of 512 and a learning rate of $10^{-4}$ are chosen while training CNNs-IV. The binary cross-entropy (see Eq. 17) is taken as the loss function on CNNs-IV. Another metric, used to measure the performance of CNNs-IV, is the binary accuracy (see Eq. 19). The other hyper-parameters are the same as the CNNs-I, CNNs-II, and CNNs-III. $\displaystyle\text{BinaryAccuracy}=\sum_{i=1}^{M}\frac{n_{(y_{i}^{\text{pred}}==y_{i}^{\text{raw}})}}{n_{y_{i}^{\text{pred}}}}.$ (19) Turning now to the experimental evidence on the inference ability of capturing relevant physic features. After obtaining the well-trained CNNs-IV with high accuracy (see Fig. S. 7), we obtain the outputs by pouring the raw 40,000 configurations $\\{\bm{X}_{1},\bm{X}_{2},\ldots,\bm{X}_{M}\\}$ into CNNs-IV. The statistical average of the outputs is calculated according to the 40 independent permeability values $\\{0.41,0.42,\ldots,0.80\\}$. The results of the correlational analysis are shown in Fig. 15. We set the horizontal dashed line as the threshold value 0.5. Hence, each curve is divided into two parts by the horizontal dashed line. The lower part indicates that the percolation system is not penetrated; while the upper part implies that the percolation system is penetrated. The crosspoint, where the horizontal dashed line and the red curve are intersected, has a permeability value of 0.594, which is very close to the theoretical value of 0.593 that is marked by the vertical dashed line in Fig. 15. Remarkably, the critical transition point can be calculated by CNNs-IV with the preset value of 0.60 for the sampling interval of 0.01. Therefore, CNNs-IV with the preset threshold value of 0.60 is the most effective model. In further studies, the preset threshold value may need to be enhanced by smaller sampling intervals for higher precision. Figure 15: The 11 curves show that the average outputs shifts when the preset threshold changes from 0.55 to 0.65. The average outputs and the threshold of phase transitions deduce from different preset threshold values on the percolation model by CNNs-IV. ## V Conclusions As machine learning approaches have become increasingly popular in phase transitions and critical phenomena, predecessors have pointed out that these approaches can capture physic characteristics. However, previous studies about identifying physical characteristics, especially order parameter and critical threshold, need to be further mutually validated. To highlight the possibility of effectiveness by machine learning methods, we conduct a much more comprehensive research than predecessors to reassess the machine learning approaches in phase transitions and critical phenomena. Our results show the effectiveness of machine learning approaches in phase transitions and critical phenomena than previous researchers. Precisely, we use CNNs-I, CNNs-II and CNNs-III to simulate the two order parameters, and the permeability values. To identify whether or not CNNs-I and CNNs-II are robust against noise, we add a proportion of the noises for the two order parameters. To validate the robustness of CNNs-I, CNNs-II and CNNs-III, we also use VAE and cVAE to generate new configurations that are slightly different from their raw configurations. After pouring the new configurations into the CNNs-I, CNNs-II, and CNNs-III, we achieve the results that these models are robust against noise. However, after we use PCA to reduce the dimension of the raw configurations and make a statistically significant linear correlation between the first principal component and the permeability values, no statistically significant linear correlations are found between the first principal component and the two order parameters. Clearly, the first principal component fails to be regarded as an order parameter in the two-dimensional percolation model. To identify order parameter, we use the fourth kinds of CNNs, i.e., CNNs-IV. The results show that CNNs-IV can identify new order parameter when the preset threshold value is 0.60. Surprisingly, we find that the critical transition point value is 0.594 by CNNs-IV. Although these machine learning methods are valid to explore the physical characteristics in the percolation model, the current study may still have some inevitable limitations that prevent us from making an overall judgement by these methods on the other models of phase transitions and critical phenomena. In other words, it must be acknowledged that this research is based on the two-dimensional percolation model.We are not sure of the usefulness of applying our methods to the other models. Consequently, our methods in this study may open an opportunity to other models on phase transitions and critical phenomena for further research. ###### Acknowledgements. The authors gratefully thank Yicun Guo for revising the manuscript. We also thank Jie-Ping Zheng and Li-Ying Yu for helpful discussions and comments. The work of S.Cheng, H.Zhang and Y.-L.Shi is supported by Joint Funds of the National Natural Science Foundation of China (U1839207). The work of F.He and K.-D.Zhu is supported by the National Natural Science Foundation of China (No.11274230 and No.11574206) and Natural Science Foundation of Shanghai (No.20ZR1429900). ## References * Tanaka and Tomiya [2017] A. Tanaka and A. Tomiya, Detection of phase transition via convolutional neural networks, Journal of the Physical Society of Japan 86, 063001 (2017). * Carrasquilla and Melko [2017] J. Carrasquilla and R. G. Melko, Machine learning phases of matter, Nature Physics 13, 431 (2017). * Van Nieuwenburg _et al._ [2017] E. P. Van Nieuwenburg, Y.-H. Liu, and S. D. Huber, Learning phase transitions by confusion, Nature Physics 13, 435 (2017). * Shiina _et al._ [2020] K. Shiina, H. Mori, Y. Okabe, and H. K. Lee, Machine-learning studies on spin models, Scientific reports 10, 1 (2020). * Suchsland and Wessel [2018] P. Suchsland and S. Wessel, Parameter diagnostics of phases and phase transition learning by neural networks, Physical Review B 97, 174435 (2018). * Huembeli _et al._ [2018] P. Huembeli, A. Dauphin, and P. Wittek, Identifying quantum phase transitions with adversarial neural networks, Physical Review B 97, 134109 (2018). * Wetzel and Scherzer [2017] S. Wetzel and M. Scherzer, Machine learning of explicit order parameters: From the ising model to su(2) lattice gauge theory, Physical Review B 96 (2017). * Wetzel [2001] S. J. Wetzel, Unsupervised learning of phase transitions: From principal component analysis to variational autoencoders, Physical Review E 96, 022140 (2001). * Deng _et al._ [2017] D.-L. Deng, X. Li, and S. D. Sarma, Machine learning topological states, Physical Review B 96, 195145 (2017). * Bachtis _et al._ [2020] D. Bachtis, G. Aarts, and B. Lucini, Mapping distinct phase transitions to a neural network, Physical Review E 102, 053306 (2020). * Zhao and Fu [2019] X. Zhao and L. Fu, Machine learning phase transition: An iterative proposal, Annals of Physics 410, 167938 (2019). * Ni _et al._ [2019] Q. Ni, M. Tang, Y. Liu, and Y.-C. Lai, Machine learning dynamical phase transitions in complex networks, Physical Review E 100, 052312 (2019). * Ch’ng _et al._ [2018] K. Ch’ng, N. Vazquez, and E. Khatami, Unsupervised machine learning account of magnetic transitions in the hubbard model, Physical Review E 97, 013306 (2018). * Wetzel [2017] S. J. Wetzel, Discovering phases, phase transitions, and crossovers through unsupervised machine learning: A critical examination, Physical Review E 95, 062122 (2017). * Wang [2016] L. Wang, Discovering phase transitions with unsupervised learning, Physical Review B 94, 195105 (2016). * Zhang _et al._ [2018] W. Zhang, J. Liu, and T.-C. Wei, Machine learning of phase transitions in the percolation and xy models, Physical Review E 99 (2018). * Yu and Lyu [2020] W. Yu and P. Lyu, Unsupervised machine learning of phase transition in percolation, Physica A: Statistical Mechanics and its Applications 559, 125065 (2020). * D’Angelo and Böttcher [2020] F. D’Angelo and L. Böttcher, Learning the ising model with generative neural networks, Physical Review Research 2, 023266 (2020). * Janet _et al._ [2020] J. P. Janet, H. Kulik, Y. Morency, and M. Caucci, _Machine Learning in Chemistry_ (2020). * Plant and Barton [2020] D. Plant and A. Barton, Machine learning in precision medicine: lessons to learn, Nature Reviews Rheumatology 17 (2020). * Hull [2021] I. Hull, Machine learning and economics (2021) pp. 61–86. * Buchanan [2020] M. Buchanan, Machines learn from biology, Nature Physics 16, 238 (2020). * Poorvadevi _et al._ [2020] R. Poorvadevi, G. Sravani, and V. Sathyanarayana, An effective mechanism for detecting crime rate in chennai location using supervised machine learning approach, International Journal of Scientific Research in Computer Science, Engineering and Information Technology , 326 (2020). * Cheng _et al._ [2021] S. Cheng, X. Qiao, Y. Shi, and D. Wang, Machine learning for predicting discharge fluctuation of a karst spring in north china, Acta Geophysica , 1 (2021). * Kashiwa _et al._ [2019] K. Kashiwa, Y. Kikuchi, and A. Tomiya, Phase transition encoded in neural network, Progress of Theoretical and Experimental Physics 2019, 083A04 (2019). * Arai _et al._ [2018] S. Arai, M. Ohzeki, and K. Tanaka, Deep neural network detects quantum phase transition, Journal of the Physical Society of Japan 87, 033001 (2018). * Xu _et al._ [2019] R. Xu, W. Fu, and H. Zhao, A new strategy in applying the learning machine to study phase transitions, arXiv preprint arXiv:1901.00774 (2019). * Kingma and Welling [2014] D. P. Kingma and M. Welling, Auto-encoding variational bayes, CoRR abs/1312.6114 (2014). * Agarap [2018] A. F. Agarap, Deep learning using rectified linear units (relu), arXiv preprint arXiv:1803.08375 (2018). * Kingma and Ba [2014] D. P. Kingma and J. Ba, Adam: A method for stochastic optimization, arXiv preprint arXiv:1412.6980 (2014). * Hartigan and Wong [1979] J. A. Hartigan and M. A. Wong, Algorithm as 136: A k-means clustering algorithm, Journal of the royal statistical society. series c (applied statistics) 28, 100 (1979). * [32] https://www.tensorflow.org/api_docs/python/tf/keras/initializers/HeNormal. * L [2] https://www.tensorflow.org/api_docs/python/tf/keras/regularizers/L2.
# Rationality of even-dimensional intersections of two real quadrics Brendan Hassett Department of Mathematics Brown University Box 1917 151 Thayer Street Providence, RI 02912 USA brendan<EMAIL_ADDRESS>, János Kollár Department of Mathematics Princeton University Fine Hall, Washington Road Princeton NJ 08544-1000 USA<EMAIL_ADDRESS>and Yuri Tschinkel Courant Institute New York University New York, NY 10012 USA Simons Foundation 160 Fifth Avenue New York, NY 10010 USA<EMAIL_ADDRESS> ###### Abstract. We study rationality constructions for smooth complete intersections of two quadrics over nonclosed fields. Over the real numbers, we establish a criterion for rationality in dimension four. ## 1\. Introduction Consider a geometrically rational variety $X$, smooth and projective over a field $k$. Is $X$ rational over $k$? A necessary condition is that $X(k)\neq\emptyset$, which is sufficient in dimension one, as well as for quadric hypersurfaces and Brauer-Severi varieties of arbitrary dimension. When the dimension of $X$ is at most two, rationality over $k$ was settled by work of Enriques, Manin, and Iskovksikh [Isk79]. Rationality is encoded in the Galois action on the geometric Néron-Severi group – varieties with rational points that are ‘minimal’ in the sense of birational geometry need not be rational. In dimension three, recent work [HT19a, HT19c, KP20a, BW20, BW19] has clarified the criteria for rationality: one also needs to take into account principal homogeneous spaces over the intermediate Jacobian, reflecting which curve classes are realized over the ground field. The case of complete intersections of two quadrics was an important first step in understanding the overall structure [HT19b]; rationality in dimension three is equivalent to the existence of a line over $k$ [HT19a, BW19, KP20a]. These developments stimulate investigations in higher dimensions [KP20b]; the examples considered are rational provided there are rational points. In this paper, we focus on the case of four-dimensional complete intersections of two quadrics, especially over the real numbers $\mathbb{R}$. Here we exhibit rational examples without lines and explore further rationality constructions. ###### Theorem 1.1. A smooth complete intersection of two quadrics $X\subset\mathbb{P}^{6}$ over $\mathbb{R}$ is rational if and only if $X(\mathbb{R})$ is nonempty and connected. In general, a projective variety $X$ that is rational over $\mathbb{R}$ has connected nonempty real locus $X(\mathbb{R})$. The point of Theorem 1.1 is that this necessary condition is also sufficient. ###### Corollary 1.2. A smooth complete intersection of two quadrics $X\subset\mathbb{P}^{6}$ is rational over $\mathbb{R}$ if and only if there exists a unirational parametrization $\mathbb{P}^{4}\dashrightarrow X$, defined over $\mathbb{R}$, of odd degree. Indeed, odd degree rational maps are surjective on real points, which guarantees that $X(\mathbb{R})$ is connected. Smooth complete intersections of two quadrics, of dimension at least two, are unirational provided they have a rational point; see Section 3.1 for references and discussion. We also characterize rationality in dimension six, with the exception of one isotopy class that remains open (see Section 6.2). Here is the roadmap of this paper. In Section 2 we recall basic facts about quadrics in even-dimensional projective spaces and their intersections. All interesting cohomology is spanned by the classes of projective subspaces in $X$ of half-dimension, and the Galois group acts on these classes via symmetries of the primitive part of this cohomology, a lattice for the root system $D_{2n+3}$. In Section 3 we present several rationality constructions. The isotopy classification, using Krasnov’s invariants, is presented in Section 4; we draw connections with the Weyl group actions. In Section 5 we focus attention on cases where rationality is not obvious, e.g., due to the presence of a line. In Section 6 we prove Theorem 1.1 and discuss the applicability and limitations of our constructions in dimensions four and six. We speculate on possible extensions to more general fields in Section 7. Acknowledgments: The first author was partially supported by NSF grant 1701659 and Simons Foundation Award 546235, the second author by the NSF grant DMS-1901855, and the third author by NSF grant 2000099. ## 2\. Geometric background ### 2.1. Roots and weights Let $D_{2n+3}$ be the root lattice of the corresponding Dynkin diagram, expressed in the standard Euclidean lattice $\left<L_{1},\ldots,L_{2n+3}\right>,\quad\quad L_{i}\cdot L_{j}=\delta_{ij}$ as the lattice generated by simple roots $\displaystyle R_{1}=L_{1}-L_{2},R_{2}=L_{2}-L_{3},$ $\displaystyle\ldots,R_{2n+1}=L_{2n+1}-L_{2n+2},$ $\displaystyle R_{2n+2}=L_{2n+2}-L_{2n+3},\quad$ $\displaystyle R_{2n+3}=L_{2n+2}+L_{2n+3}.$ Its discriminant group is cyclic of order four, generated by $\tfrac{1}{4}(2(R_{1}+2R_{2}+\cdots+(2n+1)R_{2n+1})+(2n+1)R_{2n+2}+(2n+3)R_{2n+3}).$ Multiplication by $-1$ acts on the discriminant via $\pm 1$. The outer automorphisms of $D_{2n+3}$ also act via automorphisms of $D_{2n+3}$ acting on the discriminant via $\pm 1$, e.g., exchanging $R_{2n+2}$ and $R_{2n+3}$ and keeping the other roots fixed. The Weyl group $W(D_{2n+3})$ acts in the basis $\\{L_{i}\\}$ via signed permutations with an even number of $-1$ entries. The outer automorphisms act via signed permutations with no constraints on the choice of signs, e.g., $L_{i}\mapsto L_{i},\quad i=1,\ldots,2n+2,\quad\quad L_{2n+3}\mapsto- L_{2n+3}.$ The odd and even half-spin representations have weights indexed by subsets $I\subset\\{1,2,\ldots,2n+3\\}$, with $|I|$ odd or even, written $w_{I}=\tfrac{1}{2}(\sum_{i\in I}L_{i}-\sum_{j\in I^{c}}L_{j}).$ The odd and even weights are exchanged by outer automorphisms. ### 2.2. Planes In this section, we assume that the ground field is algebraically closed of characteristic zero. Let $X\subset\mathbb{P}^{2n+2}$ be a smooth complete intersection of two quadrics. We will identify subvarieties in $X$ with their classes in the cohomology of $X$ when no confusion may arise. Let $h$ denote the hyperplane section and consider the primitive cohomology of $X$ under the intersection pairing. Reid [Rei72, 3.14] shows that $(h^{n})^{\perp}\simeq(-1)^{n}D_{2n+3}.$ In other words, the primitive sublattice of $H^{2n}(X,\mathbb{Z}(n))$ – the Tate twist of singular cohomology for the underlying complex variety – may be identified with the root lattice. This is the target of the cycle class map $\operatorname{CH}^{n}(X)\rightarrow H^{2n}(X,\mathbb{Z}(n))$ so the sign convention is natural. ###### Remark 2.1 (Caveat on signs). When $X$ is defined over $\mathbb{R}$, codimension-$n$ subvarieties $Z\subset X$ defined over $\mathbb{R}$ yield classes in $H^{2n}(X,\mathbb{Z}(n))$ that are invariant under complex conjugation. However, the corresponding classes in $H^{2n}(X,\mathbb{Z})$ are multiplied by $(-1)^{n}$. When we mention invariant classes, it is with respect to the former action. Given a plane $P\simeq\mathbb{P}^{n}\subset X$, we have $(P\cdot P)_{X}=c_{n}$ where [Rei72, 3.11] $c_{0}=1,c_{1}=-1,c_{2}=2,c_{3}=-2,\ldots,c_{n}=(-1)^{n}(\lfloor\tfrac{n}{2}\rfloor+1).$ The projection of $P$ into rational primitive cohomology takes the form $P-\tfrac{1}{4}h^{n}$ which has self-intersection $c_{n}-1/4$. The corresponding element $w_{P}\in D_{2n+3}$ has $w_{P}\cdot w_{P}=(2n+3)/4.$ By [Rei72, Cor. 3.9], we obtain bijections $\\{w_{P}\\}_{P\simeq\mathbb{P}^{n}\subset X}=\\{w_{I}\\}_{|I|\text{ has fixed parity }}.$ Note that the residual intersections to $\mathbb{P}^{n}\subset X$ $X\cap\mathbb{P}^{n+2}=\mathbb{P}^{n}\cup S$ give cubic scrolls $S\subset X$; these realize the weights of opposite parity. By [Rei72, Th. 3.8], two planes $P_{1}$ and $P_{2}$ are disjoint if and only if $w_{P_{1}}\cdot w_{P_{2}}=(-1)^{n+1}/4.$ For example, if $n=1$ and $w_{P_{1}}$ is identified with $(L_{1}-L_{2}-L_{3}-L_{4}-L_{5})/2$ then the relevant weights are $(L_{1}+L_{2}+L_{3}-L_{4}-L_{5})/2,\quad\ldots,\quad(L_{1}-L_{2}-L_{3}+L_{4}+L_{5})/2$ and $(-L_{1}+L_{2}+L_{3}+L_{4}-L_{5})/2,\quad\ldots,\quad(-L_{1}-L_{2}+L_{3}+L_{4}+L_{5})/2,$ a total of $10=\binom{5}{2}$ such lines. When $n=2$ and $w_{P_{1}}$ is identified with $(L_{1}-L_{2}-\cdots-L_{7})/2$ then the relevant weights are $(L_{1}+L_{2}+L_{3}+L_{4}+L_{5}-L_{6}-L_{7})/2,\ldots,(L_{1}-L_{2}-L_{3}+L_{4}+L_{5}+L_{6}+L_{7})/2$ and $(-L_{1}+L_{2}+L_{3}+L_{4}-L_{5}-L_{6}-L_{7})/2,\ldots,(-L_{1}-L_{2}-L_{3}-L_{4}+L_{5}+L_{6}+L_{7})/2,$ a total of $\binom{6}{4}+\binom{6}{3}=35=\binom{7}{3}$ planes. The planes $P_{1}$ and $P_{2}$ meet at a point if and only if $w_{P_{1}}\cdot w_{P_{2}}=(-1)^{n}\tfrac{3}{4}.$ If they meet along an $r$-plane then an excess intersection computation gives [Rei72, 3.10] $P_{1}\cdot P_{2}=(-1)^{r}(\lfloor\tfrac{r}{2}\rfloor)+1)$ and $w_{P_{1}}\cdot w_{P_{2}}=(-1)^{r+n}(\lfloor\tfrac{r}{2}\rfloor+1)-(-1)^{n}\tfrac{1}{4}.$ In particular, they meet along an $(n-1)$-plane when $w_{P_{1}}\cdot w_{P_{2}}=-(\lfloor\tfrac{n-1}{2}\rfloor+1)-(-1)^{n}\tfrac{1}{4};$ for a fixed $w_{P_{1}}$ there are $2n+3$ planes $P_{2}$ meeting $P_{1}$ in this way. For example, if $n=1$ and $w_{P_{1}}=(L_{1}-L_{2}-L_{3}-L_{4}-L_{5})/2$ then the possibilities for $w_{P_{2}}$ are $\displaystyle(L_{1}+L_{2}+L_{3}+L_{4}+L_{5})/2,\quad(-L_{1}-L_{2}+L_{3}+L_{4}+L_{5})/2,$ $\displaystyle(-L_{1}+L_{2}-L_{3}+L_{4}+L_{5})/2,\quad(-L_{1}+L_{2}+L_{3}-L_{4}+L_{5})/2,$ $\displaystyle(-L_{1}+L_{2}+L_{3}+L_{4}-L_{5})/2.$ ### 2.3. Quadrics We retain the notation of Section 2.2. Our next task is to analyze quadric $n$-folds $Q\subset X$, i.e., $Q$ a degree-two hypersurface in $\mathbb{P}^{n+1}$. Let $\\{\mathcal{Q}_{t}\\},t\in\mathbb{P}^{1}$ denote the pencil of quadric hypersurfaces cutting out $X$. The degeneracy locus $D:=\\{t\in\mathbb{P}^{1}:\mathcal{Q}_{t}\text{ singular }\\}$ consists of $2n+3$ points; since $X$ is smooth, each has rank $2n+2$. The Hilbert scheme of quadric $n$-folds $Q\subset X$ is isomorphic to the relative Fano variety of $(n+1)$-planes $\mathcal{F}(\mathcal{Q}/\mathbb{P}^{1})=\\{\Pi\simeq\mathbb{P}^{n+1}\subset\mathcal{Q}_{t}\text{ for some }t\in\mathbb{P}^{1}\\},$ which consists of $2(2n+3)$ copies of the connected isotropic Grassmannian $\operatorname{OGr}(n+1,2n+2)$. Given a quadric $Q$, its projection to rational primitive cohomology $Q-\tfrac{1}{2}h^{n}$ corresponds to an element $w_{Q}\in D_{2n+3},\quad w_{Q}\in\\{\pm L_{1},\ldots,\pm L_{2n+3}\\}.$ In particular, we have $Q\cdot Q=\begin{cases}2&\text{ if $n$ even}\\\ 0&\text{ if $n$ odd.}\end{cases}$ Residuation in a complete intersection of linear forms $Q\cup Q^{\prime}=X\cap h^{n}$ reverses signs, i.e., $w_{Q}=-w_{Q^{\prime}}$. On the other hand, if $Q_{1}$ and $Q_{2}$ are not residual then (2.1) $Q_{1}\cdot Q_{2}=1.$ We summarize this as follows: ###### Proposition 2.2. The signed permutation representation of $W(D_{2n+3})$ is realized via the action on classes $[Q]$, where $Q\subset X$ is a quadric $n$-fold. Note that there are $2^{2n+1}(2n+3)$ reducible quadrics – unions of two $n$-planes meeting in an $(n-1)$-plane – with $2^{2n}$ reducible quadrics in each copy of the isotropic Grassmannian. ## 3\. Rationality constructions We now work over an arbitrary field $k$ of characteristic zero. ### 3.1. General considerations Let $X\subset\mathbb{P}^{d+2}$ denote a smooth complete intersection of two quadrics of dimension at least two. Recall the following: * • If $X(k)\neq\emptyset$ then $X$ is unirational over $k$ and has Zariski dense rational points [CTSSD87, Rem. 3.28.3]. * • If there is a line $\ell\subset X$ defined over $k$ then projection induces a birational map $\pi_{\ell}:X\stackrel{{\scriptstyle\sim}}{{\dashrightarrow}}\mathbb{P}^{d}$. For reference, we recall Amer’s theorem [Lee07, Th. 2.2]: ###### Theorem 3.1. Let $k$ be a field of characteristic not two, $F_{1}$ and $F_{2}$ quadrics over $k$, and $\mathcal{Q}_{t}=\\{F_{1}+tF_{2}\\}$ the associated pencil of quadrics over $k(t)$. Then $X=\\{F_{1}=F_{2}=0\\}$ has an $r$-dimensional isotropic subspace over $k$ if and only if $\mathcal{Q}_{t}$ has an $r$-dimensional isotropic subspace over $k(t)$. We apply this for $k=\mathbb{R}$, where $X\subset\mathbb{P}^{d+2}$ is a smooth complete intersection of two quadrics and $\mathcal{Q}\rightarrow\mathbb{P}^{1}$ is the associated pencil. Recall Springer’s Theorem: A quadric hypersurface $\mathcal{Q}$ over a field $L$ has a rational point if it admits a rational point over some odd-degree extension of $L$. Applying this to the pencil $\mathcal{Q}\rightarrow\mathbb{P}^{1}$ associated with a complete intersection of two quadrics, with Amer’s Theorem, yields: ###### Proposition 3.2. If $d\geq 1$ and $X$ contains a subvariety of odd degree over $k$ then $X(k)\neq\emptyset$. We can prove a bit more: ###### Proposition 3.3. If $d\geq 2$ and $X$ has a curve of odd degree defined over $k$ then $X$ is rational over $k$. ###### Proof. Recall that double projection from a sufficiently general rational point $x\in X(k)$ yields a diagram $X\stackrel{{\scriptstyle\sim}}{{\dashrightarrow}}Y\rightarrow\mathbb{P}^{1}$ where $Y$ is a quadric bundle of relative dimension $d-1$. A curve $C\subset X$ of odd degree yields an multisection of this bundle of odd degree. Thus $Y\rightarrow\mathbb{P}^{1}$ has a section by Springer’s Theorem and its generic fiber $Y_{t}$ is rational over $k(\mathbb{P}^{1})$. It follows that $X$ is rational over the ground field. ∎ ###### Remark 3.4. The pencil defining $X$ gives a quadric bundle $\mathcal{Q}\rightarrow\mathbb{P}^{1}$ of relative dimension $d+1$. We apply Witt’s decomposition theorem to $[\mathcal{Q}_{t}]$ and $[Y_{t}]$, understood as quadratic forms over $k(\mathbb{P}^{1})=k(t)$, to obtain $[\mathcal{Q}_{t}]=[Y_{t}]\oplus\left(\begin{matrix}0&1\\\ 1&0\end{matrix}\right).$ Thus a section of $Y\rightarrow\mathbb{P}^{1}$ yields an isotropic line of $\mathcal{Q}\rightarrow\mathbb{P}^{1}$, and Theorem 3.1 implies that $X$ contains a line defined over $k$. ###### Corollary 3.5. (see appendix by Colliot-Thélène [HT19b, Th. A5]) Let $X\subset\mathbb{P}^{d+2}$ denote a smooth complete intersection of two quadrics of dimension at least two. Suppose there exists an irreducible positive-dimensional subvariety $W\subset X$ of odd degree, defined over $k$. Then $X$ is rational over $k$. Given these results, we focus on proving rationality in cases where $X$ does not contain lines or positive-dimensional subvarieties of odd degree. ### 3.2. Rationality using half-dimensional subvarieties We now turn to even-dimensional intersections of two quadrics $X\subset\mathbb{P}^{2n+2},\quad n\geq 1.$ Throughout, we assume that $X(k)\neq\emptyset,$ and thus $X$ is $k$-unirational and $k$-rational points are Zariski dense. Construction I: Suppose that * • $X$ has a pair of conjugate disjoint $n$-planes $P,\bar{P}$, defined over a quadratic extension $K$ of $k$. Projecting from a general $x\in X(k)$ gives a birational map $X\stackrel{{\scriptstyle\sim}}{{\dashrightarrow}}X^{\prime}\subset\mathbb{P}^{2n+1},$ where $X^{\prime}$ is a cubic hypersurface. Since $X(k)\subset X$ is Zariski dense, we may assume that the images of $P$ and $\bar{P}$ in $X^{\prime}$ remain disjoint. The ‘third point’ construction gives a birational map $\mathbf{R}_{K/k}(P)\stackrel{{\scriptstyle\sim}}{{\dashrightarrow}}X^{\prime},$ where the source variety is the restriction of scalars. We conclude that $X$ is rational over $k$. This construction appears in [CTSSD87, Th. 2.4]. Construction II: Suppose that * • $X$ has a pair of conjugate disjoint quadric $n$-folds $Q,\bar{Q}$, defined over a quadratic extension $K$ of $k$, and meeting transversally at one point. Projecting from $x\in Q\cap\bar{Q}$, which is a $k$-rational point $X$, gives a birational map $X\stackrel{{\scriptstyle\sim}}{{\dashrightarrow}}X^{\prime}\subset\mathbb{P}^{2n+1},$ where $X^{\prime}$ is a cubic hypersurface. The proper transforms $Q^{\prime},\bar{Q}^{\prime}\subset X^{\prime}$ are disjoint unless there exists a line $x\in\ell\subset\mathbb{P}^{2n+2}$ with $\\{x\\}\subsetneq\ell\cap Q,\ell\cap\bar{Q}$ as schemes. We may assume that $\ell\not\subset X$ as we already know $X$ is rational in this case. Thus the only possibility is $\ell\cap Q=\ell\cap\bar{Q}$ as length-two subschemes, which is precluded by the intersection assumption. Repeating the argument for Construction I thus gives rationality. Construction III: Suppose that * • $X$ contains a quadric $Q$ of dimension $n$, defined over $k$. Projection gives a fibration $q:\operatorname{Bl}_{Q}(X)\rightarrow\mathbb{P}^{n}$ with fibers quadrics of dimension $n$. Now suppose that $X$ contains a second $n$-fold $T$ with the property that $\deg(T)-T\cdot Q$ is odd, i.e., a multisection for $q$ of odd degree. It follows that the generic fiber of $q$ is rational and thus $Y$ is rational over $k$. When $\dim(X)=4$ a number of $T$ might work, e.g., * • a plane disjoint from $Q$, * • a second quadric meeting $Q$ in one point, * • a quartic or a sextic del Pezzo surface meeting $Q$ in one or three points, * • a degree 8 K3 surface meeting $Q$ in one, three, five, or seven points. Construction IV: Suppose that * • $\dim(X)=4$ and $X$ contains a quartic scroll $T$, defined over $k$. Geometrically, $T$ is the image of the ruled surface ${\mathbb{F}}_{0}=\mathbb{P}^{1}\times\mathbb{P}^{1}\hookrightarrow\mathbb{P}^{5}$ under the linear series of bidegree $(1,2)$. Over $\mathbb{R}$ we are interested in cases where $T=\mathbb{P}^{1}\times C$ with $C$ a nonsplit conic. We do not want to force $X$ to have lines! (Note that quartic scrolls geometrically isomorphic to ${\mathbb{F}}_{2}$ contain lines defined over the ground field and thus are not useful for our purposes.) On projecting from a point $x\in X(k)$ we get a cubic fourfold $X^{\prime}\subset\mathbb{P}^{5},$ containing a quartic scroll. The Beauville-Donagi construction [BD85] – concretely, take the image under the linear system of quadrics vanishing along $T$ – shows that $X^{\prime}$ is birational to a quadric hypersurface thus rational over $k$. Recall that an $n$-dimensional smooth variety $W\subset\mathbb{P}^{2n+1}$ is said to have one apparent double point if a generic point is contained in a unique secant to $W$. Construction V: Suppose that $X$ contains a variety $W$ defined over $k$ of dimension $n\geq 2$ such that: * • $W$ spans a $\mathbb{P}^{2n+1}$ and has one apparent double point; or * • $W$ has a rational point $w$ such that projecting from $w$ maps $W$ birationally to a variety with one apparent double point. Then $X$ is rational over $k$. As before, one projects from a rational point to a get cubic hypersurface $X^{\prime}\subset\mathbb{P}^{2n+1}$. Cubic hypersurfaces containing varieties $W$ with one apparent double point are rational [Rus00, Prop. 9]. Indeed, intersecting secant lines of $W$ with $X^{\prime}$ yields $\operatorname{Sym}^{2}(W)\dashrightarrow X^{\prime},$ which is birational when each point lies on a unique secant to $W$. Quartic scrolls in $\mathbb{P}^{5}$ have one apparent double point so Construction IV is a special case of Construction V. ## 4\. Isotopy classification We review the classification of smooth complete intersections of two quadrics $X\subset\mathbb{P}^{2n+2}$ over $\mathbb{R}$, following [Kra18]. Express $X=\\{F_{1}=F_{2}=0\\}$ where $F_{1}$ and $F_{2}$ are real quadratic forms. We continue to use $D$ for the degeneracy locus of the associated pencil $\mathcal{Q}_{t}=\\{t_{1}F_{1}+t_{2}F_{2}=0\\}$. Let $r=|D(\mathbb{R})|$ which is odd with $r\leq 2n+3$. Consider the signatures $(I^{+},I^{-})$ of the forms $s_{1}F_{1}+s_{2}F_{2},\quad(s_{1},s_{2})\in\mathbb{S}^{1}=\\{(s_{1},s_{2})\in\mathbb{R}^{2}:s_{1}^{2}+s_{2}^{2}=1\\}.$ Record these at the $2r$ points lying over $D$, in order as we trace the circle counterclockwise. We label each of these points with $\pm$ depending on whether the positive part $I^{+}$ of the signature increases or decreases as we cross the point. Each point of $D(\mathbb{R})$ yields a pair of antipodal points on $\mathbb{S}^{1}$ labelled with opposite signs. For example, for $n=0$ and $r=3$ admissible sequences of signatures and $\pm$’s include $(0,2)(1,1)(2,0)(2,0)(1,1)(0,2)\quad(+,+,+,-,-,-).$ and $(1,1)(1,1)(1,1)(1,1)(1,1)(1,1)\quad(+,-,+,-,+,-).$ Suppose the sequence of $\pm$’s has maximal unbroken chains of $+$’s of lengths $r_{1},r_{2},\ldots,r_{2s+1}$ where $r=r_{1}+r_{2}+\cdots+r_{2s+1}.$ The number of terms is odd because antipodal points have opposite signs. In the examples above we have $3=3$ and $3=1+1+1$. Our invariant is the sequence $(r_{1},\ldots,r_{2s+1})$ up to cyclic permutations and reversals – a complete isotopy invariant of $X$ over $\mathbb{R}$ [Kra18]. We derive a sequence of $\pm 1$’s of length $k$ from this invariant as follows: For each point of $D(\mathbb{R})$, record the sign of the discriminant of the associated rank-$(2n+2)$ quadratic form, determined by the parity of $(I^{+}-I^{-})/2$. In the examples above, we obtain $(-1,+1,-1)$ and $(+1,+1,+1)$. The number of $-1$’s is always even. The analysis in Section 2.3 shows that complex conjugation acts on $H^{2n}(X,\mathbb{Z}(n))$ in the basis $\\{L_{1},\ldots,L_{2n+3}\\}$ as a signed permutation of order two. This is a direct sum of blocks $(+1),(-1),\pm\left(\begin{matrix}0&1\\\ 1&0\end{matrix}\right).$ Actually, we may assume the sign is positive in the third case after conjugating by $\left(\begin{matrix}0&-1\\\ 1&0\end{matrix}\right).$ Suppose there are $a$ blocks $(+1)$, $2b$ blocks $(-1)$, and $c$ blocks of the third kind, with $a+2b+2c=2n+3$. These correspond to the conjugacy classes of involutions $\iota\in W(D_{2n+3})$ [Kot00, §3.2,3.3]. We have $r=a+2b$, reflecting the number of points of $D(\mathbb{R})$ with positive and negative discriminants respectively, and $2c=2n+3-r$, reflecting the number of complex- conjugate pairs in $D(\mathbb{C})\setminus D(\mathbb{R})$. The passage from isotopy classes to conjugacy classes in $W(D_{2n+3})$ results in a loss of information. We give an example for real quartic del Pezzo surfaces $X\subset\mathbb{P}^{4}$. ###### Example 4.1. The isotopy class $(5)$ has singular members with signatures $(0,4)(1,3)(2,2)(3,1)(4,0)(4,0)(3,1)(2,2)(1,3)(0,4)$ with involution given the diagonal $5\times 5$ matrix $\operatorname{diag}(1,-1,{\bf{1}},-1,1),$ where the bolded $\bf{1}$ corresponds to a degenerate fiber $\mathcal{Q}_{t}$ whose rulings sweep out quadric curves (conics) on $X$ defined over $\mathbb{R}$. (There is only one pair of such conics.) Here $X(\mathbb{R})=\emptyset$ as it is contained in an anisotropic quadric threefold. The isotopy class $(2,2,1)$ has singular members with signatures $(1,3)(2,2)(2,2)(2,2)(3,1)(3,1)(2,2)(2,2)(2,2)(1,3)$ with involution $\operatorname{diag}(-1,{\bf{1}},{\bf{1}},{\bf{1}},-1).$ This has the same Galois action but contains three pairs of conics defined over $\mathbb{R}$, represented by the bolded $\bf{1}$’s. These are distinguished cohomologically by the arrow $\mathbb{Z}^{3}\simeq H^{0}(G,\operatorname{Pic}(X_{\mathbb{C}}))\rightarrow H^{2}(G,\Gamma(\mathcal{O}^{*}_{X_{\mathbb{C}}}))=\operatorname{Br}(\mathbb{R})\simeq\mathbb{Z}/2\mathbb{Z}$ of the Hochschild-Serre spectral sequence. ###### Proposition 4.2. Fix a conjugacy class $[\iota=\iota_{abc}]$ of involutions in $W(D_{2n+3})$. Consider the isotopy classes of $X(\mathbb{R})\subset\mathbb{P}^{2n+2}$ such that complex conjugation acts by $\iota$. The possible isotopy classes correspond to shuffles of $(\underbrace{1,\ldots,1}_{a\text{ times}},\underbrace{-1,\ldots,-1}_{2b\text{ times}})$ up to cyclic permutations and reversals. ###### Proof. Observe first that points in $D(\mathbb{C})\setminus D(\mathbb{R})$ are irrelevant to the Krasnov invariant, assuming the dimension is given. So it makes sense they are not relevant in the enumeration. We have already seen that sequences of the prescribed form arise from each isotopy class; we present the reverse construction. Choose such a sequence, e.g. $(1,1,1,-1,1,-1,-1,1,-1).$ The key observation is that local maxima and minima of $I^{+}-I^{-}$ – which necessarily occur at smooth points – arise precisely between points of the degeneracy locus where signs do not change. We indicate smooth fibers achieving maxima/minima with $\mid$, e.g., $1\mid 1\mid 1,-1,1,-1\mid-1,1,-1$ or equivalently $\mid 1\mid 1,-1,1,-1\mid-1,1,-1,1\mid$ after cyclic permutation. Lifting to the double cover $\mathbb{S}^{1}$ entails concatenating two such expressions: $\mid 1\mid 1,-1,1,-1\mid-1,1,-1,1\mid 1\mid 1,-1,1,-1\mid-1,1,-1,1\mid$ From this, we read off the points of $D(\mathbb{R})$ on which $I^{+}$ increases and decreases $+----++++-++++----$ which determines the Krasnov invariant – $(1,4,4)$ in this example. ∎ ## 5\. Applying quadratic forms ### 5.1. Quadric fibrations over real curves Let $C$ be a smooth projective geometrically connected curve over $\mathbb{R}$ with function field $K=\mathbb{R}(C)$. Let $Q\subset\mathbb{P}^{d+1}$ be a smooth (rank $d+2$) quadric hypersurface over $K$ and $F_{i}(Q)$ the $i$-dimensional isotropic subspaces, so that $F_{0}(Q)=Q$ and $F_{m}(Q)$ is empty when $2m>d$. If $d=2m$ then $F_{m}(Q)$ has two geometrically connected components; otherwise it is connected. Suppose that $\pi:\mathcal{Q}\rightarrow C$ is a regular projective model of $Q$, such that the fibers are all quadric hypersurfaces of rank at least $d+1$. The locus $D\subset C$ corresponding to fibers of rank $d+1$ is called the degeneracy locus. Fundamental results of Witt – see [CT96, §2] and [Sch96] for a modern formulation in terms of local-global principles – assert that: * • if $d>0$ then $Q(K)\neq\emptyset$ if $\mathcal{Q}_{c}=\pi^{-1}(c)$ has a smooth real point for each $c\in C(\mathbb{R})$; * • if $d>2$ then $F_{1}(Q)(K)\neq\emptyset$ if $F_{1}(\mathcal{Q}_{c})$ has a smooth real point for each $c\in C(\mathbb{R})$. We can translate these into conditions on the signatures of the smooth fibers * • if $d>0$ then $Q(K)\neq\emptyset$ if $\mathcal{Q}_{c}$ is not definite for any $c\in(C\setminus D)(\mathbb{R})$; * • if $d>2$ then $F_{1}(Q)(K)\neq\emptyset$ if $\mathcal{Q}_{c}$ does not have signatures $(d+2,0),(d+1,1),(1,d+1)$ or $(0,d+2)$ for any $c\in(C\setminus D)(\mathbb{R})$. In other words, we have points and lines over $K$ if the fibers permit them. This reflects a general principle: Suppose $\mathcal{X}$ is regular and has a flat proper morphism $\varpi:\mathcal{X}\rightarrow C$ to a curve $C$, all defined over $\mathbb{R}$. The local-global and reciprocity obstructions to sections of $\varpi$ are reflected in the absence of continuous sections $C(\mathbb{R})\rightarrow\mathcal{X}(\mathbb{R})$ for induced map of the underlying real manifolds [Duc98]. ### 5.2. Implications of Amer’s Theorem Let $X\subset\mathbb{P}^{d+2}$ be a smooth complete intersection of two quadrics over $\mathbb{R}$ and $\mathcal{Q}\rightarrow\mathbb{P}^{1}$ the associated pencil of quadrics. The results of Section 5.1 imply that $\mathcal{Q}\rightarrow\mathbb{P}^{1}$ has a section unless the Krasnov invariant is $(d+3)$; the variety of lines $F_{1}(\mathcal{Q}/\mathbb{P}^{1})\rightarrow\mathbb{P}^{1}$ has a section unless the Krasnov invariant is $(d+3),(d+2-e,e,1)\text{ with }1\leq e\leq\tfrac{d+2}{2},\quad(d+1).$ Thus the Krasnov invariant determines which dimensional linear subspaces and quadrics appear on $X$: ###### Proposition 5.1. Let $X\subset\mathbb{P}^{d+2}$ be a smooth complete intersection of two quadrics defined over $\mathbb{R}$. The only isotopy classes of $X$ that fail to contain a line are: * • $(d+3)$ where $X(\mathbb{R})=\emptyset$; * • $(d+2-e,e,1)$ with $1\leq e\leq\tfrac{d+2}{2}$; * • $(d+1)$. The case $(d+1,1,1)$ has disconnected real locus $X(\mathbb{R})$ [Kra18, p. 117]; thus $X$ cannot be rational over $\mathbb{R}$. The cases $(d+2-e,e,1)$ with $2\leq e\leq\tfrac{d+2}{2}$ are connected. ### 5.3. Quadric $n$-folds over $\mathbb{R}$ Assume now that $X$ has even dimension $d=2n$. We may read off from the invariant $(r_{1},\ldots,r_{2s+1})$ which classes of quadric $n$-folds $Q\subset X_{\mathbb{C}}$ are realized by quadrics defined over $\mathbb{R}$. Fix a smooth real quadric hypersurface ${\bf Q}=\\{F=0\\}\subset\mathbb{P}^{2n+1}.$ The following conditions are equivalent: * • the geometric components of the variety of maximal isotropic subspaces $\operatorname{OGr}(\bf{Q})$ are defined over $\mathbb{R}$; * • the discriminant of $F$ is positive; * • the $I^{+}(F)-I^{-}(F)$ is divisible by four. This means that complex conjugation fixes the class of a maximal isotropic subspace. We also have equivalence among: * • there is a maximal isotropic subspace $\mathbb{P}^{n}\subset\bf{Q}$ defined over $\mathbb{R}$; * • the signature of $F$ is $(n+1,n+1)$. Thus quadric $n$-folds $Q\subset X\subset\mathbb{P}^{2n+2}$ defined over $\mathbb{R}$ correspond to rulings of degenerate fibers $\mathcal{Q}_{t},t\in D(\mathbb{R})$ where $\mathcal{Q}_{t}$ has signature $(n+1,n+1)$. As in Example 4.1, the corresponding $(+1)$-blocks in the complex conjugation involution $\iota\in W(D_{2n+3})$ will be designated $\bf{1}$, in boldface. ### 5.4. Analysis of the remaining even-dimensional isotopy classes We continue to assume that $X$ has even dimension $d=2n$, focusing on the isotopy classes without lines. The cases $(2n+2-e,e,1)=(e,1,2n+2-e),\quad 2\leq e\leq n+1$ have degeneracy consisting of $2n+3$ real points. The signatures of nonsingular members are $\displaystyle(1,2n+2)\ldots(e+1,2n+2-e)(e,2n+3-e)(e+1,2n+2-e)\ldots$ $\displaystyle(2n+1,2)(2n+2,1)\ldots(2n+2-e,e+1)(2n+3-e,e)$ $\displaystyle(2n+2-e,e+1)\ldots(2,2n+1)$ For $(n+1,n+1,1)$ the resulting signed permutation matrix is the diagonal matrix (5.1) $\operatorname{diag}(\underbrace{(-1)^{n},\ldots,-1,{\mathbf{1}}}_{n+1\text{ terms }},{\mathbf{1}},\underbrace{{\mathbf{1}},-1,\ldots,(-1)^{n}}_{n+1\text{ terms }}),$ with the bolded $\bf{1}$’s corresponding to singular fibers with signature $(n+1,n+1)$. The number of $+1$’s $a=\begin{cases}n+2&\text{if $n$ odd }\\\ n+3&\text{if $n$ even.}\end{cases}$ For $e\neq n+1$ we have $\operatorname{diag}(\underbrace{(-1)^{n},\ldots,(-1)^{n+e-1}}_{e\text{ terms}},(-1)^{n+e-1},\underbrace{(-1)^{n+(2n+2-e)-1},\ldots,(-1)^{n}}_{2n+2-e\text{ terms}}).$ Note that $(-1)^{n+e-1}=(-1)^{n+(2n+2-e)-1}$ so the three middle terms are equal. There is exactly one $\bf{1}$ corresponding to the singular fiber with signature $(n+1,n+1)$. The number of $+1$’s is given (5.2) $a=\begin{cases}n&\text{if $n,e$ odd }\\\ n+2&\text{if $n$ odd and $e$ even}\\\ n+3&\text{if $n$ even and $e$ odd}\\\ n+1&\text{if $n,e$ even.}\end{cases}$ For case $(2n+1)$ the signatures of nonsingular members are $\displaystyle(2,2n+1)\ldots(2n,3)(2n+1,2)(2n+2,3)\ldots(2,2n+1)$ The signed permutation matrix has one factor $\left(\begin{matrix}0&1\\\ 1&0\end{matrix}\right)$ and diagonal entries $((-1)^{n-1},\ldots,-1,1,-1,\ldots,(-1)^{n-1}).$ The number of positive terms is $a=\begin{cases}n&\text{if $n$ odd}\\\ n+1&\text{if $n$ even.}\end{cases}$ ## 6\. Application of the constructions ### 6.1. Proof of Theorem 1.1 Rationality is evident for isotopy classes of varieties that contain a line defined over $\mathbb{R}$. Propostion 5.1 enumerates the remaining cases $(5),(1,3,3),(1,2,4).$ These are covered by the following propositions. ###### Proposition 6.1. Let $X\subset\mathbb{P}^{2n+2}$ be a smooth complete intersection of two quadrics over $\mathbb{R}$ with invariant $(2n+1)$. Then $X$ is rational. ###### Proof. The analysis in Section 5.4 indicates that complex conjugation exchanges two classes of quadric $n$-folds associated to complex conjugate points in $D(\mathbb{C})\setminus D(\mathbb{R})$. Denote these by $[Q]$ and $[\bar{Q}]$ – recall from (2.1) that $[Q]\cdot[\bar{Q}]=1.$ Choosing a suitably general complex representation $Q\subset X_{\mathbb{C}}$, the intersection $Q\cap\bar{Q}$ is proper. Then $Q\cap\bar{Q}$ consists of a single rational point of $X$ with multiplicity one. In particular, the hypotheses of Construction II are satisfied. ∎ ###### Proposition 6.2. Let $X\subset\mathbb{P}^{2n+2}$ be a smooth complete intersection of two quadrics over $\mathbb{R}$ with invariant $(2n+2-e,e,1)=(e,1,2n+2-e),\quad 2\leq e\leq n+1.$ Assume that either $e$ is even or $e=n+1$. Then $X$ is rational. ###### Proof. Assume first that $e=n+1$. It follows from (5.1) in Section 5.4 that $X$ admits three classes $\begin{array}[]{c|cccc}&h^{n}&Q_{1}&Q_{2}&Q_{3}\\\ \hline\cr h^{n}&4&2&2&2\\\ Q_{1}&2&1+(-1)^{n}&1&1\\\ Q_{2}&2&1&1+(-1)^{n}&1\\\ Q_{3}&2&1&1&1+(-1)^{n}\end{array}$ with each $Q_{i}$ realized by a quadric $n$-fold defined over $\mathbb{R}$. Construction III gives rationality in this case. Now assume that $e$ is even. The formula (5.2) shows that the numbers of $+1$’s and $-1$’s appearing in the $\iota\in W(D_{2n+3})$ associated with complex conjugation are as close as possible. If $n$ is even then we have $n+1$ of the former and $n+2$ of the latter; when $n$ is odd we have $n+2$ of the former and $n+1$ of the letter. Given an $n$-plane $P\subset X_{\mathbb{C}}$, the formulas in Section 2.2 yield $w_{P}\cdot w_{\bar{P}}=(-1)^{n+1}/4$ whence $P$ and $\bar{P}$ are disjoint in $X_{\mathbb{C}}$. Thus we may apply Construction I to conclude rationality. ∎ ###### Remark 6.3. Remark 3.4 implies that $X$ does not admit curves (or surfaces!) of odd degree defined over $\mathbb{R}$. These would force the existence of lines defined over $\mathbb{R}$, which do not exist in this isotopy class. ### 6.2. Remaining six-dimensional case To settle the rationality of six-dimensional complete intersections of two quadrics $X\subset\mathbb{P}^{6}$, there is one remaining case in the Krasnov classification: $(1,3,5)$. The sequence of signatures of nonsingular elements of $\\{\mathcal{Q}_{t}\\}$ is: $\displaystyle(1,8)(2,7)(3,6)(4,5)(5,4)(6,3)(5,4)(6,3)(7,2)$ $\displaystyle(8,1)(7,2)(6,3)(5,4)(4,5)(3,6)(4,5)(3,6)(2,7).$ The signed permutation is the diagonal matrix $\operatorname{diag}(-1,1,-1,{\bf{1}},-1,-1,-1,1,-1)$ and the invariant cycles are: $\begin{array}[]{c|cccc}&h^{3}&Q_{1}&Q_{2}&Q_{3}\\\ \hline\cr h^{3}&4&2&2&2\\\ Q_{1}&2&0&1&1\\\ Q_{2}&2&1&0&1\\\ Q_{3}&2&1&1&0\end{array}$ Here $Q_{1}$ corresponds to the singular fiber of signature $(4,4)$ and $Q_{2}$ and $Q_{3}$ correspond to the singular fibers of signatures $(2,6)$ and $(6,2)$. If $P\subset X$ is a three-plane then $w_{P}\cdot w_{\bar{P}}=-3/4$ and $P$ and $\bar{P}$ meet in a single point. ## 7\. Extensions and more general fields We work over a field $k$ of characteristic zero. In this section, we give further examples of rationality constructions for $2n$-dimensional intersections of two quadrics over $k$, relying on special subvarieties of dimension $n$. ### 7.1. Dimension four: intersection computations Given $X\subset\mathbb{P}^{6}$, a smooth complete intersection of two quadrics, we have $\displaystyle c_{t}({\mathcal{T}}_{X})$ $\displaystyle\equiv(1+7ht+21h^{2}t^{2})/(1+2ht)^{2}\pmod{t^{3}}$ $\displaystyle\equiv(1+7ht+21h^{2}t^{2})(1-2ht+4h^{2}t^{2})^{2}\pmod{t^{3}}$ $\displaystyle\equiv 1+3ht+5h^{2}t^{2}\pmod{t^{3}}$ If $T\subset X$ is a smooth projective geometrically connected surface then $\displaystyle c_{t}({\mathcal{N}}_{T/X})$ $\displaystyle=(1+3ht+5h^{2}t^{2})/(1-K_{T}t+\chi(T)t^{2})$ $\displaystyle=1+(3h+K_{T})t+(5h^{2}+3hK_{T}+K_{T}^{2}-\chi(T))t^{2},$ where $\chi(T)$ is the topological Euler characteristic. The expected dimension of the deformation space of $T$ in $X$ is $\displaystyle\chi({\mathcal{N}}_{T/X})$ $\displaystyle=2-\tfrac{1}{2}(3h+K_{T})K_{T}+\frac{1}{2}(3h+K_{T})^{2}$ $\displaystyle\hskip 142.26378pt-(5h^{2}+3hK_{T}+K_{T}^{2}-\chi(T))$ $\displaystyle=2\chi(\mathcal{O}_{T})+\tfrac{1}{2}(-h^{2}-3hK_{T})-K_{T}^{2}+\chi(T)).$ For example, * • if $T=\mathbb{P}^{2}$ is embedded as a plane then $(T\cdot T)_{X}=2$ and $T$ is rigid; * • if $T$ is a quadric then $(T\cdot T)_{X}=2$ and $T$ moves in a three-parameter family; * • if $T$ is a quartic scroll then $(T\cdot T)_{X}=6$ and moves in a five- parameter family; * • if $T$ is a sextic del Pezzo surface then $(T\cdot T)_{X}=12$ and $T$ moves in an eight-parameter family. ### 7.2. Dimension four: surfaces with one apparent double point Recall that Construction V gives the rationality of fourfolds admitting a surface with one apparent double point. A classical result asserted by Severi – see [CMR04, Th. 4.10] for a modern proof – characterizes smooth surfaces $T\subset\mathbb{P}^{5}$ with one apparent double point, i.e., surfaces that acquire one singularity on generic projection into $\mathbb{P}^{4}$: * • $\deg(T)=4$: $T$ is a quartic scroll $T\simeq\mathbb{P}(\mathcal{O}_{\mathbb{P}^{1}}(2)^{2}),\quad\mathbb{P}(\mathcal{O}_{\mathbb{P}^{1}}(1)\oplus\mathcal{O}_{\mathbb{P}^{1}}(3));$ * • $\deg(T)=5$: $T$ is a quintic del Pezzo surface. Thus Construction V says that a smooth complete intersection of two quadrics $X\subset\mathbb{P}^{6}$ is rational if it contains a quartic scroll, a quintic del Pezzo surface, or a sextic del Pezzo surface with a rational point. Rationality always holds when there are positive-dimensional subvarieties of odd degree (see Section 3.1), so we focus attention to the first case. We seek criteria for the existence of a quartic scroll $T\subset X$, defined over $k$. Clearly, the class $[T]$ must be Galois-invariant; however, Galois- invariant classes need not be represented by cycles over $k$. The intersection computations above imply that $[T]=[Q_{2}]+[Q_{3}],$ where $Q_{2}$ and $Q_{3}$ represent quadric surfaces in $X$, defined over the algebraic closure. Assume that the class $[Q_{2}]+[Q_{3}]$ is Galois invariant and represents algebraic cycles defined over the ground field. We look for quartic scrolls $T\subset X$ with class $[T]=[Q_{2}]+[Q_{3}]$. ###### Remark 7.1. Over $k=\mathbb{R}$, case $(1,2,4)$ has signed permutation $\operatorname{diag}(1,-1,-1,-1,{\bf{1}},-1,1).$ The intersection form on the invariant part of $H^{4}(X_{\mathbb{C}},\mathbb{Z})$ takes the form $\begin{array}[]{c|cccc}&h^{2}&Q_{1}&Q_{2}&Q_{3}\\\ \hline\cr h^{2}&4&2&2&2\\\ Q_{1}&2&2&1&1\\\ Q_{2}&2&1&2&1\\\ Q_{3}&2&1&1&2\end{array}$ Here we assume $Q_{1}$ (and $h^{2}-Q_{1}$) are associated with the element of $\\{\mathcal{Q}_{t}\\}$ with signature $(3,3)$ and $Q_{2}$ and $Q_{3}$ are contributed by the elements with signatures $(1,5)$ and $(5,1)$. While $Q_{1}$ is definable over $\mathbb{R}$, $Q_{2}$ and $Q_{3}$ are not definable over $\mathbb{R}$ (see Section 5.3). The requisite real cycles exist in $[T]=[Q_{2}]+[Q_{3}]$. This follows from the exact sequence in [CT15, §4], using the rationality of $X$ over $\mathbb{R}$ to ensure the vanishing of the unramified cohomology. It would be interesting to deduce this directly using cohomological machinery [Kah96]. However, we do not know whether such $X$ contain quartic scrolls, in general. Let $M$ denote the moduli space of quartic scrolls in a fixed cohomology class on $X$. There is a morphism $\begin{array}[]{rcl}M&\rightarrow&(\mathbb{P}^{6})^{\vee}\\\ T&\mapsto&\operatorname{span}(T)\end{array}$ assigning to each scroll the hyperplane it spans. Hyperplane sections $Y=H\cap X$ containing such scrolls are singular by the Lefschetz hyperplane theorem. Computations in Macaulay2 indicate that a generic such $Y$ has four ordinary singularities. If a complete intersection of two quadrics $Y\subset\mathbb{P}^{5}$ contains a quartic scroll, it contains two families of such scrolls, each parametrized by $\mathbb{P}^{3}$: These arise from residual intersections in quadrics in $I_{T}(2)/I_{Y}(2).$ Thus the residual family has class $2h^{2}-[T]=(h^{2}-[Q_{2}])+(h^{2}-[Q_{3}]).$ The hyperplane sections of $X$ with four singularities should be parametrized by a reducible surface with distinguished component $\Sigma\subset(\mathbb{P}^{6})^{\vee}$. We speculate that $\Sigma$ is a quartic del Pezzo surface, constructed as follows: Consider the pencil of quadrics $\mathcal{Q}_{t}$ defining $X$ and fix the pair of rank-six quadrics $\mathcal{Q}_{t_{2}},\mathcal{Q}_{t_{3}}$ whose maximal isotropic subspaces sweep out $Q_{2}$ and $Q_{3}$. Let $v_{i}\in\mathcal{Q}_{t_{i}}$ denote the vertices and $\ell$ the line they span, which is defined over $k$ even when $t_{2}$ and $t_{3}$ are conjugate over $k$. Projecting from $\ell$ gives a degree-four cover $X\rightarrow\mathbb{P}^{4}.$ Geometrically, the covering group is the Klein four-group and the branch locus consists of two quadric hypersurfaces $Y_{2},Y_{3}$ intersecting in a degree- four del Pezzo surface $S_{23}$. Is $\Sigma\simeq S_{23}$ over $k$? ### 7.3. Dimension six: Threefolds with one apparent double point Construction V indicates that the existence of a threefold $W\subset X$ with one apparent double point yields rationality. The following classification [CMR04] builds on constructions of Edge [Edg32]: * • $\deg(W)=5$: a scroll in planes associated with two lines and twisted cubic, or one line and two conics; * • $\deg(W)=6$: an Edge variety constructed as a residual intersection $Q\cap(\mathbb{P}^{1}\times\mathbb{P}^{3})=\Pi_{1}\cup\Pi_{2}\cup W,$ where $Q$ is a quadric hypersurface and the $\Pi_{i}\simeq\mathbb{P}^{3}$ are fibers of the Segre variety under the first projection; * • $\deg(W)=7$: an Edge variety constructed as a residual intersection $Q\cap(\mathbb{P}^{1}\times\mathbb{P}^{3})=\Pi\cup W$ with the notation as above; * • $\deg(W)=8$: a scroll in lines over $\mathbb{P}^{2}$ of the form $\mathbb{P}(\mathcal{E})$ (one-dimensional quotients of $\mathcal{E}$) where $\mathcal{E}$ is a rank-two vector bundle given as an extension $0\rightarrow\mathcal{O}_{\mathbb{P}^{2}}\rightarrow\mathcal{E}\rightarrow\mathcal{I}_{p_{1},\ldots,p_{8}}(4)\rightarrow 0$ for eight points $p_{1},\ldots,p_{8}\in\mathbb{P}^{2}$, no four collinear or seven on a conic. Given that rationality follows when there are positive-dimensional subvarieties of odd degree (see Section 3.1) we focus on the varieties of even degree. ### 7.4. Dimension six: Degree six Edge variety Let $W\subset\mathbb{P}^{7}$ denote an Edge variety arising as follows: Consider the Segre fourfold $\mathbb{P}^{1}\times\mathbb{P}^{3}\subset\mathbb{P}^{7}$ and take the residual intersection to two copies of $\mathbb{P}^{3}$ $\\{0,\infty\\}\times\mathbb{P}^{3}\subset\mathbb{P}^{1}\times\mathbb{P}^{3}$ in a quadric hypersurface. The resulting threefold $W\simeq\mathbb{P}^{1}\times\Sigma,$ where $\Sigma\subset\mathbb{P}^{3}$ is a quadric hypersurface. Note that the ideal of $W\subset\mathbb{P}^{7}$ is generated by nine quadratic forms. Complete intersections of two quadrics $W\subset Y\subset\mathbb{P}^{7}$ depend on five parameters. A Magma computation shows that a generic such $Y$ has eight ordinary singularities. Suppose we have an embedding $W\hookrightarrow X$, where $X\subset\mathbb{P}^{8}$ is a smooth complete intersection of two quadrics. For fixed $X$, the Hilbert scheme of such threefolds has dimension eight. Realizing $W\simeq\mathbb{P}^{1}\times\mathbb{P}^{1}\times\mathbb{P}^{1}$ we have $\displaystyle c_{1}({\mathcal{N}}_{W/X})$ $\displaystyle=$ $\displaystyle 3(h_{1}+h_{2}+h_{3}),$ $\displaystyle c_{2}({\mathcal{N}}_{W/X})$ $\displaystyle=$ $\displaystyle 8(h_{1}h_{2}+h_{1}h_{3}+h_{2}h_{3}),$ $\displaystyle c_{3}({\mathcal{N}}_{W/X})$ $\displaystyle=$ $\displaystyle 4h_{1}h_{2}h_{3}.$ The Riemann-Roch formula gives $\chi(N_{W/X})=8$. Since $(W\cdot W)_{X}=4$, the primitive class $([W]-\tfrac{3}{2}h^{3})^{2}=4-18+9=-5.$ ###### Remark 7.2. Suppose that $X$ is defined over $\mathbb{R}$, and corresponds to the $(1,2,6)$ case, with signed permutation matrix (see Section 5.4) $\operatorname{diag}(-1,1,1,1,-1,{\bf{1}},-1,1,-1).$ The invariant cycles are: $\begin{array}[]{c|cccccc}&h^{3}&Q_{1}&Q_{2}&Q_{3}&Q_{4}&Q_{5}\\\ \hline\cr h^{3}&4&2&2&2&2&2\\\ Q_{1}&2&0&1&1&1&1\\\ Q_{2}&2&1&0&1&1&1\\\ Q_{3}&2&1&1&0&1&1\\\ Q_{4}&2&1&1&1&0&1\\\ Q_{5}&2&1&1&1&1&0\end{array}$ Here $Q_{1}$ corresponds to the singular fiber of signature $(4,4)$ and the classes $Q_{2},\ldots,Q_{5}$ correspond to the singular fibers of signatures $(2,6)$ and $(6,2)$. The class $[Q_{2}]+[Q_{3}]+[Q_{4}]+[Q_{5}]-[Q_{1}]$ has degree six and self-intersection four. Is it represented by cycles defined over $\mathbb{R}$? Does it admit an Edge variety of degree six over $\mathbb{R}$? Over more general $k$ where the requisite cycles exist? ### 7.5. Dimension six: Degree eight variety Let $\mathcal{E}$ be a stable rank-two vector bundle on $\mathbb{P}^{2}$ with invariants $c_{1}(\mathcal{E})=4L$ and $c_{2}(\mathcal{E})=8L^{2}$. Note that $\Gamma(\mathcal{E})$ has dimension eight, giving an inclusion $V:=\mathbb{P}(\mathcal{E}^{\vee})\subset\mathbb{P}^{7}.$ We have a tautological exact sequence $0\rightarrow\mathcal{O}_{V}(-\xi)\rightarrow\mathcal{E}^{\vee}_{V}\rightarrow Q\rightarrow 0,$ whence $0\rightarrow Q(\xi)\rightarrow T_{V}\rightarrow T_{\mathbb{P}^{2}}\rightarrow 0.$ Thus we have the following $\displaystyle\xi^{2}-4L\xi+8L^{2}$ $\displaystyle=0$ $\displaystyle c(Q(\xi))$ $\displaystyle=1+(2\xi-4L)+(\xi^{2}-4L\xi+8L^{2}),$ $\displaystyle c({\mathcal{T}}_{V})$ $\displaystyle=(1+3L+3L^{2})(c(Q(\xi))$ $\displaystyle=1+(2\xi-L)+(\xi^{2}-4L\xi+8L^{2}+6\xi L-12L^{2}+3L^{2})$ and we find $\displaystyle c_{1}({\mathcal{N}}_{V/X})$ $\displaystyle=$ $\displaystyle 3\xi+L,$ $\displaystyle c_{2}({\mathcal{N}}_{V/X})$ $\displaystyle=$ $\displaystyle 5\xi^{2}-\xi L+2L^{2},$ $\displaystyle c_{3}({\mathcal{N}}_{V/X})$ $\displaystyle=$ $\displaystyle 3\xi^{3}-3\xi^{2}L+2L^{2}\xi-9L^{3}.$ Note that $\deg(L^{3})=0,\quad\deg(L^{2}\xi)=1,\quad\deg(L\xi^{2})=4,\text{ and }\deg(\xi^{3})=8,$ so we conclude that $(V\cdot V)_{X}=14.$ The primitive class $[V]-2h^{3}$ has self-intersection $14-4\cdot 8+16=-2$ which means that $[V]=h^{3}+[Q_{2}]+[Q_{3}],\quad(Q_{2}\cdot Q_{3})=1,$ where $Q_{2}$ and $Q_{3}$ are classes of quadric threefolds in $X$, defined over the algebraic closure. (Up to the action of the Weyl group $W(D_{9})$ this is the only possibility.) Returning to the only remaining case in dimension six where rationality over $\mathbb{R}$ remains open (see Section 6.2): ###### Question 7.3. Let $X\subset\mathbb{P}^{8}$ be a smooth complete intersection of two quadrics over $\mathbb{R}$ in isotopy class $(1,3,5)$. Which classes of codimension- three cycles $X$ are realized over $\mathbb{R}$? Are there varieties with one apparent double point, defined over $\mathbb{R}$, representing these classes? ## References * [BD85] Arnaud Beauville and Ron Donagi. La variété des droites d’une hypersurface cubique de dimension $4$. C. R. Acad. Sci. Paris Sér. I Math., 301(14):703–706, 1985\. * [BW19] Olivier Benoist and Olivier Wittenberg. Intermediate Jacobians and rationality over arbitrary fields, 2019. arXiv:1909.12668. * [BW20] Olivier Benoist and Olivier Wittenberg. The Clemens-Griffiths method over non-closed fields. Algebr. Geom., 7(6):696–721, 2020. * [CMR04] Ciro Ciliberto, Massimiliano Mella, and Francesco Russo. Varieties with one apparent double point. J. Algebraic Geom., 13(3):475–512, 2004. * [CT96] Jean-Louis Colliot-Thélène. Groupes linéaires sur les corps de fonctions de courbes réelles. J. Reine Angew. Math., 474:139–167, 1996. * [CT15] Jean-Louis Colliot-Thélène. Descente galoisienne sur le second groupe de Chow: mise au point et applications. Doc. Math., (Extra vol.: Alexander S. Merkurjev’s sixtieth birthday):195–220, 2015. * [CTSSD87] Jean-Louis Colliot-Thélène, Jean-Jacques Sansuc, and Peter Swinnerton-Dyer. Intersections of two quadrics and Châtelet surfaces. I. J. Reine Angew. Math., 373:37–107, 1987. * [Duc98] Antoine Ducros. L’obstruction de réciprocité à l’existence de points rationnels pour certaines variétés sur le corps des fonctions d’une courbe réelle. J. Reine Angew. Math., 504:73–114, 1998. * [Edg32] W. L. Edge. The number of apparent double points of certain loci. Proc. Cambridge Philos. Soc., 28:285–299, 1932. * [HT19a] Brendan Hassett and Yuri Tschinkel. Cycle class maps and birational invariants, 2019. arXiv:1908.00406 to appear in Communications on Pure and Applied Mathematics. * [HT19b] Brendan Hassett and Yuri Tschinkel. Rationality of complete intersections of two quadrics, 2019. arXiv:1903.08979 with an appendix by J.L. Colliot-Thélène, to appear in L’Enseignement Mathématique. * [HT19c] Brendan Hassett and Yuri Tschinkel. Rationality of Fano threefolds of degree 18 over nonclosed fields, 2019\. arXiv:1910.13816 to appear in the Schiermonnikoog rationality volume. * [Isk79] V. A. Iskovskih. Minimal models of rational surfaces over arbitrary fields. Izv. Akad. Nauk SSSR Ser. Mat., 43(1):19–43, 237, 1979. * [Kah96] Bruno Kahn. Applications of weight-two motivic cohomology. Doc. Math., 1:No. 17, 395–416, 1996. * [Kot00] Robert E. Kottwitz. Involutions in Weyl groups. Represent. Theory, 4:1–15, 2000. * [KP20a] Alexander Kuznetsov and Yuri Prokhorov. Rationality of Fano threefolds over non-closed fields, 2020. arXiv:1911.08949. * [KP20b] Alexander Kuznetsov and Yuri Prokhorov. Rationality of Mukai varieties over non-closed fields, 2020. arXiv:2003.10761. * [Kra18] V. A. Krasnov. On the intersection of two real quadrics. Izv. Ross. Akad. Nauk Ser. Mat., 82(1):97–150, 2018. translation in Izv. Math. 82 (2018), no. 1, 91–-139. * [Lee07] David B. Leep. The Amer-Brumer theorem over arbitrary fields, 2007. available at http://www.ms.uky.edu/~leep/Amer-Brumer_theorem.pdf. * [Rei72] Miles Reid. The complete intersection of two or more quadrics. PhD thesis, Trinity College, Cambridge, 1972. available at http://homepages.warwick.ac.uk/~masda/3folds/qu.pdf. * [Rus00] Francesco Russo. On a theorem of Severi. Math. Ann., 316(1):1–17, 2000. * [Sch96] Claus Scheiderer. Hasse principles and approximation theorems for homogeneous spaces over fields of virtual cohomological dimension one. Invent. Math., 125(2):307–365, 1996.
# Inferring solar differential rotation through normal-mode coupling using Bayesian statistics ∗Samarth G. Kashyap Department of Astronomy and Astrophysics Tata Institute of Fundamental Research Mumbai, India ∗Srijan Bharati Das Department of Geosciences Princeton University Princeton, New Jersey, USA Shravan M. Hanasoge Department of Astronomy and Astrophysics Tata Institute of Fundamental Research Mumbai, India Martin F. Woodard NorthWest Research Associates Boulder Office, 3380 Mitchell Lane Boulder, Colorado, USA Jeroen Tromp Department of Geosciences and Program in Applied & Computational Mathematics Princeton University Princeton, New Jersey, USA (Received December 23, 2021; Revised January 18, 2021; Accepted January 21, 2021) ###### Abstract Normal-mode helioseismic data analysis uses observed solar oscillation spectra to infer perturbations in the solar interior due to global and local-scale flows and structural asphericity. Differential rotation, the dominant global- scale axisymmetric perturbation, has been tightly constrained primarily using measurements of frequency splittings via “$a$-coefficients”. However, the frequency-splitting formalism invokes the approximation that multiplets are isolated. This assumption is inaccurate for modes at high angular degrees. Analysing eigenfunction corrections, which respect cross coupling of modes across multiplets, is a more accurate approach. However, applying standard inversion techniques using these cross-spectral measurements yields $a$-coefficients with a significantly wider spread than the well-constrained results from frequency splittings. In this study, we apply Bayesian statistics to infer $a$-coefficients due to differential rotation from cross spectra for both $f$-modes and $p$-modes. We demonstrate that this technique works reasonably well for modes with angular degrees $\ell=50-291$. The inferred $a_{3}-$coefficients are found to be within $1$ nHz of the frequency splitting values for $\ell>200$. We also show that the technique fails at $\ell<50$ owing to the insensitivity of the measurement to the perturbation. These results serve to further establish mode coupling as an important helioseismic technique with which to infer internal structure and dynamics, both axisymmetric (e.g., meridional circulation) and non-axisymmetric perturbations. Sun: helioseismology — Sun: oscillations — Sun: interior — differential rotation — MCMC ††journal: ApJS††thanks: Both authors have contributed equally to this study. ## 1 Introduction The strength and variation of observed solar activity is governed by the spatio-temporal dependence of flow fields in the convective envelope (Charbonneau, 2005; Fan, 2009). Thus, understanding the physics that governs the evolution and sustenance of the activity cycle of the Sun necessitates imaging its internal layers. While differential rotation has the most significant imprint on Dopplergram images (Schou et al., 1998), signatures due to weaker effects, such as meridional circulation (Giles et al., 1997; Basu et al., 1999; Zhao & Kosovichev, 2004; Gizon et al., 2020) and magnetic fields (Gough, 1990; Dziembowski & Goode, 2004; Antia et al., 2013), are also noticeable. The ability to image these weaker effects therefore critically depends on an accurate measurement of the dominant flows. This makes inferring the strength of the dominant flows along with assigning appropriate statistical uncertainties an important area of study. Differences between normal modes of the Sun and those predicted using standard solar models may be used to constrain solar internal properties. The standard models are typically adiabatic, hydrodynamic, spherically symmetric and non- rotating, also referred to as SNRNMAIS (Lavely & Ritzwoller, 1992; Christensen-Dalsgaard et al., 1996). The usual labelling convention, using 3 quantum numbers, $(n,\ell,m)$, where $n$ denotes radial order, $\ell$ the angular degree, and $m$ the azimuthal order, are used to uniquely identify normal modes. Departures of solar structure from the SNRNMAIS are modelled as small perturbations (Christensen–Dalsgaard, 2003), which ultimately manifest themselves as observable shifts (or splittings) in the eigenfrequencies and distortions in the eigenfunctions (Woodard, 1989). The distorted eigenfunctions may be expressed as a linear combination of reference eigenfunctions and are said to be coupled with respect to the reference. Observed cross-spectra of spherical-harmonic time series corresponding to full-disk Dopplergrams are used to measure eigenfunction distortion. In the present study, we use observational data from the _Helioseismic Magnetic Imager_ (HMI) onboard the _Solar Dynamics Observatory_ (Schou et al., 2012). Different latitudes of the Sun rotate at different angular velocities, with the equator rotating faster than the poles (Howard et al., 1984; Ulrich et al., 1988). To an observer in a frame co-rotating at a specific rotation rate $\bar{\Omega}$ of the Sun, this latitudinal rotational shear is the most significant perturbation to the reference model. This large-scale toroidal flow $\Omega\,(r,\theta)$ is well approximated as being time-independent (shown to vary less than 5% over the last century in Gilman, 1974; Howard et al., 1984; Basu & Antia, 2003) and zonal, with variations only along the radius $r$ and co-latitude $\theta$. Very low $\ell\leq 5$ modes penetrate the deepest layers of the Sun and were used in earlier attempts to constrain the rotation rate in the core and radiative interior (Claverie et al., 1981; Chaplin et al., 1999; Eff-Darwich et al., 2002; Couvidat et al., 2003; Chaplin et al., 2004). However, observed solar activity is believed to be governed by the coupling of differential rotation and magnetic fields in the bulk of the convection zone (Miesch, 2005). Subsequently, studies using intermediate $\ell\leq 100$ (Duvall & Harvey, 1984; Brown & Morrow, 1987; Brown et al., 1989; Libbrecht, 1989; Duvall et al., 1996) and modes with relatively high $\ell\leq 250$ (Thompson et al., 1996; Kosovichev et al., 1997; Schou et al., 1998) yielded overall convergent results for the rotation profile. Among other features of the convection zone (Howe, 2009), these studies established the presence of shear layers at the base of the convection zone (the tachocline) and below the solar surface. Most of these studies used measurements of frequency splittings in a condensed convention known as $a$-coefficients (Ritzwoller & Lavely, 1991). The azimuthal and temporal independence make differential rotation particularly amenable to inversion via $a$-coefficients. The assumption behind this formalism is that multiplets, identified by $(n,\ell)$, are well separated in frequency from each other, known as the ‘isolated multiplet approximation’. This assumption holds true when differential rotation is the sole perturbation under consideration (Lavely & Ritzwoller, 1992), even at considerably high $\ell$. We therefore state at the outset that estimates of $a$-coefficients determined from frequency splitting serve as reliable measures of differential rotation (Chatterjee & Antia, 2009). Nevertheless, the estimation of non- axisymmetric perturbations requires a rigorous treatment honoring the cross coupling of multiplets (Hanasoge et al., 2017; Das et al., 2020). In such cases, measuring changes to the eigenfunctions is far more effective than, for instance, the $a$-coefficient formalism. As a first step, it is therefore important to explore the potential of eigenfunction corrections to infer differential rotation. The theoretical modeling of eigenfunction corrections for given axisymmetric – zonal and meridional – flow fields may be traced back to Woodard (1989), followed up by further investigations Woodard (2000); Gough & Hindman (2010); Vorontsov (2007); Schad et al. (2011). Schad et al. (2013) and Schad & Roth (2020) used observables in the form of mode-amplitude ratios to infer meridional circulation and differential rotation, respectively. In this study, we adopt the closed-form analytical expression for correction coefficients first proposed by Vorontsov (2007) and subsequently verified to be accurate up to angular degrees as high as $\ell\leq 1000$ Vorontsov (2011), henceforth V11. The method of using cross-spectral signals to fit eigenfunction corrections was first applied by Woodard et al. (2013), henceforth W13, to infer differential rotation and meridional circulation. A simple least-squares fitting, assuming a unit covariance matrix, was used for inversions in W13. Their results of odd $a$-coefficients (which encodes differential rotation), even though qualitatively similar, show a considerably larger spread than the results from frequency splittings. Moreover, the authors of W13 note that the inferred meridional flow was “less satisfactory” [than their zonal flow estimates]. Cross spectra are dominated by differential rotation, a much larger perturbation than meridional circulation. Although zonal and meridional flows are measured in different cross-spectral channels, the inference of meridional flow is affected by differential rotation through leakage. Thus, the accurate determination of odd $a$-coefficients is critical to the inference of meridional flow. The relatively large spread in inferences of differential rotation obtained by W13 may be due to (a) a poorly conditioned minimizing function with multiple local minima surrounding the expected (frequency splitting) minima, (b) a relative insensitivity of various modes to differential rotation, resulting in a flat minimizing function close to the expected minima, (c) an inaccurate estimation of the minimizing function on account of assuming a unit-data covariance matrix, and/or (d) eigenfunction corrections only yielding accurate results in the limit of large $\ell$ ($>250$), where the isolated-multiplet approximation starts worsening. In this study, we investigate the above issues and explore the potential of using eigenfunction corrections as a means to infer differential rotation using tools from Bayesian statistics. We apply the Markov Chain Monte Carlo (MCMC) algorithm (Metropolis & Ulam, 1949; Metropolis et al., 1953) using a minimizing function calculated in the L2 norm, adequately weighted by data variance. We do not bias the MCMC sampler in light of any previous measurement, effectively using an uninformed prior. The results inferred, therefore, are an independent measurement constrained only by observed cross spectra. Since Bayesian inference is a probabilistic approach to parameter estimation, we obtain joint probability-density functions in the $a$-coefficient space. This allows us to rigorously compute uncertainties associated with the measurements. We compare and qualify the results obtained with independent measurements from frequency splitting and those obtained using similar cross-spectral analysis in W13. Further, we report the inadequacy of this method for low angular-degree modes on account of poor sensitivity of spectra to rotation via $a$-coefficients. The structure of this paper is as follows. We establish mathematical notations and describe the basic physics of normal-mode helioseismology in Section 2.1. The governing equations which we use for modeling cross spectra using eigenfunction-correction coefficients are outlined in Section 2.2. Section 3 elaborates the steps for computing the observed cross spectra and building the misfit function and estimating data variance for performing the MCMC. Results are discussed in Section 4. Using the $a$-coefficients inferred from MCMC, cross-spectra are reconstructed in Section 4.1. A discussion on sensitivity of the current model to the model parameters is presented in Section 4.2. The conclusions from this work are reported in Section 5. ## 2 Theoretical Formulation ### 2.1 Basic Framework and Notation For inferring flow profiles in the solar interior, we begin by considering the system of coupled hydrodynamic equations, namely, $\displaystyle\partial_{t}\rho$ $\displaystyle=$ $\displaystyle-\bm{\nabla}\cdot(\rho\,\mathbf{v}),$ (1) $\displaystyle\rho(\partial_{t}\mathbf{v}+\mathbf{v}\cdot\bm{\nabla}\mathbf{v})$ $\displaystyle=$ $\displaystyle-\bm{\nabla}P-\rho\bm{\nabla}\phi,$ (2) $\displaystyle\partial_{t}P$ $\displaystyle=$ $\displaystyle-\mathbf{v}\cdot\bm{\nabla}P-\gamma\,P\,\bm{\nabla}\cdot\mathbf{v},$ (3) where $\rho$ is the mass density, $\mathbf{v}$ the material velocity, $P$ the pressure, $\phi$ the gravitational potential and $\gamma$ the ratio of specific heats determined by an adiabatic equation of state. The eigenstates of the Sun are modeled as linear combinations of the eigenstates of a standard solar model. Here we use model S as this reference state, which is discussed in Christensen-Dalsgaard et al. (1996). In absence of background flows, $\tilde{{\mathbf{v}}}={\bf 0}$, the zeroth-order hydrodynamic equations trivially reduce to the hydrostatic equilibrium $\bm{\nabla}\tilde{P}+\tilde{\rho}\bm{\nabla}\tilde{\phi}=0$. Hereafter, all zeroth-order static fields, unperturbed mode eigenfrequencies, eigenfunctions and amplitudes corresponding to the reference model will be indicated using tilde (to maintain consistency with notation used in W13). In response to small perturbations to the static reference model, the system exhibits oscillations $\mbox{\boldmath$\bf\xi$}(\mathbf{r},t)$. These oscillations may be decomposed into resonant “normal modes” of the system, labeled by index $k$, with characteristic frequency $\tilde{\omega}_{k}$ and spatial pattern $\tilde{}\mbox{\boldmath$\bf\xi$}_{k}$, as follows: $\bm{\xi}(\bm{r},t)=\sum_{k}\tilde{\Lambda}_{k}(t)\,\tilde{\bm{\xi}}_{k}(\bm{r})\exp(i\tilde{\omega}_{k}t),$ (4) where $\tilde{\Lambda}_{k}$ are the respective mode amplitudes and $\bm{r}=(r,\theta,\phi)$ denote spherical-polar coordinates. Linearizing eqns. (1)–(3) about the hydrostatic background model gives (for a detailed derivation refer to Christensen–Dalsgaard, 2003) $\mathcal{L}_{0}\tilde{\bm{\xi}}_{k}=-\bm{\nabla}(\tilde{\rho}c_{s}^{2}\,\bm{\nabla}\cdot\tilde{\bm{\xi}}_{k}-\tilde{\rho}g\,\tilde{\bm{\xi}}_{k}\cdot\hat{\bm{e}}_{r})-g\,\hat{\bm{e}}_{r}\bm{\nabla}\cdot(\tilde{\rho}\,\tilde{\bm{\xi}}_{k})=\tilde{\rho}\,\tilde{\omega}_{k}^{2}\,\tilde{\bm{\xi}_{k}}.$ (5) Here $\tilde{\rho}(r),c_{s}(r)$, and $g(r)$ denote density, sound speed, and gravity (directed radially inward) respectively of the reference solar model, and $\mathcal{L}\,_{0}$ is the self-adjoint unperturbed wave operator. This ensures that the eigenfrequencies $\tilde{\omega}_{k}$ are real and eigenfunctions $\tilde{\mbox{\boldmath$\bf\xi$}}_{k}$ are orthogonal. Introducing flows and other structure perturbations through the operator $\delta\mathcal{L}\,$ (e.g., magnetic fields or ellipticity) modifies the unperturbed wave equation (5) to $\tilde{\rho}\,\omega_{k}^{2}\,\bm{\xi}_{k}=\left(\mathcal{L}\,_{0}+\delta\mathcal{L}\,\right)\bm{\xi}_{k},$ (6) where $\omega_{k}=\tilde{\omega}_{k}+\delta\omega_{k}$ and $\mbox{\boldmath$\bf\xi$}_{k}=\sum_{k^{\prime}}c_{k^{\prime}}\tilde{\mbox{\boldmath$\bf\xi$}}_{k^{\prime}}$ are the eigenfrequency and eigenfunction associated with the perturbed wave operator $\mathcal{L}\,_{0}+\delta\mathcal{L}\,$. The Sun, a predominantly hydrodynamic system, is thus treated as a fluid body with vanishing shear modulus (Dahlen & Tromp, 1998). This is unfavourable for sustaining shear waves and therefore the eigenfunctions of the reference model are very well approximated as spheroidal (Chandrasekhar & Kendall, 1957), $\displaystyle\tilde{\bm{\xi}}_{k}(r,\theta,\phi)={}_{n}U{}_{\ell}(r)\,Y_{\ell m}(\theta,\phi)\,\hat{\bm{e}}_{r}+{}_{n}V{}_{\ell}(r)\,\bm{\nabla}_{1}Y_{\ell m}(\theta,\phi).$ (7) $\bm{\bm{\nabla}}_{1}=\hat{\bm{e}}_{\theta}\,\partial_{\theta}+\hat{\bm{e}}_{\phi}\,(\sin\theta)^{-1}\partial_{\phi}$ is the dimensionless lateral covariant derivative operator. Suitably normalized eigenfunctions $\tilde{\mbox{\boldmath$\bf\xi$}}_{k}$ and $\tilde{\mbox{\boldmath$\bf\xi$}}_{k^{\prime}}$, where $k^{\prime}=(n^{\prime},\ell^{\prime},m^{\prime})$, satisfy the orthonormality condition $\int_{\odot}\mathrm{d}^{3}\mathbf{r}\,\rho\,\bm{\tilde{\xi}}_{k^{\prime}}^{*}\cdot\bm{\tilde{\xi}}_{k}=\delta_{n^{\prime}n}\,\delta_{\ell^{\prime}\ell}\,\delta_{m^{\prime}m}.$ (8) Since we observe only half the solar surface, orthogonality cannot be used to extract each mode separately. Windowing in the spatial domain results in spectral broadening, where contributions from neighbouring modes seep into the observed mode signal $\varphi^{\ell m}(\omega)$, as described by the leakage matrix (Schou & Brown, 1994), $\varphi^{\ell m}(\omega)=\sum_{k^{\prime}}L^{\ell m}_{k^{\prime}}\,\Lambda^{k^{\prime}}(\omega)=\sum_{k^{\prime}}\tilde{L}^{\ell m}_{k^{\prime}}\,\ \tilde{\Lambda}^{k^{\prime}}(\omega).$ (9) Here, leakage matrices $L^{\ell m}_{k^{\prime}},\tilde{L}^{\ell m}_{k^{\prime}}$ and amplitudes $\Lambda^{k^{\prime}}(\omega),\tilde{\Lambda}^{k^{\prime}}(\omega)$ of the observed surface velocity field ${\mathbf{v}}(\omega)$ correspond to the bases of perturbed ($\mbox{\boldmath$\bf\xi$}_{k^{\prime}}$) and unperturbed eigenfunctions ($\tilde{\mbox{\boldmath$\bf\xi$}}_{k^{\prime}}$), respectively, ${\mathbf{v}}=\sum_{k^{\prime}}\Lambda^{k^{\prime}}\mbox{\boldmath$\bf\xi$}_{k^{\prime}}=\sum_{k^{\prime}}\tilde{\Lambda}^{k^{\prime}}\tilde{\mbox{\boldmath$\bf\xi$}}_{k^{\prime}}.$ (10) Since leakage falls rapidly with increasing spectral distance $(|\ell-\ell^{\prime}|,|m-m^{\prime}|)$, Eqn. (9) demonstrates the entangling of modes in spectral proximity to $(\ell,m)$. The presence of a zeroth-order flow field $\tilde{{\mathbf{v}}}$ in Eqns. (1)–(3) gives rise to perturbed eigenfunctions $\mbox{\boldmath$\bf\xi$}_{k}$ and therefore introduces correction factors $c_{k}^{k^{\prime}}$ with respect to the unperturbed eigenfunctions $\tilde{\mbox{\boldmath$\bf\xi$}}_{k^{\prime}}$. $\mbox{\boldmath$\bf\xi$}_{k}=\sum_{k^{\prime}}c_{k}^{k^{\prime}}\tilde{\mbox{\boldmath$\bf\xi$}}_{k^{\prime}}.$ (11) Figure 1: Differential rotation induces 3D distortions in radial eigenfunction of unperturbed mode $(n,l)=(2,150)$ for $m=10,75,140$ at radii $r/R_{\odot}=0.95,1.0$. Each column in the upper panel correspond to 2D surfaces for the undistorted eigenfunctions $\tilde{\bm{\xi}}_{nlm}$ in the left slice and differences between distorted and undistorted eigenfunctions $\hat{r}\cdot(\bm{\xi}_{nlm}-\tilde{\bm{\xi}}_{nlm})$ in the right slice. The middle panel shows the difference in the radial variation of eigenfunctions for a chosen $(\theta_{0},\phi_{0})=(67.8^{\circ},177.6^{\circ})$. The lower panels indicate the magnitudes of the coupling coefficients that induce eigenfunction distortion, as in Eqn. (11). The self-coupling coefficients $c^{\ell m}_{\ell m}$ (i.e., $p=0$), being the most dominant, are not shown, in order to highlight the contributions of cross-coupling coefficients ($p\neq 0$.) Using this, the statistical expectation of the cross-spectral measurement is expressed as in Eqns. (14)–(17) of W13, $\langle\varphi^{\ell^{\prime}m^{\prime}}\varphi^{\ell m}\rangle=\sum_{i,j,k}\tilde{L}^{\ell^{\prime}m^{\prime}}_{j}\,\tilde{L}^{\ell m*}_{k}\,c^{j}_{i}\,c^{k*}_{i}\,\langle|\Lambda^{i}(\omega)|^{2}\rangle,$ (12) where $\langle|\Lambda^{i}(\omega)|^{2}\rangle$ denotes Lorentzians centered at resonant frequencies $\omega=\omega_{i}$ corresponding to the perturbed modes $\mbox{\boldmath$\bf\xi$}_{i}$. ### 2.2 Eigenfunction corrections due to axisymmetric flows This study uses the fact that eigenfunction-correction factors $c_{k}^{k^{\prime}}$ in Eqn. (11) carry information about the flow field $\tilde{{\mathbf{v}}}$. Although this problem was first addressed by Woodard (1989), a rigorous treatment using perturbative analysis of mode coupling was only presented in V11. In this section, we outline the governing equations for the eigenfunction-correction factors $c_{k}^{k^{\prime}}$ due to differential rotation and meridional circulation as shown in V11. Upon introducing flows, the model-S eigenfunctions are corrected as follows: $\mbox{\boldmath$\bf\xi$}_{\ell}=\sum_{\ell^{\prime}}c_{\ell}^{\ell^{\prime}}\,\tilde{\mbox{\boldmath$\bf\xi$}}_{\ell^{\prime}}+\delta\mbox{\boldmath$\bf\xi$}_{\ell}=\sum_{p=0,\pm 1,\pm 2,...}c_{\ell}^{\ell+p}\,\tilde{\mbox{\boldmath$\bf\xi$}}_{\ell+p}+\delta\mbox{\boldmath$\bf\xi$}_{\ell},$ (13) where $p=\ell^{\prime}-\ell$ is used to label the offset (in angular degrees) of the neighbouring mode contributing to the distortion of the erstwhile unperturbed eigenfunction $\mbox{\boldmath$\bf\xi$}_{\ell}$ — visual illustration may be found in Figure 1. Correction factors $c_{\ell}^{\ell+p}$ solely from modes with the same radial orders and azimuthal degrees are considered in Eqn. (13) and therefore labels $n$ and $m$ are suppressed. $c_{\ell,m}^{\ell+p,m^{\prime}}=0$ for $m\neq m^{\prime}$ since differential rotation and meridional circulation are axisymmetric (see selection rules imposed due to Wigner 3-$j$ symbols in Appendix A of V11). Corrections from modes belonging to a different radial order $n$ are accumulated in $\delta\mbox{\boldmath$\bf\xi$}$. Following V11 and W13, subsequent treatment ignores terms in $\delta\mbox{\boldmath$\bf\xi$}$ since it is considered to be of the order of the perturbation $\delta\mathcal{L}\,$ or smaller (rendering them at least second order in perturbed quantities). This is because the correction factor $c_{n\ell}^{n^{\prime}\ell^{\prime}}$ is non-trivial only if modes ${}_{n}\mathrm{S}_{\ell}\,$ and ${}_{n^{\prime}}\mathrm{S}_{\ell^{\prime}}\,$ are proximal in frequency space as well as the angular degree $s$ of the perturbing flow satisfies the relation $|\ell^{\prime}-\ell|\leq s$. For modes belonging to different dispersion branches $(n\neq n^{\prime})$, with either $\ell$ or $\ell^{\prime}$ being moderately large ($>50$) the prior conditions are not satisfied, since, for differential rotation, the largest non-negligible angular degree of perturbation is $s=5$. As shown in V11, using eigenfunction perturbations as in Eqn. (13) and eigenfrequency perturbations $\omega_{\ell}=\tilde{\omega}_{\ell}+\delta\omega_{\ell}$, the wave equation (6) reduces to an eigenvalue problem of the form $\mathbf{Z}\,\bm{\mathcal{C}}_{\ell}=\delta\omega_{\ell}\,\bm{\mathcal{C}}_{\ell},$ (14) where $\bm{\mathcal{C}}_{\ell}=\\{...,c_{\ell}^{\ell-1},c_{\ell}^{\ell},c_{\ell}^{\ell+1},...\\}$ are eigenvectors corresponding to the $(P\times P)$ self-adjoint matrix ${\mathbf{Z}}$ and $P=\mathrm{max}(|\ell^{\prime}-\ell|)$ denotes the largest offset of a contributing mode $\ell^{\prime}$ from $\ell$ according as Eqn. (13). From detailed considerations of first- and second-order quasi-degenerate perturbation theory, V11 showed that the following closed-form expression for correction coefficients is accurate up to angular degrees as high as $\ell=1000$: $c_{\ell}^{\ell+p}=\tfrac{1}{\pi}\int_{0}^{\pi}\cos\left[pt-\sum_{k=1,2,...}\tfrac{2}{k}\mathrm{Re}(b_{k})\sin{(kt)}\right]\times\mathrm{exp}\left[i\sum_{k=1,2,...}\tfrac{2}{k}\mathrm{Im}(b_{k})\cos{(kt)}\right]\mathrm{d}t,\qquad p=0,\pm 1,...,$ (15) where the convenient expressions for real and imaginary parts of $b_{k}$ are $\displaystyle\mathrm{Re}(b_{k})$ $\displaystyle=$ $\displaystyle\ell\left(\frac{\partial\tilde{\omega}}{\partial\ell}\right)^{-1}_{n}\sum_{s+k=\mathrm{odd}}(-1)^{\frac{s-k+1}{2}}\frac{(s-k)!!(s+k)!!}{(s+k)!}\times P_{s}^{k}\left(\frac{m}{\ell}\right)\langle\Omega_{s}\rangle_{n\ell},\quad k=1,2,...$ (16) $\displaystyle\mathrm{Im}(b_{k})$ $\displaystyle=$ $\displaystyle k\ell\left(\frac{\partial\tilde{\omega}}{\partial\ell}\right)^{-1}_{n}\sum_{s+k=\mathrm{even}}(-1)^{\frac{s-k+2}{2}}\left(\frac{2s+1}{4\pi}\right)^{1/2}\frac{(s-k-1)!!(s+k-1)!!}{(s+k)!}\times P_{s}^{k}\left(\frac{m}{\ell}\right)\langle\frac{v_{s}}{r}\rangle_{n\ell},\quad k=1,2,....$ (17) We consider only odd-$s$ dependencies of $\Omega$. The even-$s$ correspond to North-South (NS) asymmetry in differential rotation and are estimated to be weak at the surface (NS asymmetry coefficients are estimated to be an order of magnitude smaller than their symmetric counterparts; Mdzinarishvili et al., 2020). The contribution of even-$s$ components to the real part of $b_{k}$ can thus be ignored. For the asymptotic limit of high-degrees, $\mathrm{Re}(b_{k})=\ell\left(\frac{\partial\tilde{\omega}}{\partial\ell}\right)_{n}^{-1}\sum_{s+k=odd}(-1)^{\frac{k-2}{2}}\frac{s!(s-k)!!(s+k)!!}{(s+k)!s!!s!!}\times a^{n\ell}_{s}\,P_{s}^{k}\left(\frac{m}{\ell}\right),\quad k=2,4,...,$ (18) $a^{n\ell}_{s}\approx(-1)^{\frac{s-1}{2}}\frac{s!!s!!}{s!}\langle\Omega_{s}\rangle_{n\ell},\quad s=1,3,...$ (19) Figure 1 illustrates the distortion of eigenfunctions due to an equatorially symmetric differential rotation (using frequency splitting estimates of $a_{3}$ and $a_{5}$ coefficients). It can be seen that differences between distorted eigenfunctions $\bm{\xi}_{nlm}$ and their undistorted counterparts $\tilde{\bm{\xi}}_{nlm}$ are at around the $50\%$ level for some azimuthal orders. The correction coefficients, given by $c^{\ell+p,m}_{\ell,m}$, are shown in the bottom panel of Figure 1. Since the largest contribution to $\bm{\xi}_{\ell}$ comes from $\tilde{\bm{\xi}}_{\ell}$, $c^{\ell,m}_{\ell,m}(\gtrsim 0.8)$ are not plotted to highlight the corrections from neighbouring modes with $p\neq 0$. Visual inspection shows that $c^{\ell+p,m}_{\ell,m}$ have non-zero elements at $p=\pm 2,\pm 4$, as expected from selection rules due to the rotation field $\Omega_{s}(r)$ for $s=3,5$. High $\ell$ eigenfunctions are predominantly large close to the surface. Consequently, we see that their distortions are much larger at shallower than deeper depths. We choose to plot three cases — low, intermediate, and high $m$. For the extreme cases of $m=0$ and $m=\ell$, $c^{\ell+p,m}_{\ell,m}\sim 0$, since for odd $s$ and even $k$, $P_{s}^{k}(\mu)$ vanishes at $\mu=0,1$. Thus these eigenfunctions remain undistorted under an equatorially symmetric differential rotation. For sake of completeness, it may be mentioned that the finite $c^{\ell+p,m}_{\ell,m}$ for $p\neq 0$ seemingly disqualifies the frequency- splitting measurements, which assume isolated multiplets — meaning $c^{\ell+p,m}_{\ell,m}=\delta_{p,0}$. However, it does not necessarily imply that the isolated multiplet approximation is poor at these angular degrees. If the eigenfunction error $\delta\bm{\xi}_{k}$ incurred on neglecting cross- coupling is of order $\mathcal{O}(\epsilon)$ then it can be shown (see Chapter 8 of Freidberg, 2014; Cutler, 2017) that the error in estimating eigenfrequency $\delta\omega_{k}$ is at most of order $\mathcal{O}(\epsilon^{2})$, where $\epsilon$ is small. To illustrate this further, if the error in estimating eigenfunction distortion on neglecting cross coupling $(p\neq 0)$ is written as $\epsilon\,\bm{\xi}_{\ell+p}$, then from inspecting Eqn. (13), we see that $\epsilon\sim|c_{\ell}^{\ell+p}|$. Upon investigating the $(n,\ell)=(2,150)$ case presented in Figure 1 for $p\neq 0$, we find $c_{\ell}^{\ell+p}\lesssim\mathcal{O}(10^{-1})$. The equivalent error incurred in eigenfrequency estimation may be computed according to the discussion in Section 4.4. This yields $\delta\omega/\omega\lesssim\mathcal{O}(10^{-2})$ in the range $150\leq\ell\leq 250$ thereby confirming the above argument for $\epsilon\sim 10^{-1}$. Given the leakage matrices and Lorentzians, the forward problem of modeling $\langle\varphi^{\ell^{\prime}m^{\prime}}\varphi^{\ell m*}\rangle$ requires constructing eigenfunction corrections $c^{\ell+p}_{\ell}$ using the $a$-coefficients in Eqn. (18) and the poloidal flow in Eqn. (17). Thus, for axisymmetric flows, the cross spectra for moderately large $\ell_{1}$ and $\ell_{2}$ from Eqn. (12) may be written more explicitly as $\langle\varphi^{\ell_{1},m_{1}}\,\varphi^{\ell_{2},m_{1}*}\rangle=\sum_{p,p^{\prime},\ell,m}\,\tilde{L}_{\ell+p,m}^{\ell_{1},m_{1}}\,\tilde{L}_{\ell+p^{\prime},m}^{\ell_{2},m_{2}}\,c_{\ell,m}^{\ell+p,m}\,c_{\ell,m}^{\ell+p^{\prime},m*}\,\langle|\Lambda^{\ell,m}(\omega)|^{2}\rangle.$ (20) The leakage matrices $\tilde{L}_{\ell+p,m}^{\ell_{1},m_{1}}$ impose bounds on the farthest modes that leak into mode amplitude $\varphi^{\ell m}$. This is because $\tilde{L}_{\ell+p,m}^{\ell_{1},m_{1}}$ is non-zero only when $\ell+p\in[\ell_{1}-\delta\ell,\ell_{1}+\delta\ell]$ and $m\in[m_{1}-\delta m,m_{1}+\delta m]$, where $\delta\ell$ and $\delta m$ are the farthest spectral offsets. Thus, for a given $\ell$, we must determine the correction coefficients $c_{\ell,m}^{\ell+p,m}$ such that $p\in[\ell_{1}-\delta\ell-\ell,\ell_{1}+\delta\ell-\ell]$. Similar bounds on $p^{\prime}$ in $c_{\ell,m}^{\ell+p^{\prime},m}$ are imposed by the second leakage matrix $\tilde{L}_{\ell+p^{\prime},m}^{\ell_{2},m_{2}}$, namely, $\langle\varphi^{\ell_{1},m_{1}}\,\varphi^{\ell_{1}+\Delta\ell,m_{1}*}\rangle=\sum_{p,p^{\prime},\ell,m}\,\tilde{L}_{\ell+p,m}^{\ell_{1},m_{1}}\,\tilde{L}_{\ell+p^{\prime},m}^{\ell_{1}+\Delta\ell,m_{1}}\,c_{\ell,m}^{\ell+p,m}\,c_{\ell,m}^{\ell+p^{\prime},m*}\,\langle|\Lambda^{\ell,m}(\omega)|^{2}\rangle.$ (21) Being significantly weaker than differential rotation, we neglect the contribution of meridional circulation (Imada & Fujiyama, 2018; Gizon et al., 2020) to the eigenfunction corrections. ## 3 Data Analysis Figure 2: Cross-spectral signal for $\ell=200$, $\Delta\ell=2$ and $n=0$. Panel (a, b): Observed cross-spectrum corresponding to $m^{+}$ and $m^{-}$. Panel (c, d): Derotated cross spectrum corresponding to $m^{+}$ and $m^{-}$. Panel (e, f): $D^{\ell,\Delta\ell,\pm}_{n}$. The baseline is indicated by the dashed blue line. The blue dots represent observations from the five 72-day time series and the red curve corresponds to the expectation value of the cross-spectrum. We use the full-disk 72-day gap-filled spherical-harmonic time series $\varphi^{\ell m}(t)$, which are recorded at a cadence of 45 seconds by HMI (Larson & Schou, 2015). The data are available for harmonic degrees in the range $\ell\leq 300$. The time series is transformed to the frequency domain to obtain $\varphi^{\ell m}(\omega)$. The negative-frequency components are associated with the negative $m$ components using the symmetry relation (Appendix A) $\varphi^{\ell,-|m|}(\omega)=(-1)^{|m|}\varphi^{\ell,|m|*}(-\omega).$ (22) The ensemble average of the cross spectrum is computed by averaging five continuous 72-day time series, which corresponds to 360 days of helioseismic data. The eigenfrequencies of the unperturbed model $\tilde{\omega}_{n\ell m}$ are degenerate in $m$, i.e., $\tilde{\omega}_{n\ell m}=\tilde{\omega}_{n\ell 0}$. Rotation breaks spherical symmetry and lifts the degeneracy in $m$. As in W13, we show the cross-spectrum for $n=0$ and $\ell=200$, $\Delta\ell=2$ in Figure 2. The effect of rotation is visible through the inclination of the ridges in the $m-\nu$ spectrum, as seen in Panels (a, b) of the figure. The multiple vertical ridges are due to leakage of power. The cross spectra are derotated and stacked about the central frequency, corresponding to $m=0$, which is shown in Panels (c, d) of Figure 2. In order to improve the signal-to-noise ratio, the stacked cross spectrum is summed over azimuthal order $m$. This quantity is used to determine the extent of coupling, denoted by $D^{\ell,\Delta\ell,\pm}_{n}$. The $-$ ($+$) signs indicates summation over negative (positive) $m$. For notational convenience, we define $m^{+}$ when referring to $m\geq 0$ and $m^{-}$ to denote $m\leq 0$. The operation of stacking (derotating) the original spectra is denoted by $\mathcal{S}_{m}$. Since differential rotation affects only the real part of the cross-spectrum (see Eqn. 15), $D^{\ell,\Delta\ell,\pm}_{n}$ refers to the real part of the cross-spectrum. $D_{n}^{\ell,\Delta\ell,\pm}(\omega)=\left\langle\sum_{m^{\pm}}\mathcal{S}_{m}\left(\text{Re}\left[\varphi^{\ell m}(\omega)\varphi^{\ell+\Delta\ell,m*}(\omega)\right]\right)\right\rangle.$ (23) The cross-spectral model is a combination of Lorentzians and is based on Eqn. (21). The HMI-pipeline analysis provides us with mode amplitudes and linewidths for multiplets $(n,\ell)$. The $m$ dependence of frequency, $\omega_{n\ell m}-\omega_{nl0}$, is encoded in 36 frequency-splitting coefficients $(a^{nl}_{1},a^{nl}_{2},...,a^{nl}_{36})$. These values are used to construct the Lorentzians for the model, which is denoted by $M^{\ell,\Delta\ell,\pm}$ and expressed as $M^{\ell,\Delta\ell,\pm}_{n}(\omega)=\sum_{m\pm}\mathcal{S}_{m}\left(\sum_{p,p^{\prime},\ell,m^{\prime}}\,\tilde{L}_{\ell+p,m^{\prime}}^{\ell_{1},m}\,\tilde{L}_{\ell+p^{\prime},m^{\prime}}^{\ell_{1}+\Delta\ell,m}\,c_{\ell,m^{\prime}}^{\ell+p,m^{\prime}}\,c_{\ell,m^{\prime}}^{\ell+p^{\prime},m^{\prime}*}\,\langle|\Lambda^{\ell,m^{\prime}}_{n}(\omega)|^{2}\rangle\right).$ (24) As seen in Panels (e, f) of Figure 2, the cross spectra sit on a non-zero baseline. This is a non-seismic background and hence is explicitly fitted for before further analysis of the data. The complete model of the cross spectrum involves leakage from the power spectrum, eigenfunction coupling, as well as the non-seismic background, i.e., the data $D^{\ell,\Delta\ell,\pm}_{n}(\omega)$ is modelled as $M^{\ell,\Delta\ell,\pm}_{n}(\omega)+b^{\ell,\Delta\ell,\pm}_{n}(\omega)$. The baseline $b_{n}^{\ell,\Delta\ell,\pm}(\omega)$ is computed by considering 50 frequency bins on either side, far from resonance, and fitting a straight line through them, in a least-squares sense. The model $M^{\ell,\Delta\ell,\pm}_{n}(\omega)$ depends on the $a^{nl}_{3}$ and $a^{nl}_{5}$ splitting coefficients via the eigenfunction-correction coefficients $c^{\ell+p}_{\ell}$. A Bayesian-analysis approach is used to estimate the values $(a^{nl}_{3},a^{nl}_{5})$, using MCMC, described in Section 3.1. The misfit function that quantifies the goodness of a chosen model is given by $\Xi_{n}=\sum_{l,\omega,\pm}\left(\frac{D^{\ell,\Delta\ell,\pm}_{n}(\omega)-(M^{\ell,\Delta\ell,\pm}_{n}(\omega)+b^{\ell,\Delta\ell,\pm}_{n}(\omega))}{\sigma^{\ell,\Delta\ell,\pm}_{n}(\omega)}\right)^{2},$ (25) where $[\sigma^{\ell,\Delta\ell,\pm}_{n}(\omega)]^{2}$ denotes the variance of the data $D^{\ell,\Delta\ell,\pm}_{n}(\omega)$ and is given by $[\sigma^{\ell,\Delta\ell,\pm}_{n}(\omega)]^{2}=\left\langle\left(\sum_{m\pm}\mathcal{S}_{m}\left[\phi^{\ell,m}(\omega)\,\phi^{\ell+\Delta\ell,m*}(\omega)\right]-D^{\ell,\Delta\ell,m\pm}\right)^{2}\right\rangle.$ (26) ### 3.1 Bayesian Inference: MCMC Bayesian inference is a statistical method to determine the probability distribution functions (PDF) of the inferred model parameters. For data $D$ and model parameters $a$, the posterior PDF $p(a|D)$, which is the conditional probability of the model given data, may be constructed using the likelihood function $p(D|a)$ and a given prior PDF of the model parameters $p(a)$. The prior encapsulates information about what is already known about the model parameters $a$. $p(a|D)\propto p(D|a)p(a).$ (27) The constant of proportionality is the normalization factor for the posterior probability distribution, which may be difficult to compute. The sampling of these PDFs is performed using MCMC, which involves performing a biased random walk in parameter space. Starting from an initial guess of parameters, a random change is performed. The move is accepted or rejected based on the ratio of the posterior probability at the two locations. Hence, the normalization factor is superfluous to the MCMC method. Bayesian MCMC analysis has been used quite extensively in astrophysical problems (Saha & Williams, 1994; Christensen & Meyer, 1998; Sharma, 2017, and references therein) and terrestrial seismology (Sambridge & Mosegaard, 2002, and references therein). However, the use of MCMC in global helioseismology has been limited as compared to terrestrial seismology (Jackiewicz, 2020). The aim of the current calculation is the estimation of $(a_{3}^{n\ell},a_{5}^{n\ell})$ that best reproduce the observed cross-spectra from the model, given by Eqn. (24), where it is seen that the coupling coefficients $c^{\ell+p}_{\ell}$ depend on $(a_{3}^{n\ell},a_{5}^{n\ell})$. However, because of leakage, neighbouring $\ell$ corresponding to the spectrum in question also contribute to the cross-spectrum. Hence, the spectrum of $(\ell,\Delta\ell)$ depends on $(a_{3}^{n\ell^{\prime}},a_{5}^{n\ell^{\prime}})$ for $\ell^{\prime}\in[\ell-\delta\ell,\ell+\Delta\ell+\delta\ell]$. Since we only consider mode leakage at the same radial order $n$, we are forced to simultaneously estimate all the $(a_{3}^{n\ell},a_{5}^{n\ell})$ for a given $n$. For instance, at $n=0$, we have 52 modes with $\ell<250$, and 94 spectra corresponding to $\Delta\ell=2,4$, for both $m^{+}$ and $m^{-}$ branches. In this case, there are 52 $(a_{3}^{0\ell},a_{5}^{0\ell})$ pairs that need to be estimated and 188 spectra which need to be modeled. Performing inversions on a high dimensional, jagged landscape is a challenge as the fine tuning of regularization is tedious. However, since we have a model which encodes the dependence of the $a$ coefficients on the cross-spectrum, we could “brute- force” the estimation of parameters. The utility of MCMC is that it enables us to sample the entire parameter space. Since the inference of the posterior PDF depends strongly on the prior, it is instructive to use an uninformed or flat prior. For the MCMC simulations, we use the Python package emcee by Foreman-Mackey et al. (2013). The package is based on the affine invariant ensemble sampler by Goodman & Weare (2010). Multiple random walkers are used to sample high- dimensional parameter spaces efficiently. We use a flat prior for all $a_{3}^{n\ell}$ and $a_{5}^{n\ell}$ given by $\displaystyle p(a_{3})=\frac{1}{20}\qquad 15\leq a_{3}\leq 35\qquad\text{and}\qquad p(a_{5})=\frac{1}{16}\qquad-16\leq a_{5}\leq 0,$ (28) and zero everywhere else for all $(\ell,n)$. This is motivated by the results of frequency splittings. For modes near the surface, i.e., for low values of $\nu_{n\ell}/\ell$, $a_{3}$ has been measured to be nearly $22$ nHz and $a_{5}$ is $-4$ nHz. The likelihood function is defined as $p(D|a)=\exp(-\Xi_{n}),$ (29) where $\Xi_{n}$ is the misfit given by Eqn. (25). Flat priors enable us to sample the likelihood function in the given region in parameter space. We perform MCMC inversions for $n=0,1,..8$ and find that the likelihood function is unimodal in all model parameters. For the sake of illustration, a smaller computation is presented in Appendix B. Radial order $n$ | Range of $\ell$ for $(a_{3},a_{5})$ ---|--- 0 | 192–241, 241–281, 271–289 1 | 80–120, 110–150, 140–183 2 | 60–100, 90–130, 120–161 3 | 43–73, 73–113, 103–145 4 | 40–80, 70–110, 100–140 5 | 46–86, 76–116, 106–146 6 | 58–98, 88–128, 118–138 7 | 64–104, 94–114 8 | 73–103 Table 1: List of modes $(n,\ell)$ used in MCMC. These are marked as black dot in Figure 1. Figure 3: Classification of modes. ## 4 Results and Discussion The MCMC analysis is performed for each radial order separately. The current model only considers leakage between modes of the same radial order and hence the ideal way of estimating the parameters would be to estimate all $(a_{3},a_{5})$ at a given radial order by modelling all the cross-spectra at the same radial order. However, this makes the problem computationally very demanding as the MCMC method used requires at least $2k+1$ random walkers for $k$ different parameters to be fit. To work around this, we break the entire set of parameters into chunks of 40 pairs, while ensuring an overlap of 10 pairs between the chunks. In Table 1, we list the set of $\ell$’s for which MCMC sampling is performed and parameters are estimated. Figure 1 marks the multiplets $(n,\ell)$ available from the HMI pipeline. The multiplets whose modes are used for this study are labelled as black dots. The red dots, which are located at lower $\ell$, correspond to those modes which have contributions from neighbouring radial orders within the temporal- frequency window. This gets worse for $\ell<20$, where contributions from neighbouring radial orders may be seen even near central peaks. Modelling these spectra would require including coupling across radial orders, which is not the case in the present analysis. Thus we only use modes corresponding to $\nu_{n\ell}/\ell<45$. Figure 1 also marks unused HMI-resolved modes as blue dots on either side of the black dots (used modes). This is because we consider only modes that may be fully modelled with parameters available from the HMI pipeline. Modelling a the degree $\ell$ requires mode parameters corresponding to modes from $(\ell-\delta\ell)$ to $(\ell+\delta\ell)$. The existence of unresolved modes (with no mode-parameter information from the HMI pipeline) in this region means that modelling is incomplete, i.e., there would be peaks in the observed spectrum that are missed by the model. Hence, such modes are not considered for the present work. For any given radial order, the first $\delta\ell$ and the last $\delta\ell$ modes cannot be modelled and thus we see blue points on either side of the set of black dots in Figure 1. The results of the MCMC analysis at all the radial orders are combined and presented in Figure 6. We note that that the confidence intervals become larger for higher $\nu/\ell$. The reasons for this are discussed in Section 4.2. Estimates of $a$-coefficients are largely in agreement with the splitting coefficients — although the most probable values of the coupling-derived parameters are different from their splitting counterparts, they predominantly lie within the 1-$\sigma$ confidence interval. The confidence intervals of $a_{3}$ and $a_{5}$ are nearly the same size. We obtain better results, in terms of the spread in the inferred $a_{3}$-coefficients, than W13. This may be attributed to the consideration of data variance as well as simultaneous fitting for model parameters using a Bayesian approach. For instance, the spread of $a_{3}$ in the range $0<\nu/\ell\lesssim 40$ is seen to be in the range 7.5–30 nHz in W13, whereas our estimates are in the range 15–26 nHz. The present method allows us to quantify the 1-$\sigma$ confidence interval around the most probable values for estimated $a$-coefficients, whereas W13 have shown only inversion values of $a$-coefficients without their respective uncertainties. However, we also note that the estimates of $a_{5}$ from Bayesian analysis are comparable to the least-squares inversions of W13. ### 4.1 Reconstructed power and cross spectra Figure 4: Cross spectrum for $\ell=222$ and $\Delta\ell=0,2,4$. The upper panels correspond to $m^{+}$ and lower panels to $m^{-}$. The black curve shows observed data. The blue curve is the model before considering eigenfunction coupling and the red curve corresponds to model constructed using parameters estimated from MCMC. Figure 5: Cross spectrum for $\ell=70$ and $\Delta\ell=0,2,4$. The upper panels correspond to $m^{+}$ and lower panels to $m^{-}$. The black curve shows observed data. The blue curve is the model before considering eigenfunction coupling and the red curve corresponds to the model constructed using parameters estimated from MCMC. The $a$-coefficients obtained from the MCMC analysis are used to reconstruct cross-spectra, e.g., Figure 4 shows the cross spectrum for $(n=0,\ell=222)$. It may be seen that, before considering eigenfunction corrections (in the absence of differential rotation), the spectrum shown in blue is considerably different — in both magnitude and sign — from the observed data. After including eigenfunction corrections, which have been estimated from MCMC, we see that the model is in close agreement with the data. In the intermediate-$\ell$ range, we show cross-spectra for $(n=4,\ell=70)$ in Figure 5. The corrections due to eigenfunction distortion are markedly less significant when compared to $(\ell=222,n=0)$, demonstrating loss of sensitivity of the model to the coupling coefficients. Figure 6: Inferred $a_{3}$ and $a_{5}$ coefficients from MCMC are shown as black dots with 1-$\sigma$ confidence intervals. The values from frequency splitting are shown in red. ### 4.2 Sensitivity of $a$-coefficients to differential rotation Mode coupling has diminished sensitivity in estimating $a$-coefficients for low-$\ell$ modes. The coupling coefficients $c_{\ell}^{\ell+p}$ depend on the real and imaginary parts of $b_{k}$. Differential rotation contributes to only the real part of $b_{k}$ (Eqn. [16]) and the dependence on $\ell$ appears through the factor $\ell(\partial\omega_{nl}/\partial\ell)^{-1}$. The plot of eigenfrequencies $\omega_{n\ell}$ against $\ell$ is known to flatten for higher $\ell$. Hence, $\partial\omega_{n\ell}/\partial\ell$ is large for small $\ell$ and small for large $\ell$ (see Figure 1 in Rhodes et al., 1997). This results in $b_{k}$ being small for low $\ell$ and its magnitude increases with $\ell$, causing this decreased sensitivity to low $\ell$. The lower sensitivity implies that the misfit function $S$ is flatter at lower $\ell$. To demonstrate this, we compute $S$ over $\ell=80$–245 for a range of values of $a-$coefficients and determine how wide or flat $S$ is in the neighbourhood of the optimal solution. Figure 7 shows that the misfit is wide for $\ell=80$ and it becomes sharper with increasing $\ell$. As the highest-resolved mode for $n=1$ corresponds to $\ell=179$, we consider the radial order $n=0$ in order to study this in an extended region of $\ell$. The first two panels show the colour map of the misfit function. Near the optimal value $a^{n\ell}_{s}/a^{n\ell}_{FS}=1$, the synthetic misfit falls to $0$. This is possible as the synthetic data is noise free and it can be completely modeled. The misfit increases on either side of the optimum value. The second panel shows the scaled misfit for HMI data, which is close to $1$ at the optimum, increasing on either side of the optimal value. We see the dark patch become wider at lower $\ell$, indicating the flatness of the misfit function for low $\ell$. The likelihood function, which is defined to be $\exp(-S)$, is approximated as a Gaussian in the vicinity of the optimum. The width of this Gaussian is treated as a measure of the width of the misfit function $S$, with wider misfit implying lower sensitivity to $a-$coefficients. This is shown in the third panel of Figure 7, where we see a decreasing trend in misfit width, indicating that the sensitivity of mode coupling increases with $\ell$. Figure 7: Sensitivity of spectral fitting to $a$-coefficients as a function of angular degree $\ell$. Top panel shows the variation of misfit between synthetic data calculated using frequency-splitting $a$-coefficients $a_{s,\mathrm{FS}}^{n\ell}$ and synthetic spectra computed from a scaled set of $a$ coefficients $a_{s}^{n\ell}$. A well-defined minimum along $a_{s}^{n\ell}/a_{s,\mathrm{FS}}^{n\ell}=1.0$, which broadens towards smaller $\ell$, shows a drop in sensitivity of the spectra to variations in $a$ coefficients, as predicted by theory. Middle panel shows the sensitivity of $a$ coefficients, but now computed using the misfit between HMI and synthetic spectra computed from a scaled set of $a$ coefficients $a_{s}^{n\ell}$. While it has the same qualitative drop in $a$-coefficient sensitivity for decreasing $\ell$, the ridge of the minimum (darkest patch) is seen to deviate from $a_{s,\mathrm{FS}}^{n\ell}$. Bottom panel shows in black the effective variance of misfit for each $\ell$. The narrowing confinement of the data misfit towards higher $\ell$ is seen as a decreasing effective variance with increasing $\ell$. The red line shows the increase in the factor $\ell/(\partial\omega_{n\ell}/\partial\ell)$ that enhances sensitivity at higher $\ell$, as predicted by Eqn. (18). The areas corresponding to radial orders $n=0,1$ are indicated on top of each plot. ### 4.3 Scaling factor for synthetic spectra The model constructed using mode parameters obtained from the HMI pipeline needs to be scaled to match the observations. This scaling factor has to be empirically determined. Since there is no well-accepted convention to estimate this factor, it is worthwhile to explore different methods of its estimating. We employ three different methods to infer the scale factor and show that the results are nearly identical. * • Consider all the power spectra for a given radial order and perform a least- squares fitting for the scale factor $N_{0}$. * • Fit for the scale factor $N_{\ell}$ as a function of the spherical harmonic degree $\ell$ by considering all power spectra at a given radial order. * • Include the scale factor as an independent parameter to be estimated in the MCMC analysis. Figure 8 shows that all the independent ways of estimating the scale factor are within 5% of each other, indicating robustness. Figure 8: The red line corresponds to $N_{0}$. The gray region corresponds to 5% error from $N_{0}$. The gray points correspond to $N_{\ell}$ and the solid black lines are from each MCMC simulation. The right-most panel shows the histogram of all the gray points, taken from all radial orders. ### 4.4 How good is the isolated multiplet approximation for $\ell\leq 300$? Figure 9: The relative offset of $L_{2}^{\text{QDPT}}$ as compared to that of $L_{2}^{\text{DPT}}$ (see Eqn [30,31]) under the perturbation of an axisymmetric differential rotation $\Omega(r,\theta)$ as observed in the Sun. An increase in intensity of the color scale indicates worsening of the isolated multiplet approximation. The measures of offset are plotted for the HMI-resolved multiplets shown in Figure 1. Estimation of $a$ coefficients through frequency splitting measurements assumes validity of the isolated multiplet approximation using degenerate perturbation theory (DPT). However, an inspection of the distribution of the multiplets in $\nu-\ell$ space (as shown in Fig. [1]) shows that it is natural to expect this approximation to worsen with increasing $\ell$. This necessitates carrying out frequency estimation respecting cross-coupling of modes across multiplets, also known as quasi-degenerate perturbation theory (QDPT). A detailed discussion on DPT and QDPT in the context of differential rotation can be found in Ritzwoller & Lavely (1991) and Lavely & Ritzwoller (1992). In this section we discuss the goodness of the isolated multiplet approximation in estimating $a_{n\ell}$ due to $\Omega(r,\theta)$ for all the HMI-resolved modes shown in Figure 1. A similar result but for $\ell\leq 30$ was presented in Appendix G of Das et al. (2020). In Figure 9 we color code multiplets to indicate the departure of frequency shifts obtained from QDPT $\delta{}_{n}\omega{}_{\ell m}^{Q}$ as compared to shifts obtained from DPT $\delta{}_{n}\omega{}_{\ell m}^{D}$. Strictly speaking, carrying out the eigenvalue problem in the QDPT formalism causes garbling of the quantum numbers —$n$, $\ell$, and $m$ are no longer good quantum numbers— and prevents a one-to-one mapping of unperturbed to perturbed modes. This prohibits an explicit comparison of frequency shifts on a singlet-by-singlet basis. However, modes belonging to the same multiplet can still be identified visually and grouped together. So, to quantify the departure of $\delta{}_{n}\omega{}_{\ell m}^{Q}$ from $\delta{}_{n}\omega{}_{\ell m}^{D}$ we calculate the Frobenius norm of these frequency shifts corresponding to each multiplet: $\displaystyle L_{2}^{\text{QDPT}}$ $\displaystyle=$ $\displaystyle\sqrt{\sum_{m}(\delta{}_{n}\omega{}_{\ell m}^{\text{Q}})^{2}}\qquad\text{for cross-coupling,}$ (30) $\displaystyle L_{2}^{\text{DPT}}$ $\displaystyle=$ $\displaystyle\sqrt{\sum_{m}(\delta{}_{n}\omega{}_{\ell m}^{\text{D}})^{2}}\qquad\text{for self-coupling.}$ (31) The color scale intensity in Figure 9 indicates the relative offset of $L_{2}^{\text{QDPT}}$ as compared to $L_{2}^{\text{DPT}}$ for a multiplet $(n,\ell)$ marked as an ‘o’. Larger offset indicates the degree of worsening of the isolated multiplet approximation. We find that the largest error incurred using DPT instead of QPDT is 0.27% this is found to be at $\ell=300$. This clearly shows that even for the $f$ mode (which is the most susceptible to errors) the frequency splitting $a$-coefficients are exceptionally accurate. ## 5 Conclusion Most of what is currently known about solar differential rotation is derived from from $a$-coefficients using frequency splitting measurements. Inferring these $a$-coefficients involves invoking the isolated multiplet approximation based on degenerate perturbation theory. Although this approximation works well even for high $\ell\leq 300$ modes, reasons motivating the need to investigate the possibility of erroneous $a$-coefficients from frequency splitting measurements at even higher $\ell$ stem from a combination of two effects, namely, the increasing proximity of modes (in frequency) along the same radial branch, and spectral-leakage from neighbouring modes. Partial visibility of the Sun causes broadening of peaks in the spectral domain, referred to as mode leakage (Schou & Brown, 1994; Hanasoge, 2018). This causes proximal modes at high $\ell$ to widen and resemble continuous ridges in observed spectra. As a result, spectral-peak identification for frequency- splitting measurements are harder and increasingly inaccurate. Moreover, since the $a$-coefficient formalism breaks down for non-axisymmetric perturbations, considering techniques which respect cross-coupling becomes indispensable. Thus, mode coupling becomes more relevant in these regimes, and it is important to investigate the potential of mode-coupling techniques as compared to frequency splittings. Hence, this study was directed towards answering the following broad questions. (i) Can mode-coupling via MCMC use information stored in eigenfunction distortions to constrain differential rotation as accurately as frequency splittings? This would also serve to compare the potential of a Bayesian approach with the least square inversion performed in W13. (ii) Can this technique further increase the accuracy of $a_{n\ell}$ at $\ell\geq 150$? We already know that higher $\ell$ estimates are increasingly precise and accurate from W13. (iii) What are the uncertainties in estimating $a_{n\ell}$ using mode-coupling theory and do they fall within 1-$\sigma$ of frequency splitting estimates? (iv) Why are mode-coupling results poorer in the low $\ell$ regime? This is seen in earlier studies, which aimed to go deeper into the convection zone and obtained significantly imprecise and inaccurate results (Woodard et al., 2013; Schad & Roth, 2020). The approach in this study is broadly based on the theoretical formulations from V11 and modelling from W13. However, the novelty of the current work lies in 3 main aspects. (a) The MCMC analysis enabled exploration of the complete parameter space, and it was found that the chosen misfit function is unimodal in nature, for all degrees $\ell$ and radial orders $n$. This establishes that the method of normal-mode coupling does return a unique value of $(a_{3},a_{5})$. (b) Leakage of power occurs for modes in the same radial order $n$ and hence the determination of $(a_{3},a_{5})$ in a consistent manner would involve simultaneous estimation of splitting coefficients for all $\ell$ and the same radial order. However, the number of parameters is large and hence we break it into chunks of 40 pairs of $(a_{3},a_{5})$ per MCMC, with an overlap of $N_{o}$ pairs of the parameters between two different chunks. To settle on a reasonable value of $N_{o}$, we perform a simple experiment. From MCMC simulations with different overlap numbers $N_{o}=\\{0,2,4,6,8,10\\}$, we find that for $N_{o}>6$, the inferred $a$-coefficients vary less than 1-$\sigma$ and therefore reasonably stable for larger $N_{o}$. Hence, we choose the modal overlap number $N_{o}=10$ for computation at all radial orders. (c) Since a large number of splitting coefficients are determined simultaneously, a corresponding number of spectra is used. Hence, estimation of the data variance becomes critical in order to appropriately weight different data points according to their noise levels. These improvements lead to a better estimate of differential rotation using mode coupling. The inference of rotation at lower $\ell$ ($<50$) suffers for two reasons. (a) Low sensitivity of the model to the $a$-coefficients. (b) Proximity of modes of radial orders $(n+1)$ and $(n-1)$ to modes at radial order $n$. Since the current model only accounts for leakage of power within the same radial order, a chosen frequency window in data would contain peaks from neighboring radial orders, which are not modelled. Hence, an improvement might be achieved at lower $\ell$ by modelling the interaction of modes of different radial orders. Finally, in this study we also show that even though frequency splitting is much more precise for low $\ell\leq 150$, mode coupling estimates of differential rotation improves at high $\ell\geq 200$. Therefore, it is expected that mode-coupling would be comparable to (or possibly more accurate than) frequency splitting for very high $\ell\geq 300$. This would then allow one to compare mode-coupling estimates of shallow, small-scale structures with results from methods in local helioseismology. Going this high in angular degree for mode-coupling, however, introduces some challenges: (a) The computation of leakage matrices for high $\ell$ is very expensive. (b) $\partial\omega/\partial\ell$ decreases as $\ell$ grows and the spectrum becomes a continuous ridge in frequency space making it harder to resolve the modes completely. In conclusion, there remains scope for improvement and related lines of study. In this study, we have ignored the even-$s$ components of $\Omega_{s}$, which are the NS-asymmetric components of differential rotation. These components have been estimated to be small at the surface and are anticipated to be small in the interior. However, this assumption may be premature given that prior estimates of interior rotation-asymmetries are based on non-seismic surface measurements. Since the V11 formalism is capable of accommodating the estimation of even-$s$ components as well, this could be the focus of a future investigation. Additionally, the current analysis was performed after summing up the stacked cross-spectrum. Although this was done to improve the signal- to-noise ratio, the spectrum at different azimuthal orders $m$ are not identical. Hence a more complete computation would involve the misfit computed using the full spectrum as a function of $m$. This may possibly lead to better results of the $a$-coefficients, as there exists structure in the azimuthal order (see Fig. [2]), which is lost after summation. The authors of this study are grateful to Jesper Schou (Max Planck Institute for Solar System Research) for numerous insightful discussions as well as detailed comments that helped us improve the quality of the manuscript. The authors thank the anonymous referee for valuable suggestions that helped improve the text and figures in this manuscript. ## Appendix A Spherical harmonics symmetry relations Consider a time-varying, real-valued scalar field on a sphere $\phi(\theta,\phi,t)$. The spherical harmonic components are given by $\phi^{l,|m|}(t)=\int_{\Omega}{\mathrm{d}}\Omega Y^{*l,|m|}(\theta,\phi)\phi(\theta,\phi,t)=(-1)^{|m|}\int_{\Omega}{\mathrm{d}}\Omega Y^{l,-|m|}\phi(\theta,\phi,t)=(-1)^{|m|}\phi^{*l,-|m|}(t)$ (32) where $d\Omega$ is the area element, the integration being performed over the entire surface of the sphere. After performing a temporal Fourier transform, we have $\phi^{l,|m|}(\omega)=\frac{1}{\sqrt{2\pi}}\int_{-\infty}^{\infty}{\mathrm{d}}te^{-i\omega t}\phi^{l,|m|}(t)=(-1)^{|m|}\frac{1}{\sqrt{2\pi}}\int_{-\infty}^{\infty}{\mathrm{d}}te^{-i\omega t}\phi^{*l,-|m|}(t)$ (33) $\phi^{*l,-|m|}(\omega)=\frac{1}{\sqrt{2\pi}}\int_{-\infty}^{\infty}{\mathrm{d}}te^{i\omega t}\phi^{*l,-|m|}(t)=(-1)^{|m|}\phi^{l,|m|}(-\omega)\implies\phi^{l,-|m|}(\omega)=(-1)^{|m|}\phi^{*l,|m|}(-\omega)$ (34) ## Appendix B MCMC: An illustrative Case We present an MCMC estimation of $a$-coefficients using a smaller set of modes (and hence model parameters). The smaller number of model parameters lets us present all the marginal probabilities in a single plot. The MCMC walkers are shown in Figure 10. In spite of using a flat prior, the likelihood function is sharp enough to bias the walkers to move towards the region of optimal solution within $\sim 500$ iterations. It can be seen that different walkers start off randomly at different locations in parameter space and ultimately converge to the same region around the optimal solution. After removing the iterations from the “burn-in” period, where the walkers are still exploring a larger parameter space, histograms are plotted and marginal probability distributions are obtained. Figure 11 shows one such estimation of $(a_{3},a_{5})$ for $n=0$ and $\ell$ in the range $200$ to $202$. It can be seen that the marginal posterior probability distributions for each of the parameters are unimodal. This tells us that the currently defined misfit function has a unique minimum. Note that this distribution was obtained using a flat prior and hence the resulting posterior distributions are essentially sampling the likelihood function. It is also worth noting that for the range of $\ell$’s chosen, the confidence intervals are $<1$ nHz. Figure 10: Each parameter is shown in a different figure to indicate the value as a function of the Markov Chain step number. The first few “burn-in” values are discarded and only the values beyond the vertical line are considered to obtain the probability distributions. Figure 11: Cross-correlation of model parameters and the marginal probability of the model parameters. ## References * Antia et al. (2013) Antia, H. M., Chitre, S. M., & Gough, D. O. 2013, MNRAS, 428, 470, doi: 10.1093/mnras/sts040 * Basu & Antia (2003) Basu, S., & Antia, H. M. 2003, The Astrophysical Journal, 585, 553, doi: 10.1086/346020 * Basu et al. (1999) Basu, S., Antia, H. M., & Tripathy, S. C. 1999, ApJ, 512, 458, doi: 10.1086/306765 * Brown et al. (1989) Brown, T. M., Christensen-Dalsgaard, J., Dziembowski, W. A., et al. 1989, ApJ, 343, 526, doi: 10.1086/167727 * Brown & Morrow (1987) Brown, T. M., & Morrow, C. A. 1987, ApJ, 314, L21, doi: 10.1086/184843 * Chandrasekhar & Kendall (1957) Chandrasekhar, S., & Kendall, P. C. 1957, The Astrophysical Journal, 126, 457 * Chaplin et al. (2004) Chaplin, W. J., Sekii, T., Elsworth, Y., & Gough, D. O. 2004, MNRAS, 355, 535, doi: 10.1111/j.1365-2966.2004.08338.x * Chaplin et al. (1999) Chaplin, W. J., Christensen-Dalsgaard, J., Elsworth, Y., et al. 1999, MNRAS, 308, 405, doi: 10.1046/j.1365-8711.1999.02691.x * Charbonneau (2005) Charbonneau, P. 2005, Living Reviews in Solar Physics, 2, 2 * Chatterjee & Antia (2009) Chatterjee, P., & Antia, H. M. 2009, ApJ, 707, 208, doi: 10.1088/0004-637X/707/1/208 * Christensen–Dalsgaard (2003) Christensen–Dalsgaard, J. 2003, Lecture Notes on Stellar Oscillations, 5th edn. * Christensen & Meyer (1998) Christensen, N., & Meyer, R. 1998, Phys. Rev. D, 58, 082001, doi: 10.1103/PhysRevD.58.082001 * Christensen-Dalsgaard et al. (1996) Christensen-Dalsgaard, J., Dappen, W., Ajukov, S. V., et al. 1996, Science, 272, 1286 * Claverie et al. (1981) Claverie, A., Isaak, G. R., McLeod, C. P., van der Raay, H. B., & Roca Cortes, T. 1981, Nature, 293, 443, doi: 10.1038/293443a0 * Couvidat et al. (2003) Couvidat, S., García, R. A., Turck-Chièze, S., et al. 2003, ApJ, 597, L77, doi: 10.1086/379698 * Cutler (2017) Cutler, C. 2017, Using eigenmode-mixing to measure or constrain the Sun’s interior B-field. https://arxiv.org/abs/1706.07404 * Dahlen & Tromp (1998) Dahlen, F. A., & Tromp, J. 1998, Theoretical Global Seismology (Princeton University Press) * Das et al. (2020) Das, S. B., Chakraborty, T., Hanasoge, S. M., & Tromp, J. 2020, ApJ, 897, 38, doi: 10.3847/1538-4357/ab8e3a * Duvall et al. (1996) Duvall, Jr., T. L., D’Silva, S., Jefferies, S. M., Harvey, J. W., & Schou, J. 1996, Nature, 379, 235, doi: 10.1038/379235a0 * Duvall & Harvey (1984) Duvall, Jr., T. L., & Harvey, J. W. 1984, Nature, 310, 19 * Dziembowski & Goode (2004) Dziembowski, W. A., & Goode, P. R. 2004, ApJ, 600, 464, doi: 10.1086/379708 * Eff-Darwich et al. (2002) Eff-Darwich, A., Korzennik, S. G., & Jiménez-Reyes, S. J. 2002, ApJ, 573, 857, doi: 10.1086/340747 * Fan (2009) Fan, Y. 2009, Living Reviews in Solar Physics, 6, 4, doi: 10.12942/lrsp-2009-4 * Foreman-Mackey et al. (2013) Foreman-Mackey, D., Hogg, D. W., Lang, D., & Goodman, J. 2013, PASP, 125, 306, doi: 10.1086/670067 * Freidberg (2014) Freidberg, J. P. 2014, Ideal MHD (Cambridge University Press), doi: 10.1017/CBO9780511795046 * Giles et al. (1997) Giles, P. M., Duvall, Jr., T. L., Scherrer, P. H., & Bogart, R. S. 1997, Nature, 390, 52 * Gilman (1974) Gilman, P. A. 1974, ARA&A, 12, 47, doi: 10.1146/annurev.aa.12.090174.000403 * Gizon et al. (2020) Gizon, L., Cameron, R. H., Pourabdian, M., et al. 2020, Science, 368, 1469, doi: 10.1126/science.aaz7119 * Goodman & Weare (2010) Goodman, J., & Weare, J. 2010, Communications in Applied Mathematics and Computational Science, 5, 65, doi: 10.2140/camcos.2010.5.65 * Gough & Hindman (2010) Gough, D., & Hindman, B. W. 2010, ApJ, 714, 960, doi: 10.1088/0004-637X/714/1/960 * Gough (1990) Gough, D. O. 1990, in Lecture Notes in Physics, Berlin Springer Verlag, Vol. 367, Progress of Seismology of the Sun and Stars, ed. Y. Osaki & H. Shibahashi, 283, doi: 10.1007/3-540-53091-6 * Hanasoge (2018) Hanasoge, S. 2018, ApJ, 861, 46, doi: 10.3847/1538-4357/aac3e3 * Hanasoge et al. (2017) Hanasoge, S. M., Woodard, M., Antia, H. M., Gizon, L., & Sreenivasan, K. R. 2017, MNRAS, 470, 1404, doi: 10.1093/mnras/stx1298 * Howard et al. (1984) Howard, R., Gilman, P. I., & Gilman, P. A. 1984, ApJ, 283, 373, doi: 10.1086/162315 * Howe (2009) Howe, R. 2009, Living Reviews in Solar Physics, 6, 1. https://arxiv.org/abs/0902.2406 * Imada & Fujiyama (2018) Imada, S., & Fujiyama, M. 2018, ApJ, 864, L5, doi: 10.3847/2041-8213/aad904 * Jackiewicz (2020) Jackiewicz, J. 2020, Sol. Phys., 295, 137, doi: 10.1007/s11207-020-01667-3 * Kosovichev et al. (1997) Kosovichev, A. G., Schou, J., Scherrer, P. H., et al. 1997, Sol. Phys., 170, 43 * Larson & Schou (2015) Larson, T. P., & Schou, J. 2015, Sol. Phys., 290, 3221, doi: 10.1007/s11207-015-0792-y * Lavely & Ritzwoller (1992) Lavely, E. M., & Ritzwoller, M. H. 1992, Philosophical Transactions of the Royal Society of London Series A, 339, 431, doi: 10.1098/rsta.1992.0048 * Libbrecht (1989) Libbrecht, K. G. 1989, ApJ, 336, 1092, doi: 10.1086/167079 * Mdzinarishvili et al. (2020) Mdzinarishvili, T., Shergelashvili, B., Japaridze, D., et al. 2020, Advances in Space Research, 65, 1843 , doi: https://doi.org/10.1016/j.asr.2020.01.015 * Metropolis et al. (1953) Metropolis, N., Rosenbluth, A. W., Rosenbluth, M. N., Teller, A. H., & Teller, E. 1953, J. Chem. Phys., 21, 1087, doi: 10.1063/1.1699114 * Metropolis & Ulam (1949) Metropolis, N., & Ulam, S. 1949, Journal of the American Statistical Association, 44, 335. http://www.jstor.org/stable/2280232 * Miesch (2005) Miesch, M. S. 2005, Living Reviews in Solar Physics, 2, 1 * Rhodes et al. (1997) Rhodes, E. J., J., Kosovichev, A. G., Schou, J., Scherrer, P. H., & Reiter, J. 1997, Sol. Phys., 175, 287, doi: 10.1023/A:1004963425123 * Ritzwoller & Lavely (1991) Ritzwoller, M. H., & Lavely, E. M. 1991, ApJ, 369, 557, doi: 10.1086/169785 * Saha & Williams (1994) Saha, P., & Williams, T. B. 1994, AJ, 107, 1295, doi: 10.1086/116942 * Sambridge & Mosegaard (2002) Sambridge, M., & Mosegaard, K. 2002, Reviews of Geophysics, 40, 1009, doi: 10.1029/2000RG000089 * Schad & Roth (2020) Schad, A., & Roth, M. 2020, ApJ, 890, 32, doi: 10.3847/1538-4357/ab65ec * Schad et al. (2011) Schad, A., Timmer, J., & Roth, M. 2011, ApJ, 734, 97, doi: 10.1088/0004-637X/734/2/97 * Schad et al. (2013) —. 2013, ApJ, 778, L38, doi: 10.1088/2041-8205/778/2/L38 * Schou & Brown (1994) Schou, J., & Brown, T. M. 1994, A&AS, 107, 541 * Schou et al. (1998) Schou, J., Antia, H. M., Basu, S., et al. 1998, ApJ, 505, 390, doi: 10.1086/306146 * Schou et al. (2012) Schou, J., Scherrer, P. H., Bush, R. I., et al. 2012, Sol. Phys., 275, 229, doi: 10.1007/s11207-011-9842-2 * Sharma (2017) Sharma, S. 2017, ARA&A, 55, 213, doi: 10.1146/annurev-astro-082214-122339 * Thompson et al. (1996) Thompson, M. J., Toomre, J., Anderson, E. R., et al. 1996, Science, 272, 1300, doi: 10.1126/science.272.5266.1300 * Ulrich et al. (1988) Ulrich, R. K., Boyden, J. E., Webster, L., et al. 1988, Sol. Phys., 117, 291, doi: 10.1007/BF00147250 * Vorontsov (2007) Vorontsov, S. V. 2007, MNRAS, 378, 1499, doi: 10.1111/j.1365-2966.2007.11894.x * Vorontsov (2011) Vorontsov, S. V. 2011, Monthly Notices of the Royal Astronomical Society, 418, 1146, doi: 10.1111/j.1365-2966.2011.19564.x * Woodard et al. (2013) Woodard, M., Schou, J., Birch, A. C., & Larson, T. P. 2013, Sol. Phys., 287, 129, doi: 10.1007/s11207-012-0075-9 * Woodard (1989) Woodard, M. F. 1989, ApJ, 347, 1176, doi: 10.1086/168206 * Woodard (2000) —. 2000, Sol. Phys., 197, 11, doi: 10.1023/A:1026508211960 * Zhao & Kosovichev (2004) Zhao, J., & Kosovichev, A. G. 2004, ApJ, 603, 776, doi: 10.1086/381489
# Spin Hall effect of light under arbitrarily polarized and unpolarized light Minkyung Kim Department of Mechanical Engineering, Pohang University of Science and Technology (POSTECH), Pohang 37673, Republic of Korea Dasol Lee Department of Mechanical Engineering, Pohang University of Science and Technology (POSTECH), Pohang 37673, Republic of Korea Junsuk Rho Department of Mechanical Engineering, Pohang University of Science and Technology (POSTECH), Pohang 37673, Republic of Korea Department of Chemical Engineering, Pohang University of Science and Technology (POSTECH), Pohang 37673, Republic of Korea<EMAIL_ADDRESS> ###### Abstract The spin Hall effect of light (SHEL), which refers to a spin-dependent and transverse splitting at refraction and reflection phenomena, inherently depends on the polarization states of the incidence. Most of the previous research have focused on a horizontally or vertically polarized incidence, in which the analytic formula of the shift is well-formulated and SHEL appears symmetrically in both shift and intensity. However, the SHEL under an arbitrarily polarized or unpolarized incidence has remained largely unexplored. Whereas the SHEL under other polarization is sensitive to incident polarization and is asymmetrical, here we demonstrate that the SHEL is independent of the incident polarization and is symmetrical in shift if Fresnel coefficients of the two linear polarization are the same. The independence of the shift with respect to the incident polarization is proved both analytically and numerically. Moreover, we prove that under an unpolarized incidence composed of a large number of completely random polarization states, the reflected beam is split in half into two circularly polarized components that undergo the same amount of splitting but in opposite directions. This result means that under unpolarized incidence, the SHEL occurs exactly the same as under horizontally or vertically polarized incidence. We believe that the incident-polarization-independent SHEL can broaden the applicability of the SHEL to cover optical systems in which polarization is ill-defined. ## 1 Introduction Spin-orbit coupling is a universal phenomenon that can be observed in a variety of fields in physics including classical mechanics 1, quantum physics 2, and photonics 3. In particular, there exist diverse kinds of spin-orbit interactions in photonics because of rich physics enabled by the two electric and magnetic vector fields 4. Among them, one interesting spin-orbit related feature is the spin Hall effect of light (SHEL) 5, 6, 7, also known as an Imbert-Fedorov shift 8, 9, which reveals itself as a transverse and spin- dependent splitting of a finitely-thick beam at an optical interface. A physical mechanism that underpins the spin-dependent shift is the transverse nature of light, $\mathbf{k}\cdot\mathbf{E}=0$, that makes the incidence contain a bundle of slightly differently defined polarization bases 10. Despite its long history tracing back to the mid-19th century, the SHEL has regained a booming interest recently especially in photonics and metamaterials communities. A variety of nanophotonic devices and metamaterials have been proposed to enlarge the spin-dependent shift 11, 12, 13, 14, 15, 16, 17, 18, to increase the efficiency 19, and to exploit the SHEL to identify geometric parameters 20, 21, electric and magnetic properties 22, 23, and chemical reactions 24, 25 with high precision. Except for the studies of asymmetric SHEL, in which the magnitude of the shift is spin-dependent 26, 27, 28, 29, 30, most previous studies have focused only on horizontally or vertically polarized incidence. By symmetry, a horizontally or vertically polarized light is split at the optical interface into equal amplitudes of left circularly polarized (LCP) and right circularly polarized (RCP) beams, which shift by the same magnitude but with opposite signs 6. However, under an arbitrary incident polarization, neither the magnitude of the shift nor the intensity of the two circularly polarized components is symmetrical in general; in other words, the arbitrarily polarized incidence is split into LCP and RCP unevenly in both shift and intensity. Furthermore, the shift is sensitive to the polarization states of the incidence, resulting in a different amount of splitting as the incident polarization varies. Until very recently, a well-defined polarization state has been regarded as a prerequisite of spin-orbit related phenomena. However, it has been reported recently that even unpolarized light can induce a transverse spin 31. Here, we demonstrate that if reflection coefficients for $s$ and $p$ polarizations are equal to each other, the shift is degenerate for any arbitrarily polarized incidence and that the SHEL appears symmetrically in both shift and intensity under unpolarized incidence, as it does under a horizontally or vertically polarized incidence. First, we prove both analytically and numerically that the SHEL is independent of the polarization state of the incidence when the two linear polarizations have the same reflection coefficients. In such a case, LCP and RCP components of the reflected beam are shifted evenly, by the same amount but in opposite directions. The intensities of the LCP and RCP components are generally asymmetrical under an arbitrarily given incident polarization, but they also become symmetrical when the incidence is unpolarized, i.e., when the incidence is a superposition state of a vast number of completely random polarizations. We show that when the reflection coefficients are degenerate, the SHEL appears symmetrically in both shift and intensity, thereby having the whole symmetries of the SHEL under horizontally or vertically polarized incidence even under unpolarized incidence. In stark contrast to the previous studies of the SHEL where the incidence has a well- defined single polarization state, our work will extend the field of the SHEL to include unpolarized sources. ## 2 Results and Discussion ### 2.1 The analytical proof of incident-polarization-independent shift This section is devoted to proving the incident-polarization-independent shift theoretically by using a wave packet model. To do so, we examine how an incidence characterized by a single arbitrary polarization transforms at an interface through the reflection. An incident Gaussian beam propagating in the $x_{I}$-$z_{I}$ plane along the $z_{I}$-axis can be expressed in momentum space as $\bm{\psi}_{I}=\begin{pmatrix}\psi_{I}^{H}\\\ \psi_{I}^{V}\end{pmatrix}\psi_{0},$ (1) where the superscripts $H$ and $V$ correspond to horizontal and vertical polarization respectively, $\begin{pmatrix}\psi^{H}_{I}&\psi^{V}_{I}\end{pmatrix}^{T}$ is the Jones vector of the incidence, and $\psi_{0}$ is a Gaussian term given as $\psi_{0}=\frac{\omega_{0}}{\sqrt{2\pi}}\exp(-\frac{\omega_{0}^{2}}{4}(k_{x}^{2}+k_{y}^{2})),$ (2) where $\omega_{0}$ is a beam waist, $k_{0}$ is the incident wave vector, and $k_{x}\equiv k_{Ix}=-k_{Rx}$ and $k_{y}\equiv k_{Iy}=k_{Ry}$ are $x$\- and $y$-components of the wave vector. The Jones vectors of the incident and reflected beams represented in a linear basis are related to each other 32 $\begin{pmatrix}\psi^{H}_{R}\\\ \psi^{V}_{R}\end{pmatrix}=\begin{pmatrix}r_{p}&\frac{k_{y}}{k_{0}}(r_{p}+r_{s})\cot{\theta_{i}}\\\ -\frac{k_{y}}{k_{0}}(r_{p}+r_{s})\cot{\theta_{i}}&r_{s}\end{pmatrix}\begin{pmatrix}\psi^{H}_{I}\\\ \psi^{V}_{I}\end{pmatrix},$ (3) where the subscripts $I$ and $R$ correspond to the incident and reflected beam respectively, $r_{s}$ and $r_{p}$ are Fresnel reflection coefficients for $s$ and $p$ polarization, and $\theta_{i}$ is an incident angle. The reflected beam is composed of the Jones vector and the Gaussian part similarly to Eq. 1, $\bm{\psi}_{R}=\begin{pmatrix}\psi_{R}^{H}\\\ \psi_{R}^{V}\end{pmatrix}\psi_{0}.$ (4) The reflected beam can be represented in a circular basis by using a basis transformation $\begin{pmatrix}\psi^{+}_{R}\\\ \psi^{-}_{R}\end{pmatrix}=\frac{1}{\sqrt{2}}\begin{pmatrix}1&i\\\ 1&-i\end{pmatrix}\begin{pmatrix}\psi^{H}_{R}\\\ \psi^{V}_{R}\end{pmatrix},$ (5) where the superscripts $+$ and $-$ denote LCP and RCP respectively. Since the wave vector component along the transverse axis is much smaller than the wave number ($k_{y}/k_{0}\ll 1$), the first-order Taylor expansion of $1\pm x\approx\exp(\pm x)$ can be applied. Then, Eq. 3 can be expressed as $\begin{pmatrix}\psi^{+}_{R}\\\ \psi^{-}_{R}\end{pmatrix}=\frac{1}{\sqrt{2}}\begin{pmatrix}r_{p}\exp(+ik_{y}\triangle_{H})&ir_{s}\exp(+ik_{y}\triangle_{V})\\\ r_{p}\exp(-ik_{y}\triangle_{H})&-ir_{s}\exp(-ik_{y}\triangle_{V})\end{pmatrix}\begin{pmatrix}\psi^{H}_{I}\\\ \psi^{V}_{I}\end{pmatrix},$ (6) where $\displaystyle\triangle_{H}$ $\displaystyle=\frac{\cot{\theta_{i}}}{k_{0}}(1+\frac{r_{s}}{r_{p}}),$ $\displaystyle\triangle_{V}$ $\displaystyle=\frac{\cot{\theta_{i}}}{k_{0}}(1+\frac{r_{p}}{r_{s}}).$ (7) Then the shift can be obtained by using a position operator $\mathbf{r}=i\hbar\partial_{\mathbf{p}}=i\partial_{\mathbf{k}}$ to calculate an average value of $y$ position of the reflected beam as $\delta^{\pm}=\text{Re}\frac{\expectationvalue{i\partial_{k_{y}}}{\psi_{R}^{\pm}\psi_{0}}}{\bra{\psi_{R}^{\pm}\psi_{0}}\ket{\psi_{R}^{\pm}\psi_{0}}}.$ (8) If the polarization state of the incidence is either horizontal ($\psi_{I}^{H}=1,\psi_{I}^{V}=0$) or vertical ($\psi_{I}^{H}=0,\psi_{I}^{V}=1$), then $\psi_{R}^{\pm}$ are eigenstates of the position operator $i\partial_{k_{y}}$ with eigenvalues of $\mp\triangle_{H}$ and $\mp\triangle_{V}$ respectively. On the other hand, $i\partial_{k_{y}}\psi_{0}=-\psi_{0}k_{y}\omega_{0}^{2}/2$, and this odd function has no contribution in the indefinite integral. Therefore, the shift can be obtained by taking the eigenvalues of $\psi_{R}^{\pm}$, so Eq. 8 gives the well-known formulas 33, 6 $\displaystyle\delta_{H}^{\pm}$ $\displaystyle=\mp\frac{\cot{\theta_{i}}}{k_{0}}\text{Re}(1+\frac{r_{s}}{r_{p}}),$ $\displaystyle\delta_{V}^{\pm}$ $\displaystyle=\mp\frac{\cot{\theta_{i}}}{k_{0}}\text{Re}(1+\frac{r_{p}}{r_{s}}).$ (9) Under an incidence that is polarized neither horizontally nor vertically, the shift should be calculated by substituting $\psi_{R}^{\pm}$ (Eq. 6) into Eq. 8 and by performing integration in momentum space. Interestingly, when the Fresnel coefficients of the two linear polarizations are equal ($r_{s}=r_{p}\equiv r$), the two equations in Eq. 7 are degenerate ($\triangle_{H}=\triangle_{V}\equiv\triangle$), then $\psi_{R}^{\pm}$ become eigenstates of the operator $i\partial_{k_{y}}$ with eigenvalues of $\mp\triangle$ for any incident polarization. Consequently, similarly to the instance under a horizontally or vertically polarized incidence, Eq. 8 reduces to $\delta^{\pm}=\mp\text{Re}(\triangle)$, which is equivalent to Eq. 9 but under an arbitrarily polarized incidence. This formula of the shift contains neither $\psi^{H}_{I}$ nor $\psi^{V}_{I}$ but only depends on $\theta_{i}$. The independence of $\delta^{\pm}$ with respect to the incident polarization clearly shows that an arbitrarily polarized incidence is shifted by the same distance, regardless of the polarization states of the incidence. Another important attribute of the SHEL that originates from $r_{s}=r_{p}$ is that the splitting occurs symmetrically in shift ($\lvert\delta^{+}\rvert=\lvert\delta^{-}\rvert$) under an arbitrarily polarized incidence. This result is opposed to the conventional wisdom that a circularly or elliptically polarized incidence is split asymmetrically into LCP and RCP 26, 27, 28, 29, 30. Although the condition of $r_{s}=r_{p}$ seems to imply the polarization- independent reflection, it is not true. Instead, $r_{s}=r_{p}$ is associated with the $\pi$ phase shift between the electric fields of $s$\- and $p$-polarized reflected beam because of the sign convention 4. Polarization- independent reflection ($r_{s}=-r_{p}$) is another possible solution of $\triangle_{H}=\triangle_{V}$, but is excluded because it leads to zero shift. Here we consider only the reflection type of the SHEL for demonstration but this scheme is also applicable to the transmission type by matching $t_{s}=-t_{p}$. ### 2.2 The intensities of circularly polarized components of the reflected beam under arbitrarily polarized incidence To understand how the SHEL occurs systematically, not only the shifts but also the intensities of the LCP and RCP components of the reflected beam should be considered. For a given incident polarization, the intensities can be calculated as $I_{R}^{\pm}=\frac{\bra{\psi_{R}^{\pm}\psi_{0}}\ket{\psi_{R}^{\pm}\psi_{0}}}{\bra{\psi_{I}^{H}\psi_{0}}\ket{\psi_{I}^{H}\psi_{0}}+\bra{\psi_{I}^{V}\psi_{0}}\ket{\psi_{I}^{V}\psi_{0}}}.$ (10) Considering that both $\psi_{I}^{H}$ and $\psi_{I}^{V}$ are complex, the intensities of the two circularly polarized beams are generally not symmetrical ($I_{R}^{+}\neq I_{R}^{-}$). Especially, when $r_{s}=r_{p}$, Eq. 10 can be simplified to $I_{R}^{\pm}=\lvert r\rvert^{2}\Big{[}\frac{1}{2}\pm\lvert\psi_{I}^{H}\rvert\lvert\psi_{I}^{V}\rvert\sin\Big(\text{arg}(\psi_{I}^{H})-\text{arg}(\psi_{I}^{V})\Big{missing})\Big{]}.$ (11) This result shows that even when $r_{s}=r_{p}$ is satisfied, $I_{R}^{+}=I_{R}^{-}$ if the incidence is linearly polarized, and $I_{R}^{+}\neq I_{R}^{-}$ otherwise. It also conforms to our intuition in that only linearly polarized light is a superposition of equal amounts of LCP and RCP whereas the others such as elliptically and circularly polarized light are not. ### 2.3 The SHEL under arbitrarily polarized incidence and unpolarized incidence Figure 1: Schematics of the SHEL under various polarization states of incidence in a general case ($r_{s}\neq r_{p}$). (a) The SHEL in real space and (b) corresponding intensity profiles of the reflected beam along the transverse axis when the incidence is horizontally or vertically polarized. The two circularly polarized reflected beams are symmetrical in shift ($\lvert\delta^{+}\rvert=\lvert\delta^{-}\rvert$) and intensity ($I_{R}^{+}=I_{R}^{-}$). (c) The SHEL in real space and (d) corresponding intensity profiles of the reflected beam when the incidence has an arbitrary polarization. Both symmetries are preserved under linear polarization but are broken ($\lvert\delta^{+}\rvert\neq\lvert\delta^{-}\rvert$ and $I_{R}^{+}\neq I_{R}^{-}$) under circularly or elliptically polarized incidence. For clear visualization, only the central wave vector component of the incident and reflected beams are plotted, instead of the wave packets with finite beam waist. The width of the central wave vectors of the reflected beams shown in Fig. 1a and c represent their intensities. Using Eq. 8 and Eq. 11, which provide the shift and intensities of the circularly polarized components of the reflected beam respectively, we now illustrate how the SHEL occurs under a given incident polarization. Firstly, a general case of $r_{s}\neq r_{p}$ is considered. Under a horizontally or vertically polarized incidence, as is well-known from many previous publications 6, 33, 34, the reflected beam is split in half into LCP and RCP ($I_{R}^{+}=I_{R}^{-}$), which undergoes the equal amount of the shift along the opposite direction ($\lvert\delta^{+}\rvert=\lvert\delta^{-}\rvert$) (Fig. 1a and b). This symmetrical splitting originates from the high symmetries of the incident polarization state and maintains under other linear polarization. In contrast, under a circularly or elliptically polarized incidence, the polarization breaks the symmetries and thus both the shift and intensity are asymmetrical ($\lvert\delta^{+}\rvert\neq\lvert\delta^{-}\rvert$ and $I_{R}^{+}\neq I_{R}^{-}$) (Fig. 1c and d). Figure 2: Schematics of the SHEL for various polarization states of incidence when $r_{s}=r_{p}$. (a) The SHEL in real space and (b) corresponding intensity profiles of the reflected beam along the transverse axis under arbitrarily polarized incidence. The shift is incident-polarization-independent and is symmetrical ($\lvert\delta^{+}\rvert=\lvert\delta^{-}\rvert$). Intensity is symmetrical ($I_{R}^{+}=I_{R}^{-}$) only under a linear polarization, and is asymmetrical ($I_{R}^{+}\neq I_{R}^{-}$) under a circularly or elliptically polarized incidence. (c) The SHEL in real space and (d) corresponding intensity profiles of the reflected beam under unpolarized incidence. Both the shift and intensity are symmetrical ($\lvert\delta^{+}\rvert=\lvert\delta^{-}\rvert$ and $I_{R}^{+}=I_{R}^{-}$). For clear visualization, only the central wave vector component of the incident and reflected beams are plotted, instead of the wave packets with finite beam waist. The width of the central wave vectors of the reflected beams shown in Fig. 2a and c represent their intensities. In contrast, if $r_{s}=r_{p}$, the shift becomes symmetrical ($\lvert\delta^{+}\rvert=\lvert\delta^{-}\rvert$) under arbitrarily polarized incidence (Fig. 2a and b) as proved in the previous section. More importantly, any incidence is split by the same amount regardless of its polarization state. However, despite the degeneracy in the shift, the intensities of the LCP and RCP components are symmetrical ($I_{R}^{+}=I_{R}^{-}$) only under a linear polarization but are polarization-dependent and are asymmetrical ($I_{R}^{+}\neq I_{R}^{-}$) in a general circularly or elliptically polarized incidence. The intensities can be made symmetrical by using unpolarized incidence. Here, the unpolarized state refers to a superposition of a vast number of randomly polarized light. Then the total incidence has no phase difference between the two components of the Jones vector due to the randomness and hence the sine term in Eq. 11 averages to zero. Therefore, the unpolarized incidence is split in half into LCP and RCP ($I_{R}^{+}=I_{R}^{-}$), the shifts of which have the same magnitude ($\lvert\delta^{+}\rvert=\lvert\delta^{-}\rvert$) just as a horizontally or vertically polarized incidence does (Fig. 2c and d). On the other hand, when $r_{s}$ and $r_{p}$ have nonzero phase difference, the SHEL has asymmetrical intensity ($I_{R}^{+}\neq I_{R}^{-}$) under unpolarized incidence. This phenomenon occurs because when $r_{s}\neq r_{p}$, one can still obtain the formula of the intensities of the LCP and RCP components of the reflected beam akin to Eq. 11, but the sine term of that formula includes not only the phase difference between $\psi_{I}^{H}$ and $\psi_{I}^{V}$ but also that between $r_{s}$ and $r_{p}$. It prevents $I_{R}^{\pm}$ from being degenerate, resulting in asymmetrical intensities under unpolarized incidence. ### 2.4 Numerical demonstration of the incident-polarization-independent SHEL To confirm the incident-polarization-independent SHEL numerically, a Monte Carlo simulation is performed by generating $N$ numbers of randomly oriented polarization states, then using each of them as an incidence. We consider a paraxial regime in which the polarization is defined in a two-dimensional sense. The $n$-th incidence ($n\in\\{1,2,...,N\\}$) has a well-defined polarization that corresponds to the Stokes parameters ($S_{1}^{(n)},S_{2}^{(n)},S_{3}^{(n)}$) distributed randomly on a Poincaré sphere (Fig. 3a, left), and have a degree of paraxial two-dimensional polarization equal to unity ($P_{2D}^{(n)}=\sqrt{\sum_{i=1}^{3}\big{(}S_{i}^{(n)}\big{)}^{2}}/S_{0}^{(n)}=1$). The averaged Stokes parameters ($\sum_{n=1}^{N}{S_{i}^{(n)}/N}$ for $i=1,2,3$) converge to zero as $N$ increases and are all less than $5\times 10^{-3}$ when $N=1000$ (Fig. 3a, right). The total incidence satisfies the condition of two- dimensional unpolarized light, that is 31, $(S_{0},S_{1},S_{2},S_{3})\propto(1,0,0,0)$ and $P_{2D}=0$. Figure 3: The SHEL under arbitrarily polarized incidence. (a) Poincaré sphere representing $N$ randomly distributed polarization states of incidence (left) and the averaged Stokes parameters (right). (b) Amplitude and (c) phase of $r_{s}$ and $r_{p}$ at an interface between an isotropic ($\varepsilon_{1}=2$) and anisotropic ($\varepsilon_{2x}=\varepsilon_{2z}=4.1,\varepsilon_{2y}=1$) media. Yellow dashed lines indicate $\theta_{0}$, an incident angle at which $r_{s}=r_{p}$ is satisfied. (d, e) Scatter plots of (d) $\delta^{+(n)}/\lambda$ and (e) $\delta^{-(n)}/\lambda$, where $\lambda$ is the wavelength, under $N$ independent incidences at the interface between the isotropic and anisotropic media. Solid curves represent $\delta_{H,V}^{\pm}/\lambda$ obtained by Eq. 9. (f) Standard deviation of the $N$ values of $\delta^{\pm(n)}/\lambda$ at each $\theta_{i}$. For simulation, $N=1000,w_{0}=10^{3}\lambda,z_{R}=10\lambda$ are used. To investigate the polarization dependency of the SHEL under arbitrarily polarized incidence, an interface between an isotropic ($\varepsilon_{1}$) and anisotropic ($\varepsilon_{2x}=\varepsilon_{2z}\neq\varepsilon_{2y}$) media is considered. Whereas $r_{s}$ and $r_{p}$ are coupled to each other by Fresnel equations 4 at an interface between two isotropic media, $s$\- and $p$-polarized light experience the anisotropic medium, in which the optic axis lies perpendicular to the incident plane, independently. When $\varepsilon_{2x}=\varepsilon_{2z}>\varepsilon_{1}>\varepsilon_{2y}$, light propagating from the isotropic medium to the anisotropic one has degenerate Fresnel coefficients for $s$ and $p$ polarization ($r_{s}=r_{p}$) at a certain angle. This attribute originates from the occurrence of total internal reflection under the $s$-polarized light and the resultant $\pi$ phase shift in the Fresnel coefficients. We use the material parameters given as $\varepsilon_{1}=2,\varepsilon_{2x}=\varepsilon_{2z}=4.1,\varepsilon_{2y}=1$. In such a case, $r_{s}=r_{p}$ is satisfied at a certain angle defined as $\theta_{0}$ (Fig. 3b and c). A feasible design of such a medium and details of the reflection at the boundary of an anisotropic medium can be found in Supporting Information. Because the analytic formula of the shift is known explicitly only for a horizontally or vertically polarized incidence, the shift under arbitrarily polarized incidence is obtained by taking the average $y$ position of the reflected beam profile numerically. Thus, we first obtain the spatial distribution of the reflected Gaussian beam. By combining Eq. 1-5 and then applying a Taylor expansion to reflection coefficients $r_{p,s}\approx r_{p,s}(\theta_{i})+\frac{k_{Ix}}{k_{0}}\frac{\partial r_{p,s}}{\partial\theta_{i}},$ (12) the field distribution of the reflected beam can be found in momentum space. Then the spatial distribution of the reflected beam can be obtained by applying a Fourier transform $\tilde{\psi}^{\pm}_{R}(x_{R},y_{R},z_{R})=\iint{dk_{Rx}dk_{Ry}\psi^{\pm}_{R}(k_{Rx},k_{Ry})\psi_{0}\exp\big(i(k_{Rx}x_{R}+k_{Ry}y_{R}+k_{Rz}z_{R})\big{missing})}.$ (13) Each circularly polarized component of a reflected beam under a horizontally and vertically polarized incidence has a spatial profile of $\displaystyle\tilde{\psi}^{\pm}_{R,H}=$ $\displaystyle\frac{1}{\sqrt{2\pi}\omega_{0}}\frac{z_{0}}{z_{0}+iz_{R}}\exp(-\frac{k_{0}}{2}\frac{x_{R}^{2}+y_{R}^{2}}{z_{0}+iz_{R}})$ $\displaystyle\times\Big{[}r_{p}-i\frac{x_{R}}{z_{0}+iz_{R}}\dot{r_{p}}\mp\frac{y_{R}\cot{\theta_{i}}}{z_{0}+iz_{R}}(r_{p}+r_{s})\mp i\frac{x_{R}y_{R}\cot{\theta_{i}}}{(z_{0}+iz_{R})^{2}}(\dot{r_{p}}+\dot{r_{s}})\Big{]}\exp(ik_{r}z_{R}),$ $\displaystyle\tilde{\psi}^{\pm}_{R,V}=$ $\displaystyle\frac{\pm i}{\sqrt{2\pi}\omega_{0}}\frac{z_{0}}{z_{0}+iz_{R}}\exp(-\frac{k_{0}}{2}\frac{x_{R}^{2}+y_{R}^{2}}{z_{0}+iz_{R}})$ $\displaystyle\times\Big{[}r_{s}-i\frac{x_{R}}{z_{0}+iz_{R}}\dot{r_{s}}\mp\frac{y_{R}\cot{\theta_{i}}}{z_{0}+iz_{R}}(r_{p}+r_{s})\mp i\frac{x_{R}y_{R}\cot{\theta_{i}}}{(z_{0}+iz_{R})^{2}}(\dot{r_{p}}+\dot{r_{s}})\Big{]}\exp(ik_{r}z_{R}),$ (14) where the second subscript corresponds to incident polarization, the dot notation indicates the first derivative with respect to $\theta_{i}$, and $z_{0}=k_{0}\omega_{0}^{2}/2$ is the Rayleigh length. These four field distributions correspond to the elements of a matrix that transforms the Jones vector of the $n$-th incident polarization to the $n$-th reflected beam profile, i.e., $\tilde{\psi}^{\pm(n)}_{R}=\tilde{\psi}^{\pm}_{R,H}\psi_{I}^{H(n)}+\tilde{\psi}^{\pm}_{R,V}\psi_{I}^{V(n)}.$ (15) Eq. 15 provides the spatial profiles of the two circularly polarized components of the reflected beam for each of the given incident polarization. We conduct $N$ independent calculations with parameters given as: $N=1000,w_{0}=10^{3}\lambda,z_{R}=10\lambda$ where $\lambda$ is the wavelength. In the simulation, $\lambda$ is set as 600 nm but has no influence on the results. The shifts that each of the given incidence undergoes are obtained by calculating beam centroids of the reflected beam as $\delta^{\pm(n)}=\frac{\expectationvalue{y_{R}}{\tilde{\psi}_{R}^{\pm(n)}}}{\bra{\tilde{\psi}_{R}^{\pm(n)}}\ket{\tilde{\psi}_{R}^{\pm(n)}}},$ (16) and are plotted as black dots in Fig. 3d and e, after normalized by $\lambda$. Because the shift depends on the incident polarization, the scatter plots exhibit many different $\delta^{\pm(n)}/\lambda$ at a fixed $\theta_{i}$. However, at $\theta_{0}$ at which $r_{s}=r_{p}$ is satisfied and thus $\delta_{H}^{\pm}=\delta_{V}^{\pm}$, the values of $\delta^{\pm(n)}/\lambda$ all appear at the intersection point regardless of the incident polarization (Fig. 3d and e, yellow dashed lines). The convergence of the shift for all incident polarization is also confirmed by the zero standard deviation of $\delta^{\pm(n)}/\lambda$ at $\theta_{0}$ (Fig. 3f). The standard deviation is nonzero at all $\theta_{i}$ other than $\theta_{0}$. The large standard deviation near the Brewster angle results from the diverging shift under horizontally polarized incidence. In several consecutive Monte Carlo simulations in which the $N$ polarization states are randomly redistributed, the fine details of the scatter plot in Fig. 3d and e change, but the overall tendency remains unaltered. Besides $\theta_{0}$, there exists another intersection of $\delta^{\pm}_{H,V}/\lambda$ near 60∘, but the shift is incident-polarization-dependent because $\triangle_{H}\neq\triangle_{V}$ at this angle and hence $\psi_{R}^{\pm}$ are not eigenstates of $i\partial_{k_{y}}$. Figure 4: Shift distribution represented in a Poincaré sphere when (a) $\theta_{i}=\theta_{0}$, (b) $\theta_{i}=20^{\circ}\neq\theta_{0}$, and (c) $\theta_{i}=40^{\circ}\neq\theta_{0}$. The location of a point on the sphere manifests the polarization states of the $n$-th incidence; the color represents $\delta^{\pm(n)}/\lambda$. Top and bottom corresponds to the shift of LCP and RCP components respectively. The boundary of the sphere is omitted for clarity. Fig. 3d and e show the distributions of the shift for a given $\theta_{i}$, but do not provide the correspondence between the polarization state and the shift. Thus, we examine the distribution of $\delta^{\pm(n)}/\lambda$ in Poincaré sphere for three different $\theta_{i}$ (Fig. 4). The location of a point on the sphere manifests the polarization states of the $n$-th incidence, and the color represents $\delta^{\pm(n)}/\lambda$. As confirmed by Fig. 3d-f, the shift under an arbitrarily polarized incidence are all degenerate and are also symmetrical for LCP and RCP components when $\theta_{i}=\theta_{0}$ (Fig. 4a). However, at other $\theta_{i}=20^{\circ}\neq\theta_{0}$, the shift depends on the incident polarization, showing variations of the shift over the sphere (Fig. 4b). This tendency is more noticeable under circular and elliptical polarization than under linear polarization. In contrast, at $\theta_{i}=40^{\circ}$, incidence that has Stokes parameters near $(1,0,0)$ produce shifts with large deviation (Fig. 4c). This trend occurs because of the diverging shift under the horizontally polarized light near the Brewster angle. Furthermore, the shift distribution at $\theta_{i}\neq\theta_{0}$ shows that the splitting is not symmetrical ($\lvert\delta^{+(n)}\rvert\neq\lvert\delta^{-(n)}\rvert$) under arbitrarily polarized incidence if $r_{s}\neq r_{p}$ (Fig. 4b and c). ### 2.5 Numerical demonstration of the SHEL under unpolarized incidence Figure 5: Field profiles of the reflected Gaussian beam. Spatial distributions of (a) $\lvert\tilde{\psi}_{R}^{+}\rvert^{2}$, (b) $\lvert\tilde{\psi}_{R}^{-}\rvert^{2}$, and (c) their difference $\lvert\tilde{\psi}_{R}^{-}\rvert^{2}-\lvert\tilde{\psi}_{R}^{+}\rvert^{2}$ at $\theta_{0}$. (d) Spatial distribution of $\lvert\tilde{\psi}_{R}^{-}\rvert^{2}-\lvert\tilde{\psi}_{R}^{+}\rvert^{2}$ at $\theta_{i}=50^{\circ}\neq\theta_{0}$. (e, f) Intensity profiles in an arbitrary unit (A.U.) of the two circularly polarized components of the reflected beam (blue: $\lvert\tilde{\psi}_{R}^{+}\rvert^{2}$, red: $\lvert\tilde{\psi}_{R}^{-}\rvert^{2}$) and their difference (yellow) along the transverse axis at $x_{R}=0$ at (e) $\theta_{0}$ and (f) $\theta_{i}=50^{\circ}\neq\theta_{0}$. For better visualization, $\lvert\tilde{\psi}_{R}^{-}\rvert^{2}-\lvert\tilde{\psi}_{R}^{+}\rvert^{2}$ are exaggerated in Fig. 5e and f. Because of the linearity of Eq. 15, the total reflected beam under unpolarized incidence can be obtained by taking a superposition of all $\tilde{\psi}^{\pm(n)}_{R}$ for a set of given incident polarizations: $\tilde{\psi}^{\pm}_{R}=\sum_{n=1}^{N}\tilde{\psi}^{\pm(n)}_{R}/N$. For completeness, the intensity distributions of the two circularly polarized components of the total reflected beam and their difference are presented in Fig. 5. The splitting of the LCP and RCP is not readily apparent (Fig. 5a and b), but their intensity difference, which is proportional to $S_{3}$, shows the symmetric and spin-dependent shift along the transverse direction at $\theta_{0}$ (Fig. 5c). This SHEL under unpolarized incidence can be understood as a result of the superposition of the spin-dependent splittings under $N$ differently polarized incidences, in which the magnitude of the splittings are all degenerate. This field profile is obtained under unpolarized incidence but resembles that under a horizontally or vertically polarized incidence 35, 13. In contrast, at $\theta_{i}=50^{\circ}\neq\theta_{0}$, this splitting exhibits significantly distinct features (Fig. 5d) for the following reasons. Firstly, the shift is incident-polarization-dependent; the $N$ differently polarized incidences are split by different amounts of displacement, then superposed. Secondly, the $\lvert\tilde{\psi}_{R}^{-}\rvert^{2}-\lvert\tilde{\psi}_{R}^{+}\rvert^{2}$ distribution of a single sign is attributed to the asymmetrical splitting of intensity ($I_{R}^{+}\neq I_{R}^{-}$) as a result of the phase difference between $r_{s}$ and $r_{p}$. If the shift is not large enough, the asymmetry in intensity leads to a single sign of $S_{3}$ distribution (Fig. 5d). For better understanding, the intensities along $x_{R}=0$ are shown in Fig. 5e and f. Whereas the unpolarized incidence is split symmetrically into two circularly polarized components at $\theta_{0}$ (Fig. 5e), the SHEL at $\theta_{i}\neq\theta_{0}$ is asymmetrical in both shift and intensity (Fig. 5f). ## 3 Conclusion The SHEL is known as a symmetrical splitting of a refracted or reflected beam into two circularly polarized beams along the transverse axis at an optical interface, but such symmetries only appear when the incidence is horizontally or vertically polarized. Here we show that the two circularly polarized components of the reflected beam undergo incident-polarization-independent and symmetrical splitting under an arbitrary incident polarization if $r_{s}=r_{p}$. The polarization independence of the shift is proved theoretically, then numerically using a Monte Carlo simulation. The intensities of the circularly polarized components of the reflected beams are symmetrical under unpolarized incidence, thereby appearing exactly the same as under horizontally or vertically polarized incidence. Lastly, an interface at which $r_{s}=r_{p}$ is suggested by adjoining isotropic and anisotropic dielectric media. In contrast to the previous research on the SHEL in which a well-defined polarization state is preliminary, the incident-polarization- independent SHEL can widen the boundaries and possible applications of the SHEL to cover unpolarized or ill-defined polarized sources. ## 4 Method ### 4.1 Monte Carlo simulation For the random polarization states, the elements of the Jones vector of the $n$-th incidence are defined as $\psi_{I}^{H,V(n)}=\lvert\psi_{I}^{H,V(n)}\rvert\exp(i\phi^{H,V(n)})$ where the magnitude $\lvert\psi_{I}^{H,V(n)}\rvert$ and the phase $\phi^{H,V(n)}$ are real scalars that are randomly chosen in $(0,1)$ and $(-\pi,\pi)$ respectively for $n\in\\{1,...,N\\}$. Then the Jones vector $\begin{pmatrix}\psi_{I}^{H(n)}&\psi_{I}^{V(n)}\end{pmatrix}^{T}$ is normalized by its magnitude $\sqrt{\lvert\psi_{I}^{H(n)}\rvert^{2}+\lvert\psi_{I}^{V(n)}\rvert^{2}}$. ### 4.2 Reflection coefficients at an interface between isotropic and anisotropic media At an interface between the isotropic (medium 1, permittivity $\varepsilon_{1}$) and anisotropic (medium 2, permittivity $\varepsilon_{2}=\text{diag}(\varepsilon_{2x},\varepsilon_{2y},\varepsilon_{2z})$) media, the Fresnel reflection coefficients can be obtained by solving Maxwell’s equations: $\displaystyle r_{s}=$ $\displaystyle\frac{\sqrt{\varepsilon_{1}-\beta^{2}}-\sqrt{\varepsilon_{2y}-\beta^{2}}}{\sqrt{\varepsilon_{1}+\beta^{2}}+\sqrt{\varepsilon_{2y}-\beta^{2}}},$ $\displaystyle r_{p}=$ $\displaystyle\frac{\sqrt{\varepsilon_{1}-\beta^{2}}/\varepsilon_{1}-\sqrt{\varepsilon_{2x}-\beta^{2}\varepsilon_{2x}/\varepsilon_{2z}}/\varepsilon_{2x}}{\sqrt{\varepsilon_{1}-\beta^{2}}/\varepsilon_{1}+\sqrt{\varepsilon_{2x}-\beta^{2}\varepsilon_{2x}/\varepsilon_{2z}}/\varepsilon_{2x}}$ (17) where $\beta=\sqrt{\varepsilon_{1}}\sin\theta_{i}$ is the propagation constant. ## References * 1 A. Persson. How Do We Understand the Coriolis Force? _Bull. Am. Meteorol. Soc._ 79(7), 1373–1386 (1998). * 2 L. H. Thomas. The motion of the spinning electron. _Nature_ 117(2945), 514–514 (1926). * 3 K. Y. Bliokh, F. J. Rodríguez-Fortuño, F. Nori and A. V. Zayats. Spin–orbit interactions of light. _Nat. Photonics_ 9(12), 796–808 (2015). * 4 M. Born and E. Wolf. Principles of optics: electromagnetic theory of propagation, interference and diffraction of light. Cambridge University Press (2013). * 5 M. Onoda, S. Murakami and N. Nagaosa. Hall effect of light. _Phys. Rev. Lett._ 93(8), 083901 (2004). * 6 O. Hosten and P. Kwiat. Observation of the spin Hall effect of light via weak measurements. _Science_ 319(5864), 787–790 (2008). * 7 X. Ling, X. Zhou, K. Huang, Y. Liu, C.-W. Qiu, H. Luo and S. Wen. Recent advances in the spin Hall effect of light. _Rep. Prog. Phys._ 80(6), 066401 (2017). * 8 F. I. Fedorov. To the theory of total reflection. _J. Opt._ 15(1), 014002–1 (2013). * 9 C. Imbert. Calculation and Experimental Proof of the Transverse Shift Induced by Total Internal Reflection of a Circularly Polarized Light Beam. _Phys. Rev. D_ 5, 787–796 (1972). * 10 K. Y. Bliokh and A. Aiello. Goos–Hänchen and Imbert–Fedorov beam shifts: an overview. _Journal of Optics_ 15(1), 014001 (2013). * 11 X. Yin, Z. Ye, J. Rho, Y. Wang and X. Zhang. Photonic spin Hall effect at metasurfaces. _Science_ 339(6126), 1405–1407 (2013). * 12 O. Takayama, J. Sukham, R. Malureanu, A. V. Lavrinenko and G. Puentes. Photonic spin Hall effect in hyperbolic metamaterials at visible wavelengths. _Opt. Lett._ 43(19), 4602–4605 (2018). * 13 M. Kim, D. Lee, T. H. Kim, Y. Yang, H. J. Park and J. Rho. Observation of enhanced optical spin Hall effect in a vertical hyperbolic metamaterial. _ACS Photonics_ 6(10), 2530–2536 (2019). * 14 M. Kim, D. Lee, B. Ko and J. Rho. Diffraction-induced enhancement of optical spin Hall effect in a dielectric grating. _APL Photonics_ 5(6), 066106 (2020). * 15 W. Zhu and W. She. Enhanced spin Hall effect of transmitted light through a thin epsilon-near-zero slab. _Opt. Lett._ 40(13), 2961–2964 (2015). * 16 X. Jiang, J. Tang, Z. Li, Y. Liao, L. Jiang, X. Dai and Y. Xiang. Enhancement of photonic spin Hall effect via bound states in the continuum. _J. Phys. D: Appl. Phys._ 52(4), 045401 (2018). * 17 H. Dai, L. Yuan, C. Yin, Z. Cao and X. Chen. Direct Visualizing the Spin Hall Effect of Light via Ultrahigh-Order Modes. _Phys. Rev. Lett._ 124, 053902 (2020). * 18 X. Zhou and X. Ling. Enhanced Photonic Spin Hall Effect Due to Surface Plasmon Resonance. _IEEE Photon. J._ 8(1), 1–8 (2016). * 19 M. Kim, D. Lee, H. Cho, B. Min and J. Rho. Spin Hall Effect of Light with Near-Unity Efficiency in the Microwave. _Laser Photonics Rev._ n/a(n/a), 2000393. * 20 B. Wang, K. Rong, E. Maguid, V. Kleiner and E. Hasman. Probing nanoscale fluctuation of ferromagnetic meta-atoms with a stochastic photonic spin Hall effect. _Nat. Nanotechnol._ 15, 450–456 (2020). * 21 X. Zhou, X. Ling, H. Luo and S. Wen. Identifying graphene layers via spin Hall effect of light. _Appl. Phys. Lett._ 101(25), 251602 (2012). * 22 S. Chen, X. Ling, W. Shu, H. Luo and S. Wen. Precision Measurement of the Optical Conductivity of Atomically Thin Crystals via the Photonic Spin Hall Effect. _Phys. Rev. Appl._ 13, 014057 (2020). * 23 T. Li, Q. Wang, A. Taallah, S. Zhang, T. Yu and Z. Zhang. Measurement of the magnetic properties of thin films based on the spin Hall effect of light. _Opt. Express_ 28(20), 29086–29097 (2020). * 24 R. Wang, J. Zhou, K. Zeng, S. Chen, X. Ling, W. Shu, H. Luo and S. Wen. Ultrasensitive and real-time detection of chemical reaction rate based on the photonic spin Hall effect. _APL Photonics_ 5(1), 016105 (2020). * 25 J. Liu, K. Zeng, W. Xu, S. Chen, H. Luo and S. Wen. Ultrasensitive detection of ion concentration based on photonic spin Hall effect. _Appl. Phys. Lett._ 115(25), 251102 (2019). * 26 X. Zhou and X. Ling. Unveiling the photonic spin Hall effect with asymmetric spin-dependent splitting. _Opt. Express_ 24(3), 3025–3036 (2016). * 27 W. Zhu, J. Yu, H. Guan, H. Lu, J. Tang, J. Zhang, Y. Luo and Z. Chen. The upper limit of the in-plane spin splitting of Gaussian beam reflected from a glass-air interface. _Sci. Rep._ 7(1), 1–9 (2017). * 28 M. Jiang, H. Lin, L. Zhuo, W. Zhu, H. Guan, J. Yu, H. Lu, J. Tan and Z. Chen. Chirality induced asymmetric spin splitting of light beams reflected from an air-chiral interface. _Opt. Express_ 26(6), 6593–6601 (2018). * 29 X. Zhou, L. Xie, X. Ling, S. Cheng, Z. Zhang, H. Luo and H. Sun. Large in-plane asymmetric spin angular shifts of a light beam near the critical angle. _Opt. Lett._ 44(2), 207–210 (2019). * 30 Y. Wu, L. Sheng, L. Xie, S. Li, P. Nie, Y. Chen, X. Zhou and X. Ling. Actively manipulating asymmetric photonic spin Hall effect with graphene. _Carbon_ 166, 396 – 404 (2020). * 31 J. Eismann, L. Nicholls, D. Roth, M. Alonso, P. Banzer, F. Rodríguez-Fortuño, A. Zayats, F. Nori and K. Bliokh. Transverse spinning of unpolarized light. _Nat. Photonics_ (2020). * 32 H. Luo, X. Zhou, W. Shu, S. Wen and D. Fan. Enhanced and switchable spin Hall effect of light near the Brewster angle on reflection. _Phys. Rev. A_ 84, 043806 (2011). * 33 Y. Qin, Y. Li, H. He and Q. Gong. Measurement of spin Hall effect of reflected light. _Opt. Lett._ 34(17), 2551–2553 (2009). * 34 X. Zhou, X. Ling, Z. Zhang, H. Luo and S. Wen. Observation of spin Hall effect in photon tunneling via weak measurements. _Sci. Rep._ 4, 7388 (2014). * 35 K. Y. Bliokh, C. T. Samlan, C. Prajapati, G. Puentes, N. K. Viswanathan and F. Nori. Spin-Hall effect and circular birefringence of a uniaxial crystal plate. _Optica_ 3(10), 1039–1047 (2016).
# Strongly coupled Yukawa plasma layer in a harmonic trap Hong Pan Department of Physics, Boston College, Chestnut Hill, Massachusetts, 02467, USA Gabor J. Kalman Department of Physics, Boston College, Chestnut Hill, Massachusetts, 02467, USA Peter Hartmann Institute for Solid State Physics and Optics, Wigner Research Centre for Physics, P.O.Box. 49, H-1525 Budapest, Hungary ###### Abstract Observations made in dusty plasma experiments suggest that an ensemble of electrically charged solid particles, confined in an elongated trap, develops structural inhomogeneities. With narrowing the trap the particles tend to form layers oriented parallel with the trap walls. In this work we present theoretical and numerical results on the structure of three-dimensional many- particle systems with screened Coulomb (Yukawa) inter-particle interaction in the strongly coupled liquid phase, confined in one-dimensional harmonic trap, forming quasi-2D configurations. Particle density profiles are calculated by means of the hypernetted chain approximation (HNC), showing clear signs of layer formation. The mechanism behind the formation of layer structure is discussed and a method to predict the number of layers is presented. Molecular dynamics (MD) simulations provide validation of the theoretical results and detailed microscopic insights. ###### pacs: 52.27.Gr, 52.27.Lw Strongly coupled charged particle systems (e.g. plasmas), where the inter- particle interaction energy exceeds the thermal kinetic energy of the constituent particles, can often be approximated by the one component plasma (OCP) model [1]. In cases when the dynamics of one dominant particle species decouples from that of the other components, it might be sufficient to provide a detailed description only for the former, while approximating the contribution of rest with a continuous neutralizing background. In the case of warm dense matter, for instance, the ions can follow classical particle trajectories, while the electrons experience quantum degeneracy and realize a homogeneous background [2, 3]. In the case of dusty plasmas it is the very different time scales and charge states that decouple the dust dynamics (with characteristic times in the order of 10 ms and $10^{4}$ elementary charges per dust) from the microscopic interactions with the electrons and ions present in the gas discharge (with typical times in the ns to $\mu$s range and unit charges). In both cases, only one of the plasma components (the ions in warm dens matter, and the dust particles in dusty plasmas) needs to be traced explicitly, the contribution of the background is reflected in the particular shape of the inter-particle interaction. If the polarizability of the isotropic neutralizing background is taken into account, the system can be approximated by the Yukawa one component plasma (YOCP) model. In this case the electrostatic pairwise interaction between the particles becomes screened, which screening can be approximated by the Dedye-Hückel mechanism resulting in an exponential decay superimposed to the bare Coulomb potential. Since the pioneering experiments published in 1994 [4, 5, 6] laboratory dusty plasmas have been extensively used to gain insight into the microscopic details of macroscopic and collective phenomena. In large area radio frequency (RF) discharges, using micrometer sized monodisperse spherical dust particles it is easy to form a single layer of highly charged dust grains. This ensemble can form large (consisting of tens of thousands of dust grains) structures with the particles ordered in triangular lattices. Simply by changing discharge conditions (e.g. the RF power) or by other forms of external energy coupling (e.g. laser heating [7] or DC pulsing [8]) the system can undergo a solid to liquid transition and stabilize in the strongly coupled liquid state. However, the apparent single layer structure does not mean that the system is strictly two-dimensional. The vertical confinement is defined by the interplay of gravity and the electric field in the RF sheath, ultimately forming a potential well experienced by each dust particle. Both the vertical equilibrium position and the effective trap frequency depend on the charge-to- mass ($q/m$) ratio of the dust grains and the discharge conditions. At finite temperature the dust particles experience small amplitude vertical oscillations, forming quasi-2D configurations. Early experimental investigations on quasi-2D system have been published by Lin I [9, 10] observing anomalous diffusion, as well as layering and slow dynamics. Theoretical work on quasi-2D systems include studies of in-plane and out-of-plane polarized wave propagation and structural transitions into multi- layered configurations [11, 12, 13, 14, 15] building heavily on numerical simulations. An analytical approach targeting the particle density distribution in the trap would provide deeper insight into the double or triple layer formation, as observed in the experiments. A similar problem has been studied by Wrighton [16, 17, 18] for spherical confinement comparing mean field and hypernetted chain calculations to numerical results. In Klumov’s work [19], crystallization of quasi-2D system under both harmonic and hard wall confinement was studied, in which the wetting phenomenon was observed in the hard wall case. Further theoretical studies on radially, as well as linearly confined liquids [20, 21, 22, 23, 24, 25] include the prediction of the radial density distribution, mechanisms of oscillatory density profiles and ensuing solvation forces, in both the weak and strong coupling regimes. In this work, we study the evolution of the density profile of quasi-2D Yukawa systems in a one-dimensional harmonic trap in the strongly coupled liquid state. The system is infinite in the $x$ and $y$ directions, where the particles can move freely. A harmonic trapping potential is applied in the $z$ direction. Let $n_{\rm s}$ denote the surface density of the quasi-2D layer projected onto the $x$-$y$ plane. In this case we define $a$ as the Wigner- Seitz radius in the projected plane as $a^{2}=1/(\pi n_{\rm s})$. The Yukawa interaction potential energy and the harmonic trap potential are, respectively $\varphi(r)=q^{2}\frac{e^{-\kappa\,r/a}}{r},~{}~{}~{}~{}~{}{\rm and}~{}~{}~{}V(z)=\frac{m\omega_{\rm t}^{2}z^{2}}{2},$ (1) where $r$ is the three-dimensional inter-particle distance, $\kappa$ is the dimensionless Yukawa screening parameter, $\omega_{\rm t}$ is the trap frequency, $q$ and $m$ are the electric charge and mass, equal for all particles. The strength of the electrostatic coupling can be characterized with the Coulomb coupling parameter, nominally expressing the ratio of the potential to kinetic energies per particle, and is defined as $\Gamma=\beta q^{2}/a,$ (2) where the thermodynamic $\beta=1/(k_{\rm B}T)$. For Yukawa systems an effective coupling $\Gamma_{\rm eff}(\kappa)<\Gamma$ [12, 26] can be introduced, that depends on the Yukawa screening parameter a does more accurately characterize the state of a particular system. The strongly coupled domain is defined by $\Gamma\gg 1$. Using the nominal 2D plasma frequency $\omega_{\rm p}^{2}=2\pi q^{2}n_{s}/(ma)$ to define the frequency and time units we introduce the dimensionless parameter $t$ characterizing the strength of the trapping potential, as $t^{2}=\frac{\omega_{\rm t}^{2}}{\omega_{\rm p}^{2}}=\frac{m\omega_{\rm t}^{2}a}{2\pi q^{2}n_{\rm s}}=\frac{m\omega_{\rm t}^{2}a^{3}}{2q^{2}}.$ (3) Within the frame of the quasi-2D YOCP model the behavior of a system is fully determined by the three dimensionless parameters $\Gamma$, $\kappa$ and $t$. Both the equilibrium properties and the dynamics of the system have been studied by molecular dynamics (MD) simulation and by theoretical analysis. Here we report on the equilibrium studies. Our MD simulations trance the trajectories of 10 000 particles in an external trapping potential defined by $V(z)$ and periodic boundary conditions in $x$ and $y$ in a cubic simulation domain. Initial positions are assigned to the particles based on a simple initial barometric estimate. Inter-particle forces are summed for all particles within a radius of $R\approx 44a$ for each particle. Before conducting any measurements the system is given enough time (approximately of 4 000 plasma oscillation cycles) to reach equilibrium during the initial thermalization phase using the velocity back-scaling thermostat. This is verified by observing the temperature stability after the thermaization is turned off. During the measurements data is collected and averaged over approx. 2 000 plasma oscillation cycles (100 000 time-steps). In order to obtain a theoretical description of the density distribution within the trap we invoke the density functional theory (DFT) [27, 28]. By minimizing the grand potential, a general, but quite complex expression for $n(r)$, the 3D number density has been reported by Evans [28] (eq. 26 in chapter 3), see also the Appendix of [16] in terms of $c\left(r,r^{\prime};[n(r)]\right)$, the direct correlation function (DCF) of the system. The notation emphasizes that the DCF is a unique, albeit unknown functional of the density profile. $c(r,r^{\prime})$ is also connected with $h(r,r^{\prime})$, the pair correlation function through the Ornstein-Zernike equation (OZ). $h(r,r^{\prime})=c(r,r^{\prime})+\int c(r,r^{\prime\prime})n(r^{\prime\prime})h(r^{\prime\prime},r^{\prime})~{}{\rm d}r^{\prime\prime}$ (4) The latter is related to the pair distribution function through $g(r,r^{\prime})=h(r,r^{\prime})+1$. The density profile is determined by the self-consistency relation [27] $n(r)\propto\exp\left[-\beta V(r)+\int n(r^{\prime})c_{0}(r-r^{\prime};n)~{}{\rm d}r^{\prime}\right],$ (5) where $c_{0}$ is the DCF of a reference system with uniform density. The structure of eq. (5) tells us that $c(r,r^{\prime})$ plays the role of the effective interaction potential in the system (note that $\varphi(r)$ does not appear explicitly in eq. (5)), $-\beta\varphi_{\rm eff}(r-r^{\prime})=c_{0}(r-r^{\prime})$ (6) In order to solve $c$ and $h$ simultaneously, the OZ relation provides the first equation, the second relation is derived by applying the hypernetted chain (HNC) approximation [29, 27]. HNC has been successfully used in various problems relating to strongly coupled Coulomb and Yukawa systems [30]. The HNC approximation is based on the neglect of the so-called “bridge” (or irreducible, i.e. not derivable from the combined operations of “parallel connection” and “series connection” of Mayer diagrams) diagrams: their contribution is assumed to be negligible in the case of long range potentials. As a result, one obtains the general basic relationship $g(r,r^{\prime})=\exp\left[-\beta\varphi(r-r^{\prime})+h(r,r^{\prime})-c(r,r^{\prime})\right],$ (7) which in combination with the OZ equation (4), provides a solution for $h(r,r^{\prime})$ and $c(r,r^{\prime})$. Restricting the solutions to homogeneous and isotropic functions, DCF satisfies the simpler variant of the OZ equation $h(r)=c(r)+{\bar{n}}\int c(|r-r^{\prime}|)h(r^{\prime})~{}{\rm d}r^{\prime}.$ (8) In equation (7), when $r$ is large enough, $g(r)\rightarrow 1$, $h(r)\rightarrow 0$, the expression reduces to $-\beta\varphi_{(}r-r^{\prime})=c(r-r^{\prime})$. Using this asymptotic formula for the whole range of $r$, one arrives to the mean field (MF) approximation. Another self-consistent approach to calculating DCF for homogeneous fluids was described in [31]. Since the system is uniform in the $x$-$y$ plane, the density is non-uniform only along the $z$ direction, the parametrization can be simplified to $n(r)=n(z)$, with the normalization condition $n_{s}=\int n(z)~{}{\rm d}z$. The integral in equation (5) can be split into$z$ part and radial part separately. We can write eq. (5) as $n(z)\propto\exp\left[-\beta V(z)+2\pi\iint n(z^{\prime})c_{0}(z,z^{\prime},\rho^{\prime};n)\rho^{\prime}~{}{\rm d}\rho^{\prime}~{}{\rm d}z^{\prime}\right].$ (9) We can rewrite eq. (9) in terms of the dimensionless quantities $\tilde{n}=na^{2}$, $\tilde{z}=z/a$, $\tilde{U}(z)=\frac{\beta}{\Gamma}U(z)$, etc. as $\displaystyle\tilde{n}(\tilde{z})$ $\displaystyle=$ $\displaystyle\tilde{n_{s}}\frac{\exp[-\Gamma\tilde{U}(\tilde{z})]}{\int_{-\infty}^{\infty}\exp[-\Gamma\tilde{U}(\tilde{z})]~{}{\rm d}\tilde{z}}$ (10) $\displaystyle\tilde{U}(\tilde{z})$ $\displaystyle=$ $\displaystyle t^{2}\tilde{z}^{2}-\tilde{W}(\tilde{z})$ $\displaystyle\tilde{W}(\tilde{z})$ $\displaystyle=$ $\displaystyle\frac{2\pi}{\Gamma}\iint\tilde{n}(\tilde{z}^{\prime})c_{0}(\tilde{z},\tilde{z}^{\prime},\tilde{\rho}^{\prime};n)\tilde{\rho}^{\prime}~{}{\rm d}\tilde{\rho}^{\prime}~{}{\rm d}\tilde{z}^{\prime}$ $\rho$ being the 2D distance between two projected particle positions in the $x$-$y$ plane. Then, provided that $c_{0}(r-r^{\prime};n)$ is known one can obtain the density profile by the iterative solution of eq. (10). In view of eq. (6) the simplest approximation for $c_{0}(z-z^{\prime},\rho)$ is to ignore correlations and set $\varphi_{\rm eff}(r,r^{\prime})=\varphi(r-r^{\prime})$, i.e. $c_{0}(r-r^{\prime})=-\beta\varphi(r-r^{\prime})$ (11) This is tantamount to a mean field approximation. Substituting (11) into (10), (10) reduces to $\displaystyle U(z)=t^{2}z^{2}+2\iint n(z^{\prime})\frac{\exp[-\kappa d(z,z^{\prime},\rho^{\prime})]}{d(z,z^{\prime},\rho^{\prime})}\rho^{\prime}~{}{\rm d}\rho^{\prime}~{}{\rm d}z^{\prime}$ $\displaystyle d(z,z^{\prime},\rho^{\prime})=\sqrt{(z-z^{\prime})^{2}+\rho^{\prime 2}}.$ (12) In the sequel we drop the symbol for the dimensionless quantities. It should be kept in mind that all the length variables are in the unit of the Wigner- Seitz radius $a$. In the following, for simplicity, we only discuss the results for $\kappa=0.4$. The results of the MF calculation and their comparison with the results of the MD simulation are given in Fig. 1. If there is no particle-particle interaction, then $\Gamma\rightarrow 0$, the profile is Gaussian $n(z)\propto\exp[-\Gamma t^{2}z^{2}]$, mapping the Maxwell distribution of non-interacting particles. Proceeding to higher $\Gamma$ values, we expect the MF method to gradually fail to reproduce the numerical data as it is only valid for weak coupling (low $\Gamma$, i.e. low density or high temperature). Inspecting the MF profiles in Fig. 1, covering moderate and strong $\Gamma$ values, we observe that those reasonably match the MD results at low $\Gamma$ values, while fail spectacularly at high $\Gamma$-s. In particular MF does not provide the non-monotonic behavior related to layer formation. In fact, it has been proved analytically that the MF density profiles have to be monotonic on both sides of the maximum [32], and thus MF is unable to predict the formation of multiple layers. Figure 1: Density profile comparison between the MD simulation and the MF approximation (continuous line) for different $\Gamma$ and t values. (a) $\Gamma=8$, $t=0.2$, (b) $\Gamma=16$, $t=0.1$, (c) $\Gamma=64$, $t=0.2$, (d) $\Gamma=64$, $t=0.4$ Next we calculate the DCF from the HNC approximation, then derive the density profile. The $\bar{n}$ in eq. (8) is chosen as the average number density of the plasma between the two layer boundaries. The boundary is defined arbitrarily as the points where the density value is one percent of the maximum density in the layer. This, of course requires the knowledge of the density, whose determination is the purpose of the calculation. All this lends itself to an iteration scheme. The resulting protocol for the numerical calculation is portrayed in Fig. 2. The algorithm is based on work published in [33]. Figure 2: Numerical iteration loop for calculating density profile. Figure 3: Density profile comparison between MD simulation (continuous line) and the HNC approximation, for different trapping strengths. $\Gamma=32$ Figure 4: Density profile comparison between MD simulation (continuous line) and the HNC approximation, for different trapping strengths. $\Gamma=64$ The major improvement of the HNC over the MF calculation is that it correctly reproduces the splitting of the system into multiple layers. Figs. 3 and 4 show comparisons with MD data at moderately high coupling values for cases when multi-peak profiles form. The similar multi-peak profile was reported in previous works [34, 20, 23, 24, 35, 36, 37]. Generally, both $t$ and $\Gamma$ can affects the density profile formation, but the acting mechanisms are different. From Figs. 3 and 4 we can see that when $t$ decreases, the layer becomes wider, developing more peaks (layers) in the density profile. The coupling parameter $\Gamma$ only affect the density modulation amplitude. The relation between the trapping strength and the density profile was also studied in [38, 39]. MD results in Figs. 5 and 6 show density profiles for a set of $\Gamma$ values. With increasing $\Gamma$, the system develops very sharp density peaks with deep minima between the peaks, a sign of the formation of crystal-like ordering (no exchange of particle between layers). As $\Gamma$ decreases, the particles experience larger vertical oscillation amplitudes, and the sides of neighboring peaks in the density profile fuse. The position and the number of the density peaks is mostly independent of the coupling, the distribution in the liquid state resembles that of the solid with lower amplitude of the density modulation. | 3 | 3 | 3 | 4 | 4 | 5 | 5 | 6 | 6 | 7 | peak number ---|---|---|---|---|---|---|---|---|---|---|--- 0.34 | 0.3 | 0.22 | 0.2 | 0.14 | 0.12 | 0.09 | 0.08 | 0.07 | 0.06 | t N | $\sqrt{N\pi}$ | 3.5 | 4.2 | 6 | 7 | 10 | 12 | 15 | 17.5 | 19 | 22 | layer width 2 | 2.5066 | 3.5 | 4.2 | 6 | 7 | 10 | 12 | 15 | 17.5 | 19 | 22 | $\frac{\mbox{layer width}}{N-1}$ 3 | 3.0700 | 1.75 | 2.1 | 3 | 3.5 | 5 | 6 | 7.5 | 8.75 | 9.5 | 11 4 | 3.5449 | 1.1667 | 1.4 | 2 | 2.3333 | 3.3333 | 4 | 5 | 5.8333 | 6.3333 | 7.3333 5 | 3.9633 | 0.875 | 1.05 | 1.5 | 1.75 | 2.5 | 3 | 3.75 | 4.375 | 4.75 | 5.5 6 | 4.3416 | 0.7 | 0.84 | 1.2 | 1.4 | 2 | 2.4 | 3 | 3.5 | 3.8 | 4.4 7 | 4.6895 | 0.5833 | 0.7 | 1 | 1.1667 | 1.6667 | 2 | 2.5 | 2.9167 | 3.1667 | 3.6667 8 | 5.0133 | 0.5 | 0.6 | 0.8571 | 1 | 1.4286 | 1.7143 | 2.1429 | 2.5 | 2.7143 | 3.1429 Table 1: Number of layers with different trapping strength, $\Gamma=512$, $\kappa=0.4$, the row with green color corresponds to the predicted $N$ value. Based on the idea that a multiple layers may be regarded as a slab extracted from a 3D lattice [40, 41, 42], one may determine the number of layers by a simple algorithm. Assuming there are $N$ layers, for simplicity a planar square lattice is associated to each layer. In this case the unit square side length is $x_{N}=\sqrt{\pi N}$. The inter-layer distance is $d_{N}=\frac{w}{N-1}$, where $w$ is the distance between the two outermost peaks. Comparing $x_{N}$ and $d_{N}$, the lowest value of $N$ at which $x_{N}>d_{N}$ is the prediction for the number of layers (see Table 1). In [40] the total energy of the system with different number of layers, including the particle interaction and the trapping energy is calculated. Despite operating with different physical quantities, the principles of the two methods are equivalent. In [43], the shell structure of the density profile in a confined colloidal particle system was studied by Monte Carlo simulation. Figure 5: Evolution of the density profiles with increasing $\Gamma$ values as determined by MD simulations. $t=0.1$ Figure 6: Evolution of the density profiles with increasing $\Gamma$ values as determined by MD simulations. $t=0.2$ Figure 7: HNC (lines) and MD (symbols) density profile comparison at high $\Gamma$. $\Gamma=300$. (a) $t=0.05$, (b) $t=0.07$. A similar configuration was studied in [22] applying both the MF and HNC formalism, however, as those results are restricted to the weak coupling regime no layer formation was reported. In that work the correlation energy, based on local density approximation (LDA), and with this the excess chemical potential was calculated. In anther study [44] the density profile of a polyelectrolyte system in a harmonic trap was calculated. A similar shell structure formation was observed. In [45] a similar DFT method was applied on hard-core Yukawa dusty plasma in a spherically harmonic trap. In [46] a study on van der Waals fluids with different confining potentials was presented including the modulation of the density profiles. In our calculations, relying on the HNC approximation, the stability of the solution scheme appears to be limited to $\Gamma\leq 300$ (for $\kappa=0.4$), at higher couplings no converged solution could be found. For this reason comparing the HNC and MD result at the strongest coupling is performed at $\Gamma=300$. Fig. 7 shows that even though the HNC result underestimates the density modulation amplitude, it predicts correctly the number and position of the peaks. In conclusion, in this paper, we have analyzed the density profile of quasi-2D Yukawa plasma in harmonic trap in the strongly coupled liquid regime. We utilize the HNC approximation to calculate the density profile, which matches MD simulation result very well at low to moderate couplings. At stronger coupling, near the solidification the HNC calculations underestimate the density modulation amplitudes. We have presented a method to predict the number of layers in the solid phase based on the comparison between the in- plane inter-particle distance and the inter-layer distance. We confirm the validity of this prediction in the liquid state as well. An outlook into future research directions include: (1) theoretical models beyond the standard HNC approximation could be applied to derive the direct correlation function $c(r)$; (2) a method for the more accurate incorporation of correlation effect could be worked out to extend applicability to higher $\Gamma$-s, where the system is in an intermediate state between liquid and solid phases; (3) application of the current model to anharmonic confinement potentials, like hard-wall traps. ###### Acknowledgements. The authors thank Jeff Wrighton’s help on numerical calculation in this paper. The authors are grateful for financial support from the Hungarian Office for Research, Development, and Innovation NKFIH grants K-134462 and K-132158. ## References * Baus and Hansen [1980] M. Baus and J.-P. Hansen, Statistical mechanics of simple coulomb systems, Physics Reports 59, 1 (1980). * Clérouin _et al._ [2016] J. Clérouin, P. Arnault, C. Ticknor, J. D. Kress, and L. A. Collins, Unified concept of effective one component plasma for hot dense plasmas, Phys. Rev. Lett. 116, 115003 (2016). * Wang _et al._ [2020] Z.-Q. Wang, J. Tang, Y. Hou, Q.-F. Chen, X.-R. Chen, J.-Y. Dai, X.-J. Meng, Y.-J. Gu, L. Liu, G.-J. Li, Y.-S. Lan, and Z.-G. Li, Benchmarking the effective one-component plasma model for warm dense neon and krypton within quantum molecular dynamics simulation, Phys. Rev. E 101, 023302 (2020). * Chu and I [1994] J. H. Chu and L. I, Direct observation of coulomb crystals and liquids in strongly coupled rf dusty plasmas, Phys. Rev. Lett. 72, 4009 (1994). * Thomas _et al._ [1994] H. Thomas, G. E. Morfill, V. Demmel, J. Goree, B. Feuerbacher, and D. Möhlmann, Plasma crystal: Coulomb crystallization in a dusty plasma, Phys. Rev. Lett. 73, 652 (1994). * Melzer _et al._ [1994] A. Melzer, T. Trottenberg, and A. Piel, Experimental determination of the charge on dust particles forming coulomb lattices, Physics Letters A 191, 301 (1994). * Nosenko _et al._ [2006] V. Nosenko, J. Goree, and A. Piel, Laser method of heating monolayer dusty plasmas, Physics of Plasmas 13, 032106 (2006). * Donkó _et al._ [2017] Z. Donkó, P. Hartmann, P. Magyar, G. J. Kalman, and K. I. Golden, Higher order structure in a complex plasma, Physics of Plasmas 24, 103701 (2017). * Juan and I [1998] W. Juan and L. I, Anomalous diffusion in strongly coupled quasi-2d dusty plasmas, Phys. Rev. Lett. 80, 3073 (1998). * Teng _et al._ [2003] L. Teng, P. Tu, and L. I, Microscopic observation of confinement-induced layering and slow dynamics of dusty-plasma liquids in narrow channels, Phys. Rev. Lett. 90, 245004 (2003). * Bystrenko [2003] O. Bystrenko, Structural transitions in one-dimensionally confined one-component plasmas, Phys. Rev. E 67, 025401 (2003). * Hartmann _et al._ [2005] P. Hartmann, G. J. Kalman, Z. Donkó, and K. Kutasi, Equilibrium properties and phase diagram of two-dimensional yukawa systems, Phys. Rev. E 72, 026409 (2005). * Qiao and Hyde [2005] K. Qiao and T. W. Hyde, Structural phase transitions and out-of-plane dust lattice instabilities in vertically confined plasma crystals, Phys. Rev. E 71, 026406 (2005). * Henning _et al._ [2007] C. Henning, P. Ludwig, A. Filinov, A. Piel, and M. Bonitz, Ground state of a confined yukawa plasma including correlation effects, Phys. Rev. E 76, 036404 (2007). * Donkó _et al._ [2009] Z. Donkó, P. Hartmann, and G. J. Kalman, Two-dimensional dusty plasma crystals and liquids, Journal of Physics: Conference Series 162, 012016 (2009). * Wrighton _et al._ [2009] J. Wrighton, J. W. Dufty, H. Kählert, and M. Bonitz, Theoretical description of coulomb balls: Fluid phase, Phys. Rev. E 80, 066405 (2009). * Bruhn _et al._ [2011] H. Bruhn, H. Kählert, T. Ott, M. Bonitz, J. Wrighton, and J. W. Dufty, Theoretical description of spherically confined, strongly correlated yukawa plasmas, Phys. Rev. E 84, 046407 (2011). * Wrighton _et al._ [2012] J. Wrighton, H. Kählert, T. Ott, P. Ludwig, H. Thomsen, J. Dufty, and M. Bonitz, Charge correlations in a harmonic trap, Contributions to Plasma Physics 52, 45 (2012). * Klumov and Morfill [2008] B. A. Klumov and G. Morfill, Effect of confinement on the crystallization of a dusty plasma in narrow channels, JETP Letters 87, 409 (2008). * Vanderlick _et al._ [1989] T. K. Vanderlick, L. E. Scriven, and H. T. Davis, Molecular theories of confined fluids, The Journal of Chemical Physics 90, 2422 (1989). * Yu [2009] Y.-X. Yu, A novel weighted density functional theory for adsorption, fluid-solid interfacial tension, and disjoining properties of simple liquid films on planar solid surfaces, The Journal of Chemical Physics 131, 024704 (2009). * Girotto _et al._ [2014] M. Girotto, A. P. dos Santos, T. Colla, and Y. Levin, Yukawa particles in a confining potential, The Journal of Chemical Physics 141, 014106 (2014). * Nygård [2016] K. Nygård, Local structure and density fluctuations in confined fluids, Current Opinion in Colloid & Interface Science 22, 30 (2016). * Mansoori and Rice [2014] G. A. Mansoori and S. A. Rice, Confined fluids: Structure, properties and phase behavior, in _Advances in Chemical Physics_ (John Wiley & Sons, Ltd, 2014) Chap. 5, pp. 197–294. * Henning _et al._ [2006] C. Henning, H. Baumgartner, A. Piel, P. Ludwig, V. Golubnichiy, M. Bonitz, and D. Block, Ground state of a confined yukawa plasma, Phys. Rev. E 74, 056403 (2006). * Ott _et al._ [2014] T. Ott, M. Bonitz, L. G. Stanton, and M. S. Murillo, Coupling strength in coulomb and yukawa one-component plasmas, Physics of Plasmas 21, 113704 (2014). * Han [2013] Theory of simple liquids, in _Theory of Simple Liquids (Fourth Edition)_, edited by J. P. Hansen and I. R. McDonald (Academic Press, Oxford, 2013) fourth edition ed. * Eva [1992] Fundamentals of inhomogenous fluids, in _Fundamentals of Inhomogenous Fluids_ , edited by D. Henderson (Marcel Dekker, NY, 1992). * Rowlinson [1965] J. S. Rowlinson, The equation of state of dense systems, Reports on Progress in Physics 28, 169 (1965). * Ng [1974] K. Ng, Hypernetted chain solutions for the classical one‐component plasma up to $\Gamma=7000$, The Journal of Chemical Physics 61, 2680 (1974). * Rickayzen _et al._ [1994] G. Rickayzen, P. Kalpaxis, and E. Chacon, A self‐consistent approach to a density functional for homogeneous fluids, The Journal of Chemical Physics 101, 7963 (1994). * Pan [2019] H. Pan, _Static and dynamic properties of strongly coupled quasi-2D Yukawa plasma layers_ , Ph.D. thesis, Morrissey College of Arts and Sciences, Boston College (2019). * Springer _et al._ [1973] J. F. Springer, M. A. Pokrant, and F. A. Stevens, Integral equation solutions for the classical electron gas, The Journal of Chemical Physics 58, 4863 (1973). * Snook and van Megen [1980] I. K. Snook and W. van Megen, Solvation forces in simple dense fluids. i, The Journal of Chemical Physics 72, 2907 (1980). * Ingebrigtsen and Dyre [2014] T. S. Ingebrigtsen and J. C. Dyre, The impact range for smooth wall–liquid interactions in nanoconfined liquids, Soft Matter 10, 4324 (2014). * Ebner _et al._ [1980] C. Ebner, M. A. Lee, and W. F. Saam, Nonuniform one-dimensional classical fluids: Theory versus monte carlo experiments, Phys. Rev. A 21, 959 (1980). * Kierlik and Rosinberg [1991] E. Kierlik and M. L. Rosinberg, Density-functional theory for inhomogeneous fluids: Adsorption of binary mixtures, Phys. Rev. A 44, 5025 (1991). * Ott _et al._ [2008] T. Ott, M. Bonitz, Z. Donkó, and P. Hartmann, Superdiffusion in quasi-two-dimensional yukawa liquids, Phys. Rev. E 78, 026409 (2008). * Almarza _et al._ [2009] N. G. Almarza, C. Martín, and E. Lomba, Phase behavior of the hard-sphere maier-saupe fluid under spatial confinement, Phys. Rev. E 80, 031501 (2009). * Totsuji _et al._ [1997] H. Totsuji, T. Kishimoto, and C. Totsuji, Structure of confined yukawa system (dusty plasma), Phys. Rev. Lett. 78, 3113 (1997). * Totsuji _et al._ [1996] H. Totsuji, T. Kishimoto, Y. Inoue, C. Totsuji, and S. Nara, Yukawa system (dusty plasma) in one-dimensional external fields, Physics Letters A 221, 215 (1996). * Oguz _et al._ [2009] E. C. Oguz, R. Messina, and H. Löwen, Multilayered crystals of macroions under slit confinement, J. Phys.: Condens. Matter 21, 424110 (2009). * Ricci _et al._ [2007] A. Ricci, P. Nielaba, S. Sengupta, and K. Binder, Ordering of two-dimensional crystals confined in strips of finite width, Phys. Rev. E 75, 011405 (2007). * Dutta and Jho [2016] S. Dutta and Y. S. Jho, Shell formation in short like-charged polyelectrolytes in a harmonic trap, Phys. Rev. E 93, 012503 (2016). * Gu _et al._ [2012] F. Gu, H.-J. Wang, and J.-T. Li, Density functional theory for the ground state of spherically confined dusty plasma, Phys. Rev. E 85, 056402 (2012). * Tang and Wu [2004] Y. Tang and J. Wu, Modeling inhomogeneous van der waals fluids using an analytical direct correlation function, Phys. Rev. E 70, 011201 (2004).
# Increasing Cluster Size Asymptotics for Nested Error Regression Models Ziyang Lyu and A.H. Welsh Mathematical Sciences Institute and Research School of Finance Actuarial Studies and Statistics Australian National University (today) ###### Abstract This paper establishes asymptotic results for the maximum likelihood and restricted maximum likelihood (REML) estimators of the parameters in the nested error regression model for clustered data when both of the number of independent clusters and the cluster sizes (the number of observations in each cluster) go to infinity. Under very mild conditions, the estimators are shown to be asymptotically normal with an elegantly structured covariance matrix. There are no restrictions on the rate at which the cluster size tends to infinity but it turns out that we need to treat within cluster parameters (i.e. coefficients of unit-level covariates that vary within clusters and the within cluster variance) differently from between cluster parameters (i.e. coefficients of cluster-level covariates that are constant within clusters and the between cluster variance) because they require different normalisations and are asymptotically independent. Key words: asymptotic independence; maximum likelihood estimator; mixed model; REML estimator; variance components. ## 1 Introduction Regression models with nested errors (also called random intercept or homogeneous correlation models) are widely used in applied statistics to model relationships in clustered data; they were introduced for survey data, by Scott and Holt (1982) and Battese et al. (1988), and, for longitudinal data, by Laird and Ware (1982). The models are usually fitted (see Harville (1977)) by assuming normality and computing maximum likelihood or restricted maximum likelihood (REML) estimators. As these estimators are nonlinear, asymptotic results provide an important way to understand their properties and then to construct approximate inferences about the unknown parameters. The usual asymptotic results applied to these estimators from Hartley and Rao (1967), Anderson (1969), Miller (1977), Das (1979), Cressie and Lahiri (1993), and Richardson and Welsh (1994) increase the number of clusters while keeping the size of each cluster fixed or bounded. However, there are many applications, particularly with survey data, with large cluster sizes; for example, Arora and Lahiri (1997) give an example with $43$ clusters and cluster sizes ranging from $95$ to $633$ and such examples are common in analysing poverty data (Pratesi, 2016). In addition, there are theoretical problems (e.g. in prediction, see Jiang (1998)) for which both the number of clusters and the cluster sizes need to increase. Therefore, in this paper, we study the asymptotic properties of normal-theory maximum likelihood and REML estimators of the parameters in the nested error regression model as both the number of clusters and the cluster sizes tend to infinity. Suppose that we observe on the $j$th unit in the $i$th cluster the vector $[y_{ij},\mathbf{x}_{ij}^{T}]^{T}$, where $y_{ij}$ is a scalar response variable and $\mathbf{x}_{ij}$ is a vector of explanatory variables or covariates, $j=1,\ldots,m_{i}$, $i=1,\ldots,g$. The nested error regression model specifies that $y_{ij}=\beta_{0}+\mathbf{x}_{ij}^{T}\boldsymbol{\beta}_{s}+\alpha_{i}+e_{ij},\qquad j=1,\ldots,m_{i},\,i=1,\ldots,g,$ (1) where $\beta_{0}$ is the intercept, $\boldsymbol{\beta}_{s}$ is the slope parameter, $\alpha_{i}$ is a random effect representing a random cluster effect and $e_{ij}$ is an error term. We assume that the $\\{\alpha_{i}\\}$ and $\\{e_{ij}\\}$ are all mutually independent with mean zero and variances (called the variance components) $\sigma_{\alpha}^{2}$ and $\sigma_{e}^{2}$, respectively; we do not assume normality. This regression model treats clusters as independent with constant (i.e. homogeneous) correlation within clusters. It is a particular, simple linear mixed model that is widely used in fields such as small area estimation (see Rao and Molina (2015)) to model and make predictions from clustered data, so our results are immediately useful. In addition, its simplicity allows us to use elementary methods to gain insight into exactly what is going on and obtain explicit, highly interpretable results as the cluster sizes increase. These arguments and results form the basis for how to proceed to more complicated cases, with multiple variance components. When the random effects and errors are normally distributed, the likelihood for the parameters and the REML criterion can be obtained analytically. Irrespective of whether normality holds or not, we refer to these functions as the likelihood and the REML criterion for the model (1) and the values of the parameters that maximise them as maximum likelihood and REML estimators, respectively. For our results, we make very simple assumptions: essentially finite “$4+\delta$” moments for the random effects and errors (instead of normality) and, allowing the explanatory variables to be fixed or random, conditions analogous to finite “$2+\delta$” moments for the explanatory variables. We allow $g\to\infty$ and $\min_{1\leq i\leq g}m_{i}\to\infty$ without any restriction on the rates. We obtain asymptotic representations for both the maximum likelihood and REML estimators that give the influence functions of these estimators, are very useful for deriving results when we combine these estimators with other estimators, and lead to central limit theorems for these estimators and asymptotic inferences for the unknown parameters. The normalisation is by a diagonal matrix which is easy to interpret. These results provide new and striking insights. First, we need to separate and treat within cluster parameters (i.e. coefficients of unit-level covariates that vary within clusters and the within cluster variance $\sigma_{e}^{2}$) differently from between cluster parameters (i.e. coefficients of cluster-level covariates that are constant within clusters and the between cluster variance $\sigma_{\alpha}^{2}$). We make explicit the fact that the information for within cluster parameters grows with $n=\sum_{i=1}^{g}m_{i}$ and the information for between cluster parameters grows with $g$ so they require different normalisations. The asymptotic variance matrix which we obtain explicitly has a very tidy and easy to interpret block diagonal structure. Second, there are good reasons for centering the within cluster covariates about their cluster means and then including the cluster means as contextual effect variables in the between cluster covariates (see for example Yoon and Welsh (2020) for references) but our asymptotic results (which include both cases) show that increasing cluster size has asymptotically the same effect as the centering (although without increasing the number of between cluster parameters) and also asymptotically orthogonalises the variance components. These apparently simple insights are new and not available from the existing literature. The few results in the literature that allow both the number of clusters and the cluster size to go to infinity do not give the same insights as our results. Jiang (1996) proved consistency and asymptotic normality of the maximum likelihood and REML estimators for a wide class of linear mixed models allowing increasing cluster sizes. He later showed this condition is required for studying the empirical distribution of the empirical best linear unbiased predictors (EBLUPs) of the random effects (Jiang, 1998). Xie and Yang (2003) obtained results for generalized estimating equation regression parameter estimators with increasing cluster size which potentially relate to our estimators, but their estimators do not include the variance components so the results do not apply to our estimators. The difficulties with trying to apply general results to particular models like (1) are that it can be difficult to understand the conditions and interpret the main result. To illustrate, increasing cluster size in Jiang (1996) is a part of other complicated assumptions and, for particular examples, he needed further conditions on the way the cluster size increases, making it difficult to see whether there is any restriction on the relationship between the cluster size and the number of clusters and leaving open questions of whether the conditions are minimal or not. Also, although Jiang did give some nested model examples which satisfy his main invariant class $AI^{4}$ condition, this condition is quite complicated. In terms of their main results, both Jiang (1996) and Xie and Yang (2003) normalise the estimators by the product of general (nondiagonal) matrices, producing results which are difficult to interpret and do not provide the insights our results provide. We introduce notation to describe the maximum likelihood and REML estimators for the parameters in (1)-(2), specify the conditions and state our main results in Section 2. We discuss the results in Section 3 and give the proofs in Section 4. ## 2 Results We gain important insights by partitioning the vector of covariates $\mathbf{x}_{ij}$ into the $p_{w}$-vector $\mathbf{x}_{ij}^{(w)}$ of within cluster covariates and the $p_{b}$-vector $\mathbf{x}_{i}^{(b)}$ of between cluster covariates. As noted in the Introduction, it is also often useful to center the within cluster covariates about their cluster means and then expand the between cluster covariate vector to include the cluster means of the within cluster covariates. Specifically, for a single within cluster covariate $x_{ij}$, we can make the regression function either $\beta_{0}+x_{ij}\beta_{2}$ or the centered form $\beta_{0}+\bar{x}_{i}\beta_{1}+(x_{ij}-\bar{x}_{i})\beta_{2}$. This centering ensures that $\sum_{j=1}^{m_{i}}\mathbf{x}_{ij}^{(w)}=\boldsymbol{0}_{[p_{w}:1]}$ for all $i=1,\ldots,g$, where $\boldsymbol{0}_{[p:q]}$ denotes the $p\times q$ matrix of zeros, and as it orthogonalises the between and within covariates, has advantages for interpreting and fitting the model (Yoon and Welsh, 2020) as well as increasing flexibility. We leave this as choice for the modeller; our analysis handles both cases as well as the cases in which there are no within cluster or no between cluster covariates because they are all special cases of the model (1) which we re-express as $y_{ij}=\beta_{0}+\mathbf{x}_{i}^{(b)T}\boldsymbol{\beta}_{1}+\mathbf{x}_{ij}^{(w)T}\boldsymbol{\beta}_{2}+\alpha_{i}+e_{ij},\qquad j=1,\ldots,m_{i},\,i=1,\ldots,g,$ (2) where $\beta_{0}$ is the unknown intercept, $\boldsymbol{\beta}_{1}$ is the unknown between cluster slope parameter and $\boldsymbol{\beta}_{2}$ is the unknown within cluster slope parameter. We treat the covariates as fixed; when they are random, we condition on them, though we omit this from the notation. We assume throughout that the true model that describes the data generating mechanism is (2) with general parameter $\boldsymbol{\omega}=[\boldsymbol{\beta}_{0},\boldsymbol{\beta}_{1}^{T},\sigma_{\alpha}^{2},\boldsymbol{\beta}_{2}^{T},\sigma_{e}^{2}]^{T}$, true parameter $\dot{\boldsymbol{\omega}}=[\dot{\beta}_{0},\dot{\boldsymbol{\beta}}_{1}^{T},\dot{\sigma}_{\alpha}^{2},\dot{\boldsymbol{\beta}}_{2}^{T},\dot{\sigma}_{e}^{2}]^{T}$ and take all expectations under the true model. The order of the parameters in $\boldsymbol{\omega}$ and $\dot{\boldsymbol{\omega}}$ groups the between parameters and the within parameters together and simplifies the presentation of our results. To simplify notation, let $\tau_{i}=m_{i}/(\sigma_{e}^{2}+m_{i}\sigma_{a}^{2})$ with true value $\dot{\tau}_{i}$, $m_{L}=\min_{1\leq i\leq g}m_{i}$, $\begin{split}&\bar{y}_{i}=\frac{1}{{m_{i}}}\sum_{j=1}^{m_{i}}y_{ij},\quad\bar{\mathbf{x}}_{i}^{(w)}=\frac{1}{m_{i}}\sum_{j=1}^{m_{i}}\mathbf{x}_{ij}^{(w)},\quad S_{w}^{y}=\sum_{i=1}^{g}\sum_{j=1}^{m_{i}}(y_{ij}-\bar{y}_{i})^{2},\\\ &\mathbf{S}_{w}^{xy}=\sum_{i=1}^{g}\sum_{j=1}^{m_{i}}(\mathbf{x}_{ij}^{(w)}-\bar{\mathbf{x}}_{i}^{(w)})(y_{ij}-\bar{y}_{i}),\quad\text{and}\quad\\\ &\mathbf{S}_{w}^{x}=\sum_{i=1}^{g}\sum_{j=1}^{m_{i}}(\mathbf{x}_{ij}^{(w)}-\bar{\mathbf{x}}_{i}^{(w)})(\mathbf{x}_{ij}^{(w)}-\bar{\mathbf{x}}_{i}^{(w)})^{T}.\end{split}$ The log-likelihood for the parameters in the model (after discarding constant terms) is $\begin{split}l(\boldsymbol{\omega})&=\frac{1}{2}\sum_{i=1}^{g}\log(\tau_{i})-\frac{n-g}{2}\log\sigma_{e}^{2}-\frac{1}{2\sigma_{e}^{2}}(S_{w}^{y}-2\mathbf{S}_{w}^{xyT}\boldsymbol{\beta}_{2}+\boldsymbol{\beta}_{2}^{T}\mathbf{S}_{w}^{x}\boldsymbol{\beta}_{2})\\\ &\qquad-\frac{1}{2}\sum_{i=1}^{g}\tau_{i}(\bar{y}_{i}-\beta_{0}-\mathbf{x}_{i}^{(b)T}\boldsymbol{\beta}_{1}-\bar{\mathbf{x}}_{i}^{(w)T}\boldsymbol{\beta}_{2})^{2}.\end{split}$ (3) To maximize $l(\boldsymbol{\omega})$ and find the maximum likelihood estimator $\hat{\boldsymbol{\omega}}$ of $\boldsymbol{\omega}$, we differentiate (3) with respect to $\boldsymbol{\omega}$ to obtain the estimating function $\boldsymbol{\psi}(\boldsymbol{\omega})$ and then solve the estimating equation $\boldsymbol{0}_{[p_{b}+p_{w}+3:1]}=\boldsymbol{\psi}(\boldsymbol{\omega})$. The components of $\boldsymbol{\psi}(\boldsymbol{\omega})$ are $\begin{split}&l_{\beta_{0}}(\boldsymbol{\omega})=\sum_{i=1}^{g}\tau_{i}(\bar{y}_{i}-\beta_{0}-\mathbf{x}_{i}^{(b)T}\boldsymbol{\beta}_{1}-\bar{\mathbf{x}}_{i}^{(w)T}\boldsymbol{\beta}_{2}),\\\ &\mathbf{l}_{\boldsymbol{\beta}_{1}}(\boldsymbol{\omega})=\sum_{i=1}^{g}\tau_{i}\mathbf{x}_{i}^{(b)}(\bar{y}_{i}-\beta_{0}-\mathbf{x}_{i}^{(b)T}\boldsymbol{\beta}_{1}-\bar{\mathbf{x}}_{i}^{(w)T}\boldsymbol{\beta}_{2}),\\\ &l_{\sigma_{\alpha}^{2}}(\boldsymbol{\omega})=-\frac{1}{2}\sum_{i=1}^{g}\tau_{i}+\frac{1}{2}\sum_{i=1}^{g}\tau_{i}^{2}(\bar{y}_{i}-\beta_{0}-\mathbf{x}_{i}^{(b)T}\boldsymbol{\beta}_{1}-\bar{\mathbf{x}}_{i}^{(w)T}\boldsymbol{\beta}_{2})^{2},\\\ &\mathbf{l}_{\boldsymbol{\beta}_{2}}(\boldsymbol{\omega})=\frac{1}{\sigma_{e}^{2}}\mathbf{S}_{w}^{xy}-\frac{1}{\sigma_{e}^{2}}\mathbf{S}_{w}^{x}\boldsymbol{\beta}_{2}+\sum_{i=1}^{g}\tau_{i}\bar{\mathbf{x}}_{i}^{(w)}(\bar{y}_{i}-\beta_{0}-\mathbf{x}_{i}^{(b)T}\boldsymbol{\beta}_{1}-\bar{\mathbf{x}}_{i}^{(w)T}\boldsymbol{\beta}_{2}),\\\ &l_{\sigma_{e}^{2}}(\boldsymbol{\omega})=-\frac{1}{2}\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}-\frac{n-g}{2\sigma_{e}^{2}}+\frac{1}{2\sigma_{e}^{4}}(S_{w}^{y}-2\mathbf{S}_{w}^{xyT}\boldsymbol{\beta}_{2}+\boldsymbol{\beta}_{2}^{T}\mathbf{S}_{w}^{x}\boldsymbol{\beta}_{2})\\\ &\qquad\qquad+\frac{1}{2}\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{2}(\bar{y}_{i}-\beta_{0}-\mathbf{x}_{i}^{(b)T}\boldsymbol{\beta}_{1}-\bar{\mathbf{x}}_{i}^{(w)T}\boldsymbol{\beta}_{2})^{2}.\end{split}$ (4) Let $\boldsymbol{\psi}(\boldsymbol{\omega})^{T}=[\boldsymbol{\psi}^{(b)}(\boldsymbol{\omega})^{T},\boldsymbol{\psi}^{(w)}(\boldsymbol{\omega})^{T}]$, where $\boldsymbol{\psi}^{(b)}(\boldsymbol{\omega})^{T}=[l_{\beta_{0}}(\boldsymbol{\omega}),\mathbf{l}_{\boldsymbol{\beta}_{1}}(\boldsymbol{\omega})^{T},l_{\sigma_{\alpha}^{2}}(\boldsymbol{\omega})]$ are the estimating functions for the between cluster parameters and $\boldsymbol{\psi}^{(w)}(\boldsymbol{\omega})^{T}=[\mathbf{l}_{\boldsymbol{\beta}_{2}}(\boldsymbol{\omega})^{T},l_{\sigma_{e}^{2}}(\boldsymbol{\omega})]$ are the estimating functions for the within cluster parameters. The derivatives of the estimating functions which we write as $\nabla\boldsymbol{\psi}(\boldsymbol{\omega})$ and their expected values under the model are given in the Appendix. To control the estimating function and derive the asymptotic properties of $\hat{\boldsymbol{\omega}}$ from the estimating equation, we impose the following condition. Condition A * 1. The model (2) holds with true parameters $\dot{\boldsymbol{\omega}}$ inside the parameter space $\Omega$. * 2. The number of clusters $g\to\infty$ and the minimum number of observations per cluster $m_{L}\to\infty$. * 3. The random variables $\\{\alpha_{i}\\}$ and $\\{e_{ij}\\}$ are independent and identically distributed and there is a $\delta>0$ such that $\operatorname{E}|\alpha_{i}|^{4+\delta}<\infty$ and $\operatorname{E}|e_{ij}|^{4+\delta}<\infty$ for all $i=1,\ldots,g$ and $j\in\mathcal{S}_{i}$. * 4. Suppose that the limits $\mathbf{c}_{1}=\lim_{g\rightarrow\infty}g^{-1}\sum_{i=1}^{g}\mathbf{x}_{i}^{(b)}$, $\mathbf{C}_{2}=\lim_{g\rightarrow\infty}g^{-1}\sum_{i=1}^{g}\mathbf{x}_{i}^{(b)}\mathbf{x}_{i}^{(b)^{T}}$ and $\mathbf{C}_{3}=\lim_{g\to\infty}\lim_{m_{L}\to\infty}n^{-1}\mathbf{S}_{w}^{x}$ exist and the matrices $\mathbf{C}_{2}$ and $\mathbf{C}_{3}$ are positive definite. Suppose further that $\lim_{g\to\infty}\lim_{m_{L}\to\infty}\frac{1}{g}\sum_{i=1}^{g}|\bar{\mathbf{x}}_{i}^{(w)}|^{2}<\infty$, and there is a $\delta>0$ such that $\lim_{g\to\infty}g^{-1}\sum_{i=1}^{g}|\mathbf{x}_{i}^{(b)}|^{2+\delta}<\infty$ and $\lim_{g\to\infty}\lim_{m_{L}\to\infty}n^{-1}\sum_{i=1}^{g}\sum_{j=1}^{m_{i}}|\mathbf{x}_{ij}^{(w)}-\bar{\mathbf{x}}_{i}^{(w)}|^{2+\delta}<\infty$. These are very mild conditions which are often satisfied in practice. Conditions A3 and A4 ensure that limits needed to ensure the existence of the asymptotic variance of the estimating function exist and that we can establish a Lyapounov condition and hence a central limit theorem for the estimating function. They also ensure that minus the appropriately normalised second derivative of the estimating function converges to $\mathbf{B}$ given in (11) below. Unlike in the case of fixed $m_{i}$, A4 does not involve unknown parameters through the weights $\dot{\tau}_{i}$. Our main result is the following theorem which we prove in Section 4. ###### Theorem 1. Suppose Condition A holds. Then, as $g,m_{L}\to\infty$, there is a solution $\hat{\boldsymbol{\omega}}$ to the estimating equations $\boldsymbol{0}_{[p_{b}+p_{w}+3:1]}=\boldsymbol{\psi}(\boldsymbol{\omega})$, satisfying $|\mathbf{K}^{1/2}(\hat{\boldsymbol{\omega}}-\dot{\boldsymbol{\omega}})|=O_{p}(1)$, where $\mathbf{K}=\operatorname{diag}(g,g\boldsymbol{1}_{p_{b}}^{T},g,n\boldsymbol{1}_{p_{w}}^{T},n)$ with $\boldsymbol{1}_{p}$ the $p$ vector of ones. Moreover, $\hat{\boldsymbol{\omega}}$ has the asymptotic representation $\mathbf{K}^{1/2}(\hat{\boldsymbol{\omega}}-\dot{\boldsymbol{\omega}})=\mathbf{B}^{-1}\mathbf{K}^{-1/2}\boldsymbol{\xi}+o_{p}(1),$ (5) where $\mathbf{B}$ is given by (11) below and $\boldsymbol{\xi}=[\xi_{\beta_{0}},\boldsymbol{\xi}_{\boldsymbol{\beta}_{1}}^{T},\xi_{\sigma_{\alpha}^{2}},\boldsymbol{\xi}_{\boldsymbol{\beta}_{2}}^{T},\xi_{\sigma_{e}^{2}}]^{T}$ has components $\begin{split}&\xi_{\beta_{0}}=\frac{1}{\dot{\sigma}_{\alpha}^{2}}\sum_{i=1}^{g}\alpha_{i},\qquad\boldsymbol{\xi}_{\boldsymbol{\beta}_{1}}=\frac{1}{\dot{\sigma}_{\alpha}^{2}}\sum_{i=1}^{g}\mathbf{x}_{i}^{(b)}\alpha_{i},\qquad\xi_{\sigma_{\alpha}^{2}}=\frac{1}{2\dot{\sigma}_{\alpha}^{4}}\sum_{i=1}^{g}(\alpha_{i}^{2}-\dot{\sigma}_{\alpha}^{2}),\\\ &\boldsymbol{\xi}_{\boldsymbol{\beta}_{2}}=\frac{1}{\dot{\sigma}_{e}^{2}}\sum_{i=1}^{g}\sum_{j=1}^{m_{i}}(\mathbf{x}_{ij}^{(w)}-\bar{\mathbf{x}}_{i}^{(w)})e_{ij}\qquad\mbox{and}\qquad\xi_{\sigma_{e}^{2}}=\frac{1}{2\dot{\sigma}_{e}^{4}}\sum_{i=1}^{g}\sum_{j=1}^{m_{i}}(e_{ij}^{2}-\dot{\sigma}_{e}^{2}).\end{split}$ It follows that $\mathbf{K}^{1/2}(\hat{\boldsymbol{\omega}}-\dot{\boldsymbol{\omega}})\xrightarrow{D}N(\boldsymbol{0},\mathbf{C}),$ where $\begin{split}\mathbf{C}&=\left[\begin{matrix}\dot{\sigma}_{\alpha}^{2}d&\dot{\sigma}_{\alpha}^{2}\mathbf{d}_{1}^{T}&\operatorname{E}\alpha_{1}^{3}&\boldsymbol{0}_{[1:p_{w}]}&0\\\ \dot{\sigma}_{\alpha}^{2}\mathbf{d}_{1}&\dot{\sigma}_{\alpha}^{2}\mathbf{D}_{2}&\boldsymbol{0}_{[p_{b}:1]}&\boldsymbol{0}_{[p_{b}:p_{w}]}&\boldsymbol{0}_{[p_{b}:1]}\\\ \operatorname{E}\alpha_{1}^{3}&\boldsymbol{0}_{[1:p_{b}]}&\operatorname{E}\alpha_{1}^{4}-\dot{\sigma}_{\alpha}^{4}&\boldsymbol{0}_{[1:p_{w}]}&0\\\ \boldsymbol{0}_{[p_{w}:1]}&\boldsymbol{0}_{[p_{w}:p_{b}]}&\boldsymbol{0}_{[p_{w}:1]}&\dot{\sigma}_{e}^{2}\mathbf{C}_{3}^{-1}&\boldsymbol{0}_{[p_{w}:1]}\\\ 0&\boldsymbol{0}_{[1:p_{b}]}&0&0&\operatorname{E}e_{ij}^{4}-\dot{\sigma}_{e}^{4}\end{matrix}\right]\end{split}$ with $d=1//(1-\mathbf{c}_{1}^{T}\mathbf{C}_{2}^{-1}\mathbf{c}_{1})$, $\mathbf{d}_{1}=-\mathbf{c}_{1}^{T}\mathbf{C}_{2}^{-1}/(1-\mathbf{c}_{1}^{T}\mathbf{C}_{2}^{-1}\mathbf{c}_{1})$ and $\mathbf{D}_{2}=\mathbf{C}_{2}^{-1}+\mathbf{C}_{2}^{-1}\mathbf{c}_{1}\mathbf{c}_{1}^{T}\mathbf{C}_{2}^{-1}/(1-\mathbf{c}_{1}^{T}\mathbf{C}_{2}^{-1}\mathbf{c}_{1})$. We now consider REML estimation. To describe REML, we group the parameters into the regression parameters $\boldsymbol{\beta}=[\beta_{0},\boldsymbol{\beta}_{1}^{T},\boldsymbol{\beta}_{2}^{T}]$ and variance components $\boldsymbol{\theta}=[\sigma_{\alpha}^{2},\sigma_{e}^{2}]^{T}$.The REML criterion function is obtained by replacing the regression parameters $\boldsymbol{\beta}$ in the log-likelihood (3) by their maximum likelihood estimators for each fixed $\boldsymbol{\theta}$ to produce a profile log- likelihood for $\boldsymbol{\theta}$ and then adding an adjustment term. Let $\mathbf{z}_{i}=[1,\mathbf{x}_{i}^{(b)T},\bar{\mathbf{x}}_{i}^{(w)T}]^{T}$, $\mathbf{w}=[0,\boldsymbol{0}_{[1:p_{b}]},\mathbf{S}_{w}^{xyT}]^{T}$ and $\mathbf{W}=\mbox{block diag}(0,\boldsymbol{0}_{[p_{b}:p_{b}]},\mathbf{S}_{w}^{x})$. Then, for each fixed $\boldsymbol{\theta}$, we solve the estimating equations in (4) for $\boldsymbol{\beta}$ to obtain $\begin{split}\hat{\boldsymbol{\beta}}(\boldsymbol{\theta})&=\boldsymbol{\Delta}(\boldsymbol{\theta})^{-1}\Big{(}\sum_{i=1}^{g}\tau_{i}\mathbf{z}_{i}\bar{y}_{i}+\sigma_{e}^{-2}\mathbf{w}\Big{)},\quad\mbox{ with }\quad\boldsymbol{\Delta}(\boldsymbol{\theta})=\sum_{i=1}^{g}\tau_{i}\mathbf{z}_{i}\mathbf{z}_{i}^{T}+\sigma_{e}^{-2}\mathbf{W},\end{split}$ and the REML criterion function is given by $l_{R}(\boldsymbol{\theta};\mathbf{y})=l(\hat{\boldsymbol{\beta}}(\boldsymbol{\theta}),\boldsymbol{\theta};\mathbf{y})-\frac{1}{2}\log\left\\{|\boldsymbol{\Delta}(\boldsymbol{\theta})|\right\\}.$ The REML estimator $\hat{\boldsymbol{\theta}}_{R}$ of $\boldsymbol{\theta}$ is the maximiser of the REML criterion function $l_{R}(\boldsymbol{\theta};\mathbf{y})$; we call $\hat{\boldsymbol{\beta}}_{R}=\hat{\boldsymbol{\beta}}(\hat{\boldsymbol{\theta}}_{R})$ the REML estimator of $\boldsymbol{\beta}$ and write $\hat{\boldsymbol{\omega}}_{R}=(\hat{\beta}_{R0},\hat{\boldsymbol{\beta}}_{R1}^{T},\hat{\sigma}_{R\alpha}^{2},\hat{\boldsymbol{\beta}}_{R2}^{T},\hat{\sigma}_{Re}^{2})^{T}$. Since $\boldsymbol{\Delta}(\boldsymbol{\theta})$ does not depend on $\boldsymbol{\beta}$, the REML estimator is also the maximiser of the adjusted log-likelihood $l_{A}(\boldsymbol{\beta},\boldsymbol{\theta};\mathbf{y})=l(\boldsymbol{\omega};\mathbf{y})-\frac{1}{2}\log\left\\{|\Delta(\boldsymbol{\theta})|\right\\}.$ That is, we can find the REML estimator in one step instead of two (Patefield, 1977) by maximising $l_{A}(\boldsymbol{\beta},\boldsymbol{\theta};\mathbf{y})$. In either case, the estimating function is $\boldsymbol{\psi}_{A}(\boldsymbol{\omega})=[l_{A\beta_{0}}(\boldsymbol{\omega}),\,\mathbf{l}_{A\beta_{1}}(\boldsymbol{\omega})^{T},\,l_{A\sigma_{\alpha}^{2}}(\boldsymbol{\omega}),\,\mathbf{l}_{A\beta_{2}}(\boldsymbol{\omega})^{T},\,l_{A\sigma_{e}^{2}}(\boldsymbol{\omega})]^{T}$. The derivatives $l_{A\beta_{0}}(\boldsymbol{\omega})=l_{\beta_{0}}(\boldsymbol{\omega})$, $\mathbf{l}_{A\beta_{1}}(\boldsymbol{\omega})=\mathbf{l}_{\beta_{1}}(\boldsymbol{\omega})$ and $\mathbf{l}_{A\beta_{2}}(\boldsymbol{\omega})=\mathbf{l}_{\beta_{2}}(\boldsymbol{\omega})$, while $\begin{split}&l_{A\sigma_{\alpha}^{2}}(\boldsymbol{\omega})=l_{\sigma_{\alpha}^{2}}(\boldsymbol{\omega})-\frac{1}{2}\mbox{trace}\Big{\\{}\boldsymbol{\Delta}(\boldsymbol{\theta})^{-1}\frac{\partial\boldsymbol{\Delta}(\boldsymbol{\theta})}{\partial\sigma_{\alpha}^{2}}\Big{\\}}\quad\mbox{ with }\quad\frac{\partial\boldsymbol{\Delta}(\boldsymbol{\theta})}{\partial\sigma_{\alpha}^{2}}=-\sum_{i=1}^{g}\tau_{i}^{2}\mathbf{z}_{i}\mathbf{z}_{i}^{T}\\\ &l_{A\sigma_{e}^{2}}(\boldsymbol{\omega})=l_{\sigma_{e}^{2}}(\boldsymbol{\omega})-\frac{1}{2}\mbox{trace}\Big{\\{}\boldsymbol{\Delta}(\boldsymbol{\theta})^{-1}\frac{\partial\boldsymbol{\Delta}(\boldsymbol{\theta})}{\partial\sigma_{e}^{2}}\Big{\\}}\quad\mbox{ with }\\\ &\qquad\frac{\partial\boldsymbol{\Delta}(\boldsymbol{\theta})}{\partial\sigma_{e}^{2}}=-\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{2}\mathbf{z}_{i}\mathbf{z}_{i}^{T}-\sigma_{e}^{-4}\mathbf{W}.\end{split}$ We show that the REML estimator is asymptotically equivalent to the maximum likelihood estimator by showing that the contribution from the adjustment terms to the estimating function is asymptotically negligible. This yields the following theorem which we prove in Section 4. ###### Theorem 2. Suppose Condition A holds. Then, as $g,m_{L}\to\infty$, there is a solution $\hat{\boldsymbol{\omega}}_{R}$ to the adjusted likelihood estimating equations satisfying $|\mathbf{K}^{1/2}(\hat{\boldsymbol{\omega}}_{R}-\dot{\boldsymbol{\omega}})|=O_{p}(1)$ and $\mathbf{K}^{\frac{1}{2}}(\hat{\boldsymbol{\omega}}_{R}-\hat{\boldsymbol{\omega}})=\ o_{p}(1),$ so Theorem 1 applies to the REML estimator. ## 3 Discussion Theorems 1 and 2 establish the asymptotic equivalence, asymptotic representations and asymptotic normality for the maximum likelihood and REML estimators of the parameters in the nested error regression model under very mild conditions when both the number of clusters and the cluster sizes increase to infinity. In this section we interpret and discuss these results before pointing out possible directions for future work. We can estimate $\dot{\boldsymbol{\omega}}$ consistently when $g\to\infty$ with bounded cluster sizes but we need to let $m_{L}\to\infty$ to estimate the random effects $\\{\alpha_{i}\\}$ consistently (Jiang, 1998). If $m_{L}\to\infty$ but $g$ is held fixed, we can estimate the within cluster variance $\dot{\sigma}_{e}^{2}$ consistently but not the between cluster variance $\dot{\sigma}_{\alpha}^{2}$. These considerations motivate allowing both $g\to\infty$ and $m_{L}\to\infty$. The asymptotic representation shows that the influence function of the maximum likelihood and REML estimators under the model is given by the summands of $\mathbf{B}^{-1}\boldsymbol{\xi}$. Explicitly, at a point $[\alpha_{i},e_{ij},\mathbf{x}^{(b)T}_{i},\mathbf{x}_{ij}^{(w)T}]^{T}$ (which we suppress in the notation), the influence function is the $(p_{b}+p_{w}+3)$-vector function $\boldsymbol{\lambda}=[\lambda_{\beta_{0}},\boldsymbol{\lambda}_{\boldsymbol{\beta}_{1}}^{T},\lambda_{\sigma_{\alpha}^{2}},\boldsymbol{\lambda}_{\boldsymbol{\beta}_{2}}^{T},\lambda_{\sigma_{e}^{2}}]^{T}$, where $\begin{split}&\lambda_{\beta_{0}}=\\{(1-\mathbf{c}_{1}^{T}\mathbf{C}_{2}^{-1}\mathbf{x}_{i}^{(b)})/(1-\mathbf{c}_{1}^{T}\mathbf{C}_{2}^{-1}\mathbf{c}_{1})\\}\alpha_{i},\\\ &\boldsymbol{\lambda}_{\boldsymbol{\beta}_{1}}=\\{\mathbf{C}_{2}^{-1}\mathbf{x}_{i}^{(b)}+(1-\mathbf{c}_{1}^{T}\mathbf{C}_{2}^{-1}\mathbf{x}_{i}^{(b)})/(1-\mathbf{c}_{1}^{T}\mathbf{C}_{2}^{-1}\mathbf{c}_{1})\\}\mathbf{x}_{i}^{(b)}\alpha_{i},\\\ &\lambda_{\sigma_{\alpha}^{2}}=\alpha_{i}^{2}-\dot{\sigma}_{\alpha}^{2},\qquad\boldsymbol{\lambda}_{\boldsymbol{\beta}_{2}}=\mathbf{C}_{3}^{-1}(\mathbf{x}_{ij}^{(w)}-\bar{\mathbf{x}}_{i}^{(w)})e_{ij}\qquad\mbox{and}\qquad\lambda_{\sigma_{e}^{2}}=e_{ij}^{2}-\dot{\sigma}_{e}^{2}.\end{split}$ These expressions are not easy to obtain directly because the between and within parameters are estimated at different rates. As is well-known, the estimators are not robust because the influence function is unbounded in the covariates, random effect and error. The central limit theorem allows us to construct asymptotic confidence intervals for the parameters in the model. An asymptotic $100(1-\gamma)\%$ confidence interval for $\dot{\beta}_{1k}$ is $[\hat{\beta}_{1k}-\Phi^{-1}(1-\gamma/2)\hat{\sigma}_{\alpha}d_{kk}^{(b)1/2}/g^{1/2},\,\,\hat{\beta}_{1k}+\Phi^{-1}(1-\gamma/2)\hat{\sigma}_{\alpha}d_{kk(b)}^{1/2}/g^{1/2}],$ where $d_{kk}^{(b)}$ is the $k$th diagonal element of $\hat{\mathbf{C}}_{2}^{-1}+\hat{\mathbf{C}}_{2}^{-1}\hat{\mathbf{c}}_{1}\hat{\mathbf{c}}_{1}^{T}\hat{\mathbf{C}}_{2}^{-1}/(1-\hat{\mathbf{c}}_{1}^{T}\hat{\mathbf{C}}_{2}^{-1}\hat{\mathbf{c}}_{1})$ with $\hat{\mathbf{c}}_{1}=g^{-1}\sum_{i=1}^{g}\mathbf{x}_{i}^{(b)}$ and $\hat{\mathbf{C}}_{2}=g^{-1}\sum_{i=1}^{g}\mathbf{x}_{i}^{(b)}\mathbf{x}_{i}^{(b)^{T}}$, and an asymptotic $100(1-\gamma)\%$ confidence interval for $\dot{\beta}_{2r}$ is $[\hat{\beta}_{2r}-\Phi^{-1}(1-\gamma/2)\hat{\sigma}_{e}d_{rr}^{(w)1/2}/n^{1/2},\,\,\hat{\beta}_{2r}+\Phi^{-1}(1-\gamma/2)\hat{\sigma}_{e}d_{rr}^{(w)1/2}/n^{1/2}],$ where $d_{rr}^{(w)}$ is the $r$th diagonal element of $(\mathbf{S}_{w}^{x}/n)^{-1}$. Setting the confidence interval on the $\log$ scale and then backtransforming, an asymptotic $100(1-\gamma)\%$ confidence interval for $\dot{\sigma}_{\alpha}$ is $\begin{split}[\hat{\sigma}_{\alpha}\exp\\{-\Phi^{-1}(1-\gamma/2)&(\hat{\mu}_{4\alpha}-\hat{\sigma}_{\alpha}^{4})^{1/2}/2g^{1/2}\hat{\sigma}_{\alpha}^{2}\\},\,\\\ &\hat{\sigma}_{\alpha}\exp\\{\Phi^{-1}(1-\gamma/2)(\hat{\mu}_{4\alpha}-\hat{\sigma}_{\alpha}^{4})^{1/2}/2g^{1/2}\hat{\sigma}_{\alpha}^{2}\\}],\end{split}$ where $\hat{\mu}_{4\alpha}=g^{-1}\sum_{i=1}^{g}(\bar{y}-\hat{\beta}_{0}-\mathbf{x}_{i}^{(b)^{T}}\hat{\boldsymbol{\beta}}_{1}-\bar{\mathbf{x}}_{i}^{(w)^{T}}\hat{\boldsymbol{\beta}}_{2})^{4}$ estimates $\operatorname{E}\alpha_{1}^{4}$. Squaring the endpoints gives an asymptotic $100(1-\gamma)\%$ confidence interval for $\dot{\sigma}_{\alpha}^{2}$. The results show explicitly that the between and within parameters are estimated at different rates and the form of $\mathbf{C}$ shows that, even without assuming normality, the maximum likelihood and REML estimators of the within parameters are asymptotically independent of the estimators of the between parameters. That is, the two sets of parameters are asymptotically orthogonal. The within cluster regression parameter is asymptotically orthogonal to the within cluster variance and the between cluster slope parameter is asymptotically orthogonal to the between cluster variance, but the intercept is only asymptotically orthogonal to the between cluster variance when the random effect distribution is symmetric. When the cluster sizes are fixed, the maximum likelihood and REML estimators all converge to the true parameters at the same rate ($g^{-1/2}$) and the expression for their asymptotic variance is much more complicated. Appending a subscript $m$ to emphasise that the cluster sizes are fixed at their upper bounds, the asymptotic variance of the estimators is $g^{-1}\mathbf{C}_{m}=g^{-1}\mathbf{B}_{m}^{-1}\mathbf{A}_{m}\mathbf{B}_{m}^{-1}$, where $\mathbf{B}_{m}=-\lim_{g\to\infty}g^{-1}\operatorname{E}\nabla\psi(\dot{\boldsymbol{\omega}})$ (which we can obtain from (12)) and $\mathbf{A}_{m}=\lim_{g\to\infty}g^{-1}\operatorname{Var}\\{\psi(\dot{\boldsymbol{\omega}})\\}$. We require assumptions on the convergence of weighted means and weighted products of covariates to ensure the existence of $\mathbf{B}_{m}$ and $\mathbf{A}_{m}$. Under these assumptions, in general, $\mathbf{B}_{m}$ is block diagonal for $[\boldsymbol{\beta}^{T},\boldsymbol{\theta}^{T}]^{T}$, although it is not block diagonal for $\boldsymbol{\omega}$ because the $(\sigma_{\alpha}^{2},\sigma_{e}^{2})$ term is nonzero. When $\bar{\mathbf{x}}_{i}^{(w)}=\boldsymbol{0}_{[p_{w}:1]}$ for all $i=1,\ldots,g$, the $(\boldsymbol{\beta}_{2},\beta_{0})$ and $(\boldsymbol{\beta}_{2},\boldsymbol{\beta}_{1})$ terms are zero, but this does not affect the $(\sigma_{\alpha}^{2},\sigma_{e}^{2})$. The matrix $\mathbf{A}_{m}$ involves third and fourth moments and is rarely evaluated in the non-normal case; general expressions are given in Field et al. (2008) and expressions specific to the model (2) are available from the authors on request. It is in general not block diagonal for $[\boldsymbol{\beta}^{T},\boldsymbol{\theta}^{T}]^{T}$ unless both $\operatorname{E}(\alpha_{1}^{3})=0$ and $\operatorname{E}(e_{11}^{3})=0$. The centering condition makes the covariance of $\mathbf{l}_{\boldsymbol{\beta}_{2}}(\dot{\boldsymbol{\omega}})$ with all the other components of $\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})$ equal zero, but not the covariance between $l_{\beta_{0}}(\dot{\boldsymbol{\omega}})$ or $\mathbf{l}_{\boldsymbol{\beta}_{2}}(\dot{\boldsymbol{\omega}})$ with $l_{\sigma_{\alpha}^{2}}(\dot{\boldsymbol{\omega}})$ or $l_{\sigma_{e}^{2}}(\dot{\boldsymbol{\omega}})$. These limits are not block diagonal for $\boldsymbol{\omega}$ even when both $\operatorname{E}(\alpha_{1}^{3})=0$ and $\operatorname{E}(e_{11}^{3})=0$ (because the covariance between $l_{\sigma_{\alpha}^{2}}(\dot{\boldsymbol{\omega}})$ and $l_{\sigma_{e}^{2}}(\dot{\boldsymbol{\omega}})$ is nonzero). Of course, the nonzero terms in $\mathbf{C}_{m}$ involve limits of weighted averages which differ from those in $\mathbf{C}$. In working with the model (2), we have both between cluster and within cluster regression parameters to estimate. If we have no between cluster covariates,we discard $\boldsymbol{\beta}_{1}$, while if we have no within cluster covariates, we discard $\boldsymbol{\beta}_{2}$. The results for these cases can be obtained as special cases of the general results by deleting the components of vectors and the rows and columns of matrices corresponding to the discarded parameter. If there are no between cluster covariates in the model (there is no $\boldsymbol{\beta}_{1}$ in the model), we drop rows and columns $2$ to $p_{b}+1$ from $\mathbf{C}$. If there are no within cluster covariates (there is no $\boldsymbol{\beta}_{2}$ in the model), we drop rows and columns $p_{b}+3$ to $p_{b}+p_{w}+2$ from $\mathbf{C}$. There is a corresponding simplification to Condition A4. We have treated the covariates in the model as fixed, conditioning on them when they are random. As noted by Yoon and Welsh (2020), when the covariates are random, it makes sense to treat them as having a similar covariance structure to the response. That is, $\mathbf{x}_{i}^{(b)}$ are independent with mean $\boldsymbol{\mu}_{x}^{(b)}$ and variance $\boldsymbol{\Sigma}_{x}^{(b)}$, and the $\mathbf{x}_{ij}^{(w)}$ are independent in different clusters but correlated within clusters with mean $\boldsymbol{\mu}_{x}^{(w)}$, variance $\boldsymbol{\Upsilon}_{x}^{(w)}+\boldsymbol{\Sigma}_{x}^{(w)}$ and within cluster covariance $\boldsymbol{\Upsilon}_{x}^{(w)}$. The two types of covariates can be correlated. Condition A holds if both covariates have finite $2+\delta$ moments. We have $\mathbf{c}_{1}=\boldsymbol{\mu}_{x}^{(b)}$,$\mathbf{C}_{2}=\boldsymbol{\Sigma}_{x}^{(b)}+\boldsymbol{\mu}_{x}^{(b)}\boldsymbol{\mu}_{x}^{(b)T}$ and $\mathbf{C}_{3}=\boldsymbol{\Sigma}_{x}^{(w)}$. If $\mathbf{x}_{i}^{(b)}$ contains $\bar{\mathbf{x}}_{i}^{(w)}$, the terms $\boldsymbol{\mu}_{x}^{(b)}$ and variance $\boldsymbol{\Sigma}_{x}^{(b)}$ contain $\operatorname{E}\bar{\mathbf{x}}_{i}^{(w)}=\boldsymbol{\mu}_{x}^{(w)}$, $\operatorname{Var}(\bar{\mathbf{x}}_{i}^{(w)})=\boldsymbol{\Upsilon}_{x}^{(w)}+m_{i}^{-1}\boldsymbol{\Sigma}_{x}^{(w)}\to\boldsymbol{\Upsilon}_{x}^{(w)}$, as $m_{L}\to\infty$, and the covariance between $\mathbf{x}_{i}^{(b)}$ and $\bar{\mathbf{x}}_{i}^{(w)}$. One motivation for allowing the cluster size to increase with the number of clusters is that, as we have noted, this is required for consistent prediction of the random effects (Jiang, 1998). We have not considered prediction of the random effects explicitly in this paper but will do so in follow up work. The present paper makes an important step towards tackling prediction for sample survey applications because our results allow subsampling within clusters. In particular if the model (2) holds for the finite population, then noninformative subsampling of units within clusters ensures that the sample data satisfy the same model and hence that we can apply Theorems 1 and 2. The model we have considered is a simple linear mixed model. It is of interest to extend our results to more general linear mixed models and indeed to generalized linear mixed models. It is clear that we can extend the hierarchical structure of the model and allow for more variance components. The effect is to increase the sets of parameters so that there is a set for each level in the hierarchy. The estimators in each level converge at different rates and the limit distribution has a diagonal block for each level in the hierarchy. The maximum likelihood and REML estimators are not the only estimators of interest for the parameters of linear mixed models. Other estimators (including robust estimators) are available and it is also of interest to derive their asymptotic properties. We expect that the form of the asymptotic covariance matrices for these estimators will be block diagonal with a separate block for the parameters at each level in the hierarchy, just as we found for the maximum likelihood and REML estimators. Finally, Jiang (1996) also allowed the number of covariates to increase asymptotically and showed that the maximum likelihood and REML estimators have different asymptotic properties in this case. This is also an interesting problem to consider in the framework of this paper. ## 4 Proofs The proofs of Theorems 1 and 2 are presented in Subsection 4.1. The supporting lemmas used in these proofs are then proved in Subsections 4.2 and 4.3. ### 4.1 Proofs of Theorems 1 and 2 Proof. Write $\begin{split}\mathbf{K}^{-1/2}\psi(\boldsymbol{\omega})&=\mathbf{K}^{-1/2}\boldsymbol{\xi}-\mathbf{B}\mathbf{K}^{1/2}(\boldsymbol{\omega}-\dot{\boldsymbol{\omega}})+T_{1}+T_{2}(\boldsymbol{\omega})+T_{3}(\boldsymbol{\omega}),\end{split}$ where ${\mathbf{B}}=\lim_{g\to\infty}\lim_{m_{L}\to\infty}-\mathbf{K}^{-1/2}\operatorname{E}\nabla\psi(\boldsymbol{\omega})\mathbf{K}^{-1/2}$, $T_{1}=\mathbf{K}^{-1/2}\\{\psi(\dot{\boldsymbol{\omega}})-\xi\\}$, $T_{2}(\boldsymbol{\omega})=\mathbf{K}^{-1/2}\operatorname{E}\\{\psi(\boldsymbol{\omega})-\psi(\dot{\boldsymbol{\omega}})\\}+\mathbf{B}\mathbf{K}^{1/2}(\boldsymbol{\omega}-\dot{\boldsymbol{\omega}})$ and $T_{3}(\boldsymbol{\omega})=\mathbf{K}^{-1/2}[\psi(\boldsymbol{\omega})-\psi(\dot{\boldsymbol{\omega}})-\operatorname{E}\\{\psi(\boldsymbol{\omega})-\psi(\dot{\boldsymbol{\omega}})\\}]$. If we can show that $|T_{1}|=o_{p}(1)$, $\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}|T_{2}(\boldsymbol{\omega})|=o(1)$ and $\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}|T_{3}(\boldsymbol{\omega})|=o_{p}(1)$, respectively, then uniformly on $\mathcal{N}$, we have $\mathbf{K}^{-1/2}\psi(\boldsymbol{\omega})=\mathbf{K}^{-1/2}\boldsymbol{\xi}-\mathbf{B}\mathbf{K}^{1/2}(\boldsymbol{\omega}-\dot{\boldsymbol{\omega}})+o_{p}(1).$ (6) Multiplying by $(\boldsymbol{\omega}-\dot{\boldsymbol{\omega}})^{T}\mathbf{K}^{1/2}$, $(\boldsymbol{\omega}-\dot{\boldsymbol{\omega}})^{T}\psi(\boldsymbol{\omega})=(\boldsymbol{\omega}-\dot{\boldsymbol{\omega}})^{T}\boldsymbol{\xi}-(\boldsymbol{\omega}-\dot{\boldsymbol{\omega}})^{T}\mathbf{K}^{1/2}\mathbf{B}\mathbf{K}^{1/2}(\boldsymbol{\omega}-\dot{\boldsymbol{\omega}})+o_{p}(1).$ Since $B$ is positive definite, the right-hand side of is negative for $M$ sufficiently large. Therefore, according to Result 6.3.4 of Ortega and Rheinboldt (1973), a solution to the estimating equations exists in probability and satisfies $|\mathbf{K}^{1/2}(\hat{\boldsymbol{\omega}}-\dot{\boldsymbol{\omega}})|=O_{p}(1)$, so $\hat{\boldsymbol{\omega}}\in\mathcal{N}$. This allows us to substitute $\hat{\boldsymbol{\omega}}$ for $\boldsymbol{\omega}$ in (6) and rearrange the terms to obtain the asymptotic representation for $\hat{\boldsymbol{\omega}}$; the central limit theorem follows from the asymptotic representation and the central limit theorem for $\boldsymbol{\xi}$ that we establish in Lemma 1. It remains to show that that remainder terms in (4.1) are of smaller order and can be ignored. In Lemma 2, we establish $|T_{1}|=o_{p}(1)$ by showing that the result holds for each component of $T_{1}={\mathbf{K}}^{-1/2}\\{{\boldsymbol{\psi}}(\dot{\boldsymbol{\omega}})-\boldsymbol{\xi}\\}$ by applying Chebychev’s inequality and calculating the variances of the components. Our approach to handling $T_{2}(\boldsymbol{\omega})$ and $T_{3}(\boldsymbol{\omega})$ is inspired by Bickel (1975) who applied similar arguments to one-step regression estimators. The approach was extended to maximum likelihood and REML estimators in linear mixed models by Richardson and Welsh (1994); the bounds we use require more care with increasing cluster size. For $T_{2}(\boldsymbol{\omega})$, we have $\begin{split}\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}|T_{2}(\boldsymbol{\omega})|&\leq\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}\left|\mathbf{K}^{-1/2}\left[\operatorname{E}\left\\{\boldsymbol{\psi}(\boldsymbol{\omega})-\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})\right\\}-\operatorname{E}\nabla\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})(\boldsymbol{\omega}-\dot{\boldsymbol{\omega}})\right]\right|\\\ &\quad+\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}|\mathbf{K}^{-1/2}\operatorname{E}\nabla\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})(\boldsymbol{\omega}-\dot{\boldsymbol{\omega}})+\mathbf{B}\mathbf{K}^{1/2}(\boldsymbol{\omega}-\dot{\boldsymbol{\omega}})|\\\ &\leq M\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}\left\|\mathbf{K}^{-1/2}\left\\{\operatorname{E}\nabla\boldsymbol{\psi}(\boldsymbol{\Omega})-\operatorname{E}\nabla\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})\right\\}\mathbf{K}^{-1/2}\right\|\\\ &\quad+M\|-\mathbf{K}^{-1/2}\operatorname{E}\nabla\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})\mathbf{K}^{-1/2}-\mathbf{B}\|\\\ &\leq M\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}\left\|\mathbf{K}^{-1/2}\left\\{\operatorname{E}\nabla\boldsymbol{\psi}(\boldsymbol{\Omega})-\operatorname{E}\nabla\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})\right\\}\mathbf{K}^{-1/2}\right\|+M\|\mathbf{B}_{n}-\mathbf{B}\|,\end{split}$ where the rows of $\boldsymbol{\Omega}$ are possibly different but lie between $\boldsymbol{\omega}$ and $\dot{\boldsymbol{\omega}}$ and $\mathbf{B}_{n}=-\mathbf{K}^{-1/2}\operatorname{E}\nabla\psi(\dot{\boldsymbol{\omega}})\mathbf{K}^{-1/2}$. In Lemma 4 we show that $\|\mathbf{B}_{n}-\mathbf{B}\|=o(1)$ and in Lemma 5, $\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}\left\|\mathbf{K}^{-1/2}\left\\{\operatorname{E}\nabla\boldsymbol{\psi}(\boldsymbol{\Omega})-\operatorname{E}\nabla\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})\right\\}\mathbf{K}^{-1/2}\right\|=o(1).$ Finally, to handle $T_{3}(\boldsymbol{\omega})$, decompose $\mathcal{N}=\\{\boldsymbol{\omega}:|\mathbf{K}^{1/2}(\boldsymbol{\omega}-\dot{\boldsymbol{\omega}})|\leq M\\}$ into the set of $N=O(g^{\frac{1}{4}})$ smaller cubes $\mathcal{C}=\\{\mathcal{C}(\mathbf{t}_{k})\\}$, where $\mathcal{C}(\mathbf{t})=\\{\boldsymbol{\omega}:|\mathbf{K}^{1/2}(\mathbf{t}-\dot{\boldsymbol{\omega}})|\leq Mg^{-1/4}\\}$. We first show that $|T_{2}(\boldsymbol{\omega})|=o(1)$ holds over the set of indices $\mathbf{t}_{k}=(t_{k1},t_{k2},t_{k3},t_{k4},t_{k5})^{T}$ for the cubes in $\mathcal{C}$ and then that the difference between taking the supremum over a fine grid of points and over $\mathcal{N}$ is small. Using Chebychev’s inequality, for any $\eta>0$, we have $\begin{split}\operatorname{Pr}&\left(\max_{1\leq k\leq N}|\mathbf{K}^{-1/2}[\boldsymbol{\psi}(\mathbf{t}_{k})-\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})-\operatorname{E}\\{\boldsymbol{\psi}(\mathbf{t}_{k})-\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})\\}]|>\eta\right)\\\ &\leq\sum_{k=1}^{N}\operatorname{Pr}\left(|\mathbf{K}^{-1/2}[\boldsymbol{\psi}(\mathbf{t}_{k})-\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})-\operatorname{E}\\{\boldsymbol{\psi}(\mathbf{t}_{k})-\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})\\}]|>\eta\right)\\\ &\leq\eta^{-2}\sum_{k=1}^{N}\operatorname{E}|\mathbf{K}^{-1/2}[\boldsymbol{\psi}(\mathbf{t}_{k})-\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})-\operatorname{E}\\{\boldsymbol{\psi}(\mathbf{t}_{k})-\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})\\}]|^{2}\\\ &=\eta^{-2}g^{-1}\sum_{k=1}^{N}\operatorname{E}|\boldsymbol{\psi}^{(b)}(\mathbf{t}_{k})-\boldsymbol{\psi}^{(b)}(\dot{\boldsymbol{\omega}})-\operatorname{E}\\{\boldsymbol{\psi}^{(b)}(\mathbf{t}_{k})-\boldsymbol{\psi}^{(b)}(\dot{\boldsymbol{\omega}})\\}|^{2}\\\ &\qquad+\eta^{-2}n^{-1}\sum_{k=1}^{N}\operatorname{E}|\boldsymbol{\psi}^{(w)}(\mathbf{t}_{k})-\boldsymbol{\psi}^{(w)}(\dot{\boldsymbol{\omega}})-\operatorname{E}\\{\boldsymbol{\psi}^{(w)}(\mathbf{t}_{k})-\boldsymbol{\psi}^{(w)}(\dot{\boldsymbol{\omega}})\\}|^{2}\\\ &=\eta^{-2}g^{-1}\sum_{k=1}^{N}\mbox{trace}[\operatorname{Var}\\{\boldsymbol{\psi}^{(b)}(\mathbf{t}_{k})-\boldsymbol{\psi}^{(b)}(\dot{\boldsymbol{\omega}})\\}]\\\ &\qquad+\eta^{-2}n^{-1}\sum_{k=1}^{N}\mbox{trace}[\operatorname{Var}\\{\boldsymbol{\psi}^{(w)}(\mathbf{t}_{k})-\boldsymbol{\psi}^{(w)}(\dot{\boldsymbol{\omega}})\\}].\end{split}$ We show in Lemma 3 that the variances $\operatorname{Var}\\{\psi^{(b)}(\mathbf{t}_{k})-\psi^{(b)}(\dot{\boldsymbol{\omega}})\\}$ and $\operatorname{Var}\\{\psi^{(w)}(\mathbf{t}_{k})-\psi^{(w)}(\dot{\boldsymbol{\omega}})\\}$ are uniformly bounded by $L$, say, so $\begin{split}\operatorname{Pr}\Big{(}\max_{1\leq k\leq N}|\mathbf{K}^{-1/2}[\boldsymbol{\psi}(\mathbf{t}_{k})-&\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})-\operatorname{E}\\{\boldsymbol{\psi}(\mathbf{t}_{k})-\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})\\}]|>\eta\Big{)}\\\ &\leq\eta^{-2}LN\\{g^{-1}(2+p_{b})+n^{-1}(1+p_{w})\\}=o(1),\end{split}$ using the fact that $N=O(g^{1/4})$. Using Taylor expansion, we get $\begin{split}\underset{1\leq k\leq N}{\max}\underset{\boldsymbol{\omega}\in\mathcal{C}(\mathbf{t}_{k})}{\sup}&|\mathbf{K}^{-1/2}[\boldsymbol{\psi}(\boldsymbol{\omega})-\boldsymbol{\psi}(\mathbf{t}_{k})-\operatorname{E}\\{\boldsymbol{\psi}(\boldsymbol{\omega})-\boldsymbol{\psi}(\mathbf{t}_{k})\\}]|\\\ &=\underset{1\leq k\leq N}{\max}\underset{\boldsymbol{\omega}\in\mathcal{C}(\mathbf{t}_{k})}{\sup}|\mathbf{K}^{-1/2}[\nabla\boldsymbol{\psi}(\boldsymbol{\Omega}_{k})(\boldsymbol{\omega}-\mathbf{t}_{k})-\operatorname{E}\nabla\boldsymbol{\psi}(\boldsymbol{\Omega}_{k})(\boldsymbol{\omega}-\mathbf{t}_{k})]|\\\ &\leq M\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}g^{-1/4}\|\mathbf{K}^{-1/2}\\{\nabla\boldsymbol{\psi}(\boldsymbol{\Omega})-\operatorname{E}\nabla\boldsymbol{\psi}(\boldsymbol{\Omega})\\}\mathbf{K}^{-1/2}\|,\end{split}$ where the rows of $\boldsymbol{\Omega}_{k}$ are between $\mathbf{t}_{k}$ and $\boldsymbol{\omega}$. The result follows from Lemma 6. $\Box$ Proof. Let $\mathbf{K}_{\boldsymbol{\beta}}=\mbox{diag}(g,g\boldsymbol{1}_{p_{b}},n\boldsymbol{1}_{p_{w}})$ and write $\begin{split}|g^{-1/2}\\{l_{A\sigma_{\alpha}^{2}}(\boldsymbol{\omega})-l_{\sigma_{\alpha}^{2}}(\boldsymbol{\omega})\\}|&\leq\frac{1}{2g^{1/2}}|\mbox{trace}\Big{\\{}\mathbf{K}_{\boldsymbol{\beta}}^{1/2}\boldsymbol{\Delta}(\boldsymbol{\theta})^{-1}\mathbf{K}_{\boldsymbol{\beta}}^{1/2}\mathbf{K}_{\boldsymbol{\beta}}^{-1/2}\frac{\partial\boldsymbol{\Delta}(\boldsymbol{\theta})}{\partial\sigma_{\alpha}^{2}}\mathbf{K}_{\boldsymbol{\beta}}^{-1/2}\Big{\\}}|\\\ |n^{-1/2}\\{l_{A\sigma_{e}^{2}}(\boldsymbol{\omega})-l_{\sigma_{e}^{2}}(\boldsymbol{\omega})\\}|&\leq\frac{1}{2n^{1/2}}|\mbox{trace}\Big{\\{}\mathbf{K}_{\boldsymbol{\beta}}^{1/2}\boldsymbol{\Delta}(\boldsymbol{\theta})^{-1}\mathbf{K}_{\boldsymbol{\beta}}^{1/2}\mathbf{K}_{\boldsymbol{\beta}}^{-1/2}\frac{\partial\boldsymbol{\Delta}(\boldsymbol{\theta})}{\partial\sigma_{e}^{2}}\mathbf{K}_{\boldsymbol{\beta}}^{-1/2}\Big{\\}}|.\end{split}$ Then from Lemma 4 and the arguments establishing the convergence of $\mathbf{B}_{n}$ to $\mathbf{B}$, we can show that uniformly in $\boldsymbol{\omega}\in\mathcal{N}$ as $g,m_{L}\rightarrow\infty$, the matrices $\mathbf{K}_{\boldsymbol{\beta}}^{-1/2}\boldsymbol{\Delta}(\boldsymbol{\theta})\mathbf{K}_{\boldsymbol{\beta}}^{-1/2}$, $\mathbf{K}_{\boldsymbol{\beta}}^{-1/2}\frac{\partial\boldsymbol{\Delta}(\boldsymbol{\theta})}{\partial\sigma_{\alpha}^{2}}\mathbf{K}_{\boldsymbol{\beta}}^{-1/2}$ and $\mathbf{K}_{\boldsymbol{\beta}}^{-1/2}\frac{\partial\boldsymbol{\Delta}(\boldsymbol{\theta})}{\partial\sigma_{e}^{2}}\mathbf{K}_{\boldsymbol{\beta}}^{-1/2}$ all converge to $(p_{b}+p_{w}+1)\times(p_{b}+p_{w}+1)$ matrices with finite elements. Consequently, both $|g^{-1/2}\\{l_{A\sigma_{\alpha}^{2}}(\boldsymbol{\omega})-l_{\sigma_{\alpha}^{2}}(\boldsymbol{\omega})\\}|=o_{p}(1)$ and $|n^{-1/2}\\{l_{A\sigma_{e}^{2}}(\boldsymbol{\omega})-l_{\sigma_{e}^{2}}(\boldsymbol{\omega})\\}|=o_{p}(1)$ uniformly in $\boldsymbol{\omega}\in\mathcal{N}$, and the result follows from Theorem 1. $\Box$ ### 4.2 Lemmas for the estimating function $\boldsymbol{\psi}$ We prove a central limit theorem for $\boldsymbol{\xi}$ and that $T_{1}=o_{p}(1)$ (i.e $\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})$ can be approximated by $\boldsymbol{\xi}$). We also prove that the variances of the components of $\boldsymbol{\psi}(\boldsymbol{\omega})-\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})$ are uniformly bounded ###### Lemma 1. Suppose Condition A holds. Then, as $g,m_{L}\to\infty$, $\mathbf{K}^{-1/2}\boldsymbol{\xi}\xrightarrow{D}N(\boldsymbol{0},\mathbf{A})$, where $\begin{split}&\mathbf{A}=\\\ &\left[\begin{matrix}1/\dot{\sigma}_{\alpha}^{2}&\mathbf{c}_{1}^{T}/\dot{\sigma}_{\alpha}^{2}&\operatorname{E}\alpha_{1}^{3}/(2\dot{\sigma}_{\alpha}^{6})&\boldsymbol{0}_{[1:p_{w}]}&0\\\ \mathbf{c}_{1}/\dot{\sigma}_{\alpha}^{2}&\mathbf{C}_{2}/\dot{\sigma}_{\alpha}^{2}&\mathbf{c}_{1}\operatorname{E}\alpha_{1}^{3}/(2\dot{\sigma}_{\alpha}^{6})&\boldsymbol{0}_{[p_{b}:p_{w}]}&\boldsymbol{0}_{[p_{b}:1]}\\\ \operatorname{E}\alpha_{1}^{3}/(2\dot{\sigma}_{\alpha}^{6})&\mathbf{c}_{1}^{T}\operatorname{E}\alpha_{1}^{3}/(2\dot{\sigma}_{\alpha}^{6})&(\operatorname{E}\alpha_{1}^{4}-\dot{\sigma}_{\alpha}^{4})/(4\dot{\sigma}_{\alpha}^{8})&\boldsymbol{0}_{[1:p_{w}]}&0\\\ \boldsymbol{0}_{[p_{w}:1]}&\boldsymbol{0}_{[p_{w}:p_{b}]}&\boldsymbol{0}_{[p_{w}:1]}&\mathbf{C}_{3}/\dot{\sigma}_{e}^{2}&\boldsymbol{0}_{[p_{w}:1]}\\\ 0&\boldsymbol{0}_{[1:p_{b}]}&0&\boldsymbol{0}_{[1:p_{w}]}&(\operatorname{E}e_{11}^{4}-\dot{\sigma}_{e}^{4})/(4\dot{\sigma}_{e}^{8})\end{matrix}\right].\end{split}$ (7) ###### Proof. The components of $\boldsymbol{\xi}$ are sums of independent random variables with zero means and finite variances. Since $\\{\alpha_{i}\\}$ and $\\{e_{ij}\\}$ are independent, it is straightforward to compute $\operatorname{Var}\\{\mathbf{K}^{-1/2}\boldsymbol{\xi}\\}=\mathbf{A}_{n}$, and then from Condition A4, as $g,\,m_{L}\rightarrow\infty$, to show that $\mathbf{A}_{n}\rightarrow\mathbf{A}$. Partition $\boldsymbol{\xi}$ into $\boldsymbol{\xi}^{(b)}$ containing the first $p_{b}+2$ elements (corresponding to the between parameters) and $\boldsymbol{\xi}^{(w)}$ containing the remaining $p_{w}+1$ elements (corresponding to the within parameters). Partition $\mathbf{A}$ conformably into the block diagonal matrix with diagonal blocks $\mathbf{A}_{11}$ and $\mathbf{A}_{22}$, where $\mathbf{A}_{11}$ is $(p_{b}+2)\times(p_{b}+2)$ and $\mathbf{A}_{22}$ is $(p_{w}+1)\times(p_{w}+1)$. We prove that $g^{-1/2}\boldsymbol{\xi}^{(b)}\xrightarrow{D}N(\boldsymbol{0},\mathbf{A}_{11})$ and $n^{-1/2}\boldsymbol{\xi}^{(w)}\xrightarrow{D}N(\boldsymbol{0},\mathbf{A}_{22})$, and the result then follows from the fact that $\boldsymbol{\xi}^{(b)}$ and $\boldsymbol{\xi}^{(w)}$ are independent. Write $\boldsymbol{\xi}^{(b)}=\sum_{i=1}^{g}\boldsymbol{\xi}_{i}^{(b)}$, where the summands $\boldsymbol{\xi}_{i}^{(b)}=[\xi_{i\beta_{0}},\boldsymbol{\xi}_{i\boldsymbol{\beta}_{1}}^{T},\xi_{i\sigma_{\alpha}^{2}}]^{T}$, and let $\mathbf{a}$ be a fixed $(p_{b}+2)$-vector satisfying $\mathbf{a}^{T}\mathbf{a}=1$. Then $g^{-1/2}\mathbf{a}^{T}\boldsymbol{\xi}^{(b)}$ is a sum of independent scalar random variables with mean zero and finite variance. It follows from the $c_{r}$-inequality and Conditions A3 - A4 that Lyapunov’s condition holds. Consequently $g^{-1/2}\mathbf{a}^{T}\boldsymbol{\xi}^{(b)}$ converges in distribution to $N(0,\mathbf{a}^{T}\mathbf{A}_{11}\mathbf{a})$, as $g\to\infty$ and the result follows from the Cramer-Wold device (Billingsley, 1999, p 49). The proof that $n^{-\frac{1}{2}}\boldsymbol{\xi}^{(w)}$ converges to $N(0,\mathbf{A}_{22})$, as $g,m_{L}\to\infty$, is similar. ∎ In the proofs of Lemmas 2-6, we use the following simple bounds which we gather here for convenience. Uniformly on $\boldsymbol{\omega}\in\mathcal{N}$, there exist fixed constants $0<L_{1}<L_{2}<\infty$ such that both $m_{i}L_{2}^{-1}\leq m_{i}\dot{\sigma}_{\alpha}^{2}\leq\dot{\sigma}_{e}^{2}+m_{i}\dot{\sigma}_{\alpha}^{2}=m_{i}\dot{\sigma}_{\alpha}^{2}\\{\dot{\sigma}_{e}^{2}/m_{i}\dot{\sigma}_{\alpha}^{2}+1\\}\leq m_{i}L_{1}^{-1}$ and for $g$ sufficiently large, $m_{i}L_{2}^{-1}\leq\sigma_{e}^{2}+m_{i}\sigma_{\alpha}^{2}\leq m_{i}L_{1}^{-1}$ hold. It follows that uniformly both in $\boldsymbol{\omega}\in\mathcal{N}$ and $1\leq i\leq g$, $\begin{split}&L_{1}\leq\tau_{i},\dot{\tau}_{i}\leq L_{2},\qquad|\tau_{i}-\dot{\tau}_{i}|\leq L_{2}^{2}M(m_{L}^{-1}n^{-1/2}+g^{-1/2})\leq O(g^{-1/2})\\\ &|\tau_{i}^{2}(1-\dot{\tau}_{i}^{-1}\tau_{i})|=O(g^{-1/2}),\quad|\tau_{i}^{2}-\dot{\tau}_{i}^{2}|\leq|\tau_{i}-\dot{\tau}_{i}||\tau_{i}+\dot{\tau}_{i}|=O(g^{-1/2}).\end{split}$ (8) We also require the moments of $\bar{e}_{i}=m_{i}^{-1}\sum_{j=1}^{m_{i}}e_{ij}$ which are $\begin{split}&\operatorname{E}\bar{e}_{i}=0,\qquad\operatorname{Var}\bar{e}_{i}=m_{i}^{-1}\dot{\sigma}_{e}^{2},\qquad\operatorname{E}(\bar{e}_{i}^{3})=m_{i}^{-2}\operatorname{E}e_{11}^{3},\\\ &\operatorname{E}(\bar{e}_{i}^{4})=m_{i}^{-2}3\dot{\sigma}_{e}^{4}+m_{i}^{-3}\\{\operatorname{E}e_{11}^{4}-3\dot{\sigma}_{e}^{4}\\}\leq m_{i}^{-2}3\dot{\sigma}_{e}^{4}+m_{i}^{-3}\operatorname{E}e_{11}^{4};\end{split}$ (9) see for example (Cramér, 1946, p 345). These imply that $\operatorname{Var}(\alpha_{i}+\bar{e}_{i})=(\dot{\sigma}_{\alpha}^{2}+m_{i}^{-1}\dot{\sigma}_{e}^{2})=\dot{\tau}_{i}^{-1}$. ###### Lemma 2. Suppose Condition A holds. Then $|T_{1}|=o_{p}(1)$. ###### Proof. We establish the result for each component of $T_{1}={\mathbf{K}}^{-1/2}\\{{\boldsymbol{\psi}}(\dot{\boldsymbol{\omega}})-\boldsymbol{\xi}\\}$. Write $\bar{y}_{i}=\dot{\beta}_{0}+\mathbf{x}_{i}^{(b)T}\dot{\boldsymbol{\beta}}_{1}+\bar{\mathbf{x}}_{i}^{(w)T}\dot{\boldsymbol{\beta}}_{2}+\alpha_{i}+\bar{e}_{i}=\mathbf{z}_{i}^{T}\dot{\boldsymbol{\beta}}+\alpha_{i}+\bar{e}_{i}$, where $\bar{e}_{i}=m_{i}^{-1}\sum_{j=1}^{m_{i}}e_{ij}$, and $y_{ij}-\bar{y}_{i}=(\mathbf{x}_{ij}^{(w)}-\bar{\mathbf{x}}_{i}^{(w)})^{T}\dot{\boldsymbol{\beta}}_{2}+e_{ij}-\bar{e}_{i}$. Then $\begin{split}\mathbf{l}_{\boldsymbol{\beta}_{1}}(\dot{\boldsymbol{\omega}})-\boldsymbol{\xi}_{\boldsymbol{\beta}_{1}}&=\sum_{i=1}^{g}(\dot{\tau}_{i}-1/\dot{\sigma}_{\alpha}^{2})\mathbf{x}_{i}^{(b)}\alpha_{i}+\sum_{i=1}^{g}\dot{\tau}_{i}\mathbf{x}_{i}^{(b)}\bar{e}_{i}.\end{split}$ The $k$th components of the two sums in the last line have mean zero and variances $\operatorname{Var}\\{\sum_{i=1}^{g}(\dot{\tau}_{i}-1/\dot{\sigma}_{\alpha}^{2})x_{ik}^{(b)}\alpha_{i}\\}=\sum_{i=1}^{g}m_{i}^{-2}\dot{\tau}_{i}^{2}\dot{\sigma}_{e}^{4}x_{ik}^{(b)2}\dot{\sigma}_{\alpha}^{2}=O(m_{L}^{-2}g)$ and $\operatorname{Var}\\{\sum_{i=1}^{g}\dot{\tau}_{i}x_{ik}^{(b)}\bar{e}_{i(s)}\\}=\sum_{i=1}^{g}m_{i}^{-1}\dot{\tau}_{i}^{2}\dot{\sigma}_{e}^{2}x_{ik}^{(b)2}=O(m_{L}^{-1}g),$ respectively, using (8) and (9). It follows that $\mathbf{l}_{\boldsymbol{\beta}_{1}}(\dot{\boldsymbol{\omega}})-\boldsymbol{\xi}_{\boldsymbol{\beta}_{1}}=o_{p}(g^{1/2})$ and, by essentially the same argument, $l_{\beta_{0}}(\dot{\boldsymbol{\omega}})-\xi_{\beta_{0}}=o_{p}(g^{1/2})$. For the estimating equation for the between variance component, write $\begin{split}l_{\sigma_{\alpha}^{2}}(\dot{\boldsymbol{\omega}})-\xi_{\sigma_{\alpha}^{2}}=\frac{1}{2}\sum_{i=1}^{g}(\dot{\tau}_{i}^{2}-1/\dot{\sigma}_{\alpha}^{4})(\alpha_{i}^{2}-\dot{\sigma}_{\alpha}^{2})+\frac{1}{2}\sum_{i=1}^{g}\dot{\tau}_{i}^{2}(2\bar{e}_{i}\alpha_{i}+\bar{e}_{i}^{2}-\dot{\sigma}_{e}^{2}/m_{i}).\end{split}$ From (8) and (9), the variances of the sums are $O(m_{L}^{-2}g)$ and $O(m_{L}^{-1}g)$ so $l_{\sigma_{\alpha}^{2}}(\dot{\boldsymbol{\omega}})-\xi_{\sigma_{\alpha}^{2}}=o_{p}(g^{1/2})$. Next, we can write $\mathbf{S}_{w}^{xy}=\mathbf{S}_{w}^{x}\dot{\boldsymbol{\beta}}_{2}+\sum_{i=1}^{g}\sum_{j=1}^{m_{i}}(\mathbf{x}_{ij}^{(w)}-\bar{\mathbf{x}}_{i}^{(w)})e_{ij}$ so, for the within slope parameter, $\begin{split}\mathbf{l}_{\boldsymbol{\beta}_{2}}(\dot{\boldsymbol{\omega}})-\boldsymbol{\xi}_{\boldsymbol{\beta}_{2}}&=\sum_{i=1}^{g}\dot{\tau}_{i}\bar{\mathbf{x}}_{i}^{(w)}(\alpha_{i}+\bar{e}_{i})=o_{p}(n^{-1/2}),\end{split}$ because $g=o(n)$. For the within variance component, expanding $S_{w}^{y}$, we show that $\begin{split}&S_{w}^{y}-2\dot{\boldsymbol{\beta}}_{2}^{T}\mathbf{S}_{w}^{xy}+\dot{\boldsymbol{\beta}}_{2}^{T}\mathbf{S}_{w}^{x}\dot{\boldsymbol{\beta}}_{2}=(n-g)\dot{\sigma}_{e}^{2}+\sum_{i=1}^{g}\sum_{j=1}^{m_{i}}(e_{ij}^{2}-\dot{\sigma}_{e}^{2})+o_{p}(n^{1/2}).\end{split}$ (10) It then follows that $\begin{split}l_{\sigma_{e}^{2}}(\dot{\boldsymbol{\omega}})-\xi_{\sigma_{e}^{2}}&=\frac{1}{2}\sum_{i=1}^{g}m_{i}^{-1}\dot{\tau}_{i}^{2}(\alpha_{i}^{2}-\dot{\sigma}_{\alpha}^{2}+2\alpha_{i}\bar{e}_{i}+\bar{e}_{i}^{2}-\dot{\sigma}_{e}^{2}/m_{i})+o_{p}(n^{1/2}).\end{split}$ Since $\operatorname{E}(\alpha_{i}^{2}-\dot{\sigma}_{\alpha}^{2}+2\alpha_{i}\bar{e}_{i}+\bar{e}_{i}^{2}-\dot{\sigma}_{e}^{2}/m_{i})^{2}=\operatorname{E}(\alpha_{1}^{2}-\dot{\sigma}_{\alpha}^{2})^{2}+m_{i}^{-1}4\dot{\sigma}_{\alpha}^{2}\dot{\sigma}_{e}^{2}+\operatorname{Var}(\bar{e}_{i}^{2})$, we have $l_{\sigma_{e}^{2}}(\dot{\boldsymbol{\omega}})-\xi_{\sigma_{e}^{2}}=o_{p}(n^{1/2})$, which completes the proof. ∎ ###### Lemma 3. Suppose Condition A holds. Then, there exists a finite constant $L$ such that $\begin{split}&\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}\operatorname{Var}\\{l_{\beta_{0}}(\boldsymbol{\omega})-l_{\beta_{0}}(\dot{\boldsymbol{\omega}})\\}\leq L,\quad\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}\operatorname{Var}\\{l_{\beta_{1k}}(\boldsymbol{\omega})-l_{\beta_{1k}}(\dot{\boldsymbol{\omega}})\\}\leq L,\\\ &\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}\operatorname{Var}\\{l_{\sigma_{\alpha}^{2}}(\boldsymbol{\omega})-l_{\sigma_{\alpha}^{2}}(\dot{\boldsymbol{\omega}})\\}\leq L,\quad\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}\operatorname{Var}\\{l_{\beta_{2r}}(\boldsymbol{\omega})-l_{\beta_{2r}}(\dot{\boldsymbol{\omega}})\\}\leq L,\\\ &\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}\operatorname{Var}\\{l_{\sigma_{e}^{2}}(\boldsymbol{\omega})-l_{\sigma_{e}^{2}}(\dot{\boldsymbol{\omega}})\\}\leq L,\,\,\,\,k=1,\ldots,p_{b},r=1,\ldots,p_{w}.\end{split}$ ###### Proof. We write out $\boldsymbol{\psi}(\boldsymbol{\omega})-\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})$, take the variance (which eliminates all no-stochastic terms) and then bound the variance uniformly on $\boldsymbol{\omega}\in\mathcal{N}$ using (8) and (9). For example,we have $\begin{split}l_{\beta_{0}}(\boldsymbol{\omega})-l_{\beta_{0}}(\dot{\boldsymbol{\omega}})&=\sum_{i=1}^{g}\tau_{i}\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta})+\sum_{i=1}^{g}(\tau_{i}-\dot{\tau}_{i})(\alpha_{i}+\bar{e}_{i})\end{split}$ so $\begin{split}\operatorname{Var}\\{l_{\beta_{0}}(\boldsymbol{\omega})-l_{\beta_{0}}(\dot{\boldsymbol{\omega}})\\}&=\sum_{i=1}^{g}(\tau_{i}-\dot{\tau}_{i})^{2}/\dot{\tau}_{i}\\\ &\leq 2L_{1}^{-1}L_{2}^{2}M^{2}\sum_{i=1}^{g}(m_{i}^{-2}n^{-1}+g^{-1})\leq 4L_{1}^{-3}M^{2}.\end{split}$ The argument for each of the remaining terms is similar ∎ ### 4.3 Lemmas for the derivative of $\boldsymbol{\psi}$ A key part of the proof of Theorem 1 is using the mean value theorem to obtain a linear approximation for $\boldsymbol{\psi}(\boldsymbol{\omega})$. We apply the mean value theorem to each (real) element of ${\boldsymbol{\psi}}(\boldsymbol{\omega})$ so we need to allow different arguments (i.e. values of $\boldsymbol{\omega}$) in each row of the derivative matrix. Let $\boldsymbol{\Omega}$ be a $(p_{b}+p_{w}+3)\times(p_{b}+p_{w}+3)$ matrix and write $\nabla{\boldsymbol{\psi}}(\boldsymbol{\Omega})$ to mean that each row of the derivative $\nabla{\boldsymbol{\psi}}$ is evaluated at the corresponding row of $\boldsymbol{\Omega}$. We also partition $\nabla{\boldsymbol{\psi}}(\boldsymbol{\Omega})$ into submatrices conformably with the between cluster and within cluster parameters. Leting $\boldsymbol{\Omega}^{(b)}$ contain the first $p_{b}+2$ rows and $\boldsymbol{\Omega}^{(w)}$ the remaining $p_{w}+1$ rows of $\boldsymbol{\Omega}$, we can write $\begin{split}\nabla{\boldsymbol{\psi}}(\boldsymbol{\Omega})=\left[\begin{matrix}\nabla{\boldsymbol{\psi}^{(bb)}}({\boldsymbol{\Omega}}^{(b)})&\nabla{\boldsymbol{\psi}}^{(bw)}({\boldsymbol{\Omega}}^{(b)})\\\ \nabla{\boldsymbol{\psi}}^{(wb)}({\boldsymbol{\Omega}}^{(w)})&\nabla{\boldsymbol{\psi}}^{(ww)}({\boldsymbol{\Omega}}^{(w)})\end{matrix}\right].\end{split}$ The arguments to $\nabla\boldsymbol{\psi}^{(wb)}$ and $\nabla\boldsymbol{\psi}^{(bw)}$ are potentially different but, when they are the same, these matrices are the transposes of each other. When the rows of $\boldsymbol{\Omega}$ all equal $\boldsymbol{\omega}^{T}$, we simplify the notation by replacing $\boldsymbol{\Omega}$ and its submatrices by $\boldsymbol{\omega}$. (We discard the transpose because there is no ambiguity in doing so and the notation looks unnecessarily complicated when it is retained.) Again discarding the transpose, we also use $\boldsymbol{\omega}$ as a generic symbol to represent any of the rows of $\boldsymbol{\Omega}$ when the specific choice of row is not important. ###### Lemma 4. Suppose Condition A holds. Then, as $g,m_{L}\to\infty$, $\left\|\mathbf{B}_{n}-\mathbf{B}\right\|=o(1)$, where $\mathbf{B}_{n}=-\mathbf{K}^{-1/2}\operatorname{E}\nabla\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})\mathbf{K}^{-1/2}$ and $\mathbf{B}=\left[\begin{matrix}1/\dot{\sigma}_{\alpha}^{2}&\mathbf{c}_{1}^{T}/\dot{\sigma}_{\alpha}^{2}&0&\boldsymbol{0}_{[1:p_{w}]}&0\\\ \mathbf{c}_{1}/\dot{\sigma}_{\alpha}^{2}&\mathbf{C}_{2}/\dot{\sigma}_{\alpha}^{2}&\boldsymbol{0}_{[p_{b}:1]}&\boldsymbol{0}_{[p_{b}:p_{w}]}&\boldsymbol{0}_{[p_{b}:1]}\\\ 0&\boldsymbol{0}_{[1:p_{b}]}&1/(2\dot{\sigma}_{\alpha}^{4})&\boldsymbol{0}_{[1:p_{w}]}&0\\\ \boldsymbol{0}_{[p_{w}:1]}&\boldsymbol{0}_{[p_{w}:p_{b}]}&\boldsymbol{0}_{[p_{w}:1]}&\mathbf{C}_{3}/\dot{\sigma}_{e}^{2}&\boldsymbol{0}_{[p_{w}:1]}\\\ 0&\boldsymbol{0}_{[1:p_{b}]}&0&\boldsymbol{0}_{[1:p_{w}]}&1/(2\dot{\sigma}_{e}^{4})\end{matrix}\right].$ (11) We have $\mathbf{A}=\mathbf{B}$ under normality, but not otherwise. ###### Proof. From the expressions for the elements of $\operatorname{E}\nabla{\boldsymbol{\psi}}(\boldsymbol{\Omega})$ given in the Appendix, we have $\begin{split}&\mathbf{B}_{n}=\\\ &\left[\begin{matrix}g^{-1}\sum_{i=1}^{g}\dot{\tau}_{i}&g^{-1}\sum_{i=1}^{g}\dot{\tau}_{i}\mathbf{x}_{i}^{(b)T}&0&\mathbf{f}^{T}&0\\\ g^{-1}\sum_{i=1}^{g}\dot{\tau}_{i}\mathbf{x}_{i}^{(b)}&g^{-1}\sum_{i=1}^{g}\dot{\tau}_{i}\mathbf{x}_{i}^{(b)}\mathbf{x}_{i}^{(b)T}&\boldsymbol{0}_{[p_{b}:1]}&\mathbf{H}&\boldsymbol{0}_{[p_{b}:1]}\\\ 0&\boldsymbol{0}_{[1:p_{b}]}&(2g)^{-1}\sum_{i=1}^{g}\dot{\tau}_{i}^{2}&\boldsymbol{0}_{[1:p_{w}]}&r\\\ \mathbf{f}&\mathbf{H}^{T}&\boldsymbol{0}_{[p_{w}:1]}&\mathbf{S}_{w}^{x}/n\dot{\sigma}_{e}^{2}+\mathbf{P}&\boldsymbol{0}_{[p_{w}:1]}\\\ 0&\boldsymbol{0}_{[1:p_{b}]}&r&\boldsymbol{0}_{[1:p_{w}]}&q\end{matrix}\right],\end{split}$ (12) where $\mathbf{f}=(gn)^{-1/2}\sum_{i=1}^{g}\dot{\tau}_{i}\bar{\mathbf{x}}_{i}^{(w)T}$, $\mathbf{H}=(gn)^{-1/2}\sum_{i=1}^{g}\dot{\tau}_{i}\mathbf{x}_{i}^{(b)}\bar{\mathbf{x}}_{i}^{(w)T}$, $\mathbf{P}=n^{-1}\sum_{i=1}^{g}\dot{\tau}_{i}\bar{\mathbf{x}}_{i}^{(w)}\bar{\mathbf{x}}_{i}^{(w)T}$ and $q=(2n)^{-1}\sum_{i=1}^{g}m_{i}^{-2}\dot{\tau}_{i}^{2}+(n-g)/(2n\dot{\sigma}_{e}^{4})$ and $r=(4gn)^{-1/2}\sum_{i=1}^{g}m_{i}^{-1}\dot{\tau}_{i}^{2}$. It is straightforward to show from Conditions A3-A4 and the fact that $\dot{\tau}_{i}\rightarrow 1/\dot{\sigma}_{\alpha}^{2}$ uniformly in $1\leq i\leq g$ as $m_{L}\to\infty$ that $\begin{split}|g^{-1}\sum_{i=1}^{g}\dot{\tau}_{i}\bar{x}_{ik}^{(b)}-c_{1k}/\dot{\sigma}_{\alpha}^{2}|&\leq\max_{1\leq i\leq g}|\dot{\tau}_{i}-1/\dot{\sigma}_{\alpha}^{2}|g^{-1}\sum_{i=1}^{g}|\bar{x}_{ik}^{(b)}|\\\ \quad&+|g^{-1}\sum_{i=1}^{g}\bar{x}_{ik}^{(w)}-c_{1k}|/\dot{\sigma}_{\alpha}^{2}=o(1).\end{split}$ Similar arguments can be applied to establish the convergence of the terms the remaining terms in $-g^{-1}\operatorname{E}\nabla{\boldsymbol{\psi}^{(bb)}}({\boldsymbol{\Omega}}^{(b)})$ and $-n^{-1}\operatorname{E}\nabla{\boldsymbol{\psi}}^{(ww)}({\boldsymbol{\Omega}}^{(w)})$. Finally, similar arguments can be used to show that $(n/g)^{1/2}$ times the entries in the off-diagonal blocks $(ng)^{-1/2}\operatorname{E}\nabla{\boldsymbol{\psi}}^{(bw)}({\boldsymbol{\Omega}}^{(b)})$ and $(ng)^{-1/2}\operatorname{E}\nabla{\boldsymbol{\psi}}^{(wb)}({\boldsymbol{\Omega}}^{(w)})$ converge and then using the fact that $g=o(n)$ to show that the entries in the off-diagonal blocks converge to zero. ∎ The convergence result for the expected derivative of the estimating equation that we require in order to handle $T_{2}(\boldsymbol{\omega})$ is established in Lemma 5. ###### Lemma 5. Suppose Condition A holds. Then, as $g,m_{L}\to\infty$, $\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}\left\|\mathbf{K}^{-1/2}\left\\{\operatorname{E}\nabla\boldsymbol{\psi}(\boldsymbol{\Omega})-\operatorname{E}\nabla\boldsymbol{\psi}(\dot{\boldsymbol{\omega}})\right\\}\mathbf{K}^{-1/2}\right\|=o(1).$ ###### Proof. It is enough to show the uniform convergence to zero of the elements of $g^{-1}\\{\operatorname{E}\nabla\boldsymbol{\psi}^{(bb)}(\boldsymbol{\Omega}^{(b)})-\operatorname{E}\nabla\boldsymbol{\psi}^{(bb)}(\dot{\boldsymbol{\omega}})\\}$, $g^{-1/2}n^{-1/2}\\{\operatorname{E}\nabla\boldsymbol{\psi}^{(bw)}(\boldsymbol{\Omega}^{(b)})-\operatorname{E}\nabla\boldsymbol{\psi}^{(bw)}(\dot{\boldsymbol{\omega}})\\}$ and $n^{-1}\\{\operatorname{E}\nabla\boldsymbol{\psi}^{(ww)}(\boldsymbol{\Omega}^{(w)})-\operatorname{E}\nabla\boldsymbol{\psi}^{(ww)}(\dot{\boldsymbol{\omega}})\\}$. These are all deterministic matrices so the result is obtained by directly bounding the components of these matrices. In addition to the bounds (8), we also use the fact that, uniformly in $\boldsymbol{\omega}\in\mathcal{N}$, $|\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta})|\leq Mg^{-1/2}\\{(1+|\mathbf{x}_{i}^{(b)}|)+(g/n)^{1/2}|\bar{\mathbf{x}}^{(w)}_{i}|\\}$ to obtain bounds of the form $|g^{-1}\sum_{i=1}^{g}\tau_{i}^{2}\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta})|\leq L_{2}^{2}Mg^{-1}\sum_{i=1}^{g}g^{-1/2}\\{(1+|\mathbf{x}_{i}^{(b)}|)+(g/n)^{1/2}|\bar{\mathbf{x}}^{(w)}_{i}|\\}=O(g^{-1/2}).$ Combining these bounds, we can show that, uniformly in $\boldsymbol{\omega}\in\mathcal{N}$, $g^{-1}\\{\operatorname{E}\nabla\boldsymbol{\psi}^{(bb)}(\boldsymbol{\Omega}^{(b)})-\operatorname{E}\nabla\boldsymbol{\psi}^{(bb)}(\dot{\boldsymbol{\omega}})\\}=O(g^{-1/2})$, $g^{-1/2}n^{-1/2}\\{\operatorname{E}\nabla\boldsymbol{\psi}^{(bw)}(\boldsymbol{\Omega}^{(b)})-\operatorname{E}\nabla\boldsymbol{\psi}^{(bw)}(\dot{\boldsymbol{\omega}})\\}=O(n^{-1/2})$ and $n^{-1}\\{\operatorname{E}\nabla\boldsymbol{\psi}^{(ww)}(\boldsymbol{\Omega}^{(w)})-\operatorname{E}\nabla\boldsymbol{\psi}^{(ww)}(\dot{\boldsymbol{\omega}})\\}=O(n^{-1/2})$ and the result follows. ∎ The final result we require in order to handle $T_{3}(\boldsymbol{\omega})$ and complete the proof of Theorem 1 is given in Lemma 6. ###### Lemma 6. Suppose Condition A holds. As $g,m_{L}\to\infty$, $\begin{split}&\underset{\boldsymbol{\omega}\in\mathcal{N}}{\sup}g^{-1/4}\|\mathbf{K}^{-1/2}\\{\nabla\boldsymbol{\psi}(\boldsymbol{\Omega})-\operatorname{E}\nabla\boldsymbol{\psi}(\boldsymbol{\Omega})\\}\mathbf{K}^{-1/2}\|=o_{p}(1).\end{split}$ ###### Proof. Arguing as in the proof of Lemma 5, it is enough to show the uniform convergence to zero of the elements of $g^{-5/4}\\{\nabla\boldsymbol{\psi}^{(bb)}(\boldsymbol{\Omega}^{(b)})-\operatorname{E}\nabla\boldsymbol{\psi}^{(bb)}(\boldsymbol{\Omega}^{(b)})\\}$, $g^{-3/4}n^{-1/2}\\{\nabla\boldsymbol{\psi}^{(bw)}(\boldsymbol{\Omega}^{(b)})-\operatorname{E}\nabla\boldsymbol{\psi}^{(bw)}(\boldsymbol{\Omega}^{(b)})\\}$ and $g^{-1/4}n^{-1}\\{\nabla\boldsymbol{\psi}^{(ww)}(\boldsymbol{\Omega}^{(w)})-\operatorname{E}\nabla\boldsymbol{\psi}^{(ww)}(\boldsymbol{\Omega}^{(w)})\\}$. We use the bounds (8) and the fact that, by direct calculation of means and variances, we have $\begin{split}&\sum_{i=1}^{g}(1+|\mathbf{x}_{i}^{(b)}|)|\alpha_{i}+\bar{e}_{i}|=O_{p}(g),\quad\sum_{i=1}^{g}|\bar{\mathbf{x}}_{i}^{(w)}||\alpha_{i}+\bar{e}_{i}|=O_{p}(g)\quad\mbox{and}\\\ &\sum_{i=1}^{g}|(\alpha_{i}+\bar{e}_{i})^{2}-\dot{\sigma}_{\alpha}^{2}-m_{i}^{-1}\dot{\sigma}_{e}^{2}|=O_{p}(g).\end{split}$ For the derivatives with respect to the variance components, we have $\begin{split}|l_{\sigma_{\alpha}^{2}\sigma_{\alpha}^{2}}(\boldsymbol{\omega})-&\operatorname{E}l_{\sigma_{\alpha}^{2}\sigma_{\alpha}^{2}}(\boldsymbol{\omega})|\leq L_{2}^{3}\sum_{i=1}^{g}|(\alpha_{i}+\bar{e}_{i})^{2}-\dot{\sigma}_{\alpha}^{2}-m_{i}^{-1}\dot{\sigma}_{e}^{2}|\\\ &+2L_{2}^{3}M\sum_{i=1}^{g}\\{g^{-1/2}(1+|\mathbf{x}_{i}^{(b)}|)+n^{-1/2}|\bar{\mathbf{x}}_{i}^{(w)}|\\}|\alpha_{i}+\bar{e}_{i}|=O_{p}(g),\end{split}$ $\begin{split}|l_{\sigma_{\alpha}^{2}\sigma_{e}^{2}}(\boldsymbol{\omega})-&\operatorname{E}l_{\sigma_{\alpha}^{2}\sigma_{e}^{2}}(\boldsymbol{\omega})|\leq m_{L}^{-1}L_{2}^{3}\sum_{i=1}^{g}|(\alpha_{i}+\bar{e}_{i})^{2}-\dot{\sigma}_{\alpha}^{2}-m_{i}^{-1}\dot{\sigma}_{e}^{2}|\\\ &+m_{L}^{-1}2L_{2}^{3}M\sum_{i=1}^{g}\\{g^{-1/2}(1+|\mathbf{x}_{i}^{(b)}|)+n^{-1/2}|\bar{\mathbf{x}}_{i}^{(w)}|\\}|\alpha_{i}+\bar{e}_{i}|=O_{p}(m_{L}^{-1}g).\end{split}$ and $\begin{split}|l_{\sigma_{e}^{2}\sigma_{e}^{2}}(\boldsymbol{\omega})&-\operatorname{E}l_{\sigma_{e}^{2}\sigma_{e}^{2}}(\boldsymbol{\omega})|\leq\sigma_{e}^{-6}|2(\dot{\boldsymbol{\beta}}_{2}-\boldsymbol{\beta}_{2})^{T}\sum_{i=1}^{g}\sum_{j=1}^{m_{i}}(\mathbf{x}_{ij}^{(w)}-\bar{\mathbf{x}}_{i}^{(w)})e_{ij}+\sum_{i=1}^{g}\sum_{j=1}^{m_{i}}(e_{ij}^{2}-\dot{\sigma}_{e}^{2})\\\ &-\sum_{i=1}^{g}({m_{i}}\bar{e}_{i}^{2}-\dot{\sigma}_{e}^{2})|+m_{L}^{-2}L_{2}^{3}\sum_{i=1}^{g}|(\alpha_{i}+\bar{e}_{i})^{2}-\dot{\sigma}_{\alpha}^{2}-m_{i}^{-1}\dot{\sigma}_{e}^{2}|\\\ &+2m_{L}^{-2}L_{2}^{3}\sum_{i=1}^{g}\\{g^{-1/2}(1+|\mathbf{x}_{i}^{(b)}|)+n^{-1/2}|\bar{x}_{ik}^{(w)}|\\}|\alpha_{i}+\bar{e}_{i}|=O_{p}(n)\end{split}$ because $g<n$ implies $m_{L}^{-2}g^{1/2}<m_{L}^{-2}g<n$. ∎ ## Appendix A Appendix: The derivative and expected derivative of $\boldsymbol{\psi}$ For the first row in $\nabla{\boldsymbol{\psi}}^{(bb)}({\boldsymbol{\Omega}}^{(b)})$, we have $\begin{split}&l_{\beta_{0}\beta_{0}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}\tau_{i},\quad\mathbf{l}_{\beta_{0}\boldsymbol{\beta}_{1}}(\boldsymbol{\omega})^{T}=-\sum_{i=1}^{g}\tau_{i}\mathbf{x}_{i}^{(b)T},\quad l_{\beta_{0}\sigma_{\alpha}^{2}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}\tau_{i}^{2}(\bar{y}_{i}-\mathbf{z}_{i}^{T}\boldsymbol{\beta});\end{split}$ for rows $k=2,\ldots,p_{b}+1$ in $\nabla{\boldsymbol{\psi}}^{(bb)}({\boldsymbol{\Omega}}^{(b)})$, we have $\begin{split}&l_{\beta_{1k}\beta_{0}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}\tau_{i}\mathbf{x}_{ik}^{(b)},\quad\mathbf{l}_{\beta_{1k}\boldsymbol{\beta}_{1}}(\boldsymbol{\omega})^{T}=-\sum_{i=1}^{g}\tau_{i}\mathbf{x}_{ik}^{(b)}\mathbf{x}_{i}^{(b)T},\\\ &l_{\beta_{1k}\sigma_{\alpha}^{2}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}\tau_{i}^{2}\mathbf{x}_{ik}^{(b)}(\bar{y}_{i}-\mathbf{z}_{i}^{T}\boldsymbol{\beta});\end{split}$ and for the $(p_{b}+2)$th row in $\nabla{\boldsymbol{\psi}}^{(bb)}({\boldsymbol{\Omega}}^{(b)})$, we have $\begin{split}&l_{\sigma_{\alpha}^{2}\beta_{0}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}\tau_{i}^{2}(\bar{y}_{i}-\mathbf{z}_{i}^{T}\boldsymbol{\beta}),\quad\mathbf{l}_{\sigma_{\alpha}^{2}\boldsymbol{\beta}_{1}}(\boldsymbol{\omega})^{T}=-\sum_{i=1}^{g}\tau_{i}^{2}\mathbf{x}_{i}^{(b)T}(\bar{y}_{i}-\mathbf{z}_{i}^{T}\boldsymbol{\beta}),\\\ &l_{\sigma_{\alpha}^{2}\sigma_{\alpha}^{2}}(\boldsymbol{\omega})=\frac{1}{2}\sum_{i=1}^{g}\tau_{i}^{2}-\sum_{i=1}^{g}\tau_{i}^{3}(\bar{y}_{i}-\mathbf{z}_{i}^{T}\boldsymbol{\beta})^{2}.\end{split}$ The rows of $\nabla{\boldsymbol{\psi}}^{(bw)}({\boldsymbol{\Omega}}^{(b)})$ are $\begin{split}&\mathbf{l}_{\beta_{0}\boldsymbol{\beta}_{2}}(\boldsymbol{\omega})^{T}=-\sum_{i=1}^{g}\tau_{i}\bar{\mathbf{x}}_{i}^{(w)T},\quad l_{\beta_{0}\sigma_{e}^{2}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{2}(\bar{y}_{i}-\mathbf{z}_{i}^{T}\boldsymbol{\beta});\\\ &\mathbf{l}_{\boldsymbol{\beta}_{1k}\boldsymbol{\beta}_{2}}(\boldsymbol{\omega})^{T}=-\sum_{i=1}^{g}\tau_{i}x_{ik}^{(b)}\bar{\mathbf{x}}_{i}^{(w)T},\quad l_{\beta_{1k}\sigma_{e}^{2}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{2}\mathbf{x}_{ik}^{(b)}(\bar{y}_{i}-\mathbf{z}_{i}^{T}\boldsymbol{\beta});\quad\\\ &\mathbf{l}_{\sigma_{\alpha}^{2}\boldsymbol{\beta}_{2}}(\boldsymbol{\omega})^{T}=-\sum_{i=1}^{g}\tau_{i}^{2}\bar{\mathbf{x}}_{i}^{(w)T}(\bar{y}_{i}-\mathbf{z}_{i}^{T}\boldsymbol{\beta}),\quad\\\ &l_{\sigma_{\alpha}^{2}\sigma_{e}^{2}}(\boldsymbol{\omega})=\frac{1}{2}\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{2}-\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{3}(\bar{y}_{i}-\mathbf{z}_{i}^{T}\boldsymbol{\beta})^{2};\end{split}$ $k=1,\ldots,p_{b}$, and, finally, the rows of $\nabla{\boldsymbol{\psi}}^{(ww)}({\boldsymbol{\Omega}}^{(w)})$ are $\begin{split}&\mathbf{l}_{\beta_{2k}\boldsymbol{\beta}_{2}}(\boldsymbol{\omega})^{T}=-\frac{1}{\sigma_{e}^{2}}\mathbf{S}_{wk}^{xT}-\sum_{i=1}^{g}\tau_{i}\bar{x}_{ik}^{(w)}\bar{\mathbf{x}}_{i}^{(w)T};\\\ &l_{\beta_{2k}\sigma_{e}^{2}}(\boldsymbol{\omega})=-\frac{1}{\sigma_{e}^{4}}\\{S_{wk}^{xy}-\mathbf{S}_{wk}^{xT}\boldsymbol{\beta}_{2}\\}-\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{2}\bar{x}_{ik}^{(w)}(\bar{y}_{i}-\mathbf{z}_{i}^{T}\boldsymbol{\beta});\\\ &\mathbf{l}_{\sigma_{e}^{2}\boldsymbol{\beta}_{2}}(\boldsymbol{\omega})^{T}=-\frac{1}{\sigma_{e}^{4}}\\{\mathbf{S}_{w}^{xyT}-\boldsymbol{\beta}_{2}^{T}\mathbf{S}_{w}^{x}\\}-\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{2}\bar{\mathbf{x}}_{i}^{(w)T}(\bar{y}_{i}-\mathbf{z}_{i}^{T}\boldsymbol{\beta}),\\\ \ &l_{\sigma_{e}^{2}\sigma_{e}^{2}}(\boldsymbol{\omega})=\frac{1}{2}\sum_{i=1}^{g}m_{i}^{-2}\tau_{i}^{2}+\frac{n-g}{2\sigma_{e}^{4}}-\frac{1}{\sigma_{e}^{6}}(S_{w}^{y}-2\boldsymbol{\beta}_{2}^{T}\mathbf{S}_{w}^{xy}+\boldsymbol{\beta}_{2}^{T}\mathbf{S}_{w}^{x}\boldsymbol{\beta}_{2})\\\ &\qquad\qquad-\sum_{i=1}^{g}m_{i}^{-2}\tau_{i}^{3}(\bar{y}_{i}-\mathbf{z}_{i}^{T}\boldsymbol{\beta})^{2},\end{split}$ $k=1,\ldots,p_{w}$. Here we have written $\mathbf{S}_{wk}^{xT}$ for the $k$th row of $\mathbf{S}_{w}^{x}$ so $\mathbf{S}_{w}^{x}=[\mathbf{S}_{w1}^{x},\ldots,\mathbf{S}_{wp_{w}}^{x}]^{T}$ and $x_{ik}^{(b)}$, $\bar{x}_{ik}^{(w)}$and $S_{wk}^{xy}$ for the $k$th element of $\mathbf{x}_{i}^{(b)}$, $\bar{\mathbf{x}}_{i}^{(w)}$ and $\mathbf{S}_{wk}^{xy}$, respectively, so $\mathbf{x}_{i}^{(b)}=[x_{ik}^{(b)}]$, $\bar{\mathbf{x}}_{ik}^{(w)}=[\bar{x}_{ik}^{(w)}]$ and $\mathbf{S}_{wk}^{xy}=[S_{wk}^{xy}]$. When we need to address the elements of $\mathbf{S}_{w}^{x}$, we write $\mathbf{S}_{w}^{x}=[S_{wkr}^{x}]$. We calculate the expected derivative matrix using $\operatorname{E}(\bar{y}_{i}-\mathbf{z}_{i}^{T}\boldsymbol{\beta})^{2}=\left\\{\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta})\right\\}^{2}+\dot{\tau}_{i}^{-1}$, $\operatorname{E}(\mathbf{S}_{w}^{xy})=\mathbf{S}_{w}^{x}\boldsymbol{\beta}_{2}$ and $\operatorname{E}(S_{w}^{y}-2\boldsymbol{\beta}_{2}^{T}\mathbf{S}_{w}^{xy}+\boldsymbol{\beta}_{2}^{T}\mathbf{S}_{w}^{x}\boldsymbol{\beta}_{2})=(\dot{\boldsymbol{\beta}}_{2}-\boldsymbol{\beta}_{2})^{T}\mathbf{S}_{w}^{x}(\dot{\boldsymbol{\beta}}_{2}-\boldsymbol{\beta}_{2})+(n-g)\dot{\sigma}_{e}^{2}$. The first row of $\operatorname{E}\nabla{\boldsymbol{\psi}}^{(bb)}({\boldsymbol{\Omega}}^{(b)})$ is $\begin{split}&\operatorname{E}l_{\beta_{0}\beta_{0}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}\tau_{i},\quad\operatorname{E}\mathbf{l}_{\beta_{0}\boldsymbol{\beta}_{1}}(\boldsymbol{\omega})^{T}=-\sum_{i=1}^{g}\tau_{i}\mathbf{x}_{i}^{(b)T},\\\ &\operatorname{E}l_{\beta_{0}\sigma_{\alpha}^{2}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}\tau_{i}^{2}\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta});\end{split}$ rows $k=2,\ldots,p_{b}+1$ of $\operatorname{E}\nabla{\boldsymbol{\psi}}^{(bb)}({\boldsymbol{\Omega}}^{(b)})$ are $\begin{split}&\operatorname{E}l_{\beta_{1k}\beta_{0}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}\tau_{i}\mathbf{x}_{ik}^{(b)},\quad\operatorname{E}\mathbf{l}_{\beta_{1k}\boldsymbol{\beta}_{1}}(\boldsymbol{\omega})^{T}=-\sum_{i=1}^{g}\tau_{i}\mathbf{x}_{ik}^{(b)}\mathbf{x}_{i}^{(b)T},\\\ &\operatorname{E}l_{\beta_{1k}\sigma_{\alpha}^{2}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}\tau_{i}^{2}\mathbf{x}_{ik}^{(b)}\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta});\end{split}$ and the $(p_{b}+2)$th row of $\operatorname{E}\nabla{\boldsymbol{\psi}}^{(bb)}({\boldsymbol{\Omega}}^{(b)})$ is $\begin{split}&\operatorname{E}l_{\sigma_{\alpha}^{2}\beta_{0}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}\tau_{i}^{2}\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta}),\quad\operatorname{E}\mathbf{l}_{\sigma_{\alpha}^{2}\boldsymbol{\beta}_{1}}(\boldsymbol{\omega})^{T}=-\sum_{i=1}^{g}\tau_{i}^{2}\mathbf{x}_{i}^{(b)T}\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta}),\\\ &\operatorname{E}l_{\sigma_{\alpha}^{2}\sigma_{\alpha}^{2}}(\boldsymbol{\omega})=\frac{1}{2}\sum_{i=1}^{g}\tau_{i}^{2}(1-2\dot{\tau}_{i}^{-1}\tau_{i})-\sum_{i=1}^{g}\tau_{i}^{3}\\{\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta})\\}^{2}.\end{split}$ The first $p_{w}$ columns of $\operatorname{E}\nabla{\boldsymbol{\psi}}^{(bw)}({\boldsymbol{\Omega}}^{(b)})$ are $\begin{split}&\operatorname{E}\mathbf{l}_{\beta_{0}\boldsymbol{\beta}_{2}}(\boldsymbol{\omega})^{T}=-\sum_{i=1}^{g}\tau_{i}\bar{\mathbf{x}}_{i}^{(w)T},\quad\operatorname{E}\mathbf{l}_{\boldsymbol{\beta}_{1k}\boldsymbol{\beta}_{2}}(\boldsymbol{\omega})^{T}=-\sum_{i=1}^{g}\tau_{i}x_{ik}^{(b)}\bar{\mathbf{x}}_{i}^{(w)T};\\\ &\operatorname{E}\mathbf{l}_{\sigma_{\alpha}^{2}\boldsymbol{\beta}_{2}}(\boldsymbol{\omega})^{T}=-\sum_{i=1}^{g}\tau_{i}^{2}\bar{\mathbf{x}}_{i}^{(w)T}\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta}),\end{split}$ $k=1,\ldots,p_{b}$. The last column of $\operatorname{E}\nabla{\boldsymbol{\psi}}^{(bw)}({\boldsymbol{\Omega}})$ is $\begin{split}&\operatorname{E}l_{\beta_{0}\sigma_{e}^{2}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{2}\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta});\\\ &\operatorname{E}l_{\boldsymbol{\beta}_{1k}\sigma_{e}^{2}}(\boldsymbol{\omega})=-\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{2}x_{ik}^{(b)}\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta});\\\ &\operatorname{E}l_{\sigma_{\alpha}^{2}\sigma_{e}^{2}}(\boldsymbol{\omega})=\frac{1}{2}\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{2}(1-2\dot{\tau}_{i}^{-1}\tau_{i})-\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{3}\left\\{\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta})\right\\}^{2},\end{split}$ $k=1,\ldots,p_{b}$, and, finally, the rows of $\operatorname{E}\nabla{\boldsymbol{\psi}}^{(ww)}({\boldsymbol{\Omega}})$ are $\begin{split}&\operatorname{E}\mathbf{l}_{\beta_{2k}\boldsymbol{\beta}_{2}}(\boldsymbol{\omega})^{T}=-\frac{1}{\sigma_{e}^{2}}\mathbf{S}_{wk}^{xT}-\sum_{i=1}^{g}\tau_{i}\bar{x}_{ik}^{(w)}\bar{\mathbf{x}}_{i}^{(w)T},\,\,\,\,k=1,\ldots,p_{w};\\\ &\operatorname{E}l_{\beta_{2k}\sigma_{e}^{2}}(\boldsymbol{\omega})=-\frac{1}{\sigma_{e}^{4}}\mathbf{S}_{wk}^{xT}(\dot{\boldsymbol{\beta}}_{2}-\boldsymbol{\beta}_{2})-\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{2}\bar{x}_{ik}^{(w)}\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta}),\,\,\,\,k=1,\ldots,p_{w};\\\ &\operatorname{E}\mathbf{l}_{\sigma_{e}^{2}\boldsymbol{\beta}_{2}}(\boldsymbol{\omega})^{T}=-\frac{1}{\sigma_{e}^{4}}(\dot{\boldsymbol{\beta}}_{2}-\boldsymbol{\beta}_{2})^{T}\mathbf{S}_{w}^{x}-\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{2}\bar{\mathbf{x}}_{ik}^{(w)T}\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta})\\\ &\operatorname{E}l_{\sigma_{e}^{2}\sigma_{e}^{2}}(\boldsymbol{\omega})=\frac{1}{2}\sum_{i=1}^{g}m_{i}^{-1}\tau_{i}^{2}(1-2\dot{\tau}_{i}^{-1}\tau_{i})+\frac{n-g}{2\sigma_{e}^{4}}\big{(}1-2\frac{\dot{\sigma}_{e}^{2}}{\sigma_{e}^{2}}\big{)}\\\ &\qquad\qquad\qquad\qquad-\frac{1}{\sigma_{e}^{6}}(\dot{\boldsymbol{\beta}}_{2}-\boldsymbol{\beta}_{2})^{T}\mathbf{S}_{w}^{x}(\dot{\boldsymbol{\beta}}_{2}-\boldsymbol{\beta}_{2})-\sum_{i=1}^{g}m_{i}^{-2}\tau_{i}^{3}\\{\mathbf{z}_{i}^{T}(\dot{\boldsymbol{\beta}}-\boldsymbol{\beta})\\}^{2}.\end{split}$ ## References * Anderson [1969] T.W. Anderson. _Statistical inference for Covariance Matrices with Linear Structure_. Academic Press, New York, 1969. * Arora and Lahiri [1997] Vipin Arora and P. Lahiri. On the superiority of the Bayesian method over the BLUP in small area estimation problems. _Statistica Sinica_ , 7:1053–1063, 1997. * Battese et al. [1988] George E. Battese, Rachel M. Harter, and Wayne A. Fuller. An error-components model for prediction of county crop areas using survey and satellite data. _Journal of the American Statistical Association_ , 83:28–36, 1988. * Bickel [1975] Peter J. Bickel. One-step Huber estimates in the linear model. _Journal of the American Statistical Association_ , 70:428–434, 1975. * Billingsley [1999] P. Billingsley. _Convergence of Probability Measures_. John Wiley & Sons, New York, 2nd edition, 1999. * Cramér [1946] Harald Cramér. _Mathematical Methods of Statistics_. Princeton University Press, 1946. * Cressie and Lahiri [1993] Noel Cressie and Soumendra Nath Lahiri. The asymptotic distribution of REML estimators. _Journal of Multivariate Analysis_ , 45:217–233, 1993. * Das [1979] K. Das. Asymptotic optimality of restricted maximum likelihood estimates for the mixed model. _Calcutta Statistical Association Bulletin_ , 28:125–142, 1979. * Field et al. [2008] C. A. Field, Pang Zhen, and A. H. Welsh. Bootstrapping data with multiple levels of variation. _Canadian Journal of Statistics_ , 36:521–539, 2008. * Hartley and Rao [1967] Herman O. Hartley and J. N. K. Rao. Maximum-likelihood estimation for the mixed analysis of variance model. _Biometrika_ , 54:93–108, 1967. * Harville [1977] David A. Harville. Maximum likelihood approaches to variance component estimation and to related problems. _Journal of the American Statistical Association_ , 72:320–338, 1977. * Jiang [1996] Jiming Jiang. REML estimation: asymptotic behavior and related topics. _The Annals of Statistics_ , 24:255–286, 1996. * Jiang [1998] Jiming Jiang. Asymptotic properties of the empirical BLUP and BLUE in mixed linear models. _Statistica Sinica_ , 8:861–885, 1998. * Laird and Ware [1982] Nan M. Laird and James H. Ware. Random-effects models for longitudinal data. _Biometrics_ , 38:963–974, 1982. * Miller [1977] John J. Miller. Asymptotic properties of maximum likelihood estimates in the mixed model of the analysis of variance. _The Annals of Statistics_ , 5:746–762, 1977. * Ortega and Rheinboldt [1973] James M. Ortega and Werner C. Rheinboldt. _Iterative Solution of Nonlinear Equations in Several Variables_. Academic Press, New York, 1973. * Patefield [1977] W. M. Patefield. On the maximized likelihood function. _Sankhyā: The Indian Journal of Statistics, Series B_ , 39:92–96, 1977. * Pratesi [2016] Monica Pratesi. _Analysis of Poverty Data by Small Area Estimation_. John Wiley & Sons, New York, 2016. * Rao and Molina [2015] J. N. K. Rao and Isabel Molina. _Small Area Estimation_. John Wiley & Sons, New York, 2015. * Richardson and Welsh [1994] A. M. Richardson and A. H. Welsh. Asymptotic properties of restricted maximum likelihood (REML) estimates for hierarchical mixed linear models. _Australian Journal of statistics_ , 36:31–43, 1994. * Scott and Holt [1982] Alastair J. Scott and D. Holt. The effect of two-stage sampling on ordinary least squares methods. _Journal of the American Statistical Association_ , 77:848–854, 1982. * Xie and Yang [2003] Minge Xie and Yaning Yang. Asymptotics for generalized estimating equations with large cluster sizes. _The Annals of Statistics_ , 31:310–347, 2003. * Yoon and Welsh [2020] Hwan-Jin Yoon and A. H. Welsh. On the effect of ignoring correlation in the covariates when fitting linear mixed models. _Journal of Statistical Planning and Inference_ , 204:18–34, 2020.
# Study to improve the performance of interferometer with ultra-cold atoms Xiangyu Dong1, Shengjie Jin1, Hongmian Shui1, Peng Peng1, Xiaoji Zhou1,⋆ 1State Key Laboratory of Advanced Optical Communication System and Network, Department of Electronics, Peking University, Beijing 100871, China > Ultra-cold atoms provide ideal platforms for interferometry. The macroscopic > matter-wave property of ultra-cold atoms leads to large coherent length and > long coherent time, which enable high accuracy and sensitivity to > measurement. Here, we review our efforts to improve the performance of the > interferometer. We demonstrate a shortcut method for manipulating ultra-cold > atoms in an optical lattice. Compared with traditional ones, this shortcut > method can reduce manipulation time by up to three orders of magnitude. We > construct a matter-wave Ramsey interferometer for trapped motional quantum > states and significantly increase its coherence time by one order of > magnitude with an echo technique based on this method. Efforts have also > been made to enhance the resolution by multimode scheme. Application of a > noise-resilient multi-component interferometer shows that increasing the > number of paths could sharpen the peaks in the time-domain interference > fringes, which leads to a resolution nearly twice compared with that of a > conventional double-path two-mode interferometer. With the shortcut method > mentioned above, improvement of the momentum resolution could also be > fulfilled, which leads to atomic momentum patterns less than 0.6 $\hbar > k_{L}$. To identify and remove systematic noises, we introduce the methods > based on the principal component analysis (PCA) that reduce the noise in > detection close to the $1/\sqrt{2}$ of the photon-shot noise and separate > and identify or even eliminate noises. Furthermore, we give a proposal to > measure precisely the local gravity acceleration within a few centimeters > based on our study of ultracold atoms in precision measurements. Keywords: Precision Measurement, Ultra-cold atoms, Atomic interferometer, Gravity measurements PACS: 42.50.Dv, 67.10.Ba, 07.60.Ly, 91.10.Pp ## 1 Introduction Precision measurement is the cornerstone of the development of modern physics. Atom-based precision measurement is an important part. In recent years, ultra- cold atoms have attracted extensive interest in different fields (?, ?, ?, ?), because of their macroscopic matter-wave property (?, ?). This property will lead to large coherent length and long coherent time, which enable high fringe contrast (?, ?, ?). Hence, with these advantages, ultra-cold atoms provide ideal stages for precision measurement (?, ?) and have numerous applications (?, ?, ?, ?, ?, ?, ?, ?), ranging from inertia measurements (?, ?, ?) to precision time keeping (?, ?). For example, a Bose-Einstein condensate is used as an atomic source for a high precision sensor, which is released into free fall for up to 750 ms and probed with a $T=130$ ms Mach-Zehnder atom interferometer based on Bragg transitions (?). A trapped geometry is realized to probe gravity by holding ultra-cold cesium atoms for 20 seconds (?), which suppresses the phase variance due to vibrations by three to four orders of magnitude, overcoming the dominant noise source in atom-interferometric gravimeters. With ultra-cold atoms in an optical cavity, the detection of weak force can achieve a sensitivity of 42 $\rm{yN}/\sqrt{\rm{Hz}}$, which is a factor of 4 above the standard quantum limit (?). With the superiority mentioned above, it is obvious to consider an interferometer by ultra-cold atoms to further precision measurements. Interferometers with atoms propagating in free fall are ideally suited for inertia measurements (?, ?, ?, ?, ?, ?, ?). Meanwhile, with atoms held in tight traps or guides, they are better to measure weak localized interactions. For example, a direct measurement to the Casimir-Polder force is performed by I. Carusotto et al. in 2005, which is as large as $10^{-4}$ gravity (?). However, ultra-cold atoms still get some imperfections needed to be surmounted when combining with interferometry. The macroscopic matter-wave property is generated simultaneously with non-linear atom-atom interactions. Phase diffusion caused by interactions limits the coherence time, and ultimately restricts the sensitivity of interferometers. Besides the interrogation time, the momentum splitting as well as the path number also has an impact on the sensitivity. It has been demonstrated by experiments of multipath interferometers that, interferometric fringes can be sharpened due to the higher-harmonic phase contributions of the multiple energetically equidistant Zeeman states (?, ?), whereas a decrease in the average number of atoms per path causes a greater susceptibility to shot noise. Equilibrium between these parameters could lead to an optimal resolution. In addition, we should also pay attention to the signal analysis procedure as the interferometric information is mainly extracted from the signal detected. The resulting resolution severely relies on the probing system. In this review, we mainly introduce our experimental developments that study these fundamental and important issues to improve the performance of interferometer with ultra-cold atoms. The main developments are concentrated in three aspects: increasing coherent time, using multimode scheme and reducing systematic noises. A. Enhanced resolution by increasing coherent time. We introduce an effective and fast (few microseconds) methods, for manipulating ultra-cold atoms in an optical lattice (OL), which can be used to construct the atomic interferometer and increase the coherent time to finally get a higher resolution. This shortcut loading method is a designed pulse sequence, which can be used for preparing and manipulating arbitrary pure states and superposition states. Another advantage of this method is that the manipulation time is much shorter than traditional methods (100 ms$\to$100 $\rm{\mu s}$). Based on this shortcut method, we constructed an echo-Ramsey interferometer (RI) with motional Bloch states (at zero quasi-momentum on S- and D-bands of an OL) (?). Thanks to the rapidity of shortcut methods, more time could be used for the RI process. We identified the mechanisms that reduced the RI contrast, and greatly increased the coherent time (1.3 ms $\to$14.5 ms) by a quantum echo process, which eliminated the influence of contrast attenuation mechanisms mostly. B. Enhanced resolution by multimode scheme. Several efforts have been made to avoiding the decays of interferometric resolution because of the experimental noises. We demonstrated that the improvement of the phase resolution could be accomplished by a noise-resilient multi-component interferometric scheme. With the relative phase of different components remaining stable, increasing the number of paths could sharpen the peaks in the interference fringes, which led to a resolution nearly twice compared with that of a conventional double-path two-mode interferometer. Moreover, improvement of the momentum resolution was fulfilled with optical lattice pulses. We got results of atomic momentum patterns with intervals less than the double recoil momentum. The momentum pattern exhibited 10 main peaks. C. Enhanced resolution by removing the systematic noise. The method to identify and remove systematic noises for ultra-cold atoms is also introduced in this paper. For improving the quality of absorption image, which is the basic detection result in ultra-cold atoms experiments, we developed an optimized fringe removal algorithm (OFRA), making the noise close to the theoretical limit as $1/\sqrt{2}$ of the photon-shot noise. Besides, for the absorption images after preprocessing by OFRA, we applied the principal component analysis to successfully separate and identify noises from different origins of leading contribution, which helped to reduce or even eliminate noises via corresponding data processing procedures. Furthermore, based on our study of ultracold atoms in precision measurements, we demonstrated a scheme for potential compact gravimeter with ultra-cold atoms in a small displacement. The text structure is as follows. In Sec.2, a shortcut method manipulating ultra-cold atoms in an optical lattice and an Echo-Ramsey interferometry with motional quantum states are introduced, which can increase the coherent time. In Secs.3, we prove that the resolution can be increased using a double-path multimode interferometer with spinor Bose-Einstein condensates (BECs) or an optical pulse, both of which can be classified into multimode scheme. In Secs.4, methods for identifying and reducing the systematic noises for ultra- cold atoms are demonstrated. Finally, we give a proposal on gravimeter with ultra-cold atoms in Secs.5. ## 2 Enhanced resolution by increasing coherent time Figure 1: Schematic diagram of the shortcut method (take ground state preparation as an example). (a) At the beginning, the BECs are formed in a weak harmonic trap. (b) Time sequence of shortcut method. (c) Mapping the shortcut process onto the Bloch sphere. The track $A\to C\to|S\rangle$ and track $A\to B\to E\to M\to|S\rangle$ represent one pulse and two pulses shortcut process, respectively. (d) After this shortcut process, the desired states of an 3D optical lattice are prepared. (e) Band structure of 1D OL with different quasi-momentum $q$ when $V_{0}=10\;E_{r}$. Reproduced with permission from Ref. (?). The macroscopic coherent properties of ultra-cold atoms (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?) are conducive to precise measurement. To make full use of the advantages of ultra-cold atomic coherence properties, one method is to reduce the manipulation time and another is to suppress the attenuation of coherence. Firstly, we demonstrated a shortcut process for manipulating BECs trapped in an OL (?, ?, ?). By optimizing the parameters of the pulses, which constitute the sequence of the shortcut process, We can get extremely high fidelity and robustness for manipulating BECs into the desired states, including the ground state, excited states, and superposition states of a one, two or three- dimensional OLs. Another advantage of this method is that the manipulation time is much shorter than that in traditional methods (100 ms$\to$100 $\rm{\mu s}$). This shortcut is composed of optical lattice pulses and intervals that are imposed on the system before the lattice is switched on. The time durations and intervals in the sequence are optimized to transfer the initial state to the target state with high fidelity. This shortcut procedure can be completed in several tens of microseconds, which is shorter than the traditional method (usually hundreds of milliseconds). It can be applied to the fast manipulation of the superposition of Bloch states. Then, based on this method, we constructed an echo-Ramsey interferometer (RI) with motional Bloch states (at zero quasi-momentum on S- and D-bands of an OL) (?). The key to realizing a RI is to design effective $\pi$\- and $\pi/2$ pulses, which can be obtained by the shortcut method (?, ?, ?). Thanks to the rapidity of shortcut methods, more time can be used for the RI process. We identified the mechanisms that reduced the RI contrast, and greatly increased the coherent time (1.3 ms $\to$14.5 ms) by a quantum echo process, which eliminated the influence of most contrast attenuation mechanisms. ### 2.1 Shortcut manipulating ultra-cold atoms Efficient and fast manipulation of BECs in OLs can be used for precise measurements, such as constructing atom-based interferometers and increasing the coherent time of these interferometers. Here we demonstrate an effective and fast (around 100 $\rm{\mu s}$) method for manipulating BECs from an arbitrary initial state to a desired OL state. This shortcut method is a designed pulse sequence, in which the parameters, such as duration and interval of each step, are optimized to maximize fidelity and robustness of the final state. With this shortcut method, the pure Bloch states with even or odd parity and superposition states of OLs can be prepared and manipulated. In addition, the idea can be extended to the case of two- or three-dimensional OLs. This method has been verified by experiments many times and is very consistent with the theoretical analysis (?, ?, ?, ?, ?, ?, ?, ?). We used the simplest one-dimensional standing wave OL to demonstrate the design principle of this method. The OL potential is $V_{OL}(x)=V_{0}\cos^{2}{kx}$, where $V_{0}$ is used to characterize the depth of the OL. Supposing that the target state $\left|{\psi_{a}}\right\rangle$ is in the OL with depth $V_{0}$, $m$-step preloading sequence has been applied on the initial state $\left|\psi_{i}\right\rangle$. The final states $\left|{\psi_{f}}\right\rangle$ is given by $\centering|\psi_{f}\rangle=\prod_{j=1}^{m}\hat{Q}_{j}|\psi_{i}\rangle,\@add@centering$ (1) where $\hat{Q}_{j}=e^{-i\hat{H}_{j}t_{j}}$ is the evolution operator of the $j$th step. By maximizing the fidelity $Fidelity=|\langle\psi_{a}|\psi_{f}\rangle|^{2},$ (2) we can get the optimal parameters $\hat{H}_{j}$ and $t_{j}$. This preprocess is called a shortcut method, which can be used for loading atoms into different bands of an optical lattice. For example, the shortcut loading ultra-cold atoms into S-band in a one-dimensional optical lattice is shown in Fig. 1. By setting different initial state and target state, different time sequences can be designed to manipulate atoms, to build different interferometers, which greatly saves the coherent time. Based on this shortcut method, we can prepare exotic quantum states (?, ?, ?) and construct interferometer with motional quantum states of ultra-cold atoms (?). ### 2.2 Increasing coherent time in a Ramsey interferometry with motional Bloch states of ultra-cold atoms Suppressing the decoherence mechanism in the atomic interferometer is beneficial for increasing coherence time and improving the measurement accuracy. Here we demonstrated an echo method can increase the coherent time for Ramsey interferometry with motional Bloch states (at zero quasi-momentum on S- and D-bands of an OL) of ultra-cold atoms (?). The RI can be applied to the measurement of quantum many-body effects. The key challenge for the construction of this RI is to achieve $\pi$\- and $\pi/2$-pulses, because there is no selection rule for Bloch states of OLs. The $\pi$\- or $\pi/2$-pulse sequences can be obtained by the shortcut method (?, ?, ?, ?, ?), which precisely and rapidly manipulates the superposition of BECs at the zero quasi-momentum on the 1st and 3rd Bloch bands. Retaining the OL, we observed the interference between states and measured the decay of coherent oscillations. We identified the mechanisms that reduced the RI contrast: thermal fluctuations, laser intensity fluctuation, transverse expansion induced by atomic interaction, and the nonuniform OL depth. Then, we greatly increased the coherent time (1.3 ms $\to$14.5 ms) by a quantum echo process, which eliminated the influence of most contrast attenuation mechanisms. #### 2.2.1 Ramsey interferometer in an optical lattice This RI starts from BECs of ${}^{87}\rm{Rb}$ at the temperature 50 nK similar to our previous work (?, ?, ?, ?, ?, ?, ?, ?, ?). Then a 1D standing wave OL is formed (Fig. 2). After a shortcut sequence, the BEC is transferred into the ground band of the OL, denoted as $\phi_{S,0}$. Figure 2: Experimental configuration for a Ramsey interferometer in a $V_{0}=10\;E\mathrm{r}$ lattice: (a) The BEC is divided into discrete pancakes in $yz$ plane by an 1-dimensional optical lattice along $x$ axis with a lattice constant $d=426$ nm. (b) Band energies for the S-band and the D-band. (c) Time sequences for the Ramsey interferometry. The atoms are first loaded into the S band of OL, followed by the RI sequence: $\pi/2$ pulse, holding time $t_{OL}$, and the second $\pi/2$ pulse. Finally band mapping is used to detect the atom number in the different bands. (d) The used pulse sequences designed by an optimised shortcut method. Reproduced with permission from Ref. (?). Figure 3: (a) Change of $p_{\mathrm{D}}$, the population of atoms in the D-band, over time $t_{OL}$ with temperature $T=50$ nK. (b) Influence of different mechanisms on the RI. (c) Characteristic time $\tau$ for the different number of $\pi$ pulse $n$ and different temperatures. The circles, squares, and diamonds represent the experimental results and lines are fitting curves. Reproduced with permission from Ref. (?). Fig. 2(b) illustrates that the RI is constructed with Bloch states $\phi_{i,q}$, which includes the ground band $|S\rangle$, the third band $|D\rangle$, and their superposition state $\psi=a_{S}|S\rangle+a_{D}|D\rangle$ (denoted as $\binom{a_{S}}{a_{D}}$). It is difficult to realize the interferometer with this pseudo-spin system, because there is no selection rule for Bloch states of OLs. However, thanks to the existence of a coherent, macroscopic matter-wave, a $\pi/2$-pulse for BECs in an OL can be obtained, where the Bloch states $|S\rangle$ and $|D\rangle$ are to be manipulated to $|\psi_{1}\rangle$=$(|S\rangle+|D\rangle)/\sqrt{2}$ and $|\psi_{2}\rangle$=$(-|S\rangle+|D\rangle)/\sqrt{2}$, respectively. Fig. 2(d) shows the $\pi/2$ pulse we used for the RI (?, ?, ?) with fidelities of $98.5\%$ and $98.0\%$ respectively (?, ?, ?, ?). Fig. 2(c) illustrates the whole process of RI. First, the BEC is transferred to $|S\rangle$ of an OL. Then, a $\pi/2$-pulse $\hat{L}(\pi/2)$ is applied to atoms, where $\hat{L}(\pi/2)\binom{1}{0}=\frac{1}{\sqrt{2}}\binom{1}{1}$. After holding time $t_{OL}$ and another $\pi/2$-pulse, the state is: $\psi_{f}=\hat{L}(\pi/2)\hat{Q}(t_{OL})\hat{L}(\pi/2)\psi_{i},$ (3) where $\hat{L}(\alpha)=(\cos\frac{\alpha}{2}-i\sin\frac{\alpha}{2})\hat{\sigma}_{y}$. The operator $\hat{Q}(t_{OL})=(\cos\omega t+i\sin\omega t)\hat{\sigma}_{z}$ ($\omega$ corresponds to the energy gap between S and D bands at zeros quasi- momentum). Fig. 3(a) depicts the results $p_{D}(t_{OL})=N_{D}/(N_{S}+N_{D})$ at different $t_{OL}$ for the RI process ($\hat{R}(\pi/2)-\hat{U}(t_{OL})-\hat{R}(\pi/2)$). $N_{\mathrm{S}}$ ($N_{\mathrm{D}}$) represents the number of atoms in S-band (D-band). The period of the oscillation of $p_{D}$ is $41.1\pm 1.0\;\mu s$. This period related to the band gap and the theoretical value is $40.8\;\mu s$. From Fig. 3(a), we can see that the amplitude, or the contrast $C(t_{OL})$, decreases with the increase of $t_{OL}$, where $p_{D}(t_{OL})=[1+C(t_{OL})\cos(\omega t_{OL}+\phi)]/2.$ (4) We defined a characteristic time $\tau$, which corresponds to the time when the $C(t_{OL})$ decreases to $1/e$. Temperature can affect the length of $\tau$. #### 2.2.2 Contrast decay mechanisms To improve RI’s coherent time and performance, we should analyze the mechanisms that cause RI signal attenuation. By solving the Gross-Pitaevskii equation(GPE), which considers the mechanism that may lead to decay, we can get the process of contrast decay in theory. In Fig. 3(b), The following mechanisms are introduced in turn: the effect of the imperfection of the $\pi/2$ pulse (brown dashed line), inhomogeneity of laser wavefront (blue dotted line), the transverse expansion caused by the many-body interaction (blue dashed line), laser intensity fluctuation (the dash-dotted line), and the thermal fluctuations (the orange solid line). Fig. 3(b) illustrates that the theoretical (the orange solid line) and experimental (black dots) curves of the final result are very consistent. #### 2.2.3 An echo-Ramsey interferometer with motional Bloch states of BECs In order to extend the coherence time $\tau$, we proposed a quantum echo method. The echo process refers to a designed $\pi$ pulse ($\hat{L}(\pi)$) that flips the atomic populations of the two bands. So the evolution operator of Echo-RI is $\hat{L}(\pi/2)[\hat{Q}(t_{OL}/2n)\hat{L}(\pi)\hat{Q}(t_{OL}/2n)]^{n}\hat{L}(\pi/2)$, where $n$ is the number of the $\pi$ pulse inserted between the two $\pi/2$ pulses. Table 1: The effects for the contrast decay. Decay factor | Beam inhomogeneity | Echo recovery ---|---|--- $Dephasing$ | Momentum dispersion | Yes $Collision$ | Unbalance of population | Yes $Decoherence$ | fluctuation | No Fig. 3(c) illustrates the characteristic time $\tau$ for different $n$ and temperatures. And the effects for the contrast decay are listed in Table. 1. It can be seen from Fig. 3(c) that the interferometer with the longest characteristic time (14.5 ms) was obtained when $n\geq 6$ and $T=50$ nK. ## 3 Enhanced resolution by multimode scheme As an essential indicator, the resolution evaluates the performance of interferometers. The resolution is theoretically restricted to shot-noise limit, or sub-shot noise limit (?, ?), however, it will decay easily due to other experimental noises, with those upper limits beyond reach. Therefore, we have made several efforts to increase the resolution in practice. Improvement of the phase resolution was accomplished by a noise-resilient multi-component interferometric scheme. With the relative phase of different components remaining stable, increasing the number of paths could sharpen the peaks in the interference fringes, which leads to a resolution nearly twice compared with that of a conventional double-path two-mode interferometer with hardly any attenuation in visibility. Moreover, improvement of the momentum resolution is fulfilled with optical lattice pulses. Under the condition of 10 $\rm{E_{R}}$ OL depth, atomic momentum patterns with interval less than the double recoil momentum can be achieved, exhibiting 10 main peaks, respectively, where the minimum one we have given was 0.6 $\hbar k_{L}$. The demonstration of these techniques is shown in the next four subsections. ### 3.1 Time evolution of two-component Bose-Einstein condensates with a coupling drive For the multicomponent interferometer, it is necessary to study the interference characteristics of multi-component ultra-cold atoms. Here we introduced a basic method to deal with this problem, which simulates the time evolution of the relative phase in two-component Bose-Einstein condensates with a coupling drive (?). We considered a two-component Bose-Einstein condensate system with weak nonlinear interatomic interactions and coupling drive. In the formalism of the second quantization, the Hamiltonian of such a system can be written as $\hat{H}=\hat{H}_{1}+\hat{H}_{2}+\hat{H}_{int}+\hat{H}_{driv},$ (5) $\hat{H}_{i}=\int dx\Psi_{i}^{\dagger}(x)[-\frac{\hbar^{2}}{2m}\nabla^{2}+V_{i}(x)+U_{i}(x)\Psi_{i}^{\dagger}(x)\Psi_{i}(x)]\Psi_{i}(x),$ (6) $\hat{H}_{int}=U_{12}\int dx\Psi_{1}^{\dagger}(x)\Psi_{2}^{\dagger}(x)\Psi_{1}(x)\Psi_{2}(x),$ (7) $\hat{H}_{driv}=\int dx[\Psi_{1}^{\dagger}(x)\Psi_{2}(x)e^{i\omega_{rf}t}+\Psi_{1}(x)\Psi_{2}^{\dagger}(x)e^{-i\omega_{rf}t}],$ (8) where $i=1$ and $2$. Then the interference between two BEC’s is $I(t)=\frac{1}{2}N+\frac{1}{2}(N_{1}-N_{2})\cos{\omega_{rf}t}+\frac{1}{2}e^{-A(t)}\sin{\omega_{rf}t}\mathcal{R}(t).$ (9) Previous analysis can be used to simulate the time evolution of the relative phase in two-component Bose-Einstein condensates with a coupling drive, as well as to study the interference of multi-component ultra-cold atoms. This simulation would help to construct a multimode interferometer of a spinor BEC (see Subsecs. 3.3). ### 3.2 Parallel multicomponent interferometer with a spinor Bose-Einstein condensate Figure 4: (a) One typical interference picture. These spatial interference fringes come from the five sub-magnetic states of $\left|F=2\right\rangle$ hyperfine level. (b1-5) Density distributions corresponding to different sub- magnetic components respectively, where the points are the experimental data and the curves are fitting results according to the empirical expression (?, ?, ?). (c) Average of 15 consecutive experimental shots with a visibility reduction to zero for the chosen state $\left|m_{F}=-1\right\rangle$. Reproduced with permission from Ref. (?). Figure 5: (a1)-(a4) Histograms of relative phases distributions $(\phi_{2}-\phi_{1},\phi_{-2}-\phi_{-1},\phi_{2}-\phi_{-2},and\;\phi_{1}-\phi_{-1})$ respectively. These relative phases show good reproducibility, for the first two are concentrated at about $0^{o}$, while the latter two are concentrated at about $180^{o}$ in 61 consecutive experimental shots; (b) Relative phase distributions of 41 consecutive experimental shots with $t_{0}=3.6$ ms. Distributions of relative phases $\phi_{2}-\phi_{1}$ and $\phi_{1}-\phi_{-1}$ are shown in (b1) and (b2); (c)When $t_{0}=3.5$ ms, distributions of relative phases $\phi_{2}-\phi_{1}$ and $\phi_{1}-\phi_{-1}$ are shown in (c1) and (c2). The polar plots of relative phase vs visibility (shown as angle vs radius) are shown as these insets, respectively, where the value of visibility is an average of the visibility involved in calculation. Reproduced with permission from Ref. (?). Revealing the wave-particle duality, Young’s double-slit interference experiment plays a critical role in the foundation of modern physics. Other than quantum mechanical particles such as photons or electrons which had been proved in this stunning achievement, ultra-cold atoms with long coherent time have got the potential of precision measurements when utilizing this interferometric structure. Here we have demonstrated a parallel multi-state interferometer structure (?) in a higher spin atom system (?, ?, ?), which was achieved by using our spin-2 BEC of ${}^{87}\rm{Rb}$ atoms. The experimental scheme is described as following. After the manufacture of Bose-Einstein condensates in an optical-magnetic dipole trap, we switched off the optical harmonic trap and populate the condensates from $\left|F=2,m_{F}=2\right\rangle$ state to $\left|m_{F}=2\right\rangle$ and $\left|m_{F}=1\right\rangle$ sub-magnetic level equally. After the evolution in a gradient magnetic field for time $t_{1}$, these two wave packets were converted again into multiple $m_{F}$ states ($m_{F}=\pm 2,\pm 1,0$) as our spin states, leading to the so-called parallel path. All these states were allowed to evolve for another period time $t_{2}$, then the time-of-flight (TOF) stage $t_{3}$ for absorption imaging. Spatial interference fringes had been observed in all the spin channels. Here, we used the technique of spin projection with Majorana transition (?, ?, ?) by switching off the magnetic field pulses nonadiabatically to translate the atoms into different Zeeman sublevels. The spatial separation of atom cloud in different Zeeman states was reached by Stern-Gerlach momentum splitting in the gradient magnetic field. A typical picture after 26 ms TOF is shown in Fig. 4(a). Fig. 4(b1-b5) are the density distributions for each interference fringes. To reach the maximal visibility, we studied the correlation between the interference fringes’ visibility and the time interval applying Stern-Gerlach process. Though separated partially, the interfering wave pockets must overlap in a sort of way. The optimal visibility was about 0.6, corresponding to $t_{1}=210\;\rm{\mu s}$ and $t_{2}=1300\;\rm{\mu s}$. We also measured the fringe frequencies of different components, which exhibited a weak dependence on $m_{F}$. Special attention is required in Fig. 4(c). After an average of 15 consecutive CCD shots in repeated experiments, the interference fringe almost disappeared for the chosen state $\left|m_{F}=-1\right\rangle$. This result manifested the phase difference between the two copies of each component in every experimental run is evenly distributed. The poor phase repeatability could be attributed to uncontrollable phase accumulation in Majorana transitions. However, the relative phase across the spin components remained the same after more than 60 continuous experiments, just as Fig. 5(a) illustrates. Furthermore, evidence has been spotted that the relative phase can be controlled by changing the time $t_{0}$ before the first Majorana transition, as shown in Fig. 5(b)(c), paving a way towards noise-resilient multicomponent parallel interferometer or multi-pointer interferometric clocks (?). ### 3.3 Implementation of a double-path multimode interferometer using a spinor Bose-Einstein condensate Figure 6: (a1-a3) Single-shot spatial interference pattern with five interference modes after TOF = 26 ms. Fringes of each mode are (a1)in phase (a2)partially in phase (a3)complementary in space. (b1-b3) Black points are the experimental data by integrating the image in panels (a1-a3) along the z direction. Red solid lines are fitted by Thomas-Fermi Distribution (?). Visibilities are 0.55, 0.24, and 0.05, respectively. (c) Schematic of the spatial interference image. $\Delta\phi(T_{d},T_{N})$ is the relative phase between adjacent mode fringes. The fringe in each color represents the interference between the two wave packets of a single mode. Reproduced with permission from Ref. (?). Figure 7: Dependence of the visibility on the number of modes N and initial relative phase $\phi_{m_{F}}$ of the same mode in two paths. (a) Dependence on N in a situation that $\phi_{m_{F}}$ are all zero. FWHM of the N-mode fringe is 2/N times that of the two-mode fringe. (b) Dependence on $\phi_{m_{F}}$ using $N=4$ as an example. The green dashed line, red solid line, and purple dotted line show the fringes with $(\phi_{1},\phi_{2},\phi_{3},\phi_{4})=(0,0,0,0)$, $(0,0,\pi,\pi)$, and $(0.7\pi,0.2\pi,0.5\pi,\pi)$, respectively. Reproduced with permission from Ref. (?). The experiment described above was achieved by Stern-Gerlach momentum splitting, separating the wave pockets in different spin states or Zeeman sub- magnetic states in space. The conclusion that relative phases across the spin components remain stable gives us an inspiration to carry on the double-path multimode matter wave interferometer scheme. With the number of paths increased, it will suppress the noise and improve the resolution (?, ?, ?, ?, ?, ?) compared with the conventional double-path single-mode structure. The results show that resolution of the phase measurements is increased nearly twice in time domain interferometric fringes (?). The experimental procedure is similar to the previous one. The major difference lies in the splitting stage, during the optical harmonic trap participating in the preparation of the condensates is not going to switch off until the TOF stage, thus the Stern-Gerlach process in the gradient magnetic field mentioned above cannot significantly split the wave packets. With different momentum atomic clouds are spatially separated only for tens of nanometers, approximately 1% of the BEC size, thus well overlapped (?). As a result, multi-modes from two paths will interfere in one region instead of five. Another difference lies in the second spin projection with non-adiabatic Majorana transition. Here we replace it with a radio frequency pulse for its higher efficiency as a 1 to 5 beam splitter, although that we still use it to transfer the initial condensates into $\left|m_{F}=2\right\rangle$ and $\left|m_{F}=1\right\rangle$ sub-magnetic levels. The performance of Majorana transition is better than RF pulse as a 1 to 2 beam splitter. There are also some changes with experimental parameters that count a little and we would not discuss them here. Hence the global view of our interferometer is as follows: The magnetic sublevels are considered as modes in the interferometer, each has its own different phase evolution rates in gradient magnetic field. The double path configuration is made up of Majorana transition as well as the evolution of the first two $m_{F}$ superposed states during time $T_{d}$, makes up (path I, path II). RF pulse leading to the multiple $m_{F}$ superposed states together with their evolution in time $T_{N}$ forms the multi-modes configuration. During the TOF stage, atomic clouds expand and interfere with each other. Owing to the different state-dependent phase evolution rate $\omega_{m_{F}}^{(I,II)}$, the absorption image shows something more than spatial interference fringes, which is a periodic dependence of the visibility on phase evolution time as the function $V_{N}(T_{d},T_{N})$. We refer to it as the time domain interference. Fig. 6(a) shows a group of absorption images with various combinations of $T_{d}$, $T_{N}$. The observed fringe is a superposition of the interference fringes of different modes. Consequently, the visibility depends on the relative phase $\Delta\phi(T_{d},T_{N})$ between the interference fringes of each mode [Fig. 6(c)] and can also be modulated. By carefully analyzing with expression $V_{N}(T_{d},T_{N})=\langle\Psi^{(I)}|\Psi^{(II)}\rangle$[34], we can acquire the expression of the relative phase between two adjacent components: $\displaystyle\begin{split}\Delta\phi(T_{d},T_{N})&=(\Delta\phi_{m_{F}}-\Delta\theta_{m_{F}-1})\\\ &=\Delta\omega T_{N}+\Delta\theta\end{split}$ (10) where $\Delta\omega$ is the relative phase evolution rate between the two paths, $\Delta\theta$ is the relative initial phase introduced through the double path stage $T_{d}$. Yet we have already demonstrated that the visibility $V_{N}$ is modulated with the period $2\pi/\Delta\omega$ along with how the time domain fringe emerges theoretically. A remarkable feature of the multi-modes interferometer is the enhancement of resolution, which is defined as (fringe period)/(full width at half maximum). We have investigated the resolution of the time domain fringe experimentally and theoretically. It can be influenced by parameters like modes number and initial phase, which is $R(N,\phi_{m_{F}})$. $\phi_{m_{F}}$ refers to the initial relative phase of $m_{F}$ states accumulated in double paths $T_{d}$. Fig. 7 is the numerical results considering an arbitrary number of modes. Fig. 7(a) is under the condition that the phases $\phi_{m_{F}}$ are all the same for any modes. In that case, if we denote $\Delta\omega T_{N}=2n\pi/N$, then the visibility achieves $V_{N}=1$ when n is the multiple of N and a major peak is observed in this case. A remarkable feature of our interferometer is the enhancement of resolution by $N/2$ times without any changes in visibility nor periodical time. It is the harmonics that cause the peak width to decrease with the number of modes increasing in this case (?). Fig. 7(b) indicates $\phi_{m_{F}}$ varies from mode to mode for comparison. Neither the maximum visibility $V_{N}=1$ nor the minimum could be reached. Meanwhile, the time domain fringe shows more than one main peak in one period. Therefore, the initial phase $\phi_{m_{F}}$ needs to be well controlled to achieve the highest possible visibility and clear interference fringe in the time domain. We also experimentally study the time domain fringes. The experimental data (not depicted here) coincides with the numerical results of Fig. 7(b) red line, testifying its superiority to the resolution of the phase measurement. Moreover, the relative phase evolution rate $\Delta\omega$ can be controlled by adjusting the difference between the two paths accumulated in $T_{d}$ stage (?, ?, ?). With enhanced resolution, the sensitivity of interferometric measurements of physical observables can also be improved by properly assigning measurable quantities to the relative phase between two paths, as long as the modes do not interact with each other (?, ?, ?). ### 3.4 Atomic momentum patterns with narrower interval Figure 8: (a) Shortcut method for loading atoms: (a1) after the first two pulses and the $30$ ms holding time in the OL and the harmonic trap, the state becomes the superposition of the Bloch states in S-band with quasi-momenta taking the values throughout the FBZ, and is denoted by $\left|{\psi\left(0\right)}\right\rangle$. Then $1$ms band mapping is added. (a2) The single pulse acted on the superposed state $\left|{\psi\left(0\right)}\right\rangle$. (b) the superposed Bloch states of S-band spreading in the FBZ (black circles). The top Patterns in (c) and (d) are the TOF images in experiments. The lower part of (c) and (d) depicts the atomic distributions in experiments (red circles) and theoretical simulations (blue solid lines). There are seven peaks in (c) and ten peaks in (d) with $q=\pm 3\;\hbar k_{L}$. Reproduced with permission from Ref. (?). For ultra-cold atoms used in precise measurement, improving the precision of momentum manipulation is also conducive to improving the measurement resolution. The method to get atomic momentum patterns with narrower interval has been proposed and verified by experiments (?). Here we applied the shortcut pulse to realize the atomic momentum distribution with high resolutions for a superposed Bloch states spreading in the ground band of an OL. While difficult to prepare this superposition of Bloch states, it can be overcome by the shortcut method. First, the atoms are loaded in the superposition of S- and D-bands $(|S,q=0\rangle+|D,q=0\rangle)/\sqrt{2}$, where $q$ is the quasi-momentum. Fig. 8(a1) depicts the loading sequence. The atoms in S and D bands Collision between atoms in S and D bands will cause the atoms to gradually transfer to S band with non-zero quasi-momentum. After $30$ ms, as shown in Fig. 8(b), atoms cover the entire ground band from $q=-\hbar k_{L}$ to $\hbar k_{L}$. The momentum distribution of the initial state is a Gaussian-like shape. After an OL standing-wave pulse, which is similar to that in the shortcut process, the different patterns with the narrower interval can be obtained. The standing-wave pulse sequence is shown in Fig. 8(a2). Fig. 8(c) and (d) show the different designs for patterns of multi modes with various numbers of peaks under OL depth 10 $E_{r}$, where the top figures are the absorption images after the pulse and a $25$ ms TOF. The red circles are the experimental results of the atomic distribution along the x-axis from the TOF images. These results are very close to the numerical simulation result (blue lines). For the numerical simulation, we can get the initial superposition of states by fitting the experimental distributions (Fig. 8(b)). Fig. 8(c) and (d) depict the atomic momentum distribution with resolutions of $0.87\;\hbar k_{L}$ interval (seven peaks within $q=\pm 3\;\hbar k_{L}$) and $0.6\;\hbar k_{L}$ (ten peaks within $q=\pm 3\;\hbar k_{L}$) interval, respectively. The superposed states with different quasi-momenta in the ground band cause the narrow interval (far less than double recoil momentum) between peaks, which is useful to improve the resolution of atom interferometer (?). ## 4 Enhanced resolution by removing the systematic noise Noise identification as well as removal is crucial when extracting useful information in ultra-cold atoms absorption imaging. In general concept, systematic noises of cold atom experiments originate from two sources, one is the process of detection, such as optical absorption imaging; the other is the procedure of experiments, such as the instability of experimental parameters. Here we provided an OFRA scheme, reducing the noise to a level near the theoretical limit as $1/\sqrt{2}$ of the photo-shot noise. When applying the PCA, we found that the noise origins, which mainly come from the fluctuations of atom number and spatial positions, much fewer than the data dimensions of TOF absorption images. These images belong to BECs in one dimensional optical lattice, where the data dimension is actually the number of image pixels. If the raw TOF data can be preprocessed with normalization and adaptive region extraction methods, these noises can be remarkably attenuated or even wiped out. PCA of the preprocessed data exhibits a more subtle noises structure. When we compare the practical results with the numerical simulations, the few dominant noise components reveal a strong correlation with the experimental parameters. These encouraging results prove that the OFRA as well as PCA can be a promising tool for analysis in interferometry with higher precision (?, ?). ### 4.1 Optimized fringe removal algorithm for absorption images Optical absorption imaging is an important detection technique to obtain information from matter waves experiments. By comparing the recorded detection light field with the light field in the presence of absorption, we can easily attain the atoms’ spatial distribution. However, due to the inevitable differences between two recorded light field distributions, detection noises are unavoidable. Figure 9: Comparison between ordinary method and OFRA method. The integral of the atomic distribution in the red box in (a1) and (b1) correspond to (a2) and (b2). The atomic distribution (blue dots) is fitted by a bi-modal function to extract the temperature of atoms, which is shown in (c). Reproduced with permission from Ref. (?). Therefore, we have demonstrated an OFRA scheme to generate an ideal reference light field. With the algorithm, noise generated by the light field difference could be eliminated, leading to a noise close to the theoretical limit (?). The OFRA scheme is based on the PCA, we confirmed its validity by experiments of triangular optical lattices. The experimental configuration has been described in our prior work (?, ?). When the experiment was in process, the depth of the lattice was adiabatically raised to a final value, followed by a hold time of 20 ms to keep the atoms in the lattice potential before the optical absorption imaging. There are several parameters to characterize the triangle lattice system (?), among which the visibility, the condensate fraction, and the temperature matter. Fig. 9 (a2) and Fig. 9 (b2) are bimodal fitting to the scattering peaks by summing up the atomic distribution within the red box in the direction perpendicular to the center. Here Fig. 9 (a) stands for the common way of calculation and 9(b) for the OFRA. The bi-modal curve is composed of two parts: a Gaussian distribution for the thermal component and an inverse parabola curve for the condensed atoms. For each part, the column densities along the imaging axis can be written as $\displaystyle n_{th}(x)$ $\displaystyle=$ $\displaystyle\frac{n_{th}(0)}{g_{2}(1)}g_{2}[\exp(-(x-x_{0})^{2}/\sigma_{T}^{2})],$ (11) $\displaystyle n_{c}(x)$ $\displaystyle=$ $\displaystyle n_{c}(0)\max[1-\frac{(x-x_{0})^{2}}{\chi^{2}}].$ In the formula there are 5 parameters accounting for the bi-mode fitting, the amplitude of two components $n_{th}(0)$ and $n_{c}(0)$, the width of two components $\sigma_{T}$, $\chi$ and the center position $x_{0}$ of the atomic cloud. The Bose function is defined as $g_{j}(z)=\sum_{i}z^{i}/i^{j}$. In practice, we performed the least-squares fitting of $n_{th}(x)+n_{c}(x)$ to the real distribution obtained from the imaging. From the fit, we can get the atom number and width of the two components separately. Note that the measurement in Fig. 9 (9c) is performed at different lattice depths. For each lattice depth, 30 experiments have been performed to acquire the statistical results. The temperature is given as $T=1/2M\sigma_{T}^{2}/t_{TOF}^{2}/k_{B}$, where $M$ described the atom mass and $t_{TOF}$ width of the thermal part (?). For the number of condensed atoms, the fitting outcome is less affected by the fringe shown in Fig. 9 (a). Whereas the influence of the fringe on the fitting of temperature is much more evident. The temperature is proportional to the width of the Gaussian distribution $\sigma_{T}$ as mentioned above. Fig. 9(c) shows the temperature extracted from the TOF absorption images with and without the OFRA separately, namely Fig. 9(a1) and 9(b1). Fig. 9(a2) and 9(b2) are the corresponding integrated one-dimensional atomic distributions for each method. Fig. 9 depicts that the temperature we get with the common way of calculation has a large error of 400 nK, extraordinary higher than the initial BEC temperature of 90 nK. The turning on the procedure of lattice potential would indeed lead to a limited heating effect, nevertheless the proportion of condensed atom should be reduced significantly considering our system has been heated up by 4 times. This is still not consistent with the observation. However, the temperature is measured with much small variance at a much reasonable value if we dive into the OFRA scheme. For example, the measured temperature is 123.5 nK for a lattice depth of $V=4\;E_{r}$, with 183.9 nK for $V=9\;E_{r}$. Comparison between these two results illustrates that only by using the fringe removal algorithm we can get a reliable result, especially in the case of small atom numbers when fitting physical quantities such as the temperature. In conclusion, with this algorithm, we can measure parameters with higher contradiction to the conventional methods. The OFRA scheme is easy to implement in absorption imaging-based matter-wave experiments as well. There is no need to do any changes to the experimental system, only some algorithmic modifications matter. ### 4.2 Extraction and identification of noise patterns for ultracold atoms in an optical lattice Furthermore, on the basis of the absorption images after preprocessing by OFRA, the PCA method is used to identify the external noise fluctuation of the system caused by the imperfection of the experimental system. The noise can be reduced or even eliminated by the corresponding data processing program. It makes the task more difficult that these external systematic noises are often coupled, covered by nonlinear effects and a large number of pixels. PCA provides a good method to solve this problem (?, ?, ?, ?, ?). PCA can decompose the fluctuations in the experimental data into eigenmodes and provide an opportunity to separate the noises from different sources. For BEC in a one-dimensional OL, it was proved that PCA could be applied to the TOF images, where it successfully separated and recognizing noises from different main contribution sources, and reduced or even eliminated noises by data processing programs (?). The purpose of PCA is to use the smallest set of orthogonal vectors, called principal components (PCs) to approximate the variations of data while preserving the information of datasets as much as possible. The PCs correspond to the fluctuations of the experimental system, which can help to distinguish the main features of fluctuations. In the experimental system of BECs, the data are usually TOF images. A specific TOF image $A_{i}$ can be represented by the sum of the average value of the images and its fluctuation: $\displaystyle A_{i}=\bar{A}+\sum\varepsilon_{ij}P_{j},$ (12) Here $P_{j}$ is the different eigenmodes of fluctuation, $\varepsilon_{ij}$ is the weight of the eigenmodes $P_{j}$. Taking the BEC experiment (?, ?, ?, ?) in an OL as an example, we demonstrated the protocol of PCA for extraction noise. A TOF image for ultra-cold atoms in experiments can be represented as a $h\times w$ matrix. The PCA progress, shown in Fig. 10, will be applied to the images: Figure 10: Process diagram of PCA method. (a) Transform the $h\times w$ matrix (raw images) into a $1\times hw$ vector, denoted by $A_{i}$. And stack these vectors together. (b) Calculate the mean vector $\bar{A}=\frac{1}{n}\sum_{i=1}^{n}A_{i}$ and the fluctuations $\delta_{i}=A_{i}-\bar{A}$. Leaving only the fluctuation term in the matrix. (c) Stack $\delta_{i}$ together to form a matrix $X=\left[\delta_{1},\delta_{2},\cdots,\delta_{n}\right]$. Then the covariance matrix $S$ is obtained by $S=\frac{1}{n-1}X\cdot X^{T}$. (d) Decompose covariance matrix so that ${{V}^{-1}{S}{V}={D}}$ where D is a diagonal matrix. (e) Transform eigenvectors of interest back to a new TOF image. Reproduced with permission from Ref. (?). (1) Transform the $h\times w$ matrix into a $1\times hw$ vector, denoted by $A_{i}$. (2) Express $A_{i}$ as the sum of the average value $\bar{A}$ and the fluctuation $\delta_{i}$, where $\bar{A}=\frac{1}{n}\sum_{i=1}^{n}A_{i}$ and $\delta_{i}=A_{i}-\bar{A}$. (3) Stack $\delta_{i}$ together to form a matrix $X=\left[\delta_{1},\delta_{2},\cdots,\delta_{n}\right]$. Then the covariance matrix $S$ is obtained by $S=\frac{1}{n-1}X\cdot X^{T}$. (4) Decompose covariance matrix with ${{V}^{-1}{S}{V}={D}}$, where $V$ is the matrix of eigenvectors. Figure 11: PCA results of the TOF images. (a) Example of a raw TOF image. (b)-(f) correspond to the fluctuation in atom number (b), atom position (c)(d), peak width (e) and normal phase fraction (f), respectively. (a2), (b2), (d2), (e2), and (f2) are the integrated results of atom distributions along x direction. (c2) is for the atom distribution along z direction. The blue lines are the experimental results, and the orange lines are the simulation results. Reproduced with permission from Ref. (?). Fig. 11 shows the PCA results of the TOF images. Fig. 11(b)-(f) correspond to the first to fifth PCs, respectively. The first PC is from the fluctuation in the atom number. The normalization process can be applied to reducing the fluctuation in the atomic number. Because for a macroscopic wave function $\Psi{\left(\textbf{r}\right)}=\sqrt{N}\phi{\left(\textbf{r}\right)}$ (?), we usually concentrate on the relative density distribution, instead of the $\sqrt{N}$. After the normalization, the impact of this PC becomes very small. The second and third PCs correspond to the position fluctuations along the $z$\- and $x$-directions, respectively. The fluctuation in spatial position of the TOF images originates from the vibration of the system structure, such as the OL potential, trapping potential, and imaging system. We used a dynamic extraction method to eliminate the fluctuation in spatial position. We chose a region whose center is also the center of the density distribution. We first set a criterion to determine the center of the density distribution in the extraction area, and then used this center as the center of the new area to extract the new one. We repeated this process until the region to be extracted becomes stable. The fourth PC is from the fluctuation in the width of the Bragg peaks in the TOF images. The final PC shown in Fig. 11(f1) comes from the normal phase fraction fluctuation. By studying the first five feature images, we have identified the physical origins of several PCs leading to the main contributions. We numerically simulated this understanding using the GPE with external fluctuation terms, and got very consistent results (?). It is helpful to understand the physical origins of PCs in designing a pretreatment to reduce or even eliminate fluctuations in atom number, spatial position and other sources. Even in the absence of any knowledge of the system, the PCA method is very effective to analyze the noise, so that it can be applied to interferometers with higher precision (?, ?). ## 5 Proposal on gravity measurements Figure 12: Schematic of the experimental protocol. ${}^{87}\rm{Rb}$ atoms are evaporation cooled as a Bose-Einstein Condensate in the $\left|F=2,m_{F}=0\right\rangle$ initial state. However, they are transferred to $\left|F=1,m_{F}=0\right\rangle$ as the two arms of interferometry, followed by $P$ sequence of accelerate optical lattice pulse to maintain the atoms against gravity: When atoms fall to a velocity of $q\times v_{Recoil}$, they acquire a velocity of $2q\times v_{Recoil}$ upwards. The delay $T_{Bloch}$ is chosen as $T_{Bloch}=2qv_{Recoil}/g$ to eliminate the fall caused by gravity. Still, the probe beam should consist a laser light resonant with the $F=1$ ground state to upper levels, thus we can take absorption photos of $F=1$ population for analysis. Inertia measurements (?, ?, ?, ?, ?, ?, ?), especially those for gravity acceleration $g$, have always drawn lots of attention. Until now, the performance of atom interferometry has reached a sensitivity of $8\times 10^{-9}$ at 1 second (?, ?), pushing forward the determination of the Newtonian gravitational constant G (?, ?) or the verification of equivalence principle (?, ?). Yet the bulky size of these quantum sensors strictly restricts their application for on-site measurements. Therefore, based on our previous study of ultra-cold atoms in precision measurements, we intend to precisely measure the local gravity acceleration with our ${}^{87}\rm{Rb}$ Bose-Einstein Condensates in a small displacement. Note that this conception bases on previous research of Perrier Cladé (?). Fig. 12 illustrates the protocol of this BEC gravimeter. Instead of the Mach- Zender method $(\pi/2-\pi-\pi/2)$ widely used in free-falling or atomic fountain gravimeters, we utilize the Ramsey-Bordé approach by two pairs of $\pi/2$ Raman pulses to get a small volume. Application of the Doppler- sensitive Raman beam rather than the 6.8 GHz microwave field provides a far more efficient way to realize larger momentum splitting, which will significantly boost the interference resolution. Raman light pulse can also attain the effects of velocity distribution. Consequently, the first pair of $\pi/2$ pulses selects the initial velocity while the second pair can measure the final distribution. It should be noticed that right after the velocity selection step, a cleaning light pulse resonant with the $D_{2}$ line will shine on the condensate to clear away atoms remaining in $F=2$ state, leading to the two arms in interferometry. The critical feature lies in the evolving stage between the two pairs of $\pi/2$ pulses. By periodically inverting the velocity, the two arms shall replicate parabolic trajectory in a confined volume, without any decreasing of interrogation time. Choosing appropriate parameters, displacement of ultra- cold atoms can be limited in a few centimeters, at least 1 order less than that of a Mach-Zender gravimeter. This assumption is accomplished by a succession of Bloch oscillation (BO) in a pulsed accelerated optical lattice which transfers many photon recoils to the condensates (?, ?, ?, ?). Here the application of ultra-cold atoms instead of optical molasses (?) makes it more efficient when loading the atoms in the first Brillouin zone adiabatically, owing to their wave function consistency and narrow velocity distribution. This pulsed accelerated optical lattice should be manufactured along the direction of gravity with higher lattice depth, to minimize the effect of Landau-Zener Tunneling loss. To deduce the value of g, we may scan the evolving time between the two pairs of $\pi/2$ pulses, keeping the Raman frequency of each pair of $\pi/2$ pulses fixed. When the time interval equals $2Pqv_{Recoil}/g$, where $P$ is the number of pulses and $q$ is the number of recoil velocities($v_{Recoil}$) obtained by a single pulse (shown in Fig. 12), the two arms are in phase and the value of g can be extrapolated. Here the absorption image is used to extract the interference information, due to the number of atoms being one order less than that obtained by the conventional method. We also give a qualitative analysis of this compact BEC gravimeter. Besides its enormous potential in transportable instruments, prospective sensitivity maybe even better. This encouraging outlook can be attributed to a longer coherent time of ultra-cold atoms where the phase shift scales quadratically. The smaller range of movement possesses other superiorities. Systematic errors stemming from the gradients of residual magnetic fields and light fields become negligible, especially for Gouy phase and wave-front aberrations (?, ?, ?). Furthermore, the value of the gravity acceleration is averaged over a smaller height compared with the Mach-Zender ones. Finally, this vertical Bloch oscillation technique offers a remarkable ability to coherently and efficiently transfer photon momenta (?), though decoherence induced by the inhomogeneity of the optical lattice must be taken into consideration. In conclusion, we believe that this compact BEC gravimeter will have a sensitivity of a few tens of ${\mu}$Gal at least. The falling distance will be no more than 2 centimeters. Further improvement should be possible by performing atom chip-assisted BEC preparation as well as the interaction- suppressing mechanism (?). Gravity measurements with sub-${\mu}$Gal accuracies in miniaturized, robust devices are sure to come in the future. ## 6 Conclusion In summary, we review our recent experimental developments on the performance of interferometer with ultra-cold atoms. First, we demonstrated a method for effective preparation of a BEC in different bands of an optical lattice within a few tens of microseconds, reducing the loading time by up to three orders of magnitude as compared to adiabatic loading. Along with this shortcut method, a Ramsey interferometer with band echo technique is employed to atoms within an OL, enormously extend the coherence time by one order of magnitude. Efforts to boost the resolution with multimode scheme is made as well. Application of a noise-resilient multi-component interferometric scheme shows that increasing the number of paths could sharpen the peaks in the time-domain interference fringes, which leads to a resolution nearly twice compared with that of a conventional double-path two-mode interferometer. We can somehow boost the momentum resolution meanwhile. The patterns in the momentum space have got an interval far less than the double recoil momentum, where the narrowest one is given as $0.6\;\hbar k_{L}$. However, these advancements are inseparable with our endeavor to optimize data analysis based on the PCA. Extrinsic systematic noise for absorption imaging can be reduced efficiently. A scheme for potential compact gravimeter with ultra-cold atoms has been proposed. We believe it will tremendously shrink the size of a practical on-site instrument, promoting another widely used quantum-based technique. ## Acknowledgment This work is supported by the National Basic Research Program of China (Grant No. 2016YFA0301501), the National Natural Science Foundation of China (Grants No. 61727819, No. 11934002, No. 91736208, and No. 11920101004), and the Project funded by China Postdoctoral Science Foundation. ## References * 1. Victoria Xu, Matt Jaffe, Cristian D. Panda, Sofus L. Kristensen, Logan W. Clark, and Holger Müller. Probing gravity by holding atoms for 20 seconds. Science, 366(6466):745, 2019. * 2. N. R. Cooper, J. Dalibard, and I. B. Spielman. Topological bands for ultracold atoms. Rev. Mod. Phys., 91(1):015005, 2019. * 3. Linxiao Niu, Shengjie Jin, Xuzong Chen, Xiaopeng Li, and Xiaoji Zhou. Observation of a dynamical sliding phase superfluid with $p$-band bosons. Phys. Rev. Lett., 121(26):265301, 2018. * 4. Anton Mazurenko, Christie S. Chiu, Geoffrey Ji, Maxwell F. Parsons, Márton Kanász-Nagy, Richard Schmidt, Fabian Grusdt, Eugene Demler, Daniel Greif, and Markus Greiner. A cold-atom fermi-hubbard antiferromagnet. Nature, 545(7655):462, 2017. * 5. Franco Dalfovo, Stefano Giorgini, Lev P Pitaevskii, and Sandro Stringari. Theory of bose-einstein condensation in trapped gases. Rev. Mod. Phys., 71(3):463, 1999. * 6. Immanuel Bloch, Jean Dalibard, and Wilhelm Zwerger. Many-body physics with ultracold gases. Rev. Mod. Phys., 80(3):885, 2008. * 7. Stuart S. Szigeti, Samuel P. Nolan, John D. Close, and Simon A. Haine. High-precision quantum-enhanced gravimetry with a bose-einstein condensate. Phys. Rev. Lett., 125:100402, Sep 2020. * 8. K. S. Hardman, P. J. Everitt, G. D. McDonald, P. Manju, P. B. Wigley, M. A. Sooriyabandara, C. C. N. Kuhn, J. E. Debs, J. D. Close, and N. P. Robins. Simultaneous precision gravimetry and magnetic gradiometry with a bose-einstein condensate: A high precision, quantum sensor. Phys. Rev. Lett., 117(13):138501, 2016. * 9. Kyle S. Hardman, Carlos C. N. Kuhn, Gordon D. McDonald, John E. Debs, Shayne Bennetts, John D. Close, and Nicholas P. Robins. Role of source coherence in atom interferometry. Phys. Rev. A, 89(2):023626, 2014. * 10. Jun. Ye, Sebastian Blatt, Martin M. Boyd, Seth M. Foreman, Eric R. Hudson, Tetsuya Ido, Benjamin Lev, Andrew D. Ludlow, Brian C. Sawyer, Benjamin Stuhl, and Tanya Zelinsky. Precision measurement based on ultracold atoms and cold molecules. Int. J. Mod. Phys. D, 16(12b):2481, 2007. * 11. T. Berrada, S. van Frank, R. Bücker, T. Schumm, J. F. Schaff, and J. Schmiedmayer. Integrated mach-zehnder interferometer for bose-einstein condensates. Nat. Commun., 4(1):2077, 2013. * 12. Sydney Schreppler, Nicolas Spethmann, Nathan Brahms, Thierry Botter, Maryrose Barrios, and Dan M. Stamper-Kurn. Optically measuring force near the standard quantum limit. Science, 344(6191):1486, 2014. * 13. Wei Xiong, Xiaoji Zhou, Xuguang Yue, Xuzong Chen, Biao Wu, and Hongwei Xiong. Critical correlations in an ultra-cold bose gas revealed by means of a temporal talbot-lau interferometer. Laser Phys. Lett., 10(12):125502, nov 2013. * 14. Andrei Derevianko and Hidetoshi Katori. Colloquium: Physics of optical lattice clocks. Rev. Mod. Phys., 83(2):331, 2011. * 15. S. L. Campbell, R. B. Hutson, G. E. Marti, A. Goban, N. Darkwah Oppong, R. L. McNally, L. Sonderhouse, J. M. Robinson, W. Zhang, B. J. Bloom, and J. Ye. A fermi-degenerate three-dimensional optical lattice clock. Science, 358(6359):90, 2017. * 16. Xiaoji Zhou, Xia Xu, Xuzong Chen, and Jingbiao Chen. Magic wavelengths for terahertz clock transitions. Phys. Rev. A, 81:012115, Jan 2010. * 17. E. R. Moan, R. A. Horne, T. Arpornthip, Z. Luo, A. J. Fallon, S. J. Berl, and C. A. Sackett. Quantum rotation sensing with dual sagnac interferometers in an atom-optical waveguide. Phys. Rev. Lett., 124:120403, Mar 2020. * 18. A. Gauguet, B. Canuel, T. Lévèque, W. Chaibi, and A. Landragin. Characterization and limits of a cold-atom sagnac interferometer. Phys. Rev. A, 80:063604, Dec 2009. * 19. J. Le Gouët, T. E. Mehlstäubler, J. Kim, S. Merlet, A. Clairon, A. Landragin, and F. Pereira Dos Santos. Limits to the sensitivity of a low noise compact atomic gravimeter. Appl. Phys. B, 92(2):133, 2008. * 20. M. Schmidt, A. Senger, M. Hauth, C. Freier, V. Schkolnik, and A. Peters. A mobile high-precision absolute gravimeter based on atom interferometry. Gyroscopy Navig., 2(3):170, 2011. * 21. J. K. Stockton, K. Takase, and M. A. Kasevich. Absolute geodetic rotation measurement using atom interferometry. Phys. Rev. Lett., 107:133001, Sep 2011. * 22. G. Tackmann, P. Berg, C. Schubert, S. Abend, M. Gilowski, W. Ertmer, and E. M. Rasel. Self-alignment of a compact large-area atomic sagnac interferometer. New J. Phys., 14(1):015002, 2012. * 23. P. A. Altin, M. T. Johnsson, V. Negnevitsky, G. R. Dennis, R. P. Anderson, J. E. Debs, S. S. Szigeti, K. S. Hardman, S. Bennetts, G. D. McDonald, L. D. Turner, J. D. Close, and N. P. Robins. Precision atomic gravimeter based on bragg diffraction. New J. Phys., 15(2):023009, 2013. * 24. I. Carusotto, L. Pitaevskii, S. Stringari, G. Modugno, and M. Inguscio. Sensitive measurement of forces at the micron scale using bloch oscillations of ultracold atoms. Phys. Rev. Lett., 95:093202, Aug 2005. * 25. M Weitz, T Heupel, and TW Hänsch. Multiple beam atomic interferometer. Phys. Rev. Lett., 77(12):2356, 1996. * 26. Jovana Petrovic, Ivan Herrera, Pietro Lombardi, Florian Schaefer, and Francesco S Cataliotti. A multi-state interferometer on an atom chip. New J. Phys., 15(4):043002, 2013. * 27. Dong Hu, Linxiao Niu, Shengjie Jin, Xuzong Chen, Guangjiong Dong, Jörg Schmiedmayer, and Xiaoji Zhou. Ramsey interferometry with trapped motional quantum states. Commun. Phys., 1(1):29, 2018. * 28. Xiaoji Zhou, Shengjie Jin, and Jörg Schmiedmayer. Shortcut loading a bose-einstein condensate into an optical lattice. New J. Phys., 20(5):055005, 2018. * 29. Bo Lu, Thibault Vogt, Xinxing Liu, Xu Xu, Xiaoji Zhou, and Xuzong Chen. Cooperative scattering measurement of coherence in a spatially modulated bose gas. Phys. Rev. A, 83:051608, May 2011. * 30. Zhongkai Wang, Linxiao Niu, Peng Zhang, Mingxuan Wen, Zhen Fang, Xuzong Chen, and Xiaoji Zhou. Asymmetric superradiant scattering and abnormal mode amplification induced by atomic density distortion. Opt. Express, 21(12):14377, Jun 2013. * 31. Thibault Vogt, Bo Lu, XinXing Liu, Xu Xu, Xiaoji Zhou, and Xuzong Chen. Mode competition in superradiant scattering of matter waves. Phys. Rev. A, 83:053603, May 2011. * 32. Xiaoji Zhou, Fan Yang, Xuguang Yue, T. Vogt, and Xuzong Chen. Imprinting light phase on matter-wave gratings in superradiance scattering. Phys. Rev. A, 81:013615, Jan 2010. * 33. Xiaoji Zhou, Xu Xu, Lan Yin, W. M. Liu, and Xuzong Chen. Detecting quantum coherence of bose gases in optical lattices by scattering light intensity in cavity. Opt. Express, 18(15):15664, Jul 2010. * 34. Xiaoji Zhou, Jiageng Fu, and Xuzong Chen. High-order momentum modes by resonant superradiant scattering. Phys. Rev. A, 80:063608, Dec 2009. * 35. Xu Xu, Xiaoji Zhou, and Xuzong Chen. Spectroscopy of superradiant scattering from an array of bose-einstein condensates. Phys. Rev. A, 79:033605, Mar 2009. * 36. Juntao Li, Xiaoji Zhou, Fan Yang, and Xuzong Chen. Superradiant rayleigh scattering from a bose-einstein condensate with the incident laser along the long axis. Phys. Lett. A, 372(26):4750, 2008. * 37. Rui Guo, Xiaoji Zhou, and Xuzong Chen. Enhancement of motional entanglement of cold atoms by pairwise scattering of photons. Phys. Rev. A, 78:052107, Nov 2008. * 38. Fan Yang, Xiaoji Zhou, Juntao Li, Yuankai Chen, Lin Xia, and Xuzong Chen. Resonant sequential scattering in two-frequency-pumping superradiance from a bose-einstein condensate. Phys. Rev. A, 78:043611, Oct 2008. * 39. Bo Lu, Xiaoji Zhou, Thibault Vogt, Zhen Fang, and Xuzong Chen. Laser driving of superradiant scattering from a bose-einstein condensate at variable incidence angle. Phys. Rev. A, 83:033620, Mar 2011. * 40. Shumpei Masuda, Katsuhiro Nakamura, and Adolfo del Campo. High-fidelity rapid ground-state loading of an ultracold gas into an optical lattice. Phys. Rev. Lett., 113(6):063003, 2014. * 41. Xi Chen, A. Ruschhaupt, S. Schmidt, A. del Campo, D. Guéry-Odelin, and J. G. Muga. Fast optimal frictionless atom cooling in harmonic traps: Shortcut to adiabaticity. Phys. Rev. Lett., 104(6):063002, 2010. * 42. Xinxing Liu, Xiaoji Zhou, Wei Xiong, Thibault Vogt, and Xuzong Chen. Rapid nonadiabatic loading in an optical lattice. Phys. Rev. A, 83(6):063402, 2011. * 43. Yueyang Zhai, Xuguang Yue, Yanjiang Wu, Xuzong Chen, Peng Zhang, and Xiaoji Zhou. Effective preparation and collisional decay of atomic condensates in excited bands of an optical lattice. Phys. Rev. A, 87(6):063638, 2013. * 44. Xuguang Yue, Yueyang Zhai, Zhongkai Wang, Hongwei Xiong, Xuzong Chen, and Xiaoji Zhou. Observation of diffraction phases in matter-wave scattering. Phys. Rev. A, 88:013603, Jul 2013. * 45. Xinxing Liu, Xiaoji Zhou, Wei Zhang, Thibault Vogt, Bo Lu, Xuguang Yue, and Xuzong Chen. Exploring multiband excitations of interacting bose gases in a one-dimensional optical lattice by coherent scattering. Phys. Rev. A, 83:063604, Jun 2011. * 46. Yueyang Zhai, Peng Zhang, Xuzong Chen, Guangjiong Dong, and Xiaoji Zhou. Bragg diffraction of a matter wave driven by a pulsed nonuniform magnetic field. Phys. Rev. A, 88:053629, Nov 2013. * 47. Wei Xiong, Xuguang Yue, Zhongkai Wang, Xiaoji Zhou, and Xuzong Chen. Manipulating the momentum state of a condensate by sequences of standing-wave pulses. Phys. Rev. A, 84(4):043616, 2011. * 48. Zhongkai Wang, Baoguo Yang, Dong Hu, Xuzong Chen, Hongwei Xiong, Biao Wu, and Xiaoji Zhou. Observation of quantum dynamical oscillations of ultracold atoms in the $f$ and $d$ bands of an optical lattice. Phys. Rev. A, 94(3):033624, 2016. * 49. Baoguo Yang, Shengjie Jin, Xiangyu Dong, Zhe Liu, Lan Yin, and Xiaoji Zhou. Atomic momentum patterns with narrower intervals. Phys. Rev. A, 94(4):043607, 2016. * 50. Linxiao Niu, Dong Hu, Shengjie Jin, Xiangyu Dong, Xuzong Chen, and Xiaoji Zhou. Excitation of atoms in an optical lattice driven by polychromatic amplitude modulation. Opt. Express, 23(8):10064, 2015. * 51. Dong Hu, Linxiao Niu, Baoguo Yang, Xuzong Chen, Biao Wu, Hongwei Xiong, and Xiaoji Zhou. Long-time nonlinear dynamical evolution for $p$-band ultracold atoms in an optical lattice. Phys. Rev. A, 92(4):043614, 2015. * 52. Xinxin Guo, Wenjun Zhang, Zhihan Li, Hongmian Shui, Xuzong Chen, and Xiaoji Zhou. Asymmetric population of momentum distribution by quasi-periodically driving a triangular optical lattice. Opt. Express, 27(20):27786, 2019. * 53. Peng-Ju Tang, Peng Peng, Xiang-Yu Dong, Xu-Zong Chen, and Xiao-Ji Zhou. Implementation of full spin-state interferometer. Chinese Phys. Lett., 36(5):050301, 2019. * 54. Linxiao Niu, Pengju Tang, Baoguo Yang, Xuzong Chen, Biao Wu, and Xiaoji Zhou. Observation of quantum equilibration in dilute bose gases. Phys. Rev. A, 94(6):063603, 2016. * 55. Luca Pezzé and Augusto Smerzi. Entanglement, nonlinear dynamics, and the heisenberg limit. Phys. Rev. Lett., 102:100401, Mar 2009. * 56. B. Lücke, M. Scherer, J. Kruse, L. Pezzé, F. Deuretzbacher, P. Hyllus, O. Topic, J. Peise, W. Ertmer, J. Arlt, L. Santos, A. Smerzi, and C. Klempt. Twin matter waves for interferometry beyond the classical limit. Science, 334(6057):773, 2011. * 57. Wei-Dong Li, X. J. Zhou, Y. Q. Wang, J. Q. Liang, and Wu-Ming Liu. Time evolution of the relative phase in two-component bose-einstein condensates with a coupling drive. Phys. Rev. A, 64:015602, Jun 2001. * 58. J. E. Simsarian, J. Denschlag, Mark Edwards, Charles W. Clark, L. Deng, E. W. Hagley, K. Helmerson, S. L. Rolston, and W. D. Phillips. Imaging the phase of an evolving bose-einstein condensate wave function. Phys. Rev. Lett., 85:2040, Sep 2000. * 59. Shimon Machluf, Yonathan Japha, and Ron Folman. Coherent stern-gerlach momentum splitting on an atom chip. Nat. Commun., 4:2424, 2013. * 60. Yair Margalit, Zhifan Zhou, Shimon Machluf, Daniel Rohrlich, Yonathan Japha, and Ron Folman. A self-interfering clock as a ”which path” witness. Science, 349(6253):1205, 2015. * 61. Pengju Tang, Peng Peng, Zhihan Li, Xuzong Chen, Xiaopeng Li, and Xiaoji Zhou. Parallel multicomponent interferometer with a spinor bose-einstein condensate. Phys. Rev. A, 100:013618, Jul 2019. * 62. Meng Han, Peipei Ge, Yun Shao, Qihuang Gong, and Yunquan Liu. Attoclock photoelectron interferometry with two-color corotating circular fields to probe the phase and the amplitude of emitting wave packets. Phys. Rev. Lett., 120:073202, Feb 2018. * 63. Alexander D. Cronin, Jörg Schmiedmayer, and David E. Pritchard. Optics and interferometry with atoms and molecules. Rev. Mod. Phys., 81:1051, Jul 2009. * 64. M. R. Andrews, C. G. Townsend, H.-J. Miesner, D. S. Durfee, D. M. Kurn, and W. Ketterle. Observation of interference between two bose condensates. Science, 275(5300):637, 1997. * 65. Xiuquan Ma, Lin Xia, Fang Yang, Xiaoji Zhou, Yiqiu Wang, Hong Guo, and Xuzong Chen. Population oscillation of the multicomponent spinor bose-einstein condensate induced by nonadiabatic transitions. Phys. Rev. A, 73:013624, Jan 2006. * 66. E. Majorana. Atomi orientati in campo magnetico variabile. Nuovo Cimento, 9:43, 1932. * 67. Lin Xia, Xu Xu, Rui Guo, Fan Yang, Wei Xiong, Juntao Li, Qianli Ma, Xiaoji Zhou, Hong Guo, and Xuzong Chen. Manipulation of the quantum state by the majorana transition in spinor bose-einstein condensates. Phys. Rev. A, 77:043622, Apr 2008. * 68. Pengju Tang, Xiangyu Dong, Wenjun Zhang, Yunhong Li, Xuzong Chen, and Xiaoji Zhou. Implementation of a double-path multimode interferometer using a spinor bose-einstein condensate. Phys. Rev. A, 101:013612, Jan 2020. * 69. Gregor Weihs, Michael Reck, Harald Weinfurter, and Anton Zeilinger. All-fiber three-path mach–zehnder interferometer. Opt. Lett., 21(4):302, Feb 1996. * 70. M. W. Mitchell, J. S. Lundeen, and A. M. Steinberg. Super-resolving phase measurements with a multiphoton entangled state. Nature, 429:161, May 2004. * 71. J. Chwedeńczuk, F. Piazza, and A. Smerzi. Multipath interferometer with ultracold atoms trapped in an optical lattice. Phys. Rev. A, 87:033607, Mar 2013. * 72. Tania Paul and Tabish Qureshi. Measuring quantum coherence in multislit interference. Phys. Rev. A, 95(4):042110, 2017. * 73. Igor Pikovski, Magdalena Zych, Fabio Costa, and Časlav Brukner. Time dilation in quantum systems and decoherence. New J. Phys., 19(2):025011, 2017. * 74. C Fort, P Maddaloni, F Minardi, M Modugno, and M Inguscio. Spatial interference of coherent atomic waves by manipulation of the internal quantum state. Opt. Lett., 26(14):1039, 2001. * 75. Jonas Söderholm, Gunnar Björk, Björn Hessmo, and Shuichiro Inoue. Quantum limits on phase-shift detection using multimode interferometers. Phys. Rev. A, 67:053803, May 2003. * 76. J. Chwedeńczuk. Quantum interferometry in multimode systems. Phys. Rev. A, 96:032320, Sep 2017. * 77. Stephen R. Segal, Quentin Diot, Eric A. Cornell, Alex A. Zozulya, and Dana Z. Anderson. Revealing buried information: Statistical processing techniques for ultracold-gas image analysis. Phys. Rev. A, 81:053601, May 2010. * 78. Sheng-wey Chiow, Tim Kovachy, Hui-Chun Chien, and Mark A. Kasevich. $102\hbar k$ large area atom interferometers. Phys. Rev. Lett., 107:130403, Sep 2011. * 79. Linxiao Niu, Xinxin Guo, Yuan Zhan, Xuzong Chen, Wuming Liu, and Xiaoji Zhou. Optimized fringe removal algorithm for absorption images. Appl. Phys. Lett., 113(14):144103, 2018. * 80. Shengjie Jin, Xinxin Guo, Peng Peng, Xuzong Chen, Xiaopeng Li, and Xiaoji Zhou. Finite temperature phase transition in a cross-dimensional triangular lattice. New J. Phys., 21(7):073015, 2019. * 81. C. Becker, P. Soltan-Panahi, J. Kronjäger, S. Dörscher, K. Bongs, and K. Sengstock. Ultracold quantum gases in triangular optical lattices. New J. Phys., 12(6):065025, 2010. * 82. M. Gatzke, G. Birkl, P. S. Jessen, A. Kastberg, S. L. Rolston, and W. D. Phillips. Temperature and localization of atoms in three-dimensional optical lattices. Phys. Rev. A, 55:R3987, Jun 1997. * 83. Martin D Levine. Feature extraction: A survey. Proc. IEEE, 57(8):1391, 1969. * 84. Romain Dubessy, Camilla De Rossi, Thomas Badr, Laurent Longchambon, and Helene Perrin. Imaging the collective excitations of an ultracold gas using statistical correlations. New J. Phys., 16:122001, Dec 2014. * 85. Andrea Alberti, Carsten Robens, Wolfgang Alt, Stefan Brakhane, Michal Karski, Rene Reimann, Artur Widera, and Dieter Meschede. Super-resolution microscopy of single atoms in optical lattices. New J. Phys., 18:053010, May 2016. * 86. C. F. Ockeloen, A. F. Tauschinsky, R. J. C. Spreeuw, and S. Whitlock. Detection of small atom numbers through image processing. Phys. Rev. A, 82:061606, Dec 2010. * 87. Shuyang Cao, Pengju Tang, Xinxin Guo, Xuzong Chen, Wei Zhang, and Xiaoji Zhou. Extraction and identification of noise patterns for ultracold atoms in an optical lattice. Opt. Express, 27(9):12710, 2019. * 88. Oliver Penrose and Lars Onsager. Bose-einstein condensation and liquid helium. Phys. Rev., 104:576, Nov 1956. * 89. T. L. Gustavson, A. Landragin, and M. A. Kasevich. Rotation sensing with a dual atom-interferometer sagnac gyroscope. Classical Quant. Grav., 17(12):2385, 2000. * 90. Holger Müller, Sheng-wey Chiow, Sven Herrmann, Steven Chu, and Keng-Yeow Chung. Atom-interferometry tests of the isotropy of post-newtonian gravity. Phys. Rev. Lett., 100:031101, Jan 2008. * 91. Anne Louchet-Chauvet, Tristan Farah, Quentin Bodart, André Clairon, Arnaud Landragin, Sébastien Merlet, and Franck Pereira Dos Santos. The influence of transverse motion within an atomic gravimeter. New J. Phys., 13(6):065025, 2011. * 92. J. B. Fixler, G. T. Foster, J. M. McGuirk, and M. A. Kasevich. Atom interferometer measurement of the newtonian constant of gravity. Science, 315(5808):74, 2007. * 93. G. Lamporesi, A. Bertoldi, L. Cacciapuoti, M. Prevedelli, and G. M. Tino. Determination of the newtonian gravitational constant using atom interferometry. Phys. Rev. Lett., 100:050801, Feb 2008. * 94. Michael A. Hohensee, Holger Müller, and R. B. Wiringa. Equivalence principle and bound kinetic energy. Phys. Rev. Lett., 111:151102, Oct 2013. * 95. Michael A. Hohensee, Steven Chu, Achim Peters, and Holger Müller. Equivalence principle and gravitational redshift. Phys. Rev. Lett., 106:151102, Apr 2011. * 96. Manuel Andia, Raphaël Jannin, Francois Nez, Francois Biraben, Saïda Guellati-Khélifa, and Pierre Cladé. Compact atomic gravimeter based on a pulsed and accelerated optical lattice. Phys. Rev. A, 88:031605, Sep 2013. * 97. Ekkehard Peik, Maxime Ben Dahan, Isabelle Bouchoule, Yvan Castin, and Christophe Salomon. Bloch oscillations of atoms, adiabatic rapid passage, and monokinetic atomic beams. Phys. Rev. A, 55:2989, Apr 1997. * 98. Maxime Ben Dahan, Ekkehard Peik, Jakob Reichel, Yvan Castin, and Christophe Salomon. Bloch oscillations of atoms in an optical potential. Phys. Rev. Lett., 76:4508, Jun 1996. * 99. S. R. Wilkinson, C. F. Bharucha, K. W. Madison, Qian Niu, and M. G. Raizen. Observation of atomic wannier-stark ladders in an accelerating optical potential. Phys. Rev. Lett., 76:4512, Jun 1996. * 100. Rémy Battesti, Pierre Cladé, Saïda Guellati-Khélifa, Catherine Schwob, Benoît Grémaud, François Nez, Lucile Julien, and François Biraben. Bloch oscillations of ultracold atoms: A tool for a metrological determination of $h/{m}_{\mathrm{rb}}$. Phys. Rev. Lett., 92:253001, Jun 2004. * 101. A. Peters, K. Y. Chung, and S. Chu. High-precision gravity measurements using atom interferometry. Metrologia, 38(1):25, 2001. * 102. Pierre Cladé, Estefania de Mirandes, Malo Cadoret, Saïda Guellati-Khélifa, Catherine Schwob, François Nez, Lucile Julien, and François Biraben. Precise measurement of $h/{m}_{\mathrm{rb}}$ using bloch oscillations in a vertical optical lattice: Determination of the fine-structure constant. Phys. Rev. A, 74:052109, Nov 2006. * 103. S. Abend, M. Gebbe, M. Gersemann, H. Ahlers, H. Müntinga, E. Giese, N. Gaaloul, C. Schubert, C. Lämmerzahl, W. Ertmer, W. P. Schleich, and E. M. Rasel. Atom-chip fountain gravimeter. Phys. Rev. Lett., 117:203003, Nov 2016.
# Re-evaluation of Spin-Orbit Dynamics of Polarized $e^{+}e^{-}$ Beams in High Energy Circular Accelerators and Storage Rings: an approach based on a Bloch equation††thanks: Based on a talk at IAS, Hong Kong, January 17, 2019. Also available as article: _Int. J. Mod. Phys._ , vol. A35, Nos. 15 & 16, 2041003, 2020. Moreover available as DESY Report 20-137. ###### Abstract We give an overview of our current/future analytical and numerical work on the spin polarization in high-energy electron storage rings. Our goal is to study the possibility of polarization for the CEPC and FCC-ee. Our work is based on the so-called Bloch equation for the polarization density introduced by Derbenev and Kondratenko in 1975. We also give an outline of the standard approach, the latter being based on the Derbenev-Kondratenko formulas. Keywords: electron storage rings, spin-polarized beams, polarization density, FCC, CEPC, stochastic . differential equations, method of averaging. Klaus Heinemann 111Corresponding author. Department of Mathematics and Statistics, University of New Mexico, Albuquerque, NM 87131, USA <EMAIL_ADDRESS> Daniel Appelö 222Now at Michigan State University, USA. (email: <EMAIL_ADDRESS> Department of Applied Mathematics, University of Colorado Boulder, Boulder, CO 80309-0526, USA <EMAIL_ADDRESS> Desmond P. Barber Deutsches Elektronen-Synchrotron (DESY) Hamburg, 22607, Germany and: Department of Mathematics and Statistics, University of New Mexico Albuquerque, NM 87131, USA <EMAIL_ADDRESS> Oleksii Beznosov Department of Mathematics and Statistics, University of New Mexico, Albuquerque, NM 87131, USA <EMAIL_ADDRESS> James A. Ellison Department of Mathematics and Statistics, University of New Mexico, Albuquerque, NM 87131, USA <EMAIL_ADDRESS> PACS numbers:29.20.db,29.27.Hj,05.10.Gg ###### Contents 1. 1 Introduction 2. 2 Sketching the standard approach based on the Derbenev-Kondratenko formulas 3. 3 The Bloch equation and the Reduced Bloch equation in the laboratory frame 4. 4 The Reduced Bloch equation in the beam frame 5. 5 The Effective Reduced Bloch equation in the beam frame 6. 6 Next steps 7. 7 Acknowledgement ## 1 Introduction This paper is an update on a talk by K. Heinemann at the IAS Mini-Workshop on Beam Polarization in Future Colliders on January 17, 2019, in Hong Kong [1]. Our ultimate goal is to examine the possibility of high polarization for CEPC and FCC-ee. We will first briefly review the “standard” approach which is based on the Derbenev-Kondratenko formulas [2]. These formulas rely, in part, on plausible assumptions grounded in deep physical intuition. So the following question arises: do the Derbenev-Kondratenko formulas tell full story? In fact there is an alternative approach based on a Bloch-type equation for the polarization density [3] which we call 333Note that in previous work we sometimes called it the full Bloch equation. the Bloch equation (BE) and which we believe can deliver more information than the standard approach even if the latter includes potential correction terms [4]. So we aim to determine the domain of applicability of the Derbenev-Kondratenko formulas and the possibility in theory of polarization at the CEPC and FCC-ee energies. Of course both approaches focus on the equilibrium polarization and the polarization time. We use the name “Bloch” to reflect the analogy with equations for magnetization in condensed matter [5]. This paper concentrates on the Bloch approach. The cost of the numerical computations in the Bloch approach is considerable since the polarization density depends on six phase-space variables plus the time variable so that the numerical solution of the BE, the BE being a system of three PDEs in seven independent variables, is a nontrivial task which cannot be pursued with traditional approaches like the finite difference method. However we see at least five viable methods: 1. 1. Approximating the BE by an effective BE via the Method of Averaging and solving the effective BE via spectral phase-space discretization, e.g., a collocation method, plus an implicit-explicit time discretization. 2. 2. Solving the system of stochastic differential equations (SDEs), which underlies the BE, via Monte-Carlo spin tracking. See Ref. 6 for the system of SDEs underlying the BE. 3. 3. Solving the Fokker-Planck equation, which underlies the BE, via the Gram- Charlier method. 4. 4. Solving the BE via a deep learning method. 5. 5. Solving the system of SDEs in a way that allows connections with the Derbenev- Kondratenko formulas to be established. We will dwell on Method 1 in this paper. We plan to validate this method by one of the other four methods. More details on Method 1 can be found in Ref. 6. The method of averaging we use is discussed in Refs. 7-12. One hope tied to Method 1 is that the effective BE gives analytical insights into the spin- resonance structure of the bunch. Note that Methods 1-4 are independent of the standard approach. In particular they do not rely on the invariant spin field. Note also that Methods 1-3 and 5 are based on knowing the system of SDEs, which underlies the BE. For details of this system of SDEs, see the invited ICAP18 paper of Ref. 6. Regarding Method 2 there is a large literature on the numerical solution of SDEs, see Refs. 13, 14 and references in Ref. 15. By neglecting the spin-flip terms and the kinetic-polarization term in the BE one obtains an equation that we call the Reduced Bloch equation (RBE). The RBE approximation is sufficient for computing the radiative depolarization rate due to stochastic orbital effects and it shares the terms with the BE that are challenging to discretize. For details on our phase-space discretization and time discretization of the RBE, see Refs. 6,16,17 and 18. We proceed as follows. In Section 2 we sketch the standard approach. In Section 3 we present, for the laboratory frame, the BE and its restriction, the RBE. In Section 4 we discuss the RBE in the beam frame and in Section 5 we show how, in the beam frame, the effective RBE is obtained via the method of averaging. In Section 6 we describe ongoing and future work. ## 2 Sketching the standard approach based on the Derbenev-Kondratenko formulas We define the “time” $\theta=2\pi s/C$ where $s$ is the distance around the ring and $C$ is the circumference. We denote by $y$ a position in six- dimensional phase space of accelerator coordinates which we call beam-frame coordinates. In particular, following Ref. 19, $y_{6}$ is the relative deviation of the energy from the reference energy. Then if, $f=f(\theta,y)$ denotes the normalized $2\pi$-periodic equilibrium phase-space density at $\theta$ and $y$ and $\vec{P}_{loc}=\vec{P}_{loc}(\theta,y)$ denotes the local polarization vector of the bunch we have $\displaystyle\int dy\;f(\theta,y)=1\;,\quad\int dy\;f(\theta,y)\vec{P}_{loc}(\theta,y)=\vec{P}(\theta)\;,$ (1) where $\vec{P}(\theta)$ is the polarization vector of the bunch at $\theta$. For a detailed discussion about $\vec{P}_{loc}$, see, e.g., Ref. 20. Here and in the following we use arrows on three-component column vectors. Central to the standard approach is the invariant spin field (ISF) $\hat{n}=\hat{n}(\theta,y)$ defined as a normalized periodic solution of the Thomas-BMT-equation in phase space, i.e., $\displaystyle\partial_{\theta}\hat{n}=L_{\rm Liou}(\theta,y)\hat{n}+\Omega(\theta,y)\hat{n}\;,$ (2) such that 1. 1. $\Big{|}\hat{n}(\theta,y)\Big{|}=1$, 2. 2. $\hat{n}(\theta+2\pi,y)=\hat{n}(\theta,y)$, and where $L_{\rm Liou}$ denotes the Hamiltonian part of the Fokker-Planck operator $L_{\rm FP}^{y}$, the latter being introduced in Section 3 below. For some of our work on the ISF see Refs. 21 and 22. The unit vector of the ISF on the closed orbit is denoted by $\hat{n}_{0}(\theta)$ and it is easily obtained as an eigenvector of the one-turn spin-transport map on the closed orbit [19]. There are many methods for computing the ISF but none are trivial (for a recent technique see Ref. 23). In fact the existence, in general, of the invariant spin field is a mathematical issue which is only partially resolved, see, e.g., Ref. 21. The standard approach assumes that a function $P_{\rm DK}=P_{\rm DK}(\theta)$ exists such that $\displaystyle\vec{P}_{loc}(\theta,y)\approx P_{\rm DK}(\theta)\hat{n}(\theta,y)\;.$ (3) Thus, by (1) and (3), $\displaystyle\vec{P}(\theta)=\int dy\;f(\theta,y)\vec{P}_{loc}(\theta,y)\approx P_{\rm DK}(\theta)\int dy\;f(\theta,y)\hat{n}(\theta,y)\;.$ (4) The approximation (3) leads to [2] $\displaystyle P_{\rm DK}(\theta)=P_{\rm DK}(\infty)(1-e^{-\theta/\tau_{\rm DK}})+P_{\rm DK}(0)e^{-\theta/\tau_{\rm DK}}\;,$ (5) where $\tau_{\rm DK}$ and $P_{\rm DK}(\infty)$ are given by the Derbenev- Kondratenko formulas $\displaystyle P_{\rm DK}(\infty):=\frac{\tau_{0}^{-1}}{\tau_{\rm DK}^{-1}}\;,$ (6) $\displaystyle\tau_{\rm DK}^{-1}:=\frac{5\sqrt{3}}{8}\frac{r_{e}\gamma_{0}^{5}\hbar}{m}\frac{C}{4\pi^{2}}\int_{0}^{2\pi}\;d\theta\Big{\langle}\frac{1}{|R|^{3}}[1-\frac{2}{9}(\hat{n}\cdot{\hat{\beta}})^{2}+\frac{11}{18}\Big{|}\partial_{y_{6}}\hat{n}\Big{|}^{2}]\Big{\rangle}_{\theta}\;,$ (7) $\displaystyle\tau_{0}^{-1}:=\frac{r_{e}\gamma_{0}^{5}\hbar}{m}\frac{C}{4\pi^{2}}\int_{0}^{2\pi}\;d\theta\Big{\langle}\frac{1}{|R|^{3}}\hat{b}\cdot[\hat{n}-\partial_{y_{6}}\hat{n}]\Big{\rangle}_{\theta}\;,$ (8) with * • $\Big{\langle}\cdots\Big{\rangle}_{\theta}\equiv\int dy\;f(\theta,y)\cdots$ * • $\hat{b}=\hat{b}(\theta,y)\equiv$ normalized magnetic field, $\hat{\beta}=\hat{\beta}(\theta,y)\equiv$ normalized velocity vector, $\gamma_{0}\equiv$ Lorentz factor of the reference particle, $R(\theta,y)\equiv$ radius of curvature in the external magnetic field, $r_{e}\equiv$ classical electron radius, $m\equiv$ rest mass of electrons or positrons. By (4) and for large $\theta$ $\displaystyle\vec{P}(\theta)\approx P_{\rm DK}(\infty)\int dy\;f(\theta,y)\hat{n}(\theta,y)\;,$ (9) where $P_{\rm DK}(\infty)$ is given by (6) and where the rhs of (9) is the approximate equilibrium polarization vector. Note that the latter is $2\pi$-periodic in $\theta$ since $f(\theta,y)$ and $\hat{n}(\theta,y)$ are $2\pi$-periodic in $\theta$. Defining $\displaystyle\tau_{dep}^{-1}:=\frac{5\sqrt{3}}{8}\frac{r_{e}\gamma_{0}^{5}\hbar}{m}\frac{C}{4\pi^{2}}\int_{0}^{2\pi}\;d\theta\Big{\langle}\frac{1}{|R|^{3}}\frac{11}{18}\Big{|}\partial_{y_{6}}\hat{n}\Big{|}^{2}\Big{\rangle}_{\theta}\;,$ (10) we can write (7) as $\displaystyle\tau_{\rm DK}^{-1}=\tau_{dep}^{-1}+\frac{5\sqrt{3}}{8}\frac{r_{e}\gamma_{0}^{5}\hbar}{m}\frac{C}{4\pi^{2}}\int_{0}^{2\pi}\;d\theta\Big{\langle}\frac{1}{|R|^{3}}[1-\frac{2}{9}(\hat{n}\cdot{\hat{\beta}})^{2}]\Big{\rangle}_{\theta}\;.$ (11) For details on (6), (7), (8), (10) and (11) see, e.g., Refs. 24 and 19. We now briefly characterize the various terms in the Derbenev-Kondratenko formulas. First, $\tau_{dep}^{-1}$ is the radiative depolarization rate. Secondly, the term $\frac{r_{e}\gamma_{0}^{5}\hbar}{m}\frac{C}{4\pi^{2}}\int_{0}^{2\pi}\;d\theta\Big{\langle}\frac{1}{|R|^{3}}\hat{b}\cdot\hat{n}\Big{\rangle}_{\theta}$ in $\tau_{0}^{-1}$ and the term $\frac{5\sqrt{3}}{8}\frac{r_{e}\gamma_{0}^{5}\hbar}{m}\frac{C}{4\pi^{2}}\int_{0}^{2\pi}\;d\theta\Big{\langle}\frac{1}{|R|^{3}}\Big{\rangle}_{\theta}$ in $\tau_{\rm DK}^{-1}$ cover the Sokolov-Ternov effect. Lastly, the term $-\frac{r_{e}\gamma_{0}^{5}\hbar}{m}\frac{C}{4\pi^{2}}\int_{0}^{2\pi}\;d\theta\Big{\langle}\frac{1}{|R|^{3}}\hat{b}\cdot[\partial_{y_{6}}\hat{n}]\Big{\rangle}_{\theta}$ in $\tau_{0}^{-1}$ covers the kinetic polarization effect and the term in $\tau_{\rm DK}^{-1}$ which is proportional to $2/9$ covers the Baier-Katkov correction. We now sketch three approaches for computing $P_{\rm DK}(\infty)$ via the Derbenev-Kondratenko formulas. All three approaches use (6) but they differ in how $\tau_{0}^{-1}$ and $\tau_{\rm DK}^{-1}$ are computed. * (i) Compute $\tau_{0}^{-1}$ via (8) and $\tau_{\rm DK}^{-1}$ via (7) by computing $f$ and $\hat{n}$ as accurately as needed. * (ii) Approximate $\tau_{0}^{-1}$ by neglecting the usually-small kinetic polarization term in (8) and by approximating the remaining term in (8) by replacing $\hat{n}$ by $\hat{n}_{0}$. Compute $\tau_{\rm DK}^{-1}$ via (11) where $\tau_{dep}^{-1}$ is not computed via (10) but via Monte-Carlo spin tracking and where the remaining terms in (11) are approximated by using the $\hat{n}_{0}$-axis. 444Prominent Monte-Carlo spin tracking codes are SLICKTRACK by D.P. Barber [19], SITROS by J. Kewisch [19], Zgoubi by F. Meot [25], PTC/FPP by E. Forest [26], and Bmad by D. Sagan [27]. This approach provides a useful first impression avoiding the computation of $f$ and $\hat{n}$. For more details on this approach see Ref. 19. Monte-Carlo tracking can also be extended beyond integrable orbital motion to include, as just one example, beam-beam forces. Note that Monte-Carlo tracking just gives an estimate of $\tau_{dep}^{-1}$ but it does not provide an explanation. Nevertheless, insights into sources of depolarization can be obtained by switching off terms in the Thomas-BMT equation. In principal such diagnoses can also be applied in approach (i). Such investigations can the systematized under the heading of “spin matching” [19]. * (iii) Compute $\tau_{0}^{-1}$ via (8) and $\tau_{\rm DK}^{-1}$ via (7) by linear approximation in orbit and spin variables via the so-called SLIM formalism [19]. Approach (ii) is the most practiced while approach (i) is only feasible if one can compute $f$ and $\hat{n}$ as accurately as needed (which is not easy!). Approach (iii), which was historically the first, is very simple and is often used for ballparking $P_{\rm DK}(\infty)$. Since the inception of the Derbenev-Kondratenko formulas correction terms to the rhs of (10) have been suspected. See Refs. 4, 28 as well as Z. Duan’s contribution to this workshop. These correction terms, associated with so-called resonance crossing, in turn associated with large energy spread, are not as well understood as the rhs of (10), partly because of their peculiar form. Nevertheless, careful observation of spin motion during the Monte-Carlo tracking in approach (ii), might provide a way to investigate their existence and form. ## 3 The Bloch equation and the Reduced Bloch equation in the laboratory frame In the previous section we used the beam frame and we will do so later. However the BE was first presented in Ref. 3 for the laboratory frame and in that frame it also has its simplest form. In this section we focus on the laboratory frame. In a semiclassical probabilistic description of an electron or positron bunch the spin-orbit dynamics is described by the spin-$1/2$ Wigner function $\rho$ (also called the Stratonovich function) written as $\displaystyle\rho(t,z)=\frac{1}{2}[f_{lab}(t,z)I_{2\times 2}+\vec{\sigma}\cdot\vec{\eta}_{lab}(t,z)]\;,$ (12) with $z=(\vec{r},\vec{p})$ where $\vec{r}$ and $\vec{p}$ are the position and momentum vectors of the phase space and $t$ is the time, and where $f_{lab}$ is the phase-space density of particles normalized by $\int dzf_{lab}(t,z)=1$, $\vec{\eta}_{lab}$ is the polarization density of the bunch and $\vec{\sigma}$ is the vector of the three Pauli matrices. As explained in Ref. 20, $\vec{\eta}_{lab}$ is proportional to the spin angular momentum density. In fact it is given by $\vec{\eta}_{lab}(t,z)=f_{lab}(t,z)\vec{P}_{loc,lab}(t,z)$ where $\vec{P}_{loc,lab}$ is the local polarization vector. Thus $f_{lab}=Tr[\rho]$ and $\vec{\eta}_{lab}=Tr[\rho\vec{\sigma}]$. The polarization vector $\vec{P}_{lab}(t)$ of the bunch is $\vec{P}_{lab}(t)=\int dz\vec{\eta}_{lab}(t,z)=\int dzf_{lab}(t,z)\vec{P}_{loc,lab}(t,z)$. Then, by neglecting collective effects and after several other approximations, the phase-space density evolves according to Ref. 3 via $\displaystyle\partial_{t}f_{lab}=L_{FP}^{lab}(t,z)f_{lab}\;.$ (13) Using the units as in Ref. 3 the Fokker-Planck operator $L_{FP}^{lab}$ is defined by $\displaystyle L_{FP}^{lab}(t,z):=L_{Liou}^{lab}(t,z)+\vec{F}_{rad}(t,z)+\vec{Q}_{rad}(t,z)+\frac{1}{2}\sum_{i,j=1}^{3}\partial_{p_{i}}\partial_{p_{j}}{\cal E}_{ij}(t,z)\;,$ (14) where $\displaystyle L_{Liou}^{lab}(t,z):=-\partial_{\vec{r}}\cdot\frac{1}{m\gamma(\vec{p})}\vec{p}-\partial_{\vec{p}}\cdot[e\vec{E}(t,\vec{r})+\frac{e}{m\gamma(\vec{p})}(\vec{p}\times\vec{B}(t,\vec{r}))]\;,$ (15) $\displaystyle\vec{F}_{rad}(t,z):=-\frac{2}{3}\frac{e^{4}}{m^{5}\gamma(\vec{p})}|\vec{p}\times\vec{B}(t,\vec{r})|^{2}\vec{p}\;,$ (16) $\displaystyle\vec{Q}_{rad}(t,z):=\frac{55}{48\sqrt{3}}\sum_{j=1}^{3}\;\frac{\partial[\lambda(t,z)\vec{p}p_{j}]}{\partial p_{j}}\;,$ (17) $\displaystyle{\cal E}_{ij}(t,z):=\frac{55}{24\sqrt{3}}\lambda(t,z)p_{i}p_{j}\;,\quad\lambda(t,z):=\hbar\frac{|e|^{5}}{m^{8}\gamma(\vec{p})}|\vec{p}\times\vec{B}(t,\vec{r})|^{3}\;,$ (18) $\displaystyle\gamma(\vec{p}):=\frac{1}{m}\sqrt{|\vec{p}|^{2}+m^{2}}\;,$ (19) and with $e$ being the electric charge of the electron or positron and $\vec{E}$ and $\vec{B}$ being the external electric and magnetic fields. The Fokker-Planck operator $L_{FP}^{lab}$ whose explicit form is taken from Ref. 3 is a linear second-order partial differential operator and, with some additional approximations, is commonly used for electron synchrotrons and storage rings, see Ref. 29 and Section 2.5.4 in Ref. 19. As usual, since it is minuscule compared to all other forces, the Stern-Gerlach effect from the spin onto the orbit is neglected in (13). The polarization density $\vec{\eta}_{lab}$ evolves via eq. 2 in Ref. 3, i.e., via that which we call the Bloch equation, namely $\displaystyle\partial_{t}\vec{\eta}_{lab}=L_{\rm FP}^{lab}(t,z)\vec{\eta}_{lab}+M(t,z)\vec{\eta}_{lab}$ $\displaystyle\quad-[1+\partial_{\vec{p}}\cdot\vec{p}]\lambda(t,z)\frac{1}{m\gamma(\vec{p})}\frac{\vec{p}\times\vec{a}(t,z)}{|\vec{a}(t,z)|}f_{lab}(t,z)\;,$ (20) where $\displaystyle M(t,z):=\Omega^{lab}(t,z)-\lambda(t,z)\frac{5\sqrt{3}}{8}[I_{3\times 3}-\frac{2}{9m^{2}\gamma^{2}(\vec{p})}\vec{p}\vec{p}^{T}]\;,$ (21) $\displaystyle\vec{a}(t,z):=\frac{e}{m^{2}\gamma^{2}(\vec{p})}(\vec{p}\times\vec{B}(t,\vec{r}))\;.$ (22) The BE was derived in Ref. 3 from the semiclassical approximation of quantum electrodynamics and it is a generalization, to the whole phase space, of the Baier-Katkov-Strakhovenko equation which just describes the evolution of polarization along a single deterministic trajectory [30]. Note also that, while the BE was new in 1975, the orbital Fokker-Planck equation (13) was already known thanks to research of the 1950s, e.g., Schwinger’s paper on quantum corrections to synchrotron radiation [31]. The skew-symmetric matrix $\Omega^{lab}(t,z)$ takes into account the Thomas-BMT spin-precession effect. Thus in the laboratory frame the Thomas-BMT-equation (2) reads as $\displaystyle\partial_{t}\hat{n}_{lab}=L_{Liou}^{lab}(t,z)\hat{n}_{lab}+\Omega^{lab}(t,z)\hat{n}_{lab}\;.$ (23) The quantum aspect of (13) and (20) is embodied in the factor $\hbar$ in $\lambda(t,z)$. For example $\vec{Q}_{rad}$ is a quantum correction to the classical radiation reaction force $\vec{F}_{rad}$. The terms $-\lambda(t,z)\frac{5\sqrt{3}}{8}\vec{\eta}_{lab}$ and $-\lambda(t,z)\frac{1}{m\gamma(\vec{p})}\frac{\vec{p}\times\vec{a}(t,z)}{|\vec{a}(t,z)|}f_{lab}(t,z)$ take into account spin flips due to synchrotron radiation and encapsulate the Sokolov-Ternov effect. The term $\lambda(t,z)\frac{5\sqrt{3}}{8}\frac{2}{9m^{2}\gamma^{2}(\vec{p})}\vec{p}\vec{p}^{T}\vec{\eta}_{lab}$ encapsulates the Baier-Katkov correction, and the term $\partial_{\vec{p}}\cdot\vec{p}\;\lambda(t,z)\frac{1}{m\gamma(\vec{p})}\frac{\vec{p}\times\vec{a}(t,z)}{|\vec{a}(t,z)|}f_{lab}(t,z)$ encapsulates the kinetic-polarization effect. The only terms in (20) which couple the three components of $\vec{\eta}_{lab}$ are the Thomas-BMT term and the Baier-Katkov correction term. As mentioned above, there exists a system of SDEs underlying (20) (for details, see Ref. 6). In particular, $f_{lab}$ and $\vec{\eta}_{lab}$ are related to a spin-orbit density ${\cal P}_{lab}={\cal P}_{lab}(t,z,\vec{s})$ via $\displaystyle f_{lab}(t,z)=\int_{{\mathbb{R}}^{3}}\;d\vec{s}\;{\cal P}_{lab}(t,z,\vec{s})\;,$ (24) $\displaystyle\vec{\eta}_{lab}(t,z)=\int_{{\mathbb{R}}^{3}}\;d\vec{s}\;\vec{s}\;{\cal P}_{lab}(t,z,\vec{s})\;,$ (25) where ${\cal P}_{lab}$ satisfies the Fokker-Planck equation corresponding to the system of SDEs in Ref. 6. These SDEs can be used as the basis for a Monte- Carlo spin tracking algorithm, i.e., for Method 2 mentioned in Section 1 above. This would extend the standard Monte-Carlo spin tracking algorithms, which we mentioned in Section 2 above, by taking into account all physical effects described by (20), i.e., the Sokolov-Ternov effect, the Baier-Katkov correction, the kinetic-polarization effect and, of course, spin diffusion. If we ignore the spin-flip terms and the kinetic-polarization term in the BE then (20) simplifies to the RBE $\displaystyle\partial_{t}\vec{\eta}_{lab}=L_{FP}^{lab}(t,z)\vec{\eta}_{lab}+\Omega^{lab}(t,z)\vec{\eta}_{lab}\;.$ (26) The RBE models spin diffusion due to the effect of the stochastic orbital motion on the spin and thus contains those terms of the BE which are related to the radiative depolarization rate $\tau_{dep}^{-1}$. This effect is clearly seen in the SDEs (see, e.g., (28) and (29)). ## 4 The Reduced Bloch equation in the beam frame In the beam frame, i.e., in the accelerator coordinates $y$ of Section 2, the RBE (26) becomes $\displaystyle\partial_{\theta}\vec{\eta}=L_{\rm FP}^{y}(\theta,y)\vec{\eta}+\Omega(\theta,y)\vec{\eta}\;.$ (27) Because the coefficients of $L_{\rm FP}^{y}$ are $\theta$-dependent, the RBE (27) is numerically and analytically quite complex. So we first approximate it by treating the synchrotron radiation as a perturbation. Then, in order to solve it numerically to determine the long-time behavior that we need, we address the system of SDEs underlying (27) and apply the refined averaging technique presented in Ref. 32 (see also 7), for the orbital dynamics, and extend it to include spin. The averaged SDEs are then used to construct an approximate RBE which we call the effective RBE. The system of SDEs underlying (27) reads as 555We denote the random dependent variables like $Y$ in (28) by capital letters to distinguish them from independent variables like $y$ in (27). $\displaystyle\frac{dY}{d\theta}=(A(\theta)+\epsilon_{R}\delta A(\theta))Y+\sqrt{\epsilon_{R}}\sqrt{\omega(\theta)}e_{6}\xi(\theta)\;,$ (28) $\displaystyle\frac{d\vec{S}}{d\theta}=[\Omega_{0}(\theta)+\epsilon_{S}C(\theta,Y)]\vec{S}\;,$ (29) where the orbital dynamics has been linearized in $Y$ and $\Omega=\Omega_{0}+\epsilon_{S}C$ has been linearized in $Y$ so that $\displaystyle C(\theta,Y)=\sum_{j=1}^{6}\;C_{j}(\theta)Y_{j}\;.$ (30) Also, $A(\theta)$ is a Hamiltonian matrix representing the nonradiative part of the orbital dynamics and $Y$ has been scaled so that $\epsilon_{R}$ is the size of the orbital effect of the synchrotron radiation. Thus $\epsilon_{R}\delta A(\theta)$ represents the orbital damping effects due to synchrotron radiation and the cavities, $\sqrt{\epsilon_{R}}\xi(\theta)$ represents the associated quantum fluctuations, $\xi$ is the white noise process and $e_{6}:=(0,0,0,0,0,1)^{T}$. In the spin equation (29), $\Omega_{0}$ is the closed-orbit contribution to $\Omega$ so that $\epsilon_{S}C(\theta,Y)$ is what remains and $C(\theta,Y)$ is chosen $O(1)$. Hence $\epsilon_{S}$ estimates the size of $\Omega-\Omega_{0}$. Both $\Omega_{0}(\theta)$ and $C(\theta,Y)$ are, of course, skew-symmetric $3\times 3$ matrices. We are interested in the situation where $\epsilon_{R}$ and $\epsilon_{S}$ are small in some appropriate sense. Eqs. (28) and (29) can be obtained by transforming the system of SDEs in Ref. 6 from the laboratory frame to the beam frame [33]. However, since in this section we only deal with the RBE (not with the BE), (28) and (29) can also be found in older expositions on spin in high-energy electron storage rings, e.g., Ref. 34. Note that these expositions make approximations as for example with the linearity of (28) in $Y$ and the linearity of $C(\theta,Y)$ in $Y$. With (28) and (29) the evolution equation for the spin-orbit joint probability density ${\cal P}={\cal P}(\theta,y,\vec{s})$ is the following spin-orbit Fokker-Planck equation $\displaystyle\partial_{\theta}{\cal P}=L_{\rm FP}^{y}(\theta,y){\cal P}-\partial_{\vec{s}}\cdot\Biggl{(}\biggl{(}\Omega(\theta,y)\vec{s}\biggr{)}{\cal P}\Biggr{)}\;,$ (31) where $L_{\rm FP}^{y}$ is the orbital Fokker-Planck operator. The phase-space density $f$ and the polarization density $\vec{\eta}$ corresponding to ${\cal P}$ are defined by $\displaystyle f(\theta,y)=\int_{{\mathbb{R}}^{3}}\;d\vec{s}\;{\cal P}(\theta,y,\vec{s})\;,\quad\vec{\eta}(\theta,y)=\int_{{\mathbb{R}}^{3}}\;d\vec{s}\;\vec{s}\;{\cal P}(\theta,y,\vec{s})\;,$ (32) which are the beam-frame analogs of (24) and (25). The local polarization vector $\vec{P}_{loc}$ from Section 2 above is related to $f$ and $\vec{\eta}$ by $\displaystyle\vec{\eta}(\theta,y)=f(\theta,y)\vec{P}_{loc}(\theta,y)\;.$ (33) The RBE (27) follows from (31) by differentiating (32) w.r.t. $\theta$ and by using the Fokker-Planck equation for ${\cal P}$. This proves that (28) and (29) is the system of SDEs which underlie the RBE (27). For (27), see also Ref. 20. ## 5 The Effective Reduced Bloch equation in the beam frame The effective RBE is, by definition, an approximation of the RBE (27) obtained by approximating the system of SDEs (28) and (29) using the method of averaging, see Refs. 7-12. We call the system of SDEs underlying the effective RBE the effective system of SDEs. We now discuss first-order averaging in the case where $\epsilon:=\epsilon_{S}=\epsilon_{R}$ is small. To apply the method of averaging to (28) and (29) we must transform them to a standard form for averaging, i.e., we must transform the variables $Y,\vec{S}$ to slowly varying variables. We do this by using a fundamental solution matrix $X$ of the unperturbed $\epsilon=0$ part of (28), i.e., $\displaystyle X^{\prime}=A(\theta)X\;,$ (34) and a fundamental solution matrix $\Phi$ of the unperturbed $\epsilon=0$ part of (29), i.e., $\displaystyle\Phi^{\prime}=\Omega_{0}(\theta)\Phi\;.$ (35) We thus transform $Y$ and $\vec{S}$ into the slowly varying $U$ and $\vec{T}$ via $\displaystyle Y(\theta)=X(\theta)U(\theta)\;,\quad\vec{S}(\theta)=\Phi(\theta)\vec{T}(\theta)\;.$ (36) Hence (28) and (29) are transformed to $\displaystyle U^{\prime}=\epsilon{\cal D}(\theta)U+\sqrt{\epsilon}\sqrt{\omega(\theta)}X^{-1}(\theta)e_{6}\xi(\theta)\;,$ (37) $\displaystyle\vec{T}^{\prime}=\epsilon{\mathfrak{D}}(\theta,U)\vec{T}\;,$ (38) where ${\cal D}$ and ${\mathfrak{D}}$ are defined by $\displaystyle{\cal D}(\theta):=X^{-1}(\theta)\delta A(\theta)X(\theta)\;,$ (39) $\displaystyle{\mathfrak{D}}(\theta,U):=\Phi^{-1}(\theta)C(\theta,X(\theta)U)\Phi(\theta)\;.$ (40) Of course, (37) and (38) carry the same information as (28) and (29). Now, applying the method of averaging to (37) and (38), we obtain the following effective system of SDEs $\displaystyle V^{\prime}=\epsilon\bar{\cal D}V+\sqrt{\epsilon}{\cal B}(\xi_{1},...,\xi_{k})^{T}\;,$ (41) $\displaystyle\vec{T}_{a}^{\prime}=\epsilon\bar{\mathfrak{D}}(V)\vec{T}_{a}\;,$ (42) where the bar denotes $\theta$-averaging, i.e., the operation $\lim_{T\rightarrow\infty}(1/T)\int_{0}^{T}d\theta\cdots$. Moreover $\xi_{1},...,\xi_{k}$ are statistically independent versions of the white noise process and ${\cal B}$ is a $6\times k$ matrix which satisfies ${\cal B}{\cal B}^{T}=\bar{\cal E}$ with $k=rank(\bar{\cal E})$ and where $\bar{\cal E}$ is the $\theta$-average of $\displaystyle{\cal E}(\theta)=\omega(\theta)X^{-1}(\theta)e_{6}e_{6}^{T}X^{-T}(\theta)\;.$ (43) For physically reasonable $A$ and $\Omega$ the fundamental matrices $X$ and $\Phi$ are quasiperiodic functions whence ${\cal D},{\mathfrak{D}}(\cdot,U)$ and ${\cal E}$ are quasiperiodic functions so that their $\theta$ averages $\bar{\cal D},\bar{\mathfrak{D}}(V)$ and $\bar{\cal E}$ exist. Our derivation of (41) from (37) is discussed in some detail in Ref. 6. We are close to showing that $U=V+O(\epsilon)$ on $\theta$-intervals of length $O(1/\epsilon)$ and it seems likely that this error is valid for $0\leq\theta<\infty$, because of the radiation damping. This is a refinement of Ref. 32 and assumes a non-resonance condition. Since the sample paths of $U$ are continuous and $U$ is slowly varying it seems likely that $\vec{T}_{a}$ is a good approximation to $\vec{T}$ and we are working on the error analysis. Spin-orbit resonances will be an important focus in the construction of $\bar{\mathfrak{D}}(V)$ from (40) which contains both the orbital frequencies in $X$ and the spin precession frequency in $\Phi$. Since, by definition, the effective system of SDEs underly the effective RBE, the latter can be obtained from the former in the same way as we obtained (27) from (28) and (29) (recall the discussion after (32)). Thus the evolution equation for the spin-orbit probability density ${\cal P}_{V}={\cal P}_{V}(\theta,{\rm v},\vec{t})$ is the following Fokker-Planck equation: $\displaystyle\partial_{\theta}{\cal P}_{V}=L^{V}_{\rm FP}({\rm v}){\cal P}_{V}-\epsilon\partial_{\vec{t}}\cdot\Biggl{(}\biggl{(}\bar{\mathfrak{D}}({\rm v})\vec{t}\biggr{)}{\cal P}_{V}\Biggr{)}\;,$ (44) where $\displaystyle L^{V}_{\rm FP}({\rm v})=-\epsilon\sum_{j=1}^{6}\partial_{{\rm v}_{j}}(\bar{\cal D}{\rm v})_{j}+\frac{\epsilon}{2}\sum_{i,j=1}^{6}\bar{\cal E}_{ij}\partial_{{\rm v}_{i}}\partial_{{\rm v}_{j}}\;.$ (45) The polarization density $\vec{\eta}_{V}$ corresponding to ${\cal P}_{V}$ is defined by $\displaystyle\vec{\eta}_{V}(\theta,{\rm{\rm v}})=\int_{{\mathbb{R}}^{3}}\;d\vec{t}\;\vec{t}\;{\cal P}_{V}(\theta,{\rm v},\vec{t})\;,$ (46) so that, by (44), the effective RBE is $\displaystyle\partial_{\theta}\vec{\eta}_{V}=L^{V}_{\rm FP}({\rm v})\vec{\eta}_{V}+\epsilon\bar{\mathfrak{D}}({\rm v})\vec{\eta}_{V}\;.$ (47) This then is the focus of our approach in Method 1. For more details on this section, see Refs. 6, 17 and 18. ## 6 Next steps * • Further development of Bloch-equation approach (numerical and theoretical), i.e., of Method 1 and with a realistic lattice. * • Development of validation methods, i.e., Methods 2-4. Note that Method 2 is an extension of the standard Monte-Carlo spin tracking algorithms and for that matter we will study Refs. 13, 14 and 15. * • Investigating the connection between the Bloch-equation approach and the standard approach based on the Derbenev-Kondratenko formulas, and studying the potential for correction terms [4] to $\tau_{\rm DK}^{-1}$ by using the RBE. ## 7 Acknowledgement This material is based upon work supported by the U.S. Department of Energy, Office of Science, Office of High Energy Physics, under Award Number DE- SC0018008. ## References * [1] K. Heinemann, Re-evaluation of Spin-Orbit Dynamics of Polarized $e^{+}e^{-}$ Beams in High Energy Circular Accelerators and Storage Rings: Bloch equation approach, Invited talk at IAS Mini-Workshop on Beam Polarization in Future Colliders, Hong Kong, Jan. 17, 2019. See also: http://iasprogram.ust.hk/hep/2019 and K. Heinemann et al., _Int. J. Mod. Phys._ , vol. A35, Nos. 15 & 16, 2041003, 2020. * [2] Ya.S. Derbenev, A.M. Kondratenko, Polarization kinetics of particles in storage rings, _Sov. Phys. JETP_ , vol. 37, p. 968, 1973. * [3] Ya.S. Derbenev, A.M. Kondratenko, Relaxation and equilibrium state of electrons in storage rings, _Sov. Phys. Dokl._ , vol. 19, p. 438, 1975. * [4] Z. Duan, M. Bai, D.P. Barber, Q. Qin, A Monte-Carlo simulation of the equilibrium beam polarization in ultra-high energy electron (positron) storage rings, _Nucl. Instr. Meth._ , vol. A793, pp. 81-91, 2015. See also arXiv:physics/1505.02392v2. * [5] F. Bloch, Nuclear Induction, _Phys. Rev._ , vol. 70, p. 460, 1946. * [6] K. Heinemann, D. Appelö, D. P. Barber, O. Beznosov, J.A. Ellison, The Bloch equation for spin dynamics in electron storage rings: computational and theoretical aspects. Invited paper for ICAP18. _Int. J. Mod. Phys._ , vol. A34, 1942032, 2019. * [7] J.A. Ellison, H. Mais, G. Ripken, K. Heinemann, Details of Orbital Eigen-analysis for Electron Storage Rings. (An extension of the article in Ref. 32. In preparation). * [8] J. A. Ellison and H-Jeng Shih, The Method of Averaging in Beam Dynamics in Accelerator Physics Lectures at the Superconducting Super Collider, AIP Conference Proceedings 326, edited by Yiton T. Yan, James P. Naples and Michael J. Syphers (1995). * [9] J. A. Ellison, K. A. Heinemann, M. Vogt and M. Gooden, Planar undulator motion excited by a fixed traveling wave: Quasiperiodic Averaging, normal forms and the FEL pendulum, _Phys. Rev. ST Accel. Beams_ , vol. 16, 090702, 2013. An earlier version is on the archive at arXiv:1303.5797 (2013) and published as DESY report 13-061. * [10] J.A. Sanders, F. Verhulst, J. Murdock, Averaging Methods in Nonlinear Dynamical Systems, 2nd Edition, Springer, New York, 2007. * [11] J. Murdock, Perturbations: Theory and Methods, SIAM, Philadelphia, 1999. * [12] R. Cogburn and J.A. Ellison, A Stochastic Theory of Adiabatic Invariance, _Communications of Mathematical Physics_ , vol. 148, pp. 97-126, 1992. Also: A four-thirds law for phase randomization of stochastically perturbed oscillators and related phenomena, _Communications of Mathematical Physics_ , vol. 166, pp. 317-336, 1994. * [13] P.E. Kloeden, E. Platen, Numerical Solution of Stochastic Differential Equations, Appl. Math., Vol. 23, Springer. Third printing, 1999 (first edition 1992). * [14] P.E. Kloeden, E. Platen, H. Schurz, Numerical Solution of SDEs Through Computer Experiments, Universitext, Springer. Third corrected printing, 2003 (first edition 1994). * [15] E. Platen, N. Bruti-Liberati, Numerical Solution of Stochastic Differential Equations with Jumps in Finance, Springer, 2010. * [16] O. Beznosov, From Wave Propagation to Spin Dynamics: Mathematical and Computational Aspects, PhD Dissertation, Department of Mathematics and Statistics, University of New Mexico (in progress). * [17] K. Heinemann, D. Appelö, D. P. Barber, O. Beznosov, J. A. Ellison, Spin dynamics in modern electron storage rings: Computational and theoretical aspects, ICAP18, Key West, Oct 19–23, 2018. * [18] O. Beznosov, J. A. Ellison, K. Heinemann, D. P. Barber, D. Appelö, Spin dynamics in modern electron storage rings: Computational aspects, ICAP18, Key West, Oct 19–23, 2018. * [19] D.P. Barber, G. Ripken, _Handbook of Accelerator Physics and Engineering_. Eds. A.W. Chao and M. Tigner, 1st edition, 3rd printing, World Scientific, 2006\. See also the 2nd edition: Eds. A.W. Chao, K.H.Mess, M. Tigner and F. Zimmermann, World Scientific, 2013 as well as arXiv:physics/9907034v2. * [20] K. Heinemann, D.P. Barber, Spin transport, spin diffusion and Bloch equations in electron storage rings, _Nucl. Instr. Meth._ , vol. A463, pp. 62–67, 2001. Erratum-ibid.A469:294. * [21] J.A. Ellison, K. Heinemann, Polarization Fields and Phase Space Densities in Storage Rings: Stroboscopic Averaging and the Ergodic Theorem, _Physica D_ , vol. 234, p. 131, 2007. * [22] D. Barber, J. A. Ellison, K. Heinemann, Quasiperiodic spin-orbit motion and spin tunes in storage rings, _Phys. Rev. ST Accel. Beams_ , vol. 7, 124002, 2004. * [23] D. Sagan, A Superconvergent Algorithm for Invariant Spin Field Stroboscopic Calculations David Sagan, IPAC18, Vancouver, April 29-May 4, 2018. * [24] S.R. Mane, Yu. M. Shatunov, K. Yokoya, Spin-polarized charged particle beams in high-energy. accelerators, _Rep. Prog. Phys._ , vol. 68, p. 1997, 2005. * [25] F. Meot, Zgoubi. http://sourceforge.net/projects/zgoubi * [26] E. Forest, From Tracking Code to Analysis, Springer, 2016. * [27] D. Sagan, Bmad, a subroutine library for relativistic charged-particle dynamics. See: https://www.classe.cornell.edu/bmad * [28] Y.S. Derbenev, A.M. Kondratenko, Diffusion of Particle Spins in Storage Rings, _Sov. Phys. JETP_ , vol. 35, p. 230, 1972. * [29] M. Sands, The physics of electron storage rings, SLAC-121, 1970. J. Jowett, Introductory Statistical Mechanics for electron storage rings, SLAC-PUB-4033, 1986. * [30] V.N. Baier, V.M. Katkov, V.M. Strakhovenko, Kinetics of Radiative Polarization, _Sov. Phys. JETP_ , vol. 31, p. 908, 1970. * [31] J. Schwinger, The quantum correction in the radiation by energetic accelerated electrons, _Proc.Nat.Acad.Sci._ , vol. 40, p. 132, 1954. * [32] J.A. Ellison, H. Mais, G. Ripken, Orbital Eigen-analysis for Electron Storage Rings, in: Section 2.1.4 of _Handbook of Accelerator Physics and Engineering_. Eds. A.W. Chao, K.H. Mess, M. Tigner, F. Zimmermann, 2nd edition, World Scientific, 2013. * [33] K. Heinemann, D. P. Barber, O. Beznosov, J. A. Ellison. In preparation. * [34] D.P. Barber, K. Heinemann, H. Mais, G. Ripken, A Fokker-Planck treatment of stochastic particle motion within the framework of a fully coupled six-dimensional formalism for electron - positron storage rings including classical spin motion in linear approximation, DESY-91-146, 1991.
[figure]style=plain,subcapbesideposition=top # Kinetic modelling of three-dimensional shock/laminar separation bubble instabilities in hypersonic flows over a double wedge Saurabh S. Sawant1<EMAIL_ADDRESS>V. Theofilis2,3 D. A. Levin1 1Department of Aerospace, University of Illinois at Urbana-Champaign, 104 S. Wright St, Champaign, Illinois, USA 2 School of Engineering, University of Liverpool, The Quadrangle, Brownlow Hill, L69 3GH, UK 3 Escola Politecnica, Universidade São Paulo, Av. Prof. Mello Moraes 2231, CEP 5508-900, São Paulo- SP, Brasil ###### Abstract Linear global instability of the three-dimensional (3-D), spanwise-homogeneous laminar separation bubble (LSB) induced by shock-wave/boundary-layer interaction (SBLI) in a Mach 7 flow of nitrogen over a $30^{\circ}-55^{\circ}$ double wedge is studied. At these conditions corresponding to a freestream unit Reynolds number, $Re_{1}=$5.2\text{\times}{10}^{4}$$ m-1, the flow exhibits rarefaction effects and comparable shock-thicknesses to the size of the boundary-layer at separation. This, in turn, requires the use of the high- fidelity Direct Simulation Monte Carlo (DSMC) method to accurately resolve unsteady flow features. We show for the first time that the LSB sustains self-excited, small- amplitude, 3-D perturbations that lead to spanwise-periodic flow structures not only in and downstream of the separated region, as seen in a multitude of experiments and numerical simulations, but also in the internal structure of the separation and detached shock layers. The spanwise-periodicity length and growth rate of the structures in the two zones are found to be identical. It is shown that the linear global instability leads to low-frequency unsteadiness of the triple point formed by the intersection of separation and detached shocks, corresponding to a Strouhal number of $St\sim 0.02$. Linear superposition of the spanwise-homogeneous base flow and the leading 3-D flow eigenmode provides further evidence of the strong coupling between linear instability in the LSB and the shock layer. ###### keywords: Authors should not enter keywords on the manuscript, as these must be chosen by the author during the online submission process and will then be added during the typesetting process (see Keyword PDF for the full list). Other classifications will be added at the same time. ## 1 Introduction Laminar SBLI has been a topic of extensive study since the best part of last century. The early experimental and theoretical work primarily focused on the upstream influence of disturbances in boundary layers, as can be found in seminal contributions such as Czarnecki & Mueller (1950); Liepmann et al. (1951); Lighthill & Newman (1953); Lighthill (1953, 2000); Chapman et al. (1958); Stewartson (1964). In subsequent research, triple deck theory (Stewartson & Williams, 1969; Smith, 1986; Neiland, 2008) was developed and used to understand boundary layer instability mechanisms that lead to separation in supersonic and hypersonic flows over compression ramps at moderate to high Reynolds numbers (Rizzetta et al., 1978; Cowley & Hall, 1990; Smith & Khorrami, 1991; Cassel et al., 1995; Korolev et al., 2002; Fletcher et al., 2004). More recent topics of study on shock-induced LSB include 3-D effects (Lusher & Sandham, 2020), unsteadiness and underlying instability mechanisms (Sansica et al., 2016), and the coupling between LSB and shock structure (Tumuklu et al., 2018b; Sawant et al., 2019). Experimental investigations of hypersonic SBLI primarily exist on compression ramps in flow regimes from laminar to turbulent and are at large $Re_{1}\sim O(10^{6}-10^{7})$ m-1 (Holden, 1963, 1978; Needham, 1965b, a; Elfstrom, 1971, 1972; Hankey Jr & Holden, 1975; Simeonides & Haase, 1995; Schneider, 2004; Roghelia et al., 2017; Chuvakhov et al., 2017). Experiments on flows over double wedges can be found at moderate Reynolds number, $Re_{1}\sim O(10^{5}-10^{6})$ m-1, but are limited, exhibit far more complicated SBLI than compression ramps, and suffer from test times significantly lower than the characteristic times involved in the development of instabilities and unsteadiness. Schrijer et al. (2006, 2009) performed experiments at $Re_{1}\sim O(10^{7})$ m-1 and $M=7$ in a Ludwieg tube facility and observed an unsteady flow exhibiting Edney type-$VI$ and $V$ interactions on 15∘-30∘ and 15∘-45∘ double wedge configurations, respectively. In a moderate Reynolds number regime, $Re_{1}\sim O(10^{5})$ m-1, Hashimoto (2009) performed experiments in a free piston shock tunnel, where the flow of air on a double wedge was tested for 300 $\mu$s at $Re_{1}=0.18-$3.5\text{\times}{10}^{6}$$ m-1 and $M=7$, and was found to exhibit Edney type-$V$ interactions. Recently, Swantek & Austin (2015) have performed experiments on a 30∘-25∘ double wedge in Hypervelocity Expansion Tube (HET) facility to test air and nitrogen at $Re_{1}=0.44-$4.6\text{\times}{10}^{6}$$ m-1 and $M=4.01-7.14$, where the test times ranged from 361-562 $\mu$s. Experiments of Knisely & Austin (2016) include nitrogen flow over the 30∘-55∘ double wedge geometry considered in this work at higher $Re_{1}=0.435-$1.1\text{\times}{10}^{6}$$ m-1 and $M=6.64-7.14$, where, in addition to HET facility, the T5 reflected shock tunnel was used that allows for a longer test time of 1 ms. On the numerical side, existing studies in laminar and transitional regimes primarily use compressible Navier-Stokes equations to understand different aspects of SBLI in hypersonic flows over compression ramps such as three- dimensionality of an LSB (Rudy et al., 1991), stability of hypersonic boundary layers (Balakumar et al., 2005), development of 3-D instability in the form of spanwise periodic striations of the LSB (Egorov et al., 2011; Dwivedi et al., 2019), and formation of secondary vortices and their fragmentation inside an LSB (Gai & Khraibut, 2019). Efforts such as these have been extended to simulate laminar SBLI over double wedge geometries (Knight et al., 2017). Durna et al. (2016) simulated a 2-D Mach 7 flow of nitrogen over a double wedge at the 2 MJ low enthalpy conditions of Swantek and Austin ($Re_{1}=$1.1\text{\times}{10}^{6}$$) to study the effect of the aft wedge angle on the flow characteristics with additional, recently included 3-D effects (Durna & Celik, 2020). Sidharth et al. (2018) carried out global stability analysis and Direct Numerical Simulation (DNS) of a Mach 5 perfect gas flow at $Re_{1}=$1.36\text{\times}{10}^{6}$$ over double ramps with forward and aft angles of 12∘ and 12∘-22∘, respectively. For aft angle of 20∘ and greater, they observed a linear instability of the 2-D separation bubble in the absence of upstream perturbations and associated that with streamwise streaks in wall temperature near the reattachment region. Recently, Reinert et al. (2020) simulated 3-D flows at Mach 7 over a 30∘-55∘ double wedge at the experimental conditions of Swantek & Austin (2015) and Knisely & Austin (2016) for much longer flow times than the duration of the experiments and reported unsteady asymmetric 3-D separation bubble. However, despite extensive numerical and experimental work, the physics of complicated SBLI formed in a hypersonic flow over double-wedge configurations is not well-understood. In this work, we investigate questions about the instability mechanism of a 3-D LSB, the coupling between the shock structure and LSB, and the low-frequency unsteadiness of the shock structure. We focus on the linear instability of a shock-induced, 3-D LSB formed in a Mach 7 nitrogen flow over a 30∘-55∘ double wedge configuration at a freestream unit Reynolds number of $Re_{1}=$5.2\text{\times}{10}^{4}$$ corresponding to an altitude of about 60 km. We will show that even for this lower density free stream condition, that is typically not studied, our fully-resolved, kinetic DSMC simulations of this complex flow allow us to study the strong coupling between the separation shock and the LSB. In our previous work, a 2-D (spanwise independent) flow over the same configuration and freestream conditions was simulated by Tumuklu et al. (2019), who demonstrated that the flow reaches a steady-state in $\sim 0.9~{}ms$ after the leading damped global modes, recovered by the residuals algorithm (Theofilis et al., 2000) and proper orthogonal decomposition (POD), have decayed. Yet we will show in our 3-D treatment using this 2-D, steady base flow of Tumuklu et al. (2019), that indeed the flow is linearly unstable to self-excited, small-amplitude, spanwise-homogeneous perturbations and will ultimately transition to turbulence. An important goal of the research discussed in this paper is to understand the effect of three-dimensionality on the coupling between the LSB and shock structure. The accurate modelling of the internal shock structure is an essential feature of relevance to the study of linear instability in high- speed boundary layer flow, where coupling between the separation bubble and the shock has been demonstrated through the amplitude function of the underlying global modes in a number of studies (e.g. Crouch et al., 2007). In the hypersonic regime, Tumuklu et al. (2018a, b) used DSMC to study the effect of unit Reynolds number, $Re_{1}=0.935-$3.74\text{\times}{10}^{5}$~{}m^{-1}$, on laminar SBLI in a Mach 16 nitrogen flow over an axisymmetric double cone configuration. The authors demonstrated a strong coupling between oscillations of the shock structure and instability of the laminar separated flow region through the spatial structure of the amplitude functions as well as Kelvin- Helmholtz waves formed at the contact surface downstream of the triple point. In this work, we focus on the coupling mechanism between a fully 3-D LSB and shock, and show that the instability in the LSB as well as inside the strong gradient region of shocks are intimately related. To capture the complex physics of SBLI with highest fidelity, we use the DSMC method. Continuous developments spanning the past fifty decades have resulted in this method being well-suited for the study of the physics of unsteady laminar SBLI to deliver accurate results in five critical aspects: (a) computations of molecular thermal fluctuations (Garcia, 1986; Bruno, 2019; Sawant et al., 2020), (b) calculation of anisotropic stresses and heat fluxes in strong shock layers (M $>>$ 1.6) (Bird, 1970; Cercignani et al., 1999), (c) prediction of rarefaction effects such as velocity slip and temperature jump (Moss, 2001; Moss & Bird, 2005; Tumuklu et al., 2018a), (d) quantification of translational, rotational, and vibrational nonequilibrium (Sawant et al., 2018), and (e) time-accurate evolution of self-excited perturbations (Tumuklu et al., 2018a, b, 2019). As a result, the method is gaining momentum in the study of hydrodynamic instabilities (Bird, 1998; Stefanov et al., 2002a, b, 2007; Kadau et al., 2004, 2010; Gallis et al., 2015, 2016). In our flow, even though the freestream Knudsen number of $3.0\text{\times}{10}^{-3}$ is continuum, the local Knudsen number in the shock-LSB region is much higher due to the steep gradients in macroscopic flow parameters. These non-continuum features, also known as local rarefaction zones, are crucial to understanding the coupling between the shock and LSB, as this work will demonstrate. Furthermore, the high-fidelity kinetic modelling of these regions is crucial because the thicknesses of shocks and the boundary-layer in the separation region are comparable. The time-accuracy of the DSMC method in modelling unsteady evolution of 3-D perturbations allows for quantification of the low-frequency unsteadiness of the shock structure. This phenomenon has been extensively investigated in turbulent SBLI at $Re_{1}\sim O(10^{6}-10^{8})$ using DNS and large eddy simulation (LES) (see Pirozzoli & Grasso, 2006; Touber & Sandham, 2009; Piponniau et al., 2009; Grilli et al., 2012; Priebe & Martín, 2012; Clemens & Narayanaswamy, 2014; Gaitonde, 2015; Priebe et al., 2016; Pasquariello et al., 2017), where numerical studies report a Strouhal number associated with unsteadiness within a range of 0.01 to 0.05, consistent with findings of many experiments (Dussauge et al., 2006). However, in hypersonic, 3-D laminar SBLI, such investigations are sparse. Tumuklu et al. (2018b) observed a similar Strouhal number of $\sim$0.08, corresponding to the bow shock oscillation in their axisymmetric flow over a double cone simulated using DSMC. In the 3-D, Mach 7, finite-span double-wedge flow simulation of Reinert et al. (2020), however, such unsteadiness was not observed for conditions at a freestream flow enthalpy of 8 MJ, although, the Reynolds number of their case was a factor of 8 higher. In this work, we show that our 8 MJ enthalpy, Mach 7, spanwise-periodic flow over the same configuration (at eight-times lower density) exhibits low-frequency unsteadiness after the onset of linear instability. Finally, topology analysis of separated flows is an important way to characterize and compare complex 3-D flows for different input conditions and shapes (see, e.g. Lighthill, 1963; Tobak & Peake, 1982; Perry & Hornung, 1984a, b; Perry & Chong, 1987; Dallmann, 1983, 1985). Topologies of 3-D flow constructed from the linear superposition of the leading stationary eigenmode due to a linear instability of an LSB and 2-D base flow were analyzed by Rodríguez & Theofilis (2010) in the incompressible regime and Robinet (2007); Boin et al. (2006) in the compressible regime involving oblique SBLI. In the analysis presented here, we estimate the changes in the 3-D wall-streamline topology for an increasing amplitude of the 3-D perturbations based on a linear combination of the 2-D base flow and 3-D perturbations. In this simplified approach, we will demonstrate that the wall-streamline signature is very different depending on whether the coupling is considered. The paper is organized as follows: section 2 describes the methodology, which includes a brief description of the DSMC method in section 2.1 and details about numerical models, the DSMC solver, the input conditions, and flow initialization in section 2.2. The features of 2-D base flow are described in section 3. Section 4 is devoted to the key findings of this paper. Section 4.1 describes the linear instability mechanism and its spatial origin through a detailed discussion of boundary layer profiles. The correlation between the shock and the separation bubble is explained in section 4.2. The surface rarefaction effects are described in section 4.3, whereas the isocontours of spanwise periodic flow structures are discussed in section 4.4. The topology of the LSB is discussed in section 5, first without taking into account the coupling between the bubble and the shock in section 5.1 and then the effect of their coupling in section 5.2. The important findings are summarized in section 6. ## 2 Methodology ### 2.1 The DSMC algorithm The equation for the evolution of velocity distribution function of molecules, $f(t,\vec{r},\vec{v})$ with respect to time $t$, position vector $\vec{r}$, and instantaneous velocity vector $\vec{v}$, is written as, $\begin{split}\frac{\partial f}{\partial t}+(\vec{v}\cdot\nabla)f+\left(\frac{\vec{F}}{m}\cdot\nabla_{v}\right)f=\left[\frac{df}{dt}\right]_{coll}\\\ \end{split}$ (1) where $\nabla$ and $\nabla_{v}$ are gradient operators in physical and velocity spaces, respectively. The first, second, and third terms on the left- hand side describe the change of $f$ with time, change due to convection of molecules in physical space, and that in the velocity space, respectively. The latter can happen due to the action of external conservative force per unit mass $\vec{F}/m$, such as gravity or electric field, which are ignored in this work. The right-hand side (RHS) term accounts for changes in $f$ in an element of space-velocity phase-space due to intermolecular collisions. For a thorough description, see Vincenti & Kruger (1965). The DSMC method (see Bird, 1994) decouples the advection of molecules and their intermolecular collisions. Each simulated particle represents $F_{n}$ amount of actual gas molecules and is advected for a discrete timestep. Based on the choice of boundary conditions, particles are introduced, removed, or reflected from the domain boundaries and interacted with the embedded surfaces using gas-surface collision models for the duration of the timestep. They are then mapped to an adaptively refined collision mesh ($C$-mesh) that encompasses the flow domain and ensures the spatial proximity of particles that are potential candidates for binary collisions. Next follows a collision scheme, which selects particle pairs that are collided based on the appropriate (elastic/inelastic) collision cross-section and are assigned with post-collisional instantaneous velocities and internal energies. Macroscopic flow parameters of interest such as pressure, velocities, etc., can be derived from the microscopic properties of simulated particles using statistical relations of kinetic theory. These parameters are represented on the sampling mesh ($S$-mesh), which has coarser cells than the $C$-mesh. Note that the unique characteristic of the DSMC method, the advection-collision decoupling, is justified if the local cell size of $C$-Mesh, $\Delta x$, is smaller than the local mean-free-path of molecules, $\lambda$, and the timestep, $\Delta t$, is lower than the mean-collision-time, $\tau$. A sufficient number of instantaneous particles in the smallest collision cells must also be ensured for unbiased collisions. Conveniently, the satisfaction of only these three numerical criteria leads to accurate modelling of internal structure of shocks, their mutual interaction, and surface rarefaction effects. This warrants the use of DSMC for detailed modelling of SBLIs at high altitudes, compared to ad-hoc techniques of modelling shocks in computational fluid dynamics (CFD) simulations that fall short of accurately capturing the internal structure of shocks. ### 2.2 The numerical implementation and flow initialization The fulfillment of the numerical criteria demands a large number of computational particles and collision cells. To overcome this challenge, we have previously developed an octree-based, 3-D DSMC solver known as Scalable Unstructured Gas-dynamic Adaptive mesh-Refinement (SUGAR-3D). See Sawant et al. (2018) for a comprehensive account of the implementation strategies, validation, and performance studies of the solver. In summary, the code takes advantage of message-passing-interface (MPI) for parallel communication between processors, adaptive mesh refinement (AMR) of coarser Cartesian octree cells to achieve spatial fidelity at regions of strong gradients, a cutcell algorithm to correctly capture physics in the vicinity of embedded surfaces, a domain decomposition strategy based on Morton-Z space-filling-curves, capability of parallel input/output, inclusion of thermal nonequilibrium models, and numerous run-time memory reduction strategies. In the octree-based AMR framework, the $C$-mesh is formed from a user-defined, uniform Cartesian grid. The cells of this grid are known as ‘root’ cells, which are recursively subdivided into eight parts until the local cell-size is smaller than the local mean-free-path. Note that a subdivision based on the above criterion is performed only if there are at least 32 particles in a collision cell. The satisfaction of both of these criteria in the presented flow over a double wedge requires $\sim$60 billion computational particles and $\sim$4.5 billion collision cells of an adaptively refined octree grid. See the appendix of Sawant et al. (2019) for the details of convergence study. Table 1: Freestream and numerical parameters for the Mach 7 nitrogen flow. Parameters | Values ---|--- Unit Reynolds number, $Re_{1}$ | $5.22\text{\times}{10}^{4}$ Knudsen numbera | $3.2\text{\times}{10}^{-3}$ Number density, $n_{1}$/(m3) | 1022 Streamwise velocity, $u_{x,1}$/(m.s-1) | 3812 Equilibrium translational temperature, $T_{tr,1}$/(K) | 710 Surfaceb temperature, $T_{s}$/(K) | 298.5 Species mass, $m$/(kg) | $4.65\text{\times}{10}^{-26}$ Species diameter, $d$/(m) | $4.17\text{\times}{10}^{-10}$ Viscosity index, $\omega$ | 0.745 Reference temperature, Tr/(K) | 273 Parker model parameters, Zr,∞ and T∗/(K) | 18.5 and 91 Vibrational characteristic temperature, $\theta$/(K) | 3371 Domain size, ($L_{x}$,$L_{y}$,$L_{z}$)/(mm) | (80, 28.8, 80) Number of octree and sampling cells along ($X$,$Y$,$Z$) | (400, 144, 400) Number of gas-surface interaction cells along ($X$,$Y$,$Z$) | (25, 10, 25) Fnc | $6.1\text{\times}{10}^{7}$ Timestep, $\Delta t$/(ns) | 5 Adaptive mesh refinement interval /($\mu$s) | 5 Relaxation probability computation interval /($\mu$s) | 1 * • a Based on the length of the lower wedge, 50.8 mm. * • b Surface is fully accommodated (Bird, 1994), i.e., isothermal. * • c Number of actual molecules represented by a computational particle. The DSMC specific input and numerical parameters used in this work are listed in table 1. Note that the Cartesian coordinates are used with streamwise, spanwise, streamwise-normal directions as $X$, $Y$, and $Z$, respectively. The code uses the majorant frequency scheme (MFS) of Ivanov & Rogasinsky (1988) derived using the Kac stochastic model for the selection of collision pair and the variable hard sphere (VHS) model for elastic collisions. Appendix A describes an essential modification to the computation of maximum collision cross-section used in the MFS scheme for accurate spectral analysis of unsteady flows simulated on adaptively refined grids. For rotational relaxation, the Borgnakke & Larsen (1975) model is employed with rates by Parker (1959) and DSMC correction factors (Lumpkin III et al., 1991; Gimelshein et al., 2002) to account for the temperature dependence of the rotational probability. For vibrational relaxation, the semi-empirical expression of Millikan & White (1963) is used with the high-temperature correction of Park (1984). For this work, the SUGAR solver is also employed with spanwise-periodic boundary conditions as follows. Suppose a particle, during its discrete movement, intersects the spanwise domain boundary, $Y=0$ or $Y=L_{y}$, within a period, $\delta t$, smaller than the timestep. In that case, its spanwise position index is changed to the periodically opposite $Y$ boundary index, i.e., $Y=L_{y}$ or $Y=0$, respectively. After this translation, the particle continues its movement for the remaining period, $\Delta t-\delta t$. This simple algorithm is implemented in SUGAR’s parallel framework by ensuring that the processors containing a portion of the flow domain adjacent to any $Y$-boundary must also store the information of processors containing the periodically opposite portion of the domain. Such information includes the ‘location code array’ and the triangulated panels of the embedded surface. The location code arrays are special arrays used in the efficient particle mapping strategy based on the Morton-based space-filling-curve approach. See (Sawant et al., 2018) for details of these arrays and optimized gas-surface interaction strategies employed in the SUGAR code. The 2-D, steady-state solution of the flow over a double wedge, previously simulated by Tumuklu et al. (2019), is extruded in the spanwise direction ($Y$) with as many replicas as the number of spanwise octrees. See figure 1 for understanding the simulation domain setup in the $X-Z$ plane, where $X$ and $Z$ are streamwise and streamwise-normal directions. The spanwise $Y$ boundaries are periodic. From the inlet boundary, $X=0$, inward-directed ($X>0$) local Maxwellian flow is introduced at an average number density, bulk velocity, and temperature of $n_{1}$, $u_{x,1}$, and $T_{tr,1}$, respectively. Particles with the same properties are also introduced within one mean-free- path distance from the $Z$ boundaries, such that the streamlines of the flow are parallel to the $Z$-boundaries. If particles move out of the domain from either $X$ or $Z$ boundaries, they are deleted. The chosen spanwise extent of $L_{y}$=28.8 mm was estimated from a preliminary simulation with a span length of 72 mm for 30 flow times and was expected to contain four spanwise periodic structures. However, this turned out to be an underestimate, because when linear instability was detected after 50 flow times, the flow was found to exhibit a much larger spanwise wavelength. The spanwise extent of the current simulation is long enough to capture one linearly growing periodic structure. Contours and isocontours detailing spanwise periodic structures are shown with two periodic wavelengths for clarity. Note that a flow time, $T$, is defined as the time it takes for the flow to traverse a length of the separation bubble in the base (or mean) flow, $L_{s}=40$ mm, at a freestream velocity of $u_{x,1}$ where $L_{s}$ is defined as a straight-line distance from the separation point, $P_{s}$, to the reattachment point, $P_{R}$. Note that the spanwise-periodic simulation takes $\sim$5 hours per flow time using 19.2k Intel Xeon Platinum 8280 (“Cascade Lake”) processors of the Frontera supercomputer (2019). ## 3 Features of Two-dimensional Base flow Figure 1 shows the typical features of an Edney-IV type SBLI (Edney, 1968) in the base (or mean) flow, which is similar to that observed on double cones (Druguet et al., 2005; Babinsky & Harvey, 2011). The base flow macroscopic parameters are denoted by the subscript ‘$b$’. For details of the time evolution of 2-D SBLI interaction over the double wedge, see the work of Tumuklu et al. (2019). In summary, these features are formed by the interaction of a leading-edge attached (oblique) and detached (bow) shocks generated by the lower and upper wedge surfaces, respectively. This interaction generates a transmitted shock that impinges on the upper wedge surface and increases the pressure and heat flux at the reattachment (or impingement) location, $P_{R}$. The induced adverse pressure gradient results in the separation of the supersonic boundary layer on the lower wedge surface at $P_{S}$ and the formation of flow recirculation zone in the vicinity of the intersection of two surfaces, also known as the hinge. Inside the separation bubble, a shear layer is represented by the line contour of $u_{x}=0$ from $P_{S}$ to $P_{R}$. The separation zone significantly alters the SBLI system, such that the compression waves generated at the separation coalesce into a separation shock that interacts with the attached and the detached shocks at triple points. Two contact surfaces, $C_{1}$ and $C_{2}$, are formed downstream of triple points $T_{1}$ and $T_{2}$, respectively. The former is between two supersonic streams formed downstream of the separation shock, and the latter is between the lower supersonic and upper, hotter subsonic flow formed downstream of the detached shock. The transmitted shock is also affected by the contact surface $C_{1}$ and causes the reattachment point to move downstream, and the separation bubble to increase in size. A reflected shock is formed downstream of the transmitted shock to guide the supersonic stream along the upper wedge surface. If the upper wedge surface were longer, such interaction would have resulted in a $\lambda$-shock pattern, which was observed on the double cone by Tumuklu et al. (2018a, b). Instead, the flow encounters the corner of the upper wedge and goes through the Prandtl-Meyer expansion. [] [] Figure 1: (a) SBLI features shown in the magnitude of mass density gradient of the base flow, $|\nabla\rho_{b}|$, normalized by $\rho_{1}L_{s}^{-1}$, where $\rho_{1}$ is freestream mass density. Contour levels are shown in (b). (b) On same flowfield, overlay of wall-normal directions $S$ and $R$, and numerical probes $b$ inside separation bubble ($X$=48.496 mm, $Z$=24.270 mm), $r$ at reattachment (64.396, 44.358), $s$ in the separation shock (44.165, 32.597), $c$ near contact surface (65.191, 56.593), $t$ at the triple point $T_{2}$ (48.347, 41.624), $f$ in the freestream (39.212, 49.722). $S$ and $R$ directions intersect the $X$-axis at 62 and 127 mm, respectively. The initial 2-D SBLI system moves slightly downstream within the first 30 flow times because of low spanwise relaxation that leads to a decrease in pressure downstream of the primary shocks. This spanwise relaxation is induced by the thermal fluctuations of spanwise velocity about zero in the spanwise periodic simulation. This is consistent with the fact that all macroscopic quantities fluctuate about their mean (Landau & Lifshitz, 1980, chapter XII). A strictly imposed zero bulk velocity in the purely 2-D solution is unrealistic in that it does not account for such thermal fluctuations. The new 2-D flow state is defined by spanwise and temporally averaging the solution between 48 to 60 flow times. This is referred to as the base state, which fosters the growth of linear instability, detectable after 50 flow times. Note that the DSMC-derived instantaneous data at 90.5 flow times, shown in this work, _i.e._ , the boundary-layer profiles shown in section 4.1, the perturbation flow field contours shown in section 4.2, isocontours shown in section 4.4, and the perturbation field used for superposition in section 5, are noise-filtered using the POD method (see Appendix B). In spite of molecular fluctuations, DSMC allows for the detection of the onset of instability. Statistical mechanics predicts the standard deviation in the fluctuations of the directed bulk velocity such as $u_{x}$ in a gas at local equilibrium as, $\sqrt{R\langle T_{tr}\rangle/\langle N\rangle}$, where $R$, $\langle T_{tr}\rangle$, $\langle N\rangle$ are the gas constant, average translational temperature and average number of particles (Hadjiconstantinou et al., 2003; Landau & Lifshitz, 1980, chapter XII). Similarly, we can estimate the level of spanwise fluctuations about the spanwise average in a 2-D flow at local equilibrium conditions exhibiting small-amplitude, self- excited fluctuations by calculating $\sqrt{R\langle T_{tr}\rangle_{s}/\langle N\rangle_{s}}$. Subscript ‘$s$’ attached to the averaged quantities denote a spanwise average. If the DSMC-computed standard deviation is greater than the equilibrium estimate, then the fluctuations are not entirely thermal but are due to self-excited linear instability. The only exception is the finite thick region of shock layers, where additional fluctuations are present due to strong translational nonequilibrium (Sawant et al., 2020). This test was used as a first confirmation of the onset of linear instability at approximately 50 flow times, when the self-excited fluctuations in the separation bubble became slightly but noticeably larger than the thermal fluctuations. ## 4 Three-dimensional Instability Mechanisms ### 4.1 Linear instability: growth rate and spatial origin A linear instability responsible for making the 2-D base flow unstable to self-excited spanwise-homogeneous perturbations is verified in figure 2. Figure 2 shows the good comparison of the temporal evolution of perturbation rotational temperature $\tilde{T}_{rot}$, obtained from DSMC and a 2-D linear function that fits the DSMC solution. Note that the perturbation part of a macroscopic flow variable $Q\in(n,u_{x},u_{y},u_{z},T_{tr},T_{rot},T_{vib})$ is given by subtracting the 2-D base flow state $Q_{b}(x,z)$ as, $\epsilon\tilde{Q}(x,y,z,t)=Q(x,y,z,t)-Q_{b}(x,z)$ (2) Note that $\epsilon<<1$, which indicates the the perturbation is small. $\tilde{n}$ is the perturbation number density, $\tilde{u}_{x},\tilde{u}_{y},\tilde{u}_{z}$ are perturbation velocities in the $X$, $Y$, and $Z$ directions, and $\tilde{T}_{tr},\tilde{T}_{rot},\tilde{T}_{vib}$ are perturbation translational, rotational, and vibrational temperatures, respectively. A DSMC- computed perturbation flow parameter $\tilde{Q}\in(\tilde{n},\tilde{u}_{x},\tilde{u}_{y},\tilde{u}_{z},\tilde{T}_{tr},\tilde{T}_{rot},\tilde{T}_{vib})$ is fitted by a linear function written as, $\centering\begin{split}\tilde{Q}(x,y,z,t)&=\hat{Q}(x,z)\exp{(i\Theta)}+c.c.\\\ \end{split}\@add@centering$ (3) where $\hat{Q}(x,z)$ is a spanwise homogeneous amplitude function, and $\Theta$ is a phase function of the linear perturbation that has the form, $\centering\Theta=\beta y-\Omega t\@add@centering$ (4) $\beta=2\pi/L_{y}$ is a real spatial wavenumber indicating spanwise wavelength of the mode, $\Omega=\omega_{r}+i\omega_{i}$ is a complex parameter, whose real part indicates frequency and the imaginary part is the growth rate in time $t$, and $c.c.$ indicates complex conjugation so that $\tilde{Q}$ is real. A 2-D linear fit is performed using the generalized least-squares method using Python’s LMFIT (Version 1.0.1) module, which gives the mean value of unknown fit parameters, $\omega_{i}$, $\hat{Q}$, $\omega_{r}$, and $1\sigma$-uncertainty (standard error) in these parameters. These are listed in table 2. Note that by keeping $\omega_{r}$ as an unknown resulted in a small number for $\omega_{r}$ and imposing it as $\omega_{r}=0$ did not change the value of other three fit parameters, indicating that the linearly growing mode is stationary. [] [] [] Figure 2: (a) At probe $b$ inside the separation bubble, (left) temporal evolution of DSMC-derived perturbation rotational temperature, $\tilde{T}_{rot}$, normalized by freestream temperature, $T_{tr,1}$, and (right) 2-D linear fit. (b) Comparison of linear fits of all residuals at a spanwise location that corresponds to the peak. For $\tilde{T}_{rot}$, it is indicated at $Y/L_{y}$=0.88. Same holds true for $\tilde{T}_{tr}$ and $\tilde{T}_{vib}$. For $\tilde{n}$, $\tilde{u}_{x}$, and $\tilde{u}_{z}$, it is at 1.38, whereas for $\tilde{u}_{y}$, it is at 1.13. (c) Comparison of the linear fit of $\tilde{u}_{y}$ through the peak at probes $b$, $s$, $r$, $c$. For $b$ and $s$, the peak location is at $Y/L_{y}$=1.13, whereas for $r$ and $c$, it is at 0.63. Similar linear fits are performed on other DSMC-computed macroscopic flow parameters, and a 1-D extracted curve passing through the peak spanwise structure, such as that marked in figure 2 by a dashed line, is compared in figure 2. All curves are parallel to each other, indicating similar growth rates. Also, figure 2 shows the comparison of curve-fitting functions through the peak structure of $\tilde{u}_{y}$ of probe $b$ with probes at other important locations, $s$, $r$, and $c$. Nearly parallel curves are observed, which indicates that linear growth is global. By comparing the absolute values of the amplitude of $\tilde{u}_{y}$, it is seen that probes $r$, $b$, $s$, $c$ have largest to lowest amplitude, indicating decreasing magnitude of perturbation. The average of the mean growth rate for each parameter listed in table 2 is $\omega_{i}=5.0$ kHz, with bounds of +0.16% and -0.16%. A maximum deviation of 11.4% is observed at probe $c$. Table 2: 2-D linear curve fit parameters in equations 8 and 4 corresponding to figures 2 and 2. Perturbation parametera | Growth rate $\omega_{i}$/(kHz) | Amplitude $\hat{Q}$ ---|---|--- $\tilde{n}$ | 4.91 $\pm$ 0.06% | -5.013e+19 $\pm$ 0.24% $\tilde{u}_{x}$ | 4.90 $\pm$ 0.07% | -0.1613 $\pm$ 0.30% $\tilde{u}_{z}$ | 4.95 $\pm$ 0.08% | -0.1108 $\pm$ 0.33% $\tilde{T}_{tr}$ | 4.88 $\pm$ 0.04% | 0.5111 $\pm$ 0.17% $\tilde{T}_{rot}$ | 4.88 $\pm$ 0.05% | 0.5128 $\pm$ 0.19% $\tilde{T}_{vib}$ | 5.15 $\pm$ 0.11% | 0.1560 $\pm$ 0.51% $\tilde{u}_{y}$ | 4.89 $\pm$ 0.10% | 0.0762 $\pm$ 0.43% $\tilde{u}_{y}$ (at $s$) | 5.12 $\pm$ 0.26% | 0.03648 $\pm$ 1.14% $\tilde{u}_{y}$ (at $r$) | 4.77 $\pm$ 0.11% | -0.0914 $\pm$ 0.46% $\tilde{u}_{y}$ (at $c$) | 5.55 $\pm$ 0.66% | -0.0092 $\pm$ 3.20% * • a Probe locations other than $b$ are explicitly denoted. In comparison, Tumuklu et al. (2019), using the POD analysis, had found a least damped eigenmode of $-5.88$ kHz that leads the 2-D (spanwise independent) solution to reach steady state, unlike we find here. Also, our growth rate is larger than that obtained by Sidharth et al. (2018), which is consistent with their finding that a larger growth rate is expected for a larger angle difference between the upper and lower wedges. They performed a Mach 5 hypersonic flow of calorically perfect gas and obtained a nondimensional growth rate of approximately $7.5\text{\times}{10}^{-4}$ for a 12∘-20∘ double wedge (angle difference of 8∘). Following their nondimensionalization, where the growth rate is multiplied by the $\delta_{99}$ boundary-layer thickness at separation equal to 3.35 mm, and divided by the freestream velocity downstream of the leading-edge shock derived from the inviscid shock theory (Anderson, 2003) for observed shock angle of 41∘, $u_{x,2}=2930.8$, we obtain a value of $0.0057$. Now we turn to the question of the spatial origin of the linear instability and answer whether these spanwise structures seen in figure 2 start upstream, at or inside the separation bubble by comparing the boundary layer profiles at wall-normal directions $d_{1}$ to $d_{10}$ shown in figure 3. These are denoted in figure 3 on top of the contours of pressure gradient magnitude, $|\nabla p_{b}|$, in the base flow, which identifies the location of shock structure and the recirculation zone. The shear layer ($u_{x}=0$) and the separation and reattachment points are also overlaid. Along each wall-normal direction, three boundary layer profiles are shown–one in the base flow and two at $T$=90.5 on spanwise locations $Y/L_{y}$=0.88 (A) and 1.38 (B). These spanwise locations correspond to a spanwise peak and a trough of the local- streamwise (or wall-tangential) velocity so that the maximum spanwise deviation at $T=90.5$ from the base flow state can be assessed. For profiles corresponding to the lower wedge, $d_{1}$ to $d_{6}$, the local-streamwise velocity, denoted as $u_{t,l}$, is plotted as a function of wall-normal height $H_{l}$. Subscript $`t^{\prime}$ stands for the wall-tangential (or local- streamwise) component and $`l^{\prime}$ is associated with the lower wedge surface. For profiles corresponding to upper wedge, $d_{7}$ to $d_{10}$, the local-streamwise velocity, denoted as $u_{t,u}$, is plotted as a function of wall-normal height $H_{u}$. Similarly, subscript $`l^{\prime}$ is associated with the lower wedge surface. Note that $H_{l}$ and $H_{u}$ are zero at the respective surfaces. The boundary layer profiles just upstream of separation shock ($d_{1}$), at the separation ($d_{2}$), and just downstream of separation ($d_{3}$) are shown in figure 3. At $d_{1}$, all profiles overlap, indicating that the flow is 2-D upstream of the separation. Along $d_{2}$, at the separation, the absolute maximum difference of 0.72% of the freestream velocity, $u_{x,1}$, is seen between $A$ and $B$ profiles at $H_{l}/(0.1L_{y})=0.29$, which indicates spanwise modulation. The difference decreases above this height but remains nonzero even inside the shock layer, indicating the origin of linear instability inside the interaction region of the separation shock layer with the LSB. Profiles $A$ and $B$ also differ from the base state profile, indicating deviation from the base flow. Along $d_{3}$, just inside the separation bubble, $A$ and $B$ profiles deviate from each other by a maximum of 1% at $H_{l}/(0.1L_{y})=0.7$. Further inside the separation bubble, along directions $d_{4}$, $d_{5}$, and $d_{6}$, similar profiles are shown in figures 3 and 3, where the latter figure is a zoom of the rectanular boxed region denoted in the former. The absolute maximum deviation between $A$ and $B$ profiles increases along the local streamwise direction. At $d_{4}$, $d_{5}$ and $d_{6}$, it is 1.34, 1.92, 2.52% at locations $H_{l}/(0.1L_{y})=0.88,1.11,1.57$, respectively, For $d_{5}$ and $d_{6}$ directions, these profiles are on either side of their respective base profiles, indicating spanwise modulation about the base flow. On the upper wedge surface, the boundary layer profiles are shown along $d_{7}$ to $d_{10}$ in figures 3 and 3, where the latter figure is a zoom of the rectangular boxed region denoted in the former. The difference between $A$ and $B$ is even larger on the upper wedge, indicating larger amplitude of spanwise perturbations. At $d_{7}$, $d_{8}$, $d_{9}$, it is 2.8, 3.33, 3.34% at locations $H_{u}/(0.1L_{y})=0.6,0.59,0.62$, respectively, At $d_{10}$ at the reattachment location, the maximum difference decreases to 2.46% at $H_{u}/(0.1L_{y})=0.44$. The generalized inflection point (GIP) is also denoted on each boundary layer profile (open circle). Profiles $d_{1}$, $d_{2}$ and $d_{10}$ have only one GIP, whereas profiles $d_{3}$ to $d_{9}$, inside the separation bubble, have two GIPs. The GIP closest to the wall is induced in the recirculating flow between the shear layer and the surface. The GIP located farthest from the wall is induced between the shear layer and the supersonic flow outside the separation bubble. From profiles $d_{3}$ to $d_{6}$, the upper inflection point moves further away from the wall as the distance between the top enclosure of the bubble and the wall increases. The lower inflection point, more clearly seen in the respective zooms, also moves away from the surface as the distance between the shear layer and the wall increases. From profiles $d_{7}$ to $d_{9}$, both inflection points move closer to the wall. Additionally, notice that each profile exhibits a non-zero local streamwise velocity at the wall, the magnitude of which is maximum before the separation, lowest inside the separation zone on the lower wedge, and relatively larger on the upper wedge. This variation is explained by the rarefaction effects at the wall, more details of which are provided in section 4.3. [] [] [] [] [] [] Figure 3: (a) Base flow pressure gradient magnitude, $|\nabla p_{b}|$, normalized by $p_{1}L_{s}^{-1}$, where $p_{1}$ is freestream pressure. Overlaid wall-normal directions $d_{1}$ to $d_{10}$ are shown at a local streamwise distance from the hinge normalized by the length of separation $L_{s}$ as -0.625, -0.5225, -0.375, -0.25, -0.125, 0, 0.125, 0.25, 0.375, 0.51, respectively. (b) Local streamwise velocity tangential to lower wedge surface $u_{t,l}$ normalized by freestream velocity, $u_{x,1}$, versus wall- normal height at locations $d_{1}$, $d_{2}$, $d_{3}$. Insert shows zoom of the marked rectangular box. (c) Similar profiles at locations $d_{4}$, $d_{5}$, $d_{6}$. (d) Zoom of the rectanular region marked in (c). (e) Local streamwise velocity tangential to upper wedge surface, $u_{t,u}$, normalized by $u_{x,1}$ versus wall-normal height at locations $d_{7}$, $d_{8}$, $d_{9}$, $d_{10}$. (f) Zoom of the rectanular region marked in (e). Legends for (b) to (f):( ) base state profile, ( ) profile on an $X-Z$ slice passing through location $A$ ($Y/L_{y}$=0.88) at $T=90.5$, ( ) profile on an $X-Z$ slice passing through $B$ ($Y/L_{y}$=1.38) at $T=90.5$. ### 4.2 Correlation between the shock and separation bubble [] [] [] Figure 4: (a) Contours of $\tilde{u}_{y}$ normalized by $u_{x,1}$ at $T$=90.5 on a plane defined along wall-normal direction $S$, marked in figure 1. $X$ and $Y$ axes are the normalized span and the wall-normal height, respectively. (b) Perturbation pressure gradient magnitude, $|\nabla\tilde{p}|$, normalized by $p_{1}L_{s}^{-1}$ along direction $S$ as a function of wall-normal height at two spanwise locations $A$ and $B$. (c) Contours of $\tilde{u}_{y}$ normalized by $u_{x,1}$ at $T$=90.5 on plane defined along wall-normal direction $R$, marked in figure 1. Overlaid line contours: ( ) $|\nabla\tilde{p}|=$6.121\text{\times}{10}^{-3}$$ and ( ) $\tilde{\omega}_{y}$=0. The self-excited linear instability leads to the presence of spanwise periodic flow structures in perturbation flow parameters with a spanwise wavelength of $L_{y}$. Figure 4 shows the contours of spanwise perturbation velocity, $\tilde{u}_{y}$, at $T$=90.5 in the wall-normal planes $S$ and $R$ denoted in figure 1. On the $S$-plane, the spanwise periodic flow structures inside the separation bubble are seen between the surface ($H_{l}$=0) and the upper envelope of the separation bubble at $H_{l}=0.15L_{y}$ where the spanwise vorticity, $\tilde{\omega}_{y}$, is zero, as shown in figure 4. These structures have elliptical cross-sections with major and minor axes of lengths roughly equal to 0.4$L_{y}$ and 0.2$L_{y}$, respectively. Note that the upper envelope of the bubble also has a spanwise sinusoidal shape. The overlaid line contours of zero spanwise vorticity between $0<H_{l}<0.1L_{y}$ that are elliptical in shape shows a 90∘ phase shift in its spanwise mode and that of the spanwise velocity, _i.e._ , the center of the circular structure of $\tilde{\omega}_{y}$ is at $Y$=0.88$L_{y}$, inbetween a peak and a trough of $\tilde{u}_{y}$. The spanwise vorticity of the flow is negative inside these elliplical shaped contour lines of $\tilde{\omega}_{y}=0$, i.e. the flow rolls down the surface, and it is positive outside this zone and below the $\tilde{\omega}_{y}$=0 contour line at $H_{l}=0.15L_{y}$, i.e. the flow rolls up the surface. This shows that the flow moves in the spanwise direction while swirling about the spanwise axis ($Y$). Further away from the wall, figure 4 shows, for the first time, the spanwise periodic flow structures inside the strong gradient region of the separation shock ($0.36<H_{l}/L_{y}<0.44$). These structures are in phase with structures inside the separation bubble and they have the same periodicity length. This is consistent with the boundary-layer profiles shown in the previous section that showed the origin of linear instability inside the separation shock layer and the linear stability analysis that showed identical growth rate inside the LSB (probe $b$) and the separation shock (probe $s$). Note that the approximate boundary of the finite shock is marked by dashed horizontal lines corresponding to the isocontour line of normalized perturbation pressure gradient magnitude, $|\nabla\tilde{p}|=$0.612\text{\times}{10}^{-2}$$. To justify the choice of this value, figure 4 shows the variation of $|\nabla\tilde{p}|$ as a function of wall-normal height, $H_{l}$, along the $S$-plane at two spanwise locations, $A$ ($Y/L_{y}$=0.88) and $B$ ($Y/L_{y}$=1.38). The rapid increase of $|\nabla\tilde{p}|$ at $H_{l}=0.36L_{y}$ is indicative of the separation shock, inside of which the value of $|\nabla\tilde{p}|$ far exceeds that in the vicinity of the surface. Note that the thickness of the shock layer, $0.083L_{y}=2.39$ mm, is comparable the boundary-layer thickness at separation, $\delta_{99}=3.35$ mm. The locations $A$ and $B$ correspond to the peak and trough of the sinusoidal modulation of $|\nabla\tilde{p}|$ inside the separation shock. The difference between the two profiles also highlights the spanwise changes inside the shock layer. In the $R$-plane at the reattachment, a similar contour plot of $\tilde{u}_{y}$ is shown in figure 4, which exhibits spanwise periodic structures inside the reattached boundary layer. Such structures also exist in the vicinity of contour line $\tilde{\omega}_{y}$=0 at $H_{u}=0.36L_{y}$, which indicates the presence of a contact surface $C_{2}$ downstream of the triple point $T_{2}$ at the intersection of separation and detached shocks. Further away from the wall, the contour lines of $|\nabla\tilde{p}|$ at $H_{u}=0.61L_{y}$ and $0.677L_{y}$ indicate the approximate layer of detached shock, which is slightly smaller in thickness than the separation shock because the detached shock strength is higher. The spanwise structures inside this shock are not as noticeable as the separation shock. Additionally, figure 5 shows that the spanwise structures inside the separation bubble are present in the contours of all other perturbation flow parameters. Interestingly, inside the separation shock, all flow parameters exhibit spanwise modulations, as shown in the inserts of respective figures. The minimum (negative) and maximum (positive) values of spanwise structures in $\tilde{u}_{t,l}$, $\tilde{u}_{n,l}$, and $\tilde{n}$ are at spanwise location $Y/L_{y}$=0.88 ($A$) and 1.38 ($B$), respectively. All three perturbation temperatures have primary spanwise structures adjacent to the wall having minimum and maximum values at spanwise locations $Y/L_{y}$=1.38 and 0.88, respectively, i.e., 180∘ out of phase with that of velocities and number density. $\tilde{T}_{tr}$ and $\tilde{T}_{rot}$ also exhibit secondary structures right above the primary structures within $0.1<H_{l}<0.15$. Such secondary structures are also seen in $\tilde{u}_{n,l}$ and $\tilde{n}$, but are farther along the height within $0.2<H_{l}<0.35$. Furthermore, the onset of global linear instability ($T$=50) in the separation bubble is followed by the low-frequency unsteadiness of the shock structure. Figure 6 shows the spatio-temporal variation of normalized perturbation number density, $\tilde{n}$, at the triple point $T_{2}$ formed by the intersection of the detached and separation shocks. To capture one cycle of unsteadiness, the simulation had to be continued much longer up to $T$=165. Figure 6 shows that the triple point starts to oscillate at $T$=70 and its motion remains 2-D up to approximately $T$=85, as there is no variation in $\tilde{n}$ along the spanwise direction within this period. Afterword, however, linear instability begins at the triple point, which results in spanwise modulation of $\tilde{n}$. After $T$=100, we can see the presence of both the linear instability and the low-frequency unsteadiness at the triple point, where we see spanwise structures changing in time. These features are more clearly seen in figure 6 at spanwise locations $A$ and $B$. The period of oscillation is 54 $T$, which corresponds to the Strouhal number $St$ of 0.0185, defined based on the length of the separation bubble in the base flow, $L_{s}=$40 mm, and the freestream velocity, $u_{x,1}=$3812 m.s-1 as, $St=\frac{fL_{s}}{u_{x,1}}$ (5) This number is within the low-frequency range, $0.01\leq St\leq 0.05$, reported in the literature (see section 1). [] [] [] [] [] [] Figure 5: Contours of perturbation macroscopic flow parameters at $T$=90.5 on a plane defined along $S$, same as figure 4. (a) number density $\tilde{n}$ (b) local streamwise velocity (i.e., direction perpendicular to $S$), $u_{t,l}$, (c) wall-normal velocity (in the direction of $S$), $u_{n,l}$, (d) translational temperature, $\tilde{T}_{tr}$, (e) rotational temperture,$\tilde{T}_{rot}$, (f) vibrational temperature, $\tilde{T}_{vib}$. All quantities are normalized by freestream values, i.e., number density by $n_{1}$, velocities by $u_{x,1}$, and temperatures by $T_{tr,1}$. [] [] Figure 6: (a) At probe $t$ in the vicinity of the triple point $T_{2}$, denoted in figure 1, the temporal evolution of DSMC-derived perturbation number density, $\tilde{n}$, normalized by $n_{1}$, indicating low-frequency unsteadiness. (b) Normalized $\tilde{n}$ at spanwise locations $A$ and $B$, also marked in (a), indicating the period of unsteadiness. ### 4.3 Rarefaction effects in the surface parameters To understand the flow behaviour near the wall, figure 7 shows surface parameters at two spanwise locations $A$ ($Y/L_{y}=0.88$) and $B$ ($Y/L_{y}=1.38$) at the latest timestep $T$=90.5 and in the base state. Figure 7 shows local-streamwise (tangential) and spanwise velocity slips, $V_{t}$ and $V_{l}$, respectively, and figure 7 shows the local mean-free-path adjacent to the wall, $\lambda$, and the translational temperature jump at the surface, $T_{s}$. Velocity slip and temperature jump are rarefaction effects that are proportional to the Knudsen layer in the vicinity of the wall (Kogan, 1969; Chambre & Schaaf, 1961). Within this layer, two classes of molecules coexist–those reflected from the wall (in our case, diffusely), and those impinging on the wall which enters this layer from the outside region. As a result, the average velocity and temperature of the gas are different from the respective velocity and temperature of the wall. The Knudsen layer is approximately on the order of $\lambda$, the profile of which is noisy because it is obtained on the adaptively refined $C$-mesh. Note that $\lambda$ is inversely proportional to number density, $n$, and proportional to the translational temperature, $T_{tr}^{\omega-0.5}$, where $\omega=0.745$ is the viscosity index of the gas. Figure 7 shows a maximum tangential velocity slip of 2.16% of the freestream velocity at the leading edge ($X$=10 mm), which decreases along the local streamwise direction to 0.6% at $X$=32 mm. Tumuklu et al. (2019) had obtained a maximum velocity slip of 2.45% at the leading edge in their 2-D flow simulation of nitrogen over a double wedge. A large slip at the leading edge is due to the increased rarefaction of gas induced by steep gradients of the leading edge shock. It can be seen from figure 7 that $\lambda$ adjacent to the wall also follows the same behavior as $V_{t}$ in the local streamwise direction, although they are not exactly proportional to each other by a constant factor. Just upstream of the separation, $P_{S}$, within a region from $X$=32 to 36 mm, the local streamwise velocity, $u_{t,l}$, as well as $V_{t}$ decrease rapidly and become zero at the separation point, $P_{S}$ ($X$=36 mm). $\lambda$ also decreases within this region as there is a rapid increase in number density, $n$, and a decrease in translational temperature, $T_{tr}$, near the wall (not shown). Inside the recirculation zone, from $P_{S}$ to $P_{R}$, the point of reattachment, $V_{t}$ is negative because the flow impinging on the wall is opposite to the local streamwise direction. $V_{t}$ and $\lambda$ remain constant on the lower wedge, where the latter is about 3.69% of the freestream mean-free-path, $\lambda_{1}$. On the upper wedge, $V_{t}$ increases in magnitude and so does $\lambda$, as $n$ decreases and $T_{tr}$ increases in the local streamwise direction. From $P_{R}$ to the upper corner of the wedge, $V_{t}$ continues to increase similar to $\lambda$ as the rates of decrease of $n$ and increase of $T_{tr}$ are larger. At the location of expansion on the shoulder, $V_{t}$ decreases a bit before it plateaus. The profiles of $V_{t}$ at $A$, $B$, and the base state, are similar to each other, indicating no significant change so far due to linearly growing mode. The lateral slip, $V_{l}$, also remains within 0.078% on the entire surface of the wedge. [] [] [] [] Figure 7: Surface macroscopic flow parameters in the base state and at the latest time at two spanwise locations $A$ ($Y/L_{y}=0.88$) and $B$ ($Y/L_{y}=1.38$). Note that the base state profiles are time-averaged from 48 to 53 flow times and those at $A$ and $B$ from 85 to 90 flow times. (a) Surface velocity slips $V_{t}$ and $V_{y}$, normalized by $u_{x,1}$. (b) $\lambda$ adjacent to the wall normalized by freestream mean-free-path $\lambda_{1}$, and temperature jump $T_{s}$, normalized by $T_{tr,1}$. (c) The heat transfer and pressure coefficients, $C_{h}$ and $C_{p}$, respectively. (d) A zoom of the boxed regions marked in (c). The translational temperature jump, $T_{s}$ follows a similar behavior as $V_{t}$, where it is maximum at the leading-edge of the wedge and decreases up to the recirculation region, in which it remains constant on the lower wedge and increases on the upper wedge. From $P_{R}$ to the upper corner of the wedge, the rate of increase of temperature jump is larger, whereas on the shoulder, it plateaus. Also, no difference is seen in the profiles of $T_{s}$ at $A$, $B$, and the base state. Figure 7 shows the surface heat flux and pressure coefficients, $C_{h}$ and $C_{p}$, respectively. Similar to local streamwise velocity and temperature slips, $C_{h}$ is maximum at the leading edge of the wedge, decreases along the local streamwise direction, and remains at a nearly constant minimum value from the separation to the hinge. On the upper wedge surface, it increases rapidly up to the upper corner of the wedge, while the rate of increase is larger beyond $X$=61 mm. The pressure coefficient, $C_{p}$, is constant on the lower wedge, which increases sharply between $X$=32 to 38 mm, which is the local streamwise region in the vicinity of the separation point. Inside the recirculation zone on the lower wedge, $C_{p}$ is nearly constant but increases rapidly on the upper wedge up to the top corner of the wedge, where it is maximum. On the shoulder of the wedge, both coefficients decrease significantly. These coefficients are similar in value for profiles $A$, $B$, and the base state, yet figure 7 shows a zoom of the boxed region marked in figure 7, to highlight small differences in these profiles on the lower wedge surface inside the recirculation zone. $C_{h}$ is at most 11.8% higher for $A$ and 10.43% lower for $B$ than the base state, indicating spanwise modulation about the base state. $C_{p}$ is at most 0.852% higher for $A$ than $B$, while both profiles are higher than the base state, indicating a small overall increase in pressure. ### 4.4 Spanwise periodic flow structures [] [] [] [] [] [] Figure 8: At $T$=90.5, isocontour surfaces of $\tilde{u}_{y}$ normalized by $u_{x,1}$ and vorticity components $\tilde{\omega}_{x}$, $\tilde{\omega}_{y}$, $\tilde{\omega}_{z}$ normalized by the local vorticity mangitude. (a) $\tilde{u}_{y}$ in side view along with overlaid cut-boundaries $P$ and $Q$ with arrows attached to them that denote normal vectors [-0.7193 $\hat{i}$ \+ 0.6946 $\hat{k}$] and [0.7193 $\hat{i}$ \- 0.6946 $\hat{k}$], respectively. If extended, the cut-boundaries would intersect the $X$ axis at 21.3 and 14 mm, respectively. (b) $\tilde{u}_{y}$ on the normal side of cut-boundary $P$, (c) $\tilde{u}_{y}$ on the normal side of cut-boundary $Q$, (d) $\tilde{\omega}_{x}$, (e) $\tilde{\omega}_{y}$, (f) $\tilde{\omega}_{z}$. Isocontours of all vorticity components are shown on the normal side of cut- boundary $P$. In summary, the 2-D base flow is unstable to self-excited, small-amplitude, spanwise-homogeneous perturbations, and a linearly growing stationary global mode is observed, which is characterized by spanwise periodic structures in the perturbation flow fields. The spanwise perturbation velocity, $\tilde{u}_{y}=u_{y}$, which was zero at the beginning of the simulation, attains a sinusoidally varying amplitude not only inside the separation bubble but also inside shock layers and downstream of triple points. This section shows the spanwise periodic sinusoidal flow structures in $\tilde{u}_{y}$ and vorticity components. Between the wedge surface and the cut-boundary $P$, marked in figure 8, the spanwise periodic structures are shown in figure 8. The cut-boundary $P$ cuts through the outer isosurface of $\tilde{u}_{y}=0.07\%$ of $u_{x,1}$ to reveal core structures having a larger magnitude of $\tilde{u}_{y}=0.14\%$. The spanwise structures are seen to extend downstream of the reattachment and on the shoulder of the wedge. The global mode is also present in the subsonic and supersonic regions downstream of separation and detached shocks, respectively, as seen from figure 8 upstream of the cut-boundary $Q$ marked in figure 8. Such global behavior is expected due to the strong coupling of shocks and the separation bubble. Finally, the spanwise mode in the isocontours of $X$, $Y$, and $Z$ perturbation vorticity components are also shown in figures 8, 8, and 8, respectively. The $X$ and $Z$ components are in phase with each other and 90∘ out of phase with the $Y$ component. ## 5 Topology of Three-dimensional Laminar Separation Bubble This section investigates the changes in wall-streamlines and three- dimensionality inside the separation bubble by linearly superposing to the 2-D normalized base flow field, $Q_{b}$, a 3-D normalized perturbation field, $\tilde{Q}$, with a small amplitude, $\epsilon$, ranging from 0.005 to 0.1, using equation 2. Note that the velocity field of the base flow is normalized by the $X$-directional freestream velocity component, $u_{x,1}$. The perturbation velocity field at $T=90.5$ is normalized in two ways–by the absolute maximum component of velocity inside the separation bubble, i.e., inside the zone marked in figure 9 (section 5.1) and by the absolute maximum component of velocity in the entire flow field, which is located in the detached shock near the triple point $T_{2}$ (section 5.2). This distinction will highlight why one cannot draw conclusions about flow topology by decoupling the shock and a separation bubble. In the former case, the absolute maximum values of normalized $X$, $Y$, and $Z$ perturbation velocities inside the zone marked in figure 9 are 0.954, 0.455, and 1, respectively. In the latter case, these are 1, 0.0565, and 0.518, respectively. Figure 9: At $Y/L_{y}=0.38$, the $X$-perturbation velocity, $\tilde{u}_{x}$, normalized by the absolute maximum component of velocity inside the zone marked by a dashed line. Note that the zone is extended in the entire span. ### 5.1 Analysis without the coupling of shock and separation bubble [] [] [] [] Figure 10: Wall streamlines in the flow constructed by superposition of scaled 2-D base flow with scaled linear perturbations having amplitude $\epsilon$. (a) $\epsilon=0.005$, (b) $\epsilon=0.01$, (c) $\epsilon=0.05$, (d) $\epsilon=0.1$. Figure 10 shows profiles of wall-streamlines in the superposed flow field for different amplitudes, $\epsilon$. The signature observed in figure 10 typically results from small-amplitude spanwise homogeneous perturbations to the 2-D separation bubble, as was shown by Rodríguez & Theofilis (2010) in an incompressible flow. A series of critical points are formed on the separation and reattachment lines between which the wall-streamlines are slightly bent in the spanwise direction, indicating three-dimensionality of the separated flow. At a saddle point of separation, $S_{s}$, on the line of separation, the flow is attracted in the local streamwise direction and is diverted in the spanwise direction. In the middle of two $S_{s}$ points on the line of separation, a node point of separation, $N_{s}$, is formed, where the flow coming from both saddle points meets and leaves in the wall-normal direction. On the reattachment line, a node point of attachment, $N_{a}$, is formed, where the flow coming from the wall-normal direction is diverted in the spanwise and local streamwise directions. Between two $N_{a}$ points, a saddle point of attachment, $S_{a}$, is formed where the flow coming from the spanwise direction is diverted in the local streamwise direction. Figure 10 shows a similar pattern for a larger amplitude of $\epsilon=0.01$; however, now the node and saddle points on the separation and reattachment lines are not colinear in the local streamwise direction. As a result, the flow exhibits two new saddle points near the hinge, as seen in figure 10 for $\epsilon=0.05$. At the larger amplitude of $\epsilon=0.1$, the two saddle points are aligned with the node points of separation and reattachment; however, in the vicinity of the line connecting the saddle point of separation and reattachment, two counter-rotating foci, $F_{1}$ and $F_{2}$, are formed. Further increase in amplitude may lead to the merging of points $S_{a}$ and $N_{s}$ on the lower wedge and $S_{s}$ and $N_{a}$ on the upper wedge such that the node points on the separation and reattachment lines will disappear. Such a signature would resemble a simple $U$-shaped separation, first classified by Perry & Hornung (1984b). However, these speculations are beyond the purview of linear analysis. We will also see in section 5.2 that such topology cannot be studied without accounting for the perturbations in the shock. [] [] [] [] Figure 11: Three-dimensionality of the flow inside the separation bubble shown using volume streamlines. A volume line is a streamline traveling through 3-D volume data rather than being confined to a surface (Tecplot-360, 2020 R1). As in the earlier figure, the flow is constructed by superposition of scaled 2-D base flow with scaled linear perturbations having amplitude $\epsilon$. (a) $\epsilon=0.005$, (b) $\epsilon=0.01$, (c) $\epsilon=0.05$, (d) $\epsilon=0.1$. The increasing three-dimensionality of the separation bubble is seen in figure 11 for superpositions with the above amplitudes. Comparison of figures 11 and 11 shows increasing spanwise modulation of recirculating streamlines from $\epsilon=0.005$ to $\epsilon=0.01$, while the axis of rotation remains parallel to the spanwise-direction ($Y$). For $\epsilon=0.05$ in figure 11, the streamlines become 3-D, where the axes of rotation are seen to deviate from $Y$, and spanwise modulation is increased. For $\epsilon=0.1$ in figure 11, the streamlines are fully 3-D, where the axes of rotation diverge significantly from $Y$, so much that at some locations it is perpendicular to $Y$. ### 5.2 Analysis with the coupling of shock and separation bubble [] [] [] [] Figure 12: Streamwise velocity, $u_{x}$, obtained from the superposition of normalized base and perturbed flow field showing corrugations of (a) the separation shock at $H_{y}/L_{y}$=0.41 on the $S$-plane and (b) the detached shock at $H_{u}/L_{y}$=0.65 on the $R$-plane. These locations are inside the shock layer as seen from figures 4 and 4. Corresponding legends: ( ) $\epsilon=0.005$, ( ) $\epsilon=0.01$, ( ) $\epsilon=0.05$, ( ) $\epsilon=0.1$. (c) Wall streamlines and (d) volume lines inside the separation bubble for $\epsilon=0.1$. When the perturbation velocity field is normalized by maximum perturbation velocity component inside the shock, the linear coupling of shock and the separation bubble is taken into account. Figure 12 shows the features of the superposed flow field. The spanwise corrugations of the separation and detached shocks in flow fields composed with four increasing amplitudes of linear perturbations are seen in figures 12 and 12, respectively. The three- dimensionality of the separation shock is more than the detached shock and becomes prominent for the largest amplitude of $\epsilon=0.1$. The wall- streamlines in figure 12 for $\epsilon=0.1$ reveal alternate node and saddle points on the separation and reattachment lines, where the node and saddle points on the two lines are not aligned. This topology is similar to that in figure 10 for amplitude $\epsilon=0.005$, when the effect of shock was not taken into account. Similarly, the recirculation streamline inside the separation bubble shows a low degree of spanwise modulation, where the axis of rotation is the $Y$-axis for the largest amplitude of $\epsilon=0.1$ in figure 12, similar to figure 11 for $\epsilon=0.005$ (without coupling). Therefore, the coupled analysis indicates that the deviation from two-dimensionality in the shock structure dominates the deviation in the separation bubble. As a result, the study of three-dimensionality in the topology of an LSB cannot be done by ignoring their coupling with shock structures. ## 6 Conclusion The 3-D LSB induced by a laminar SBLI on a spanwise-periodic, Mach 7 hypersonic flow of nitrogen over a $30^{\circ}-55^{\circ}$ double wedge was simulated using the massively parallel SUGAR DSMC solver using billions of computational particles and collision cells on an adaptively refined octree grid. The fully resolved kinetic solution resulted in accurate modelling of the internal structure of shocks, surface rarefaction effects, thermal nonequilibrium, and time-accurate evolution of 3-D, self-excited perturbations. This is the first simulation that analyzes the linear instability of a 2-D base flow to self-excited, small-amplitude, spanwise- homogeneous perturbations in the low Reynolds number regime. In line with the findings of Tumuklu et al. (2018b) of Mach 16 flows over axisymmetric double cone and Tumuklu et al. (2019) of a 2-D, Mach 7 flow over the double wedge, the 3-D LSB was found to be strongly coupled with the separation and detached shocks. The presence of linear instability led to the formation of spanwise periodic flow structures in 3-D perturbations of macroscopic flow parameters not only inside the LSB, but also in the internal structure of the separation shock. The spanwise periodicity length of the structures at these two zones was found to be the same and their amplitude was found to grow with an average, linear temporal growth rate of 5.0 kHz $\pm$ 0.16%. We obtained a larger value of $0.0057$ for the nondimensional growth rate compared to that of Sidharth et al. (2018) for double wedges with lower angles, which is qualitatively consistent. The boundary-layer profiles in the 2-D base flow were compared with those obtained from the 3-D flow with perturbations at $T$=90.5 at two spanwise locations corresponding to the peak and trough of the spanwise sinusoidal mode. The comparison these profiles upstream and downstream as well as at the point of separation revealed that the linear instability originates in the interaction region of the separation shock with the LSB. The difference between the peak and trough of wall-tangential velocities revealed that the amplitude of perturbations increases inside the recirculation zone from the separation to the reattachment point. All boundary-layer profiles exhibited nonzero wall-tangential velocities at the wall in the Knudsen layer region. The profiles inside the separation zone also showed the presence of two GIPs, one between the wall and shear layer and the other between the shear layer and supersonic flow outside the bubble. The onset of linear instability at $T=50$ was followed by the low-frequency unsteadiness of the triple point $T_{2}$ at $T=70$. The oscillation frequency corresponds to a Strouhal number of $St\sim 0.02$, consistent with the existing literature on turbulent SBLI, but in contrast with the 3-D, finite- span double wedge simulation of Reinert et al. (2020) at a factor of eight times higher density which did not reveal such unsteadiness. To resolve these predictions, the slow linear growth and long time-scale of low-frequency unsteadiness ($\sim 0.57$ ms) suggests that experimental test times must be significantly long to capture these effects. In addition, the long-time ($T>100$) spatio-temporal evolution of the flow at the triple point $T_{2}$ revealed for the first time the presence of spanwise corrugation as well as sinusoidal oscillations in time. Finally, the topology signature in the wall-streamlines of the 3-D flow constructed by superposition of the 2-D base flow and 3-D linear perturbations was analyzed with and without accounting for the coupling between the shocks and the LSB. For a given amplitude of perturbations, significant differences were observed in the topology with versus without coupling. The analysis with coupling also revealed an increase in the corrugation of the separation and detached shocks with increase in amplitude of 3D perturbations. These findings further emphasize that, at these conditions, the 3-D changes to the topology of an LSB cannot be studied without taking into account the coupling with the shock structure. Acknowledgements. The authors acknowledge the Texas Advanced Computing Center (TACC) at the University of Texas at Austin for providing high performance computing resources on Frontera supercomputer under the Leadership Resource Allocation (LRAC) award CTS20001 of 200k SUs that have contributed to the research results reported within this paper. This work also used the Stampede2 supercomputing resources of 400k SUs provided by the Extreme Science and Engineering Discovery Environment (XSEDE) TACC through allocation TG- PHY160006. A part of the simulation was also carried out on Blue Waters supercomputer under projects ILL-BAWV and ILL-BBBK. The Blue Waters sustained- petascale computing project is supported by the National Science Foundation (awards OCI-0725070 and ACI-1238993) the State of Illinois, and as of December, 2019, the National Geospatial-Intelligence Agency. Blue Waters is a joint effort of the University of Illinois at Urbana-Champaign and its National Center for Supercomputing Applications. In addition, the authors thank Dr. Ozgur Tumuklu for providing the 2-D steady flow solution. Funding. The research conducted in this paper is supported by the Office of Naval Research under the grant No. N000141202195 titled, “Multi-scale modelling of unsteady shock-boundary layer hypersonic flow instabilities” with Dr. Eric Marineau as the program officer. Declaration of Interests. The authors report no conflict of interest. Author ORCID. Authors may include the ORCID identifers as follows. S. Sawant, https://orcid.org/0000-0002-2931-9299; D. Levin, https://orcid.org/0000-0002-6109-283X; V. Theofilis, https://orcid.org/0000-0002-7720-3434. ## Appendix A In a typical DSMC simulation, the collision pairs selected using the MFS or the no time counter (NTC) scheme are allowed to collide with probability, $P_{c}=\frac{\sigma_{T}c_{r}}{(\sigma_{T}c_{r})_{max}}$ (6) where $\sigma_{T}=\pi d^{2}$ is the total cross-section, $d$ is the molecular diameter, and $c_{r}$ is the relative speed. The maximum collision cross- section, $(\sigma_{T}c_{r})_{max}$, is stored for each collision cell and is estimated at the beginning of the simulation to a reasonably large value. Bird estimates this number as [Sec. 11.1 Bird 1994], $(\sigma_{T}c_{r})_{max}=(\pi d_{r}^{2})300\sqrt{T_{tr}/300}$ (7) where $d_{r}$ is the reference molecular diameter. As the simulation progresses, the parameter is updated if a larger value is encountered in a collision cell. However, a problem occurs at an AMR step, where the old $C$-mesh is deleted, and a new one is constructed. For the newly created collision cells, an estimate of $(\sigma_{T}c_{r})_{max}$ is required. If the parameter value is arbitrarily guessed based on equation 7, then the instantaneous temporal signals of macroscopic parameters exhibit kinks at the timesteps when the AMR step is performed. Although these kinks decay in approximately 3 to 4 $\mu$s, they can spuriously reveal a dominant frequency equal to the inverse of the time period between two AMR steps. To avoid the corruption of instantaneous signals with such artificial perturbations, at an AMR step, each root cell stores the smallest value of $(\sigma_{T}c_{r})_{max}$ among all of its collision cells before deleting the $C$-mesh. After a new $C$-mesh is formed, the value stored in the root is assigned as the lowest estimated guess to all collision cells in a given root. Those newly formed collision cells, for which the actual value of $(\sigma_{T}c_{r})_{max}$ must be larger than that assigned as an estimate, quickly update to this value within the next 0.2 $\mu$s. This strategy avoids the kinks in the instantaneous residual. ## Appendix B This appendix shows the use of the POD method (Luchtenburg et al., 2009) to remove the statistical noise in instantaneous perturbation macroscopic flow parameter fields obtained from DSMC. The use of the POD method to reduce statistical noise in a stochastic simulation can be found in a number of resources (Grinberg, 2012; Tumuklu et al., 2019). This method performs the singular value decomposition (SVD) of the input data matrix $\mathcal{D}$ formed from the solution of any given macroscopic flow parameter such that the number of rows and columns are equal to the number of total sampling cells $N_{c}$ in the DSMC domain and the instantaneous time snapshots $N_{s}$, respectively. The SVD procedure results in the decomposition, $\centering\begin{split}\mathcal{D}&=\phi\mathcal{S}\mathcal{T}\\\ \end{split}\@add@centering$ (8) where $\phi$ is the matrix of spatial modes having dimensions $N_{c}\times N_{r}$, $N_{r}$ is the user-specified rank of the reduced SVD approximation to $\mathcal{D}$, $\mathcal{S}$ is the square diagonal matrix of singular values having dimensions $N_{r}\times N_{r}$, and $\mathcal{T}$ is the matrix of temporal modes of dimensions $N_{r}\times N_{s}$. The $i^{th}$ spatial and temporal modes are stored in the $i^{th}$ column of $\phi$ and row of $\mathcal{T}$, respectively. The singular values in $\mathcal{S}$ are arranged in decreasing order, and their square corresponds to the amount of energy in the mode. After the decomposition, a reduced-order, noise-filtered representation of $\mathcal{D}$ can be constructed by forming a new data matrix $\mathcal{D}_{2}$ from a user-specified number of ranks $N_{r2}$, which is smaller than $N_{r}$. $N_{r2}$ is chosen such that the difference between any time snapshot of $\mathcal{D}_{2}$ and that of $\mathcal{D}$ is within statistical noise. [] [] [] Figure 13: (a) Modal energy in perturbation macroscopic flow parameters based on singular values obtained from proper orthogonal decomposition. (b) Contours of unfiltered (raw DSMC data) $\tilde{u}_{y}$ normalized by $u_{x,1}$ at $T$=90.5 on a plane defined along wall-normal direction $S$ as in figure 4. Overlaid on it the contour lines of noise-filtered reconstruction of $\tilde{u}_{y}$ from the first two proper orthogonal modes. (c) Comparison of unfiltered (DSMC) and filtered (POD) $\tilde{u}_{y}$ along lines $L_{1}$ and $L_{2}$ denoted in (b). For the double wedge solution, the data matrix for each macroscopic flow parameter was formed by the number of sampling cells, $N_{c}=$23.04\text{\times}{10}^{6}$$ and number of time snapshots, $N_{s}=450$. The instantaneous snapshots were collected from $T$=48.0312 to 90.9162, at an interval of 0.0953 flow time, which corresponds to the frequency of 1 MHz. Initially, $N_{r}$=10 was chosen; however, $N_{r2}=2$ was sufficient as the modal energy of higher modes is less than 10%, as shown in figure 13. The modal energy, $E_{i}$, of the $i^{th}$ mode is defined as, $\centering\begin{split}E_{i}&=\frac{S_{i}^{2}}{\sum_{j=1}^{N_{r}}S_{j}^{2}}\end{split}\@add@centering$ (9) where $S_{i}$ is the $i^{th}$ singular value. The total modal energy of the first two modes of perturbation parameters other than $\tilde{u}_{y}$ is almost 70%. For $\tilde{u}_{y}$, this number is lower because the shock structure has little influence on its flowfield, and it is composed only of a slowly growing linear mode and statistical noise. Note that the data matrix itself requires 77.24 GBs of run time memory, larger than the typical compute nodes of supercomputing clusters. Therefore, the method was parallelized based on the Tall and Skinny QR factorization (TSQR) algorithm (Sayadi & Schmid, 2016) to overcome storage requirements and speed up the SVD procedure. Figure 13 shows the original noise-contained DSMC solution of perturbation spanwise velocity at $T$=90.5 on the $S$-plane wall-normal to the lower wedge along with the noise-filtered contour lines of the solution reconstructed using POD. The figure also shows two horizontal dashed lines $L_{1}$ and $L_{2}$ along which the DSMC data is extracted and compared in figure 13. The POD- reconstructed data exhibits the same spatial spanwise variation but contains very low statistical noise compared to the DSMC solution. ## References * Anderson (2003) Anderson, John D 2003 Modern compressible flow, with historical perspective, 3rd edn. Tata McGraw-Hill. * Babinsky & Harvey (2011) Babinsky, Holger & Harvey, John K. 2011 Shock Wave–Boundary-Layer Interactions. Cambridge University Press. * Balakumar et al. (2005) Balakumar, Ponnampalam, Zhao, Hongwu & Atkins, Harold 2005 Stability of hypersonic boundary layers over a compression corner. AIAA journal 43 (4), 760–767. * Bird (1970) Bird, G. A. 1970 Aspects of the structure of strong shock waves. The Physics of Fluids 13 (5), 1172–1177. * Bird (1994) Bird, G. A. 1994 Molecular Gas Dynamics and the Direct Simulation of Gas Flows, 2nd edn. Clarendon Press. * Bird (1998) Bird, G. A. 1998 Recent advances and current challenges for DSMC. Computers & Mathematics with Applications 35 (1-2), 1–14. * Boin et al. (2006) Boin, J-Ph, Robinet, J Ch, Corre, Ch & Deniau, H 2006 3D steady and unsteady bifurcations in a shock-wave/laminar boundary layer interaction: a numerical study. Theoretical and Computational Fluid Dynamics 20 (3), 163–180. * Borgnakke & Larsen (1975) Borgnakke, Claus & Larsen, Poul S. 1975 Statistical collision model for Monte Carlo simulation of polyatomic gas mixture. Journal of Computational Physics 18 (4), 405–420. * Bruno (2019) Bruno, Domenico 2019 Direct Simulation Monte Carlo simulation of thermal fluctuations in gases. Physics of Fluids 31 (4), 047105. * Cassel et al. (1995) Cassel, K. W., Ruban, A. I. & Walker, J. D. A. 1995 An instability in supersonic boundary-layer flow over a compression ramp. Journal of Fluid Mechanics 300, 265–285. * Cercignani et al. (1999) Cercignani, Carlo, Frezzotti, Aldo & Grosfils, Patrick 1999 The structure of an infinitely strong shock wave. Physics of fluids 11 (9), 2757–2764. * Chambre & Schaaf (1961) Chambre, Paul A. & Schaaf, Samuel A. 1961 Flow of Rarefied Gases. Princeton University Press. * Chapman et al. (1958) Chapman, Dean R, Kuehn, Donald M & Larson, Howard K 1958 Investigation of separated flows in supersonic and subsonic streams with emphasis on the effect of transition. Tech. Rep. 1356\. NACA. * Chuvakhov et al. (2017) Chuvakhov, PV, Borovoy, V Ya, Egorov, IV, Radchenko, VN, Olivier, H & Roghelia, A 2017 Effect of small bluntness on formation of Görtler vortices in a supersonic compression corner flow. Journal of Applied Mechanics and Technical Physics 58 (6), 975–989. * Clemens & Narayanaswamy (2014) Clemens, Noel T & Narayanaswamy, Venkateswaran 2014 Low-frequency unsteadiness of shock wave/turbulent boundary layer interactions. Annual Review of Fluid Mechanics 46, 469–492. * Cowley & Hall (1990) Cowley, Stephen & Hall, Philip 1990 On the instability of hypersonic flow past a wedge. Journal of Fluid Mechanics 214, 17–42. * Crouch et al. (2007) Crouch, JD, Garbaruk, A & Magidov, D 2007 Predicting the onset of flow unsteadiness based on global instability. Journal of Computational Physics 224 (2), 924–940. * Czarnecki & Mueller (1950) Czarnecki, KR & Mueller, James N 1950 Investigation at mach number 1.62 of the pressure distribution over a rectangular wing with symmetrical circular-arc section and 30-percent-chord trailing-edge flap. Tech. Rep. RM L9JO5. NACA. * Dallmann (1983) Dallmann, Uve 1983 Topological structures of three-dimensional vortex flow separation. In AIAA 16th Fluid and Plasmadynamics Conference. AIAA-83-1735. * Dallmann (1985) Dallmann, U 1985 Structural stability of three-dimensional vortex flows. In Nonlinear Dynamics of Transcritical Flows, pp. 81–102. Springer. * Druguet et al. (2005) Druguet, Marie-Claude, Candler, Graham V & Nompelis, Ioannis 2005 Effects of numerics on navier-stokes computations of hypersonic double-cone flows. AIAA journal 43 (3), 616–623. * Durna & Celik (2020) Durna, AS & Celik, Bayram 2020 Effects of double-wedge aft angle on hypersonic laminar flows. AIAA Journal 58 (4), 1689–1703. * Durna et al. (2016) Durna, Ahmet Selim, El Hajj Ali Barada, Mohamad & Celik, Bayram 2016 Shock interaction mechanisms on a double wedge at Mach 7. Physics of Fluids 28 (9), 096101. * Dussauge et al. (2006) Dussauge, Jean-Paul, Dupont, Pierre & Debiève, Jean-Francois 2006 Unsteadiness in shock wave boundary layer interactions with separation. Aerospace Science and Technology 10 (2), 85–91. * Dwivedi et al. (2019) Dwivedi, Anubhav, Sidharth, G. S., Nichols, Joseph W., Candler, Graham V. & Jovanović, Mihailo R. 2019 Reattachment streaks in hypersonic compression ramp flow: an input–output analysis. Journal of Fluid Mechanics 880, 113–135. * Edney (1968) Edney, Barry E 1968 Effects of shock impingement on the heat transfer around blunt bodies. AIAA Journal 6 (1), 15–21. * Egorov et al. (2011) Egorov, Ivan, Neiland, Vladimir & Shredchenko, Vladimir 2011 Three-dimensional flow structures at supersonic flow over the compression ramp. In 49th AIAA Aerospace Sciences Meeting. AIAA 2011-730. * Elfstrom (1971) Elfstrom, GM 1971 Turbulent separation in hypersonic flow. PhD thesis, University of London, https://spiral.imperial.ac.uk/bitstream/10044/1/16361/2/Elfstrom-GM-1971-PhD-Thesis.pdf. * Elfstrom (1972) Elfstrom, GM 1972 Turbulent hypersonic flow at a wedge-compression corner. Journal of fluid Mechanics 53 (1), 113–127. * Fletcher et al. (2004) Fletcher, A. J. P., Ruban, A. I. & Walker, J. D. A. 2004 Instabilities in supersonic compression ramp flow. Journal of Fluid Mechanics 517, 309–330. * Frontera supercomputer (2019) Frontera supercomputer 2019 System hardware and software overview,https://www.tacc.utexas.edu/systems/frontera. * Gai & Khraibut (2019) Gai, Sudhir L & Khraibut, Amna 2019 Hypersonic compression corner flow with large separated regions. Journal of Fluid Mechanics 877, 471–494. * Gaitonde (2015) Gaitonde, Datta V 2015 Progress in shock wave/boundary layer interactions. Progress in Aerospace Sciences 72, 80–99. * Gallis et al. (2016) Gallis, Michail A, Koehler, TP, Torczynski, John R & Plimpton, Steven J 2016 Direct Simulation Monte Carlo investigation of the Rayleigh-Taylor instability. Physical Review Fluids 1 (4), 043403. * Gallis et al. (2015) Gallis, Michail A, Koehler, Timothy P, Torczynski, John R & Plimpton, Steven J 2015 Direct Simulation Monte Carlo investigation of the Richtmyer-Meshkov instability. Physics of Fluids 27 (8), 084105. * Garcia (1986) Garcia, Alejandro L 1986 Nonequilibrium fluctuations studied by a rarefied-gas simulation. Physical Review A 34 (2), 1454. * Gimelshein et al. (2002) Gimelshein, NE, Gimelshein, SF & Levin, DA 2002 Vibrational relaxation rates in the Direct Simulation Monte Carlo method. Physics of Fluids 14 (12), 4452–4455. * Grilli et al. (2012) Grilli, Muzio, Schmid, Peter J., Hickel, Stefan & Adams, Nikolaus A. 2012 Analysis of unsteady behaviour in shockwave turbulent boundary layer interaction. Journal of Fluid Mechanics 700, 16–28. * Grinberg (2012) Grinberg, Leopold 2012 Proper orthogonal decomposition of atomistic flow simulations. Journal of Computational Physics 231 (16), 5542–5556. * Hadjiconstantinou et al. (2003) Hadjiconstantinou, Nicolas G, Garcia, Alejandro L, Bazant, Martin Z & He, Gang 2003 Statistical error in particle simulations of hydrodynamic phenomena. Journal of computational physics 187 (1), 274–297. * Hankey Jr & Holden (1975) Hankey Jr, WL & Holden, Michael S 1975 Two-dimensional shock wave-boundary layer interactions in high speed flows. Tech. Rep. AGARDograph No. 203. AGARD. * Hashimoto (2009) Hashimoto, Tokitada 2009 Experimental investigation of hypersonic flow induced separation over double wedges. Journal of Thermal Science 18 (3), 220–225. * Holden (1963) Holden, Michael S. 1963 Heat transfer in separated flow. PhD thesis, University of London, https://spiral.imperial.ac.uk/bitstream/10044/1/16813/2/Holden-MS-1964-PhD-Thesis.pdf. * Holden (1978) Holden, Michael S. 1978 A study of flow separation in regions of shock wave-boundary layer interaction in hypersonic flow. In AIAA 11th Fluid and Plasma Dynamics Conference. AIAA 1978-1169. * Ivanov & Rogasinsky (1988) Ivanov, M. S. & Rogasinsky, S. V. 1988 Analysis of numerical techniques of the direct simulation Monte Carlo method in the rarefied gas dynamics. Russian Journal of numerical analysis and mathematical modelling 3 (6), 453–466. * Kadau et al. (2010) Kadau, Kai, Barber, John L, Germann, Timothy C, Holian, Brad L & Alder, Berni J 2010 Atomistic methods in fluid simulation. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences 368 (1916), 1547–1560. * Kadau et al. (2004) Kadau, Kai, Germann, Timothy C, Hadjiconstantinou, Nicolas G, Lomdahl, Peter S, Dimonte, Guy, Holian, Brad Lee & Alder, Berni J 2004 Nanohydrodynamics simulations: an atomistic view of the Rayleigh–Taylor instability. Proceedings of the National Academy of Sciences 101 (16), 5851–5855. * Knight et al. (2017) Knight, Doyle, Chazot, Olivier, Austin, Joanna, Badr, Mohammad Ali, Candler, Graham, Celik, Bayram, de Rosa, Donato, Donelli, Raffaele, Komives, Jeffrey, Lani, Andrea & others 2017 Assessment of predictive capabilities for aerodynamic heating in hypersonic flow. Progress in Aerospace Sciences 90, 39–53. * Knisely & Austin (2016) Knisely, Andrew M & Austin, Joanna M 2016 Geometry and test-time effects on hypervelocity shock-boundary layer interaction. In 54th AIAA Aerospace Sciences Meeting. AIAA 2016-1979. * Kogan (1969) Kogan, Maurice N. 1969 Rarefied Gas Dynamics, 1st edn. Springer US. * Korolev et al. (2002) Korolev, G. L., Gajjar, J. S. B. & Ruban, A. I. 2002 Once again on the supersonic flow separation near a corner. Journal of Fluid Mechanics 463, 173–199. * Landau & Lifshitz (1980) Landau, LD & Lifshitz, EM 1980 Statistical Physics: Part 1, , vol. 5. Pergamon Press. * Liepmann et al. (1951) Liepmann, Hans Wolfgang, Roshko, Anatol & Dhawan, Satish 1951 On reflection of shock waves from boundary layers. Tech. Rep. NACA TN 2334. California Institute of Technology, Pasadena, CA. * Lighthill (1953) Lighthill, Michael James 1953 On boundary layers and upstream influence ii. supersonic flows without separation. Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences 217 (1131), 478–507. * Lighthill (1963) Lighthill, M. J. 1963 Attachment and separation in three-dimensional flow. In Laminar Boundary Layers, Section II 2.6, Rosenhead, L. ed., pp. 72–82. Oxford University Press. * Lighthill & Newman (1953) Lighthill, Michael James & Newman, Maxwell Herman Alexander 1953 On boundary layers and upstream influence. i. a comparison between subsonic and supersonic flows. Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences 217 (1130), 344–357. * Lighthill (2000) Lighthill, Sir James 2000 Upstream influence in boundary layers 45 years ago. Philosophical Transactions of the Royal Society of London. Series A: Mathematical, Physical and Engineering Sciences 358 (1777), 3047–3061. * LMFIT (Version 1.0.1) LMFIT Version 1.0.1 Non-linear Least-Squares Minimization and Curve-Fitting for Python,https://lmfit.github.io/lmfit-py/. * Luchtenburg et al. (2009) Luchtenburg, DM, Noack, BR & Schlegel, M 2009 An introduction to the pod galerkin method for fluid flows with analytical examples and matlab source codes. Berlin Institute of Technology MB1, Muller-Breslau-Strabe 11. * Lumpkin III et al. (1991) Lumpkin III, Forrest E., Haas, Brian L. & Boyd, Iain D. 1991 Resolution of differences between collision number definitions in particle and continuum simulations. Physics of Fluids A: Fluid Dynamics 3 (9), 2282–2284. * Lusher & Sandham (2020) Lusher, David J. & Sandham, Neil D. 2020 The effect of flow confinement on laminar shock-wave/boundary-layer interactions. Journal of Fluid Mechanics 897, A18. * Millikan & White (1963) Millikan, Roger C. & White, Donald R. 1963 Systematics of vibrational relaxation. The Journal of Chemical Physics 39 (12), 3209–3213. * Moss (2001) Moss, James 2001 Dsmc computations for regions of shock/shock and shock/boundary layer interaction. In 39th Aerospace Sciences Meeting and Exhibit, p. 1027. * Moss & Bird (2005) Moss, James N & Bird, Graeme A 2005 Direct simulation monte carlo simulations of hypersonic flows with shock interactions. AIAA journal 43 (12), 2565–2573. * Needham (1965a) Needham, David A 1965a A heat-transfer criterion for the detection of incipient separation in hypersonic flow. AIAA Journal 3 (4), 781–783. * Needham (1965b) Needham, David A 1965b Laminar separation in hypersonic flow. PhD thesis, University of London, https://spiral.imperial.ac.uk/bitstream/10044/1/11846/2/Needham-DA-1965-PhD-Thesis.pdf. * Neiland (2008) Neiland, Vladimir 2008 Asymptotic theory of supersonic viscous gas flows. Butterworth-Heinemann. * Park (1984) Park, Chul 1984 Problems of rate chemistry in the flight regimes of aeroassisted orbital transfer vehicles. In 19th Thermophysics Conference, p. 1730. * Parker (1959) Parker, J. G. 1959 Rotational and vibrational relaxation in diatomic gases. The Physics of Fluids 2 (4), 449–462. * Pasquariello et al. (2017) Pasquariello, Vito, Hickel, Stefan & Adams, Nikolaus A 2017 Unsteady effects of strong shock-wave/boundary-layer interaction at high reynolds number. J. Fluid Mech 823 (617), 014602–19. * Perry & Chong (1987) Perry, A E & Chong, M S 1987 A description of eddying motions and flow patterns using critical-point concepts. Annual Review of Fluid Mechanics 19 (1), 125–155, arXiv: https://doi.org/10.1146/annurev.fl.19.010187.001013. * Perry & Hornung (1984a) Perry, A. E. & Hornung, H. G. 1984a Some aspects of three-dimensional separation. part i. streamsurface bifurcations. In Z. Flugwiss. Weltraumforsch, , vol. 8, pp. 77–87. * Perry & Hornung (1984b) Perry, A. E. & Hornung, H. G. 1984b Some aspects of three-dimensional separation. part ii. vortex skeletons. In Z. Flugwiss. Weltraumforsch, , vol. 8, pp. 155–160. * Piponniau et al. (2009) Piponniau, Sébastien, Dussauge, Jean-Paul, Debiève, Jean-François & Dupont, Pierre 2009 A simple model for low-frequency unsteadiness in shock-induced separation. Journal of Fluid Mechanics 629, 87–108. * Pirozzoli & Grasso (2006) Pirozzoli, Sergio & Grasso, Francesco 2006 Direct numerical simulation of impinging shock wave/turbulent boundary layer interaction at M= 2.25. Physics of Fluids 18 (6), 065113\. * Priebe & Martín (2012) Priebe, Stephan & Martín, M. Pino 2012 Low-frequency unsteadiness in shock wave–turbulent boundary layer interaction. Journal of Fluid Mechanics 699, 1–49. * Priebe et al. (2016) Priebe, Stephan, Tu, Jonathan H, Rowley, Clarence W & Martín, M Pino 2016 Low-frequency dynamics in a shock-induced separated flow. Journal of Fluid Mechanics 807, 441–477. * Reinert et al. (2020) Reinert, John D, Candler, Graham V & Komives, Jeffrey R 2020 Simulations of unsteady three-dimensional hypersonic double-wedge flow experiments. AIAA Journal 58 (9), 4055–4067. * Rizzetta et al. (1978) Rizzetta, D. P., Burggraf, O. R. & Jenson, Richard 1978 Triple-deck solutions for viscous supersonic and hypersonic flow past corners. Journal of Fluid Mechanics 89 (3), 535–552. * Robinet (2007) Robinet, J.-CH. 2007 Bifurcations in shock-wave/laminar-boundary-layer interaction: global instability approach. Journal of Fluid Mechanics 579, 85–112. * Rodríguez & Theofilis (2010) Rodríguez, D. & Theofilis, V. 2010 Structural changes of laminar separation bubbles induced by global linear instability. Journal of Fluid Mechanics 655, 280–305. * Roghelia et al. (2017) Roghelia, Amit, Olivier, Herbert, Egorov, Ivan & Chuvakhov, Pavel 2017 Experimental investigation of Görtler vortices in hypersonic ramp flows. Experiments in Fluids 58 (10), 139. * Rudy et al. (1991) Rudy, David H, Thomas, James L, Kumar, Ajay, Gnoffo, Peter A & Chakravarthy, Sukumar R 1991 Computation of laminar hypersonic compression-corner flows. AIAA journal 29 (7), 1108–1113. * Sansica et al. (2016) Sansica, Andrea, Sandham, Neil D. & Hu, Zhiwei 2016 Instability and low-frequency unsteadiness in a shock-induced laminar separation bubble. Journal of Fluid Mechanics 798, 5–26. * Sawant et al. (2020) Sawant, Saurabh S., Levin, Deborah A. & Theofilis, Vassilios 2020 A kinetic approach to studying low-frequency molecular fluctuations in a one-dimensional shock, arXiv: 2012.14593. * Sawant et al. (2018) Sawant, Saurabh S, Tumuklu, Ozgur, Jambunathan, Revathi & Levin, Deborah A 2018 Application of adaptively refined unstructured grids in dsmc to shock wave simulations. Computers & Fluids 170, 197–212. * Sawant et al. (2019) Sawant, Saurabh S., Tumuklu, Ozgur, Theofilis, Vassilis & Levin, Deborah A. 2019 Linear instability of shock-dominated laminar hypersonic separated flows. In The IUTAM Transition 2019 Proceedings. Springer, (accepted), arXiv: 2101.03688. * Sayadi & Schmid (2016) Sayadi, Taraneh & Schmid, Peter J 2016 Parallel data-driven decomposition algorithm for large-scale datasets: with application to transitional boundary layers. Theoretical and Computational Fluid Dynamics 30 (5), 415–428. * Schneider (2004) Schneider, Steven P 2004 Hypersonic laminar–turbulent transition on circular cones and scramjet forebodies. Progress in Aerospace Sciences 40 (1-2), 1–50. * Schrijer et al. (2009) Schrijer, F. F. J., Caljouw, R, Scarano, F & Van Oudheusden, B. W. 2009 Three dimensional experimental investigation of a hypersonic double-ramp flow. In Shock Waves, pp. 719–724. Springer. * Schrijer et al. (2006) Schrijer, F. F. J., Van Oudheusden, B. W., Dierksheide, U & Scarano, F 2006 Quantitative visualization of a hypersonic double-ramp flow using PIV and schlieren. In 12th International Symposium on Flow Visualization. German Aerospace Center (DLR). * Sidharth et al. (2018) Sidharth, GS, Dwivedi, Anubhav, Candler, Graham V & Nichols, Joseph W 2018 Onset of three-dimensionality in supersonic flow over a slender double wedge. Physical Review Fluids 3 (9), 093901. * Simeonides & Haase (1995) Simeonides, G & Haase, W 1995 Experimental and computational investigations of hypersonic flow about compression ramps. Journal of Fluid Mechanics 283, 17–42. * Smith (1986) Smith, F. T. 1986 Steady and unsteady boundary-layer separation. Annual Review of Fluid Mechanics 18 (1), 197–220. * Smith & Khorrami (1991) Smith, F. T. & Khorrami, A. Farid 1991 The interactive breakdown in supersonic ramp flow. Journal of Fluid Mechanics 224, 197–215. * Stefanov et al. (2002a) Stefanov, S, Roussinov, V & Cercignani, C 2002a Rayleigh–Bénard flow of a rarefied gas and its attractors. I. Convection regime. Physics of Fluids 14 (7), 2255–2269. * Stefanov et al. (2002b) Stefanov, S, Roussinov, V & Cercignani, C 2002b Rayleigh–Bénard flow of a rarefied gas and its attractors. II. Chaotic and periodic convective regimes. Physics of Fluids 14 (7), 2270–2288. * Stefanov et al. (2007) Stefanov, S, Roussinov, V & Cercignani, C 2007 Rayleigh–Bénard flow of a rarefied gas and its attractors. III. Three-dimensional computer simulations. Physics of Fluids 19 (12), 124101. * Stewartson (1964) Stewartson, Keith 1964 The Theory of Laminar Boundary Layers in Compressible Fluids. Clarendon Press Oxford. * Stewartson & Williams (1969) Stewartson, Keith & Williams, PG 1969 Self-induced separation. Proceedings of the Royal Society of London. A. Mathematical and Physical Sciences 312 (1509), 181–206. * Swantek & Austin (2015) Swantek, AB & Austin, JM 2015 Flowfield establishment in hypervelocity shock-wave/boundary-layer interactions. AIAA Journal 53 (2), 311–320. * Tecplot-360 (2020 R1) Tecplot-360 2020 R1 https://www.tecplot.com/products/tecplot-360/. * Theofilis et al. (2000) Theofilis, Vassilios, Hein, Stefan & Dallmann, Uwe 2000 On the origins of unsteadiness and three-dimensionality in a laminar separation bubble. Philosophical Transactions of the Royal Society of London. Series A: Mathematical, Physical and Engineering Sciences 358 (1777), 3229–3246. * Tobak & Peake (1982) Tobak, Murray & Peake, David J 1982 Topology of three-dimensional separated flows. Annual review of fluid mechanics 14 (1), 61–85. * Touber & Sandham (2009) Touber, Emile & Sandham, Neil D 2009 Large-eddy simulation of low-frequency unsteadiness in a turbulent shock-induced separation bubble. Theoretical and Computational Fluid Dynamics 23 (2), 79–107. * Tumuklu et al. (2018a) Tumuklu, Ozgur, Levin, Deborah A. & Theofilis, Vassilis 2018a Investigation of unsteady, hypersonic, laminar separated flows over a double cone geometry using a kinetic approach. Physics of Fluids 30 (4), 046103. * Tumuklu et al. (2019) Tumuklu, Ozgur, Levin, Deborah A & Theofilis, Vassilis 2019 Modal analysis with proper orthogonal decomposition of hypersonic separated flows over a double wedge. Physical Review Fluids 4 (3), 033403. * Tumuklu et al. (2018b) Tumuklu, Ozgur, Theofilis, Vassilis & Levin, Deborah A 2018b On the unsteadiness of shock–laminar boundary layer interactions of hypersonic flows over a double cone. Physics of Fluids 30 (10), 106111. * Vincenti & Kruger (1965) Vincenti, Walter Guido & Kruger, Charles H 1965 Introduction to Physical Gas Dynamics. Wiley, New York.
# Generalized Adler-Moser Polynomials and Multiple vortex rings for the Gross- Pitaevskii equation Weiwei Ao and Yehui Huang and Yong Liu and Juncheng Wei ###### Abstract. New finite energy traveling wave solutions with small speed are constructed for the three dimensional Gross-Pitaevskii equation $i\Psi_{t}=\Delta\Psi+(1-|\Psi|^{2})\Psi,$ where $\Psi$ is a complex valued function defined on ${\mathbb{R}}^{3}\times{\mathbb{R}}$. These solutions have the shape of $2n+1$ vortex rings, far away from each other. Among these vortex rings, $n+1$ of them have positive orientation and the other $n$ of them have negative orientation. The location of these rings are described by the roots of a sequence of polynomials with rational coefficients. The polynomials found here can be regarded as a generalization of the classical Adler-Moser polynomials and can be expressed as the Wronskian of certain very special functions. The techniques used in the derivation of these polynomials should have independent interest. ## 1\. Introduction In this paper, we are interested in the existence of solutions with the shape of multiple vortex rings, to the nonlinear Schrödinger type problem (1.1) $\displaystyle i\Psi_{t}\,=\triangle\Psi+\Big{(}1-|\Psi|^{2}\Big{)}\Psi,$ where $\triangle=\partial^{2}_{y_{1}}+\partial^{2}_{y_{2}}+\partial^{2}_{y_{3}}$ is the Laplacian operator in ${\mathbb{R}}^{3}$. Equation (1.1), usually called Gross-Pitaevskii equation (GP), is a well-known mathematical model arising in various physical contexts such as nonlinear optics and Bose-Einstein condensates, see for instance [30]. Traveling wave solutions of the GP equation play important role in its long time dynamics. If $\Psi$ is a traveling wave type solution of the form $\Psi(y,t)\,=\,{\tilde{u}}\big{(}y_{1},\,y_{2},\,y_{3}-ct\big{)},$ then ${\tilde{u}({\tilde{y}}_{1},{\tilde{y}}_{2},{\tilde{y}}_{3})}$ will be a solution of the nonlinear elliptic problem (1.2) $\displaystyle\,-\,i\,c\,\frac{\partial\tilde{u}}{\partial{\tilde{y}}_{3}}\,=\,\triangle{\tilde{u}}\,+\,\Big{(}1-|{\tilde{u}}|^{2}\Big{)}{\tilde{u}}.$ The existence or nonexistence of traveling wave solutions to (1.2) with $\tilde{u}\to 1$ as $|{\tilde{y}}|\to\infty$ has attracted much attention in the literature, initiated from the work of Jones, Putterman, Roberts [22, 23], where they studied the equation from the physical point of view and obtained solutions with formal and numerical calculation. They carried out their computation in dimension two and three, and find that the solution branches in these two cases have different properties. In particular, in the energy- momentum diagram, the branch in 2D is smooth, while the branch in 3D has a cusp singularity. In any case, the solutions they found have traveling wave speed $c$ less than $\sqrt{2}$ (the sound speed in this context, appears after taking the Madelung transform for the GP equation). A natural question is whether there exist solutions whose traveling speed is larger than the sound speed. In this respect, the nonexistence of finite energy solutions with $c>\sqrt{2}$ is rigorously proved by Gravejat in [18, 19]. This result is also true for $c=\sqrt{2}$ in $\mathbb{R}^{2}$, but the higher dimensional case is still open. The first rigorous mathematical proof of the existence is carried out in [10], where solutions in 2D with small traveling speed are obtained using mountain pass theorem. Later on, the existence of small speed solutions in dimension larger than two are proved in [11], also based on the mountain pass theorem. In [10], a different approach, minimizing the action functional with fixed momentum, is applied to get the existence of solutions with large momentum in dimension $N\geq 3$. This method is further developed in [8] to all dimensions, yielding existence or nonexistence of solutions for any fixed momentum. The asymptotic profile of these solutions are also studied in the above mentioned papers. In particular, for $c$ close to $0,$ in 2D, these solutions have two vortice and around them, the solution is close to the degree one vortex solution of the the Ginzburg-Landau equation; while in 3D, the solutions have the shape of a single vortex ring, see also [12]. We also refer to the paper [7] by F. Bethuel, P. Gravejat and J. Saut and the references therein for more details and discussions. The question of existence for all traveling speed $c\in(0,\sqrt{2})$ is quite delicate. It is proved by Maris in [27] that in dimension $N>2$, one can minimize the action under a Pohozaev constraint, obtaining solutions in the full speed interval $(0,\sqrt{2})$. Unfortunately, this argument breaks down in 2D, thus leaving the problem open in this dimension. Recently, Bellazzini- Ruiz [2] proved that the existence of almost all subsonic speed in 2D, using mountain pass argument. They also recovered the results of Maris in 3D. Note that when the parameter $c=0,$ equation (1.2) reduces to the Ginzburg- Landau equation: (1.3) $\Delta u+u\left(1-\left|u\right|^{2}\right)=0.$ In $\mathbb{R}^{2}$, for each $\tau\in\mathbb{Z}\backslash\left\\{0\right\\},$ it is known that the Ginzburg-Landau equation $\left(\ref{Landau}\right)$ has a degree $\tau$ vortex solution. In the polar coordinate, it has the form $S_{\tau}\left(r\right)e^{i\tau\theta}$. The function $S_{\tau}$ is real valued and vanishes exactly at $r=0.$ It satisfies $-S_{\tau}^{\prime\prime}-\frac{1}{r}S_{\tau}^{\prime}+\frac{\tau^{2}}{r^{2}}S_{\tau}=S_{\tau}\left(1-S_{\tau}^{2}\right)\text{ in }\left(0,+\infty\right).$ This equation indeed has a unique solution $S_{\tau}$ with $S_{\tau}\left(0\right)=0$ and $S_{\tau}\left(+\infty\right)=1$ and $S_{\tau}^{\prime}\left(r\right)>0.$ See [17, 31] for a proof. Recently, based on the vortex solutions of the Ginzburg-Landau equation, multi-vortex traveling wave solutions to (1.2) were constructed in [25] using Lyapunov-Schmidt reduction method. These solutions have $\frac{n(n+1)}{2}$ pairs of vortex-anti vortex configuration, where the location of the vortex points are determined by the roots of the Adler-Moser Polynomials. It is worth pointing out that the Adler-Moser polynomials arise naturally from the rational solutions of the KdV equation. We also mention that as $c$ tends to $\sqrt{2},$ a suitable rescaled traveling waves will converge to solutions of the KP-I equation, which is an important integrable system, see [6, 13]. Interestingly, the KP-I equation is actually a two dimensional generalization of the classical KdV equation. Hence in the context of GP equation, we see the KP-I equation in the transonic limit and KdV in the small speed limit. The inherent reason behind this phenomena is still to be explored. As a related result, we would like to mention that numerical simulation has been performed in [14] to illustrate the higher energy solutions of the GP equation. Denote the degree $\pm 1$ vortex solutions of the Ginzburg-Landau equation $\left(\ref{Landau}\right)$ as $v_{+}=e^{i\theta}S_{1}\left(r\right),v_{-}=e^{-i\theta}S_{1}\left(r\right).$ To better explain our main result in this paper, let us recall the following result proved in [25], which provides a family of multi-vortex solutions in dimension $2$. ###### Theorem 1.1 ([25]). In $\mathbb{R}^{2}$, for each $n\leq 34,$ there exists $c_{0}>0,$ such that for all $c\in\left(0,c_{0}\right),$ the equation $\left(\ref{TWK0}\right)$ has a solution $u_{c}$, with $u_{c}=\prod\limits_{k=1}^{n\left(n+1\right)/2}\left[v_{+}\left(z-c^{-1}p_{k}\right)v_{-}\left(z+c^{-1}p_{k}\right)\right]+o\left(1\right),$ where $p_{k}$, $k=1,...,n\left(n+1\right)/2$ are roots of the Adler-Moser polynomials. In this paper, we construct new traveling waves for $c$ close to $0$ in 3D. The solutions will have multiple vortex rings. By our construction below, it turns out that the location of the vortex points are closely related to the following system (Balancing condition): (1.4) $\left\\{\begin{array}[c]{l}{\displaystyle\sum\limits_{j=1,j\neq k}^{m}}\frac{1}{\mathbf{a}_{k}-\mathbf{a}_{j}}-{\displaystyle\sum\limits_{j=1}^{n}}\frac{1}{\mathbf{a}_{k}-\mathbf{b}_{j}}=-n,\text{ for }k=1,...,m,\\\ -{\displaystyle\sum\limits_{j=1,j\neq k}^{n}}\frac{1}{\mathbf{b}_{k}-\mathbf{b}_{j}}+{\displaystyle\sum\limits_{j=1}^{m}}\frac{1}{\mathbf{b}_{k}-\mathbf{a}_{j}}=-m,\text{ for }k=1,...,n.\end{array}\right.$ Here $\mathbf{a}_{j},j=1,...,m,$ $\mathbf{b}_{\ell},\ell=1,...,n$, are complex numbers in the $z=x_{1}+ix_{2}$ plane. The integer $m$ actually denotes the number of positively oriented vortex rings and $n$ denotes the number of negatively oriented ones. Moreover, the solvability of our original problem is related to the nondegeneracy of the linearized operator $dF$ of the map $F$ defined by (5.2). The make our construction possible, the solution $\mathbf{a}_{j},\mathbf{b}_{\ell}$ to the system (1.4) has to satisfy some symmetric properties. We therefore introduce the following condition: ($\mathcal{M}$). $m>n$. The points $\mathbf{a}_{j},\mathbf{b}_{\ell}$, $j=1,...,m$, $\ell=1,...,n$ are all distinct. The set of points of $\\{\mathbf{a}_{1},...,\mathbf{a}_{m}\\}$ and $\\{\mathbf{b}_{1},...,\mathbf{b}_{n}\\}$ are both symmetric with respect to the $x_{1}$ axis. We use Lyapunov-Schmidt reduction method to construct multi-vortex ring solutions. Our main result is the following: ###### Theorem 1.2. Suppose $\mathbf{a}_{j},\mathbf{b}_{\ell}$, $j=1,...,m$, $\ell=1,...,n$ is a solution of (1.4) satisfying condition $\mathcal{M}$ and the linearized operator $dF$ of (5.2) is non-degenerate at this solution in the sense defined in Section 5. Then for all $\varepsilon>0$ sufficiently small, there exists an axially symmetric solution $u=u(\sqrt{y_{1}^{2}+y_{2}^{2}},y_{3})$ to equation (2.2), and $u$ has $m$ positively oriented vortex rings and $n$ negatively oriented vortex rings. The distance of the vortex rings to the axis is of the order $O(\varepsilon^{-1})$, while the mutual distance of two vortex rings are of the order $O(\varepsilon^{-1}/|\ln\varepsilon)|$. After scaling back by the factor $\varepsilon|\ln\varepsilon|$, the position of the vortex rings in the $(x_{1},x_{2})$ plane is close to suitable $x_{1}$-translation of those points $\\{\mathbf{a}_{1},...,\mathbf{a}_{m},\mathbf{b}_{1},...,\mathbf{b}_{n}\\}$. More precise description of the solutions can be found in the course of the proof. From the proof in Section 5, we can see that in the case of two positively oriented vortex rings and one negatively oriented vortex ring, there exists solutions to (1.4) and the corresponding linearized operator of (5.2) is non-degenerate. Hence one can construct traveling wave solutions with three vortex rings. We also show in Section 5.2 and Section 6 that (1.4) has solutions satisfying $\mathcal{M}$, provided that $m=n+1$. (Surprisingly, if $m-n>1$, we have not found any solutions satisfying $\mathcal{M}$.) When $m=n+1$ the location of the vortex points are determined by the roots of generating polynomials which have recurrence relations and can be explicitly written down using certain Wronskians. These generating polynomials are natural generalizations of the classical Adler-Moser polynomials. We refer to Section 6 for more details. Let us point out that traveling wave solutions of the Schrodinger map equation with single vortex ring has been constructed in [24]. In principle, our method in this paper can also be applied to this equation and other related equation such as the Euler equation. The dimension three case (with obvious extension to higher dimensions) studied in the present paper actually has some new properties compared to the 2D case. Roughly speaking, the main difference of the 2D and 3D case is the following. In 2D, the vortex location of our solutions is determined by the Adler-Moser polynomials. These polynomials can be obtained by method of integrable systems and are well studied. However, in 3D, due to the presence of additional terms in the equation, the vortex location is not determined by Adler-Moser polynomials. Indeed they are determined by a sequence of polynomials, which can be regarded as a generalization of Adler-Moser polynomials, and up to our knowledge, are new. We have to find these new generating polynomials using some techniques from the theory of integrable systems. This step is nontrivial and may have independent interest. If we rescale the Gross-Pitaevskii equation $x=\epsilon^{-1}\bar{x}$, then the distance between the locations of the vortex rings obtained in Theorem 1.2 is of the order ${\mathcal{O}}(\frac{1}{|\log\epsilon|})$. Note that this distance is much smaller than the leapfrogging region in which the distance between the vortex rings is of the order ${\mathcal{O}}(\frac{1}{\sqrt{|\log\epsilon|}})$. For the dynamics of vortex rings in the leapfrogging region for the Gross-Pitaevskii equation we refer to Jerrard-Smets [21] and the references therein. The paper is organized as follows. In Section 2, we formulate the 3D problem as a two dimensional one. In Section 3, we introduce the approximate multi- vortex ring solutions and estimate their error. Section 4 is devoted to the study of a nonlinear projected problem. This is more or less standard. The main part of the paper is Section 5 and Section 6, where we get the reduced problem for the position of the vortex points and study some generating polynomials whose roots determine the location of the vortex rings. Acknowledgement W. Ao is supported by NSFC no. 11631011, no. 11801421, and no. 12071357. Y. Liu is partially supported by NSFC no. 11971026 and “The Fundamental Research Funds for the Central Universities WK3470000014”. J. Wei is partially supported by NSERC of Canada. ## 2\. Formulation of the problem We are looking for a solution to problem (1.1) in the form $\Psi(y,t)\,=\,{\tilde{u}}\Big{(}y_{1},y_{2},y_{3}-ct\Big{)}.$ Then $\tilde{u}$ must satisfy (2.1) $-ic\frac{\partial\tilde{u}}{\partial y_{3}}=\Delta\tilde{u}+(1-|\tilde{u}|^{2})\tilde{u}.$ Let $\varepsilon>0$ be a small parameter. We would like to seek solutions with traveling speed $c=\varepsilon|\ln\varepsilon|$. Equation (1.2) then becomes (2.2) $-i\varepsilon|\ln\varepsilon|\frac{\partial\tilde{u}}{\partial y_{3}}=\Delta\tilde{u}+(1-|\tilde{u}|^{2})\tilde{u}.$ We require the solution $\tilde{u}$ satisfies $\tilde{u}(y)\to 1\mbox{ as }|y|\to\infty.$ We are interested in the solutions axially symmetric with respect to the $y_{3}$ axis. Let us introduce $x_{1}=\sqrt{y^{2}_{1}+y^{2}_{2}},x_{2}=y_{3},$ and $z=x_{1}+ix_{2},\,u(x_{1},x_{2})=\tilde{u}(y_{1},y_{2},y_{3}).$ Then we get the following equation satisfied by $u$: (2.3) $-i\varepsilon|\ln\varepsilon|\frac{\partial u}{\partial x_{2}}=\Delta_{(x_{1},x_{2})}u+\frac{1}{x_{1}}\frac{\partial u}{\partial x_{1}}+(1-|u|^{2})u,$ with boundary conditions $\frac{\partial}{\partial x_{1}}u(0,x_{2})=0,\,u\to 1\mbox{ as }|z|\to\infty.$ Observe that the problem (6.2) is invariant under the following two transformations: $u(z)\to\overline{u(\bar{z})},\,u(z)\to u(-\bar{z}).$ Thus we impose the following symmetry on the solutions $u$: $\Sigma=\\{u(z)=\overline{u(\bar{z})},\,u(z)=u(-\bar{z}).\\}$ This symmetry will play an important role in our analysis. As a conclusion, if we write $u(x_{1},x_{2})=u_{1}(x_{1},x_{2})+iu_{2}(x_{1},x_{2}),$ then $u_{1}$ and $u_{2}$ enjoy the following conditions: (2.4) $\displaystyle\begin{aligned} u_{1}(x_{1},x_{2})=u_{1}(-x_{1},x_{2}),&\qquad u_{1}(x_{1},x_{2})=u_{1}(x_{1},-x_{2}),\\\ u_{2}(x_{1},x_{2})=u_{2}(-x_{1},x_{2}),&\qquad u_{2}(x_{1},x_{2})=-u_{2}(x_{1},-x_{2}),\\\ \frac{\partial u_{1}}{\partial x_{1}}(0,x_{2})=0,&\qquad\frac{\partial u_{2}}{\partial x_{1}}(0,x_{2})=0.\end{aligned}$ We now have a two dimensional elliptic system with Neumann boundary condition $\frac{\partial u}{\partial x_{1}}(0,x_{2})=0$. Compared with the two dimensional problem studied in [25], there are two differences: Firstly, there is an extra term $\frac{1}{x_{1}}\frac{\partial u}{\partial x_{1}}$; Secondly, the coefficient in front of $\frac{\partial u}{\partial x_{2}}$ becomes $\varepsilon|\ln\varepsilon|$, instead of $\varepsilon$. Some remarks are in order. We aim to construct multi-vortex ring solutions to (2.2). For single vortex ring, one can use $v^{+}(x-p)v^{-}(x+p)$ as a good approximate solution for the equation (6.2). But for multi-vortex rings, the vortex-anti vortex pairs are not good enough because of the extra term $\frac{1}{x_{1}}\frac{\partial u}{\partial x_{1}}$ and the Neumann boundary condition. So we need to use more accurate approximate solution which we will explain in the next section. ## 3\. The approximate solution In this section, we would like to define a family of approximate solutions for the equation (6.2). ### 3.1. The first approximate solution We consider $\mathcal{K}$ distinct points $p_{j}=(p_{j,1},p_{j,2})$, $j=1,...,\mathcal{K}$, lying in the right half of the $z$ plane. Let us define $p_{j}^{*}=-\bar{p}_{j}$ for $j=1,\cdots,\mathcal{K}$. We also denote $p_{\mathcal{K}+j}=p_{j}^{*}$ for $j=1,\cdots,\mathcal{K}$. Intuitively, these points represent the location of the vortex rings. We also suppose that the set of points $\\{p_{1},...,p_{\mathcal{K}}\\}$ is symmetric with respect to the $x_{1}$ axis. Moreover, we will assume: (A1.) (3.1) $\displaystyle\rho$ $\displaystyle:=min_{\ell\neq j,1\leq\ell,j\leq\mathcal{K}}\\{|p_{\ell}-p_{j}|\\}\sim\frac{1}{\varepsilon|\ln\varepsilon|},$ and $|p_{j,1}|\sim\frac{1}{\varepsilon}$ for $j=1,\cdots,\mathcal{K}.$ n order to understand more clearly the difference between the 2D and 3D case, let us now following the strategy of [25] to define an approximate solution. Let $S=S_{1}$ be the function associated to the degree one vortex solution of the Ginzburg-Landau equation, defined in the first section. Define $u_{j}:=S(|z-p_{j}|)e^{i\tau_{j}\theta_{j}},\,\ j=1,\cdots,\mathcal{K},$ where $\theta_{j}$ is the angle around $p_{j}$, $\tau_{j}=+1\mbox{ or}-1$, corresponding to the degree $\pm 1$ vortex. We then set $u_{j}=S(|z-p_{j}|)e^{-i\tau_{j-\mathcal{K}}\theta_{j}},\,\ j=\mathcal{K}+1,\cdots,2\mathcal{K}$ where $\theta_{j}$ is the angle around $p_{j}$. The reason of defining these functions is the following: Projecting a vortex ring onto the $z$ plane, we get two circles in the right and left plane with different orientation. Here $u_{j}$ and $u_{j+\mathcal{K}}$ can be viewed as a vortex-antivortex pair. We now define the first approximate solution as (3.2) $U=\Pi_{j=1}^{2\mathcal{K}}u_{j}.$ We will see that this approximate solution is not good enough to handle the 3D case and later on we will introduce a refined approximate solution. Note that at this moment, we still haven’t decided the sign of the degree of the vortex. This will also be done later on. Since each vortex in the right half plane has a vortex in the left plane with opposite sign, we can check directly that $U\to 1$ as $|z|\to\infty$. We will see that the approximate solution satisfies the boundary and symmetry condition (2.4). In fact, by the choice of the vortex points, one has ###### Lemma 3.1. The approximate solution $U$ has the following symmetry: $U(\bar{z})=\bar{U}(z),\,\,U(z^{*})=U(z)$ where $z^{*}=-\bar{z}$. ###### Proof. This is the result by the definition of the vortex points. Since $v_{-}(z)=\overline{v_{+}(z)}$, one has $\begin{split}U(z^{*})&=\Pi_{j=1}^{\mathcal{K}}[u_{j}(z^{*}-p_{j})u_{\mathcal{K}+j}(z^{*}-p_{j}^{*})]\\\ &=\Pi_{j=1}^{\mathcal{K}}[u_{j}((z-p_{j}^{*})^{*})u_{\mathcal{K}+j}((z-p_{j})^{*})]\\\ &=\Pi_{j=1}^{\mathcal{K}}[u_{\mathcal{K}+j}(z-p_{j}^{*})u_{j}(z-p_{j})]=U(z),\end{split}$ and since the points $\\{p_{j}\\}$ are invariant with respect to the reflection across the $x_{1}$ axis, we have $\begin{split}U(\bar{z})&=\Pi_{j=1}^{\mathcal{K}}[u_{j}(\bar{z}-p_{j})u_{\mathcal{K}+j}(\bar{z}-p_{j}^{*})]\\\ &=\Pi_{j=1}^{\mathcal{K}}[\bar{u}_{j}(z-\bar{p}_{j})\bar{u}_{\mathcal{K}+j}(z-\bar{p}_{j}^{*})]=\bar{U}(z).\end{split}$ This finishes the proof. ∎ ### 3.2. The error of the first approximate solution Firstly, we estimate the error of the first approximate solution $U$. Since it satisfies the symmetry and boundary condition (2.4), one only need to consider in the domain $\\{x_{1}>0\\}$. Recall that the degree $\pm 1$ vortex satisfies $S^{\prime\prime}(r)+\frac{1}{r}S^{\prime}(r)-\frac{1}{r^{2}}S(r)+(1-S^{2})S=0.$ It has the following properties([29]): ###### Lemma 3.2. The vortex solution satisfies the following properties: * (i). $S(0)=0,\,S^{\prime}(r)>0,\,S(r)\in(0,1)$; * (ii.) $S(r)=1-\frac{1}{2r^{2}}+O(\frac{1}{r^{4}})$ as $r\to\infty$; * (iii). $S(r)=a_{0}r-\frac{a_{0}}{8}r^{3}+O(r^{5})$ as $r\to 0$ where $a_{0}$ is a positive constant. In this subsection, we are going to estimate the error caused by the first approximation $U$. Use $E_{1}$ to denote the error : $\displaystyle E_{1}=i\varepsilon|\ln\varepsilon|\frac{\partial U}{\partial x_{2}}+\Delta U+(1-|U|^{2})U+\frac{1}{x_{1}}\frac{\partial U}{\partial x_{1}}.$ We have $\begin{split}\Delta U&=\Delta(u_{1}\cdots u_{2\mathcal{K}})\\\ &=\sum_{\ell=1}^{\mathcal{K}}(\Delta u_{\ell}\Pi_{j\neq\ell}u_{j})+\sum_{j\neq\ell}\nabla u_{\ell}\cdot\nabla u_{j}\Pi_{t\neq\ell,j}u_{t}.\end{split}$ On the other hand, writing $\rho_{j}=|u_{j}|^{2}-1$, one has $\begin{split}|U|^{2}-1&=\Pi_{j=1}^{2\mathcal{K}}(1+\rho_{j})-1\\\ &=\sum_{j}\rho_{j}+\sum_{j=1}^{2\mathcal{K}}\mathcal{Q}_{j},\end{split}$ where $\mathcal{Q}_{j}=\sum_{i_{1}<i_{2}<\cdots<i_{i}}(\rho_{i_{1}}\cdots\rho_{i_{i}})$. Using the fact that $\Delta u_{j}-\rho_{j}u_{j}=0$, we get $\begin{split}\Delta U+(1-|U|^{2})U&=\sum_{\ell\neq j}(\nabla u_{\ell}\cdot\nabla u_{j}\Pi_{t\neq\ell,j}u_{t})-U\sum_{j=2}^{2\mathcal{K}}\mathcal{Q}_{j}.\end{split}$ Let $\varphi_{0}=\sum_{j=1}^{\mathcal{K}}{\tau_{j}}(\theta_{j}-\theta_{\mathcal{K}+j}),$ and $\,r_{j}=|z-p_{j}|,\,r_{\mathcal{K}+j}=|z-q_{j}|,r_{j,1}=x_{1}-p_{j,1},r_{j,2}=x_{2}-p_{j,2}.$ We have $\begin{split}\frac{1}{x_{1}}\frac{\partial U}{\partial x_{1}}&=\frac{1}{x_{1}}\Big{[}e^{i\varphi_{0}}\frac{\partial}{\partial x_{1}}(\Pi_{j=1}^{\mathcal{K}}S(r_{j})S(r_{\mathcal{K}+j}))+ie^{i\varphi_{0}}\frac{\partial\varphi_{0}}{\partial x_{1}}\Pi_{j=1}^{\mathcal{K}}S(r_{j})S(r_{\mathcal{K}+j})\Big{]}\\\ &=\Big{(}\sum_{j=1}^{2\mathcal{K}}\frac{1}{x_{1}}\frac{S^{\prime}(r_{j})}{S(r_{j})}\frac{r_{j,1}}{r_{j}}+\frac{i}{x_{1}}\frac{\partial\varphi_{0}}{\partial x_{1}}\Big{)}U.\end{split}$ Similarly, there holds $\begin{split}\frac{\partial U}{\partial x_{2}}&=\sum_{j=1}^{2\mathcal{K}}\partial_{x_{2}}u_{j}\Pi_{\ell\neq j}u_{\ell}\\\ &=\Big{(}\sum_{j=1}^{2\mathcal{K}}\frac{S^{\prime}(r_{j})}{S(r_{j})}\frac{r_{j,2}}{r_{j}}+i\frac{\partial\varphi_{0}}{\partial x_{2}}\Big{)}U.\end{split}$ Combining the above computations, we obtain $\begin{split}E_{1}&=\sum_{\ell\neq j}\frac{(\nabla u_{\ell}\cdot\nabla u_{j})}{u_{\ell}\,u_{j}}U-U\sum_{j=2}^{2\mathcal{K}}\mathcal{Q}_{j}\\\ &+\Big{(}\sum_{j=1}^{2\mathcal{K}}\frac{1}{x_{1}}\frac{S^{\prime}(r_{j})}{S(r_{j})}\frac{r_{j,1}}{r_{j}}+\frac{i}{x_{1}}\frac{\partial\varphi_{0}}{\partial x_{1}}\Big{)}U\\\ &+i\varepsilon|\ln\varepsilon|\Big{(}\sum_{j=1}^{2\mathcal{K}}\frac{S^{\prime}(r_{j})}{S(r_{j})}\frac{r_{j,2}}{r_{j}}+i\frac{\partial\varphi_{0}}{\partial x_{2}}\Big{)}U.\end{split}$ In the sequel, we denote $|p_{j,1}|$ by $d_{j}$. Direct computation yields $\begin{split}\frac{\partial\varphi_{0}}{\partial x_{2}}&=\sum_{j=1}^{\mathcal{K}}\tau_{j}\big{[}\frac{\partial\theta_{j}}{\partial x_{2}}-\frac{\partial\theta_{\mathcal{K}+j}}{\partial x_{2}}\big{]}\\\ &=\sum_{j=1}^{\mathcal{K}}\tau_{j}\Big{[}\frac{x_{1}-p_{j,1}}{r^{2}_{j}}-\frac{x_{1}-p^{*}_{j,1}}{r^{2}_{\mathcal{K}+j}}\Big{]}\\\ &=\sum_{j=1}^{\mathcal{K}}\tau_{j}\frac{2d_{j}(x_{1}^{2}-p_{j,1}^{2}-(x_{2}-p_{j,2})^{2})}{r_{j}^{2}r_{\mathcal{K}+j}^{2}}.\end{split}$ We also have $\begin{split}\frac{1}{x_{1}}\frac{\partial\varphi_{0}}{\partial x_{1}}&=\frac{1}{x_{1}}\sum_{j=1}^{\mathcal{K}}\tau_{j}\Big{[}\frac{\partial\theta_{j}}{\partial x_{1}}-\frac{\partial\theta_{\mathcal{K}+j}}{\partial x_{1}}\Big{]}\\\ &=-\frac{1}{x_{1}}\sum_{j=1}^{\mathcal{K}}\tau_{j}\Big{[}\frac{x_{2}-p_{j,2}}{r^{2}_{j}}-\frac{x_{2}-p^{*}_{j,2}}{r^{2}_{\mathcal{K}+j}}\Big{]}.\end{split}$ Observe that $\frac{1}{x_{1}}\frac{\partial\varphi_{0}}{\partial x_{1}}$ contributes to the imaginary part of the error $E_{1}$. Note that away from the vortex point $p_{j}$, this decays only at the rate $O(r_{j}^{2})$, which is not sufficient for our construction. Hence the vortex-antivortex pair is not enough to be a good approximate solution. ### 3.3. The reference vortex ring In order the get rid of these singularities, one need more accurate approximations for the vortex ring. In [21], leap frogging behavior of the vortex rings to the GP equation has been analyzed. Indeed, our construction in this paper is partly inspired by these leap frogging behavior. Following the analysis performed in [21], we introduce the potential function $A_{a}$, which satisfies the following equation: (3.3) $\left\\{\begin{array}[]{l}-div\Big{(}\frac{1}{x_{1}}\nabla(x_{1}A_{a})\Big{)}=2\pi\delta_{a}\mbox{ in }H,\\\ A_{a}=0\mbox{ on }\partial H,\end{array}\right.$ where $H=\\{(x_{1},x_{2})\in\mathbb{R}^{2},\,x_{1}>0\\}$ and $a\in H$. For the region $\\{x_{1}\leq 0\\}$, we consider the odd extension of $A_{a}$. The expression of $A_{a}$ can be integrated explicitly in terms of complete elliptic integrals (see [20, 21]). We emphasize that in the literature, there are different notations concerning the definition of complete elliptic integrals, mainly about its arguments. Let $r:=r_{a}=|z-a|$. When $r_{a}=o(|a_{1}|)$, one has the following asymptotic behavior (3.4) $A_{a}(z)=\Big{(}\ln\frac{a_{1}}{r_{a}}+3\ln 2-2\Big{)}+O\Big{(}\frac{r_{a}}{a_{1}}|\ln\frac{r_{a}}{a_{1}}|\Big{)}$ and (3.5) $\partial_{r}A_{a}=-\frac{1}{r}+O(\frac{1}{a_{1}}),$ and for $x_{1}\to 0$ (3.6) $A_{a}(x_{1},x_{2})=\frac{x_{1}a_{1}^{2}}{a_{1}^{3}+x_{2}^{2}}\mbox{ as }\frac{x_{1}}{a_{1}}\to 0.$ Up to a constant phase factor, there exists a unique unimodular map $u_{a}^{*}\in C^{\infty}(H\setminus\\{a\\},S^{1})$ such that (3.7) $x_{1}(iu_{a}^{*},\nabla u_{a}^{*})=x_{1}j(u_{a}^{*})=-\nabla^{\perp}(x_{1}A_{a}),$ where $j(u)=u\times\nabla u=(iu,\nabla u)=Re(iu\nabla\bar{u}).$ In the sense of distribution, we have $\left\\{\begin{array}[]{l}div(x_{1}j(u_{a}^{*}))=0,\\\ curl(j(u_{a}^{*}))=2\pi\delta_{a},\end{array}\right.$ and the function $u_{a}^{*}$ corresponds to a singular vortex ring centered at $a$. If we denote by $u_{a}^{*}=e^{i\varphi_{a}}$, then by (3.7), one has (3.8) $\partial_{1}\varphi_{a}=\partial_{2}A_{a},\,\partial_{2}\varphi_{a}=-\frac{1}{x_{1}}\frac{\partial(x_{1}A_{a})}{\partial x_{1}}.$ So from the definition of $\varphi_{a}$ and the boundary condition of $A_{a}$, one has $\left\\{\begin{array}[]{l}\Delta\varphi_{a}+\frac{1}{x_{1}}\frac{\partial\varphi_{a}}{\partial x_{1}}=0\mbox{ in }H,\\\ \partial_{1}\varphi_{a}(0,x_{2})=0\mbox{ on }\partial H.\end{array}\right.$ Moreover, using the relation of $\varphi_{a}$ and $A_{a}$ in (3.8), one has (3.9) $\nabla\varphi_{a}=\frac{1}{r_{a}}\nabla^{\perp}r_{a}+O(\frac{1}{x_{1}}\log\frac{a_{1}}{r_{a}})\mbox{ for }\frac{r_{a}}{a_{1}}\to 0$ and (3.10) $|\nabla\varphi_{a}|\leq\frac{C}{1+r_{a}}\mbox{ for }r_{a}\geq 1.$ So near the vortex point $a$, $\nabla\varphi_{a}$ can be viewed as a perturbation of $\nabla\theta_{a}$. ### 3.4. Improvement of the first approximate solution We will use $u_{a}^{*}$ instead of the vortex-anti vortex pair $e^{i(\theta_{j}-\theta_{\mathcal{K}+j})}$ to define a more accurate approximate vortex ring. In view of the symmetry condition (2.4), the vortex ring associated to a point $a\in H$ will defined to be $S(|z-a|)S(|z-a^{*}|)u_{a}^{*}(x)=S(|z-a|)S(|z-a^{*}|)e^{i\varphi_{a}(z)}.$ We can also decompose $\varphi_{a}$ as $\varphi_{a}(z)=\theta_{a}(z)-\theta_{a^{*}}(z)+\tilde{\varphi}_{a}.$ Note that the difference $\tilde{\varphi}_{a}$ can be analyzed around the vortex point $a$ using the asymptotic behavior of $A$. Define $\varphi_{d}=\sum_{j=1}^{\mathcal{K}}\tau_{j}\varphi_{p_{j}}=\varphi_{0}+\sum_{j=1}^{\mathcal{K}}\tau_{j}\tilde{\varphi}_{p_{j}}:=\varphi_{0}+\tilde{\varphi}_{\bf p}.$ Then our final approximation will be defined as $\mathcal{U}(x)=U(x)e^{i\tilde{\varphi}_{\bf p}}=\Pi_{j=1}^{2\mathcal{K}}S_{p_{j}}(x)e^{i\varphi_{d}}.$ Namely, we replace the function $\varphi_{0}$ by $\varphi_{d}$ in the first approximate solution. Since $\\{p_{j}\\}$ satisfies (A1), one can see that the new approximate solution will satisfy the symmetry condition (2.4). ### 3.5. Error of the final approximation Now the new error becomes $\begin{split}E_{2}&=i\varepsilon|\ln\varepsilon|\frac{\partial\mathcal{U}}{\partial x_{2}}+\Delta\mathcal{U}+(1-|\mathcal{U}|^{2})\mathcal{U}+\frac{1}{x_{1}}\frac{\partial\mathcal{U}}{\partial x_{1}}\\\ &:=E_{21}+E_{22}.\end{split}$ Here $E_{21}$ is the first term in the left hand side. We use $B_{l}(p)$ to denote the ball of radius $l$ centered at the point $p$. We have the following error estimate: ###### Lemma 3.3. There exists a constant $C$ such that for all small $\varepsilon$ and all points $p_{j}$ satisfying (A1), we have $\sum_{j=1}^{2\mathcal{K}}\|E_{2}\|_{L^{9}(B_{3}(p_{j}))}\leq C\varepsilon|\ln\varepsilon|.$ Moreover, we have $E_{2}=i\mathcal{U}[R_{1}+iR_{2}]$, with $R_{1},\,R_{2}$ real valued and $\begin{split}|R_{1}|&\leq C\sum_{j=1}^{2\mathcal{K}}\frac{O(\varepsilon^{1-\delta})}{(1+r_{j})^{3}},\\\ |R_{2}|&\leq C\sum_{j=1}^{2\mathcal{K}}\frac{O(\varepsilon^{1-\delta})}{1+r_{j}},\end{split}$ for any $\delta\in(0,1)$, if $|z-p_{j}|>1$ for all $j$. ###### Proof. We compute, in $B_{\frac{\rho}{5}}(p_{j})$, (3.11) $\begin{split}\frac{\partial\mathcal{U}}{\partial x_{2}}&=\frac{\partial\Pi_{j=1}^{2\mathcal{K}}S_{j}}{\partial x_{2}}e^{i\varphi_{d}}+i\mathcal{U}\nabla\varphi_{d}\\\ &=\big{[}\sum_{j=1}^{2\mathcal{K}}\frac{S^{\prime}(r_{j})}{S(r_{j})}\frac{r_{j,2}}{r_{j}}+i\nabla\varphi_{d}\big{]}\mathcal{U}.\end{split}$ Hence in $(\cup_{j=1}^{2\mathcal{K}}B_{3}(p_{j}))^{c}$, by (3.10), we have $\begin{split}Re\Big{[}\frac{E_{21}}{i\mathcal{U}}\Big{]}&=\varepsilon|\ln\varepsilon|\big{[}\sum_{j=1}^{2\mathcal{K}}\frac{S^{\prime}(r_{j})}{S(r_{j})}\frac{r_{j,2}}{r_{j}}\big{]}\\\ &\leq C\sum_{j=1}^{2\mathcal{K}}\frac{O(\varepsilon^{1-\delta})}{(1+r_{j})^{3}},\\\ Im\Big{[}\frac{E_{21}}{i\mathcal{U}}\Big{]}&=\varepsilon|\ln\varepsilon|[\nabla\varphi_{d}]\\\ &\leq C\sum_{j=1}^{2\mathcal{K}}\frac{O(\varepsilon^{1-\delta})}{(1+r_{j})}.\end{split}$ We also have $\|i\varepsilon|\ln\varepsilon|\,\partial_{x_{2}}\mathcal{U}\|_{L^{9}(\cup_{j}\\{r_{j}\leq 3\\})}\leq C\varepsilon|\ln\varepsilon|.$ Note that the $L^{\infty}$ norm is not bounded near $p_{j}$, due to the presence of $\ln r_{j}$ term. Next, letting $S_{j}=S(r_{j})$ and using the fact that $\Delta S_{j}-\frac{S_{j}}{r_{j}^{2}}+(1-S_{j}^{2})S_{j}=0,$ one has (3.12) $\begin{split}E_{22}&=\Delta\mathcal{U}+(1-|\mathcal{U}|^{2})\mathcal{U}+\frac{1}{x_{1}}\frac{\partial\mathcal{U}}{\partial x_{1}}\\\ &=\mathcal{U}\Big{[}\sum_{j=1}^{2\mathcal{K}}\frac{1}{r_{j}^{2}}-|\nabla\varphi_{d}|^{2}+\frac{1}{x_{1}}\sum_{j=1}^{2\mathcal{K}}\frac{S^{\prime}(r_{j})}{S(r_{j})}\partial_{x_{1}}r_{j}-\sum_{j=1}^{2\mathcal{K}}\mathcal{Q}_{j}\\\ &+2i\sum_{j=1}^{2\mathcal{K}}\frac{S^{\prime}(r_{j})}{S(r_{j})}\nabla r_{j}\cdot\nabla\varphi_{d}\Big{]}\end{split}$ where we have used the fact that $\Delta\varphi_{d}+\frac{1}{x_{1}}\frac{\partial}{\partial x_{1}}\varphi_{d}=0.$ By carefully checking the terms, using (3.4)-(3.10), away from the vortex points, one has $\begin{split}Re\Big{[}\frac{E_{22}}{i\mathcal{U}}\Big{]}&\leq C\sum_{j=1}^{2\mathcal{K}}\frac{O(\varepsilon^{1-\delta})}{(1+r_{j})^{3}},\\\ Im\Big{[}\frac{E_{22}}{i\mathcal{U}}\Big{]}&\leq C\sum_{j=1}^{2\mathcal{K}}\frac{O(\varepsilon^{1-\delta})}{(1+r_{j})}.\end{split}$ Moreover, $\|E_{22}\|_{L^{9}(\cup_{j}\\{r_{j}\leq 3\\})}\leq C\varepsilon|\ln\varepsilon|.$ Combining the estimates for $E_{21}$ and $E_{22}$, we obtain the desired estimates. ∎ ## 4\. Linear theory Now we set up the reduction procedure. The linear theory we use here will be the same one as that of [25]. We recall the framework developed there in the sequel. As usual, we shall look for a solution of (6.2) in the form: (4.1) $u:=\left(\mathcal{U}+\mathcal{U}\eta\right)\chi+\left(1-\chi\right)\mathcal{U}e^{\eta},$ where $\chi$ is a cutoff function such that $\chi(x)=\sum_{j=1}^{2\mathcal{K}}\tilde{\chi}(x-p_{j})$ and $\tilde{\chi}(s)=1$ for $s\leq 1$ and $\tilde{\chi}(s)=0$ for $s\geq 2$ and $\eta=\eta_{1}+\eta_{2}i$ is complex valued function close to $0$ in suitable norm which will be introduced below. We also assume that $\eta$ has the same symmetry as $\mathcal{U}.$ Note that near the vortice, $u$ is obtained from $\mathcal{U}$ by an additive perturbation; while away from the vortice, $u$ is of the form $\mathcal{U}e^{\eta}$. The reason of choosing the perturbation $\eta$ in the form (4.1) is explained in Section 3 of [16] and also in [25]. Essentially, the form of the perturbation far away from the origin makes it easier to handle the decay rates of the error away from the origin. The conditions imposed on $u$ in (2.4) can be transmitted to $\eta$: (4.2) $\displaystyle\begin{aligned} \eta_{1}(x_{1},x_{2})=\eta_{1}(-x_{1},x_{2}),&\qquad\eta_{1}(x_{1},x_{2})=-\eta_{1}(x_{1},-x_{2}),\\\ \eta_{2}(x_{1},x_{2})=\eta_{2}(-x_{1},x_{2}),&\qquad\eta_{2}(x_{1},x_{2})=\eta_{2}(x_{1},-x_{2}),\\\ \frac{\partial\eta_{1}}{\partial x_{1}}(0,x_{2})=0,&\qquad\frac{\partial\eta_{2}}{\partial x_{1}}(0,x_{2})=0.\end{aligned}$ In view of (4.1), we can write $u=\mathcal{U}e^{\eta}+\gamma,$ where $\gamma:=\chi\mathcal{U}\left(1+\eta-e^{\eta}\right).$ Note that $\gamma$ is localized near the vortex points and of the order $o(\eta),$ for $\eta$ small. Set $\mathcal{A}:=\left(\chi+\left(1-\chi\right)e^{\eta}\right)\mathcal{U}.$ Then $u$ can be written as $u=\mathcal{U}\eta\chi+\mathcal{A}.$ Following the computation in [25], we get $\left(1-\left|u\right|^{2}\right)u=\left(\mathcal{U}\eta\chi+\mathcal{A}\right)\left(1-\left|\mathcal{U}e^{\eta}+\gamma\right|^{2}\right).$ The equation for $\eta$ becomes (4.3) $-\mathcal{A}\mathbb{L}\left(\eta\right)=\left(1+\eta\right)\chi E_{2}\left(\mathcal{U}\right)+\left(1-\chi\right)e^{\eta}E_{2}\left(\ U\right)+N_{0}\left(\eta\right),$ where $E_{2}\left(\mathcal{U}\right)$ represents the error of the approximate solution $\mathcal{U}$, and (4.4) $\mathbb{L}\eta:=i\varepsilon\frac{\partial\eta}{\partial x_{2}}+\Delta\eta+2u^{-1}\nabla u\cdot\nabla\eta-2\left|u\right|^{2}\eta_{1}+\frac{1}{x_{1}}\frac{\partial\eta}{\partial x_{1}},$ while $N_{0}$ is $o(\eta),$ and explicitly given by $\displaystyle N_{0}\left(\eta\right)$ $\displaystyle:=\left(1-\chi\right)\mathcal{U}e^{\eta}\left|\nabla\eta\right|^{2}+i\varepsilon|\ln\varepsilon|\left(\mathcal{U}\left(1+\eta-e^{\eta}\right)\right)\partial_{x_{2}}\chi$ $\displaystyle+\frac{1}{x_{1}}\left(\mathcal{U}\left(1+\eta-e^{\eta}\right)\right)\partial_{x_{1}}\chi+2\nabla\left(\mathcal{U}\left(1+\eta-e^{\eta}\right)\right)\cdot\nabla\chi+\mathcal{U}\left(1+\eta-e^{\eta}\right)\Delta\chi$ $\displaystyle-2\mathcal{U}\left|\mathcal{U}\right|^{2}\eta\eta_{1}\chi-\left(\mathcal{A}+\mathcal{U}\eta\chi\right)\left[\left|\mathcal{U}\right|^{2}\left(e^{2\eta_{1}}-1-2\eta_{1}\right)+\left|\gamma\right|^{2}+2\operatorname{Re}\left(\mathcal{U}e^{\eta}\bar{\gamma}\right)\right].$ Let us write this equation as (4.5) $\mathbb{L}\left(\eta\right)=-{\mathcal{U}}^{-1}E_{2}\left(\mathcal{U}\right)+N\left(\eta\right),$ where $\begin{split}N(\eta)&=-\left|\mathcal{U}\right|^{2}\left(e^{2\eta_{1}}-1-2\eta_{1}\right)+\left|\nabla\eta\right|^{2}\\\ &+i\varepsilon|\ln\varepsilon|\mathcal{A}^{-1}\left(\mathcal{U}\left(1+\eta-e^{\eta}\right)\right)\partial_{x_{2}}\chi+\frac{1}{\mathcal{A}x_{1}}\left(\mathcal{U}\left(1+\eta-e^{\eta}\right)\right)\partial_{x_{1}}\chi\\\ &+2\mathcal{A}^{-1}\nabla\left(\mathcal{U}\left(1+\eta-e^{\eta}\right)\right)\cdot\nabla\chi\\\ &+\mathcal{A}^{-1}\mathcal{U}\left(1+\eta-e^{\eta}\right)\Delta\chi-\mathcal{A}^{-1}\mathcal{U}\chi\left|\nabla\eta\right|^{2}-\left|\gamma\right|^{2}-2\operatorname{Re}\left(\mathcal{U}e^{\eta}\bar{\gamma}\right)\\\ &+\mathcal{A}^{-1}\mathcal{U}\eta\chi\left[{\mathcal{U}}^{-1}E_{2}\left(\mathcal{U}\right)-2\left|\mathcal{U}\right|^{2}\eta_{1}-\left|\mathcal{U}\right|^{2}\left(e^{2\eta_{1}}-1-2\eta_{1}\right)-\left|\gamma\right|^{2}-2\operatorname{Re}\left(\mathcal{U}e^{\eta}\bar{\gamma}\right)\right].\end{split}$ This nonlinear equation, equivalent to the original GP equation, is the one we eventually want to solve. Observe that in $N\left(\eta\right)$, except $\left|\mathcal{U}\right|^{2}\left(e^{2\eta_{1}}-1-2\eta_{1}\right)-\left|\nabla\eta\right|^{2},$ other terms are all localized near the vortex points. ### 4.1. A Linear problem By the definition of our vortex configuration, one can see that the terms contain $\varepsilon|\ln\varepsilon|$ and $\frac{1}{x_{1}}$ can be viewed as small perturbation near the vortex points. Let us first consider the following linear problem: (4.6) $\mathbb{L}(\eta)=h,\,Re\int_{\mathbb{R}^{2}}\overline{\mathcal{U}\eta}Z_{\ell j}=0,\,\eta\,\mbox{ satisfies }\,(\ref{bdyofpsi}),$ where $Z_{\ell j}=\alpha_{\ell}\nabla u_{\ell}\,\tilde{\rho}_{\ell}(x),\alpha_{\ell}=\frac{\mathcal{U}}{u_{\ell}},$ and $\tilde{\rho}_{\ell}$ is a cutoff function centered at $p_{\ell}$ with support in $B_{\frac{\rho}{5}}(p_{\ell})$. We shall establish a priori estimates for this problem. The following weighted norms and linear theory has been studied in [25]. Recall that $r_{j},j=1,\cdot\cdot\cdot,\mathcal{K},$ represent the distance to the $j$-th vortex point. Let $w$ be a weight function defined by $w(z):=\left(\sum_{j=1}^{2\mathcal{K}}\left(1+r_{j}\right)^{-1}\right)^{-1}.$ This function measures the minimal distance from the point $z$ to those vortex points. We use $B_{a}\left(z\right)$ to denote the ball of radius $a$ centered at $z.$ Let $\alpha,\sigma\in\left(0,1\right)$ be small positive numbers. For complex valued function $\eta=\eta_{1}+\eta_{2}i,$ we define the following weighted norm: $\displaystyle\left\|\eta\right\|_{\ast}$ $\displaystyle=\left\|u\eta\right\|_{W^{2,9}\left(w<3\right)}+\left\|w^{1+\sigma}\eta_{1}\right\|_{L^{\infty}\left(w>2\right)}+\left\|w^{2+\sigma}(|\nabla\eta_{1}|+|\nabla^{2}\eta_{1}|)\right\|_{L^{\infty}\left(w>2\right)}$ $\displaystyle+\sup_{z\in\left\\{w>2\right\\}}\sup_{z_{1},z_{2}\in B_{w/3}\left(z\right)}\left(\frac{\left|\nabla\eta_{1}\left(z_{1}\right)-\nabla\eta_{1}\left(z_{2}\right)\right|+\left|\nabla^{2}\eta_{1}\left(z_{1}\right)-\nabla^{2}\eta_{1}\left(z_{2}\right)\right|}{w\left(z\right)^{-2-\sigma-\alpha}\left|z_{1}-z_{2}\right|^{\alpha}}\right)$ $\displaystyle+\left\|w^{\sigma}\eta_{2}\right\|_{L^{\infty}\left(w>2\right)}+\left\|w^{1+\sigma}\nabla\eta_{2}\right\|_{L^{\infty}\left(w>2\right)}+\left\|w^{2+\sigma}\nabla^{2}\eta_{2}\right\|_{L^{\infty}\left(w>2\right)}$ $\displaystyle+\sup_{z\in\left\\{w>2\right\\}}\sup_{z_{1},z_{2}\in B_{w/3}\left(z\right)}\left(w\left(z\right)^{1+\sigma+\alpha}\frac{\left|\nabla\eta_{2}\left(z_{1}\right)-\nabla\eta_{2}\left(z_{2}\right)\right|}{\left|z_{1}-z_{2}\right|^{\alpha}}\right)$ $\displaystyle+\sup_{z\in\left\\{w>2\right\\}}\sup_{z_{1},z_{2}\in B_{w/3}\left(z\right)}\left(w\left(z\right)^{2+\sigma+\alpha}\frac{\left|\nabla^{2}\eta_{2}\left(z_{1}\right)-\nabla^{2}\eta_{2}\left(z_{2}\right)\right|}{\left|z_{1}-z_{2}\right|^{\alpha}}\right).$ Basically, the norm means that the real part of $\eta$ decays like $w^{-1-\sigma}$ and its first and second derivatives decay like $w^{-2-\sigma}$. Moreover, the imaginary part of $\eta$ only decays as $w^{-\sigma}$, but its first and second derivative decay as $w^{-1-\sigma}$ and $w^{-2-\sigma}$ respectively. It is worth mentioning that the Hölder norms are taken into account in the definition because eventually we shall use the Schauder estimates. Moreover, near the vortex points, we use the $L^{p}$ norm, because the $L^{\infty}$ norm is not bounded there. On the other hand, for complex valued function $h=h_{1}+ih_{2},$ we define the following weighted Hölder norm $\displaystyle\left\|h\right\|_{\ast\ast}$ $\displaystyle:=\left\|uh\right\|_{L^{9}\left(w<3\right)}+\left\|w^{1+\sigma}h_{1}\right\|_{L^{\infty}\left(w>2\right)}$ $\displaystyle+\left\|w^{2+\sigma}\nabla h_{1}\right\|_{L^{\infty}\left(w>2\right)}+\left\|w^{2+\sigma}h_{2}\right\|_{L^{\infty}\left(w>2\right)}$ $\displaystyle+\sup_{z\in\left\\{w>2\right\\}}\sup_{z_{1},z_{2}\in B_{w/3}\left(z\right)}\left(w\left(z\right)^{2+\sigma+\alpha}\frac{\left|\nabla h_{1}\left(z_{1}\right)-\nabla h_{1}\left(z_{2}\right)\right|}{\left|z_{1}-z_{2}\right|^{\alpha}}\right)$ $\displaystyle+\sup_{z\in\left\\{w>2\right\\}}\sup_{z_{1},z_{2}\in B_{w/3}\left(z\right)}\left(w\left(z\right)^{2+\sigma+\alpha}\frac{\left|h_{2}\left(z_{1}\right)-h_{2}\left(z_{2}\right)\right|}{\left|z_{1}-z_{2}\right|^{\alpha}}\right).$ This definition tells us that the real and imaginary parts of $h$ have different decay rates. Moreover, intuitively we require $h_{1}$ to gain one more power of decay at infinity after taking one derivative. The choice of this norm is partly decided by the decay and smooth properties of $E_{2}(\mathcal{U})$. We have the following a priori estimate for solutions of the equation (4.6). ###### Lemma 4.1 (Proposition 4.5 in [25]). Let $\varepsilon>0$ be small. Suppose $\eta$ is a solution of (4.6) with $\left\|h\right\|_{\ast\ast}<\infty$. Then $\left\|\eta\right\|_{\ast}\leq C\varepsilon^{-\sigma}\left|\ln\varepsilon\right|\left\|h\right\|_{\ast\ast}$ where $C$ is a constant independent of $\varepsilon$ and $h$. We now consider the following linear projected problem: (4.7) $\left\\{\begin{array}[]{l}\mathbb{L}(\eta)=h+\sum_{j=1}^{2\mathcal{K}}\sum_{j=1}^{2}c_{\ell j}Z_{\ell j},\\\ Re\int_{\mathbb{R}^{2}}\overline{{\mathcal{U}}\eta}Z_{\ell j}\,dx=0,\\\ \eta\mbox{ satisfies }(\ref{bdyofpsi}).\end{array}\right.$ We state the following existence result: ###### Proposition 4.2. There exists constant $C$, depending only on $\alpha,\,\sigma$ such that for all $\varepsilon$ small, the following holds: if $\|h\|_{**}<\infty$, there exists a unique solution $(\eta,\\{c_{\ell j}\\})=T_{\varepsilon}(h)$ to (4.7). Furthermore, there holds $\|\eta\|_{*}\leq C\varepsilon^{-\sigma}|\ln\varepsilon|\|h\|_{**}.$ ###### Proof. The proof is similar to that of Proposition 4.1 in [16]. Instead of solving (4.7) in $\mathbb{R}^{2}$, we solve it in a bounded domain first: $\left\\{\begin{array}[]{l}\mathbb{L}(\eta)=h+\sum_{\ell=1}^{2\mathcal{K}}\sum_{j=1}^{2}c_{\ell j}Z_{\ell j},\,Re\int_{B_{M}}\overline{\mathcal{U}\eta}Z_{\ell j}\,dx=0\mbox{ in }B_{M}(0)\\\ \eta=0\mbox{ on }\partial B_{M}(0),\\\ \eta\mbox{ satisfies \, the \, condition \, (\ref{bdyofpsi})},\end{array}\right.$ where $M$ large enough. By the standard proof of a priori estimates, we also obtain the following estimates for any solution $\eta_{M}$ of the above problem with $\|\eta_{M}\|_{*}\leq C\varepsilon^{-\sigma}|\ln\varepsilon|\|h\|_{**}.$ By working in the Sobolev space $H_{0}^{1}(B_{M})$, the existence will follow by Fredholm alternatives. Now letting $M\to\infty$, we obtain a solution of the required properties. ∎ ### 4.2. Projected nonlinear problem From now on, we will denote by $T_{\varepsilon}(h)$ the solution of (4.7). We consider the following nonlinear projected problem : (4.8) $\left\\{\begin{array}[]{l}\mathbb{L}(\eta)+\frac{E_{2}(\mathcal{U})}{\mathcal{U}}+N(\eta)=\sum_{\ell=1}^{2\mathcal{K}}\sum_{j=1}^{2}c_{\ell j}Z_{\ell j},\\\ Re\int_{\mathbb{R}^{2}}\overline{\mathcal{U}\eta}Z_{\ell j}\,dx=0,\\\ \eta\mbox{ satisfies \, the \, condition }\,(\ref{bdyofpsi}).\end{array}\right.$ Using the operator $T_{\varepsilon}$ defined in Proposition 4.2, we can write the above problem as $\eta=T_{\varepsilon}(\frac{E_{2}(\mathcal{U})}{\mathcal{U}}-N(\eta)):=G_{\varepsilon}(\eta).$ Using the error estimates in Lemma 3.3, we have for $r_{j}\sim\varepsilon^{-1}$, $Re(\frac{E_{2}}{\mathcal{U}})\sim\frac{\varepsilon^{1-\delta}}{r_{j}},\,Im(\frac{E_{2}}{\mathcal{U}})\sim\frac{\varepsilon^{1-\delta}}{r_{j}^{3}}.$ More precisely, if one check the express of the error, and using the explicate expression for $A_{a}$ in Section 2, one can check by direct calculation that for $r_{j}>>\varepsilon^{-1}$, $Re(\frac{E_{2}}{\mathcal{U}})\sim\frac{\varepsilon^{1-\delta}}{r_{j}^{2}},\,Im(\frac{E_{2}}{\mathcal{U}})\sim\frac{\varepsilon^{1-\delta}}{r_{j}^{3}}.$ Taking this into account, one has $\|{\mathcal{U}}^{-1}E_{2}(\mathcal{U})\|_{**}\leq C\varepsilon^{1-\delta}$ for any $\delta>0$. Let $\eta\in B:=\\{\|\eta\|_{*}\leq C\varepsilon^{1-\beta}\\}$ for $\beta\in(\delta+\sigma,1)$. Then using the explicit form of $N(\eta)$, we have $\|G_{\varepsilon}(\eta)\|_{*}\leq C(\|N(\eta)\|_{**}+\|{\mathcal{U}}^{-1}E_{2}(\mathcal{U})\|_{**})\leq C\varepsilon^{1-\beta}$ and $\|G_{\varepsilon}(\eta)-G_{\varepsilon}(\tilde{\eta})\|_{*}\leq o(1)\|\eta-\tilde{\eta}\|_{*}$ for all $\eta,\,\tilde{\eta}\in B$. By contraction mapping theorem, we obtain the following: ###### Proposition 4.3. There exists constant $C$ and $\beta$ small, depending only on $\alpha,\,\sigma$ such that for all $\varepsilon$ small, the following holds: there exists a unique solution $(\eta_{\varepsilon,\\{p_{i}\\}},\\{c_{ij}\\})=T_{\varepsilon}(h)$ to (4.8). Furthermore, there holds $\|\eta\|_{*}\leq C\varepsilon^{1-\beta},$ and $\eta_{\varepsilon,\\{p_{i}\\}}$ is continuous in $\\{p_{i}\\}$. ## 5\. The reduced problem and the multiple vortex rings solutions ### 5.1. The reduced problem To find a real solution to problem (4.5), we solve the reduced problem by finding the positions of the vortex points $\\{p_{i}\\}$ such that the coefficients $c_{\ell j}$ in (4.8) are zero for small $\varepsilon$. In the previous section, we have deduced the existence of $\eta$ to the projected nonlinear problem: $\mathbb{L}\eta+\frac{E_{2}}{\mathcal{U}}+N(\eta)=\sum_{j}c_{j}\frac{\nabla u_{j}}{u_{j}}\tilde{\rho}_{j}(x).$ So $c_{j}=0$ is equivalent to (5.1) $Re\int_{\mathbb{R}^{2}}u_{j}[\mathcal{L}\eta+\frac{E_{2}}{\mathcal{U}}+N(\eta)]\nabla\bar{u}_{j}dx=0.$ By the relation of $u_{j}\mathbb{L}$ and $L_{0}(\phi_{j})$ where $\phi_{j}=u_{j}\eta$, $u_{j}\mathbb{L}(\eta)=L_{0}(\phi_{j})+o(\frac{1}{\rho^{2}})\phi_{j},$ where $L_{0}(\phi)=\Delta\phi+(1-S^{2})\phi-2Re(\bar{u}_{0}\phi)u_{0}.$ One has, by integration by parts, $\begin{split}Re\int_{\mathbb{R}^{2}}u_{j}\mathbb{L}\eta\,\nabla\bar{u}_{j}dx&=Re\int_{\mathbb{R}^{2}}(L_{0}+O(\frac{1}{\rho^{2}})\phi_{j}\,\nabla\bar{u}_{j}dx\\\ &=Re\int_{\mathbb{R}^{2}}\phi_{j}L_{0}(\nabla\bar{u}_{j})dx+o(\varepsilon)=o(\varepsilon),\end{split}$ and using the expression of $N(\psi)$, $Re\int_{\mathbb{R}^{2}}u_{j}N(\eta)\,\nabla\bar{u}_{j}dx=o(\varepsilon).$ We now compute $Re\int_{\mathbb{R}^{2}}u_{j}\frac{E_{2}}{\mathcal{U}}\nabla\bar{u}_{j}dx.$ Recall that one can write $\mathcal{U}$ as $u_{j}\alpha_{j}$, where $\alpha_{j}=\pi_{\ell\neq j}u_{\ell}e^{i\tilde{\varphi}_{\bf p}}$, near each vortex point $p_{j}$. By (3.11) in Section 3, we have $\begin{split}&Re\int_{\mathbb{R}^{2}}\frac{E_{21}}{\alpha_{j}}{\nabla\bar{u}_{j}}dx\\\ &=Re(i\varepsilon|\ln\varepsilon|)\int_{\mathbb{R}^{2}}\big{[}\frac{S^{\prime}(r_{j})}{S(r_{j})}\frac{r_{j,2}}{r_{j}}+i\partial_{x_{2}}\varphi_{d}\big{]}\Big{(}\frac{S^{\prime}(r_{j})}{S(r_{j})}\nabla r_{j}-i{\tau_{j}}\nabla\theta_{j}\Big{)}S^{2}(r_{j})dx\\\ &+o(\varepsilon)\\\ &=-\varepsilon|\ln\varepsilon|{\tau_{j}}\int_{\mathbb{R}^{2}}SS^{\prime}(r)\Big{(}\partial_{x_{2}}\varphi_{d}-{\tau_{j}}\frac{x_{2}}{r}\nabla\theta\Big{)}dx+o(\varepsilon)\\\ &=-\varepsilon|\ln\varepsilon|{\tau_{j}}\int_{\mathbb{R}^{2}}SS^{\prime}(r)\Big{(}\frac{x_{1}}{r^{2}}\nabla r-\frac{x_{2}}{r}\nabla\theta\Big{)}dx+o(\varepsilon)\\\ &=(-\pi\varepsilon|\ln\varepsilon|{\tau_{j}},\,0)+o(\varepsilon),\end{split}$ where we have used the estimate (3.6). On the other hand, by (3.12), $\begin{split}&Re\int_{\mathbb{R}^{2}}\frac{E_{22}}{\alpha_{j}}{\nabla\bar{u}}_{j}dx\\\ &=\int_{\mathbb{R}^{2}}\Big{(}\sum_{\ell}\frac{1}{r_{\ell}^{2}}-|\nabla\varphi_{d}|^{2}+\frac{1}{x_{1}}\sum_{\ell}\frac{S^{\prime}(r_{\ell})}{S(r_{\ell})}\partial_{x_{1}}r_{\ell}\Big{)}S(r_{j})S^{\prime}(r_{j})\nabla r_{j}dx\\\ &+2{\tau_{j}}\int\sum_{\ell}\frac{S^{\prime}(r_{\ell})}{S(r_{\ell})}\nabla r_{\ell}\cdot\nabla\varphi_{d}\nabla\theta_{j}S^{2}(r_{j})dx+o(\varepsilon)\\\ &=-\int_{\mathbb{R}^{2}}|\nabla\varphi_{d}|^{2}\nabla r_{j}S(r_{j})S^{\prime}(r_{j})dx\\\ &+\frac{1}{p_{j,1}}\int_{\mathbb{R}^{2}}(S^{\prime}(r))^{2}\partial_{x_{1}}r\nabla rdx\\\ &+2{\tau_{j}}\int_{\mathbb{R}^{2}}\nabla r_{j}\cdot\nabla\varphi_{d}\nabla\theta_{j}S(r_{j})S^{\prime}(r_{j})\,dx+o(\varepsilon)\\\ &=I_{1}+o(\varepsilon).\end{split}$ Recall the relation of $\varphi_{d}$ and $\psi$ in (3.8), one has $\nabla\varphi_{d}=\Big{(}\sum_{j=1}^{\mathcal{K}}{\tau_{j}}\partial_{2}A_{p_{j}},\,\,-\sum_{j=1}^{\mathcal{K}}{\tau_{j}}(\frac{A_{p_{j}}}{x_{1}}+\partial_{1}A_{p_{j}})\Big{)}.$ It has been shown in [20] that $A_{a}(x_{1},x_{2})=\sqrt{\frac{a_{1}}{x_{1}}}\frac{1}{\kappa}\Big{[}(2-\kappa^{2})K(\kappa^{2})-2E(\kappa^{2})\Big{]},$ where $\kappa^{2}(x)=\frac{4a_{1}x_{1}}{x_{1}^{2}+a_{1}^{2}+(x_{2}-a_{2})^{2}+2a_{1}x_{1}}$ and $K,\,E$ are the complete elliptic integrals of first and second kind, i.e., $\begin{split}K(s)&=\int_{0}^{\frac{\pi}{2}}(1-s\sin^{2}\theta)^{-\frac{1}{2}}d\theta,\\\ E(s)&=\int_{0}^{\frac{\pi}{2}}(1-s\sin^{2}\theta)^{\frac{1}{2}}d\theta.\end{split}$ They satisfy $\begin{split}K^{\prime}(s)=K(1-s),\,\,E^{\prime}(s)=E(1-s)\mbox{ for }1<s<1.\end{split}$ Note that $A_{\lambda a}(\lambda x)=A_{a}(x)$, and for $s\to 1$, $\begin{split}K(s)&=-\frac{1}{2}\ln(1-s)(1+\frac{1-s}{4})+\ln 4+O(1-s),\\\ E(s)&=1-\ln(1-s)\frac{1-s}{4}+O(1-s).\end{split}$ Moreover, as we mentioned before, when $r=|z-a|=o(|a_{1}|)$, $A_{a}(z)=\Big{(}\ln\frac{a_{1}}{r}+3\log(2)-2\Big{)}+O\Big{(}\frac{r}{a_{1}}|\ln\frac{r}{a_{1}}|\Big{)}$ and $\partial_{r}A_{a}=-\frac{1}{r}+O(\frac{1}{a_{1}}).$ Combining all these, one has $\begin{split}&-\int_{\mathbb{R}^{2}}|\nabla\varphi_{d}|^{2}\nabla r_{j}S(r_{j})S^{\prime}(r_{j})dx\\\ &+2{\tau_{j}}\int_{\mathbb{R}^{2}}\nabla r_{j}\cdot\nabla\varphi_{d}\nabla\theta_{j}S(r_{j})S^{\prime}(r_{j})\,dx\\\ &=-2{\tau_{j}}\int_{\mathbb{R}^{2}}S(r_{j})S^{\prime}(r_{j})\big{(}\nabla\theta_{j}\cdot\nabla\varphi_{d}\,\nabla r_{j}-\nabla r_{j}\cdot\nabla\varphi_{d}\,\nabla\theta_{j}\big{)}+o(\varepsilon)\\\ &=2{\tau_{j}}\Big{(}\frac{{\tau_{j}}}{p_{j,1}}\int\frac{S(r_{j})S^{\prime}(r_{j})}{r_{j}}A_{p_{j}}(x)dx+\frac{\pi}{p_{j,1}}\sum_{\ell\neq j}{\tau_{\ell}}A_{p_{\ell}}(p_{j}),\,\,0\Big{)}\\\ &+2{\tau_{j}}\pi\sum_{\ell\neq j}{\tau_{\ell}}\nabla A_{p_{\ell}}(p_{j})+o(\varepsilon)\\\ &=\frac{2\pi}{p_{j,1}}\Big{(}\ln p_{j,1}+c_{0}+\sum_{\ell\neq j}{\tau_{j}\tau_{\ell}}A_{p_{\ell}}(p_{j}),\,0\Big{)}+2\pi\sum_{\ell\neq j}{\tau_{j}\tau_{\ell}}\nabla A_{p_{\ell}}(p_{j})+o(\varepsilon)\end{split}$ where (5.2) $c_{0}=3\ln 2-2-\frac{1}{\pi}\int\frac{S(r)S^{\prime}(r)\ln r}{r}dx.$ So one has $I_{1}=\frac{2\pi}{p_{j,1}}\Big{(}\ln p_{j,1}+c_{1}+\sum_{\ell\neq j}{\tau_{j}\tau_{\ell}}A_{p_{\ell}}(p_{j}),\,0\Big{)}+2\pi\sum_{\ell\neq j}\tau_{j}\tau_{\ell}\nabla A_{p_{\ell}}(p_{j})+o(\varepsilon)$ where $c_{1}=c_{0}+\frac{1}{2}\int_{0}^{\infty}S^{\prime}(r)rdr.$ By the above estimates, $\begin{split}&Re\int_{\mathbb{R}^{2}}\frac{E_{2}}{\alpha_{j}}\partial_{x_{1}}\bar{u}_{j}dx\\\ &=-\pi\Big{[}{\tau_{j}}\varepsilon|\ln\varepsilon|-\frac{2\ln p_{j,1}}{p_{j,1}}-\frac{2c_{1}}{p_{j,1}}-2\sum_{\ell\neq j}\tau_{j}\tau_{\ell}\frac{A_{p_{\ell}}(p_{j})}{p_{j,1}}-2\sum_{\ell\neq j}\tau_{j}\tau_{\ell}\partial_{1}A_{p_{\ell}}(p_{j})\Big{]}\\\ &+o(\varepsilon)\end{split}$ and $Re\int_{\mathbb{R}^{2}}\frac{E_{2}}{\alpha_{j}}\partial_{x_{2}}\bar{u}_{j}dx=2\pi\sum_{\ell\neq j}\tau_{j}\tau_{\ell}\partial_{2}A_{p_{\ell}}(p_{j})+o(\varepsilon).$ We now have the following reduced problem: ###### Lemma 5.1. The reduced problem (5.1) is equivalent to the following system of the vortex points $\\{p_{j}\\}$: (5.3) $\begin{split}&{\tau_{j}}\varepsilon|\ln\varepsilon|-\frac{2\ln p_{j,1}}{p_{j,1}}-\frac{2c_{1}}{p_{j,1}}-2\sum_{\ell\neq j}\tau_{j}\tau_{\ell}\frac{A_{p_{\ell}}(p_{j})}{p_{j,1}}-2\sum_{\ell\neq j}\tau_{j}\tau_{\ell}\partial_{1}A_{p_{\ell}}(p_{j})=o(\varepsilon),\end{split}$ and (5.4) $2\sum_{\ell\neq j}\tau_{j}\tau_{\ell}\partial_{2}A_{p_{\ell}}(p_{j})=o(\varepsilon).$ Using the scaling invariance $A_{\lambda a}(\lambda x)=A_{a}(x),$ if we denote by $p_{j}=\frac{\tilde{p}_{j}}{\varepsilon},$ where $|\tilde{p}_{j,1}|=O(1)$, we can get the reduced problem for $\tilde{p}_{j}$: (5.5) $\begin{split}&{\tau_{j}}|\ln\varepsilon|+\frac{2\ln\varepsilon}{\tilde{p}_{j,1}}-\frac{2\ln\tilde{p}_{j,1}}{\tilde{p}_{j,1}}-\frac{2c_{1}}{\tilde{p}_{j,1}}-2\sum_{\ell\neq j}\tau_{j}\tau_{\ell}\frac{A_{\tilde{p}_{\ell}}(\tilde{p}_{j})}{\tilde{p}_{j,1}}\\\ &-2\sum_{\ell\neq j}\tau_{j}\tau_{\ell}\partial_{1}A_{\tilde{p}_{\ell}}(\tilde{p}_{j})=o(1),\end{split}$ and (5.6) $2\sum_{\ell\neq j}\tau_{j}\tau_{\ell}\partial_{2}A_{\tilde{p}_{\ell}}(\tilde{p}_{j})=o(1).$ Using the asymptotic behavior of $A_{a}(x)$ and $A^{\prime}_{a}(r)$, and recall that $|\tilde{p}_{\ell}-\tilde{p}_{j}|\sim\frac{1}{\ln\varepsilon},\,|\tilde{p}_{j,1}|\sim O(1),$ we obtain the following equivalent reduced problem: (5.7) $\left\\{\begin{array}[]{l}{\tau_{j}}|\ln\varepsilon|+\frac{2\ln\varepsilon}{\tilde{p}_{j,1}}+2\sum_{\ell\neq j}\tau_{j}\tau_{\ell}\frac{\tilde{p}_{j,1}-\tilde{p}_{\ell,1}}{|\tilde{p}_{j}-\tilde{p}_{\ell}|^{2}}=o(\ln\varepsilon),\\\ 2\sum_{\ell\neq j}\tau_{j}\tau_{\ell}\frac{\tilde{p}_{j,2}-\tilde{p}_{\ell,2}}{|\tilde{p}_{j}-\tilde{p}_{\ell}|^{2}}=o(1).\end{array}\right.$ ### 5.2. Vortex locations and their generating polynomials In this section, we construct a family of polynomials whose roots will correspond to the locations of the vortex rings. For each rescaled vortex point $\tilde{p}_{j},j=1,...,\mathcal{K},$ we have associated a degree $\tau_{j}=\pm 1.$ To analyze the reduced problem in a more precise way, let us relabel those points with $\tau_{j}=1$ by $\tilde{p}_{1}^{+},...,\tilde{p}_{m}^{+}$ and those with $\tau=-1$ will be denoted by $\tilde{p}_{1}^{-},...,\tilde{p}_{n}^{-}.$ We then write $\displaystyle\tilde{p}_{j}^{+}$ $\displaystyle=\alpha_{0}+\alpha+\frac{1}{\left|\ln\varepsilon\right|}\mathbf{a}_{j},\text{ for }j=1,...,m,$ $\displaystyle\tilde{p}_{j}^{-}$ $\displaystyle=\alpha_{0}+\alpha+\frac{1}{\left|\ln\varepsilon\right|}\mathbf{b}_{j},\text{ for }j=1,...,n.$ Here $\alpha_{0}$ is a fixed constant only depends on $m,n,$ and $\alpha=o\left(1\right)$ depends on $\varepsilon.$ Inserting these into the reduced problem (5.7), we find that, at the main order, $\left(\mathbf{a}_{1},...,\mathbf{a}_{m},\mathbf{b}_{1},...,\mathbf{b}_{n}\right)$ should satisfy the following system: $\left\\{\begin{array}[c]{l}{\displaystyle\sum\limits_{j=1,j\neq k}^{m}}\frac{1}{\mathbf{a}_{k}-\mathbf{a}_{j}}-{\displaystyle\sum\limits_{j=1}^{n}}\frac{1}{\mathbf{a}_{k}-\mathbf{b}_{j}}=\frac{1}{2}-{\alpha_{0}}^{-1},\text{ for }k=1,...,m,\\\ -{\displaystyle\sum\limits_{j=1,j\neq k}^{n}}\frac{1}{\mathbf{b}_{k}-\mathbf{b}_{j}}+{\displaystyle\sum\limits_{j=1}^{m}}\frac{1}{\mathbf{b}_{k}-\mathbf{a}_{j}}=\frac{1}{2}+{\alpha_{0}}^{-1},\text{ for }k=1,...,n.\end{array}\right.$ This can be regarded as a balancing condition between the multiple vortex rings. Adding together the $m+n$ equations in the balancing condition, we find that a necessary condition for the existence of a balancing configuration is $\left({\alpha_{0}}^{-1}-\frac{1}{2}\right)m+\left({\alpha_{0}}^{-1}+\frac{1}{2}\right)n=0.$ It follows that $\alpha_{0}=2\frac{m+n}{m-n}.$ Therefore, we are lead to consider the system (5.8) $\left\\{\begin{array}[c]{l}{\displaystyle\sum\limits_{j=1,j\neq k}^{m}}\frac{1}{\mathbf{a}_{k}-\mathbf{a}_{j}}-{\displaystyle\sum\limits_{j=1}^{n}}\frac{1}{\mathbf{a}_{k}-\mathbf{b}_{j}}=-n,\text{ for }k=1,...,m,\\\ -{\displaystyle\sum\limits_{j=1,j\neq k}^{n}}\frac{1}{\mathbf{b}_{k}-\mathbf{b}_{j}}+{\displaystyle\sum\limits_{j=1}^{m}}\frac{1}{\mathbf{b}_{k}-\mathbf{a}_{j}}=-m,\text{ for }k=1,...,n.\end{array}\right.$ To find solutions to this system, we define the generating polynomial as $P\left(x\right):={\displaystyle\prod\limits_{j=1}^{m}}\left(x-\mathbf{a}_{j}\right),\text{ \ }Q\left(x\right):={\displaystyle\prod\limits_{j=1}^{n}}\left(x-\mathbf{b}_{j}\right).$ If $\mathbf{a}_{j},\mathbf{b}_{j}$ satisfy $\left(\ref{Balance}\right),$ then (5.9) $P^{\prime\prime}Q-2P^{\prime}Q^{\prime}+PQ^{\prime\prime}+nP^{\prime}Q-mPQ^{\prime}=0.$ The case of $m=n$ has been studied in [25] . In this case, the system $\left(\ref{Balance}\right)$ is equivalent to $\left\\{\begin{array}[c]{l}{\displaystyle\sum\limits_{j=1,j\neq k}^{m}}\frac{1}{\mathbf{a}_{k}-\mathbf{a}_{j}}-{\displaystyle\sum\limits_{j=1}^{n}}\frac{1}{\mathbf{a}_{k}-\mathbf{b}_{j}}=-1,\text{ for }k=1,...,m,\\\ {\displaystyle\sum\limits_{j=1,j\neq k}^{n}}\frac{1}{\mathbf{b}_{k}-\mathbf{b}_{j}}-{\displaystyle\sum\limits_{j=1}^{m}}\frac{1}{\mathbf{b}_{k}-\mathbf{a}_{j}}=1,\text{ for }k=1,...,n.\end{array}\right.$ The polynomial solutions of this system are connected with theory of integrable system. Indeed, letting $\phi=\frac{Q}{P}\exp\left(x\right)$ and $u=2\left(\ln P\right)^{\prime\prime}.$ The equation $\left(\ref{PQ}\right)$ can be rewritten as $\phi^{\prime\prime}+u\phi=\phi.$ This equation appears as the first equation in the Lax pair of the KdV equation and has the Darboux invariance property. The polynomial solutions of $\left(\ref{PQ}\right)$ in this case are given by the Adler-Moser polynomials. From the view point of numerical computation, the equation $\left(\ref{PQ}\right)$ is indeed easier than $\left(\ref{Balance}\right).$ Note that our construction of multiple vortex ring solutions requires that all the points $\mathbf{a}_{j},j=1,...,m$ and $\mathbf{b}_{j},j=1,...,n$ are distinct from each other. Therefore we require that the polynomials $P$, $Q$ satisfy the following condition: (H1) $P,Q$ have no repeated roots. Our construction also requires the following condition: (H2) The set of points $\left\\{\mathbf{a}_{1},\cdots,\mathbf{a}_{m},\text{ }\mathbf{b}_{1},\cdots,\mathbf{b}_{n}\right\\}$ are symmetric with respect to the $x_{1}$ axis. Observe that equation $\left(\ref{PQ}\right)$ implies that if $X_{0}$ is a common root of $P$ and $Q,$ then necessarily $X_{0}$ is a repeated root of $P$ or $Q.$ We observe that due to the translation invariance of the equation in the balancing condition, we can normalize the polynomials $P,$ $Q$ as $\displaystyle P\left(x\right)$ $\displaystyle=s_{1}+s_{2}x+...+s_{m-1}x^{m-1}+x^{m},$ $\displaystyle Q\left(x\right)$ $\displaystyle=t_{1}+t_{2}x+...+t_{n-2}x^{n-2}+x^{n}.$ That is, the $x^{n-1}$ term in $Q$ can be chosen to be zero. In this section, we would like to find some solution pair $(P,Q)$ using software such as Maple. Then in the next section, we shall use techniques of integrable system to find a sequence of solution pairs, with explicit Wronskian representation. Let us consider the case of $m+n\leq 12.$ With this constraints, we find, using Maple, that there exist polynomial solutions to $\left(\ref{PQ}\right)$ satisfying (H1) and whose roots satisfy (H2), if further $\left(m,n\right)$ are one of the cases in the set $S:=\\{\left(2,1\right),\left(3,2\right),\left(4,3\right),\left(5,4\right),\left(6,5\right)\\}.$ Indeed, if $\left(m,n\right)=\left(2,1\right),$ then $\left(\ref{PQ}\right)$ has a solution of the form $P\left(x\right)=x^{2}-2x+2,\text{ }Q\left(x\right)=x.$ If $\left(m,n\right)=\left(3,2\right),$ then $\left(\ref{PQ}\right)$ has solution: $P\left(x\right)=x^{3}-2x^{2}+\frac{7}{2}x-\frac{3}{2},\text{ }Q\left(x\right)=x^{2}+1.$ If $\left(m,n\right)=\left(4,3\right),$ then $\left(\ref{PQ}\right)$ has solution: $\displaystyle P\left(x\right)$ $\displaystyle=x^{4}-2x^{3}+\frac{44}{9}x^{2}-\frac{89}{27}x+\frac{533}{324},$ $\displaystyle\text{ }Q\left(x\right)$ $\displaystyle=x^{3}+\frac{13}{6}x+\frac{13}{54}.$ If $\left(m,n\right)=\left(5,4\right),$ then $\left(\ref{PQ}\right)$ has solution: $\displaystyle P\left(x\right)$ $\displaystyle=x^{5}-2x^{4}+\frac{449}{72}x^{3}-\frac{749}{144}x^{2}+\frac{12919}{2592}x-\frac{16015}{15552},$ $\displaystyle\text{ }Q\left(x\right)$ $\displaystyle=x^{4}+\frac{61}{18}x^{2}+\frac{16}{27}x+\frac{1337}{1296}.$ When $\left(m,n\right)=\left(6,5\right),$ we have $\displaystyle P\left(x\right)$ $\displaystyle=x^{6}-2x^{5}+\frac{2269}{300}x^{4}-\frac{193279}{27000}x^{3}+\frac{10810499}{1080000}x^{2}-\frac{57115601}{16200000}x+\frac{3980046413}{2916000000},$ $\displaystyle Q\left(x\right)$ $\displaystyle=x^{5}+\frac{1669}{360}x^{3}+\frac{3607}{3600}x^{2}+\frac{1112099}{324000}x+\frac{23805769}{48600000}.$ The roots of $P,Q$ listed above are solutions of the balancing system. Here we list them in the order $\mathbf{a}_{1},...,\mathbf{a}_{m},\mathbf{b}_{1},...,\mathbf{b}_{n}$ and denote it by $\mathcal{P}_{\left(m,n\right)}.$ The numerical value can be listed as below: $\displaystyle\mathcal{P}_{\left(2,1\right)}$ $\displaystyle:\left(1+i,1-i,0\right),$ $\displaystyle\mathcal{P}_{\left(3,2\right)}$ $\displaystyle:\left(0.56,0.72+1.48i,0.72-1.48i,i,-i\right),$ $\displaystyle\mathcal{P}_{\left(4,3\right)}$ $\displaystyle:(0.393-0.57i,0.393+0.57i,0.607-1.76i,0.607+1.76i,$ $\displaystyle-0.11,0.055-1.48i,0.055+1.48i),$ $\displaystyle\mathcal{P}_{\left(5,4\right)}$ $\displaystyle:(0.255,0.322-0.938i,0.322+0.938i,0.55-1.948i,0.55+1.948i,$ $\displaystyle-0.107-0.567i,-0.107+0.567i,0.107-1.758i,0.107+1.758i),$ $\displaystyle\mathcal{P}_{\left(6,5\right)}$ $\displaystyle:(0.191-0.395i,0.191+0.395i,0.29-1.2i,0.29+1.2i,0.52-2.09i,$ $\displaystyle 0.52+2.09i,-0.145,-0.078-0.94i,-0.078+0.94i,0.15-1.95i,0.15+1.95i).$ Figure 1. $(m,n)=(2,1)$ Figure 2. $(m,n)=(4,3)$ Figure 3. $(m,n)=(6,5)$ Let us denote the pair $(P,Q)$ for $(m,n)=(j,j-1)$ as $(P_{j},Q_{j})$. Then for the above examples, we can see that $P_{j}$ is simply a translation in the $x$ variable of $Q_{j+1}$. We will see in the next section that this is true for all $m=n+1$. Next let us consider the linearized operator around the solution. Let us denote the left hand side of the $j$-th equation of (5.8) by $F_{j}.$ Then we can compute the linearization $dF$ of the map $F:\left(\mathbf{a}_{1},...,\mathbf{a}_{m},\mathbf{b}_{1},...,\mathbf{b}_{n}\right)\rightarrow\left(F_{1},...,F_{m+n}\right).$ $dF$ evaluated at the point $\mathcal{P}_{\left(m,n\right)}$ is a matrix, which can be explicitly computed. The solvability of our original reduced problem is closely related to the nondegeneracy of $dF.$ Since any translation of the $\left(\mathbf{a},\mathbf{b}\right)$ is still a solution to the balancing system, necessarily the determinant of this matrix is zero. That is, $0$ is an eigenvalue of $dF$. Observe that $\left(1,1,....,1\right)$ is an eigenvector. We call $\left(\mathbf{a}_{1},...,\mathbf{a}_{m},\mathbf{b}_{1},...,\mathbf{b}_{n}\right)$ nondegenerated, if the kernel of $dF$ is one dimensional. One can check by explicit computations that for the solutions $\mathcal{P}_{\left(m,n\right)}$ listed above, they are all nondegenerated. It is worth pointing out that if $\left(m,n\right)$ is not in $S,$ there may still have polynomials $P,Q$ satisfying $\left(\ref{PQ}\right),$ but with repeated roots. For instance, when $\left(m,n\right)=\left(4,1\right),$ it has a solution with $P\left(x\right)=x^{4}+4x^{3},\text{ \ }Q\left(x\right)=x.$ When $\left(m,n\right)=\left(5,3\right),$ it has a solution with $P\left(x\right)=x^{5}-\frac{4}{3}x^{4}+\frac{4}{3}x^{3}-\frac{8}{9}x^{2}+\frac{8}{27}x,\text{ }Q\left(x\right)=x^{3}.$ A given pair $\left(m,n\right)$ can be used in the construction of multiple vortex rings, if there exist polynomial solutions to $\left(\ref{PQ}\right)$ satisfying (H1) and (H2). In this respect, there are many questions remain to be answered. For instances, are there infinitely many such pairs? If $\left(m,n\right)$ is such a pair, is it necessarily that $m=n+1?$ Is the balancing configuration unique up to translation? These questions will be partially answered in the next section. Now let us come back to our original reduced problem (5.7) of the GP equation. For each $\left(m,n\right)\in S.$ We have a special solution $\left(\mathbf{a}_{1}^{0},...,\mathbf{a}_{m}^{0},\mathbf{b}_{1}^{0},...,\mathbf{b}_{n}^{0}\right)$ given by $\mathcal{P}_{\left(m,n\right)}.$ If we define vector $\beta$ by $\displaystyle\mathbf{a}_{j}$ $\displaystyle=\mathbf{a}_{j}^{0}+\beta_{j},j=1,...,m,$ $\displaystyle\mathbf{b}_{j}$ $\displaystyle=\mathbf{b}_{j}^{0}+\beta_{j+m},j=1,...,n,$ then the reduced problem (5.7) takes the form (5.10) $dF\left(\beta\right)=G\left(\alpha,\beta\right)+\alpha\alpha_{0}^{-2}\mathbf{e}_{1}\mathbf{,}$ where $G\left(\alpha,\beta\right)=o\left(1\right)$ as $\varepsilon\rightarrow 0,$ with higher order dependence on $\alpha,\beta,$ and $\mathbf{e}_{1}$ is a $m+n$ dimensional column vector whose first $m$ entries are equal to $-1$ and the last $n$ entries are all equal to $1.$ Note that $dF$ is in general not a symmetric matrix. However, since $dF$ is nondegenerated, the kernel of $\left(dF\right)^{T}$ is spanned by $\mathbf{e}_{2}:=\left(1,...,1\right).$ Using the fact that $m-n=1,$ we find that the projection of the right hand side of $\left(\ref{redu}\right)$ onto $\mathbf{e}_{2}$ is equal to $G\cdot\mathbf{e}_{2}-\alpha\alpha_{0}^{-2}.$ Now let us consider the projected problem (5.11) $dF\left(\beta\right)=G\left(\alpha,\beta\right)+\alpha\alpha_{0}^{-2}\mathbf{e}_{1}-\frac{G\cdot\mathbf{e}_{2}-\alpha\alpha_{0}^{-2}}{m+n}\mathbf{e}_{2}.$ Note that for each fixed small $\alpha,$ using the nondegeneracy of the solution $\left(\mathbf{a}_{1}^{0},...,\mathbf{a}_{m}^{0},\mathbf{b}_{1}^{0},...,\mathbf{b}_{n}^{0}\right),$ the projected system $\left(\ref{pro}\right)$ can be solved and a solution $\beta$ depending on $\alpha.$ With this $\beta,$ we then can solve the equation $G\cdot\mathbf{e}_{2}-\alpha\alpha_{0}^{-2}=0$ by a contraction mapping argument. Hence the reduced problem $\left(\ref{redu}\right)$ can be finally solved. Once this is done, with the help of linear theory of Section 4, arguments similar as that of [25] yield a solution to the GP equation, satisfying the conclusion of Theorem 1.2 . ## 6\. Recurrence relations and Wronskian representation of the generating polynomials In this section, we show that the generating polynomials of the balancing system discussed in the previous section have recurrence relations in the case of $m=n+1$, and can be explicitly written down using certain Wronskians. The main result of this section is the following ###### Theorem 6.1. There exists a sequence of polynomials $\mathcal{P}_{n},n=1,...,$ such that (6.1) $\mathcal{P}_{n+1}^{\prime\prime}\mathcal{P}_{n}-2\mathcal{P}_{n+1}^{\prime}\mathcal{P}_{n}^{\prime}+\mathcal{P}_{n+1}^{\prime\prime}\mathcal{P}_{n}+n\mathcal{P}_{n+1}^{\prime}\mathcal{P}_{n}-\left(n+1\right)\mathcal{P}_{n+1}\mathcal{P}_{n}^{\prime}=0,$ where $\mathcal{P}_{n}$ is of degree $n$, $\mathcal{P}_{1}=x$ and $\mathcal{P}_{2}=x^{2}-2x+2.$ Moreover, up to a constant factor(see $\left(\ref{cons}\right)$), these polynomials can be written as $\exp\left(-\frac{n\left(n-1\right)x}{2}\right)W\left(\omega_{1},...,\omega_{n}\right),$ where $W$ represents the Wronskian, $\omega_{j}=\left(x-a_{j}\right)\exp\left(\left(j-1\right)x\right),$ and $a_{1}=0,$ $a_{j+1}=a_{j}+\frac{2}{j}.$ These polynomials can be regarded as a generalization of the Adler-Moser polynomials. There are other types of generalization of the Adler-Moser polynomials, see, for instance [26]. We also refer to [3, 4, 15, 25] and the references cited therein for more discussion in this direction. Recall that in the previous section, we derived the equation (6.2) $P^{\prime\prime}Q-2P^{\prime}Q^{\prime}+PQ^{\prime\prime}+nP^{\prime}Q-\left(n+1\right)PQ^{\prime}=0.$ For $n=1,$ we have found that $P\left(x\right)=x^{2}-2x+2,$ $Q\left(x\right)=x$ is a solution. To solve this equation for general $n,$ we define $\phi=\frac{Q}{P}.$ Direction computation shows that the equation $\left(\ref{e1}\right)$ can be written as (6.3) $\phi^{\prime\prime}+\left(2\left(\ln P\right)^{\prime\prime}-\left(\ln P\right)^{\prime}\right)\phi-\left(n+1\right)\phi^{\prime}=0.$ Note that the equation in this form is different from the one considered by Adler-Moser, in the sense that we have two additional terms corresponding to $\left(\ln P\right)^{\prime}$ and $\phi^{\prime}.$ Moreover, equation (6.2) is not of the standard Hirota bilinear form. This significantly complicates the analysis. For $n=1,$ we already know that equation $\left(\ref{e3}\right)$ has the solution $\bar{\phi}\left(x\right):=\frac{Q}{P}=\frac{x}{x^{2}-2x+2}.$ It is worth pointing out, although not necessarily relevant to our later analysis, $\bar{\phi}$ is smooth in the whole line. Equation $\left(\ref{e3}\right)$ is a second order ODE, it has another solution linearly independent with $\bar{\phi}.$ One can check that $\phi^{\ast}$ defined below is such a solution. Explicitly, $\phi^{\ast}\left(x\right):=\frac{2x^{3}-10x^{2}+21x-16}{x^{2}-2x+2}\exp\left(2x\right).$ Note that $\phi^{\ast}$ can also be written as $\phi^{\ast}=\left(\int_{-\infty}^{x}\frac{\exp\left[\left(n+1\right)s\right]}{\bar{\phi}^{2}}ds\right)\bar{\phi}\left(x\right).$ Next we discuss the generalized Darboux transformation adapted to equation (6.3). The following result can be found in the last section of [28]. ###### Lemma 6.2. Suppose $\phi=\phi_{1}$ and $\phi=\phi_{2}$ are two solutions of the equation $u_{2}\phi^{\prime\prime}+u_{1}\phi^{\prime}+u_{0}\phi=0.$ Then the functions $\tilde{\phi}:=\phi_{2}^{\prime}-\frac{\phi_{1}^{\prime}\phi_{2}}{\phi_{1}}$ satisfies $\tilde{u}_{2}\tilde{\phi}^{\prime\prime}+\tilde{u}_{1}\tilde{\phi}^{\prime}+\tilde{u}_{0}\tilde{\phi}=0,$ where $\tilde{u}_{2}=u_{2},\tilde{u}_{1}=u_{1}+u_{2}^{\prime},\tilde{u}_{0}=u_{0}+u_{1}^{\prime}+2u_{2}\left(\ln\phi_{1}\right)^{\prime\prime}+u_{2}^{\prime}\left(\ln\phi_{1}\right)^{\prime}.$ To apply this lemma, we write equation $\left(\ref{e3}\right)$ as $e^{-x}\phi^{\prime\prime}+e^{-x}\left(2\left(\ln P\right)^{\prime\prime}-\left(\ln P\right)^{\prime}\right)\phi-e^{-x}\left(n+1\right)\phi^{\prime}=0.$ Let us define the new potential $\displaystyle\tilde{u}_{0}$ $\displaystyle:=e^{-x}\left(2\left(\ln P\right)^{\prime\prime}-\left(\ln P\right)^{\prime}\right)+\left(n+1\right)e^{-x}+2e^{-x}\left(\ln\phi^{\ast}\right)^{\prime\prime}-e^{-x}\left(\ln\phi^{\ast}\right)^{\prime}.$ $\displaystyle\tilde{u}_{1}$ $\displaystyle=-e^{-x}\left(n+1\right)-e^{-x},$ and the new function $\Phi_{1}:=\bar{\phi}^{\prime}-\frac{\phi^{\ast\prime}\bar{\phi}}{\phi^{\ast}}=\frac{W\left(\phi^{\ast},\bar{\phi}\right)}{\phi^{\ast}}=\frac{e^{2x}}{\phi^{\ast}}.$ Then using the generalized Darboux transformation described in the previous lemma, we have $e^{-x}\Phi_{1}^{\prime\prime}+\tilde{u}_{0}\Phi_{1}+\tilde{u}_{1}\Phi_{1}^{\prime}=0.$ That is, (6.4) $e^{-x}\Phi_{1}^{\prime\prime}+e^{-x}\left(2\left(\ln P_{3}\right)^{\prime\prime}-\left(\ln P_{3}\right)^{\prime}\right)\Phi_{1}-\left(n+2\right)e^{-x}\Phi_{1}^{\prime}=0,$ where the polynomial $P_{3}$ is defined by $P_{3}:=P\phi^{\ast}e^{-2x}=2x^{3}-10x^{2}+21x-16.$ Equation $\left(\ref{eqp3}\right)$ precisely has the form $\left(\ref{e3}\right).$ An important property is that the equation $\left(\ref{eqp3}\right)$ has another solution $\Phi_{1}^{\ast}:=\frac{36x^{4}-312x^{3}+1136x^{2}-1972x+1357}{2x^{3}-10x^{2}+21x-16}e^{3x}.$ The computations tell us that if $\mathcal{P}_{n}$ is a sequence of polynomials satisfies the conclusion of Theorem 6.1, then we expect the equation $\phi^{\prime\prime}+\left(2\left(\ln\mathcal{P}_{n+1}\right)^{\prime\prime}-\left(\ln\mathcal{P}_{n+1}\right)^{\prime}\right)\phi-\left(n+1\right)\phi^{\prime}=0,$ has two linearly independent solutions, of the form $\phi_{1}=\frac{\mathcal{P}_{n}}{\mathcal{P}_{n+1}},\phi_{2}=\frac{\mathcal{P}_{n+2}}{\mathcal{P}_{n+1}}e^{\left(n+1\right)x}.$ The Wronskian $W\left(\phi_{1},\phi_{2}\right)$ should be equal to $ce^{\left(n+1\right)x}$ for some constant $c.$ Hence we get the following recursive relations between $\mathcal{P}_{n},\mathcal{P}_{n+1},\mathcal{P}_{n+2}:$ $\left(\frac{\mathcal{P}_{n}}{\mathcal{P}_{n+1}}\right)^{\prime}\frac{\mathcal{P}_{n+2}}{\mathcal{P}_{n+1}}e^{\left(n+1\right)x}-\left(\frac{\mathcal{P}_{n}}{\mathcal{P}_{n+1}}\right)\left(\frac{\mathcal{P}_{n+2}}{\mathcal{P}_{n+1}}e^{\left(n+1\right)x}\right)^{\prime}=ce^{\left(n+1\right)x}.$ That is, $\left(\mathcal{P}_{n}^{\prime}\mathcal{P}_{n+1}-\mathcal{P}_{n}\mathcal{P}_{n+1}^{\prime}\right)\mathcal{P}_{n+2}-\mathcal{P}_{n}\left(\mathcal{P}_{n+2}^{\prime}\mathcal{P}_{n+1}-\mathcal{P}_{n+2}\mathcal{P}_{n+1}^{\prime}+\left(n+1\right)\mathcal{P}_{n+2}\mathcal{P}_{n+1}\right)=c\mathcal{P}_{n+1}^{3}.$ This can be written as $\mathcal{P}_{n}^{\prime}\mathcal{P}_{n+2}-\mathcal{P}_{n}P_{n+2}^{\prime}-\left(n+1\right)\mathcal{P}_{n}\mathcal{P}_{n+2}=c\mathcal{P}_{n+1}^{2}.$ If we normalize the polynomials $\mathcal{P}_{n}$ such that the highest order term is $x^{n}.$ Then the constant $c$ satisfies $-\left(n+1\right)=c$ We get the following recurrence relations (6.5) $\mathcal{P}_{n}^{\prime}\mathcal{P}_{n+2}-\mathcal{P}_{n}P_{n+2}^{\prime}-\left(n+1\right)\mathcal{P}_{n}\mathcal{P}_{n+2}+\left(n+1\right)\mathcal{P}_{n+1}^{2}=0.$ When we are given $\mathcal{P}_{n},\mathcal{P}_{n+1},$ the recurrence equation $\left(\ref{re}\right)$ can be integrated, and we expect that the resulted function $\mathcal{P}_{n+2}$ is a polynomial. However, in this step, we will not get a free parameter in this polynomial, because solution of the homogeneous equation has an exponential factor. To show that integrating $\left(\ref{re}\right)$ indeed yields a polynomial, we proceed to find the explicit formula of the sequence $\mathcal{P}_{n}$ which satisfies $\left(\ref{re}\right).$ Let us consider the sequence $a_{j}$ defined through the recurrence $a_{1}=0,$ and $a_{j+1}=a_{j}+\frac{2}{j}.$ Let us define functions $\omega_{j}$ by (6.6) $\omega_{j}=\left(x-a_{j}\right)\exp\left(\left(j-1\right)x\right).$ Then we define functions $\mathcal{P}_{n}$ through the Wronskian (6.7) $\mathcal{P}_{n}:=c_{n}\exp\left(-\frac{n\left(n-1\right)}{2}x\right)W\left(\omega_{1},...,\omega_{n}\right),$ where (6.8) $c_{n}=\left[\left(n-1\right)!{\displaystyle\prod\limits_{1\leq i<j\leq n-1}}\left(j-i\right)\right]^{-1}.$ The normalizing constant $c_{n}$ is used to ensure that the highest order term of $\mathcal{P}_{n}$ is $x^{n}.$ Note that $\mathcal{P}_{n}$ defined by $\left(\ref{Wron}\right)$ are indeed polynomials of degree $n,$ and its leading coefficient is a determinant of Vandermont type. ###### Lemma 6.3. $\mathcal{P}_{n}$ defined by $\left(\ref{Wron}\right)$ satisfies the three- term recurrence relation $\left(\ref{re}\right).$ ###### Proof. For national simplicity, we write $W\left(\omega_{1},...,\omega_{k}\right)$ as $W_{k}.$ Using $\left(\ref{Wron}\right)$ and the fact that $\frac{c_{n+1}^{2}}{c_{n}c_{n+2}}=n+1,$ we see that to prove $\left(\ref{re}\right),$ it suffices to prove (6.9) $W_{n}^{\prime}W_{n+2}-W_{n}W_{n+2}^{\prime}+nW_{n}W_{n+2}+\left(n+1\right)^{2}e^{x}W_{n+1}^{2}=0.$ Following Adler-Moser [1], for any function $\xi,$ we define $W_{k}\left(\xi\right):=W\left(\omega_{1},...,\omega_{k},\xi\right).$ Then we have the Jacobi identity(see [1], Lemma 1) (6.10) $\left(W_{k}\left(\xi\right)\right)^{\prime}W_{k+1}-W_{k}\left(\xi\right)W_{k+1}^{\prime}-W_{k+1}\left(\xi\right)W_{k}=0$ Direct computation tells us that $\omega_{j+1}^{\prime\prime}=j^{2}\omega_{j}\exp\left(x\right).$ Using this relation and its differentiation and the fact that $\omega_{1}=x$, we obtain $W_{k}\left(1\right)=\left(-1\right)^{k}\left(\left(k-1\right)!\right)^{2}\exp\left[\left(k-1\right)x\right]W_{k-1}.$ We then compute $\displaystyle\left(-1\right)^{k}\left(\left(W_{k}\left(1\right)\right)^{\prime}W_{k+1}-W_{k}\left(1\right)W_{k+1}^{\prime}-W_{k+1}\left(1\right)W_{k}\right)$ $\displaystyle=\left(\left[\left(k-1\right)!\right]^{2}\exp\left[\left(k-1\right)x\right]W_{k-1}\right)^{\prime}W_{k+1}$ $\displaystyle-\left(\left(k-1\right)!\right)^{2}\exp\left[\left(k-1\right)x\right]W_{k-1}W_{k+1}^{\prime}+\left(k!\right)^{2}\exp\left[kx\right]W_{k}^{2}.$ Dividing the right hand side by $\left[\left(k-1\right)!\right]^{2}\exp\left(k-1\right)x,$ we get $W_{k-1}^{\prime}W_{k+1}-W_{k-1}W_{k+1}^{\prime}+\left(k-1\right)W_{k-1}W_{k+1}+k^{2}\exp\left(x\right)W_{k}^{2}.$ By the Jacobi identity, this has to be zero. Letting $k=n+1,$ we get $\left(\ref{refi}\right).$ This finishes the proof. ∎ The conclusion of Theorem 6.1 follows immediately from Lemma 6.3 and the Darboux invariance property discussed above. Hence we have abundant candidates of balancing configurations of multiple vortex rings. In principle, the nondegeneracy of these configuration could be proved using similar idea as that of [25]. We leave this to a further study. Finally, let us comment on the reason why we restrict to the case $m=n+1.$ Indeed, our original equation in Section 5 to be solved is (6.11) $P^{\prime\prime}Q-2P^{\prime}Q^{\prime}+PQ^{\prime\prime}+nP^{\prime}Q-mPQ^{\prime}=0$ Let $Q=x,$ $n=1$ and $m\geq 3.$ Then the degree $m$ polynomial $P$ satisfying $\left(\ref{mn}\right)$ necessarily has the factor $x^{3}.$ Hence $P$ and $Q$ has a common root and can’t be used in our construction. We conjecture that when $m-n>1$, there will be no balancing configurations satisfying our requirements stated in Section 5. ## References * [1] M. Adler, J. Moser, On a class of polynomials connected with the Korteweg-de Vries equation, Comm. Math. Phys. 61 (1978), no. 1, 1–30. * [2] J. Bellazzini, D. Ruiz, Finite energy traveling waves for the Gross-Pitaevskii equation in the subsonic regime, arXiv:1911.02820. * [3] H. Aref, _Point vortex dynamics: a classical mathematics playground_ , J. Math. Phys., 48 (2007), no. 6, 065401, 23 pp. * [4] H. Aref, P. K. Newton, M. A. Sremler, T. Tokieda and D. L. Vainchtein, _Vortex crystals_ , Adv. Appl. Mech., 39 (2002), pp. 1–79. * [5] F. Bethuel, H. Brezis and F. Helein, Ginzburg-Landau vortices, Progress in Nonlinear Differential Equations and their Applications, 13\. Birkhauser Boston, Inc. Boston Ma, 1994. * [6] F. Bethuel, P. Gravejat and J. C. Saut, On the KP I transonic limit of two-dimensional Gross-Pitaevskii travelling waves, Dyn. Partial Differ. Equ. 5 (2008), no. 3, 241–280. * [7] F. Bethuel, P. Gravejat and J. Saut, Existence and properties of travelling waves for the Gross-Pitaevskii equation, Contemp. Math., 473 (2008), 55-103. * [8] F. Bethuel, P. Gravejat and J. C. Saut, Travelling waves for the Gross-Pitaevskii equation. II, Comm. Math. Phys., 285 (2009), no. 2, 567-651. * [9] F. Bethuel, G. Orlandi and D. Smets, Vortex rings for the Gross-Pitaevskii equation, J. Eur. Math. Soc. (JEMS), 6 (2004), no. 1, 17-94. * [10] F. Bethuel, J. C. Saut, Travelling waves for the Gross-Pitaevskii equation. I, Ann. Henri Poincaré, 70 (1999), no. 2, 147-238. * [11] D. Chiron, Travelling waves for the Gross-Pitaevskii equation in dimension larger than two, Nonlinear Anal. 58 (2004), no. 1-2, 175-204. * [12] D. Chiron, E. Pacherie, Smooth branch of travelling waves for the Gross-Pitaevskii equation in $\mathbb{R}^{2}$ for small speed, arXiv:1911.03433. * [13] D. Chiron, M. Maris, Rarefaction pulses for the nonlinear Schr?dinger equation in the transonic limit, Comm. Math. Phys. 326 (2014), no. 2, 329šC392. * [14] D. Chiron, C. Scheid, Multiple branches of travelling waves for the Gross-Pitaevskii equation, Nonlinearity 31 (2018), no. 6, 2809–2853. * [15] P. A. Clarkson, _Vortices and polynomials_ , Stud. Appl. Math., 123 (2009), no. 1, pp. 37–62. * [16] M. Del Pino, M. Kowalczyk and M. Musso, Variational reduction for Ginzburg -Landau vortices, Journal of Functional Analysis, 239(2) (2006), 497-541. * [17] P. C. Fife, L. A. Peletier, On the location of defects in stationary solutions of the Ginzburg-Landau equation in $\mathbb{R}^{2}$, Quart. Appl. Math., 54 (1996), no. 1,85-104. * [18] P. Gravejat, A non-existence result for supersonic travellingwaves in the Gross-Pitaevskii equation, Comm. Math. Phys., 243 (2003),93-103. * [19] P. Gravejat, Limit at infinity and nonexistence results for sonic travelling waves in the Gross-Pitaevskii equation, Differ. Int. Eqs., 17 (2004), 1213-1232. * [20] J.D.Jackson, Classical Electrodynamics., Wiley, New York, (1962). * [21] R. Jerrard, D. Smets, Leapfrogging vortex rings for the three dimensional gross-pitaevskii equationl, Annals of PDE, 4 (2016), 1-48. * [22] C. A. Jones, P. H. Roberts, Motion in a Bose condensate IV. Axisymmetric solitary waves, J. Phys. A: Math. Gen., 15 (1982), 2599-2619. * [23] C. A. Jones, S. J. Putterman and P. H. Roberts, Motions in a Bose condensate V. Stability of solitary wave solutions of nonlinear Schrodinger equations in two and three dimensions, J. Phys. A, Math. Gen., 19 (1986), 2991-3011. * [24] F. H. Lin, J. C. Wei, _Traveling wave solutions of the Schrodinger map equation_ , Comm. Pure Appl. Math., 63 (2010), no. 12, pp. 1585–1621. * [25] Y. Liu, J. C. Wei, Multi-vortex traveling waves for the Gross-Pitaevskii equation and the Adler-Moser polynomials, SIAM J. Math. Anal. 52 (2020), no. 4, 3546–3579 . * [26] I. Loutsenko, Integrable dynamics of charges related to the bilinear hypergeometric equation, Comm. Math. Phys. 242 (2003), no. 1-2, 251–275. * [27] M. Maris, Traveling waves for nonlinear Schrodinger equations with nonzero conditions at infinity, Ann. of Math. (2) 178 (2013), no. 1, 107šC182. * [28] V. B. Matveev, Darboux transformation and explicit solutions of the Kadomtcev-Petviaschvily equation, depending on functional parameters, Lett. Math. Phys. 3 (1979), no. 3, 213–216. * [29] F. Pacard, T. Riviere, _Linear and nonlinear aspects of vortices. The Ginzburg-Landau model_ , Progress in Nonlinear Differential Equations and their Applications, 39. Birkhauser Boston, Inc., Boston, MA, 2000. * [30] C. Pethick, H. Smith, Bose-Einstein condensation in dilute gases. Cambridge University Press, Cambridge, 2002. * [31] R. M. Herve and M. Herve, _Etude qualitative des solutions reelles d’une equation differentielle liee ‘a l’equation de Ginzburg–Landau_ , Ann. Inst. H. Poincaré Anal. Non Linéaire, 11 (1994), no. 4,427-440. Weiwei Ao School of Mathematics and Statistics Wuhan University, Wuhan, Hubei, China Email<EMAIL_ADDRESS> Yehui Huang School of Mathematics and Physics, North China Electric Power University, Beijing, China, Email<EMAIL_ADDRESS> Yong Liu Department of Mathematics, University of Science and Technology of China, Hefei, China, Email<EMAIL_ADDRESS> Juncheng Wei Department of Mathematics, University of British Columbia, Vancouver, B.C., Canada, V6T 1Z2 Email<EMAIL_ADDRESS>
# Knowledge Graph Completion with Text-aided Regularization Tong Chen Sirou Zhu Yiming Wen Zhaomin Zheng Language Technology Institute Carnegie Mellon University Pittsburgh, PA 15213 {tongc2, sirouz, yimingwe<EMAIL_ADDRESS> ###### Abstract Knowledge Graph Completion is a task of expanding the knowledge graph/base through estimating possible entities, or proper nouns, that can be connected using a set of predefined relations, or verb/predicates describing interconnections of two things. Generally, we describe this problem as adding new edges to a current network of vertices and edges. Traditional approaches mainly focus on using the existing graphical information that is intrinsic of the graph and train the corresponding embeddings to describe the information; however, we think that the corpus that are related to the entities should also contain information that can positively influence the embeddings to better make predictions. In our project, we try numerous ways of using extracted or raw textual information to help existing KG embedding frameworks reach better prediction results, in the means of adding a similarity function to the regularization part in the loss function. Results have shown that we have made decent improvements over baseline KG embedding methods. ## 1 Introduction The Knowledge Graph/Knowledge Base is a collection of structured tuples of information connecting entities via relations, where entities are generalized concepts or names to be appeared in head/tail positions, forming nodes/vertices in the graph and relations are verbs or predicates describing the way two entities are logically connected, forming edges in the graph. Formally, Knowledge Graph can be represented as a directed multigraph that may have multiple edges with the same vertices. Knowledge Graphs have a lot of applications, such as Structured Search, Question-Answering Systems, Recommendation Systems etc. However, most of the Knowledge Graphs are large but far from complete. In this project, we aim to provide a modern approach to create or predict missing links in the graph, formally done by link prediction: given two elements of a triple, predict missing items. Let’s formally define our task. Let $\mathcal{E}$ be be the set of entities and $\mathcal{R}$ be the set of relations. Then a KG is defined as a collection of triple facts ($\mathit{e_{s},r,e_{d}}$), and the task of knowledge graph reasoning is to find the missing entry in the triple facts. Previously, the focus of knwoledge graph embedding and completion mainly focuses on using various methods to catch the intrinsic, latent semantics of the knowledge graph, which more or less rely on the existing items or structures of the graph itself; but it is another level of concern that the information would be limited to the existing relations or semantics of the knowledge graph, so to complete and expand a knowledge graph, it would be very helpful to introduce external information from text corpora that one knowledge graph or even commonsense graph would rely on, be it general (like FreeBase or YAGO) or industry-specific (such asUMLS). Many recent joint graph and text embedding methods have been focusing on learning better knowledge graph embeddings for reasoning (Han et al. (2018)), but we consider reaching for better graph embeddings in a more language- oriented sense. In this research, we would propose a general framework of using regularization techniques to train a set of entity embeddings that can capture the nuances of relation and connection between the entities that may not lie in the knowledge graph structure itself, but more on the text corpus side; through becoming a constituent part of regularization, we make our framework more compatible to more of the commonly available knowledge graph embedding models, and have the potential to migrate to state-of-the-art models of the time. ## 2 Related Work ### 2.1 Knowledge Graph Reasoning Given a knowledge graph, the task of knowledge graph reasoning includes predicting missing links between entities, predicting missing entities, and predicting whether a graph triple is true or false. A variety of methods have been applied to the task of knowledge graph reasoning, and recently embedding- based methods are gaining popularity and yielding promising results, such as linear models in Bordes et al. (2011), matrix factorization models in Trouillon et al. (2016), convolutional neural networks in Yang et al. (2014). In spite of promising results, these models have limited interpretability. Reinforcement learning models have better interpretability, such MINERVA in Das et al. (2017), DeepPath in Xiong et al. (2017) and Multi-hop in Lin et al. (2018), which exploit policy network. ### 2.2 Open-world Knowledge Graph Completion. Using only information inside the static and structured knowledge graph limits our ability to learn representations for embeddings and relations. Therefore, (Shi & Weninger (2018)) raised the concept of Open-world Knowledge Graph Completion problem, which is to utilize external information so as to connect unseen entities to the knowledge graph. There are several attempts to utilize text corpus to improve Knowledge Graph embeddings, such as the mutual attention model by (Han et al. (2018)) and Latent Relation Language Models in Hayashi et al. (2019). However, these approaches have the following limitations: First of all, the models focuses on local information in the text corpus. Each entity is trained with just its description without considering corpus-level statistics. Secondly, the model need to align two disjoint latent spaces, the Knowledge Graph space and the text corpus space. Thirdly, the models tend to closely couple between the Knowledge Graph and Text method. Last but not least, the models typically provide no way of learning relation- specific representations of entities. ### 2.3 Text Embedding Models and Vector Sets. Word embeddings have been proved to be useful in NLP tasks as standalone features (Turian et al. (2010)). The key idea of word embedding is to use multi-dimentional vectors to represent the meanings of words. Influential models can be divided into two categories, count-based models and prediction- based models. A good example of count-based models is GloVe, introduced by Pennington et al. in 2014, which is a log-linear model trained on window based local co-occurrence information about word pairs in Pennington et al. (2014a). Popular prediction-based models include Continuous bag of words (CBOW) models, skip-gram models and transformer-based models. Based on these state-of-the-art models, pre-trained vector sets such as Word2Vec in Mikolov et al. (2013), FastText in Mikolov et al. (2018), and GloVe, are released and are ready to be used in NLP tasks. In our model, word embedding can be used both statically and dynamically. The static way is to initialize the entity weight by pretrained embedding from results in Pennington et al. (2014a), Mikolov et al. (2013), and train the entity embedding with models in Dettmers et al. (2018) and Bordes et al. (2011). The dynamic way is to combine the loss functions of word embedding model and knowledge graph model together to train the entity and relation embedding. ## 3 Text-regularized KG Completion In this research, we would propose a general framework of using regularization techniques to train a set of entity embeddings that can capture the nuances of relation and connection between the entities that may not lie in the knowledge graph structure itself, but more on the text corpus side. Many recent joint graph and text embedding methods have been focusing on learning better knowledge graph embeddings for reasoning in Han et al. (2018), but we consider reaching for better graph embeddings in a more language-oriented sense. Existing jointly training methods, like Toutanova et al. (2015), Han et al. (2018), and Hayashi et al. (2019), share the following shortcomings: * • These methods typically only focus on a neighboring/localized space that is labeled relevant to some given entity-relation-entity triples; their embeddings are only trained with the related sentence description that are labeled in the corpus. * • They do not take corpus level statistics and language model relativities into account, due to the same reason that they rely on labeled partitioned datasets. * • They need to align two disjoint latent/embedding spaces (the KG embeddings and text embeddings), which are usually not quite overlapping considering the word tokens available from both sides. * • Entities representations are typically generic, not accounting for relation variations, so no way of learning relation-specific embeddings are offered. Here we would propose a method of text-enhanced knowledge graph embedding, which uses similarity functions as regularizers towards the training loss of the knowledge graph. The underlied motivation is that, we would typically find similar descriptions and especially, overlapping words for two entities/concepts that falls into similar domains and same categories of name types; in the meantime, these entities would also share similar characteristics in their connectivity, topology, and types of relations linking to them in the knowledge graph. Thus from the description or context regarding the concept, we would be able to train a similar set of embeddings for these two concepts. For example, both Microsoft and Amazon are concept names that describes a company, and their descriptions would definitely cover the fact that they both have headquarters in Seattle, and they both have a founder (who are Bill Gates and Jeff Bezos, respectively). To learn a corpus-regularized representations for the relations and entities, we would need to build a loss function that properly trains the embedding towards better predictability. we could define a loss function $L_{(}text)=Sim(x,y)=f(x,y|text)$ on the domain $x,y\in\mathbf{R}^{|E|*|E|}$ that captures the similarities between relevant descriptions or context of entities $x$ and $y$, as shown in the figure 1. There are lots of room of decision over which type of similarity function $f$ to be used, such as word overlap on context, TF-IDF based on word pairs, dot-product of existing trained embeddings, and so on. Figure 1: Illustration of a similarity function over the corpus space. Given two entities (e.g., Carnegie Mellon University and Stanford, or Pittsburgh), we calculate the similarities between them using any defined function that is defined across the text corpus. Formally, we would define a new loss function that serves the purpose of minimizing losses from both the side of knowledge base and the one from the text: $\mathit{L}=\sum_{(\mathbf{e_{1},e_{2},r})\in KG}\lambda_{1}\times\mathit{L}_{KG}(\mathbf{e_{1},e_{2},r})+\lambda_{2}\times\mathit{L}_{Text}(\mathbf{e_{1},e_{2}})$ To exemplify one possible similarity function, we could propose a more GLoVePennington et al. (2014a)-like approach, which could make entities that have similar descriptions being closer to each other in the vector embedding space. $\mathit{L}=\sum_{(\mathbf{e_{1},e_{2},r})\in KG}\lambda_{1}\times\mathit{L}_{KG}(\mathbf{e_{1},e_{2},r})+\lambda_{2}\times(\mathbf{e_{1}^{T}e_{2}+b_{e1}+b_{e2}-X_{e_{1},e_{2}}})$ ## 4 Variants of Similarity functions Here we show some types of similarity functions that we have tested with. They include converging to generic word embeddings, similarity calculation using entity-recognized newspaper text, and ones using associated wikipedia article text. ### 4.1 Variant A: Converging to existing embeddings A Quick way to use existing fruits of extracted textual information to begin with is converging to existing word embeddings. Given the premises that we are trying to incorporate textual structures into entity embeddings, we can well assume that word embeddings should also contain information about connections between words. Considering the apparent fact that the latent structures underlying in the knowledge graphs should be drastically different than the ones extracted from text (which could be more focusing on predicting the next words that would follow the current word), we would prefer to converge the differences between the embeddings of the entities, and the ones when they are converted to words. Also considering the fact that the latent structures’ distribution and the number of dimensions should be different between the two types of embeddings, we incorporate a fully connection layer to extract and redistribute the features, and make the differences comparable. The easiest and the most prominent example would be using cosine similarity to converge the two: $\displaystyle\mathbf{Sim}(\mathbf{e_{1}},\mathbf{e_{2}})=\cos(E_{\mathbf{e_{1}}}-E_{\mathbf{e_{2}}},f(w_{\mathbf{e_{1}}}-w_{\mathbf{e_{2}}}))$ $\displaystyle f(w_{\mathbf{e_{1}}}-w_{\mathbf{e_{2}}})=NN(w_{\mathbf{e_{1}}}-w_{\mathbf{e_{2}}})$ Another would be using rank distance defined by (Santus et al. (2018)) that would extract the most prominent ranks of the embeddings and compare the differences between the two: $\displaystyle\mathbf{Sim}(\mathbf{e_{1}},\mathbf{e_{2}})=\sum_{r\in\text{intersect}}\frac{2}{(E_{\mathbf{e_{1}}}-E_{\mathbf{e_{2}}}+f(w_{\mathbf{e_{1}}}-w_{\mathbf{e_{2}}})}$ $\displaystyle f(w_{\mathbf{e_{1}}}-w_{\mathbf{e_{2}}})=NN(w_{\mathbf{e_{1}}}-w_{\mathbf{e_{2}}})$ ### 4.2 Variant B: Using Entity-tagged Text In (Pennington et al. (2014a)), word similarities can be extracted from a corpus by constructing a co-occurrence matrix. Similarly, entity similarities can be calculated in the same manner. Given a corpus, we can construct matrix $\mathbf{X}$ where $\mathbf{X_{i,j}}$ is the number of times entity $\mathbf{j}$ occurs in the context of entity $\mathbf{i}$. We select New York Time corpus as the test dataset for extracting co-occurrence matrix $\mathbf{X}$. Summing over the entities in the knowledge graph, we have our Entity-tagged text regularization term: $\displaystyle\mathit{L}_{Entity-tagged}\sum_{(\mathbf{e_{1},e_{2},r})\in KG}\mathbf{w_{i}^{T}w_{j}-logX_{i,j}}$ ### 4.3 Variant C: Using Associated Wikipedia Text As described above, similar entities should have similar text descriptions on Wikipedia and similar text structures. For example, Wikipedia documents of counties in the US should all talk about its climate, economy, population, etc, and actors should all talk about their career works and life experiences and so on. To exploit such similarity to regularize our entity embedding, we measure two correlations between the Wikipedia documents of entities, which are, first, a modified Relaxed Word Mover Distance (Kusner et al. (2015)) between two entity documents, and second, TF-IDF similarity between two entity documents. #### 4.3.1 Relaxed Word Mover Distance In (Kusner et al. (2015)), the Relaxed Word Mover Distance is calculated as $\text{RWMD}(\mathbf{e_{1},\mathbf{e_{2}}})=\sum_{i\in\mathbf{e_{1}}}c_{i}\min_{j\in\mathbf{e_{2}}}dist(i,j)$ , where $i,j$ are index of words in entity documents, $c_{i}$ is the normalized frequency of word $i$, and $dist(i,j)$ is the distance between the pre-trained word embedding of word $i$ and $j$. In our proposed method, to boost the calculation, instead of using the distance between word embedding, we use the the dot product of two word embedding to measure their similarity, and accordingly, we introduce the concept of RWMD-Gain, which is the maximum gain of transformation from the document of an entity to another and is as follows: $\displaystyle\text{RWMD- Gain}(\mathbf{e_{1},e_{2}})=\sum_{i\in\mathbf{e_{1}}}c_{i}\max_{j\in\mathbf{e_{2}}}sim(i,j)$ We employ the bidirectional gain as the semantic similarity of two entity document, which is $\displaystyle\text{Gain}(e_{1},e_{2})=\text{RWMD- Gain}(e_{1},e_{2})+\text{RWMD-Gain}(e_{2},e_{1})$ To sum over all instances in the knowledge graph, we have our RMWD regularization term: $\displaystyle\mathit{L}_{Text-RWMD}=\sum_{(\mathbf{e_{1},e_{2},r})\in KG}log(\text{Gain}(\mathbf{e_{1}},\mathbf{e_{2}}))|\mathbf{e_{1}}-\mathbf{e_{2}}|^{2}_{2}$ #### 4.3.2 TF-IDF Similarity To compute the TF-IDF similarity between two entity documents, we first convert each document $\mathit{d}$ to a TF-IDF vector $\mathbf{w_{d}}$, in which each element can be calculated by the following formula: $\displaystyle w_{i,d}=tf_{i,d}\times log(\frac{N}{df_{i}})$ where $\mathit{i}$ represents the index of the word; $tf_{i,d}$ represents the occurrence of word $\mathit{i}$ in document $\mathit{d}$; $\mathit{N}$ represents the total number of documents and $df_{i}$ represents the number of documents containing word $\mathit{i}$. The similarity of document $d_{x}$ and document $d_{y}$ is the dot product of such $\mathbf{w_{x}}$, $\mathbf{w_{y}}$. $\displaystyle Sim(x,y)=dot(\mathbf{w_{x}},\mathbf{w_{y}})$ ## 5 Experiments ### 5.1 Data Sources and formulations From (Fu et al. (2019)) we select the FB60K dataset as the base knowledge graph 111https://github.com/thunlp/OpenNRE. In the aforementioned paper, the authors studied the datasets and found that the relation distributions of the two datasets are very imbalanced; but still, It is the KG dataset that contains the most entities in the “FreeBase”-related KG variants, like FB17K, FB15K-237, etc. In the original dataset, all entity names are its Freebase ID; to properly link the partially anonymized graph to the truth raw text, we employed an openly available dictionary to convert thme from ID to its corresponding names. By linking the ID to the names, we can see that the FB60K dataset mainly contains famous or obscure locations, celebrity names, schools, sports unions, etc. On the external sources of the three variants, we would list them here: * • Variant A (using existing embeddings): we employ the GLoVe trained embeddings (Pennington et al. (2014b)) as the embeddings we would want to converge to. Specifically, since word embeddings are normally single words, while entities are mostly proper noun phrases, we generate the final embedding of the entity by calculating the average over all the constituent words that is related to the entity (an entity can have multiple words and multiple names to refer to) according to the open dictionaries. * • Variant B (using entity labeled text): we employ the NYT10 dataset that is supplied along with the FB60K dataset. Entities are labeled using common named entity recognition toolkits, and the source corpus text is excerpted from the New York Times corpus. They are also labeled what relations the sentence would contain, but currently we haven’t employed these. An analysis to the information overlap (i.e., alignment) between the corpus and the KG in Table 1. Higher CT/CE (triple-entity ratio) indicates adding corpus-edges to the KG increases the average degree more significantly, leading to more reduction in sparsity. * • Variant C (using linked Wikipedia articles): we employ the articles from the Wikipedia dump in December 2019. By linking the entity FreeBase ID to its name, and by finding these names’ English Wikipedia article, we managed to extract over 54,000 articles linking to existing entities (around 80%). We use Stanford CoreNLP Tokenizer (Manning et al. (2014)) to tokenize and lower- letter all the words in the text after we remove structured text and tables, which are text that cannot be easily consumed. We choose TransE (Bordes et al. (2011)) as the base KG embedding method as the base of comparison between all the variants. TransE is an old methods of calculating KG embeddings, but it is the easiest to implement and has still been holding decent performances despite several newer and more state-of-the- art models being introduced since. Dataset | #triples(C) | #triple(G) | #entities(C) | #entities(G) | #rel(C) | #rel(G) | S(train) | S(test) | CT/CE | CR/KR ---|---|---|---|---|---|---|---|---|---|--- FB60K-NYT10 | 172,448 | 268,280 | 63,696 | 69,514 | 57 | 1,327 | 570k | 172k | 2.71 | 0.04 Table 1: The dataset information. #triples(C) & #triples(G) denote the number of triples in the corpus and the KG respectively, and so on. S(train) denotes the number of sentences in the training corpus, while S(test) denotes the number of sentences in the testing corpus. CT/CE denotes triple-entity ratio. Lower triple-entity ratio indicates less triples per entity in average can be extracted from the corpus. CR/KR denotes corpus-relation-quantity/KG-relation- quantity ratio. Lower CR/KR indicates less information overlap between the corpus and the KG. ### 5.2 Experimental Result We run all of our proposed variants on FB60K-NYT10 dataset. Our baseline method is TransE (Bordes et al. (2011)). We add different proposed regularization terms to TransE individually to compare their performance. The experimental results are shown in Table 2. Method | Hits@1 | Hits@3 | Hits@10 | MRR | Hits@1 Filtered | Hits@3 Filtered | Hits@10 Filtered | MRR Filtered ---|---|---|---|---|---|---|---|--- TransE | 0.2144 | 0.4414 | 0.5431 | 0.3379 | 0.3392 | 0.6245 | 0.7016 | 0.4902 TransE + Cosine | 0.2303 | 0.4455 | 0.5436 | 0.3487 | 0.3957 | 0.6310 | 0.7035 | 0.5216 TransE + GloVe-NYT | 0.2112 | 0.4366 | 0.5436 | 0.3363 | 0.3378 | 0.6086 | 0.6997 | 0.4841 TransE + GloVe-Wiki | 0.2305 | 0.4474 | 0.5493 | 0.3500 | 0.3552 | 0.6170 | 0.6951 | 0.4957 TransE + GloVe-RWMD | 0.2417 | 0.4000 | 0.5147 | 0.3396 | 0.4292 | 0.5816 | 0.6787 | 0.5194 Table 2: Experimental results running on FB60K-NYT10 dataset. The best of all metrics are highlighted with bold. ### 5.3 Experimental Analysis From table 2, we can see that generally TransE + Cosine and TransE + GloVe- wiki outperform the other methods and baseline. For HR@1, TranE + GloVe-RWMD gives the best result and outperforms baseline significantly. For HR@3, HR@10, and MRR, TransE + GloVe is the best one. GloVe-RWMD considers the semantic distances, which helps finding the best suitable entity, but negatively impacts the score of the potential entities. TransE + Cosine and GloVe-wiki are more simple and intuitive, which consider only the word frequencies and proves to be more useful when considering a list of entities. The amount of injected information directly influences similarity matrix quality. Wiki is comparably larger then New York Times, and larger amount of information gives more accurate entity similarity score, so GloVe-Wiki outperforms Glove-NYT. ## 6 Conclusion From the experiments, our methods showed decent improvements compared to the baseline model. Although TransE is a prudent model, we believe the proposed regularization methods can fit to later state-of-the-art models with modest adaptations. However, due to time limitation, the hyper-parameters were not fine-tuned in the experiments so the results did not show a very significant improvement. Therefore, future work can be done on further improving the implementation of regularization, or on choosing a better set of parameters. Moreover, the design of decoders can be further improved to transform textual latent features to entity or knowledge graph latent features. After all, we believe that more information should always be better than less, but we need to find a good way to utilize it. In this project, we diversely explored text-assisted KG embedding or reasoning methods. However, given the fact that these methods all require extensive frameworks to extract external information, we think there might be unexplored while simpler ways to migrate textual latent features to enrich KG embeddings. However, further study requires us to focus on a more balanced and diverse knowledge base, which we are lacking today due to the demise of the FreeBase graph. ## References * Bordes et al. (2011) Antoine Bordes, Jason Weston, Ronan Collobert, Yoshua Bengio, et al. Learning structured embeddings of knowledge bases. In _AAAI_ , volume 6, pp. 6, 2011. * Das et al. (2017) Rajarshi Das, Shehzaad Dhuliawala, Manzil Zaheer, Luke Vilnis, Ishan Durugkar, Akshay Krishnamurthy, Alex Smola, and Andrew McCallum. Go for a walk and arrive at the answer: Reasoning over paths in knowledge bases using reinforcement learning. _arXiv preprint arXiv:1711.05851_ , 2017. * Dettmers et al. (2018) Tim Dettmers, Pasquale Minervini, Pontus Stenetorp, and Sebastian Riedel. Convolutional 2d knowledge graph embeddings. In _Thirty-Second AAAI Conference on Artificial Intelligence_ , 2018\. * Fu et al. (2019) Cong Fu, Tong Chen, Meng Qu, Woojeong Jin, and Xiang Ren. Collaborative policy learning for open knowledge graph reasoning. In _Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)_ , pp. 2672–2681, 2019. * Han et al. (2018) Xu Han, Zhiyuan Liu, and Maosong Sun. Neural knowledge acquisition via mutual attention between knowledge graph and text. In _Proceedings of AAAI_ , 2018. * Hayashi et al. (2019) Hiroaki Hayashi, Zecong Hu, Chenyan Xiong, and Graham Neubig. Latent relation language models. _arXiv preprint arXiv:1908.07690_ , 2019. * Kusner et al. (2015) Matt Kusner, Yu Sun, Nicholas Kolkin, and Kilian Weinberger. From word embeddings to document distances. In _International conference on machine learning_ , pp. 957–966, 2015. * Lin et al. (2018) Xi Victoria Lin, Richard Socher, and Caiming Xiong. Multi-hop knowledge graph reasoning with reward shaping. _arXiv preprint arXiv:1808.10568_ , 2018. * Manning et al. (2014) Christopher D. Manning, Mihai Surdeanu, John Bauer, Jenny Finkel, Steven J. Bethard, and David McClosky. The Stanford CoreNLP natural language processing toolkit. In _Association for Computational Linguistics (ACL) System Demonstrations_ , pp. 55–60, 2014. URL http://www.aclweb.org/anthology/P/P14/P14-5010. * Mikolov et al. (2013) Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. Distributed representations of words and phrases and their compositionality. In _Advances in neural information processing systems_ , pp. 3111–3119, 2013. * Mikolov et al. (2018) Tomas Mikolov, Edouard Grave, Piotr Bojanowski, Christian Puhrsch, and Armand Joulin. Advances in pre-training distributed word representations. In _Proceedings of the International Conference on Language Resources and Evaluation (LREC 2018)_ , 2018. * Pennington et al. (2014a) Jeffrey Pennington, Richard Socher, and Christopher D Manning. Glove: Global vectors for word representation. In _Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)_ , pp. 1532–1543. Citeseer, 2014a. * Pennington et al. (2014b) Jeffrey Pennington, Richard Socher, and Christopher D Manning. Glove: Global vectors for word representation. In _Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)_ , pp. 1532–1543, 2014b. * Santus et al. (2018) Enrico Santus, Hongmin Wang, Emmanuele Chersoni, and Yue Zhang. A rank-based similarity metric for word embeddings. _arXiv preprint arXiv:1805.01923_ , 2018. * Shi & Weninger (2018) Baoxu Shi and Tim Weninger. Open-world knowledge graph completion. In _Thirty-Second AAAI Conference on Artificial Intelligence_ , 2018\. * Toutanova et al. (2015) Kristina Toutanova, Danqi Chen, Patrick Pantel, Hoifung Poon, Pallavi Choudhury, and Michael Gamon. Representing text for joint embedding of text and knowledge bases. In _Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing_ , pp. 1499–1509, 2015. * Trouillon et al. (2016) Théo Trouillon, Johannes Welbl, Sebastian Riedel, Éric Gaussier, and Guillaume Bouchard. Complex embeddings for simple link prediction. In _International Conference on Machine Learning_ , pp. 2071–2080, 2016. * Turian et al. (2010) Joseph Turian, Lev-Arie Ratinov, and Yoshua Bengio. Word representations: A simple and general method for semi-supervised learning. In _Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics_ , pp. 384–394, Uppsala, Sweden, July 2010. Association for Computational Linguistics. URL https://www.aclweb.org/anthology/P10-1040. * Xiong et al. (2017) Wenhan Xiong, Thien Hoang, and William Yang Wang. Deeppath: A reinforcement learning method for knowledge graph reasoning. _arXiv preprint arXiv:1707.06690_ , 2017. * Yang et al. (2014) Bishan Yang, Wen-tau Yih, Xiaodong He, Jianfeng Gao, and Li Deng. Embedding entities and relations for learning and inference in knowledge bases. _arXiv preprint arXiv:1412.6575_ , 2014.
# Non-unitary triplet superconductivity tuned by field-controlled magnetization –URhGe, UCoGe and UTe2– Kazushige Machida Department of Physics, Ritsumeikan University, Kusatsu 525-8577, Japan ###### Abstract We report on a theoretical study on ferromagnetic superconductors, URhGe, and UCoGe and identify the pairing state as a non-unitary spin-triplet one with time reversal symmetry broken, analogous to superfluid 3He-A phase. A recently found superconductor UTe2 with almost ferromagnet is analyzed by the same manner. Through investigating their peculiar upper critical field $H_{\rm c2}$ shapes, it is shown that the pairing symmetry realized in all three compounds can be tuned by their magnetization curves under applied fields. This leads to the reentrant $H_{\rm c2}$ in URhGe, an S-shaped in UCoGe and an L-shaped $H_{\rm c2}$ in UTe2 observed under the field direction parallel to the magnetic hard $b$-axis in orthorhombic crystals in common. The identification with double chiral form: ${\bf d}(k)=(\hat{b}+i\hat{c})(k_{b}+ik_{c})$ in UTe2 naturally enables us to understand (1) multiple phases with A1, A2, and A0 phases observed under pressure, (2) the enhanced reentrant $H_{\rm c2}$ for the off-axis direction is associated with first order meta-magnetic transition, and (3) Weyl point nodes oriented along the magnetic easy $a$-axis. All three compounds are found to be topologically rich solid-state materials worth further investigation. ## I Introduction The competing orders are at the heart of the strongly correlated systems in general where multiple long-range or short range orderings, such as superconductivity (SC), ferromagnetism (FM), spin and charge density waves are emerging out of the strong interactions in condensed matter systems. This is particularly true in the case of unconventional superconductivity, which is often associated with other orderings mentioned kato1 ; kato2 . A good example is high temperature cuprate superconductors in which various coexisting or mutually repulsive orderings are found kivelson . There has been much attention focused on ferromagnetic superconductors aokireview , such as UGe2 UGe2 , URhGe URhGe , and UCoGe UCoGe in recent years. A new member of such a superconductor, UTe2 with $T_{\rm c}$=1.6K ran ; aoki2 , which is almost ferromagnetic is discovered quite recently and attracts much excitement. Those systems are contrasted with the coexisting materials of magnetism and superconductivity in (RE)Rh4B4 (RE: 4f rare earth elements) and Chevrel compounds (RE)Mo6S8 in that the 4f electrons responsible for magnetism are localized spatially and distinctive from the conduction electrons machidareview . Here the 5f electrons responsible for magnetism are more subtle in that they participate both magnetism and superconductivity. UTe2 has been investigated experimentally knebel ; daniel ; miyake ; ran2 ; metz ; mad ; tokunaga ; sonier ; nakamine ; hayes ; 1 ; 2 ; 3 ; 4 ; rosa ; aokiP ; andreyUTe2 ; kittaka and theoretically xu ; ishizuka ; shick ; nevi ; fidrysiak ; yarzhemsky ; fujimoto ; lebed ; kmiyake . Simultaneously renewed interest on the former three compounds are developing. These heavy Fermion materials belong to a strongly correlated system that is heavily governed by the 5f electrons, which form a coherent narrow band with a large mass enhancement below the Kondo temperature. Because the upper critical field $H_{\rm c2}$ in those compounds exceeds the Pauli paramagnetic limitation, a triplet or odd parity pairing state is expected to be realized aokireview . However, detailed studies of the pairing symmetry remain lacking despite of the fact that previous knowledge of the first three compounds is accumulated for over two decades. Thus now it is a good chance to understand those “old” materials URhGe and UCoGe together with the new UTe2 by seeking some common features. The prominent SC properties observed commonly in these superconductors are as follows: When $H$ is applied parallel to the magnetic hard $b$-axis in orthorhombic crystals, $H_{\rm c2}$ exhibits the reentrant behavior in URhGe, where the SC state that disappeared reappears at higher fields levy , or an S-shape in UCoGe aokiS and an L-shape in UTe2 knebel in the $H$-$T$ plane. Above the superconducting transition temperature $T_{\rm c}$, ferromagnetic transition occurs in URhGe and UCoGe. Thus, the SC state survives under a strong internal field, resulting from an exchange interaction between the conduction and the 5f electrons. However, in UTe2 “static” FM has not been detected although FM fluctuations are probed ran ; miyake ; tokunaga ; sonier above $T_{\rm c}$, i.e. there is a diverging static susceptibility along the magnetic easy $a$-axis ran ; miyake and the nuclear relaxation time $1/T_{2}$ in NMR tokunaga . The gap structure is unconventional, characterized by either a point in UTe2 aoki2 ; ran or line nodes in the others aokireview . There is clear experimental evidence for double transitions: the two successive second order SC phase transitions seen in specific heat experiments as distinctive jumps systematically change under pressure ($P$) in UTe2 daniel . A similar indication for double SC transitions in ambient pressure is observed in UCoGe at $T_{\rm c2}\sim$0.2K manago1 ; manago2 where the nuclear relaxation time $1/T_{1}T$ in NMR experiments exhibits a plateau corresponding to the “half residual density of states (DOS)” value at the intermediate $T$ below $T_{\rm c}=$0.5K. Upon further lowering of $T$, it stars decreasing again at 0.2 K. Recent specific heat $C/T$ data for several high quality samples of UTe2 aoki2 ; ran ; kittaka commonly show the residual DOS amounting to $0.5N(0)$, which is half of the normal DOS $N(0)$, while some exhibit zero residual DOS metz . Thus, this “residual” half DOS issue is currently controversial. We propose a method to resolve this issue, discussed later in this paper. To understand these three spin-polarized superconductors URhGe, UCoGe, and UTe2 in a unified way, we develop a phenomenological theory based on the assumption that the three compounds are coherently described in terms of the triplet pairing symmetry analogous to the superfluid 3He-A-phase leggett . It is instructive to remember that the A1-A2 phase transition is induced by an applied field, which is observed as the clear double specific heat jumps halperin . The originally degenerate transition temperatures for the A phase are split into the A1 and A2 phases under applied fields mermin . Therefore, to address the experimental facts mentioned above, we postulate the A-phase-like triplet pair symmetry, which responds to the spontaneous FM and/or induced moment under perpendicular external fields, to yield the A1-A2 double transitions. This scenario coherently explains the observed reentrant $H_{\rm c2}$ in URhGe, the S-shape in UCoGe, and the L-shape in UTe2 for the field direction along the magnetic hard $b$-axis in a unified way. As mentioned above, the A1-A2 phase transition in 3He A-phase halperin is controlled by the linear Zeeman effect due to the applied field, which acts to split $T_{\rm c}$ mermin . In the spin polarized superconductors, $T_{\rm c}$ is controlled by the spontaneous and/or field-induced magnetic moment, which is linearly coupled to the non-unitary order parameter. We employ the Ginzburg-Landau (GL) theory to describe these characteristic $H_{\rm c2}$ curves. We also identify the pairing symmetry by group theoretic classification machida based on our previous method ozaki1 ; ozaki2 . The pairing symmetry is a non-unitary triplet machida ; ohmi ; machida2 , where the d-vector is a complex function that points perpendicular to the magnetic easy axis in zero-field. The gap function possesses either a point or line node with a possibly chiral $p$-wave orbital form. This is maximally consistent with the SC characteristics obtained so far in UTe2, such as the STM observation mad of chiral edge states, the polar Kerr experiment hayes , which shows time reversal symmetry breaking, and other various thermodynamic measurements. The arrangement of this paper is following. We set up the theoretical framework to explain those experimental facts in the three compounds, URhGe, UCoGe and UTe2 in Section II. The theory is based on the Ginzburg-Landau theory for the order parameter with three components. The quasi-particle spectra in the triplet states are examined to understand thermodynamic behaviors for the materials. In Section III we investigate the generic phase transitions of the present pairing state under fields applied to various field directions relative to the orthorhombic crystalline axes. In order to prepare analyzing the experimental data for URhGe, UCoGe and UTe2 which exhibit a variety of the $H_{\rm c2}$ such as reentrant SC (RSC), S-shaped, and L-shaped one, the magnetization curves for three compounds are studied in detail and evaluated the curves when the experimental data are absent in Section IV. We apply the present theory to the three compounds and explain the peculiar $H_{\rm c2}$ curves observed in Section V, including the multiple phase diagrams in UTe2 under pressure. Section VI devotes to detailed discussions on the gap structure, and pairing symmetries for each material. Summary and conclusion are given in the final section VII. The present paper is a full paper version of the two short papers by the author short1 ; short2 . ## II Theoretical Framework ### II.1 Ginzburg-Landau theory In order to understand a variety of experimental phenomena exhibited by the three compounds in a common theoretical framework, we start with the most generic Ginzburg-Landau (GL) theory for a spin triplet state. This is general enough to allow us to describe the diversity of those systems. Among abundant spin triplet, or odd-parity paring states we assume an A-phase like pairing state described by the complex $\bf d$-vector with three components: $\displaystyle{\bf d}(k)=\phi(k){\vec{\eta}}=\phi(k)({\vec{\eta}}^{\prime}+i{\vec{\eta}}^{\prime\prime}).$ (1) ${\vec{\eta}}^{\prime}$ and ${\vec{\eta}}^{\prime\prime}$ are real three dimensional vectors in the spin space for Cooper pairs, and $\phi(k)$ is the orbital part of the pairing function. This is classified group-theoretically under the overall symmetry: $\displaystyle\rm{SO}(3)_{\rm spin}\times{\rm D}^{\rm orbital}_{\rm 2h}\times{\rm U}(1)$ (2) with the spin, orbital and gauge symmetry respectively machida ; annett . In this study, we adopt the weak spin-orbit coupling scheme ozaki1 ; ozaki2 which covers the strong spin-orbit (SO) case as a limit. The strength of the SO coupling depends on materials and is to be appropriately tuned relative to the experimental situations. It will turn out to be crucial to choose the weak SO coupling case in understanding the $H_{\rm c2}$ phase diagrams with peculiar shapes: This allows the $\bf d$-vector rotation under an applied field whose strength is determined by the SO coupling. Note that in the strong SO coupling scheme the $\bf d$-vector rotation field is infinite because the Cooper pair spin is locked to crystal lattices. There exists U(1)$\times$Z2 symmetry in this pairing, i.e., invariance under ${\bf d}\rightarrow-{\bf d}$ and gauge transformations. We emphasize here that this SO(3) triple spin symmetry of the pairing function is expressed by a complex three component vectorial order parameter $\vec{\eta}=(\eta_{a},\eta_{b},\eta_{c})$ in the most general. It will turn out later to be important also to describe complex multiple phase diagram, consisting of five distinctive phases, but this is a minimal framework which is necessary and sufficient. Under the overall symmetry expressed by Eq. (2) the most general Ginzburg- Landau free energy functional up to the quadratic order is written down as $\displaystyle F^{(2)}=\alpha_{0}(T-T_{\rm c0}){\vec{\eta}}\cdot{\vec{\eta}}^{\star}+b|{\vec{M}}\cdot{\vec{\eta}}|^{2}+cM^{2}{\vec{\eta}}\cdot{\vec{\eta}}^{\star}$ $\displaystyle+i\kappa{\vec{M}}\cdot{\vec{\eta}}\times{\vec{\eta}}^{\star}$ (3) with $b$ and $c$ positive constants. The last invariant with the coefficient $\kappa$ comes from the non-unitarity of the pairing function in the presence of the spontaneous moment and field induced $\vec{M}(H)$, which are to break the SO(3) spin space symmetry in Eq. (2). We take $\kappa>0$ without loss of generality. This term responds to external field directions differently through their magnetization curves. The fourth order term in the GL functional is given by machida ; annett $\displaystyle F^{(4)}={\beta_{1}\over 2}({\vec{\eta}}\cdot{\vec{\eta}}^{\star})^{2}+{\beta_{2}\over 2}|{\vec{\eta}}^{2}|^{2}.$ (4) Because the fourth order terms are written as $\displaystyle F^{(4)}={\beta_{1}\over 2}({\vec{\eta}}^{\prime}\cdot{\vec{\eta}}^{\prime}+{\vec{\eta}}^{\prime\prime}\cdot{\vec{\eta}}^{\prime\prime})^{2}+{\beta_{2}\over 2}[({\vec{\eta}}^{\prime}\cdot{\vec{\eta}}^{\prime}-{\vec{\eta}}^{\prime\prime}\cdot{\vec{\eta}}^{\prime\prime})^{2}$ $\displaystyle+4({\vec{\eta}}^{\prime}\cdot{\vec{\eta}}^{\prime\prime})^{2}]$ (5) for $\beta_{1},\beta_{2}>0$, we can find a minimum when $|{\vec{\eta}}^{\prime}|=|{\vec{\eta}}^{\prime\prime}|$ and ${\vec{\eta}}^{\prime}\perp{\vec{\eta}}^{\prime\prime}$. Notably, the weak coupling estimate machida leads to ${\beta_{1}/\beta_{2}}=-2$. Thus we have to resort to the strong coupling effects in the following arguments in order to stabilize an A1 phase. It is convenient to introduce $\displaystyle\eta_{\pm}={1\over\sqrt{2}}(\eta_{b}\pm i\eta_{c})$ (6) for ${\bf M}=(M_{a},0,0)$ where we denote the $a$-axis as the magnetic easy axis in this and next sections. From Eq. (3) the quadratic term $F^{(2)}$ is rewritten in terms of $\eta_{\pm}$ and $\eta_{a}$ as $\displaystyle F^{(2)}=\alpha_{0}\\{(T-T_{\rm c1})|\eta_{+}|^{2}+(T-T_{\rm c2})|\eta_{-}|^{2}$ $\displaystyle+(T-T_{\rm c3})|\eta_{a}|^{2}\\}$ (7) with $\displaystyle T_{\rm c1,2}=T_{\rm c0}\pm{\kappa\over\alpha_{0}}M_{a},$ $\displaystyle T_{\rm c3}=T_{\rm c0}-{b\over\alpha_{0}}M^{2}_{a}.$ (8) The actual second transition temperature is modified to $\displaystyle T^{\prime}_{\rm c2}=T_{\rm c0}-{\kappa M_{a}\over\alpha_{0}}\cdot{{\beta_{1}-\beta_{2}}\over{2\beta_{2}}}$ (9) because of the fourth order GL terms in Eq. (4). And also $T_{\rm c3}$ starts decreasing in the linear $|M_{a}|$ in stead of $M^{2}_{a}$ mentioned above just near $|M_{a}|\ll 1$. This comes from the renormalization of $T_{c3}$ in the presence of $|\eta_{+}|^{2}\propto(T_{c1}-T)$ and $|\eta_{-}|^{2}\propto(T_{c2}-T)$. Those terms give rise to the $|M_{a}|$-linear suppression of $T_{c3}$ through fourth order terms. Here we note that among the GL fourth order terms, $Re(\eta_{a}^{2}\eta_{+}\eta_{-})$ in Eq. (4) becomes important in interpreting the $H_{\rm c2}$ data later because it is independent of the signs of the GL parameters $\beta_{1}$ and $\beta_{2}$. For $1\leq{\beta_{1}/\beta_{2}}\leq 3$, $\displaystyle T^{\prime}_{c2}>T_{c2}=T_{\rm c0}\pm{\kappa\over\alpha_{0}}M_{a}.$ (10) This could lead to the modification of the otherwise symmetric phase diagram: $\displaystyle T_{c1}-T_{c0}=T_{c0}-T_{c2}.$ (11) The fourth order contribution of Eq. (9) to $T_{c2}$ may become important to quantitatively reproduce the $H$-$T$ phase diagram, such as the asymmetric L-shape $H_{\rm c2}^{b}$ observed in UTe2 knebel . Note that the ratio of the specific heat jumps to $\displaystyle{\Delta C(T_{c1})\over\Delta C(T_{c2})}={T_{c1}\over T_{c2}}\cdot{\beta_{1}\over{\beta_{1}+\beta_{2}}}.$ (12) The jump at $T_{c2}$ can be quite small for $T_{c1}\gg T_{c2}$. The FM moment $M_{a}$ acts to shift the original transition temperature $T_{\rm c0}$ and split it into $T_{c1}$, $T_{c2}$, and $T_{c3}$ as shown in Fig. 1. Here, the A1 and A2 phases correspond to $|\uparrow\uparrow>$ pair and $|\downarrow\downarrow>$ pair, respectively and the A0 phase is $|\uparrow\downarrow>+|\downarrow\uparrow>$ for the spin quantization axis parallel to the magnetization direction $M_{a}$. According to Eq. (8), $T_{c1}$ ($T_{c2}$) increases (decrease) linearly as a function of $M_{a}$ while $T_{c3}$ decreases quadratically as $M^{2}_{a}$ far away from the degeneracy point shown there (the red dot). The three transition lines meet at $M_{a}$=0 where the three components $\eta_{i}$ ($i=+,-,a$) are all degenerate, restoring SO(3) spin space symmetry. Thus away from the degenerate point at $M_{a}$=0, the A0 phase starts at $T_{\rm c3}$ quickly disappears from the phase diagram. Below $T_{\rm c2}$ ($T_{\rm c3}$) the two components $\eta_{+}$ and $\eta_{-}$ coexist, symbolically denoted by A1+A2. Note that because their transition temperatures are different, A1+A2 is not the so-called A-phase which is unitary, but generically non-unitary except at the degenerate point $M_{a}$=0 where the totally symmetric phase is realized with time reversal symmetry preserved. Likewise below $T_{\rm c3}$ all the components coexist; A1+A2+A0 realizes. Figure 1: (color online) Generic phase diagram in the $T$ and $M_{a}$ plane. $T_{\rm c1}$ ($T_{\rm c2}$) for the A1 (A2) phase increases (decreases) linearly in $M_{a}$. The third phase A0 decreases quadratically in $M_{a}$ away from the degenerate point at $M_{a}$=0. Under an applied field with the vector potential $\bf A$, the gradient GL energy is given $\displaystyle F_{grad}=\sum_{\nu=a,b,c}\\{K_{a}|D_{x}\eta_{\nu}|^{2}+K_{b}|D_{y}\eta_{\nu}|^{2}+K_{c}|D_{z}\eta_{\nu}|^{2}\\}$ (13) where $D_{j}=-i\hbar\partial_{j}-2eA_{j}/c$ and the mass terms are characterized by the coefficients $K_{j}$ ($j=a,b,c$) in D2h. We emphasize here as seen from this form of Eq. (13) that $H_{\rm c2}$ for the three components each starting at $T_{\rm cj}$ ($j=1,2,3$) intersects each other, never avoiding or leading to a level repulsion. The level repulsion may occur for the pairing states belonging to multi-dimensional representations (see for example [repulsion1, ; repulsion2, ; repulsion3, ; repulsion4, ] in UPt3). The external field $H$ implicitly comes into $T_{\rm cj}$ ($j=1,2,3$) through $M_{a}(H)$ in addition to the vector potential $A$. This gives rise to the orbital depairing mentioned above. The magnetic coupling $\kappa$, which is a key parameter to characterize materials of interest in the following, is estimated mermin by $\displaystyle\kappa=T_{\rm c}{N^{\prime}(0)\over N(0)}ln(1.14\omega/T_{\rm c})$ (14) where $N^{\prime}(0)$ is the energy derivative of the normal density of states $N(0)$ at the Fermi level and $\omega$ is the energy cut-off. This term arises from the electron-hole asymmetry near the Fermi level. $\kappa$ indicates the degree of this asymmetry, which can be substantial for a narrow band. Thus the Kondo coherent band in heavy Fermion materials, such as in our case, is expected to be important. We can estimate $N^{\prime}(0)/N(0)\sim 1/E_{\rm F}$ with the Fermi energy $E_{\rm F}$. Because $T_{\rm c}$=2mK and $E_{\rm F}$=1K in superfluid 3He, $\kappa\sim 10^{-3}$. In the present compounds $T_{\rm c}\sim$1K and $E_{\rm F}\sim T_{\rm K}$ with the $T_{\rm K}$ Kondo temperature being typically aokireview 10$\sim$50K. Thus $\kappa$ is much larger than that of superfluid 3He and is an order of $1\sim 10^{-1}$. We will assign the $\kappa$ value for each compound to reproduce the phase diagram in the following as tabulated in Table I. ### II.2 Quasi-particle spectrum for general triplet state If we choose ${\vec{\eta}}^{\prime}=\eta_{b}{\hat{b}}$ and ${\vec{\eta}}^{\prime\prime}=\eta_{c}{\hat{c}}$ with $\eta_{a}=0$ for the magnetic easy $a$-axis, the quasi-particle spectra are calculated by $\displaystyle E_{k,\sigma}=\sqrt{\epsilon(k)^{2}+(|{\vec{\eta}}|^{2}\pm|{\vec{\eta}}\times{\vec{\eta}}^{\star}|)\phi(k)^{2}}$ (15) or $\displaystyle E_{k,\sigma}=\sqrt{\epsilon(k)^{2}+\Delta_{\sigma}(k)^{2}},$ (16) where the gap functions for two branches are $\displaystyle\Delta_{\uparrow}(k)$ $\displaystyle=$ $\displaystyle|\eta_{b}+\eta_{c}|\phi(k)$ $\displaystyle\Delta_{\downarrow}(k)$ $\displaystyle=$ $\displaystyle|\eta_{b}-\eta_{c}|\phi(k)$ $\displaystyle\Delta_{0}(k)$ $\displaystyle=$ $\displaystyle|\eta_{a}|\phi(k).$ (17) Note that if $|\eta_{c}|=0$, $\Delta_{\uparrow}(k)=\Delta_{\downarrow}(k)$, which is nothing but the A-phase leggett . When $|\eta_{b}|=|\eta_{c}|$, $\Delta_{\uparrow}(k)\neq 0$ and $\Delta_{\downarrow}(k)=0$, which is the non- unitary A1 phase for $\eta_{a}=0$. The gap in one of the two branches vanishes and the other remains ungapped. Therefore, if we assume that in the normal state $N_{\uparrow}(0)=N_{\downarrow}(0)$, the A1 phase is characterized by having the ungapped DOS $N_{\downarrow}(0)=N(0)/2$ with $N(0)=N_{\uparrow}(0)+N_{\downarrow}(0)$. Generically, however, since $N_{\uparrow}(0)\neq N_{\downarrow}(0)$, that is, $N_{\uparrow}(0)>N_{\downarrow}(0)$ in the A1 phase, which is energetically advantageous than the A2 phase, the “residual DOS” is equal to $N_{\downarrow}(0)$, which is likely less-than-half rather then more-than-half physically. In the non-unitary state with the complex $\bf d$-vector, the time reversal symmetry is broken. In the most general case where all components $\eta_{a}$, $\eta_{b}$, and $\eta_{c}$ are no-vanishing, the quasi-particle spectra are calculated by diagonalizing the $4\times 4$ eigenvalue matrix. Namely in terms of Eq. (17) the spectrum is given by $\displaystyle E^{2}_{k}=\epsilon(k)^{2}+{1\over 2}{\bigl{\\{}}\Delta^{2}_{\uparrow}(k)+\Delta^{2}_{\downarrow}(k)+2\Delta^{2}_{0}(k)$ $\displaystyle\pm\sqrt{(\Delta^{2}_{\uparrow}(k)-\Delta^{2}_{\downarrow}(k))^{2}+4\Delta^{2}_{0}(k)(\Delta^{2}_{\uparrow}(k)+\Delta^{2}_{\downarrow}(k))}{\bigr{\\}}}.$ (18) It is easy to see that this spectrum is reduced to Eq. (16) when $\Delta_{0}(k)=0$. This spectrum characterizes the phase $A_{1}+A_{2}+A_{0}$ realized in UTe2 under pressure as we will see shortly. ## III Prototypes of phase transitions Let us now consider the action of the external field $H_{b}$ applied to the magnetic hard $b$-axis on the FM moment $M_{a}$, pointing parallel to the $a$-axis. The $a$-axis component of the moment $M_{a}(H_{b})$ generally decreases as it rotates toward the $b$-axis as shown in Fig. 2(b). As discussed in the next section in more details based on experimental data, it is observed in URhGe through the neutron experiment levy . Here we display the generic and typical magnetization curves of $M_{a}$ and $M_{b}$ in Fig. 2(c) where $H_{R}$ denotes a characteristic field for $M_{b}(H_{b})=M_{a}(H_{b}=0)$. The induced moment $M_{b}$ reaches the spontaneous FM moment $M_{a}$ at zero field by rotating the FM moment, implying that the FM moment points to the $b$-axis above $H_{R}$. Experimentally, it is realized by the so-called meta-magnetic transition via a first order transition in URhGe levy and UTe2 miyake or gradual change in UCoGe knafoCo . As displayed in Fig. 2(a), by increasing $H_{b}$, $T_{c1}$ ($T_{c2}$) decreases (increases) according to Eq. (8). The two transition lines $T_{c1}(H_{b})$ =$T_{c2}(H_{b})$ meet at $H_{b}=H_{R}$. As $H_{b}$ is further increased, $T_{c1}$ also increases by rotating the $\bf d$-vector direction such that the $\bf d$-vector becomes perpendicular to ${\bf M}_{b}$, which maximally gains the magnetic coupling energy $i\kappa{\vec{M}}\cdot{\vec{\eta}}\times{\vec{\eta}}^{\star}$ in Eq. (3). This process occurs gradually or suddenly, depending on the situations of the magnetic subsystem and the spin-orbit coupling that locks the $\bf d$-vector to the underlying lattices. Therefore $H_{R}$ may indicate simultaneously the $\bf d$-vector rotation. It should be noted, however, that if the spin-orbit coupling is strong, the $\bf d$-vector rotation is prevented. In this case $H^{b}_{\rm c2}$ exhibits a Pauli limited behavior as observed in UTe2 under pressure aokiP . Figure 2: (color online)(a) Prototype phase diagram in the $T$ and $H_{b}$ plane where $H_{b}$ is parallel to the magnetic hard $b$-axis and the moment $M_{a}$ points to the easy $a$-axis. The two transition lines of $T_{c1}$ and $T_{c2}$ (red curves) initially decreases and increases respectively as $H_{b}$ increases toward the degeneracy point at $T_{R}$. There the projection of the FM moment $M_{a}$ vanishes. During this process, by rotating the $\bf d$-vector to catch the magnetization $M_{b}(H_{b})$ (the green lines) instead of $M_{a}(H_{b})$, $H^{(1)}_{\rm c2}$ and $H^{(2)}_{\rm c2}$ turn around their directions. (b) Under the perpendicular field $H_{b}$ the spontaneous moment $M_{a}$ rotates toward the $b$-direction. The projection $M_{b}(H_{b})$ of $M_{a}$ on the $b$-axis increases. (c) The rotation field $H_{R}$ is indicated as the red dot where $M_{b}(H_{b})=M_{a}(H=0)$. In Fig. 3 we show prototype phase diagrams for different situations. In addition to that displayed in Fig. 3(a), which is the same as in Fig. 2(a), there is the case in which $T_{c1}$ is bent before reaching $H_{R}$ as shown in Fig. 3(b). The magnetization curve $M_{b}(H_{b})$ starting at $T_{c0}$ exceeds the decreasing $M_{a}$ at a lower field $H_{\rm CR}$ defined by $M_{b}(H_{b})=M_{a}(H_{b})$. $H^{(1)}_{\rm c2}$ turns around there by rotating the $\bf d$-vector. We will see this case in the following analysis. In the $H_{c}$ case for the field direction parallel to the another hard $c$-axis the phase diagram is shown in Fig. 3(c). Since $H_{c}$ does not much influence on $M_{a}(H_{c})$, both $H^{(1)}_{\rm c2}$ and $H^{(2)}_{\rm c2}$ are suppressed by the orbital depression of $H_{c}$. When magnetic field is applied to the magnetic easy $a$-axis, the spontaneous moment $M_{a}(H_{a})$ increases monotonically, as shown in Fig. 3(d). According to Eq. (8), $T_{c1}$ ($T_{c2}$) increases (decreases) as $H_{a}$ increases. Thus, theoretically $H_{c2}^{(1)}$ can have a positive slope at $T_{c1}$. However, the existing data on UCoGe wu indicate that it is negative as seen shortly. This is because the strong orbital depairing $H^{\prime 0}_{\rm c2}$ overcomes the positive rise of $T_{c1}$. Moreover, $H_{c2}^{(2)}$ is strongly suppressed by both $T_{c2}$ and the orbital effect $H^{\prime 0}_{\rm c2}$, resulting in a low $H^{a}_{c2}$, compared with $H^{b}_{c2}$. This $H_{c2}$ anisotropy is common in these compounds aokireview . From the above considerations, the enhanced $H^{b}_{c2}$ is observed because the higher part of the field in $H_{c2}$ belongs to $H_{c2}^{(2)}$, which has a positive slope. Figure 3: (color online) Two types (a) and (b) of the phase diagram for $H\parallel b$ with the $b$-axis (hard axis). (a) is the same as in Fig. 2(a). (b) At $H_{\rm CR}$ defined by $M_{b}(H_{b})=M_{a}(H_{b})$, $H^{(1)}_{\rm c2}$ turns round by rotating the $\bf d$-vector to catch $M_{b}$ starting from $T_{c0}$. (c) $H\parallel c$ with the $c$-axis (another hard axis). (d) $H\parallel a$ with the $a$-axis (easy axis). The green lines are the respective magnetization curves and the red curves are $H^{(1)}_{\rm c2}$ and $H^{(2)}_{\rm c2}$. Within the GL scheme it is easy to estimate $H_{\rm c2}$ as follows. We start with the $H_{\rm c2}$ expression: $H_{\rm c2}(T)=A_{0}\\{T_{c}(H_{\rm c2})-T\\}$ (19) with $A_{0}={\Phi_{0}\over 2\pi\hbar^{2}}4m\alpha_{0}$, $m$ effective mass, and $\Phi_{0}$ quantum unit flux. Here $T_{c}$ depends on $H$ though $M_{a}(H)$ as described above. Thus, the initial slope of $H^{\prime}_{\rm c2}$ at $T_{c}$ is simply given by $H^{\prime}_{\rm c2}(T)=A_{0}{dT_{c}\over dH_{\rm c2}}H^{\prime}_{\rm c2}-A_{0}.$ (20) It is seen that if ${dT_{c}/dH}=0$ for the ordinary superconductors, $H^{\prime 0}_{\rm c2}(T)=-A_{0}<0$. The slope $H^{\prime}_{\rm c2}(T)$ is always negative. However, Eq. (20) is expressed as $H^{\prime}_{\rm c2}(T)={-A_{0}\over{1-A_{0}({dT_{c}\over dH_{\rm c2}}})},$ (21) or $\displaystyle{1\over|H^{\prime}_{\rm c2}|}$ $\displaystyle=$ $\displaystyle{1\over|H^{\prime 0}_{\rm c2}|}+|{dT_{\rm c}\over dH_{\rm c2}}|$ (22) $\displaystyle=$ $\displaystyle{1\over|H^{\prime 0}_{\rm c2}|}+{1\over|{dH_{\rm c2}\over dT_{\rm c}(H)}|}.$ The condition for attaining the positive slope, $H^{\prime}_{\rm c2}(T)>0$ implies $|H^{\prime 0}_{\rm c2}|>({dH\over dT_{\rm c}})$ at $H_{\rm c2}$. This is a necessary condition to achieve S-shaped or L-shaped $H_{\rm c2}$ curves experimentally observed. This is fulfilled when $|H^{\prime 0}_{\rm c2}|$ is large enough, that is, the orbital depairing is small, $|{dT_{c}\over{dH}}|$ at $H_{\rm c2}$ is large, or the $T_{c}$ rise is strong enough. It is noted that when $1-A_{0}({dT_{c}\over dH_{\rm c2}})=0$, the $H_{\rm c2}(T)$ curve has a divergent part in its slope, which is observed in UCoGe as a part of the S-shape. It is clear from the above that when $dT_{c}/dH<0$, $|H^{\prime}_{\rm c2}(T)|<|H^{\prime 0}_{\rm c2}|$ because the two terms in Eq. (22) are added up to further depress $H_{\rm c2}(T)$. In this case the slope $|H^{\prime}_{\rm c2}|$ is always smaller than the original $|H^{\prime 0}_{\rm c2}|$ as expected. In Fig. 4 we show the changes of $H_{\rm c2}$ when the competition between the orbital suppression and $T_{c}(M)$ varies. We start from the orbital limited $H^{\rm WHH}_{\rm c2}$ curve with $T_{c}$ unchanged as a standard one. When $T_{c}(M)$ decreases with increasing $H$, the resulting $H_{\rm c2}$ is further suppressed compared with $H^{\rm WHH}_{\rm c2}$ as shown in Fig. 4(a). $T_{c}(M)$ as a function of $H$ through $M(H)$ becomes increasing as shown in Fig. 4(b), $H_{\rm c2}$ is enhanced compared to $H^{\rm WHH}_{\rm c2}$, exceeding the $H^{\rm WHH}_{\rm c2}$ value. Figure. 4(c) displays the case where $T_{c}(M)$ increases stronger than that in Fig. 4(b), $H_{\rm c2}$ has a positive slope and keeps increasing until it hits the upper limit $H^{\rm AUL}_{\rm c2}$. There exists the absolute upper limit (AUL) for $H_{\rm c2}$. Even though $T_{c}(M)$ keeps increasing with increasing $M(H)$, $H_{\rm c2}$ terminates at a certain field because a material has its own coherent length $\xi$ which absolutely limits $H^{\rm AUL}_{\rm c2}=\Phi_{0}/2\pi\xi^{2}$. Beyond $H^{\rm AUL}_{\rm c2}$ there exists no superconducting state. Figure 4: (color online) $H_{\rm c2}$ changes due to the competition between the orbital depairing and $T_{c}(M)$. (a) $T_{c}(M)$ decreases as a function of the applied field $H$. The orbital depairing is added up to further depress $H_{\rm c2}$ than $H^{\rm WHH}_{\rm c2}$. (b) $T_{c}(M)$ increases as a function of the applied field $H$, competing with the orbital depairing. The resulting $H_{\rm c2}$ is enhanced compared with $H^{\rm WHH}_{\rm c2}$. (c) $T_{c}(M)$ increases strongly as a function of the applied field $H$. $H_{\rm c2}$ has a positive initial slope and keeps growing until hitting the absolute upper limit $H^{\rm AUL}_{\rm c2}$. Then $H_{\rm c2}$ follows this boundary. There could be several types of $H^{b}_{\rm c2}$ curves for $H$ applied to the $b$-axis (hard axis), depending on several factors: (A) the magnitude of the spontaneous moment $M_{a}$, (B) its growth rate against $H_{b}$, (C) the coupling constant $\kappa$, (D) the relative position of $T_{c0}$ on the temperature axis. Possible representative $H^{b}_{\rm c2}$ curves are displayed in Figs. 5(a), 5(b) and 5(c). When the hypothetical $T_{c0}$ is situated in the negative temperature side, the realized phase is only the A1-phase at $H_{b}=0$. In high field regions SC reappears as the reentrant SC (RSC) by increasing $M_{b}(H_{b})$, which is shown in Fig. 5(a). The reentrant SC is separated from the lower SC. As shown in Fig. 5(b) the two transition temperatures, $T_{c1}$ and $T_{c2}$ are realized at $H_{b}=0$, that is, it shows double transitions at zero field, giving rise to the $A_{1}$ and $A_{2}$ phases. The three phases $A_{1}$, $A_{2}$ and $A$ appear in a finite $H_{b}$ region. $H_{\rm c2}$ could have an S-shape. This corresponds to either Figs. 3(a) or (b). When the separation between $T_{c1}$ and $T_{c2}$ becomes wider because of increasing the spontaneous moment $M_{a}$ and/or the larger magnetic coupling $\kappa$, $H_{\rm c2}$ has an L-shape as displayed in Fig. 5(c). This could happen also when the moment rotation field $T_{R}$ is situated at relatively lower field than the overall $H_{\rm c2}$. In the following we discuss those typical $H_{\rm c2}$ behaviors based on the realistic magnetization curves for each compound, reproduce the observed $H_{\rm c2}$ curves and predict the existence of the multiple phase diagram. Figure 5: (color online) Schematic typical phase diagrams for $H$ parallel to the $b$-axis with $A_{1}$, $A_{2}$ and $A$ phases, whose structure depends on the position of $T_{c0}$ and the separation of $T_{c1}$ and $T_{c2}$. The absolute upper limit $H_{\rm c2}^{\rm AUL}$ is indicated as the grey region. (a) The reentrant SC situated at high fields such as in URhGe. (b) S-shape $H_{\rm c2}$ with the double transitions from the $A_{1}$ to the $A$ phase such as in UCoGe. (c) L-shape $H_{\rm c2}$ where the high field phase is the $A_{2}$ phase such as in UTe2. ## IV Magnetization curves In order to understand their peculiar $H_{\rm c2}$ shapes and resulting pairing symmetry in three compounds, it is essential to know their magnetic responses to applied magnetic fields as mentioned above. Here we analyze their magnetism and estimate the magnetization curves of the spontaneous moment under the transverse field, which is not probed by conventional magnetization measurements. In the following we consider the cases of URhGe and UCoGe, and UTe2 with the $c$-axis and $a$-axis are the easy axes respectively as tabulated in Table I. We mainly discuss URhGe as an typical example. The concepts introduced here are applied to the other systems with appropriately changing the notation for the magnetic easy axis. Table 1: Magnetic properties and $\kappa$ values materials | Curie temp.[K] | easy axis | moment[$\mu_{B}$] | $\kappa$[K/$\mu_{B}$] ---|---|---|---|--- URhGe | 9.5 | $c$-axis | Mc=0.4 | 2.0 UCoGe | 2.5 | $c$-axis | Mc=0.06 | 1.8 UTe2 | – | $a$-axis | $\sqrt{\langle{\rm M}_{a}^{2}\rangle}$=0.48 | 6.9 ### IV.1 Rigid rotation picture: Spontaneous moment rotation When the applied field $H_{b}$ is directed to the hard axis, or the $b$-axis, the spontaneous moment $M_{c}(H_{b})$ pointing to the $c$-axis in URhGe rotates gradually toward the applied field direction. At around $H_{R}=12$T, $M_{c}(H_{b})$ quickly turns to the $b$-direction by rotating the moment as shown in Fig. 6. We define the crossing field $H_{\rm\rm CR}$ at which $M_{c}(H_{b})=M_{b}(H_{b})$. Note that $H_{R}$ and $H_{\rm\rm CR}$ are different concepts as is clear from Fig. 6 and also in UCoGe where $H_{\rm\rm CR}\sim$ a few T and $H_{R}$=45T knafoCo . Simultaneously and correspondingly, the $M_{b}(H_{b})$ moment jumps via a first order transition. Above $H_{b}>H_{R}$ the spontaneous moment is completely aligned along the $b$-axis as seen from Fig. 6. This phenomenon is often called as the meta-magnetic transition. But this is just the moment rotation since it is demonstrated that the total magnetization $\sqrt{M^{2}_{c}(H_{b})+M_{b}^{2}(H_{b})}$ hardly changes and remains a constant during this first order transition process levy . This implies that $M_{c}(H_{b})=M_{c}\cos(\alpha(H_{b}))$, and $M_{b}(H_{b})=M_{c}\sin(\alpha(H_{b}))$ with $\alpha(H_{b})$ being the rotation angle of $M_{c}(H_{b})$ from the $c$-axis. The rotation angle $\alpha(H_{b})$ is accurately measured by the neutron scattering experiment by Lévy, et allevy who construct the detailed map of the rotation angle in the $H_{b}$ and $H_{c}$ plane. This rotation process is mirrored by the magnetization curve of $M_{b}(H_{b})$ so that the projection of $M_{c}(H_{b})$ onto the $b$-axis manifests itself on $M_{b}(H_{b})$ as shown in Fig. 6. The crossing of $M_{b}(H_{b})$ and $M_{c}(H_{b})$ occurs around at $H_{\rm CR}=9\sim 10$T, corresponding to roughly $M_{c}(H_{b})/\sqrt{2}\sim M_{b}(H_{b})$. That is, $M_{c}(H_{b})$ rotates by the angle $\alpha=45^{\circ}$ from the $c$-axis at $H_{\rm\rm CR}$. This first order phase transition phenomenon in URhGe under the transverse field is neatly described by Mineev mineev using the GL theory. This is within more general framework of the so-called meta-magnetic transition theory based on the GL phenomenology wohlfarth ; shimizu ; yamada for itinerant ferromagnets. Figure 6: (color online) The ferromagnetic spontaneous moment $M_{c}(H_{b})$ rotation indicated by the green arrow under the field $H\parallel b$ in URhGe. At $H_{b}=H_{R}$, it completely orients along the $b$-axis direction via a first order transition where $M_{b}(H_{b})$ shows a jump of the magnetization. $H_{CR}$ is defined by the field $M_{c}(H_{b})=M_{b}(H_{b})$. The rotation angle $\alpha$ from the $c$-axis is measured by neutron experiment [levy, ]. The magnetization curves $M_{b}(H_{b})$ and $M_{c}(H_{c})$ are from [hardy, ]. Those considerations based on the experimental facts demonstrate to hold “a rigid moment rotation picture”. We assume this picture applicable to the other compounds too. ### IV.2 Extraction of the $M_{b}$ moment for the tilted fields from the $b$-axis data When the applied field direction is rotated from the $b$-axis toward the easy axis $c$ by the angle $\theta$, the magnetization curves are measured by Nakamura, et al nakamura . It is obvious that the measured magnetization $M(\theta)$ contains the contribution from the spontaneous moment ${\bf M}_{c}$ projected onto the applied field direction, that is, $M_{c}\sin(\theta)$. This is confirmed experimentally at least lower fields up to $H<5$T and $T$=2K aokiprivate . Thus in this situation, we can extract the $M_{b}(H)$ curves by simply subtracting the contribution $M_{c}\sin(\theta)$ from the measured data nakamura . The result is shown in Fig. 7(a). It is seen that by increasing the angle $\theta$, the first order transition field $H_{R}$ shifts to higher fields and the jump gets smaller compared to the $b$-axis case, reflecting that the moment projection onto the applied field direction decreases. This method is valid only for the small angle $\theta$ and relatively small field regions because here the $M_{c}$ moment is assumed to be fixed under the action of small field component along the $c$-axis. It may be difficult to extract reliably the $M_{b}(H)$ information for further high fields even though the tilting angle is small, and also for larger angles $\theta$. There are two factors to be taken into account, which are internally related: One is that the $c$-component magnetic field acts to prevent the moment from further rotating it toward the $b$-axis upon increasing tilting field $H$ by $\theta$ from the $b$-axis. This “rotation angle locking effect” becomes important for the field just before $H_{R}(\theta)$ where the moment ultimately rotates completely along the $b$-axis in the higher fields. The other factor to be considered is the modification of the free energy landscape of the $M_{b}$ versus $M_{c}$ space. As mentioned, the first order transition of the moment rotation is described by Mineev mineev who considers the competition between the ferromagnetic state at $M_{c}$ and the paramagnetic state with $M_{b}$ stabilized by the Zeeman effect due to the external field $H_{b}$ within a GL free energy theory. The transverse field $H_{b}$ necessarily destabilizes the second order FM phase transition at $H_{R}$ because $H_{b}$ contributes negatively to the quartic term coefficient of $M_{c}^{4}$, giving rise to a first order transition. The extra term coming from the tilting field helps to stabilize the ferromagnetic state, preventing the first order transition, thus making $H_{R}$ to higher field and the magnetization jump smaller. Thus it is not easy to extract reliably the $M_{b}(H)$ under this free energy landscape modification. In the followings, we confine our arguments for small $\theta$ and use approximate $M_{b}(H)$ forms, which are enough for our purposes to understand the peculiar $H_{c2}$. Figure 7: (color online) (a) The magnetization component of $M_{b}(H)$ in URhGe under the field direction tilted from the $b$-axis toward the $c$-axis by $\theta$, estimated from the experimental data of $M(H)$ [nakamura, ]. (b) The magnetization $M(H)$ in URhGe under the field direction tilted from the $b$-axis toward the $a$-axis by $\phi$ estimated from the experimental data (dots) of $M(H)$ [hardy, ], including magnetization curves for three $a$, $b$ and $c$-directions for reference. Figure 8: (color online) (a) The magnetization curves for three $a$, $b$, and $c$-axes in UCoGe. Here the crossing points $H^{b}_{\rm\rm CR}$ and $H^{a}_{\rm\rm CR}$ at which each curve surpasses the spontaneous moment $M_{c}(H=0)=0.06\mu_{B}$. (b) The magnetization curves of $M_{b}(H)$ for the field directions tilted from the $b$-axis toward the $c$-axis by the angle $\theta$(degrees) in UTe2. $\theta=23.7^{\circ}$ corresponds to $H\parallel(011)$ direction measured by [miyakeprivate, ]. Those are estimated by the method explained in the main text. The inset shows the magnetization curves for three $a$, $b$, and $c$-axes in UTe2. $H_{R}$ is the first order transition for the moment rotation from the $a$-axis to the $b$-axis. ### IV.3 Applied field rotation from the $b$-axis to the hard axis In the case for the tilting angle $\phi$ from the $b$-axis toward the other hard axis $a$ of URhGe, it is known levy2 that $H_{R}(\phi)$ is scaled to $H_{R}(\phi)\propto 1/\cos(\phi)$, which is also the case in UTe2 ran2 . This means that only the $M_{b}$ projection onto the $a$-axis matters to understand the magnetization process. Therefore, we can easily reconstruct the $M(\phi)$ by using the experimental data of $M_{b}(H_{b})$ except for the fact that the induced $M_{a}(\phi)$ also contributes to $M(\phi)$. This can be accomplished by an “elliptic formula” derived as follows: We start with $M_{b}(H_{b})$ and $M_{a}(H_{a})$ measured by usual magnetization experiments shown in Fig. 7(b). Assuming the linearity assumption: $M_{b}(\phi)=\chi_{b}H\cos(\phi)$ and $M_{a}(\phi)=\chi_{a}H\sin(\phi)$ with $\chi_{i}$ $(i=a,b)$ being the magnetic susceptibility, we add up the two components, $\displaystyle M(\phi)$ $\displaystyle=$ $\displaystyle M_{b}\cos\phi+M_{a}\sin\phi$ (23) $\displaystyle=$ $\displaystyle(\chi_{b}\cos^{2}(\phi)+\chi_{a}\sin^{2}(\phi))H$ $\displaystyle=$ $\displaystyle M_{b}(H_{b})\cos^{2}\phi+M_{a}(H_{a})\sin^{2}\phi.$ We call it an “elliptic formula”. Since the rotation field is given by $H_{R}(\phi)={H_{R}^{b}\over\cos(\phi)}$ (24) with $H_{R}^{b}$ the rotation field for the $b$-axis, we obtain at $H=H_{R}$ $M(\phi)=M_{b}(H_{R}){\bigl{(}}\cos\phi+{\chi_{a}\over\chi_{b}}\cdot{\sin^{2}\phi\over\cos^{2}\phi}\bigr{)}.$ (25) This formula gives the magnetization curve consisting of a straight line from $H=0$ up to $H_{R}$. The magnetization jump at $H_{R}$ is calculated by projecting the jump $\delta M_{b}$ in $M_{b}(H_{b})$, namely $\delta M_{b}\cos(\phi)$. The resulting reconstructions of $M(\phi)$ for various tilting angles are shown in Fig. 7(b). By construction, when $\phi\rightarrow 90^{\circ}$, $M(\phi)\rightarrow M_{a}(H_{a})$. We notice that the resulting $M(\phi)$ includes the contribution from $M_{a}$. Those results should be checked experimentally and will be used to reproduce the RSC in URhGe. As shown in Fig. 8(b) this idea is also applied to UTe2 where the RSC appears centered around $\theta=$35∘ from the $b$-axis toward the another hard axis $c$. As a final comment on the magnetization of UCoGe shown in Fig. 8(a), it should be mentioned that since $H_{R}\sim 45$T knafoCo , for the following discussions on this system the characteristic magnetic fields $H^{b}_{\rm\rm CR}\sim 6$T and $H^{a}_{\rm\rm CR}\sim 7$T are relevant to notice from this figure. We also note that two magnetization curves $M_{b}$ and $M_{a}$ behave similarly. It is anticipated that $H_{\rm c2}$ for the two directions should be resemble. This is indeed the case as will be seen next. ## V Application to experiments on three compounds Let us now examine the present theory to understand a variety of experiments on the three compounds, URhGe, UCoGe and UTe2. In order to clarify the essential points of the problem and for the discussions followed to be transparent, and to minimize the free adjustable parameters, we take a simplified minimal version of the present theory. It is quite easy to finely tune our theory by introducing additional parameters such as $\beta_{1}$ and $\beta_{2}$ in the GL theory Eq. (3) for each compound if necessary. We assume that $\displaystyle T_{\rm c1}=T_{\rm c0}+{\kappa}M_{a},$ $\displaystyle T_{\rm c2}=T_{\rm c0}-{\kappa}M_{a},$ $\displaystyle T_{\rm c3}=T_{\rm c0}-bM^{2}_{a}$ (26) for the spontaneous FM moment $M_{a}$ with the easy $a$-axis. We have redefined $\kappa/\alpha_{0}$ as $\kappa$ and $b/\alpha_{0}$ as $b$, ignoring the correction in Eq. (9) from the higher order GL terms. Since $\kappa$ is a converter of the units from $\mu_{B}$ to K, we further simplify the notation in that $\kappa M$ having the dimension of temperature in [K] is denoted as $M$ in [K] in the following phase diagrams as mentioned before. We use the $\kappa$ values for three compounds throughout this paper as shown in Table I where the magnetic properties are also summarized. In the following, we intend to produce the observed $H_{\rm c2}$ curves only qualitatively, not quantitatively. This is because the experimental $H_{\rm c2}$ shapes somewhat depend on the experimental methods. For example, see Fig. 1 in Ref. [wu1, ] where $H_{\rm c2}$ shapes slightly differ each other, depending on the criteria adopted either by the mid-point of the resistivity drop, the zero-resistivity, or by thermalconductivity. We here consider the sharpest curve among them when several choices are available. #### V.0.1 $H\parallel b$: Reentrant SC URhGe exhibits the ferromagnetic transition at $T_{\rm Cuire}=9.5$K where the magnetic easy axis is the $c$-axis and the FM moment $M_{c}=0.4\mu_{B}$. The superconducting transition is at $T_{c}=0.4$K under the ferromagnetic state which is persisting to the lowest $T$. When the field $H$ is applied parallel to the $b$-axis, the superconducting state reappears in a higher field region while the low field SC phase disappears at $H_{\rm c2}\sim 2$T. This reentrant superconducting state (RSC) is explained in Fig. 9, using the knowledge shown in Fig. 7. First we plot the magnetization curves for $M_{c}(H_{b})$ and $M_{b}(H_{b})$ in the $H$-$T$ plane by choosing the $\kappa=2.0{{\rm K}/\mu_{\rm B}}$ in Eq. (26) with $M_{a}$ replaced by $M_{c}$. $M_{c}(H_{b})$ starts from $T_{c1}$ and $T_{c2}$ and decreases by increasing $H_{b}$ which acts to rotate the spontaneous ferromagnetic moment toward the $b$-axis as mentioned above. Thus $T_{c1}(H_{b})=T_{c0}+{\kappa}M_{c}(H_{b})$ and $T_{c2}(H_{b})=T_{c0}-{\kappa}M_{c}(H_{b})$ decreases and increases respectively with increasing $H_{b}$ according to Eq. (26). The splitting $2{\kappa}M_{c}(H_{b})$ between $T_{c1}(H_{b})$ and $T_{c2}(H_{b})$ diminishes and meets at the rotation field $H_{\rm R}$=12T where the two transition temperatures are going to be degenerate. $M_{b}(H_{b})$ starting at $T_{c0}$ quickly increases there. Thus as shown in Fig. 9, $H_{\rm c2}$ starting at $T_{c1}$ disappears at a low field because the orbital depairing dominates over the magnetization effect as explained above. Namely, since the decrease of $T_{c1}(H_{b})$ is slow as a function of $H_{b}$, $H_{\rm c2}$ obeys the usual WHH curve, a situation similar to that shown in Fig. 4(a). Here $|H^{\prime}_{\rm c2}(M)|\gg|H^{\prime orb}_{\rm c2}|$. However, in the higher fields the upper transition temperature $T_{c1}(H_{b})$ becomes $\displaystyle T_{c1}(H_{b})=T_{c0}+{\kappa}M_{b}(H_{b})$ (27) by rotating the $\bf d$-vector so that now it is perpendicular to the $b$-axis in order to grasp the magnetization $M_{b}(H_{b})$. This $\bf d$-vector rotation field corresponds to the field where $\displaystyle T_{c1}(H_{b})=T_{c0}+{\kappa}M_{c}(H_{b})\simeq T_{c0}+{\kappa}M_{b}(H_{b}),$ (28) namely, the $M_{c}(H_{b})$ vector projection onto the $b$-axis $M_{c}/\sqrt{2}\sim M_{b}(H_{b})$ as understood from Fig. 6. Since $M_{b}(H_{b})$ is strongly enhanced at and above $H_{\rm R}$, the A1 phase reappears by following the magnetization curve $T_{c0}+{\kappa}M_{b}(H_{b})$. It ultimately hits the $H^{\rm AUL}_{\rm c2}$ boundary. The RSC finally ceases to exist beyond this boundary. This corresponds to that in Fig. 4(c). The existence of the $H^{\rm AUL}_{\rm c2}$ will be demonstrated later in Fig. 14 where we compile various $H_{\rm c2}$ data for URhGe, including those under hydrostatic pressure miyake2 and uni-axial pressure aoki-uni along the $b$-axis. Figure 9: (color online) The phase diagram for the $H_{b}$(T) versus $T$(K) plane. $M_{c}(H_{b})$ is estimated from the neutron scattering data in Ref. [levy, ] and $M_{b}(H_{b})$ comes from the magnetization curve measured in Ref. [hardy, ]. The red dots for $H_{\rm c2}$ are the experimental data points in Ref. [aokireview, ]. The red continuous line indicates $H_{\rm c2}$ which starts at $T_{c1}$ and is suppressed by the orbital depairing effect. It reappears again by following the formula $T_{c1}(H_{b})=T_{c0}+\kappa M_{b}(H_{b})$ near $H_{R}=11$T. $H^{\prime\rm orb}_{\rm c2}$ ($H^{\prime}_{\rm c2}(M)$) is the slope due to the orbital depairing ($T_{c1}(M_{b})$). #### V.0.2 $\theta$-rotation from $b$ to $c$-axis When the direction of the magnetic field turns from the $b$-axis to the easy $c$-axis, $T_{R}$ moves up to higher fields and disappears quickly around $\theta\sim 5^{\circ}$ as shown in Fig. 7(a). According to those magnetization behaviors, we construct the $H_{\rm c2}$ phase diagram in Fig. 10. It is seen that the field-direction tilting away from the $b$-axis to the $c$-axis results in the decrease of the magnetization $M_{b}(H)$, corresponding to the counter-clock wise changes of the magnetization curves in Fig. 10. Thus the RSC region shifts to higher fields with shrinking their areas and eventually disappears by entering the $H_{\rm c2}^{\rm AUL}$ region. Figure 10: (color online) Reentrant SC (Ref. [levy2, ]) for various $\theta$ values measured from the $b$-axis ($\theta=0$) toward the $c$-axis in URhGe. As $\theta$ increases (0.79∘, 1.65∘, 3.64∘, and 5.64∘), the magnetization curves (far left scale) starting at $T_{c0}$ grows slowly, pushing up the RSC regions to higher fields. The magnetization data are from Fig. 7(a) for $\theta\neq 0$ and Ref. [hardy, ] for $\theta=0$. Figure 11: (color online) Detailed RSC structures (Ref. [levy2, ]) in $T$-$H$ plane (left scale) are displayed. The triangle areas in each $\theta$ are RSC. RSC moves right as $\theta$ increases. The magnetization curve data (right scale) corrected as explained in Fig. 7(a) are originally from Ref. [nakamura, ]. The detailed phase diagram in the reentrant region is depicted in Fig. 11 where the magnetization curves of $M_{b}(H)$ in Fig. 7(a) are overlaid. According to the present theory, $H_{\rm c2}$ follows faithfully $M_{b}(H)$ in the high fields because the strong increase tendency of the magnetization $M_{b}(H)$ overcomes the orbital depression. The characteristics of those phase diagrams are; As $\theta$ increases, (1) The RSC moves up to further higher fields. (2) As $H$ further increases, within the small angles of $\theta$ up to $6^{\circ}\sim 7^{\circ}$ the RSC fades out upon entering $H_{\rm c2}^{\rm AUL}$ region. Those characteristics (1) and (2) nicely match with the experimental observations. The triangle-like shapes for RSC will be seen later in UTe2 (see Fig. 23). #### V.0.3 $\phi$-rotation from $b$ to $a$-axis When the magnetic field direction turns to the other hard $a$-axis from the $b$-axis by the angle $\phi$, the expected magnetization curves are evaluated in Fig. 7(b). Using those magnetization curves, we construct the $H_{\rm c2}$ phase diagrams for various $\phi$ values in Fig. 12. As the angle $\phi$ increases, the magnetization $M(H)$ decreases, corresponding to the clock-wise changes in Fig. 12 and the first order rotation field $H_{R}$ is pushed to higher fields simply because of the projection effect onto the $b$-axis as mentioned in section IV-C. As a consequence, the RSC moves to higher fields persisting up to higher angle $\phi$ until finally entering $H_{\rm c2}^{\rm AUL}$ region. It is confirmed experimentally that it persists at least up to $H_{\rm c2}\sim 25$T levy . According to the present results, the RSC can exist still to higher fields. This can be checked by experiments. Here we notice an important fact that in order to explain the persistence of RSC as a function of $\phi$ up to higher fields, it is essential to use the magnetization curves in Fig. 7(b) where the magnetization contains the component $M_{a}$ in addition to $M_{b}$. It is clear that only $M_{b}$ fails to reproduce the RSC phase diagram. This means that the $\bf d$-vector rotates so as to catch both components $M_{a}$ and $M_{b}$, thus the $\bf d$-vector is always perpendicular to the vectorial sum ${\bf M}_{a}+{\bf M}_{b}$. This is contrasted with the $\theta$ rotation case mentioned above where the $\bf d$-vector is perpendicular to ${\bf M}_{b}$. This intriguing anisotropy in the $\bf d$-vector rotation relative to the magnetic easy axis might be related to the underlying magnetism in URhGe and/or the spin structure of the Cooper pair symmetry assumed as $SO(3)$ originally. This spin space anisotropy should be investigated in future. In Fig. 13 we summarize the phase boundary of the RSC determined above. The band of the RSC region is tightly associated with the $H_{R}(\phi)$ curves, which are proportional to $H_{R}(\phi)\propto 1/\cos(\phi)$. This is contrasted with the lower field $H_{\rm c2}$ which is nearly independent of the angle $\phi$. The intrinsic $H_{\rm c2}$ anisotropy is quite small in URhGe. This means the importance of the magnetization rotation field $H_{R}(\phi)$, ensuring the appearance of the RSC, and pointing to the simple mechanism for the origin of RSC. It grossly follows the ${\bf M}_{b}$ projection onto the $b$-axis. This is also true for the RSC in UTe2, which will be explained shortly. The physics is common. Figure 12: (color online) RSC phase diagram in the $T$-$H$ plane for various fields rotated from the $b$-axis toward the $a$-axis by the angle $\phi$. This is constructed by using the magnetization data (right scale) shown in Fig. 7(b). When the magnetization hits the real axis $T>0$, RSC appears in high field regions. The lower field $H_{\rm c2}$ is common for all $\phi$. Figure 13: (color online) Phase boundary of the reentrance SC (RSC) as a function of the angle $\phi$ measured from the $b$-axis to the $a$-axis constructed from Fig. 12. The blue (green) line indicates the upper (lower) boundary of the RSC. The brown line is the magnetization rotation field $H_{R}(\phi)$. The dots are experimental data points by Ref. [levy2, ]. The triangles denote the lower field $H_{\rm c2}$ which is almost independent of $\phi$. #### V.0.4 Pressure effects Before starting out to analyze the experimental data taken under hydrostatic miyake2 and uni-axial pressure aoki-uni on URhGe, we summarize the relevant data for the $H_{\rm c2}$ phase diagram with the field applied to the $b$-axis in Fig. 14. Here we list up the data under hydrostatic pressure and uni-axial pressure along the $b$-axis. (1) It is clear to see that all the $H_{\rm c2}$ are limited by the common boundary $H^{\rm AUL}_{\rm c2}$. Beyond $H^{\rm AUL}_{\rm c2}$ there exists no $H_{\rm c2}$ data. (2) It is also evident to see that the $H_{R}$ data points under pressure remarkably line up along the bottom of the boundary, forming $H^{\rm AUL}_{\rm c2}$ as an envelop. In the following we utilize those experimental facts and take into account those in investigating and reconstructing the $H_{\rm c2}$ phase diagrams. In Fig. 15 we show the $H_{\rm c2}$ data points taken when $H$ is applied along the $b$-axis under uni-axial pressure $\sigma=1.0GPa$, which is listed in Fig. 14. Those data are explained in a similar way shown above. Here $H_{\rm c2}$ starting at $T_{c1}$ is strongly bent due to the sharp $M_{b}(H_{b})$ rise concomitant with the $\bf d$-vector rotation to catch $M_{b}(H_{b})$ shown by the green line in Fig. 15. Since $M_{b}(H_{b})$ starts at the temperature $T_{c0}$ midway between $T_{c1}$ and $T_{c2}$ separated by $2M_{c}$, the second transition temperature $T_{c2}$ is found to locate there where the $A_{2}$ phase begins developing while the remaining large region is occupied by the $A_{1}$ phase. Now we see the multiple phases in this situation, which is absent under the ambient pressure in URhGe. We can estimate the spontaneous moment $M_{c}$ under $\sigma$=1.0GPa as $M_{c}=0.06\mu_{B}$ on the simple assumption that $\kappa$ is unchanged under the uni-axial pressure. Figure 14: (color online) Phase diagram for $H\parallel b$ taken under hydrostatic pressure (Ref. [miyake2, ]) and uni-axial pressure along the $b$-axis (Ref. [aoki-uni, ]) on URhGe. All data at the rotation field $H_{R}$ line up along the $H^{\rm AUL}_{\rm c2}$ boundary, evidencing the existence of $H^{\rm AUL}_{\rm c2}$. Figure 15: (color online) Multiple phase diagram consisting of the $A_{1}$ and $A_{2}$ phases under uni-axial pressure $\sigma=1.0$GPa in URhGe. The data points of $H_{\rm c2}\parallel b$ are taken from Ref. [aoki-uni, ]. Two transitions at $T_{c1}$ and $T_{c2}$ separated by $2M_{c}$ are identified. $H_{R}$ is the moment rotation field found experimentally aoki-uni . The green line indicates the magnetization curve of $M_{b}$ starting at $T_{c0}$. Figure 16: (color online) Phase diagrams ($H\parallel b$) under uni-axial pressure, including the ambient pressure (a) in Fig. 9 and $\sigma=1.0$GPa (e) in Fig. 15. The data are from Ref. [aoki- uni, ]. Continuous and systematic evolution of the multiple phase diagrams with guide lines are seen. (a) $\sigma=0$GPa, (b) $\sigma=0.2$GPa, (c) $\sigma=0.5$GPa. (d) $\sigma=0.8$GPa, (e) $\sigma=1.0$GPa, and (f) $\sigma=1.2$GPa. We analyze the experimental data available under uni-axial pressure aoki-uni displayed in Fig. 16. It is seen that the continuous and systematic evolution of the multiple phase diagrams under uni-axial pressure. Namely, as uni-axial pressure $\sigma$ increases, three characteristic temperatures $T_{c1}$, $T_{c0}$ and $T_{c2}$ shifts together to higher temperatures. $T_{c2}$ appears at a finite temperature ($T>0$) around $\sigma\sim 0.8$GPa, keeping to move up with increasing further $\sigma$. The separation of $T_{c1}$ and $T_{c2}$ becomes narrow because the spontaneous moment $M_{c}$ gets diminished, corresponding to the observed Curie temperature decrease under uni-axial pressure aoki-uni (see Fig. 17(b)). We show the changes of three temperatures $T_{c1}$, $T_{c0}$ and $T_{c2}$ assigned thus in Fig. 17(a). The separation between $T_{c1}$ and $T_{c2}$ determined by $M_{c}$ diminishes simply because $M_{c}$ decreases as $\sigma$ increases. This results in $T_{c2}>0$ appearing above $\sigma>0.8$GPa, where the double transitions at $H$=0 should be observed. It is remarkable to see that upon approaching $\sigma=1.2$GPa from below, all the transition temperatures are converging toward $\sigma_{\rm cr}=1.2$GPa. This means that above this pressure, the genuine symmetric $A$ phase is realized because the symmetry breaking parameter $M_{c}$ vanishes where the spin symmetry of the pair function restores $SO(3)$ full symmetry, a situation similar to that shown in Fig. 1 (also see Fig. 25 later). At the critical pressure $\sigma_{cr}$=1.2GPa the pairing state is analogous to superfluid 3He-$A$ phase. The resulting analysis of the spontaneous moment $M_{c}$ is shown in Fig. 17(b), revealing a monotonous decrease as $\sigma$ increases. This tendency is matched with the lowering of the Curie temperature, which is observed experimentally aoki-uni . It is interesting to see the linear changes of $T_{c1}$, $T_{c0}$, $T_{c2}$, and $M_{c}$ near the critical uni-axial pressure $\sigma_{\rm cr}=1.2$GPa. This linear relationship is similar to those in UTe2 under hydrostatic pressure around the critical pressure $P_{cr}$=0.2GPa (see Fig. 25 later). Figure 17: (color online) (a) The resulting $T_{c1}$, $T_{c0}$ and $T_{c2}$ obtained from the analysis in Fig. 16 are displayed. The linear changes of those characteristic temperatures $T_{c1}$, $T_{c0}$ and $T_{c2}$ are found, corresponding to the linear decrease in $M_{c}$. The second transition $T_{c2}$ begins appearing above $\sigma>0.8$GPa where the double transitions are expected at $H$=0. (b) The resulting $M_{c}$ change as a function of uni- axial pressure $\sigma$. The observed Curie temperatures (Ref. [aoki-uni, ]) are also shown. It is consistent with the obtained decreasing tendency of $M_{c}$ as $\sigma$ increases. ### V.1 UCoGe UCoGe is another ferromagnetic superconductor worth checking our theory in the same framework for URhGe. Major differences from URhGe in the previous section lie in the fact that (1) The small spontaneous moment $M_{c}=0.06\mu_{B}$. (2) The field induced moments of $M_{b}$ and $M_{a}$ in the hard axes are comparable in magnitude as shown in Fig. 8(a). (3) The magnetization rotation field $H_{R}\sim 45$T is far above $H_{\rm c2}$. Those are contrasted with URhGe with the distinctive induced moment for $M_{b}$ that ultimately leads to the RSC. However, $H_{\rm CR}$ is situated at low fields $6\sim 8$T in UCoGe. #### V.1.1 $H\parallel b$: S-shaped $H_{\rm c2}$ and multiple phases In Fig. 18 we show the result for the phase diagram in $H\parallel b$, assuming that $\kappa=1.8{K\over\mu_{B}}$. The two transition temperatures $T_{c1}$ and $T_{c2}$ are split by $M_{c}=0.06\mu_{B}$. Under the applied field $H_{b}$, the spontaneous moment $M_{c}(H_{b})$ decreases. $T_{c1}$ and $T_{c2}$ approach each other to meet at $H^{b}_{\rm CR}\sim 6$T. Before meeting there, the upper $T_{c1}(H_{b})$ increases and catches the magnetization $M_{b}(H_{b})$ by rotating the $\bf d$-vector direction from the $c$-perpendicular direction to the $b$-perpendicular direction. This results in an S-shaped $H_{\rm c2}$ curve which eventually reaches $H^{\rm AUL}_{\rm c2}$, giving the extrapolated $H^{b}_{\rm c2}\sim 25$T. We notice here that the initial slope of $H^{b}_{\rm c2}$ is small, extrapolated to $H^{b}_{\rm c2}$ less than a few T, which is comparable to $H^{c}_{\rm c2}\sim 0.5$T. This means that the intrinsic $H_{\rm c2}$ anisotropy is within the range of the usual effective mass anisotropy. The same nearly isotropic $H_{\rm c2}$ behavior was just emphasized in URhGe (see Fig. 13). The superficial $H_{\rm c2}$ anisotropy with the order of $H^{b}_{\rm c2}/H^{c}_{\rm c2}$=25T/0.5T$\sim 50$ is an artifact due to ignoring the origin of the S-shaped $H^{b}_{\rm c2}$. This is often pointed out as one of the major mysteries in UCoGe aokireview . It is important to notice that because we identify $T_{c2}=0.2$K there must exist the phase boundary of $A_{1}$ and $A_{2}$ phases. According to thermal- conductivity measurement in Ref. [wu, ] as a function of $H_{b}$, there indeed exists an anomalous thermal-conductivity jump at 10T and low $T$ indicated as the red dot on the $H$-axis in Fig. 18. This nicely matches our identification of the $A_{2}$ phase boundary line, a situation similar to the characteristics in Fig. 4(c) and Fig. 5(b). This assignment is consistent with the $H^{c}_{\rm c2}$ phase diagram as shown shortly. Figure 18: (color online) The S-shaped phase diagram for UCoGe in $H\parallel b$. $H^{b}_{\rm c2}$ starts at $T_{c1}$ is initially depressed by the orbital depairing. At around the crossing field $H_{\rm CR}$ it turns toward higher $T$ due to the $\bf d$-vector rotation to catch $M_{b}(H_{b})$ denoted by the green line, forming the S-shape. At further high fields after hitting $H^{\rm AUL}_{\rm c2}$, $H^{b}_{\rm c2}$ follows it. The experimental data points come from [aokiS, ] and the point at $T$=0 and 10T from [wu, ]. Figure 19: (color online) (a) Weaken S-shaped $H^{a}_{\rm c2}$ for $H\parallel a$ in UCoGe because $H_{\rm\rm CR}$ moves up compared to $H^{b}_{\rm c2}$ case shown in Fig. 18. The data are from Ref. [aokiS, ]. (b) $H^{c}_{\rm c2}$ for $H\parallel c$ in UCoGe. The data [wu, ] clearly show the anomaly around 0.3T, indicating the multiple phases identified as $A_{1}$, $A_{2}$, and $A_{3}$. The magnetization curve of $M_{c}(H_{c})$ is displayed as the green dots, showing the weak rise in this scale. Both $H^{c}_{\rm c2}$ starting at $T_{c1}$ and $T_{c1}$ are thus dominated by the orbital depairing without help of the magnetization. The four points denoted by the red triangles are read off from the thermal-conductivity anomalies [taupin, ]. #### V.1.2 $H\parallel a$ As already shown in Fig. 8(a), the magnetization curves of $M_{b}$ and $M_{a}$ is quite similar. The crossing field $H^{i}_{\rm\rm CR}$ ($i=a$ and $b$) at which $M_{b}(H_{b})$ and $M_{a}(H_{a})$ reach $M_{c}=0.06\mu_{B}$ is seen to be $H^{a}_{\rm\rm CR}\sim 8$T and $H^{b}_{\rm\rm CR}\sim 6$T. Thus $H^{c}_{\rm c2}$ curve is anticipated to be similar too. Indeed the result is shown in Fig. 19(a). Even though the S-shaped $H^{b}_{\rm c2}$ is weaken, it is still seen a weak anomaly at around $H^{a}_{\rm\rm CR}\sim 8$T which is a signature that $T_{c1}(M)$ in Eq. (26) catches $M_{a}(H_{a})$ by rotating the $\bf d$-vector whose direction is perpendicular to the $c$-axis. It is now perpendicular to the $a$-axis. We also point out that the $A_{1}$ and $A_{2}$ phase diagram is essentially the same as in $H\parallel b$ and the extrapolated $H^{a}_{\rm c2}\sim$22T is comparable to $H^{b}_{\rm c2}\sim$25T. #### V.1.3 $H\parallel c$ and multiple phases We display the analysis for the phase diagram in $H\parallel c$ in Fig. 19(b). The existing experimental data clearly indicate that $H^{c}_{\rm c2}$ consists of the two parts where the $H^{c}_{\rm c2}$ enhancement is visible at low $T$ and high $H$. Thus the phase diagram is divided into the three phases, $A_{1}$, $A_{2}$ and $A_{3}$ where $A_{3}$ is genuine spin down-down pair while $A_{2}$ is a mixture of up-up and down-down pairs, or a distorted $A$ phase with different population of the two spin pairs. As indicated in Fig. 19(b) as the green dots, the magnetization curve of $M_{c}(H_{c})$ is weakly increasing in this scale. Thus the slope at $T_{c1}$ is exclusively governed by the orbital depairing, implying that this comes from the effective mass along the $c$-axis. As mentioned above, the anisotropy of the initial slopes in $H_{\rm c2}$ at $T_{c1}$ is determined by their effective mass anisotropy. #### V.1.4 Rotation $\phi$ from the $b$-axis toward the $a$-axis Finally we touch upon the case of the field rotation from the $b$-axis toward the $a$-axis by $\phi$ as shown in Fig. 20. As $H$ is turned from the $b$-axis toward the other hard $a$-axis, the crossing field $H_{\rm CR}$ increases as shown in Fig. 8(a). Since $M_{c}(H)$ becomes slowly increasing as $H$ increases, the orbital depression gets stronger and flattens the initial slopes of $H_{\rm c2}(\phi)$ at $T_{c1}$, eventually approaching $H^{a}_{\rm c2}$ as shown in Fig. 19(a). This is already realized in $\phi=11.4^{\circ}$ case seen from it. It should be pointed out again that those initial slopes at $T_{c1}$ for those $\phi$ values only slightly change, implying that the initial slope is determined by the effective masses, namely the orientational dependent Fermi velocities. So far we assumed that $\kappa=1.8{K/\mu_{B}}$ under the condition of the existence of the second transition $T_{c2}=0.2$K. But we are warned that if those suggestive signatures of the second phase $A_{2}$ coming from thermal- conductivity measurements wu ; taupin may be an artifact, then the forgoing arguments go through and are almost unchanged by taking $\kappa=3.6{K/\mu_{B}}$ without the A2 phase. Namely we are still in ambiguous situations to finally pin down the system parameters. Therefore, it is urgent to confirm or refute the existence of the second transition in order to go further from here. Figure 20: (color online) $H_{\rm c2}(\phi)$ for $\phi=0$∘, 3.2∘, 6.8∘ and 11.4∘ from the $b$-axis toward the $a$-axis in UCoGe. The data are from Ref. [aokiS, ]. As $\phi$ increases, $M_{c}$ grows slowly as a function of $H$ (the counter-clock wise rotation of the $M_{c}$ curves), pushing up $H_{\rm CR}$ to higher fields. This results in the decreases of $H_{\rm c2}(\phi)$ because the orbital suppression becomes dominant. The enhanced $H_{\rm c2}$ becomes diminished as $\phi$ increases. ### V.2 UTe2 To coherently explain a variety of physical properties of superconducting state in UTe2 accumulated experimentally in the same context of the other compounds, URhGe and UCoGe, we need a basic assumption that the ferromagnetic fluctuations are slow enough compared to the electron motion of the conduction electrons, which condense at $T_{c}$. The slow FM fluctuation moments characterized by the non-vanishing square root-mean averaged value $\sqrt{\langle(\delta M_{a})^{2}\rangle}$ over time and space $\langle\cdots\rangle$ are assumed to be able to break the spin symmetry SO(3) of the Copper pairs. In the following we denote this spontaneous and instantaneous FM moment simply $M_{a}$=$\sqrt{\langle(\delta M_{a})^{2}\rangle}$, whose magnitude is adjusted in order to best reproduce the $H_{\rm c2}$ phase diagram as we will see next. #### V.2.1 $H\parallel b$-axis We follow the same method mentioned above for URhGe and UCoGe to understand the observed L-shaped $H^{b}_{\rm c2}$ applied to the magnetic hard $b$-axis. Here we assume that $M_{a}=0.48\mu_{B}$ and $\kappa=6.9{K/\mu_{B}}$. As seen from Fig. 21, $H^{b}_{\rm c2}$ starts from $T_{c1}$=1.6K, following $M_{a}(H_{b})$ which decreases with increasing $H_{b}$ toward $H_{\rm\rm CR}$. $H_{\rm\rm CR}$ is roughly estimated from the magnetization curves shown in the inset of Fig. 8(b) as around 20T. Above $H_{b}>H_{\rm CR}$ the $\bf d$-vector rotates in order to catch the magnetization $M_{b}(H_{b})$, which strongly increases from $T_{c0}$. $H^{b}_{\rm c2}$ begins following it to grow and forms the upper part of the L-shape. It eventually reaches $H_{R}$=32T where the first order transition occurs. As shown in the inset of Fig. 8(b) the magnetization jump at $H_{R}$ amounts to 0.6$\mu_{B}$ known experimentally miyake . The reached magnetization (the horizontal green line) is deep outside $H^{\rm AUL}_{\rm c2}$ shown by dark black colored region in Fig. 21. Therefore $H^{b}_{\rm c2}$ simply stops when it hits the $H_{R}$ line. Those features nicely reproduce the experimental characteristics shown in Fig. 21. Figure 21: (color online) The L-shaped $H^{b}_{\rm c2}$ observed in [knebel, ] is shown (red dots). $H^{b}_{\rm c2}$ starting at $T_{c1}$ follows the orbital suppression plus the $M_{a}$ depression by $H_{b}$ toward $H_{\rm\rm CR}$. When it approaches the strong increasing $M_{b}(H_{b})$, the $\bf d$-vector rotates and catches $M_{b}(H_{b})$ to grow. This forms the upper part of the L-shape. In further high fields $H^{b}_{\rm c2}$ reaches $H_{R}$=32T and disappears there by hitting $H^{\rm AUL}_{\rm c2}$. The green curve denotes the magnetization curve $M_{b}(H_{b})$ shown in the inset of Fig. 8(b)[miyake, ]. #### V.2.2 $\phi$ rotation from the $b$-axis toward the $a$-axis When the field tilts from the $b$-axis toward the magnetic easy $a$-axis by the angle $\phi$, the magnetization $M_{b}(H_{b})$ growth becomes slow compared to that for the $b$-axis as shown in Fig. 22. Those counter-clock wise changes of $M_{b}(H_{b})$ for various angle $\phi$ in Fig. 22 are estimated by the method explained in section IV. Therefore, $H_{\rm c2}$ is bent upward in the upper part of their L-shaped ones while the lower parts are hardly changed because this is mainly limited by the orbital suppression. Those $H_{\rm c2}(\phi)$ curves for various $\phi$ values eventually reach their own $H^{\rm AUL}_{\rm c2}$ which depends on $\phi$, followed by the orbital suppression. Then $H_{\rm c2}(\phi)$ finally disappears abruptly by hitting $H_{R}(\phi)$. If those $H_{\rm c2}(\phi)$ curves extrapolate naively to higher fields beyond $H_{R}(\phi)$, we find $H^{\rm AUL}_{\rm c2}(\phi)$ as shown in the inset of Fig. 22, indicating that $H^{\rm AUL}_{\rm c2}(\phi)$ changes strongly within a few degrees, peaking at $H\parallel b$ sharply. We are not able to explain this peaking phenomenon at this moment. A similar peaking phenomenon is also observed in the $\theta$ side too, where the $H_{\rm c2}(\theta)$ peak occurs at $\theta\sim 35^{\circ}$. Thus the SC region in the $\phi$-$H$ plane is quite limited to small angles up to $\phi\sim 6.3^{\circ}$. As will show next, this is similar to the $\theta$ case where the high field SC $H_{\rm c2}(\theta\sim 35^{\circ})$=60T is observed in a narrow angle $\theta$ region above $T_{R}(\theta)$. Figure 22: (color online) $H_{\rm c2}(\phi)$ for $\phi$=0∘, 2∘, 4∘, 5.2∘, and 6.3∘ from the $b$-axis toward the $a$-axis. $M_{b}(H)$ grows slowly with increasing $\phi$. $H_{\rm c2}(\phi)$ curves bent over. Before hitting $H_{R}(\phi)$ which ultimately limits it, $H_{\rm c2}(\phi)$ turns around with the negative slope because they reach their own $H^{\rm AUL}_{\rm c2}(\phi)$. $M_{b}(H)$ for each $\phi$ is estimated by Eq. (25). The data (dots) are from [knebel, ]. The inset shows $H^{\rm AUL}_{\rm c2}(\phi)$ estimated by extrapolating the straight lines toward higher fields beyond $H_{R}(\phi)$. #### V.2.3 $\theta$ rotation from the $b$-axis toward the $c$-axis It is remarkable to see the extremely high $H_{\rm c2}(\theta=35^{\circ})\sim 60$T when the field is tilted from the $b$-axis toward the other magnetic hard $c$-axis ran2 . This is detached from the low field $H_{\rm c2}(\theta)\sim 8$T. This low field SC part is nearly independent of $\theta$. This $H_{\rm c2}$ isotropy was seen also in URhGe (see Fig. 13) and UCoGe. This extremely high $H_{\rm c2}(\theta=35^{\circ})$ can be understood by the present framework as follows. We begin with the $H\parallel b$ case discussed in Fig. 21. Upon increasing $\theta$, the magnetization $M_{b}(H)$ becomes slow to grow. Around $\theta=12^{\circ}$ the upper part of the L-shaped $H_{\rm c2}$ separates into two parts as shown in Fig. 23 which is observed georg . And eventually this RSC part disappears above $\theta>12^{\circ}$, leaving only the lower $H_{\rm c2}$ part at around 10T. Further increasing $\theta$, the magnetization $M_{b}(H)$ starting from $T_{c2}$ becomes relevant because as explained in Fig. 8(b), $M_{b}(H)$ becomes small and the magnetization jump also diminishes. Around $\theta=35^{\circ}$ the magnetization curves are just available for the reentrant SC to appear at higher fields above the respective $H_{R}(\theta)$. This RSC is shown in Fig. 23. This is because the state reached after the first order jump is now within the $H^{\rm AUL}_{\rm c2}$ allowed region. Thus RSC only appears within the narrow angle region centered at $\theta=35^{\circ}$. Those RSC regions are characterized by a triangle like shape as observed in [ran2, ]. This RSC shape resembles those in Figs. 9 and 10 for URhGe. Figure 23: (color online) $H_{\rm c2}(\theta)$ for various $\theta$, which is measured from the $b$-axis toward the $c$-axis. The magnetization curves of $M_{b}(H)$ starting at $T_{c2}$ and $T_{c0}$ evaluated before (see Fig. 8(b)) lead to the reentrant SC for $\theta=35^{\circ}$ in addition to the low $H_{\rm c2}$. For the lower angle of $\theta=12^{\circ}$ the two separate SC are formed. Here the $\theta=0^{\circ}$ case ($H\parallel b$) is shown for reference. It is seen that the magnetization curves only around $\theta\sim 35^{\circ}$ allow RSC to appear. #### V.2.4 Phase diagrams under pressure and multiple phases Let us examine the pressure effects on the $H_{\rm c2}$ phase diagram, which give us another testing ground to check the present scenario. In Fig. 24 (a) we show the data (dots) of $H^{b}_{\rm c2}$ for $H\parallel b$ under $P=0.4$GPa aokiP together with our analysis. It is seen that since the magnetization curve $M_{b}(H_{b})$ denoted by the green line strongly increases, $H^{b}_{\rm c2}$ started at $T_{c1}$ exhibits a bent toward higher temperatures at around $H_{\rm\rm CR}$. The two magnetization curves started from $T_{c1}$ and $T_{c2}$ meet at $H_{\rm\rm CR}$. After passing the field $H_{\rm\rm CR}$, $H^{b}_{\rm c2}$ with a positive slope heads toward $H_{R}=30$T, which is observed as the first order transition aokiP . The same feature is observed so far several times in URhGe under uni-axial pressure such as in Fig. 15 and UCoGe in Fig. 18. The second transition at $T_{c2}$ with the A2 phase is clearly found experimentally shown there detected by AC calorimetry by Aoki, et al aokiP . Moreover, the lower $H^{b}_{\rm c2}$ started from $T_{c2}$ shows an anomaly at around 5T in Fig. 24(a), suggesting the third transition $T_{c3}$. This identification is quite reasonable when we see Fig. 24(b) where the $H\parallel a$ case is displayed for the same $P=0.4$GPa. Indeed we can consistently identify $T_{c3}$ in this field orientation too. According to our theory three phases $A_{1}$, $A_{2}$, and $A_{0}$ correspond to $T_{c1}$, $T_{c2}$, and $T_{c3}$ respectively as shown there. In the high fields, we enumerate further phases $A_{4}$ and $A_{5}$. Those lower $T$ and high $H$ phases are the mixtures of the fundamental three phases $A_{1}$, $A_{2}$, and $A_{0}$ except for $A_{5}$, which is genuine $A_{0}$. For example, the $A_{4}$ phase consists of the $A_{1}$ and $A_{0}$ phases. In Fig. 24(c) we show the data of $H^{b}_{\rm c2}$ for $H\parallel b$ under $P=1.0$GPa aokiP together with our analysis. As $P$ increases, the first order transition field $H_{R}$ becomes lower, here it is $H_{R}=$20T at $P=1.0$GPa from 30T at $P=0.4$GPa. $H^{b}_{\rm c2}$ just follows a straight line due to the orbital depairing all the way up to $H_{R}$ where the magnetization $M_{b}(H_{b})$ denoted by the green line exhibits the magnetization jump. This jump is large enough to wipe out the SC state there. Thus $H^{b}_{\rm c2}$ now follows a horizontal line at $H_{R}=20$T. This is the same case as in $H^{b}_{\rm c2}$ seen in the ambient pressure (see Fig. 21). The main difference from the ambient case is that the second transition at $T_{c2}$ is now visible and observable because the FM moment $M_{a}$ diminishes under pressure and the pressure $P=0.4$GPa is situated near the critical pressure at $P=0.2$GPa (see Fig. 25). This proves the consistency of our scenario. As shown in Fig. 24(d) where at $P$=0.7GPa for $H\parallel a$ the $H_{\rm c2}$ data points are quoted from Ref. [aokiP, ], we draw the three continuous lines to connect those points. We find the missing third transition along the $T$-axis at $T_{c3}$=0.5K. Note that the tricritial point with three second order lines is thermodynamically forbidden yip . The multiple phases are enumerated, such as $A_{1}$, $A_{2}$, and $A_{0}$ at the zero-field and $A_{4}$, and $A_{5}$ at finite fields. Those phases are consisting of the coexistence of the plural fundamental three components $A_{1}$, $A_{2}$, and $A_{0}$. Namely, those are characterized by A1 at $T_{\rm c1}$, A${}_{2}\rightarrow$A1+A2 at $T_{\rm c2}$, A${}_{0}\rightarrow$ A1+A2+A0 at $T_{\rm c3}$, A${}_{4}\rightarrow$A1+A0, and A${}_{5}\rightarrow$$A_{0}$. It is understood that this phase diagram is quite exhaustive, no further state is expected in our framework. At the intersection points in Fig. 21(d) the four transition lines should always meet together according to the above general rule and thermodynamic considerations yip . The lines indicate how those three phases interact each other, by enhancing or suppressing. $T_{\rm c3}$ could be raised by the presence of the A1 and A2 phases due to the fourth order term $Re(\eta^{2}_{a}\eta_{+}\eta_{-})$ mentioned in section II. Figure 24: (color online) $H_{\rm c2}$ and the associated internal phase transition lines under hydrostatic pressure $P$ in UTe2. (a) $P$=0.4GPa and $H//b$. (b) $P$=0.4GPa and $H//a$. (c) $P$=1.0GPa and $H//b$. (d) $P$=0.7GPa and $H//a$. The data denoted by the red dots are from Ref. [aokiP, ]. $T_{c1}$ and $T_{c2}$ at $H=0$ are split by the magnetization $M_{a}$ which decreases under the applied field $H_{b}$ as shown in (a) and (c). This decrease of $M_{a}(H_{b})$ is compensated by growing of the magnetization $M_{b}(H_{b})$ as denoted by the green lines there. In Fig. 25 we compile all the data daniel ; aokiP of the phase transitions in the $T$-$P$ plane at $H=0$. As $P$ increases from $P=0$, $T_{c1}$ ($T_{c2}$) decreases (increases) to meet at the critical pressure $P_{cr}=0.2$GPa where $T_{c3}$ is also merging to converge all three transition lines. This critical pressure corresponds to the degenerate point where the symmetry breaking parameter $M_{a}$ vanishes and the three phases $A_{1}$, $A_{2}$, and $A_{0}$ becomes degenerate, restoring the full SO(3) spin symmetry at this critical point. Upon further increasing $P$, the three phases are departing from there. The three data points for $T_{c3}$ (the three red triangles on the $T_{c3}$ line in Fig. 25 are inferred from Fig. 24). The fact that $T_{c1}$ and $T_{c2}$ behave linearly in $P$ is understood as the linear relationship between $P$ and $M_{a}(P)$, leading to the linear changes of $T_{c1}$ and $T_{c2}$. This linear relationship is also seen in Fig. 17. Simultaneously a strong departure of $T_{c3}$ from the critical pressure. This is because $T_{c3}$ changes in proportion of $M_{a}^{2}$ as mentioned before (see Eq. (26)). This $T$-$P$ phase diagram is similar to that shown in Fig. 1 globally and topologically, proving that the present scenario is valid for this compound too. Figure 25: (color online) $T$-$P$ phase diagram in UTe2 with three transition temperatures $T_{c1}$, $T_{c2}$ and $T_{c3}$ corresponding to the $A_{1}$, $A_{2}$, and $A_{0}$ phases respectively. At the degenerate point of $P_{\rm\rm cr}=0.2$GPa all three phases converges. The lines for $T_{c1}$ and $T_{c2}$ as a function of $P$ indicate that the underlying symmetry breaking field $M_{a}$ changes linearly with $P$, leading to the globally quadratic variation of $T_{c3}$ from the degenerate point. The red (dark blue) round dots are from the experiment [aokiP, ] ([daniel, ]) except for the three red triangle points at $P$=0.40, 0.54 and 0.70GPa for $T_{c3}$, which are inferred from Fig. 24. ## VI Pairing symmetry ### VI.1 Gap symmetries and nodal structures The classification of the gap or orbital symmetries allowed in the present orthorhombic crystal has been done before ohmi ; annett . Among those classified pairing states, the appropriate gap function $\phi(k)$ is selected as follows: $\phi(k)=k_{a}k_{b}k_{c}$ (A1u), $\phi(k)=k_{b}$ (B1u), $\phi(k)=k_{c}$ (B2u), and $\phi(k)=k_{a}$ (B3u). The gap structure is characterized by the line nodes for those states. They are all candidates for URhGe and UCoGe as tabulated in Table II. This leads to the overall pairing function: ${\bf d}(k)=(\vec{a}\pm i\vec{b})\phi(k)$, which breaks the time reversal symmetry. This gap structure with the line nodes is consistent with the NMR experiment manago1 , reporting that 1/T1 is proportional to $T^{3}$ at low temperatures. The line nodes also suggested by other experiments on UCoGe taupin ; wu . As for UTe2, the specific heat experiments ran ; aoki2 ; metz ; kittaka exhibit $C/T\sim T^{2}$, suggesting that the gap structure is characterized by point nodes. This is also consistent with the microwave measurements 1 . Then we have to resort, an ad hoc orbital function, namely $\phi(k)=k_{b}+ik_{c}$ beyond the group-theoretical classification scheme ozaki1 ; ozaki2 , thus the resulting overall pairing function is given by ${\bf d}(k)=(\vec{b}\pm i\vec{c})(k_{b}+ik_{c})$. This pairing state is also the time reversal broken state both in spin and orbital parts. The point nodes are oriented along the $a$-axis determined by angle resolved specific heat experiment kittaka . This is characterized by the Weyl nodes analogous to superfluid 3He-A phase mizushima1 ; mizushima2 . This double chiral state both in the spin space and orbital space might be energetically advantageous because the spin and orbital moments for Cooper pairs are parallel, namely the orbital angular moment ${\bf L}$ that is spontaneously induced by this chiral state can gain the extra energy through the coupling ${\bf M}_{s}\cdot{\bf L}$ with the spontaneous magnetic moment ${\bf M}_{s}\propto{\bf d}\times{\bf d}^{\ast}$. This is consistent with the experiments by angle-resolved specific heat measurement kittaka , the STM observation mad , and the polar Kerr experiment hayes among other thermodynamic experiments nakamine . ### VI.2 Residual density of states All the compounds exhibit more or less the residual density of states at the lowest $T$ limit in the specific heat measurements aokireview ; kittaka . This is not a dirt effect of the samples used, but it is intrinsic deeply rooted to the pairing state identified as the A1 phase. In the A1 phase the superconducting DOS has intrinsically the “residual density of states”. Since $T_{c1}$ with the A1 phase is higher than $T_{c2}$ with the A2 phase, it is reasonable to expect the the DOS $N_{A_{1}}(0)$ in the A1 phase is larger than that in the A2 phase, that is, $N_{A_{1}}(0)>N_{A_{2}}(0)$ because in the Zeeman split bands, the major spin component band with larger DOS preferentially forms the higher $T_{c}$ superconducting state rather than the minority band. It is quite reasonable physically that in UTe2 at the ambient pressure the observed “residual density of states” corresponding to $N_{A_{2}}(0)$ is less than 50$\%$. Table 2: Possible Pairing Functions Compound | spin part | orbital part $\phi(k)$ ---|---|--- URhGe | $\vec{a}\pm i\vec{b}$ | $k_{a}k_{b}k_{c}$(A1u), $k_{b}$(B1u), $k_{c}$(B2u), $k_{a}$(B3u) UCoGe | $\vec{a}\pm i\vec{b}$ | $k_{a}k_{b}k_{c}$(A1u), $k_{b}$(B1u), $k_{c}$(B2u), $k_{a}$(B3u) UTe2 | $\vec{b}\pm i\vec{c}$ | $k_{b}+ik_{c}$ ### VI.3 Multiple phase diagram Our three component spin-triplet state leads intrinsically and naturally to a multiple phase diagram consisting of the A0 phase at $T_{c3}$, A1 at $T_{c1}$, and A2 at $T_{c2}$ as shown in Fig. 1 under non-vanishing symmetry breaking field due to the spontaneous moment. Depending on external conditions, such as $T$, $H$, and its direction, or pressure, etc, the structure of the multiple phase diagram is varied as explained. In fact under $P$, the successive double transitions are clearly observed in UTe2 daniel and they vary systematically in their $P$-$T$ phase diagram of Fig. 25. We see even the third transition centered around the critical pressure $P_{cr}=0.2$GPa. At the ambient pressure on UTe2 the occurrence of the second transition is debated hayes ; rosa , including the detailed internal phase lines. But they agree upon the existence of the multiple phases. As for UCoGe, the thermalconductivity experiment taupin indicates an anomaly at $T=0.2K$, which coincides roughly with our prediction shown in Figs. 18 and 19. As a function of $H(\parallel b)$, the thermalconductivity anomaly is detected as a sudden increase at $H\sim 0.6H_{\rm c2}$ (see Fig. 5 in Ref. [wu1, ]). Moreover, under $H$ parallel to the easy $c$-axis, the $H_{\rm c2}$ curve in Fig. 19(b) shows an enhancement at low $T$ indicative of the underlaying phase transition (see Fig. 2(b) in Ref. [wu, ]). According to the NMR by Manago et al manago1 ; manago2 , $1/TT_{1}$ presents a similar $T$ behavior, such as a plateau at $\sim N(0)/2$ and then sudden drop upon lowering $T$, as mentioned above. We propose to conduct further careful experiments to detect the A1-A2 transitions in this compound. In URhGe at the ambient pressure shown in Fig. 9 both low field phase and the RSC phase belong to the A1 phase. However, under the uni-axial pressure along the $b$-axis, there is a good chance to observe the second transition as explained in Figs. 15 and 16. Therefore to confirm the generic multiple phase diagram for all three compound shown in Fig. 1 is essential to establish the present scenario and also detect characteristics of each pairing state associated with those multiple phases. ### VI.4 Symmetry breaking mechanism For URhGe and UCoGe the “static” FM transitions are firmly established, there is no doubt for the spontaneous FM moment to be a symmetry breaking field. Slow FM fluctuations are found in UTe2 ran ; miyake ; tokunaga ; sonier which could be the origin of the symmetry breaking of $T_{\rm c1}\neq T_{\rm c2}$ under the assumption that FM fluctuations are slow compared to the conduction electron motion. A similar observation is made in UPt3: The fluctuating antiferromagnetism (AF) at $T_{N}=5$K is detected only through the fast probe: “nominally elastic” neutron diffraction aeppli ; trappmann and undetected through other “static” probes, such as specific heat, $\mu$SR, and NMR. Thus the AF fluctuating time scale is an order of MHz or faster. This is believed to be the origin of the double transition in UPt3 UPt3 ; sauls . In UTe2 it is essential and urgent to characterize the observed ferromagnetic fluctuations in more detail, such as fluctuation time scale, or spatial correlation. Elastic and inelastic neutron scattering experiments are ideal tools for it, which was the case in UPt3. It may be too early to discuss the pairing mechanism before confirming the non-unitary spin triplet state. There already exists an opinion appel which advocates longitudinal ferromagnetic fluctuations to help stabilizing a spin triplet state before the discoveries of those compounds. A problem of this sort is how to prove or refute it, otherwise it is not direct evidence and remains only circumstantial one. We need firm objective “evidence” for a pairing mechanism. Theory must be verifiable. ### VI.5 Common and different features As already seen, URhGe, UCoGe, and UTe2 are viewed coherently from the unified point: the non-unitary triplet state. They share the common features: (1) The unusual $H_{\rm c2}$ curves occur for the field direction parallel to the magnetic hard $b$-axis, where the magnetization curve $M_{b}(H_{b})$ exhibits the first order transition at $H_{R}$ for URhGe and UTe2, corresponding to the FM moment rotation. (2) Under pressure they show the critical point behaviors $P_{cr}=0.2$GPa for UTe2 and $\sigma_{cr}=1.2$GPa for URhGe at which the split $T_{c1}$ and $T_{c2}$ converges, leading to the $SO(3)$ spin symmetry for Cooper pairs. (3) The multiple phases, including the reentrant SC, are observed and explained in URhGe and UTe2 and expected to be confirmed for UCoGe. (4) The GL parameter $\kappa$ characterizing the strength of the symmetry breaking are tabulated in Table I, showing the similar values for three compounds. As a general tendency $\kappa$ is likely larger when the FM moment is larger because it is originated from the particle-hole asymmetry of the density of states $N(0)$ at the Fermi level. There are different features: (1) The nodal structures are points oriented along the magnetic easy $a$-axis in UTe2 while lines in URhGe and UCoGe. (2) Under the ambient pressure, $H_{\rm c2}$ curves are seemingly different as in Fig. 9 for URhGe, Fig. 18 for UCoGe, and Fig. 21 for UTe2. But it is now understood as mere differences in $T_{c0}$ or the FM moments as the symmetry breaker. From this comparison, the superconductivity in UTe2, URhGe and UCoGe should be understood by the unified view point, which is more resourceful and productive than considered differently and individually. ### VI.6 Double chiral non-unitary state in UTe2 Since UTe2 attracts much attention currently, it is worth summing up our thoughts on this system to challenge novel experiments. When combining the experimental observations of the chiral current along the wall by STM mad and the angle-resolved specific heat experiment kittaka , the double chiral non- unitary symmetry described by ${\bf d}(k)=({\hat{b}}+i{\hat{c}})(k_{b}+ik_{c})$ is quite possible: This pairing state produces the chiral current at the edges of domain walls, consistent with the former observation. And it is consistent with the polar Kerr experiment hayes which shows the broken time reversal symmetry. In this pairing state the point nodes orient along the magnetic easy $a$-axis, which is supported by the angle-resolved specific heat experiment kittaka . This experiment further indicates the unusual Sommerfeld coefficient $\gamma(H)$ in the superconducting state for $H$ along the $a$-axis. The low energy quasi- particle excitations naively expected for the point nodes miranovic is absent. This lack of the nodal excitations is understood by taking into account that $T_{c}$ depends on $H$ through the magnetization. This is indeed consistent with the notion of the field-tuned SC developed throughout the present paper. ## VII Summary and Conclusion We have discussed the superconducting properties of URhGe, UCoGe, and UTe2 in detail in terms of a non-unitary spin triplet pairing state in a unified way. The spontaneous static ferromagnetic moment in URhGe and UCoGe, and the slowly fluctuating instantaneous ferromagnetic moment in UTe2 break the spin $SO(3)$ symmetry in the degenerate triplet pairing function with three components. Those produce the various types of the $H_{\rm c2}$ curves that are observed. The possible pairing function is described by the complex ${\bf d}$-vector, whose direction is perpendicular to the magnetic easy axis at zero-field. Its direction changes under applied field parallel to the magnetic hard $b$-axis common in three compounds. This $\bf d$-vector rotation is driven by the induced magnetic moment under applied fields. Thus the SC order parameter is tunable by the magnetic field in this sense, ultimately leading to the reentrant SC in URhGe, S-shape in UCoGe, and L-shape $H_{\rm c2}$ in UTe2. As for UTe2, we can study a variety of topological properties, such as Weyl nodes associated with the point nodes, known in 3He A-phase mizushima1 ; mizushima2 , which was difficult to access experimentally and remains unexplored in the superfluid 3He. We can hope to see in UTe2 similar exotic vortices and Majorana zero modes predicted in 3He phase mizushima1 ; mizushima2 ; tsutsumi1 ; tsutsumi2 . There are several outstanding problems to be investigated in future, such as the pairing mechanism leading to the present non-unitary state where longitudinal spin fluctuations are plausible, but how to prove or to refute it. That is a question. As a next step, microscopic theory and detailed calculations are definitely needed beyond the present GL framework where the most simplified version is adopted in order to just illustrate the essential points. For example, we did not seriously attempt to produce the observed $H_{\rm c2}$ curves quantitatively because of the reasons mentioned at the beginning of section V. Thus we only scratches its surface admittedly. It is our hope that the present theory motivates ingenious experiments in this fruitful and flourishing research field. ## Acknowledgments The author is grateful for the enlightening discussions with Y. Shimizu, Y. Tokunaga, A. Miyake, T. Sakakibara, S. Nakamura, S. Kittaka, G. Knebel, A. Huxley, and K. Ishida. He would especially like to thank D. Aoki for sharing data prior to publication and stimulating discussions. He thanks K. Suzuki for helping to prepare the figures. This work is supported by JSPS KAKENHI, No.17K05553. ## References * (1) K. Machida and M. Kato, “Inherent spin-density instability in heavy-Fermion superconductivity”, Phys. Rev. Lett. 58, 1986 (1987). * (2) M. Kato and K. Machida, “Superconductivity and spin-density waves –Application to heavy-Fermion materials”, Phys. Rev. B 37, 1510 (1988). * (3) Eduardo Fradkin, Steven A. Kivelson, and John M. Tranquada, “Theory of intertwined orders in high temperature superconductors”, Rev. Mod. Phys. 87, 457 (2015). * (4) D. Aoki, K. Ishida, and J. Flouquet, “Review of U-based Ferromagnetic Superconductors: Comparison between UGe2, URhGe, and UCoGe”, J. Phys. Soc. Jpn. 88, 022001 (2019). * (5) S. S. Saxena, P. Agarwal, K. Ahilan, F. M. Grosche, R. K. W. Haselwimmer, M. J. Steiner, E. Pugh, I. R. Walker, S. R. Julian, P. Monthoux, G. G. Lonzarich, A. Huxley, I. Sheikin, D. Braithwaite, and J. Flouquet, “Superconductivity on the border of itinerant-electron ferromagnetism in UGe2”, Nature 406, 587 (2000). * (6) Dai Aoki, Andrew Huxley, Eric Ressouche, Daniel Braithwaite, Jacques Flouquet, Jean-Pascal Brison, Elsa Lhotel, and Carley Paulsen “Coexistence of superconductivity and ferromagnetism in URhGe”, Nature 416, 613 (2001). * (7) N. T. Huy, A. Gasparini, D. E. de Nijs, Y. Huang, J. C. P. Klaasse, T. Gortenmulder, A. de Visser, A. Hamann, T. Görlach, and H. v. Löhneysen, “Superconductivity on the Border of Weak Itinerant Ferromagnetism in UCoGe”, Phys. Rev. Lett. 99, 067006 (2007). * (8) S. Ran, C. Eckberg, Q-P. Ding, Y. Furukawa, T. Metz, S. H. Saha, I-L. Liu, M. Zie, H. Kim, J. Paglione, and N. P. Butch, “Nearly ferromagnetic spin-triplet superconductivity”, Science 365, 684 (2019). * (9) D. Aoki, A. Nakamura, F. Honda, D. Li, Y. Homma, Y. Shimizu, Y. J. Sato, G. Knebel, J. P. Brison, A. Pourret, D. Braithwaite, G. Lapertot, Q. Niu, M. Vališka, H. Harima, and J. Flouquet, “Unconventional superconductivity in heavy Fermion UTe2”, J. Phys. Soc. Jpn. 88, 043702 (2019). * (10) K. Machida, “Coexistence problem of magnetism and superconductivity”, Applied Phys. 35, 193 (1984). * (11) G. Knebel, W. Knafo, A. Pourret, Q. Niu, M. Vališka, D. Braithwaite, G. Lapertot, M. Nardone, A. Zitouni, S. Mishra, I. Sheikin, G. Seyfarth, J.-P. Brison, D. Aoki, and J. Flouquet, “Field-reentrant superconductivity close to a metamagnetic transition in the heavy-fermion superconductor UTe2”, J. Phys. Soc. Jpn. 88, 063707 (2019). * (12) D. Braithwaite, M. Vališka, G. Knebel, G. Lapertot, J.- P. Brison, A. Pourret, M. E. Zhitomirsky, J. Flouquet, F. Honda, and D. Aoki, “Multiple superconducting phases in a nearly ferromagnetic system”, Commun. Phys. 2, 147 (2019). * (13) A. Miyake, Y. Shimizu, Y. J. Sato, D. Li, A. Nakamura, Y. Homma, F. Honda, J. Flouquet, M. Tokunaga, and D. Aoki, “Metamagneitc transition in heavy Fermion superconductor UTe2”, J. Phys. Soc. Jpn. 88, 063706 (2019). * (14) Sheng Ran, I-Lin Liu, Yun Suk Eo, Daniel J. Campbell, Paul Neves, Wesley T. Fuhrman, Shanta R. Saha, Christopher Eckberg, Hyunsoo Kim, Johnpierre Paglione, David Graf, John Singleton, and Nicholas P. Butch, “Extreme magnetic field-boosted superconductivity”, Nature Phys. 15, 1250 (2019). * (15) T. Metz, S. Bao, S. Ran, I-L. Liu, Y. S. Eo, and W. T. Fuhrman, D. F. Agterberg, S. Anlage, N. P. Butch, and J. Paglione, “Point node gap structure of spin-triplet superconductor UTe2”, Phys. Rev. B 100, 220504 (R) (2019). * (16) Lin Jiao, Zhenyu Wang, Sheng Ran, Jorge Olivares Rodriguez, Manfred Sigrist, Ziqiang Wang, Nicholas Butch, and Vidya Madhavan, “Microscopic evidence for a chiral superconducting order parameter in the heavy fermion superconductor UTe2”, Nature, 579, 523 (2020). * (17) Yo Tokunaga, Hironori Sakai, Shinsaku Kambe, Taisuke Hattori, Nonoka Higa, Genki Nakamine, Shunsaku Kitagawa, Kenji Ishida, Ai Nakamura, Yusei Shimizu, Yoshiya Homma, DeXin Li, Fuminori Honda, and Dai Aoki, “125Te-NMR study on a single crystal of heavy fermion superconductor UTe2”, J. Phys. Soc. Jpn. 88, 073701 (2019). * (18) S. Sundar, S. Gheidi, K. Akintola, A. M. Côtè, S. R. Dunsiger, S. Ran, N. P. Butch, S. R. Saha, J. Paglione, and J. E. Sonier, “Coexistence of ferromagnetic fluctuations and superconductivity in the actinide superconductor UTe2”, Phys. Rev. B 100, 140502 (R) (2019). * (19) G. Nakamine, Shunsaku Kitagawa, Kenji Ishida, Yo Tokunaga, Hironori Sakai, Shinsaku Kambe, Ai Nakamura, Yusei Shimizu, Yoshiya Homma, Dexin Li, Fuminori Honda, and Dai Aoki, “Superconducting properties of heavy fermion UTe2 revealed by 125Te-nuclear magnetic resonance”, J. Phys. Soc. Jpn. 88, 113703 (2019). * (20) Ian M. Hayes, Di S. Wei, Tristin Metz, Jian Zhang, Yun Suk Eo, Sheng Ran, Shanta R. Saha, John Collini, Nicholas P. Butch, Daniel F. Agterberg, Aharon Kapitulnik, Johnpierre Paglione, “Weyl Superconductivity in UTe2”, arXiv:2002.02539 * (21) Seokjin Bae, Hyunsoo Kim, Sheng Ran, Yun Suk Eo, I-Lin Liu, Wesley Fuhrman, Johnpierre Paglione, Nicholas P Butch, and Steven Anlage, “Anomalous normal fluid response in a chiral superconductor”, arXiv:1909.09032. * (22) Sheng Ran, Hyunsoo Kim, I-Lin Liu, Shanta Saha, Ian Hayes, Tristin Metz, Yun Suk Eo, Johnpierre Paglione, and Nicholas P. Butch, “Enhanced spin triplet superconductivity due to Kondo destabilization”, arXiv:1909.06932. * (23) Qun Niu, Georg Knebel, Daniel Braithwaite, Dai Aoki, Gérard Lapertot, Gabriel Seyfarth, Jean-Pascal Brison, Jacques Flouquet, and Alexandre Pourret, “Fermi-surface instabilities in the heavy-Fermion superconductor UTe2”, arXiv:1907.11118. * (24) V. Hutanu, H. Deng, S. Ran, W. T. Fuhrman, H. Thoma, and N. P. Butch, “Crystal structure of the unconventional spin-triplet superconductor UTe2 at low temperature by single crystal neutron diffraction”, arXiv:1905.04377. * (25) S. M. Thomas, F. B. Santos, M. H. Christensen, T. Asaba, F. Ronning, J. D. Thompson, E. D. Bauer, R. M. Fernandes, G. Fabbris, and P. F. S. Rosa, “Evidence for a pressure-induced antiferromagnetic quantum critical point in intermediate valence UTe2”, Science Adv. 6, eabc8709 (2020). * (26) L. P. Cairns, C. R. Stevens, C. D. O’Neill, and A. Huxley, “Composition dependence of the superconducting properties of UTe2”, J. Phys. Condens. Matter, 32, 415602 (2020). * (27) Dai Aoki, Fuminori Honda, Georg Knebel, Daniel Braithwaite, Ai Nakamura, DeXin Li, Yoshiya Homma, Yusei Shimizu, Yoshiki J. Sato, Jean-Pascal Brison, and Jacques Flouquet, “Multiple Superconducting Phases and Unusual Enhancement of the Upper Critical Field in UTe2”, J. Phys. Soc. Jpn. 89, 053705 (2020). * (28) Shunichiro Kittaka, Yusei Shimizu, Toshiro Sakakibara, Ai Nakamura, Dexin Li, Yoshiya Homma, Fuminori Honda, Dai Aoki, and Kazushige Machida, “Orientation of point nodes and nonunitary triplet pairing tuned by the easy-axis magnetization in UTe2”, Phys. Rev. Research 2, 032014(R) (2020). * (29) Y. Xu, Y. Sheng, and Y. Yang, “Quasi-two-dimensional Fermi surfaces and unitary spin-triplet pairing in the heavy fermion superconductor UTe2”, Phys. Rev. Lett. 123, 217002 (2019). * (30) J. Ishizuka, S. Sumita, A. Dido, and Y. Yanase, “Insulator-metal transition and topological superconductivity in UTe2 from a first -principles calculation”, Phys. Rev. Lett. 123, 217001 (2019). * (31) A. B. Shick and W. E. Pickett, “Spin-orbit coupling induced degeneracy in the anisotropic unconventional superconductor UTe2”, Phys. Rev. B100, 134502 (2019). * (32) A. Nevidomskyy, arXiv:2001.02699. * (33) M. Fidrysiak, D. Goc-Jagło, E. Kadzielawa-Major, P. Kubiczek, and J. Spałek, “Coexistent spin-triplet superconducting and ferromagnetic phases induced by Hund’s rule coupling and electronic correlations: Effect of the applied magnetic field”, Phys. Rev. B99, 205106 (2019). * (34) V. Yarzhemsky and E. A. Teplyakov, “Time-Reversal Symmetry and the Structure of Superconducting Order Parameter of Nearly Ferromagnetic Spin-Triplet Superconductor UTe2”, arXiv:2001.02963. * (35) Kozo Hiranuma and Satoshi Fujimoto, “Paramagnetic effects of j-electron superconductivity and application to UTe2”, arXiv:2010.05112. * (36) A. G. Lebed, “Restoration of superconductivity in high magnetic fields in UTe2”, Mod. Phys. Lett. 34, 2030007 (2020). * (37) Kazumasa Miyake, “On Sharp Enhancement of Effective Mass of Quasiparticles and Coefficient of T2 Term of Resistivity around First-Order Metamagnetic Transition Observed in UTe2”, J. Phys. Soc. Jpn. 90, 024701 (2021). * (38) F. Lévy, I. Sheikin, B. Greinier, and A. D. Huxley, “Magnetic field-induced superconductivity in the ferromagnet URhGe”, Science, 309, 1343 (2005). * (39) D. Aoki, T. D. Matsuda, V. Taufour, E. Hassinger, G. Knebel, and J. Flouquet, “Extremely Large and Anisotropic Upper Critical Field and the Ferromagnetic Instability in UCoGe”, J. Phys. Soc. Jpn. 78, 113709 (2009). * (40) M. Manago, Shunsaku Kitagawa, Kenji Ishida, Kazuhiko Deguchi, Noriaki K. Sato, and Tomoo Yamamura, “Superconductivity at the pressure-induced ferromagnetic critical region in UCoGe”, J. Phys. Soc. Jpn. 88, 113704 (2019). * (41) M. Manago, private communication. * (42) A. J. Leggett, “A theoretical description of the new phases of liquid 3He”, Rev. Mod. Phys. 47, 331 (1975). * (43) W. P. Halperin, C. N. Archie, F. B. Rasmussen, T. A. Alvesalo, and R. C. Richardson, “Specific heat of normal and superfluid 3He on the melting curve”, Phys. Rev. B 13, 2124 (1976). * (44) V. Ambegaokar and N. D. Mermin, “Thermal anomalies of He3: pairing in a magnetic field”, Phys. Rev. Lett. 30, 81 (1973). * (45) K. Machida and T. Ohmi, “Phenomenological theory of ferromagnetic superconductivity”, Phys. Rev. Lett. 86, 850 (2001). * (46) Masa-aki Ozaki, Kazushige Machida, and Tetsuo Ohmi, “On p-Wave Pairing Superconductivity under Cubic Symmetry”, Prog. Theor. Phys. 74, 221 (1985). * (47) Masa-aki Ozaki, Kazushige Machida, and Tetsuo Ohmi, “On p-Wave Pairing Superconductivity under Hexagonal and Tetragonal Symmetries”, Prog. Theor. Phys. 75, 442 (1986). * (48) T. Ohmi and K. Machida, “Non-unitary superconducting state in UPt3”, Phys. Rev. Lett. 71, 625 (1993). * (49) K. Machida and T. Ohmi, “Identification of nonunitary triplet pairing in a heavy Fermion superconductor UPt3”, J. Phys. Soc. Jpn. 67, 1122 (1998). * (50) K. Machida, “Theory of Spin-polarized Superconductors –An Analogue of Superfluid 3He A-phase”, J. Phys. Soc. Jpn. 89, 033702 (2020). * (51) K. Machida, “Notes on Multiple Superconducting Phases in UTe2 –Third Transition–”, J. Phys. Soc. Jpn. 89, 0655001 (2020). * (52) J. F. Annett, “Symmetry of the order parameter for high-temperature superconductivity”, Adv. Phys. 39, 83 (1990). * (53) K. Machida, T. Ohmi, and M. Ozaki, “Anisotropy of Upper Critical Fields for d- and p-Wave Pairing Superconductivity”, J. Phys. Soc. Jpn. 54, 1552 (1985). * (54) K. Machida, M. Ozaki, and T. Ohmi, “Unconventional Superconducting Class in a Heavy Fermion System UPt3”, J. Phys. Soc. Jpn. 59, 1397 (1990). * (55) K. Machida, T. Fujita, and T. Ohmi, “Vortex Structures in an Anisotropic Pairing Superconducting State with Odd-Parity”, J. Phys. Soc. Jpn. 62, 680 (1993). * (56) K. Machida, T. Nishira, and T. Ohmi, “Orbital Symmetry of a Triplet Pairing in a Heavy Fermion Superconductor UPt3”, J. Phys. Soc. Jpn. 68, 3364 (1999). * (57) W. Knafo, T. D. Matsuda, D. Aoki, F. Hardy, G. W. Scheerer, G. Ballon, M. Nardone, A. Zitouni, C. Meingast, and J. Flouquet, “High-field moment polarization in the ferromagnetic superconductor UCoGe”, Phys. Rev. B 86, 184416 (2012). * (58) B. Wu, G. Bastien, M. Taupin, C. Paulsen, L. Howard, D. Aoki, and J.-P. Brison, “Pairing mechanism in the ferromagnetic superconductor UCoGe”, Nature Commun. 8, 14480 (2017). * (59) F. Hardy, D. Aoki, C. Meingast, P. Schweiss, P. Burger, H. v. Löhneysen, and J. Flouquet, “Transverse and longitudinal magnetic-field responses in the Ising ferromagnets URhGe, UCoGe, and UGe2”, Phys. Rev. B 83, 195107 (2011). * (60) V. P. Mineev, “Reentrant superconductivity in URhGe”, Phys. Rev. B 91, 014506 (2015). * (61) E. P. Wohlfarth and P. Rhodes, “Collective electron metamagnetism”, Phil. Mag. 7, 1817 (1962). * (62) M. Shimizu, “Itinerant electron magnetism”, J. Phys. (Paris) 43, 155 (1982). * (63) H. Yamada, “Metamagnetic transition and susceptibility maximum in an itinerant-electron system”, Phys. Rev. B 47, 11211 (1993). * (64) Shota Nakamura, Toshiro Sakakibara, Yusei Shimizu, Shunichiro Kittaka, Yohei Kono, Yoshinori Haga, Jir̆í Pospís̆il, and Etsuji Yamamoto, “Wing structure in the phase diagram of the Ising ferromagnet URhGe close to its tricritical point investigated by angle-resolved magnetization measurements”, Phys. Rev. B 96, 094411 (2017). * (65) D. Aoki, private communication. * (66) F. Lévy, I. Sheikin, B. Greinier, C. Marcenat, and A. D. Huxley, “Coexistence and interplay of superconductivity and ferromagnetism in URhGe”, J. Phys. Condensed Matter 21, 164211 (2009). * (67) A. Miyake, private communication. * (68) B. Wu, D. Aoki, and J.-P. Brison, “Vortex liquid phase in the $p$-wave ferromagnetic superconductor UCoGe”, Phys. Rev. B 98, 024517 (2018). * (69) Atsushi Miyake, Dai Aoki, and Jacques Flouquet, “Pressure Evolution of the Ferromagnetic and Field Re-entrant Superconductivity in URhGe”, J. Phys. Soc. Jpn. 78, 063703 (2009). * (70) Daniel Braithwaite, Dai Aoki, Jean-Pascal Brison, Jacques Flouquet, Georg Knebel, Ai Nakamura, and Alexandre Pourret, “Dimensionality Driven Enhancement of Ferromagnetic Superconductivity in URhGe”, Phys. Rev. Lett. 120, 037001 (2018). * (71) G. Knebel, private communication. * (72) S. K. Yip, T. Li, and P. Kumar, “Thermodynamic considerations and the phase diagram of superconducting UPt3”, Phys. Rev. B 43, 2742 (1991). * (73) M. Taupin, L. Howard, D. Aoki, and J.-P. Brison, “Superconducting gap of UGeCo probed by thermal transport”, Phys. Rev. B 90, 180501(R) (2014). * (74) T. Mizushima, Y. Tsutsumi, T. Kawakami, M. Sato, M. Ichioka, and K. Machida, “Symmetry-protected topological superfluids and superconductors –From the basics to 3He–”, J. Phys. Soc. Jpn. 85, 022001 (2016). * (75) T. Mizushima, Y. Tsutsumi, M. Sato, and K. Machida, “Symmetry protected topological superfluid He3-B”, J. Phys. Condensed Matter, 27, 113203 (2015). * (76) G. Aeppli, E. Bucher, C. Broholm, J. K. Kjems, J. Baumann, and J. Hufnagl, “Magnetic order and fluctuations in superconducting UPt3”, Phys. Rev. Lett. 60, 615 (1988). * (77) T. Trappmann, H. v. Löhneysen, and L. Taillefer, “Pressure dependence of the superconducting phases in UPt3”, Phys. Rev. B 43, 13714(R) (1991). * (78) K. Machida and M. Ozaki, “Superconducting double transition in a heavy-fermion material UPt3”, Phys. Rev. Lett. 66, 3293 (1991). * (79) J. A. Sauls, “The order parameter for the superconducting phases of UPt3”, Adv. Phys. 43, 113 (1993). * (80) D. Fay and J. Appel, “Coexistence of p-state superconductivity and itinerant ferromagnetism” Phys. Rev. B 22, 3173 (1980). * (81) P. Miranović, N. Nakai, M. Ichioka, and K. Machida, “Orientational field dependence of low-lying excitations in the mixed state of unconventional superconductors”, Phys. Rev. B 68, 052501 (2003). * (82) Y. Tsutsumi, T. Mizushima, M. Ichioka, and K. Machida, “Majorana Edge Modes of Superfluid 3He A-Phase in a Slab”, J. Phys. Soc. Jpn. 79, 113601 (2010). * (83) Y. Tsutsumi, M. Ichioka, and K. Machida, “Majorana surface states of superfluid 3He A and B phases in a slab”, Phys. Rev. B 83, 094510 (2011).
# A note on Minkowski formula of conformal Killing-Yano 2-form Xiaoxiang Chai Korea Institute for Advanced Study, Seoul 02455, South Korea <EMAIL_ADDRESS> ###### Abstract. We study the Minkowski formula of conformal Killing-Yano two-forms in a spacetime of constant curvature. We establish the spacetime Alexandrov theorem with a free boundary. ## 1\. Introduction The Minkowski formula states that for a smooth closed hypersurface $X:\Sigma\to\mathbb{R}^{n}$, (1) $(n-k)\int_{\Sigma}\sigma_{k-1}\mathrm{d}\mu=k\int_{\Sigma}\sigma_{k}\langle X,\nu\rangle.$ Here $\sigma_{k}$ is the $k$-th elementary symmetric functions of principal curvatures of $\Sigma$. It has found itself many applications in Riemannian geometry for example a proof of the celebrated Alexandrov theorem which says that an closed embedded hypersurface of constant mean curvature must be an sphere. The same ideas of proof lead to a free boundary generalization due to Wang-Xia [WX19] establishing the rigidity of spherical caps in balls of space forms. Both closed and the free boundary settings made use of a specially chosen conformal Killing vector field. Tachibana introduced the conformal Killing-Yano two-form as a generalization of conformal Killing vector field. ###### Definition 1 (Tachibana [Tac69]). A two-form $Q$ on an $(n+1)$-dimensional spacetime is called a conformal Killing-Yano 2-form if for every vector field $X,Y$ and $Z$ the following identity holds (2) $(\nabla_{X}Q)(Y,Z)+(\nabla_{Y}Q)(X,Z)=[2\langle X,Y\rangle\langle\xi,Z\rangle-\langle X,Z\rangle\langle\xi,Y\rangle-\langle Y,Z\rangle\langle\xi,X\rangle]$ where $\xi=\frac{1}{n}\operatorname{div}Q$. We call $\xi$ the associated 1-form of $Q$. In physics literature, these two forms are usually termed as hidden symmetry and can give information about the spacetime. See for example [JŁ06] and the references therein. Besides its physical significance, mathematically the conformal Killing-Yano two-forms are also interesting. In particular, they also allow a Minkowski type formula. Chen, Wang, Yau [CWY19] expressed quasilocal masses using this Minkowski formula. The authors of [WWZ17] established a spacetime version of the Alexandrov theorem for codimension two spacelike hypersurfaces via the Minkowski formula. In this work, we are going to extend results in [WWZ17] where they used only conformal Killing-Yano two-forms $r\mathrm{d}r\wedge\mathrm{d}t$. First we state the spacetime CMC condition with free boundary. ###### Definition 2. We say that $\Sigma^{2}$ in a spacetime $\mathbb{R}^{3,1}$ is CMC with free boundary if $\Sigma$ admits a null normal vector field $L$ with $\langle\vec{H},L\rangle$ is constant, $(DL)^{\bot}=0$ and $\Sigma$ meets the de Sitter sphere $\mathbf{S}^{2,1}$ orthogonally. Of course, one can allow arbitrary spacetime and boundary in the above definition. One interesting problem related to such surfaces is the uniqueness problem of a topological disk (cf. [FS15]). Without the free boundary condition, similar questions can be asked for two-spheres in $3+1$ dimensional de Sitter sphere (cf. [Che69]) . One can also ask whether a spacelike graph over $\mathbb{R}^{2}$ in $\mathbb{R}^{3,1}$ with $\langle\vec{H},L\rangle=0$ and $(DL)^{\bot}=0$ is linear which is analogous to the Bernstein problem for minimal graphs. We generalize the spacetime Alexandrov theorem to the free boundary settings via establishing a spacetime Heintz-Karcher inequality Theorem 4. We state here the theorem in the Minkowski spacetime. ###### Theorem 1. Let $\Sigma$ be a codimension two, future incoming null embedded submanifold in the $(3+1)$-dimensional Minkowski spacetime with free boundary on the de Sitter sphere $\mathbf{S}^{2,1}$. If $\Sigma$ lies in a half spacetime, and there exists a null vector field $\underline{L}$ such that along $\Sigma$ that $\langle\vec{H},\underline{L}\rangle$ is a positive constant and $(D\underline{L})^{\bot}=0$. Then $\Sigma$ lies in a shear free null hypersurface. The theorem is a direct corrollary from Theorem 4 and similar proofs as in [WWZ17, Theorem 3.14]. The article is organized as follows: In Section 2, we collect basics of spacetime of constant curvature and the conformal Killing-Yano two-forms they admit. In Section 3, we prove a spacetime Heintz-Karcher inequality with a free boundary leading to a free boundary, spacetime Alexandrov theorem. We mention briefly the generalization to higher order curvatures. Acknowledgements I would like to thank Xia Chao, Wang Ye-kai for their interest and advice in an earlier version of this work. I would also like to acknowledge the support of Korea Institute for Advanced Study under the research number MG074401. ## 2\. Conformal Killing-Yano 2-form on spacetime of constant curvature A spacetime of dimension $3+1$ can only admit 20 conformal Killing-Yano two- forms. Actually, if a spacetime admits all twenty of them, then the spacetime has to be a spacetime of constant curvature. Note that for similar statements are also true for conformal Killing vector fields. In Minkowski, de Sitter and anti-de Sitter spacetime, these two forms are found explicitly. See the works by Jezierski and Lukasik [JŁ06, Jez08]. Now we collect some basics of these spacetimes and the conformal Killing-Yano two forms that live on them. ### 2.1. Minkowski spacetime Let $(x^{0},x^{1},x^{2},x^{3})$ be the standard coordinates of the Minkowski space $\mathbb{R}^{3,1}$, define (3) $\displaystyle\mathcal{D}$ $\displaystyle=-x^{0}\mathrm{d}x^{0}+x^{1}\mathrm{d}x^{1}+x^{2}\mathrm{d}x^{2}+x^{3}\mathrm{d}x^{3},$ (4) $\displaystyle\mathcal{T}_{0}$ $\displaystyle=-\mathrm{d}x^{0},$ (5) $\displaystyle\mathcal{T}_{i}$ $\displaystyle=\mathrm{d}x^{i},$ (6) $\displaystyle\mathcal{L}_{0i}$ $\displaystyle=-x^{0}\mathrm{d}x^{i}+x^{i}\mathrm{d}x^{0}.$ The conformal Killing-Yano 2-forms on Minkowski spacetime $\mathbb{R}^{3,1}$ are (7) $\mathcal{T}_{\mu}\wedge\mathcal{T}_{\nu},\mathcal{D}\wedge\mathcal{T}_{\mu},\ast(\mathcal{D}\wedge\mathcal{T}_{\mu})\text{ and }\mathcal{D}\wedge\mathcal{L}_{\mu\nu}+\frac{1}{2}\langle\mathcal{D},\mathcal{D}\rangle\mathcal{T}_{\mu}\wedge\mathcal{T}_{\nu},$ where $\ast$ is the Hodge star operator and $\mu,\nu$ range from 0 to 3. See [JŁ06] for a calculation. Note that all are still conformal Killing-Yano 2-forms on $\mathbb{R}^{n,1}$ except $\ast(\mathcal{D}\wedge\mathcal{T}_{\mu})$. We remark that the last one in (7) can be used to prove formulas relating the center of mass (See [MT16]) and a Brown-York type quasi-local quantity by following similar procedures in [CWY19]. ### 2.2. Anti-de Sitter spacetime We recall some basics of four-dimensional anti-de Sitter spacetime. The anti- de Sitter spacetime ad$\mathbf{S}^{3,1}$ is defined to be the set in $\mathbb{R}^{3,2}$ (8) $-(y^{0})^{2}+(y^{1})^{2}+(y^{2})^{2}+(y^{3})^{2}-(y^{4})^{2}=-1$ with metric induced from $\eta=-(\mathrm{d}y^{0})^{2}+(\mathrm{d}y^{1})^{2}+(\mathrm{d}y^{2})^{2}+(\mathrm{d}y^{3})^{2}-(\mathrm{d}y^{4})^{2}$. We will use coordinates of the Poincaré ball model by setting $r=\sqrt{(x^{1})^{2}+(x^{2})^{2}+(x^{3})^{2}}$, $y^{0}=\tfrac{1+r^{2}}{1-r^{2}}\cos t$, $y^{4}=\tfrac{1+r^{2}}{1-r^{2}}\sin t$ and $y^{i}=\tfrac{2x^{i}}{1-r^{2}}$. The metric of ad$\mathbf{S}^{3,1}$ is then $-(\tfrac{1+r^{2}}{1-r^{2}})^{2}\mathrm{d}t^{2}+\tfrac{4\sum_{i}(\mathrm{d}x^{i})^{2}}{(1-r^{2})^{2}}$. It is shown in [Jez08] that the conformal Killing-Yano 2-forms in four- dimensional anti-de Sitter spacetime are (9) $\mathrm{d}y^{0}\wedge\mathrm{d}y^{i},\mathrm{d}y^{0}\wedge\mathrm{d}y^{4},\mathrm{d}y^{i}\wedge\mathrm{d}y^{4},\mathrm{d}y^{i}\wedge\mathrm{d}y^{j}$ and their Hodge duals with respect to the anti-de Sitter metric. We fix the frame $\theta^{i}=\tfrac{2}{1-r^{2}}\mathrm{d}x^{i}$ and $\theta^{0}=\tfrac{1+r^{2}}{1-r^{2}}\mathrm{d}t$. Let $\omega=\tfrac{2}{1-r^{2}}\mathrm{d}r$, then the length of $\omega$ is one. We have (10) $\mathrm{d}y^{i}=\theta^{i}+y^{i}r\omega,\mathrm{d}y^{4}=\cos t\theta^{0}+\tfrac{2r}{1-r^{2}}\omega\sin t.$ Note that $y^{4}$ and $y^{i}$ are static potentials, that is $\nabla_{i}\mathrm{d}y^{\mu}=y^{\mu}\theta^{i}$ and $\nabla_{0}\mathrm{d}y^{\mu}=-y^{\mu}\theta^{0}$ for each $\mu=0,1,\ldots,4$. Here $\nabla_{\mu}$ denotes covariant derivative with respect to the vector field $(\theta^{\mu})^{\sharp}$. Then it is easy to obtain that (11) $\operatorname{div}(\mathrm{d}y^{i}\wedge\mathrm{d}y^{4})=3(y^{i}\mathrm{d}y^{4}-y^{4}\mathrm{d}y^{i}).$ Note that $y^{i}\mathrm{d}y^{4}-y^{4}\mathrm{d}y^{i}$ is a Killing 1-form. Using the properties of Hodge operators, we find that $\operatorname{div}(\ast(\mathrm{d}y^{2}\wedge\mathrm{d}y^{3}))$ vanishes. We remark that the 2-form $\mathrm{d}y^{i}\wedge\mathrm{d}y^{4}$ can be used similarly as in [CWY15] to recover a formula relating the integrals of Ricci tensor and Brown-York type mass vector of an asymptotically hyperbolic manifold. These formulas are overlooked by the authors of [CWY15]. The original proof is due to [MTX17]. ### 2.3. de Sitter spacetime The case with de Sitter spacetime is similar to the anti-de Sitter case (See [Jez08]). We consider here the $3+1$ dimensional case i.e. $\mathbf{S}^{3,1}$. The de Sitter spacetime is the subset (12) $y_{0}^{2}+y_{1}^{2}+y_{2}^{2}+y_{3}^{2}-y_{4}^{2}=1$ in $\mathbb{R}^{4,1}$ with the metric inherited from the standard Lorentz metric of $\mathbb{R}^{4,1}$. We use the coordinate change (13) $\displaystyle y^{0}$ $\displaystyle=\tfrac{1-r^{2}}{1+r^{2}}\cosh t,$ (14) $\displaystyle y^{i}$ $\displaystyle=\tfrac{2x^{i}}{1+r^{2}}$ (15) $\displaystyle y^{4}$ $\displaystyle=\tfrac{1-r^{2}}{1+r^{2}}\sinh t,$ where $r=\sqrt{\sum_{i=1}^{3}(x^{i})^{2}}<1$. Now the metric of the de Sitter spacetime $\mathbf{S}^{3,1}$ takes the form (16) $\eta=-(\tfrac{1-r^{2}}{1+r^{2}})^{2}\mathrm{d}t^{2}+\tfrac{4}{(1+r^{2})^{2}}[(\mathrm{d}x^{1})^{2}+(\mathrm{d}x^{2})^{2}+(\mathrm{d}x^{3})^{2}].$ It is shown in [Jez08] that the conformal Killing-Yano 2-forms in four- dimensional de Sitter spacetime are (17) $\mathrm{d}y^{0}\wedge\mathrm{d}y^{i},\mathrm{d}y^{0}\wedge\mathrm{d}y^{4},\mathrm{d}y^{i}\wedge\mathrm{d}y^{4},\mathrm{d}y^{i}\wedge\mathrm{d}y^{j}$ and their Hodge duals with respect to the de Sitter metric. We fix the frame $\theta^{i}=\tfrac{2}{1+r^{2}}\mathrm{d}x^{i}$ and $\theta^{0}=\tfrac{1-r^{2}}{1+r^{2}}\mathrm{d}t$. Let $\omega=\tfrac{2}{1+r^{2}}\mathrm{d}r$, then the length of $\omega$ is one. We have (18) $\mathrm{d}y^{i}=\theta^{i}-y^{i}r\omega,\mathrm{d}y^{4}=\cosh t\theta^{0}-\tfrac{2r}{1+r^{2}}\omega\sinh t.$ Note that $y^{4}$ and $y^{i}$ are static potentials, that is $\nabla_{i}\mathrm{d}y^{\mu}=-y^{\mu}\theta^{i}$ and $\nabla_{0}\mathrm{d}y^{\mu}=y^{\mu}\theta^{0}$ for each $\mu=0,1,\ldots,4$. Here $\nabla_{\mu}$ denotes covariant derivative with respect to the vector field $(\theta^{\mu})^{\sharp}$. Then it is easy to obtain that (19) $\operatorname{div}(\mathrm{d}y^{i}\wedge\mathrm{d}y^{4})=-3(y^{i}\mathrm{d}y^{4}-y^{4}\mathrm{d}y^{i}).$ Note that $y^{i}\mathrm{d}y^{4}-y^{4}\mathrm{d}y^{i}$ is a Killing 1-form. We found also easily that $\operatorname{div}(\ast(\mathrm{d}y^{2}\wedge\mathrm{d}y^{3}))$ vanishes. ## 3\. Spacetime Alexandrov theorem with free boundary We start by proving a Minkowski formula for a codimension two spacelike hypersurface in $\mathbb{R}^{3,1}$ with boundary meeting orthogonally with the de Sitter sphere. The result is related to mean curvature only, the generalization to higher order curvatures is quite straightforward. The Minkowski spacetime is used as a prototype. First, we fix a conformal Killing-Yano 2-form in Minkowski spacetime $\mathbb{R}^{3,1}$ (20) $Q=\mathcal{D}\wedge\mathcal{L}_{0i}+\frac{1}{2}[1+\langle\mathcal{D},\mathcal{D}\rangle]e^{0}\wedge e^{i}.$ The associated 1-form is $\xi:=\tfrac{1}{n}\operatorname{div}Q=\mathcal{L}_{0i}$ since ###### Lemma 1. The divergence of the 2-form $Q=\mathcal{D}\wedge\mathcal{L}_{0i}+\tfrac{1}{2}\langle\mathcal{D},\mathcal{D}\rangle\mathrm{d}x^{0}\wedge\mathrm{d}x^{i}$ is given by $\operatorname{div}Q=3\mathcal{L}_{0i}$. Define the Minkowski unit ball (21) $\mathcal{B}^{3,1}=\\{x\in\mathbb{R}^{3,1}:\langle x,x\rangle\leqslant 1\\}.$ The boundary of $\mathcal{B}^{3,1}$ is the de Sitter sphere $\mathbf{S}^{2,1}$. It is easy to check that $\mathcal{D}^{\sharp}\lrcorner Q$ is zero along $\partial\mathcal{B}^{3,1}$, so $Q$ has no components normal to $\mathbf{S}^{2,1}$. ###### Theorem 2. Let $\Sigma$ be an immersed oriented spacelike codimension two submanifolds of the Minkowski spacetime $\mathbb{R}^{3,1}$, $\partial\Sigma$ lies in the de Sitter sphere $\mathbf{S}^{2,1}$ and $\Sigma$ meets $\mathbf{S}^{2,1}$ orthogonally. For any null vector field $\underline{L}$ of $\Sigma$, we have (22) $\int_{\Sigma}[(n-1)\langle\xi,\underline{L}\rangle+Q(\vec{H},\underline{L})+Q(\partial_{a},(D^{a}\underline{L})^{\bot})]\mathrm{d}\mu=0.$ ###### Proof. Define $\mathcal{Q}=Q(\partial_{a},\underline{L})\mathrm{d}u^{a}$ on $\Sigma$ and the proof is almost the same as Theorem 2.2 of [WWZ17]. We include their proof for convenience. Let $\underline{\chi}=\langle D_{a}\underline{L},\partial_{b}\rangle$. Consider the 1-form $\mathcal{Q}=Q(\partial_{a},\underline{L})\mathrm{d}u^{a}$, we have (23) $\displaystyle\operatorname{div}\mathcal{Q}$ $\displaystyle=\nabla_{a}\mathcal{Q}^{a}-Q(\nabla_{a}\partial_{a},\underline{L})$ (24) $\displaystyle=(D^{a}Q)(\partial_{a},\underline{L})+Q(\vec{H},\underline{L})+Q(\partial_{a},D^{a}\underline{L})$ (25) $\displaystyle=(n-1)\langle\xi,\underline{L}\rangle+Q(\vec{H},\underline{L})+\underline{\chi}_{ab}Q^{ab}+Q(\partial_{a},(D^{a}\underline{L})^{\bot})$ (26) $\displaystyle=(n-1)\langle\xi,\underline{L}\rangle+Q(\vec{H},\underline{L})+Q(\partial_{a},(D^{a}\underline{L})^{\bot}).$ Integration by parts and noting that $Q$ has no components normal to the de Sitter sphere. ∎ ### 3.1. A monotonicity formula Let $\Sigma$ be a spacelike submanifold of codimension two in a spacetime $(\mathcal{S}^{3,1},g)$ which admits a Killing-Yano two form $Q$. Here, $\mathcal{S}$ is either one of the four dimensional Minkowski, de Sitter and anti de Sitter spacetime. We require that $Q$ has no normal component normal to a support hypersurface $S$. Suppose that $\langle\vec{H},\underline{L}\rangle\neq 0$, we define the following functional (27) $\mathcal{F}(\Sigma,[\underline{L}])=(n-1)\int_{\Sigma}\frac{\langle\xi,\underline{L}\rangle}{\langle\vec{H},\underline{L}\rangle}\mathrm{d}\mu-\frac{1}{2}\int_{\Sigma}Q(L,\underline{L})\mathrm{d}\mu.$ Note $\mathcal{F}$ is invariant under the change $L\to aL$ and $\underline{L}\to\frac{1}{a}\underline{L}$. Let $\chi$ and $\underline{\chi}$ be respectively the second fundamental form with respect to $L$, $\underline{L}$; let $\underline{C}_{0}$ denote the future incoming null hypersurface of $\Sigma$. $\underline{C}_{0}$ is obtained by taking the collection of all null geodesics emanating from $\Sigma$ with initial velocity $\underline{L}$. We then extend it to a future directed null vector field along $\underline{C}_{0}$. Consider the evolution of $\Sigma$ along $\underline{C}_{0}$ by a family of immersions $F:\Sigma\times[0,T)\to\underline{C}_{0}$ satisfying (28) $\left\\{\begin{array}[]{l}\frac{\partial F}{\partial s}(x,s)=\varphi(x,s)\underline{L},\\\ F(x,0)=F_{0}(x),\\\ \Sigma\bot S\end{array}\right.$ for some positive function $\varphi(x,s)$. We have the following monotonicity property of the flow $\varphi$. ###### Theorem 3. Suppose that $\langle\vec{H},\underline{L}\rangle>0$ for some null vector field $\underline{L}$. Then $\mathcal{F}(F(\Sigma,s),[\underline{L}])$ is monotone decreasing under the flow. ###### Proof. See Theorem 3.2 of [WWZ17]. We only have to use the extra fact that $Q$ has no components normal to the de Sitter space as in the proof of Theorem 2. ∎ The monotonicity property leads to a spacetime Heintz-Karcher inequality. More, specifically, if under certain flow $\varphi$, the surface $\Sigma$ with $\langle\vec{H},\underline{L}\rangle>0$ flows into a submanifold of the time slice $\\{x^{0}=0\\}$ at $s=T$ and for $\Sigma$ (29) $\mathcal{F}(\Sigma,[L])\geqslant 0$ holds provided $\varphi(\Sigma,T)\subset\\{x^{0}=0\\}$ and $\mathcal{F}(\varphi(\Sigma,T),[L])\geqslant 0$. ###### Lemma 2. For any $\Sigma\subset\\{x^{0}=0\\}$, $\mathcal{F}(\Sigma,[L])\geqslant 0$ reduces to (30) $(n-1)\int_{\Sigma}\tfrac{x^{i}}{H}\mathrm{d}\mu\geqslant\int_{\Sigma}\langle X_{\partial_{i}},\nu\rangle\mathrm{d}\mu.$ ###### Proof. We have $\underline{L}=\partial_{t}-e_{n}$ where $e_{n}$ is a unit normal. So $\langle\vec{H},\underline{L}\rangle=H$ where $H$ is the mean curvature of $\Sigma$ in $\mathbf{B}^{n}$. We have that $\xi=\mathcal{L}_{0i}=x^{i}\mathrm{d}x^{0}$, so $\langle\xi,\underline{L}\rangle=x^{i}$. Also, (31) $Q(L,\underline{L})=Q(\partial_{t}+\nu,\partial_{t}-\nu)=2\langle X_{\partial_{i}},\nu\rangle,$ where $X_{a}=\langle X,a\rangle X+\tfrac{1}{2}(|X|^{2}+1)a$ where $a=a^{i}\partial_{i}$ is a constant vector in $\mathbb{R}^{n}$. It easily leads to (30). ∎ Note that this is precisely an inequality proven already by Wang-Xia [WX19, (5.5)] with the assumption that $\Sigma$ has positive mean curvature and lies in a half ball. Combining with their result, we have ###### Theorem 4 (spacetime Heintz-Karcher inequality). If there exists a flow $\varphi$ of a hypersurface $\Sigma$ with $\langle\vec{H},\underline{L}\rangle>0$ for some null vector field $\underline{L}$ and a free boundary on $\mathbf{S}^{2,1}$ which flows $\Sigma$ into the half unit ball of the slice $\\{x^{0}=0\\}$, then we have the inequality (32) $\int_{\Sigma}\frac{\langle\xi,\underline{L}\rangle}{\langle\vec{H},\underline{L}\rangle}\mathrm{d}\mu\geqslant\frac{1}{2(n-1)}\int_{\Sigma}Q(L,\underline{L})\mathrm{d}\mu.$ Equality occurs if and only if $\Sigma$ lies in a shear free null hypersurface with free boundary on $\mathbf{S}^{n-1,1}$. ###### Proof. Let $\Sigma_{t}=\varphi_{t}(\Sigma)$, then for each $t>0$, the equality holds. Suppose that $\Sigma_{T}\subset\\{x^{0}=0\\}$ for some $T>0$. So $\Sigma_{T}$ has to be a spherical cap orthogonal to the unit sphere in $\mathbb{R}^{n}$ according to [WX19]. In particular, under the flow $\varphi$, $\Sigma_{t}$ foliates a shear free null hypersurface $S$ with free boundary. ∎ ### 3.2. Anti-de Sitter case Theorems 2, 4 and 1 work as well in the case with $\partial\Sigma=\emptyset$. The same proof also adapts in the anti-de Sitter and de Sitter settings. We use the notations in Section 2.2. For simplicity, we set $i$ to be 1, we use the 2-forms $\mathrm{d}y^{1}\wedge\mathrm{d}y^{4}$ and $\ast(\mathrm{d}y^{2}\wedge\mathrm{d}y^{3})$ only. Note that the Hodge star operator commutes with the covariant derivative. Using this, we see easily that $\operatorname{div}(\ast(\mathrm{d}y^{2}\wedge\mathrm{d}y^{3}))$ vanishes. We use the 2-form $Q=\mathrm{d}y^{1}\wedge\mathrm{d}y^{4}+l\ast(\mathrm{d}y^{2}\wedge\mathrm{d}y^{3})$ where $l>0$ is a positive constant. We define the surface $\mathcal{B}^{3,1}$ to be the surface with distance less than $d$ from the point $t=0$, $r=0$ where $\cosh d=l$. If $Y_{1},Y_{2}\in\operatorname{ad}\mathbf{S}^{3,1}$ (using the embedding into $\mathbb{R}^{3,2}$) are two points which can be connected via a spacelike geodesic, then the distance from $Y_{1}$ to $Y_{2}$ is $\cosh d=-\eta(Y_{1},Y_{2})>0$. The boundary $S=\partial\mathcal{B}^{3,1}$ is a timelike hypersurface of dimension three of constant distance from the point $t=0$, $r=0$ and it is umbilical hence null geodesics intrinsic to $S$ are also null geodesic in ad$\mathbf{S}^{3,1}$. It is the analog of de Sitter sphere which is of constant distance to the origin in Minkowski spacetime. It is a tedious task to check that $Q$ has no component normal to $S$. We state here the spacetime Heintz-Karcher inequality and leave the spacetime Alexandrov theorem to the reader. ###### Theorem 5. (spacetime Heintz-Karcher inequality in $\mathcal{B}^{3,1}$) If there exists a flow $\varphi$ of a hypersurface $\Sigma$ with $\langle\vec{H},\underline{L}\rangle>0$ for some null vector field $\underline{L}$ and a free boundary on $S$ which flows $\Sigma$ into the half geodesic ball of the slice $\\{t=0\\}$, then we have the inequality (33) $\int_{\Sigma}\frac{\langle\xi,\underline{L}\rangle}{\langle\vec{H},\underline{L}\rangle}\mathrm{d}\mu\geqslant\frac{n}{2(n-1)}\int_{\Sigma}Q(L,\underline{L})\mathrm{d}\mu.$ Equality occurs if and only if $\Sigma$ lies in a shear free null hypersurface. ###### Proof. The proof is the same with Theorem 4. We only have to verify when $t=0$ the inequality holds. Let $\nu$ be the unit normal of $\Sigma$ in the $\\{t=0\\}$ slice. Indeed, when $t=0$, $\xi=y^{1}\mathrm{d}y^{4}$ and $\underline{L}=e_{0}-\nu=\tfrac{1-r^{2}}{1+r^{2}}\partial_{t}-\nu,$ so $\langle\xi,\underline{L}\rangle=y^{1}=\tfrac{2x^{1}}{1-r^{2}}.$ We turn to $Q(L,\underline{L})$. We have $(\mathrm{d}y^{1}\wedge\mathrm{d}y^{4})(L,\underline{L})=2\mathrm{d}y^{1}(\nu)$ and $(\mathrm{d}y^{1})^{\sharp}=\tfrac{1}{2}\partial_{1}+(x^{1}x^{j}\partial_{j}-\tfrac{1}{2}r^{2}\partial_{1}).$ And $\ast(\mathrm{d}y^{2}\wedge\mathrm{d}y^{3})=-\theta^{1}\wedge\theta^{0}+y^{2}(x^{1}\theta^{2}-x^{2}\theta^{1})\wedge\theta^{0}+y^{3}(x^{1}\theta^{3}-x^{3}\theta^{1})\wedge\theta^{0},$ so the 1-form $(\ast(\mathrm{d}y^{2}\wedge\mathrm{d}y^{3}))(\cdot,e_{0})$ is dual to $-\tfrac{1}{2}\partial_{1}+(x^{1}x^{j}\partial_{j}-\tfrac{1}{2}r^{2}\partial_{1})$. As usual, $Q(L,\underline{L})=2Q(\nu,e_{0})$. Thus, (34) $Q(L,\underline{L})=2\langle X_{\partial_{1}},\nu\rangle,$ where $X_{a}=(1+l)\left[x^{k}a_{k}x^{j}\partial_{j}-\tfrac{1}{2}(r^{2}+\tfrac{l-1}{l+1})a\right]$ with $a=a^{j}\partial_{j}$ being a constant vector in $\mathbb{R}^{n}$. Letting $l=\tfrac{1+R_{\mathbb{R}}^{2}}{1-R_{\mathbb{R}}^{2}}$, (33) reduces to also [WX19]. ∎ ###### Remark 1. It is easy to check that the higher dimensional analog of $\ast(\mathrm{d}y^{2}\wedge\mathrm{d}y^{3})$ in the $n$-dimensional anti-de Sitter spacetime $\operatorname{ad}\mathbf{S}^{n}=\\{-(y^{0})^{2}+(y^{1})^{2}+\cdots+(y^{n})^{2}-(y^{n+1})^{2}=1\\}$ is (35) $-e^{1}\wedge e^{0}+\sum_{i\neq 1}^{n}y^{i}(x^{1}e^{i}-x^{i}e^{1})\wedge e^{0}.$ ### 3.3. de Sitter case We calculate below the quantities needed for a theorem parallel to Theorem 33. We follow similar notations and omit the the statements or details. Generalizing to higher dimension is also straightforward. The conformal Killing-Yano 2-form is $Q=\mathrm{d}y^{4}\wedge\mathrm{d}y^{1}+l\ast(\mathrm{d}y^{3}\wedge\mathrm{d}y^{2})$ and its associated 1-form $\xi=\operatorname{div}Q=3y^{1}\mathrm{d}y^{4}-y^{4}\mathrm{d}y^{1}.$ Notice the order of the superscripts. Within the slice $\\{t=0\\}$, we have that $\xi=y^{1}\mathrm{d}y^{4}$ and $\underline{L}=e_{0}-\nu=\tfrac{1+r^{2}}{1-r^{2}}\partial_{t}-\nu$ and $\langle\xi,\underline{L}\rangle=\tfrac{2x^{1}}{1+r^{2}}.$ We turn to $Q(L,\underline{L})$. We have $(\mathrm{d}y^{4}\wedge\mathrm{d}y^{1})(L,\underline{L})=-2\mathrm{d}y^{1}(\nu).$ Note that $A:=-(\mathrm{d}y^{1})^{\sharp}=-\tfrac{1}{2}\partial_{1}-(\tfrac{1}{2}r^{2}\partial_{1}-x^{1}x^{j}\partial_{j})$ and $\ast(\mathrm{d}y^{3}\wedge\mathrm{d}y^{2})=\theta^{1}\wedge\theta^{0}+y^{2}(x^{1}\theta^{2}-x^{2}\theta^{1})\wedge\theta^{0}+y^{3}(x^{1}\theta^{3}-x^{3}\theta^{1})\wedge\theta^{0},$ so the 1-form $\ast(\mathrm{d}y^{3}\wedge\mathrm{d}y^{2})(\cdot,e_{0})$ is dual to $B:=\tfrac{1}{2}\partial_{1}-\tfrac{1}{2}r^{2}\partial_{1}+x^{1}x^{j}\partial_{j}.$ $A+lB$ is then (36) $X_{\partial_{1}}:=(1+l)\left[x^{1}x^{j}\partial_{j}+\tfrac{1}{2}(\tfrac{1-l}{l+1}-r^{2})\partial_{1}\right].$ Therefore $Q(L,\underline{L})=2\langle X_{\partial_{1}},\nu\rangle$. Setting $\tfrac{1-l}{1+l}=|x|^{2}$ with $0<l<1$ recovers the form of [WX19]. We have not given the support hypersurface of the boundary yet. To this end, we fix a point $O=\\{t=0,r=0\\}$, let $S$ be the hypersurface in $\mathbf{S}^{3,1}$ be a hypersurface of constant distance $d$ from the point $O$ where $\cos d=l$. It is fairly easy to check that $Q$ has no components to the hypersurface $S$. ## References * [Alm66] F. J. Almgren, Jr. Some interior regularity theorems for minimal surfaces and an extension of Bernstein’s theorem. Ann. of Math. (2), 84:277–292, 1966\. * [Cal67] Eugenio Calabi. Minimal immersions of surfaces in Euclidean spheres. J. Differential Geometry, 1:111–125, 1967. * [Che69] Shiing-shen Chern. Simple proofs of two theorems on minimal surfaces. Enseign. Math. (2), 15:53–61, 1969. * [CWY15] Po-Ning Chen, Mu-Tao Wang, and Shing-Tung Yau. Conserved Quantities in General Relativity: From the Quasi-Local Level to Spatial Infinity. Communications in Mathematical Physics, 338(1):31–80, 2015. * [CWY19] Po-Ning Chen, Mu-Tao Wang, and Shing-Tung Yau. The Minkowski formula and the quasi-local mass. Ann. Henri Poincaré, 20(3):889–904, 2019\. * [FS15] Ailana Fraser and Richard Schoen. Uniqueness theorems for free boundary minimal disks in space forms. Int. Math. Res. Not. IMRN, (17):8268–8274, 2015\. * [JŁ06] Jacek Jezierski and Maciej Łukasik. Conformal Yano-Killing tensor for the Kerr metric and conserved quantities. Classical Quantum Gravity, 23(9):2895–2918, 2006. * [Jez08] Jacek Jezierski. Conformal Yano-Killing tensors in anti-de Sitter spacetime. Classical Quantum Gravity, 25(6):65010–17, 2008\. * [MT16] Pengzi Miao and Luen-Fai Tam. Evaluation of the ADM mass and center of mass via the Ricci tensor. Proceedings of the American Mathematical Society, 144(2):753–761, 2016. * [MTX17] Pengzi Miao, Luen-Fai Tam, and Naqing Xie. Quasi-Local Mass Integrals and the Total Mass. The Journal of Geometric Analysis, 27(2):1323–1354, 2017\. * [Tac69] Shun-ichi Tachibana. On conformal Killing tensor in a Riemannian space. Tohoku Math. J. (2), 21:56–64, 1969. * [WWZ17] Mu-Tao Wang, Ye-Kai Wang, and Xiangwen Zhang. Minkowski formulae and Alexandrov theorems in spacetime. J. Differential Geom., 105(2):249–290, 2017\. * [WX19] Guofang Wang and Chao Xia. Uniqueness of stable capillary hypersurfaces in a ball. Math. Ann., 374(3-4):1845–1882, 2019.
# Martingale convergence Theorems for Tensor Splines Markus Passenbrunner Institute of Analysis, Johannes Kepler University Linz, Austria, 4040 Linz, Altenberger Strasse 69<EMAIL_ADDRESS> ###### Abstract. In this article we prove martingale type pointwise convergence theorems pertaining to tensor product splines defined on $d$-dimensional Euclidean space ($d$ is a positive integer), where conditional expectations are replaced by their corresponding tensor spline orthoprojectors. Versions of Doob’s maximal inequality, the martingale convergence theorem and the characterization of the Radon-Nikodým property of Banach spaces $X$ in terms of pointwise $X$-valued martingale convergence are obtained in this setting. Those assertions are in full analogy to their martingale counterparts and hold independently of filtration, spline degree, and dimension $d$. ###### Key words and phrases: Tensor product spline orthoprojectors, Almost everywhere convergence, Maximal functions, Radon-Nikodým property, Martingale methods ###### 2010 Mathematics Subject Classification: 41A15, 42B25, 46B22, 42C10, 60G48 ## 1\. Introduction In this article we prove pointwise convergence theorems pertaining to tensor product splines defined on $d$-dimensional Euclidean space in the spirit of the known results for martingales. We begin by discussing the situation for martingales and, subsequently, for one-dimensional splines. For martingales, we use [6] and [2] as references. Let $(\Omega,(\mathscr{F}_{n}),\mathbb{P})$ be a filtered probability space. A sequence of integrable functions $(f_{n})_{n\geq 1}$ is a _martingale_ if $\mathbb{E}(f_{n+1}|\mathscr{F}_{n})=f_{n}$ for any $n$, where we denote by $\mathbb{E}(\cdot|\mathscr{F}_{n})$ the conditional expectation operator with respect to the $\sigma$-algebra $\mathscr{F}_{n}$. This operator is the orthoprojector onto the space of $\mathscr{F}_{n}$-measurable $L^{2}$-functions and it can be extended to act on the Lebesgue-Bochner space $L^{1}_{X}$ for any Banach space $(X,\|\cdot\|)$. Observe that if $f\in L^{1}_{X}$, the sequence $(\mathbb{E}(f|\mathscr{F}_{n}))$ is a martingale. In this case, we have that $\mathbb{E}(f|\mathscr{F}_{n})$ converges almost surely to $\mathbb{E}(f|\mathscr{F})$ with $\mathscr{F}=\sigma(\cup_{n}\mathscr{F}_{n})$. A crucial step in the proof of this convergence theorem is _Doob’s maximal inequality_ $\mathbb{P}\\{\sup_{n}\|f_{n}\|>t\\}\leq\frac{\sup_{n}\|f_{n}\|_{L^{1}_{X}}}{t},\qquad t>0,$ which states that the martingale maximal function $\sup\|f_{n}\|$ is of weak type $(1,1)$. For general scalar-valued martingales, we have the following convergence theorem: any martingale $(f_{n})$ that is bounded in $L^{1}$ has an almost sure limit function contained in $L^{1}$. This limit can be identified as the Radon-Nikodým derivative of the $\mathbb{P}$-absolutely continuous part of the measure $\nu$ defined by (1.1) $\nu(A)=\lim_{m}\int_{A}f_{m}\,\mathrm{d}\mathbb{P},\qquad A\in\cup_{n}\mathscr{F}_{n}.$ This limit exists because of the martingale property of $(f_{n})$. The same convergence theorem as above holds true for $L^{1}_{X}$-bounded $X$-valued martingales $(f_{n})$, provided there exists a Radon-Nikodým derivative of the $\mathbb{P}$-absolutely continuous part of the now $X$-valued measure $\nu$ in (1.1). Banach spaces $X$ where this is always possible are said to have the _Radon-Nikodým property_ (RNP) (see Definition 2.3). The RNP of a Banach space is even characterized by martingale convergence meaning that in any Banach space $X$ without RNP, we can find a non-convergent and $L^{1}_{X}$-bounded martingale. Consider now the special case where each $\sigma$-algebra $\mathscr{F}_{n}$ is generated by a partition of a bounded interval $I\subset\mathbb{R}$ into finitely many intervals $(I_{n,i})_{i}$ of positive length as atoms of $\mathscr{F}_{n}$. In this case, $(\mathscr{F}_{n})$ is called an _interval filtration_ on $I$. Then, the characteristic functions $(\mathbbm{1}_{I_{n,i}})$ of those atoms are a sharply localized orthogonal basis of $L^{2}(\mathscr{F}_{n})$ w.r.t. Lebesgue measure $\lambda=|\cdot|$. If we want to preserve the localization property of the basis functions, but at the same time consider spaces of functions with higher smoothness, a natural candidate are spaces of piecewise polynomial functions of order $k$, given by $\displaystyle S^{k}(\mathscr{F}_{n})=\\{f:I\to\mathbb{R}\ |\ $ $f$ is $k-2$ times continuously differentiable and $\displaystyle\qquad\text{a polynomial of order $k$ on each atom of $\mathscr{F}_{n}$}\\},$ where $k$ is an arbitrary positive integer. One reason for this is that $S^{k}(\mathscr{F}_{n})$ admits a special basis, the so called B-spline basis $(N_{n,i})_{i}$, that consists of non-negative and localized functions $N_{n,i}$. Here, the term “localized” means that the support of each function $N_{n,i}$ consists of at most $k$ neighbouring atoms of $\mathscr{F}_{n}$. A second reason is that if $(\mathscr{F}_{n})$ is an increasing sequence of interval $\sigma$-algebras, then the sequence of corresponding spline spaces $S^{k}(\mathscr{F}_{n})$ is increasing as well. Note that the properties of the B-spline functions $(N_{n,i})$ imply that they do not form an orthogonal basis of $S^{k}(\mathscr{F}_{n})$ for $k\geq 2$. For more information on spline functions, see e.g. [11]. Let $P_{n}^{k}$ be the orthogonal projection operator onto $S^{k}(\mathscr{F}_{n})$ with respect to the $L^{2}$ inner product on $I$ equipped with the Lebesgue measure. Since the space $S^{1}(\mathscr{F}_{n})$ consists of piecewise constant functions, $P_{n}^{1}$ is the conditional expectation operator with respect to the $\sigma$-algebra $\mathscr{F}_{n}$ and the Lebesgue measure. In general, the operator $P_{n}^{k}$ can be written in terms of the B-spline basis $(N_{n,i})$ as (1.2) $P_{n}^{k}f=\sum_{i}\int_{I}fN_{n,i}\,\mathrm{d}\lambda\cdot N_{n,i}^{*},$ where the functions $(N_{n,i}^{*})$, contained in the spline space $S^{k}(\mathscr{F}_{n})$, are the biorthogonal (or dual) system to the B-spline basis $(N_{n,i})$. Due to the uniform boundedness of the B-spline functions $N_{n,i}$, we are able to insert functions $f$ in formula (1.2) that are contained not only in $L^{2}$, but in the Lebesgue-Bochner space $L^{1}_{X}$, thereby extending the operator $P_{n}^{k}$ to $L^{1}_{X}$. Similarly to the definition of martingales, we adopt the following notion introduced in [7]: let $(f_{n})_{n\geq 1}$ be a sequence of functions in the space $L^{1}_{X}$. We call this sequence a _$k$ -martingale spline sequence_ (adapted to $(\mathscr{F}_{n}))$ if $P_{n}^{k}f_{n+1}=f_{n},\qquad n\geq 1.$ The local nature of the B-splines and the nestedness of the spaces $(S^{k}(\mathscr{F}_{n}))_{n}$ ultimately allow us to transfer the classical martingale theorems discussed above to $k$-martingale spline sequences adapted to _arbitrary_ interval filtrations ($\mathscr{F}_{n}$) and for any positive integer $k$, just by replacing conditional expectation operators with the spline projection operators $P_{n}^{k}$. Indeed, for any positive integer $k$, we have the following results. 1. (i) (Shadrin’s theorem) There exists a constant $C$ (depending only on $k$ and not on $(\mathscr{F}_{n})$) such that $\sup_{n}\|P_{n}^{k}:L^{1}_{X}\to L^{1}_{X}\|\leq C.$ 2. (ii) (Doob’s inequality for splines) There exists a constant $C$ such that for any $k$-martingale spline sequence $(f_{n})$, $|\\{\sup_{n}\|f_{n}\|>t\\}|\leq C\frac{\sup_{n}\|f_{n}\|_{L^{1}_{X}}}{t},\qquad t>0.$ 3. (iii) (Pointwise convergence of spline projections) For any Banach space $X$ and any $f\in L^{1}_{X}$, the sequence $P_{n}^{k}f$ converges almost everywhere to some $L^{1}_{X}$-function. 4. (iv) (RNP characterization by pointwise spline convergence) For any Banach space $X$, the following statements are equivalent: 1. (a) $X$ has RNP, 2. (b) every $k$-martingale spline sequence that is bounded in $L^{1}_{X}$ converges almost everywhere to an $L^{1}_{X}$-function. We give a few comments regarding the proofs of the statements (i)–(iv) above. Property (i), for arbitrary $k$, was proved by A. Shadrin in the groundbreaking paper [12]. We also refer to the article [3] by M. v. Golitschek, who gives a substantially shorter proof of (i). It should be noted that in the case $k=1$, due to Jensen’s inequality for conditional expectations, we can choose $C=1$ in (i). Property (ii) is proved in [9]. By a standard argument for passing from a weak type (1,1) inequality of a maximal function to a.e. convergence for $L^{1}_{X}$-functions, item (iii) is proved in [9] in the case that $\cup_{n}\mathscr{F}_{n}$ generates the Borel-$\sigma$-algebra on $I$ and in [5] in general. We also identify the limit of $P_{n}^{k}f$ as $P_{\infty}f$, where $P_{\infty}$ is (the $L^{1}_{X}$-extension of) the orthogonal projector onto the $L^{2}$-closure of $\cup_{n}S^{k}(\mathscr{F}_{n})$. The implication (a)$\implies$(b) in item (iv) is also proved in [5], whereas the reverse implication (b)$\implies$(a) is shown in [7] by constructing a non-convergent $k$-martingale spline sequence with values in Banach spaces $X$ without RNP for any positive integer $k$. In this article we are concerned with similar results pertaining to tensor product spline projections. Let $d$ be a positive integer and, for $j=1,\ldots,d$, let $(\mathscr{F}_{n}^{j})$ be an interval filtration on the interval $I\subset\mathbb{R}$. Filtrations $(\mathscr{F}_{n})$ of the form $\mathscr{F}_{n}=\mathscr{F}_{n}^{1}\otimes\cdots\otimes\mathscr{F}_{n}^{d}$ will be called an _interval filtration_ on the cube $I^{d}$. Then, the atoms of $\mathscr{F}_{n}$ are of the form $A_{1}\times\cdots\times A_{d}$ with atoms $A_{j}$ in $\mathscr{F}_{n}^{j}$. For a tuple $k=(k_{1},\ldots,k_{d})$ consisting of $d$ positive integers, denote by $P_{n}^{k}$ the orthogonal projector with respect to $d$-dimensional Lebesgue measure $|\cdot|=\lambda^{d}$ onto the tensor product spline space $S^{k_{1}}(\mathscr{F}_{n}^{1})\otimes\cdots\otimes S^{k_{d}}(\mathscr{F}_{n}^{d})$. The tensor product structure of $P_{n}^{k}$ immediately allows us to conclude (i) in this case, i.e., $P_{n}^{k}$ is bounded on $L^{1}_{X}(I^{d})$ by a constant depending only on $k$ (cf. also [8, Corollary 3.1]). Similarly to the one-dimensional case above, we then introduce the following notion: ###### Definition 1.1. Let $(\mathscr{F}_{n})$ be an interval filtration on a $d$-dimensional cube $I^{d}$. A sequence of functions $(f_{n})_{n\geq 1}$ in the space $L^{1}_{X}(I^{d})$ is a _$k$ -martingale spline sequence_ (adapted to ($\mathscr{F}_{n}$)) if $P_{n}^{k}f_{n+1}=f_{n},\qquad n\geq 1.$ The implication (b)$\implies$(a) in item (iv) for martingale spline sequences on $I^{d}$ can easily be deduced from its one-dimensional version as well. Indeed, for Banach spaces $X$ without RNP we get, for any positive integer $k_{1}$, a non-convergent $X$-valued $k_{1}$-martingale spline sequence $(f_{n}^{1})$ on $I$. Then, $f_{n}(x_{1},\ldots,x_{d})=f_{n}^{1}(x_{1})$ is a non-convergent $X$-valued $(k_{1},\ldots,k_{d})$-martingale spline sequence on $I^{d}$ for any choice of positive integers $k_{2},\ldots,k_{d}$. The main objective of this article is to prove the remaining assertions (ii), (iii) and the implication (a)$\implies$(b) in item (iv) for martingale spline sequences on $I^{d}$. The basic idea in the proof of (ii) for $d=1$ (see [9, Proposition 2.3]) is the pointwise bound (1.3) $\|P_{n}^{k}f(x)\|\leq C_{k}\mathscr{M}_{\rm HL}f(x)$ of $P_{n}^{k}$ by the _Hardy-Littlewood maximal function_ (1.4) $\mathscr{M}_{\rm HL}f(x)=\sup_{J\ni x}\frac{1}{|J|}\int_{J}\|f(y)\|\,\mathrm{d}y,$ where $\sup$ is taken over all intervals $J$ that contain the point $x$. This is enough to imply (ii) for $d=1$ as it is a well known fact that $\mathscr{M}_{\rm HL}$ itself satisfies the weak type (1,1) bound $|\\{\mathscr{M}_{\rm HL}f>t\\}|\leq\frac{3}{t}\|f\|_{L^{1}_{X}},\qquad t>0.$ In dimensions $d>1$, by using this ad-hoc approach (see [8, Proposition 3.3]) one would need the _strong maximal function_ $\mathscr{M}_{\rm S}f(x)$ on the right hand side of (1.3), where $\mathscr{M}_{\rm S}f(x)$ is defined by the same formula (1.4) as $\mathscr{M}_{\rm HL}f(x)$, but where $\sup$ is taken over all $d$-dimensional axis-parallel rectangles $J\subset I^{d}$ containing the point $x$. As a matter of fact, this is not enough to derive (ii), since the best possible weak type inequality for $\mathscr{M}_{\rm S}$ is true only in the Orlicz space $L(\log L)^{d-1}$ (see [1, 4, 10]), which is a strict subset of $L^{1}$. Here we show how to employ the martingale spline structure, especially nestedness of atoms, to avoid the usage of the strong maximal function $\mathscr{M}_{\rm S}$ altogether and replace it by an intrinsic maximal function that is (as we will show) of weak type $(1,1)$. This is crucial in the proof of the statements (ii), (iii), (iv) for any dimension $d$. Those statements are in full analogy to the martingale and one-dimensional spline results. The validity of (ii) and (iii) for martingale spline sequences on $I^{d}$ solves a problem stated in [8]. The organization of this article is as follows. In Section 2 we collect a few basic facts about vector measures needed in the sequel. In Section 3, we prove items (ii) and (iii) for martingale spline sequences on $I^{d}$ (Proposition 3.1 and Theorem 3.3 respectively). In Section 4, the implication (a)$\implies$(b) of item (iv) is proved in this case (Theorem 4.1) under the restriction that $\cup_{n}\mathscr{F}_{n}$ generates the Borel-$\sigma$-algebra on $I^{d}$. In Section 5, we show this assertion for general interval filtrations on $I^{d}$ and give an explicit formula for the pointwise limit of martingale spline sequences. ## 2\. Preliminaries We refer to the book [2] by J. Diestel and J.J. Uhl for basic facts on vector valued integration, martingales, vector measures and the results that follow. Let $\Omega$ be a set, $\mathscr{A}$ an algebra of subsets of $\Omega$ and $(X,\|\cdot\|)$ a Banach space. A function $\nu:\mathscr{A}\to X$ is a _(finitely additive) vector measure_ if, whenever $E_{1},E_{2}\in\mathscr{A}$ are disjoint, we have $\nu(E_{1}\cup E_{2})=\nu(E_{1})+\nu(E_{2})$. If, in addition, $\nu(\cup_{n=1}^{\infty}E_{n})=\sum_{n=1}^{\infty}\nu(E_{n})$ in the norm topology of $X$ for all sequences $(E_{n})$ of mutually disjoint members of $\mathscr{A}$ such that $\cup_{n=1}^{\infty}E_{n}\in\mathscr{A}$, then $\nu$ is a _countably additive vector measure_. The _variation_ $|\nu|$ of a finitely additive vector measure $\nu$ is the set function $|\nu|(E)=\sup_{\pi}\sum_{A\in\pi}\|\nu(A)\|,$ where the supremum is taken over all partitions $\pi$ of $E$ into a finite number of mutually disjoint members of $\mathscr{A}$. If $\nu$ is a finitely additive vector measure, then the variation $|\nu|$ is monotone and finitely additive. The measure $\nu$ is of _bounded variation_ if $|\nu|(\Omega)<\infty$. If $\mu:\mathscr{A}\to[0,\infty)$ is a finitely additive measure and $\nu:\mathscr{A}\to X$ is a finitely additive vector measure, $\nu$ is _$\mu$ -continuous_, if $\lim_{\mu(E)\to 0}\nu(E)=0$. If $\mu_{1},\mu_{2}:\mathscr{A}\to[0,\infty)$ are two finitely additive measures on $\mathscr{A}$, $\mu_{1}$ and $\mu_{2}$ are mutually _singular_ if for each $\varepsilon>0$ there exists a set $A\in\mathscr{A}$ so that $\mu_{1}(A^{c})+\mu_{2}(A)\leq\varepsilon.$ ###### Theorem 2.1 (Lebesgue decomposition of vector measures). Let $\mathscr{A}$ be an algebra of subsets of the set $\Omega$. Let $\nu:\mathscr{A}\to X$ be a finitely additive vector measure of bounded variation. Let $\mu:\mathscr{A}\to[0,\infty)$ be a finitely additive measure. Then there exist unique finitely additive vector measures of bounded variation $\nu_{c},\nu_{s}$ so that 1. (1) $\nu=\nu_{c}+\nu_{s}$, $|\nu|=|\nu_{c}|+|\nu_{s}|$, 2. (2) $\nu_{c}$ is $\mu$-continuous, 3. (3) $|\nu_{s}|$ and $\mu$ are mutually singular. This theorem can be found in [2, Theorem 9 on p. 31]. The following theorem is part of [2, Theorem 2 on p. 27] after using [2, Proposition 15 on p. 7]. ###### Theorem 2.2 (Extension theorem). Let $\mathscr{A}$ be an algebra of subsets of a set $\Omega$ and let $\mathscr{F}$ be the $\sigma$-algebra generated by $\mathscr{A}$. Let $\nu:\mathscr{A}\to X$ be a countably additive vector measure of bounded variation. Then, $\nu$ has a unique countably additive extension $\overline{\nu}:\mathscr{F}\to X$. ###### Definition 2.3 ([2, Definition 3, p. 61]). A Banach space $X$ admits the _Radon-Nikodým property (RNP)_ if for every measure space $(\Omega,\mathscr{F})$, for every positive, finite, countably additive measure $\mu$ on $(\Omega,\mathscr{F})$ and for every $\mu$-continuous, countably additive vector measure $\nu$ of bounded variation, there exists a function $f\in L^{1}_{X}(\Omega,\mathscr{F},\mu)$ such that $\nu(A)=\int_{A}f\,\mathrm{d}\mu,\qquad A\in\mathscr{F}.$ ## 3\. Maximal functions of Tensor spline projectors Let $d$ be a positive integer and let $(\mathscr{F}_{n})=(\mathscr{F}_{n}^{1}\otimes\cdots\otimes\mathscr{F}_{n}^{d})$ be an interval filtration on $I^{d}$ for some interval $I=(a,b]$ with $a<b$ and $a,b\in\mathbb{R}$. Each $\sigma$-algebra $\mathscr{F}_{n}$ is then generated by a finite, mutually disjoint family $\\{I_{n,i}:i\in\Lambda\\}$, $\Lambda\subset\mathbb{Z}^{d}$, of $d$-dimensional rectangles of the form $I_{n,i}=\prod_{\ell=1}^{d}(a_{\ell},b_{\ell}]$ for some $a\leq a_{\ell}<b_{\ell}\leq b$. We assume that $\Lambda$ is of the form $\Lambda^{1}\times\cdots\times\Lambda^{d}$ where for each $\ell=1,\ldots,d$, $\Lambda^{\ell}$ is a finite set of consecutive integers and the rectangles $I_{n,i}$ have the property that they are ordered in the same way as $\mathbb{R}^{d}$, i.e., if $i,j\in\Lambda$ with $i_{\ell}<j_{\ell}$ then the projection of $I_{n,i}$ onto the $\ell$th coordinate axis lies to the left of the projection of $I_{n,j}$ onto the $\ell$th coordinate axis. For $x\in I^{d}$, let $A_{n}(x)$ be the uniquely determined atom (rectangle) $A\in\mathscr{F}_{n}$ so that $x\in A$. For two atoms $A,B\in\mathscr{F}_{n}$, define $d_{n}(A,B):=|i-j|_{1}$ if $A=I_{n,i}$ and $B=I_{n,j}$ and where $|w|_{1}=\sum_{\ell=1}^{d}|w_{\ell}|$ denotes the $\ell^{1}$ norm of the vector $w$. If $U=\cup_{\ell}A_{\ell}$ and $V=\cup_{\ell}B_{\ell}$ are (finite) unions of atoms in $\mathscr{F}_{n}$, we set $d_{n}(U,V)=\min_{\ell,m}d_{n}(A_{\ell},B_{m})$. Additionally, for a non- negative integer $s$, define $A_{n,s}(x)$ to be the union of all atoms $A$ in $\mathscr{F}_{n}$ with $d_{n}(A,A_{n}(x))\leq s$. Moreover, for a Borel set $B\subset I^{d}$, let $A_{n,s}(B)=\cup_{x\in B}A_{n,s}(x)$. For each $\ell=1,\ldots,d$, let $k_{\ell}$ be a positive integer. Define the tensor product spline space of order $k=(k_{1},\ldots,k_{d})$ associated to $\mathscr{F}_{n}$ as $S_{n}:=S^{k_{1}}(\mathscr{F}_{n}^{1})\otimes\cdots\otimes S^{k_{d}}(\mathscr{F}_{n}^{d}).$ The space $S_{n}$ admits the tensor product B-spline basis $(N_{n,i})_{i}$ defined by $N_{n,i}=N_{n,i_{1}}^{1}\otimes\cdots\otimes N_{n,i_{d}}^{d},$ where $(N_{n,i_{\ell}}^{\ell})_{i_{\ell}}$ denotes the B-spline basis of $S^{k_{\ell}}(\mathscr{F}_{n}^{\ell})$ that forms a partition of unity. The support $E_{n,i}=\operatorname{supp}N_{n,i}$ of $N_{n,i}$ is composed of at most $k_{1}\cdots k_{d}$ neighouring atoms of $\mathscr{F}_{n}$. Consider the orthogonal projection operator $P_{n}=P_{n}^{k}$ onto $S_{n}$ with respect to $d$-dimensional Lebesgue measure $|\cdot|=\lambda^{d}$. Using the B-spline basis and its biorthogonal system $(N_{n,i}^{*})$, the orthogonal projector $P_{n}$ is given by (3.1) $P_{n}f=\sum_{i}\int_{I^{d}}fN_{n,i}\,\mathrm{d}\lambda^{d}\cdot N_{n,i}^{*},\qquad f\in L^{1}_{X}(I^{d}).$ The dual B-spline functions $N_{n,i}^{*}$ admit the following crucial geometric decay estimate (3.2) $|N_{n,i}^{*}(x)|\leq C\frac{q^{d_{n}(E_{n,i},A_{n}(x))}}{|\operatorname{re}(E_{n,i}\cup A_{n}(x))|},\qquad x\in I^{d},$ for some constants $C$ and $q\in[0,1)$ that depend only on $k$, where $\operatorname{re}(S)$ denotes the smallest, axis-parallel rectangle containing the set $S$. This inequality was shown in [9, Theorem 1.2] for $d=1$ and if $d>1$, (3.2) is a consequence of the fact that $N_{n,i}^{*}$ is the tensor product of one-dimensional dual B-spline functions. Inserting this estimate in formula (3.1) for $P_{n}f$ and as $E_{n,i}$ consists of at most $k_{1}\cdots k_{d}$ neighbouring atoms of $\mathscr{F}_{n}$, setting $C_{k}:=C(k_{1}\cdots k_{d})q^{-|k|_{1}}$, we get the pointwise estimate (3.3) $\|P_{n}f(x)\|\leq C_{k}\sum_{A\text{ atom of }\mathscr{F}_{n}}b_{n}(q,\|f\|\,\mathrm{d}\lambda^{d},A,x),\qquad f\in L^{1}_{X}(I^{d})$ introducing the expression (3.4) $b_{n}(q,\theta,A,x)=\frac{q^{d_{n}(A,A_{n}(x))}}{|\operatorname{re}(A\cup A_{n}(x))|}\theta(A),\qquad A\text{ atom in }\mathscr{F}_{n},x\in I^{d}$ for a positive, finitely additive measure $\theta$ on the algebra $\mathscr{A}=\cup_{m}\mathscr{F}_{m}$. In view of inequality (3.3), it suffices to consider, instead of the maximal function of the projection operators $P_{n}$, the maximal functions given by (3.5) $\mathscr{M}_{K}\theta(x)=\sup_{n\geq K}\sum_{A\text{ atom of }\mathscr{F}_{n}}b_{n}(q,\theta,A,x)\qquad x\in I^{d},f\in L^{1}(I^{d})$ for any positive integer $K$ and some fixed parameter $q\in[0,1)$. If we abbreviate by $\mathscr{M}f$ the maximal function $\mathscr{M}_{1}(|f|\,\mathrm{d}\lambda^{d})$, we have the following weak type (1,1) result. ###### Proposition 3.1. The maximal function $\mathscr{M}$ is of weak type (1,1), i.e. there exists a constant $C$ depending only on the dimension $d$ and on the parameter $q<1$, so that we have the inequality $|\\{\mathscr{M}f>t\\}|\leq\frac{C}{t}\|f\|_{L^{1}},\qquad t>0,\ f\in L^{1}(I^{d}).$ ###### Proof. Set $B=I^{d}$, $K=1$ and $\theta=|f|\,\mathrm{d}\lambda^{d}$ in Theorem 3.2 below and observe that the geometric series in equation (3.6) converges. ∎ The following result about the maximal operators $\mathscr{M}_{K}$ is the focal point in our investigations. ###### Theorem 3.2. Let $(\mathscr{F}_{n})$ be an interval filtration on $I^{d}$ and let $\theta$ be a non-negative, finitely additive measure on the algebra $\mathscr{A}=\cup_{n}\mathscr{F}_{n}$. Then, for any Borel set $B\subset I^{d}$ and any positive integer $K$, (3.6) $|B\cap\\{\mathscr{M}_{K}\theta>t\\}|\leq\frac{C}{t}\cdot\sum_{s=0}^{\infty}q^{s/2}(s+1)^{d-1}\theta\big{(}A_{K,s}(B)\big{)},\qquad t>0$ for some constant $C$ depending only on $q$ and $d$. ###### Proof. Set $G_{t}=B\cap\\{\mathscr{M}_{K}\theta>t\\}$ and let $x\in G_{t}$. Then, there exists an index $n\geq K$ so that $\sum_{A\text{ atom of }\mathscr{F}_{n}}b_{n}(q,\theta,A,x)>t.$ Letting $c=(2\sum_{\ell=0}^{\infty}\rho^{\ell})^{d}=\big{(}2/(1-\rho)\big{)}^{d}<\infty$ with $\rho=q^{1/2}$, we obtain that there exists at least one atom $F$ of the $\sigma$-algebra $\mathscr{F}_{n}$ so that (3.7) $b_{n}(\rho,\theta,F,x)>t/c.$ Therefore, for $x\in G_{t}$, we choose $n_{x}<\infty$ to be the minimal index $n\geq K$ so that there exists an atom $F$ of $\mathscr{F}_{n_{x}}$ satisfying inequality (3.7). We choose a particular atom $F$ of $\mathscr{F}_{n_{x}}$ with this property which will be denoted by $F_{x}$. The collection of atoms $\\{A_{n_{x}}(x):x\in G_{t}\\}$ is nested and covers the set $G_{t}$. Thus, there exists a maximal countable subcollection $\\{A_{n_{x}}(x):x\in\Gamma\\}$ consisting of mutually disjoint sets that still covers $G_{t}$. Perform the following estimate using inequality (3.7): (3.8) $\displaystyle|G_{t}|$ $\displaystyle\leq\sum_{x\in\Gamma}|A_{n_{x}}(x)|\leq\frac{c}{t}\cdot\sum_{x\in\Gamma}\rho^{d_{n_{x}}(F_{x},A_{n_{x}}(x))}\theta(F_{x})$ $\displaystyle=\frac{c}{t}\cdot\sum_{s\geq 0}\rho^{s}\sum_{m\in\mathbb{Z}^{d}:|m|_{1}=s}\Big{(}\sum_{x\in\Gamma_{m}}\theta(F_{x})\Big{)},$ where for $m\in\mathbb{Z}^{d}$, $\Gamma_{m}$ is the set of all $x\in\Gamma$ so that, if $A_{n_{x}}(x)=I_{n_{x},i}$ and $F_{x}=I_{n_{x},j}$ for some $i,j\in\mathbb{Z}^{d}$, we have $i-j=m$. Next, we show that for each $m\in\mathbb{Z}^{d}$, the collection $\\{F_{x}:x\in\Gamma_{m}\\}$ consists of mutually disjoint sets. Assume the contrary, i.e. for some $m\in\mathbb{Z}^{d}$ there exist two points $x,y\in\Gamma_{m}$ that are different from each other with $F_{x}\cap F_{y}\neq\emptyset$. For definiteness, assume that $n_{x}\geq n_{y}$, and thus the nestedness of the $\sigma$-algebras $(\mathscr{F}_{n})$ implies $F_{x}\subseteq F_{y}$. Assume that $i,i^{\prime},j,j^{\prime}\in\mathbb{Z}^{d}$ are such that $I_{n_{x},i}=A_{n_{x}}(x),\quad I_{n_{y},i^{\prime}}=A_{n_{y}}(y),\quad I_{n_{x},j}=F_{x},\quad I_{n_{y},j^{\prime}}=F_{y}.$ Since $x,y\in\Gamma_{m}$, we know that $i-j=m=i^{\prime}-j^{\prime}$. Therefore, since $\mathscr{F}_{n_{x}}$ is finer than $\mathscr{F}_{n_{y}}$ and by the inclusion $F_{x}\subseteq F_{y}$, we have (3.9) $\operatorname{re}(F_{x}\cup A_{n_{x}}(x))\subseteq\operatorname{re}(F_{y}\cup A_{n_{y}}(x))\subseteq\operatorname{re}(F_{y}\cup A_{n_{y}}(y)).$ Moreover, this and the definition of the distance $d_{n_{y}}$ implies (3.10) $d_{n_{y}}(F_{y},A_{n_{y}}(y))\geq d_{n_{y}}(F_{y},A_{n_{y}}(x)).$ Combining (3.9) and (3.10) yields $b_{n_{y}}(\rho,\theta,F_{y},x)\geq b_{n_{y}}(\rho,\theta,F_{y},y)$; additionally, by definition of $n_{y},F_{y}$ we have the inequality $b_{n_{y}}(\rho,\theta,F_{y},y)>t/c$. Together, this implies $b_{n_{y}}(\rho,\theta,F_{y},x)>t/c.$ As $n_{x}\geq K$ is the minimal index so that such an inequality at the point $x$ is possible and $n_{x}\geq n_{y}$ we get that $n_{x}=n_{y}=:n$. Since $A_{n}(x)\cap A_{n}(y)=\emptyset$ we know that in this case $i\neq i^{\prime}$ and $x,y\in\Gamma_{m}$ implies $i-j=m=i^{\prime}-j^{\prime}$. Together, this yields $j\neq j^{\prime}$ which means $F_{x}\cap F_{y}=\emptyset$, contradicting the assumption $F_{x}\subseteq F_{y}$. Therefore, $F_{x}$ and $F_{y}$ are disjoint, concluding the proof of the fact that $\\{F_{x}:x\in\Gamma_{m}\\}$ consists of mutually disjoint sets for each $m\in\mathbb{Z}^{d}$. If $(U_{j})$ is a countable collection of disjoint members of $\mathscr{A}$ and if $U\in\mathscr{A}$ with $\cup_{j=1}^{\infty}U_{j}\subset U$, then $\sum_{j=1}^{\infty}\theta(U_{j})\leq\theta(U)$, since for finite sums this is clear by finite additivity and positivity of $\theta$ and the general case follows by passing to infinity. We apply this simple fact to the sum $\sum_{x\in\Gamma_{m}}\theta(F_{x})$ with $U=A_{K,|m|_{1}}(B)$ to obtain from (3.8) $\displaystyle|G_{t}|$ $\displaystyle\leq\frac{c}{t}\cdot\sum_{s=0}^{\infty}\rho^{s}\theta\big{(}A_{K,s}(B)\big{)}\Big{(}\sum_{|m|_{1}=s}1\Big{)}\leq\frac{2^{d}c}{t}\sum_{s=0}^{\infty}\rho^{s}(s+1)^{d-1}\theta\big{(}A_{K,s}(B)\big{)},$ which is the conclusion of the theorem. ∎ Combining Proposition 3.1 with the bound (3.3) on the operators $P_{n}$, we obtain that the maximal function of the spline projectors $P_{n}$ also satisfies a weak type (1,1) inequality (3.11) $|\\{\sup_{n}\|P_{n}f\|>t\\}|\leq\frac{C\|f\|_{L^{1}_{X}}}{t},\qquad t>0,\ f\in L^{1}_{X}(I^{d}),$ for some constant $C$ depending only on $k$. This proves Doob’s inequality (ii) on page ii for martingale spline sequences on $I^{d}$. Indeed, given a martingale spline sequence $(f_{n})$ on $I^{d}$, apply (3.11) to the function $f=f_{m}$ for a fixed positive integer $m$ and pass $m\to\infty$ to get (ii) for martingale spline sequences on $I^{d}$. As a corollary, we have the following result about almost everywhere convergence of $P_{n}f$ for $f\in L^{1}_{X}(I^{d})$, proving (iii) for tensor spline projections. ###### Theorem 3.3. Let $X$ be any Banach space and let $f\in L^{1}_{X}(I^{d})$. Then, there exists $g\in L^{1}_{X}(I^{d})$ such that $P_{n}f\to g\qquad\text{$\lambda^{d}$-almost everywhere}.$ ###### Remark. (i) The proof of Theorem 3.3 follows along the same lines as the proof of the one-dimensional case [5, Theorem 3.2] and uses standard arguments for passing from a weak type maximal inequality of the form (3.11) to almost everywhere convergence of $P_{n}f$ for $L^{1}$-functions $f$. For this argument, a dense subset of $L^{1}$ is needed, for which it is “clear” that pointwise convergence takes place. In [5, Lemma 3.1], for one-dimensional splines, this dense set is chosen to be the space of continuous functions $C(\bar{I})$ on the closure of the interval $I$. For arbitrary dimensions $d$, we can use $C(\bar{I})\otimes\cdots\otimes C(\bar{I})$ as dense subset of $L^{1}$, for which it is a consequence of the one-dimensional convergence result [5, Lemma 3.1] and its tensor product structure that $P_{n}f$ converges pointwise for $f\in C(\bar{I})\otimes\cdots\otimes C(\bar{I})$. (ii) As in the one-dimensional case, the limit function $g$ in Theorem 3.3 can be identified explicitly as the ($L^{1}_{X}$-extension of the) orthogonal projection of the function $f$ onto the closure of $\cup_{n}S_{n}$, which, in the particular case that $\cup_{n}\mathscr{F}_{n}$ generates the Borel-$\sigma$-algebra on $I^{d}$, coincides with the function $f$. We also note another immediate corollary of Theorem 3.2 that will be used later. ###### Corollary 3.4. Let $(\mathscr{F}_{n})$ be an interval filtration on $I^{d}$ and let $\theta$ be a non-negative, finitely additive measure on the algebra $\mathscr{A}=\cup_{n}\mathscr{F}_{n}$. Let $D\in\mathscr{A}$ be arbitrary and set $L_{t}:=\Big{\\{}x\in I^{d}:\limsup_{n}\sum_{A\text{ atom of }\mathscr{F}_{n}}b_{n}(q,\theta,A,x)>t\Big{\\}}.$ Let $R$ be a non-negative integer. If $B\subset D$ is a Borel set such that $A_{K,R}(B)\subset D$ for some $K$, we have (3.12) $|B\cap L_{t}|\leq\frac{C}{t}\Big{(}\theta(D)+\sum_{s>R}q^{s/2}(s+1)^{d-1}\theta(I^{d})\Big{)},\qquad t>0$ for some constant $C$ depending only on $d$ and $q$. ###### Proof. This just follows from Theorem 3.2 by noting that $L_{t}\subset\\{\mathscr{M}_{K}\theta>t\\}$ for any positive integer $K$. ∎ ###### Remark. Assume that in Corollary 3.4, the measure $\theta$ is a $\sigma$-additive Borel measure on $\bar{I}^{d}$ and replace the term $\theta(A)$ in the definition (3.4) of $b_{n}$ by the term $\theta(\overline{A})$ with the closure $\overline{A}$ of $A$ in $\bar{I}^{d}$. Then, the assertion of Corollary 3.4 still holds if we replace $\theta(D)$ and $\theta(I^{d})$ on the right hand side of (3.12) by $\theta(\overline{D})$ and $\theta(\bar{I}^{d})$ respectively. Indeed, the only modification in the proof of Theorem 3.2 is that we have to replace $F_{x}$ by $\overline{F_{x}}$ in (3.8), but this only gives an additional factor of $2^{d}$ on the right hand side of (3.6) and (3.12) since each point of $\bar{I}^{d}$ is contained in at most $2^{d}$ closures of disjoint rectangles. ## 4\. The Convergence Theorem for dense filtrations In this section, we show the remaining implication (a)$\implies$(b) of item (iv) on page iv for martingale spline sequences on $I^{d}$ in the case where $\mathscr{A}:=\cup_{n}\mathscr{F}_{n}$ generates the Borel-$\sigma$-algebra on $I^{d}$. We restrict ourselves to this special setting in this section to present the crucial arguments in a concise form. In order to lift the subsequent result from this hypothesis, we use technical arguments in the spirit of those in the proof of the one-dimensional result [5, Sections 4 and 6]. This will be presented in detail in Section 5. ###### Theorem 4.1. Let $(\mathscr{F}_{n})$ be an interval filtration on $I^{d}$ so that $\mathscr{A}=\cup_{n}\mathscr{F}_{n}$ generates the Borel-$\sigma$-algebra and let $X$ be a Banach space with RNP. Let $(g_{n})$ be an $X$-valued martingale spline sequence adapted to $(\mathscr{F}_{n})$ with $\sup_{n}\|g_{n}\|_{L^{1}_{X}}<\infty$. Then, there exists $g\in L^{1}_{X}(I^{d})$ so that $g_{n}\to g$ almost everywhere with respect to Lebesgue measure $\lambda^{d}$. ###### Remark. As for martingales (see [2]), the basic proof idea of this result is to define a vector measure $\nu$ based upon the martingale spline sequence $(g_{n})$, whose absolutely continuous part with respect to Lebesgue measure $\lambda^{d}$ has a density $g\in L^{1}_{X}$ by the RNP of $X$, which is then the a.e. limit of $g_{n}$. ###### Proof. Part I: The limit operator $T$. For $f\in S_{m}$ and $n\geq m$, since the operator $P_{n}$ is selfadjoint and using the martingale spline property of the sequence $(g_{n})$, $\displaystyle\int_{I^{d}}g_{n}\cdot f\,\mathrm{d}\lambda^{d}$ $\displaystyle=\int_{I^{d}}g_{n}\cdot P_{m}f\,\mathrm{d}\lambda^{d}=\int_{I^{d}}P_{m}g_{n}\cdot f\,\mathrm{d}\lambda^{d}=\int_{I^{d}}g_{m}\cdot f\,\mathrm{d}\lambda^{d}.$ This means in particular that for all $f\in\cup_{m}S_{m}$, the limit of $\int_{I^{d}}g_{n}\cdot f\,\mathrm{d}\lambda^{d}$ exists, so we define the linear operator $T:\cup_{m}S_{m}\to X,\qquad f\mapsto\lim_{n}\int_{I^{d}}g_{n}\cdot f\,\mathrm{d}\lambda^{d}.$ We can write $g_{n}$ in terms of this operator $T$. Indeed, by the martingale spline property of $(g_{n})$ again, (4.1) $\displaystyle g_{n}=P_{n}g_{n}$ $\displaystyle=\sum_{i}\int_{I^{d}}g_{n}N_{n,i}\,\mathrm{d}\lambda^{d}\cdot N_{n,i}^{*}$ $\displaystyle=\sum_{i}\lim_{m}\int_{I^{d}}g_{m}N_{n,i}\,\mathrm{d}\lambda^{d}\cdot N_{n,i}^{*}=\sum_{i}(TN_{n,i})N_{n,i}^{*}.$ By Alaoglu’s theorem, we may choose a subsequence $\ell_{n}$ such that the bounded sequence of measures $\|g_{\ell_{n}}\|_{X}\,\mathrm{d}\lambda^{d}$ converges in the weak*-topology on the space of Radon measures on the closure $\bar{I}^{d}$ of $I^{d}$ to some finite scalar measure $\mu$ on $\bar{I}^{d}$, i.e. $\lim_{n\to\infty}\int_{\bar{I}^{d}}f\|g_{\ell_{n}}\|\,\mathrm{d}\lambda^{d}=\int_{\bar{I}^{d}}f\,\mathrm{d}\mu,\qquad f\in C(\bar{I}^{d}).$ For a fixed positive integer $m$, we then get another subsequence of $(\ell_{n})$, again denoted by $(\ell_{n})$, so that for each atom $A$ of $\mathscr{F}_{m}$, the sequence $\|g_{\ell_{n}}\|\,\mathrm{d}\lambda^{d}$ converges to some Radon measure $\mu_{A}$ on the closure $\overline{A}$ of $A$ satisfying $\mu=\sum_{A\text{ atom of }\mathscr{F}_{m}}\mu_{A}$. Each function $f\in S_{m}$ is continuous and a polynomial in the interior $A^{\circ}$ of each atom $A\in\mathscr{F}_{m}$. Denote by $f_{A}$ the continuous function on the closure $\overline{A}$ of $A$ that coincides with $f$ on $A^{\circ}$. Then, for $\ell_{n}\geq m$ and $f\in S_{m}$ $\displaystyle\|Tf\|$ $\displaystyle=\Big{\|}\int_{I^{d}}fg_{\ell_{n}}\,\mathrm{d}\lambda^{d}\Big{\|}\leq\int_{I^{d}}|f|\|g_{\ell_{n}}\|\,\mathrm{d}\lambda^{d}$ $\displaystyle=\sum_{A\text{ atom of }\mathscr{F}_{m}}\int_{\overline{A}}|f_{A}|\|g_{\ell_{n}}\|\,\mathrm{d}\lambda^{d}\rightarrow\sum_{A\text{ atom of }\mathscr{F}_{m}}\int_{\bar{I}^{d}}|f_{A}|\,\mathrm{d}\mu_{A}$ $\displaystyle\leq\sum_{A\text{ atom of }\mathscr{F}_{m}}\int_{\bar{I}^{d}}\limsup_{s\to y}|f(s)|\,\mathrm{d}\mu_{A}(y)=\int_{\bar{I}^{d}}\limsup_{s\to y}|f(s)|\,\mathrm{d}\mu(y).$ For $f\in\cup_{n}S_{n}$ define (4.2) $\|f\|:=\int_{\bar{I}^{d}}\limsup_{s\to y}|f(s)|\,\mathrm{d}\mu(y),$ which is a seminorm on $\cup_{n}S_{n}$. As for $L^{p}$-spaces, we factor out the functions $f\in\cup_{n}S_{n}$ with $\|f\|=0$ in order to get a norm. Then, denote by $W$ the completion of $\cup_{n}S_{n}$ in this norm and extend the operator $T$ to $W$ continuously. Part II: Representing $T$ in terms of a vector measure $\nu$. Let $Q=\prod_{\ell=1}^{d}(a_{\ell},b_{\ell}]$ be an arbitrary atom of the $\sigma$-algebra $\mathscr{F}_{n}$ for some positive integer $n$. Let $\ell\in\\{1,\ldots,d\\}$ be an arbitrary coordinate direction. If the order of the polynomials $k_{\ell}$ in direction $\ell$ equals $1$ (piecewise constant case), we set $f_{m}^{\ell}=\mathbbm{1}_{(a_{\ell},b_{\ell}]}$ for $m\geq n$, which satisfies $f_{m}^{\ell}\in S^{k_{\ell}}(\mathscr{F}_{m}^{\ell})$. If $k_{\ell}>1$, we first choose an open interval $O$ and a closed interval $C$ (both in $I$) so that $C\subseteq(a_{\ell},b_{\ell}]\subseteq O$ and $|O\setminus C|\leq 1/m$. The sets $C$ and $O$ are chosen so that as many endpoints of $C$ and $O$ coincide with the corresponding endpoints of $(a_{\ell},b_{\ell}]$ as possible. Then, let $f_{m}^{\ell}\in\cup_{j}S^{k_{\ell}}(\mathscr{F}_{j}^{\ell})$ be a non- negative function that is bounded by $1$ and satisfies $\operatorname{supp}f_{m}^{\ell}\subset O\qquad\text{and}\qquad f_{m}^{\ell}\equiv 1\text{ on }C\cap I.$ Such a function exists since $\mathscr{A}$ generates the Borel-$\sigma$-algebra if one additionally notices the facts that B-splines form a partition of unity and have localized support. If we define $f_{m}=f_{m}^{1}\otimes\cdots\otimes f_{m}^{d}$, the sequence $(f_{m})$ is Cauchy in $\cup_{j}S_{j}$ with respect to the norm in (4.2) and we let $I_{Q}$ be the limit in $W$ of the sequence $(f_{m})$ satisfying $\|TI_{Q}\|=\lim_{m\to\infty}\|Tf_{m}\|\leq\mu(\overline{Q})$ (here, the closure of $Q$ is taken in $\bar{I}^{d}$). This definition of $I_{Q}$ also has the property that if $Q$ is an atom in $\mathscr{F}_{n}$ and $(Q_{j})_{j=1}^{\ell}$ is a finite sequence of disjoint atoms $Q_{j}$ in $\mathscr{F}_{n_{j}}$ with $n_{j}\geq n$ and $Q=\cup_{j=1}^{\ell}Q_{j}$, we have $I_{Q}=\sum_{j=1}^{\ell}I_{Q_{j}}$. Therefore, if $\mathscr{F}_{n}\ni A=\cup_{j=1}^{\ell}Q_{j}$ for some disjoint atoms $(Q_{j})_{j=1}^{\ell}$ in $\mathscr{F}_{n}$, it is well defined to set $I_{A}=\sum_{j=1}^{\ell}I_{Q_{j}}\in W.$ Based upon that, we define the finitely additive vector measure $\nu$ on $(I^{d},\mathscr{A})$ with values in $X$ by (4.3) $\nu(A):=T(I_{A}),\qquad A\in\mathscr{A}.$ This vector measure $\nu$ is of bounded variation, since if $\pi$ is a finite partition of $I^{d}$ into sets of $\mathscr{A}$ and if $m<\infty$ is the minimal index so that $A\in\mathscr{F}_{m}$ for all $A\in\pi$, we have $\sum_{A\in\pi}\|T(I_{A})\|\leq\sum_{Q\text{ atom in }\mathscr{F}_{m}}\|T(I_{Q})\|\leq\sum_{Q\text{ atom in }\mathscr{F}_{m}}\mu(\overline{Q})\leq 2^{d}\mu(\bar{I}^{d}),$ as each point in $\bar{I}^{d}$ is contained in at most $2^{d}$ closures of atoms of $\mathscr{F}_{m}$. Observe that for all $f\in\cup_{n}S_{n}$, we have (4.4) $\int_{I^{d}}f\,\mathrm{d}\nu=T(f).$ Indeed, each $f\in\cup_{n}S_{n}$ can be approximated uniformly by linear combinations of characteristic functions of atoms of the form $\chi_{m}:=\sum_{Q\text{ atom in }\mathscr{F}_{m}}\alpha_{Q}\mathbbm{1}_{Q}$ as $m\to\infty$, which then also has the property that $f_{m}:=\sum_{Q\text{ atom in }\mathscr{F}_{m}}\alpha_{Q}I_{Q}\to f$ in $W$ as $m\to\infty$ and thus also $Tf_{m}\to Tf$ in $X$ by the continuity of the operator $T$. As, by definition (4.3) of $\nu$, we have $\int\chi_{m}\,\mathrm{d}\nu=Tf_{m}$, equation (4.4) follows by letting $m\to\infty$. Part III: Conclusion. Continuing the calculation in equation (4.1), using the measure $\nu$ and (4.4), (4.5) $g_{n}=\sum_{i}\int_{I^{d}}N_{n,i}\,\mathrm{d}\nu\cdot N_{n,i}^{*}.$ Apply Lebesgue’s decomposition Theorem 2.1 to the measure $\nu$ with respect to $\lambda^{d}$ to get two finitely additive measures $\nu_{c},\nu_{s}$ of bounded variation with (4.6) $\nu=\nu_{c}+\nu_{s},$ where $\nu_{c}$ is $\lambda^{d}$-continuous and $|\nu_{s}|$ is singular to $\lambda^{d}$. As $\lambda^{d}$ is countably additive, so is the $\lambda^{d}$-continuous measure $\nu_{c}$ and by the extension theorem (Theorem 2.2) extends uniquely to a countably additive vector measure $\overline{\nu_{c}}$ on the Borel-$\sigma$-algebra on $I^{d}$, which, by the RNP of $X$ can be written as $\,\mathrm{d}\overline{\nu_{c}}=g\,\mathrm{d}\lambda^{d}$ for some $g\in L^{1}_{X}$. Therefore, $g_{n}=\sum_{i}\int_{I^{d}}N_{n,i}g\,\mathrm{d}\lambda^{d}\cdot N_{n,i}^{*}+\sum_{i}\int_{I^{d}}N_{n,i}\,\mathrm{d}\nu_{s}\cdot N_{n,i}^{*}.$ The first part on the right hand side of this equation equals $P_{n}g$ for the $L^{1}_{X}$ function $g$ and this converges a.e. to $g$ by Theorem 3.3 and the remark following it. The second part, denoted by $P_{n}\nu_{s}$, converges to $0$ almost everywhere, which we will now show. Let $t>0$ be arbitrary and let $G_{t}:=\\{y\in I^{d}:\limsup_{n}\|P_{n}\nu_{s}(y)\|>t\\}$. Then, let $\varepsilon>0$ be arbitrary and choose $D\in\mathscr{A}$ with the property $\lambda^{d}(D^{c})+|\nu_{s}|(D)\leq\varepsilon$, which is possible since $|\nu_{s}|$ is singular to $\lambda^{d}$. By (3.3), replacing $\|f\|\,\mathrm{d}\lambda^{d}$ with $|\nu_{s}|$, $\displaystyle\|P_{n}\nu_{s}(y)\|$ $\displaystyle\leq C_{k}\sum_{A\text{ atom of }\mathscr{F}_{n}}b_{n}(q,|\nu_{s}|,A,y)$ for some constants $C_{k}$ and $0<q<1$ depending only on $k$ with $b_{n}$ as in (3.4). Therefore, $G_{t}\subset L_{t/C_{k}}$ with $L_{u}=\Big{\\{}y\in I^{d}:\limsup_{n}\sum_{A\text{ atom of }\mathscr{F}_{n}}b_{n}(q,|\nu_{s}|,A,y)>u\Big{\\}}.$ We apply Lemma 5.1 below (with $Y=I^{d}$) to the measure $\theta=|\nu_{s}|$ on $\mathscr{A}$ and the set $D$ to get, for any $u>0$, the estimate $|L_{u}|=|D^{c}\cap L_{u}|+|D\cap L_{u}|\leq\varepsilon+C\varepsilon/u$. Since this is true for any $\varepsilon>0$, we obtain $|L_{u}|=0$ for any $u>0$. Thus, $|\\{y\in I^{d}:\limsup_{n}\|P_{n}\nu_{s}(y)\|>0\\}|=\Big{|}\bigcup_{r=1}^{\infty}G_{1/r}\Big{|}\leq\Big{|}\bigcup_{r=1}^{\infty}L_{1/(C_{k}r)}\Big{|}=\lim_{r\to\infty}|L_{1/(C_{k}r)}|=0,$ which completes the proof of the theorem. ∎ ## 5\. The convergence theorem for arbitrary filtrations Now we discuss the necessary modifications in the proof of Theorem 4.1 when the interval filtration $(\mathscr{F}_{n})$ on $I^{d}$ is allowed to be arbitrary. Assume for some Banach space $X$ with RNP, $(g_{n})$ is an $X$-valued martingale spline sequence adapted to $(\mathscr{F}_{n})$ with $\sup_{n}\|g_{n}\|_{L^{1}_{X}}<\infty$. Part I of the proof of Theorem 4.1 does not use the density of the filtration $(\mathscr{F}_{n})$ in $I^{d}$, which means that we get an operator $T:\cup_{n}S_{n}\to X$ and a finite measure $\mu$ on $\bar{I}^{d}$ satisfying (5.1) $\|Tf\|\leq\int_{\bar{I}^{d}}\limsup_{s\to y}|f(s)|\,\mathrm{d}\mu(y),\qquad f\in\cup_{n}S_{n}.$ The operator $T$ is then extended continuously to the completion $W$ of $\cup_{n}S_{n}$ w.r.t. the norm on the right hand side of (5.1). With the aid of this operator, the martingale spline sequence $(g_{n})$ is written as (5.2) $g_{n}=\sum_{i}(TN_{n,i})N_{n,i}^{*}.$ We distinguish the analysis of the convergence of $g_{n}(y)$ depending on in which coordinate direction the filtration $(\mathscr{F}_{n})$ is dense at the point $y$. To this end, for $\ell=1,\ldots,d$, we define $\Delta_{n}^{\ell}\subset\bar{I}$ to be the set of all endpoints of atoms in the $\sigma$-algebra $\mathscr{F}_{n}^{\ell}$. Next, let $U^{\ell}$ be the complement (in $\bar{I}$) of the set of all accumulation points of $\cup_{n}\Delta_{n}^{\ell}$. Note that $U^{\ell}$ is open (in $\bar{I}$), thus it can be written as a countable union of disjoint open intervals $(U^{\ell}_{j})_{j}$. Let $\displaystyle B_{j}^{\ell}=\\{a\in\partial U_{j}^{\ell}:\text{ there is no sequence of points in $U_{j}^{\ell}\cap(\cup_{n}\Delta_{n}^{\ell})$ }\text{that converges to $a$}\\}$ and define $V_{j}^{\ell}=U_{j}^{\ell}\cup B_{j}^{\ell}$ and $V^{\ell}:=\cup_{j}V_{j}^{\ell}$. ###### Lemma 5.1. Let $(\mathscr{F}_{n})$ be an interval filtration on $I^{d}$ and let $\theta$ be a non-negative, finitely additive and finite measure on $\mathscr{A}$. For $\varepsilon>0$, let $D\in\mathscr{A}$ with $\theta(D)\leq\varepsilon$ and $L_{t}:=\Big{\\{}x\in I^{d}:\limsup_{n}\sum_{A\text{ atom of }\mathscr{F}_{n}}b_{n}(q,\theta,A,x)>t\Big{\\}}.$ Then, there exists a finite constant $C$, depending only on $q$ and on $d$ so that $|D\cap L_{t}\cap Y|\leq\frac{C\varepsilon}{t},\qquad t>0,$ with $Y=(V^{1})^{c}\times\cdots\times(V^{d})^{c}$. ###### Proof. We shrink the set $D$ properly to then apply Corollary 3.4. This is done as follows. Since $D\in\mathscr{A}$, we can write it as $D=\cup_{j=1}^{L}Q_{j}$ for disjoint atoms $(Q_{j})$ of some $\sigma$-algebra $\mathscr{F}_{n}$. For each $j$, we have $Q_{j}=Q_{j}^{1}\times\cdots\times Q_{j}^{d}$ for some intervals $Q_{j}^{\ell}$, $\ell=1,\ldots,d$. Assume without restriction that for all $\ell\in\\{1,\ldots,d\\}$, the interior of the interval $Q_{j}^{\ell}$ contains at least two points from $(V^{\ell})^{c}$, since otherwise we would have $|Q_{j}\cap L_{t}\cap Y|\leq|Q_{j}\cap Y|=0$. Fix $\ell\in\\{1,\ldots,d\\}$, set $\eta=\varepsilon/(tL|I|^{d-1}d)$ and define the interval $J^{\ell}\subset Q_{j}^{\ell}$ such that 1. (1) $Q_{j}^{\ell}\setminus J^{\ell}$ has two connected components and in each one there exists a point of $(V^{\ell})^{c}$ that has positive distance to $J^{\ell}$ and to the boundary of $Q_{j}^{\ell}$, 2. (2) $|(Q_{j}^{\ell}\setminus J^{\ell})\cap(V^{\ell})^{c}|\leq\eta$. This is possible since $Q_{j}^{\ell}\cap(V^{\ell})^{c}$ contains at least two points. Then, set $Q_{j}^{\prime}=J^{1}\times\cdots\times J^{d}$ and $B=\cup_{j=1}^{L}Q_{j}^{\prime}$ and we get, by the choice of $\eta$, (5.3) $|D\cap L_{t}\cap Y|\leq|(D\setminus B)\cap Y|+|B\cap L_{t}|\leq\varepsilon/t+|B\cap L_{t}|.$ Choose the positive integer $R$ sufficiently large so that $\sum_{s>R}(s+1)^{d-1}q^{s/2}\theta(I^{d})\leq\varepsilon$. Then, there exists an integer $K$ so that $A_{K,R}(B)\subset D$, which is true by construction of $B$. Apply now Corollary 3.4 to get $|B\cap L_{t}|\leq C\varepsilon/t$, which together with (5.3) implies the assertion of the lemma. ∎ ###### Remark. Assume that in Lemma 5.1, the measure $\theta$ is a $\sigma$-additive Borel measure on $\bar{I}^{d}$ and replace the term $\theta(A)$ in the definition (3.4) of $b_{n}$ by the term $\theta(\overline{A})$ with the closure $\overline{A}$ of $A$ in $\bar{I}^{d}$. Then, the assertion of Lemma 5.1 still holds with an additional factor of $2^{d}$ on the constant $C$, since the same is true for Corollary 3.4. For a point $y=(y^{1},\ldots,y^{d})\in I^{d}$, each coordinate $y^{\ell}$ is either contained in some set $V^{\ell}_{j_{\ell}}$ or in $(V^{\ell})^{c}$. After rearranging the coordinates, we assume that $y\in F$, where $F=F^{1}\times\cdots\times F^{d}$ with $F^{\ell}=V_{j_{\ell}}^{\ell}$ if $\ell\leq s$ and $F^{\ell}=(V^{\ell})^{c}$ if $\ell>s$ for some $s\in\\{0,\ldots,d\\}$. We want to split $g_{n}(y)=\sum_{i}(TN_{n,i})N_{n,i}^{*}(y)$ into the parts where $T$ acts on functions restricted to the set $F_{\delta}$ for $\delta\in\\{0,1\\}^{d}$ with $F_{\delta}=E^{1}\times\cdots\times E^{d}$ where $E^{\ell}=F^{\ell}$ if $\delta_{\ell}=0$ and $E^{\ell}=(F^{\ell})^{c}$ if $\delta_{\ell}=1$. In order to construct elements in $W$ that correspond to the functions $N_{n,i}\mathbbm{1}_{F_{\delta}}$, we need the following lemma. ###### Lemma 5.2. For any $\ell\in\\{1,\ldots,d\\}$, let $f\in S^{k_{\ell}}(\mathscr{F}_{n}^{\ell})$ for some $n$. For any interval $V_{j}^{\ell}$, there exists a sequence $(h_{m})$ of functions $h_{m}\in S^{k_{\ell}}(\mathscr{F}_{m}^{\ell})$, open intervals $O_{m}$ and closed intervals $C_{m}$ (both in $\bar{I}$) satisfying 1. (1) $O_{m}\to V_{j}^{\ell}$ as $m\to\infty$, 2. (2) $\operatorname{supp}h_{m}\subset O_{m}$, 3. (3) $h_{m}\equiv f$ on $C_{m}\cap I$, 4. (4) The closure of $O_{m}\setminus C_{m}$ converges to the empty set as $m\to\infty$. ###### Proof. Without loss of generality, assume that $f=N_{n,i}^{\ell}$ for some integer $i$. For $m\geq n$, we can write $N_{n,i}^{\ell}=\sum_{r}\lambda_{m,r}N_{m,r}^{\ell},$ where the absolute value of each coefficient $\lambda_{m,r}$ is $\leq 1$. Set $h_{m}=\sum_{r\in\Lambda_{m}}\lambda_{m,r}N_{m,r}^{\ell},$ where the set $\Lambda_{m}$ is defined to contain precisely those indices $r$ so that the support of $N_{m,r}^{\ell}$ intersects $V_{j}^{\ell}$ but the (Euclidean) distance between the support of $N_{m,r}^{\ell}$ and $\partial U_{j}^{\ell}\setminus B_{j}^{\ell}$ is positive. The function $h_{m}$ is then contained in $S^{k_{\ell}}(\mathscr{F}_{m}^{\ell})$ and satisfies $|h_{m}|\leq 1$. With this setting, the support of $h_{m}$ is contained in $O_{m}$ for some open interval $O_{m}$ and $h_{m}\equiv N_{n,i}^{\ell}$ on some closed interval $C_{m}\subset O_{m}$. Since the endpoints of $V_{j}^{\ell}$ are accumulation points of $\cup_{n}\Delta_{n}^{\ell}$ or endpoints of $I$, the intervals $O_{m}$ and $C_{m}$ can be chosen to satisfy items (1) and (4). ∎ Let now $(h_{j,m}^{\ell})_{m}$ be the sequence of functions from Lemma 5.2 corresponding to a function $f^{\ell}\in S^{k_{\ell}}(\mathscr{F}_{n_{\ell}}^{\ell})$ for some positive integer $n_{\ell}$ and the set $V_{j}^{\ell}$. 1. (1) If $E^{\ell}=V_{j_{\ell}}^{\ell}$, set $h_{m}^{\ell}=h_{j_{\ell},m}^{\ell}$. 2. (2) If $E^{\ell}=(V_{j_{\ell}}^{\ell})^{c}$, set $h_{m}^{\ell}=f^{\ell}-h_{j_{\ell},m}^{\ell}$. 3. (3) If $E^{\ell}=V^{\ell}$, set $h_{m}^{\ell}=\sum_{j=1}^{m}h_{j,K_{m}}^{\ell}$. 4. (4) If $E^{\ell}=(V^{\ell})^{c}$, set $h_{m}^{\ell}=f^{\ell}-\sum_{j=1}^{m}h_{j,K_{m}}^{\ell}$. Then, define $h_{m}=h_{m}^{1}\otimes\cdots\otimes h_{m}^{d}$. Since $\cup_{j\geq m}\overline{V_{j}^{\ell}}$ tends to the empty set as $m\to\infty$ for each $\ell$, and due to the properties guaranteed by Lemma 5.2 of the functions $(h_{m}^{\ell})$, if $K_{m}$ is chosen sufficiently large, $h_{m}\in S_{K_{m}}$ is Cauchy in the Banach space $W$ and its limit will be denoted by $(f^{1}I_{E^{1}})\otimes\cdots\otimes(f^{d}I_{E^{d}})$. If $f^{\ell}=N_{n,i_{\ell}}^{\ell}$ is some B-spline function for all $\ell$ and some positive integer $n$, we will also write $N_{n,i}I_{F_{\delta}}$ for this limit in $W$, which (by (5.1)) satisfies (5.4) $\|T(N_{n,i}I_{F_{\delta}})\|=\lim_{m}\|Th_{m}\|\leq\liminf_{m}\mu(\overline{\operatorname{supp}h_{m}})\leq\mu\big{(}F_{\delta}\cap\overline{\operatorname{supp}N_{n,i}}\big{)}.$ This construction allows us to decompose the martingale spline sequence $g_{n}$ into $g_{n}=\sum_{\delta\in\\{0,1\\}^{d}}g_{n,\delta},\qquad\text{with }g_{n,\delta}=\sum_{i}T(N_{n,i}I_{F_{\delta}})N_{n,i}^{*}\text{ for }\delta\in\\{0,1\\}^{d}.$ We treat the sequence $(g_{n,\delta})_{n}$ for each fixed $\delta\in\\{0,1\\}^{d}$ separately. Case 1: We begin by considering the case where one of the first $s$ coordinates of $\delta$ equals one. Without restriction assume that the first coordinate of $\delta$ equals one. Write $N_{n,i}^{*}=N_{n,i_{1}}^{1*}\otimes N_{n,i_{2}}^{>1*}$, with $i=(i_{1},i_{2})$ for an integer $i_{1}$ and a $(d-1)$-tuple of integers $i_{2}$, thus $g_{n,\delta}$ can be written as (5.5) $g_{n,\delta}(y_{1},y_{2})=\sum_{i_{2}}\Big{(}\sum_{i_{1}}T(N_{n,i}I_{F_{\delta}})N_{n,i_{1}}^{1*}(y_{1})\Big{)}N_{n,i_{2}}^{>1*}(y_{2}),\qquad(y_{1},y_{2})\in F.$ Fix $y_{1}\in U_{j_{1}}^{1}$ and $t>0$. Let $\varepsilon>0$ and denote by $A_{n}^{1}(y_{1})$ the atom in $\mathscr{F}_{n}^{1}$ that contains the point $y_{1}$. Then, $\beta:=\inf_{n}|A_{n}^{1}(y_{1})|>0$ since $U_{j_{1}}^{1}$ does not contain accumulation points of $\cup_{n}\Delta_{n}^{1}$. Choose an open interval $O\supseteq V_{j_{1}}^{1}$ so that $\mu\big{(}(O\setminus V_{j_{1}}^{1})\times\bar{I}^{d-1}\big{)}\leq\varepsilon\mu(\bar{I}^{d})$. Then, choose $M$ sufficiently large so that for all $n\geq M$, we have $q^{d_{n}(A_{n}^{1}(y_{1}),B_{n})}\leq\varepsilon$ for all atoms $B_{n}$ in $\mathscr{F}_{n}^{1}$ with $B_{n}\cap O^{c}\neq\emptyset$. This is possible since the endpoints of $V_{j_{1}}^{1}$ are accumulation points of $\cup_{n}\Delta_{n}^{1}$. Split the sum over $i_{1}$ in (5.5) into indices $i_{1}$ so that $\overline{\operatorname{supp}N_{n,i_{1}}^{1}}\subseteq O$ and its complement and use the geometric decay estimate (3.2) for the dual B-splines $N_{n,i_{1}}^{1*}$ and $N_{n,i_{2}}^{>1*}$ and estimate (5.4). With the measures $\theta_{1}(A)=\frac{1}{\beta}\mu\big{(}(O\setminus V_{j_{1}}^{1})\times A\big{)},\qquad\theta_{2}(A)=\frac{\varepsilon}{\beta}\mu\big{(}\bar{I}\times A\big{)}$ satisfying $\max\\{\theta_{1}(\bar{I}^{d-1}),\theta_{2}(\bar{I}^{d-1})\\}\leq\varepsilon\mu(\bar{I}^{d})/\beta$ and the notation $\mathscr{F}_{n}^{>1}=\mathscr{F}_{n}^{2}\otimes\cdots\otimes\mathscr{F}_{n}^{d}$, we then obtain for $n\geq M$ $\|g_{n,\delta}(y_{1},y_{2})\|\leq C\sum_{\text{$A$ atom of $\mathscr{F}_{n}^{>1}$}}\big{(}b_{n}(q,\theta_{1},A,y_{2})+b_{n}(q,\theta_{2},A,y_{2})\big{)},$ where the expressions $b_{n}(q,\theta,A,y_{2})$ are defined as in (3.4), but with $\theta(A)$ replaced by $\theta(\overline{A})$. Here and in the following, the letter $C$ denotes a constant that depends only on $k,d,q$ and that may change from line to line. Then, applying Corollary 3.4 (using also the remark succeeding it) in dimension $d-1$ with $B=D=\bar{I}^{d-1}$, we estimate $\displaystyle|\\{y_{2}:\limsup_{n}\|g_{n,\delta}(y_{1},y_{2})\|>t\\}|$ $\displaystyle\leq C\frac{\varepsilon\mu(\bar{I}^{d})}{t\beta}.$ We have this inequality for any $\varepsilon>0$, which implies $|\\{y_{2}:\limsup_{n}\|g_{n,\delta}(y_{1},y_{2})\|>t\\}|=0$. As this is true for any $t>0$ and any $y_{1}\in U_{j_{1}}^{1}$, we get $g_{n,\delta}\to 0$ almost everywhere on $F$. Case 2: Next, consider the case where $\delta\neq 0$ but the first $s$ coordinates of $\delta$ equal $0$. Write $N_{n,i}^{*}=N_{n,i_{1}}^{\leq s*}\otimes N_{n,i_{2}}^{>s*}$ where $i=(i_{1},i_{2})$ for an $s$-tuple of integers $i_{1}$ and a $(d-s)$-tuple of integers $i_{2}$, thus, $g_{n,\delta}$ can be written as $\displaystyle g_{n,\delta}(y_{1},y_{2})$ $\displaystyle=\sum_{i_{2}}\Big{(}\sum_{i_{1}}T(N_{n,i}I_{F_{\delta}})N_{n,i_{1}}^{\leq s*}(y_{1})\Big{)}N_{n,i_{2}}^{>s*}(y_{2}),\qquad(y_{1},y_{2})\in F.$ Denote by $A_{m}^{\leq s}(y_{1})$ the atom $A$ in $\mathscr{F}_{m}^{1}\otimes\cdots\otimes\mathscr{F}_{m}^{s}$ with $y_{1}\in A$. If we fix $y_{1}\in U_{j_{1}}^{1}\times\cdots\times U_{j_{s}}^{s}$, we know that $\beta:=\inf_{m}|A_{m}^{\leq s}(y_{1})|>0$. Next, define $Y=F^{s+1}\times\cdots\times F^{d}$ and $Z=E^{s+1}\times\cdots\times E^{d}$. Moreover, let $\mathscr{F}_{m}^{>s}=\mathscr{F}_{m}^{s+1}\otimes\cdots\otimes\mathscr{F}_{m}^{d}$ and define the measure $\theta(A)=\mu(\bar{I}^{s}\times(A\cap Z))$. Observe that $\theta(Y)=0$, since $E^{\ell}\cap F^{\ell}=\emptyset$ for some $\ell>s$ by the form of $\delta$. Using estimate (3.2) for the dual B-spline functions and estimate (5.4) bounding the operator $T$ in terms of $\mu$, $\displaystyle\|g_{n,\delta}(y_{1},y_{2})\|$ $\displaystyle\leq\frac{C}{\beta}\sum_{\text{$A$ atom of $\mathscr{F}_{n}^{>s}$}}b_{n}(q,\theta,A,y_{2})$ where the expression $b_{n}(q,\theta,A,y_{2})$ is defined as in (3.4), but with $\theta(A)$ replaced with $\theta(\overline{A})$. Approximate $Y$ by a sequence of sets $Y_{m}\in\mathscr{F}_{m}^{>s}$ with $Y_{m}\to Y$. Then, for each $\varepsilon>0$, there exists a positive integer $m(\varepsilon)$ with $|Y\setminus Y_{m(\varepsilon)}|\leq\varepsilon$ and $\theta(Y_{m(\varepsilon)})\leq\varepsilon$. For $t>0$, apply Lemma 5.1 (and the remark succeeding it) in dimension $d-s$ with $D=Y_{m(\varepsilon)}$ to deduce (5.6) $|L_{t}\cap Y|\leq|Y_{m(\varepsilon)}\cap L_{t}\cap Y|+|Y\setminus Y_{m(\varepsilon)}|\leq\frac{C\varepsilon}{t}+\varepsilon$ with $L_{t}=\Big{\\{}y_{2}\in I^{d-s}:\limsup_{n}\sum_{\text{ $A$ atom of $\mathscr{F}_{n}^{>s}$}}b_{n}(q,\theta,A,y_{2})>t\Big{\\}}.$ Since (5.6) holds for any $\varepsilon>0$, we obtain $|L_{t}\cap Y|=0$ for any $t>0$, which gives that for any fixed $y_{1}$, $g_{n,\delta}(y_{1},y_{2})$ converges to $0$ a.e. in $y_{2}\in Y$. Summarizing and combining this with Case 1 for $\delta$, we have $g_{n,\delta}\to 0$ a.e. on $F$ as $n\to\infty$ if one of the coordinates of $\delta$ equals $1$. Case 3: It remains to consider the case $g_{n,0}$, i.e. the choice $\delta=0$. For each $\ell\leq s$, the B-splines $(N_{n,r}^{\ell})_{r}$ whose supports intersect $V_{j_{\ell}}^{\ell}$ can be indexed in such a way that for each fixed $r$, the function $N_{n,r}^{\ell}\mathbbm{1}_{V_{j_{\ell}}^{\ell}}$ converges uniformly to a function $\bar{N}_{r}^{\ell}$ as $n\to\infty$ (cf. [5, Section 4]). This is the case since the interior of $V_{j_{\ell}}^{\ell}$ does not contain any accumulation points of $\cup_{n}\Delta_{n}^{\ell}$. Depending on whether the endpoints of $V_{j_{\ell}}^{\ell}$ can be approximated from inside of $V_{j_{\ell}}^{\ell}$ by points in $\cup_{n}\Delta_{n}^{\ell}$, there are different possibilities for the index set $\Lambda^{\ell}$ of the functions $(\bar{N}_{r}^{\ell})_{r\in\Lambda^{\ell}}$. It can either be finite, infinite on one side or bi-infinite. We have the following biorthogonal functions to $(\bar{N}_{r}^{\ell})_{r}$ that admit the same geometric decay estimate (3.2) than the dual B-spline functions $N_{n,r}^{\ell*}$. This result is similar to [5, Lemma 4.2]. ###### Lemma 5.3. Let $\ell\in\\{1,\ldots,d\\}$. For each $r\in\Lambda^{\ell}$, the sequence $N_{n,r}^{\ell*}$ converges uniformly on each atom of $\mathscr{A}^{\ell}=\cup_{n}\mathscr{F}_{n}^{\ell}$ contained in $V_{j_{\ell}}^{\ell}$ to some function $\bar{N}_{r}^{\ell*}$ satisfying the estimate (5.7) $|\bar{N}_{r}^{\ell*}(y)|\leq C\frac{q^{d(A(y),E_{r})}}{|\operatorname{re}(A(y)\cup E_{r})|},\qquad y\in U_{j_{\ell}}^{\ell},$ denoting by $A(y)$ the atom of $\mathscr{A}^{\ell}$ that contains the point $y$, by $E_{r}$ the support of $\bar{N}_{r}^{\ell}$ and by $d(A(y),E_{r})$ the number of atoms in $\mathscr{A}^{\ell}$ between $A(y)$ and $E_{r}$. ###### Proof. Fix the index $r\in\Lambda^{\ell}$, the point $y\in U_{j_{\ell}}^{\ell}$ and $\varepsilon>0$. Since $r\in\Lambda^{\ell}$ is fixed, the support $E_{n,r}$ of $N_{n,r}^{\ell}$ intersects $U_{j_{\ell}}^{\ell}$ for some index $n$ and we know that $\beta=\inf_{m}|E_{m,r}|>0$. Additionally, set $\gamma=|A(y)|$. Without restriction, we assume that $\beta,\gamma\leq 1$. Next, we choose $L$ sufficiently large so that $Lq^{L}\leq\varepsilon\beta\gamma$ and, for any positive integer $n$, $d_{n}(A_{n}(y),E_{n,r})\leq L$. Moreover, choose an open interval $O\supseteq V_{j_{\ell}}^{\ell}$ satisfying $|O\setminus U_{j_{\ell}}^{\ell}|\leq\varepsilon\beta\gamma/L$. Based on that, choose $M$ sufficiently large so that each of the intervals $(\inf O,y)$ and $(y,\sup O)$ contains at least $L$ points of $\Delta_{M}^{\ell}$ and so that, for indices $\nu$ with $|\nu-r|\leq 2L$ we have (5.8) $\|N_{n,\nu}^{\ell}-N_{m,\nu}^{\ell}\|_{L^{\infty}(U_{j_{\ell}}^{\ell})}\leq\varepsilon\beta\gamma/L,\qquad m,n\geq M.$ For $n\geq m\geq M$, expand the function $N_{m,r}^{\ell*}$ in the basis $(N_{n,\nu}^{\ell*})_{\nu}$ as (5.9) $N_{m,r}^{\ell*}=\sum_{\nu}\alpha_{r\nu}N_{n,\nu}^{\ell*}.$ The coefficients $\alpha_{r\nu}$ are bounded by a constant independently of $r,\nu$ and $m,n$ as we will now see. To this end, we use the geometric decay inequality (3.2) for the dual B-spline functions $N_{m,r}^{\ell*}$ to obtain $\displaystyle|\alpha_{r\nu}|$ $\displaystyle=\Big{|}\int_{I}N_{m,r}^{\ell*}N_{n,\nu}^{\ell}\,\mathrm{d}\lambda\Big{|}\leq C\sum_{\text{ $A$ atom of $\mathscr{F}_{m}^{\ell}$}}\frac{q^{d_{m}(A,E_{m,r})}}{|\operatorname{re}(A\cup E_{m,r})|}\int_{A}N_{n,\nu}^{\ell}\,\mathrm{d}\lambda$ $\displaystyle\leq C\sum_{\text{ $A$ atom of $\mathscr{F}_{m}^{\ell}$}}\frac{q^{d_{m}(A,E_{m,r})}}{|\operatorname{re}(A\cup E_{m,r})|}|A|\leq C\sum_{\text{ $A$ atom of $\mathscr{F}_{m}^{\ell}$}}q^{d_{m}(A,E_{m,r})}\leq C.$ Denoting $f_{\nu}=N_{m,\nu}^{\ell}-N_{n,\nu}^{\ell}$, whose absolute value is bounded by $1$, $\displaystyle\delta_{r\nu}$ $\displaystyle=\int_{I}N_{m,r}^{\ell*}N_{m,\nu}^{\ell}\,\mathrm{d}\lambda=\int_{I}N_{m,r}^{\ell*}N_{n,\nu}^{\ell}\,\mathrm{d}\lambda+\int_{I}N_{m,r}^{\ell*}f_{\nu}\,\mathrm{d}\lambda=\alpha_{r\nu}+\int_{I}N_{n,r}^{\ell*}f_{\nu}\,\mathrm{d}\lambda.$ For indices $\nu$ with $|\nu-r|\leq 2L$, we now estimate this last integral, by decomposing it into the integrals $I_{1},I_{2},I_{3}$ over $U_{j_{\ell}}^{\ell}$, $O\setminus U_{j_{\ell}}^{\ell}$ and $O^{c}$, respectively. By estimate (5.8) and the fact that the integral of $N_{n,r}^{\ell*}$ is smaller than a constant $C$ by (3.2), the integral $|I_{1}|$ can be bounded by $C\varepsilon\beta\gamma/L$. For the second integral, we use the fact that the integrand is bounded by $C/\beta$ and the measure estimate for $O\setminus U_{j_{\ell}}^{\ell}$ to deduce $|I_{2}|\leq C\varepsilon\gamma/L$. For the remaining integral $I_{3}$, we note that on $O^{c}$, the function $N_{n,r}^{\ell*}$ is bounded by $Cq^{L}/\beta$, which, together with estimate (3.2) and the choice of $L$ implies $|I_{3}|\leq C\varepsilon\gamma/L$. Summarizing, this implies $|\alpha_{r\nu}-\delta_{r\nu}|\leq C\varepsilon\gamma/L,\qquad|\nu-r|\leq 2L.$ This can be used to estimate the difference between $N_{m,r}^{\ell*}$ and $N_{n,r}^{\ell*}$ for $n\geq m\geq M$ pointwise as follows $\displaystyle|N_{m,r}^{\ell*}(y)-N_{n,r}^{\ell*}(y)|$ $\displaystyle=\Big{|}\sum_{\nu}(\alpha_{r\nu}-\delta_{r\nu})N_{n,\nu}^{\ell*}(y)\Big{|}\leq C\varepsilon+\sum_{\nu:|\nu-r|>2L}|\alpha_{r\nu}N_{n,\nu}^{\ell*}(y)|,$ by using the bound $|N_{n,\nu}^{\ell*}(y)|\leq C/\gamma$. By the choice of $L$, the inequality $|\nu-r|>2L$ implies $d_{n}(A_{n}(y),E_{n,\nu})>L$ and thus, the geometric decay estimate for $N_{n,\nu}^{\ell*}$, the boundedness of $\alpha_{r\nu}$ and the choice of $L$ implies that the latter sum is bounded by $C\varepsilon$. This, in turn, implies the estimate $|N_{m,r}^{\ell*}(y)-N_{n,r}^{\ell*}(y)|\leq C\varepsilon$ and thus the convergence of $N_{n,r}^{\ell*}(y)$, which is uniform in $A(y)$ since all the estimates above only depend on $A(y)$ and not on the particular point $y$. Now, estimate (5.7) follows from the corresponding estimate of $N_{n,r}^{\ell*}$ by letting $n\to\infty$. ∎ Write $F=Z\times Y$ with $Z=F^{1}\times\cdots\times F^{s}=V_{j_{1}}^{1}\times\cdots\times V_{j_{s}}^{s}$ and $Y=F^{s+1}\times\cdots\times F^{d}$. For an $s$-tuple of integers $i_{1}=(r_{1},\ldots,r_{s})$ and a $(d-s)$-tuple of integers $i_{2}=(r_{s+1},\ldots,r_{d})$, set $N_{m,i_{1}}^{\leq s}I_{Z}=N_{m,r_{1}}^{1}I_{F^{1}}\otimes\cdots\otimes N_{m,r_{s}}^{s}I_{F^{s}}$ and $N_{n,i_{2}}^{>s}I_{Y}=N_{n,r_{s+1}}^{s+1}I_{F^{s+1}}\otimes\cdots\otimes N_{n,r_{d}}^{d}I_{F^{d}}$. The uniform convergence of $N_{m,r_{\ell}}^{\ell}\mathbbm{1}_{V_{j_{\ell}}^{\ell}}$ to $\bar{N}_{r_{\ell}}^{\ell}$ for $\ell\leq s$ as $m\to\infty$ implies that for fixed $n$ and $i_{1}$, the sequence $(N_{m,i_{1}}^{\leq s}I_{Z}\otimes N_{n,i_{2}}^{>s}I_{Y})$ converges in $W$ to some element as $m\to\infty$, which we denote by $\bar{N}_{i_{1}}^{\leq s}\otimes N_{n,i_{2}}^{>s}I_{Y}$. By the continuity of $T$, we also have $T(N_{m,i_{1}}^{\leq s}I_{Z}\otimes N_{n,i_{2}}^{>s}I_{Y})\to T(\bar{N}_{i_{1}}^{\leq s}\otimes N_{n,i_{2}}^{>s}I_{Y})$ in $X$ as $m\to\infty$. Using the expressions $T(\bar{N}_{i_{1}}^{\leq s}\otimes N_{n,i_{2}}^{>s}I_{Y})$ and the dual functions $\bar{N}_{i_{1}}^{\leq s*}=\bar{N}_{r_{1}}^{1*}\otimes\cdots\otimes\bar{N}_{r_{s}}^{s*}$ to $\bar{N}_{i_{1}}^{\leq s}$ given by Lemma 5.3, define (5.10) $u_{n}=\sum_{i_{1},i_{2}}T(\bar{N}_{i_{1}}^{\leq s}\otimes N_{n,i_{2}}^{>s}I_{Y})(\bar{N}_{i_{1}}^{\leq s*}\otimes N_{n,i_{2}}^{>s*}).$ Next, we show that the sequence $(g_{n,0})$ and the sequence $(u_{n})$ have the same a.e. limit. Indeed, for fixed $y_{1}\in U_{j_{1}}^{1}\times\cdots\times U_{j_{s}}^{s}$, the difference of those two functions has the form $\displaystyle g_{n,0}(y_{1},\cdot)-u_{n}(y_{1},\cdot)$ $\displaystyle=\sum_{i_{2}}N_{n,i_{2}}^{>s*}\Big{[}\sum_{i_{1}}T\big{(}(N_{n,i_{1}}^{\leq s}I_{Z}-\bar{N}_{i_{1}}^{\leq s})\otimes N_{n,i_{2}}^{>s}I_{Y}\big{)}N_{n,i_{1}}^{\leq s*}(y_{1})$ $\displaystyle\qquad+\sum_{i_{1}}T(\bar{N}_{i_{1}}^{\leq s}\otimes N_{n,i_{2}}^{>s}I_{Y})\big{(}N_{n,i_{1}}^{\leq s*}(y_{1})-\bar{N}_{i_{1}}^{\leq s}(y_{1})\big{)}\Big{]}.$ Denote $\mathscr{F}_{n}^{>s}=\mathscr{F}_{n}^{s+1}\otimes\cdots\otimes\mathscr{F}_{n}^{d}$. Using (3.2) for $N_{n,i_{2}}^{>s*}$ and $N_{n,i_{1}}^{\leq s*}$, Lemma 5.3, the uniform boundedness and the localized support of $\bar{N}_{i_{1}}^{\leq s}$, and the bound (5.4) of the operator $T$ in terms of the measure $\mu$, we obtain for all $\varepsilon>0$ an index $M$ so that for $n\geq M$ (5.11) $\|g_{n,0}(y_{1},y_{2})-u_{n}(y_{1},y_{2})\|\leq\sum_{\text{$A$ atom of $\mathscr{F}_{n}^{>s}$}}b_{n}(q,\theta,A,y_{2}),$ where $\theta$ is the measure given by $\theta(A)=\varepsilon\mu\big{(}\bar{I}^{s}\times(A\cap Y)\big{)}$ and the expression $b_{n}(q,\theta,A,y_{2})$ is defined as in (3.4), but with $\theta(A)$ replaced with $\theta(\overline{A})$. By Corollary 3.4 (and the remark succeeding it) with $B=D=\bar{I}^{d-s}$ we obtain $|L_{t}|\leq C\theta(\bar{I}^{d-s})/t\leq C\varepsilon\mu(\bar{I}^{d})/t$ with $L_{t}=\\{y_{2}\in I^{d-s}:\limsup_{n}\sum_{\text{$A$ atom of $\mathscr{F}_{n}^{>s}$}}b_{n}(q,\theta,A,y_{2})>t\\}.$ This implies, using also (5.11), $|\\{y_{2}\in I^{d-s}:\limsup_{n}\|g_{n,0}(y_{1},y_{2})-u_{n}(y_{1},y_{2})\|>t\\}|=0$ for any $t>0$, i.e., $g_{n,0}$ and $u_{n}$ have the same a.e. limit on $F$. Therefore, in order to identify the a.e. limit of $(g_{n,0})$ on $F$ (which, by Cases $1$ and $2$, is also the a.e. limit of $(g_{n})$), we identify the a.e. limit of $(u_{n})$. Similar to Part II in the proof of Theorem 4.1, we want to construct, for each $i_{1}$, a vector measure $\nu_{i_{1}}$ on $\mathscr{A}^{>s}=\cup_{n}\mathscr{F}_{n}^{>s}$ based on the expressions $T(\bar{N}_{i_{1}}^{\leq s}\otimes N_{n,i_{2}}^{>s}I_{Y})$. The aim is to have, for each B-spline function $N_{n,i_{2}}^{>s}$, the representation (5.12) $\int N_{n,i_{2}}^{>s}\,\mathrm{d}\nu_{i_{1}}=T(\bar{N}_{i_{1}}^{\leq s}\otimes N_{n,i_{2}}^{>s}I_{Y}).$ To this end, let $Q=\prod_{\ell=s+1}^{d}(a_{\ell},b_{\ell}]$ be an atom of the $\sigma$-algebra $\mathscr{F}_{n}^{>s}$ for some positive integer $n$. For the definition of the measure $\nu_{i_{1}}(Q)$, we approximate the characteristic function $\mathbbm{1}_{Q}$ of $Q$ by spline functions $f_{m}^{s+1}\otimes\cdots\otimes f_{m}^{d}$ contained in $\cup_{j}\big{(}S^{k_{s+1}}(\mathscr{F}_{j}^{s+1})\otimes\cdots\otimes S^{k_{d}}(\mathscr{F}_{j}^{d})\big{)}$, which will be done as follows. Let $\ell\in\\{s+1,\ldots,d\\}$. If the order of the polynomials $k_{\ell}$ in direction $\ell$ equals $1$ (piecewise constant case), we set $f_{m}^{\ell}=\mathbbm{1}_{(a_{\ell},b_{\ell}]}$ for $m\geq n$, which satisfies $f_{m}^{\ell}\in S^{k_{\ell}}(\mathscr{F}_{m}^{\ell})$. If $k_{\ell}>1$, we apply the following construction of the approximation $f_{m}^{\ell}$ of the characteristic function of the interval $(a_{\ell},b_{\ell}]$. If $a_{\ell}$ is contained in the countable set $\cup_{j}\partial U_{j}^{\ell}$ and if $a_{\ell}$ is not an endpoint of $I$, we choose $c\in(a_{\ell},a_{\ell}+1/m)$ that is not contained in $\cup_{j}\partial U_{j}^{\ell}$. Otherwise, set $c=a_{\ell}$. Similarly, if $b_{\ell}$ is contained in $\cup_{j}\partial U_{j}^{\ell}$ and if $b_{\ell}$ is not an endpoint of $I$, we choose $d\in(b_{\ell},b_{\ell}+1/m)$ that is not contained in $\cup_{j}\partial U_{j}^{\ell}$. Otherwise, set $d=b_{\ell}$. Put $J(x)=\begin{cases}V_{j}^{\ell},&\text{if }x\in U_{j}^{\ell},\\\ \emptyset,&\text{otherwise,}\end{cases}$ and define the interval $J=\Big{(}(c,d]\setminus J(c)\Big{)}\cup J(d),$ which has the property that $J\cap(V^{\ell})^{c}=(c,d]\cap(V^{\ell})^{c}$. We then choose a closed interval $C$ and an open interval $O$ (both in $I$) with $C\subseteq J\subseteq O$ and the property $|O\setminus C|\leq 1/m$. The sets $C$ and $O$ are chosen so that as many endpoints of $C$ and $O$ coincide with the corresponding endpoints of $(c,d]$ as possible. Then, let $f_{m}^{\ell}\in\cup_{j}S^{k_{\ell}}(\mathscr{F}_{j}^{\ell})$ be a non- negative function that is bounded by $1$ and satisfies $\operatorname{supp}f_{m}^{\ell}\subseteq O\qquad\text{and}\qquad f_{m}^{\ell}\equiv 1\text{ on }C\cap I.$ This is possible since if $c$ or $d$ are endpoints of $J$, they are contained in $\big{(}\cup_{j}\overline{U_{j}^{\ell}})^{c}$ and thus can be approximated from both sides with grid points $\cup_{j}\Delta_{j}^{\ell}$. Otherwise, the endpoints of $J$ are also endpoints of some set $V_{j}^{\ell}$, which are accumulation points of $\cup_{j}\Delta_{j}^{\ell}$ as well. Then, define $f_{m}=f_{m}^{s+1}\otimes\cdots\otimes f_{m}^{d}$ which gives, for each index $i_{1}$, a Cauchy sequence $\bar{N}_{i_{1}}^{\leq s}\otimes f_{m}I_{Y}$ in $W$. Its limit will be written as $\bar{N}_{i_{1}}^{\leq s}\otimes(I_{Q}\cdot I_{Y})$. Then, continuing in a similar fashion as in Part II of the proof of Theorem 4.1, we make sense of the expression $T\big{(}\bar{N}_{i_{1}}^{\leq s}\otimes(I_{A}\cdot I_{Y})\big{)}$ for any $A\in\mathscr{A}^{>s}$ and define the measure $\nu_{i_{1}}(A)=T\big{(}\bar{N}_{i_{1}}^{\leq s}\otimes(I_{A}\cdot I_{Y})\big{)}$ for $A\in\mathscr{A}^{>s}$ whose total variation satisfies $|\nu_{i_{1}}|(I^{d-s})\leq 2^{d-s}\mu(\overline{\operatorname{supp}\bar{N}_{i_{1}}^{\leq s}}\times Y)$. Additionally, for any B-spline function $N_{n,i_{2}}^{>s}$, we have equation (5.12). Now, as in Part III of the proof of Theorem 4.1, denoting by $w_{i_{1}}$ the Radon-Nikodým density of the absolutely continuous part of $\nu_{i_{1}}$ with respect to Lebesgue measure $\lambda^{d-s}$, $u_{n}(y_{1},y_{2})=\sum_{i_{1}}\bar{N}_{i_{1}}^{\leq s*}(y_{1})(P_{n}^{>s}\nu_{i_{1}})(y_{2})\to g(y_{1},y_{2}):=\sum_{i_{1}}\bar{N}_{i_{1}}^{\leq s*}(y_{1})w_{i_{1}}(y_{2})$ as $n\to\infty$ for almost every $(y_{1},y_{2})\in F$. Using the estimate from Lemma 5.3 for $\bar{N}_{i_{1}}^{\leq s*}$ and the above estimate for the total variation of the measures $\nu_{i_{1}}$, we obtain that $\|g\|_{L^{1}_{X}(F)}\leq C\cdot\mu(F)$. Thus, we have proven the following theorem: ###### Theorem 5.4. Let $(\mathscr{F}_{n})$ be an interval filtration on $I^{d}$ and let $X$ be a Banach space with RNP. Let $(g_{n})$ be an $X$-valued martingale spline sequence adapted to $(\mathscr{F}_{n})$ with $\sup_{n}\|g_{n}\|_{L^{1}_{X}}<\infty$. Then, there exists $g\in L^{1}_{X}(I^{d})$ so that $g_{n}\to g$ almost everywhere with respect to Lebesgue measure $\lambda^{d}$. ###### Remark. Employing the notation developed in this section, we emphasize that the pointwise limit $g$ has the explicit representation $g(y_{1},y_{2}):=\sum_{i_{1}}\bar{N}_{i_{1}}^{\leq s*}(y_{1})w_{i_{1}}(y_{2}),\qquad(y_{1},y_{2})\in F,$ where $\bar{N}_{i_{1}}^{\leq s*}$ are the functions given by Lemma 5.3 corresponding to $F^{1}\times\cdots\times F^{s}=V_{j_{1}}^{1}\times\cdots\times V_{j_{s}}^{s}$ and the function $w_{i_{1}}$ is the Radon-Nikodým density of the absolutely continuous part (w.r.t Lebesgue measure $\lambda^{d-s}$) of the measure $A\mapsto T\big{(}\bar{N}_{i_{1}}^{\leq s}\otimes(I_{A}\cdot I_{Y})\big{)}$ with $Y=(V^{s+1})^{c}\times\cdots\times(V^{d})^{c}$. ### Acknowledgements The author is supported by the Austrian Science Fund FWF, project P32342. ## References * [1] M. de Guzmán. An inequality for the Hardy-Littlewood maximal operator with respect to a product of differentiation bases. Studia Math., 49:185–194, 1973/74. * [2] J. Diestel and J. J. Uhl, Jr. Vector measures. American Mathematical Society, Providence, R.I., 1977. With a foreword by B. J. Pettis, Mathematical Surveys, No. 15. * [3] M. v. Golitschek. On the $L_{\infty}$-norm of the orthogonal projector onto splines. A short proof of A. Shadrin’s theorem. J. Approx. Theory, 181:30–42, 2014. * [4] B. Jessen, J. Marcinkiewicz, and A. Zygmund. Note on the differentiability of multiple integrals. Fundamenta Mathematicae, 25(1):217–234, 1935. * [5] P. F. X. Müller and M. Passenbrunner. Almost everywhere convergence of spline sequences. Israel J. Math., 240(1):149–177, 2020. * [6] J. Neveu. Discrete-parameter martingales. North-Holland Publishing Co., Amsterdam-Oxford; American Elsevier Publishing Co., Inc., New York, revised edition, 1975. Translated from the French by T. P. Speed, North-Holland Mathematical Library, Vol. 10. * [7] M. Passenbrunner. Spline characterizations of the Radon-Nikodým property. Proc. Amer. Math. Soc., 148(2):811–824, 2020. * [8] M. Passenbrunner and J. Prochno. On almost everywhere convergence of tensor product spline projections. Michigan Math. J., 68(1):3–17, 2019. * [9] M. Passenbrunner and A. Shadrin. On almost everywhere convergence of orthogonal spline projections with arbitrary knots. J. Approx. Theory, 180:77–89, 2014. * [10] S. Saks. On the strong derivatives of functions of intervals. Fundamenta Mathematicae, 25(1):235–252, 1935. * [11] L. L. Schumaker. Spline functions: basic theory. Cambridge Mathematical Library. Cambridge University Press, Cambridge, third edition, 2007. * [12] A. Shadrin. The $L_{\infty}$-norm of the $L_{2}$-spline projector is bounded independently of the knot sequence: a proof of de Boor’s conjecture. Acta Math., 187(1):59–137, 2001.
# Nonstationary Stochastic Multiarmed Bandits: UCB Policies and Minimax Regret ††thanks: This work was supported by NSF Award IIS-1734272 Lai Wei Vaibhav Srivastava L. Wei and V. Srivastava are with the Department of Electrical and Computer Engineering. Michigan State University, East Lansing, MI 48823 USA. e-mail<EMAIL_ADDRESS>e-mail<EMAIL_ADDRESS> ###### Abstract We study the nonstationary stochastic Multi-Armed Bandit (MAB) problem in which the distribution of rewards associated with each arm are assumed to be time-varying and the total variation in the expected rewards is subject to a variation budget. The regret of a policy is defined by the difference in the expected cumulative rewards obtained using the policy and using an oracle that selects the arm with the maximum mean reward at each time. We characterize the performance of the proposed policies in terms of the worst-case regret, which is the supremum of the regret over the set of reward distribution sequences satisfying the variation budget. We extend Upper-Confidence Bound (UCB)-based policies with three different approaches, namely, periodic resetting, sliding observation window and discount factor and show that they are order-optimal with respect to the minimax regret, i.e., the minimum worst-case regret achieved by any policy. We also relax the sub-Gaussian assumption on reward distributions and develop robust versions the proposed polices that can handle heavy-tailed reward distributions and maintain their performance guarantees. ###### Index Terms: Nonstationary multiarmed bandit, variation budget, minimax regret, upper- confidence bound, heavy-tailed distributions. ## I Introduction Uncertainty and nonstationarity of the environment are two of the major barriers in decision-making problems across scientific disciplines, including engineering, economics, social science, neuroscience, and ecology. An efficient strategy in such environments requires balancing several tradeoffs, including _exploration-versus-exploitation_ , i.e., choosing between the most informative and the empirically most rewarding alternatives, and _remembering- versus-forgetting_ , i.e., using more but possibly outdated information or using less but recent information. The stochastic MAB problem is a canonical formulation of the exploration- versus-exploitation tradeoff. In an MAB problem, an agent selects one from $K$ options at each time and receives a reward associated with it. The reward sequence at each option is assumed to be an unknown i.i.d random process. The MAB formulation has been applied in many scientific and technological areas. For example, it is used for opportunistic spectrum access in communication networks, wherein the arm models the availability of a channel [1, 2]. In MAB formulation of online learning for demand response[3, 4], an aggregator calls upon a subset of users (arms) who have an unknown response to the request to reduce their loads. MAB formulation has also been used in robotic foraging and surveillance [5, 6, 7, 8] and acoustic relay positioning for underwater communication [9], wherein the information gain at different sites is modeled as rewards from arms. Besides, contextual bandits are widely used in recommender systems [10, 11], wherein the acceptation of a recommendation corresponds to the rewards from an arm. The stationarity assumption in classic MAB problems limits their utility in these applications since channel usage, robot working environment and people’s preference are inherently uncertain and evolving. In this paper, we relax this assumption and study non-stationary stochastic MAB problems. Robbins [12] formulated the objective of the stochastic MAB problem as minimizing the _regret_ , that is, the loss in expected cumulative rewards caused by failing to select the best arm every time. In their seminal work, Lai and Robbins [13], followed by Burnetas and Katehakis [14], established a logarithm _problem-dependent_ asymptotic lower bound on the regret achieved by any policy, which has a leading constant determined by the underlying reward distributions. A general method of constructing UCB rules for parametric families of reward distributions is also presented in [13], and the associated policy is shown to attain the logarithm lower bound. Several subsequent UCB- based algorithms [15, 16] with efficient finite time performance have been proposed. The adversarial MAB [17] is a paradigmatic nonstationary problem. In this model, the bounded reward sequence at each arm is arbitrary. The performance of an policy is evaluated using the _weak regret_ , which is the difference in the cumulated reward of a policy compared against the best single action policy. A $\Omega(\sqrt{KT})$ lower bound on the weak regret and a near- optimal policy Exp$3$ is also presented in [17]. While being able to capture nonstationarity, the generality of the reward model in adversarial MAB makes the investigation of globally optimal policies very challenging. The nonstationary stochastic MAB can be viewed as a compromise between stationary stochastic MAB and adversarial MAB. It maintains the stochastic nature of the reward sequence while allowing some degree of nonstationarity in reward distributions. Instead of the weak regret analyzed in adversarial MAB, a strong notion of regret defined with respect to the best arm at each time step is studied in these problems. A broadly studied nonstationary problem is _piecewise stationary_ MAB, wherein the reward distributions are piecewise stationary. To deal with the remembering-versus-forgetting tradeoff, the idea of using discount factor to compute the UCB index is proposed in [18]. Garivier and Moulines [19] present and analyze Discounted UCB (D-UCB) and Sliding-Window UCB (SW-UCB), in which they compute the UCB using discounted sampling history and recent sampling history, respectively. They pointed out that if the number of change points $N_{T}$ is available, both algorithms can be tuned to achieve regret close to the $\Omega(\sqrt{KN_{T}T})$ regret lower bound. In our earlier work [20], the near optimal regret is achieved using deterministic sequencing of explore and exploit with limited memory. Other works handle the change of reward distributions in an adaptive manner by adopting change point detection techniques [21, 22, 23, 24, 25]. A more general nonstationary problem is studied in [26], wherein the cumulative maximum variation in mean rewards is subject to a variation budget $V_{T}$. Additionally, the authors in [26] establish a $\Omega((KV_{T})^{\frac{1}{3}}T^{\frac{2}{3}})$ minimax regret lower bound and propose the Rexp$3$ policy. In their subsequent work [27], they tune Exp$3$.S policy from [17] to achieve near optimal worst-case regret. Discounted Thomson Sampling (DTS) [28] has also been shown to have good experimental performance within this general framework. However, we are not aware of any analytic regret bounds for the DTS algorithm. In this paper, we follow the more general nonstationary stochastic MAB formulation in [26] and design UCB-based policies that achieve efficient performance in environments with sub-Gaussian as well as heavy-tailed rewards. We focus on UCB-based policies instead of EXP$3$-type policies because EXP$3$-type policies require bounded rewards and have large variance in cumulative rewards [17]. Additionally, by using robust mean estimator, UCB- based policies for light-tailed rewards can be extended to handle heavy-tailed reward distributions, which exist in many domains such as social networks [29] and financial markets [30]. The major contributions of this work are: * • Assuming the variation density $V_{T}/T$ is known, we extend MOSS [31] to design Resetting MOSS (R-MOSS) and Sliding-Window MOSS (SW-MOSS). Also, we show D-UCB can be tuned to solve the problem. * • With rigorous analysis, we show that R-MOSS and SW-MOSS achieve the exact order-optimal minimax regret and D-UCB achieves near-optimal worst-case regret. * • We relax the bounded or sub-Gaussian assumption on the rewards required by Rexp$3$ and SW-UCB and design policies robust to heavy-tailed rewards. We show the theoretical guarantees on the worst-case regret can be maintained by the robust policies. The remainder of the paper is organized as follows. We formulate nonstationary stochastic MAB with variation budget in Section II and review some preliminaries in Section III. In Section IV, we present and analyze three UCB policies: R-MOSS, SW-MOSS and D-UCB. We present and analyze algorithms for nonstationary heavy-tailed bandit in Section V. We complement the theoretical results with numerical illustrations in Section VI and conclude this work in Section VII. ## II Problem Formulation We consider a nonstationary stochastic MAB problem with $K$ arms and a horizon length $T$. Let $\mathcal{K}\mathrel{\mathop{\mathchar 58\relax}}=\\{1,\dots,K\\}$ be the set of arms and $\mathcal{T}\mathrel{\mathop{\mathchar 58\relax}}=\\{1,\dots,T\\}$ be the sequence of time slots. The reward sequence $\left\\{X_{t}^{k}\right\\}_{t\in\mathcal{T}}$ for each arm $k\in\mathcal{K}$ is composed of independent samples from potentially time-varying probability distribution function sequence $f_{\mathcal{T}}^{k}\mathrel{\mathop{\mathchar 58\relax}}=\left\\{f_{t}^{k}(x)\right\\}_{t\in\mathcal{T}}$. We refer to the set $\mathcal{F}_{T}^{\mathcal{K}}=\left\\{f_{\mathcal{T}}^{k}\;|\;k\in\mathcal{K}\right\\}$ containing reward distribution sequences at all arms as the _environment_. Let $\mu_{t}^{k}=\mathbb{E}[X_{t}^{k}]$. Then, the _total variation_ of $\mathcal{F}_{T}^{\mathcal{K}}$ is defined by $v\big{(}\mathcal{F}_{T}^{\mathcal{K}}\big{)}\mathrel{\mathop{\mathchar 58\relax}}=\sum_{t=1}^{T-1}\sup_{k\in\mathcal{K}}\>\mathinner{\\!\left\lvert\mu_{t+1}^{k}-\mu_{t}^{k}\right\rvert},$ (1) which captures the non-stationarity of the environment. We focus on the class of non-stationary environments that have the total variation within a _variation budget_ $V_{T}\geq 0$ which is defined by $\mathcal{E}(V_{T},T,K)\mathrel{\mathop{\mathchar 58\relax}}=\big{\\{}\mathcal{F}_{T}^{\mathcal{K}}\;|\;v\big{(}\mathcal{F}_{T}^{\mathcal{K}}\big{)}\leq V_{T}\big{\\}}.$ At each time slot $t\in\mathcal{T}$, a decision-making agent selects an arm $\varphi_{t}\in\mathcal{K}$ and receives an associated random reward $X_{t}^{\varphi_{t}}$. The objective is to maximize the expected value of the _cumulative reward_ $S_{T}\mathrel{\mathop{\mathchar 58\relax}}=\sum_{t=1}^{T}X_{t}^{\varphi_{t}}$. We assume that $\varphi_{t}$ is selected based upon past observations $\\{X_{s}^{\varphi_{s}},\varphi_{s}\\}_{s=1}^{t-1}$ following some policy $\rho$. Specifically, $\rho$ determines the conditional distribution $\mathbb{P}^{\rho}\left(\varphi_{t}=k\;|\;\\{X_{s}^{\varphi_{s}},\varphi_{s}\\}_{s=1}^{t-1}\right)$ at each time $t\in\\{1,\dots,T-1\\}$. If $\mathbb{P}^{\rho}\left(\cdot\right)$ takes binary values, we call $\rho$ deterministic; otherwise, it is called stochastic. Let the expected reward from the best arm at time $t$ be $\mu_{t}^{*}=\max_{k\in\mathcal{K}}\mu_{t}^{k}.$ Then, maximizing the expected cumulative reward is equivalent to minimizing the _regret_ defined by $R^{\rho}_{T}\mathrel{\mathop{\mathchar 58\relax}}=\sum_{t=1}^{T}\mu_{t}^{*}-\mathbb{E}^{\rho}[S_{T}]=\mathbb{E}^{\rho}\Bigg{[}\sum_{t=1}^{T}\mu_{t}^{*}-\mu_{t}^{\varphi_{t}}\Bigg{]},$ where the expectation is with respect to different realization of $\varphi_{t}$ that depends on obtained rewards through policy $\rho$. Note that the performance of a policy $\rho$ differs with different $\mathcal{F}_{T}^{\mathcal{K}}\in\mathcal{E}(V_{T},T,K)$. For a fixed variation budget $V_{T}$ and a policy $\rho$, the _worst-case regret_ is the regret with respect to the worst possible choice of environment, i.e., $R_{\textup{worst}}^{\rho}(V_{T},T,K)=\sup_{\mathcal{F}_{T}^{\mathcal{K}}\in\mathcal{E}(V_{T},T,K)}\>R_{T}^{\rho}.$ In this paper, we aim at designing policies to minimize the worst-case regret. The optimal worst-case regret achieved by any policy is called the _minimax regret_ , and is defined by $\inf_{\rho}\sup_{\mathcal{F}_{T}^{\mathcal{K}}\in\mathcal{E}(V_{T},T,K)}\>R_{T}^{\rho}.$ We will study the nonstationary MAB problem under the following two classes of reward distributions: ###### Assumption 1 (Sub-Gaussian reward). For any $k\in\mathcal{K}$ and any $t\in\mathcal{T}$, distribution $f_{t}^{k}(x)$ is $1/2$ sub-Gaussian, i.e., $\forall\lambda\in\mathbb{R}\mathrel{\mathop{\mathchar 58\relax}}\mathbb{E}\left[\exp(\lambda(X_{t}^{k}-\mu))\right]\leq\exp\left(\frac{\lambda^{2}}{8}\right).$ Moreover, for any arm $k\in\mathcal{K}$ and any time $t\in\mathcal{T}$, $\mathbb{E}\left[X_{t}^{k}\right]\in[a,a+b]$, where $a\in\mathbb{R}$ and $b>0$. ###### Assumption 2 (Heavy-tailed reward). For any arm $k\in\mathcal{K}$ and any time $t\in\mathcal{T}$, $\mathbb{E}\left[(X_{t}^{k})^{2}\right]\leq 1$. ## III Preliminaries In this section, we review existing minimax regret lower bounds and minimax policies from literature. These results apply to both sub-Gaussian and heavy- tailed rewards. The discussion is made first for $V_{T}=0$. Then, we show how the minimax regret lower bound for $V_{T}=0$ can be extended to establish the minimax regret lower bound for $V_{T}>0$. To this end, we review two UCB algorithms for the stationary stochastic MAB problem: UCB1 and MOSS. In the later sections, they are extended to design a variety of policies to match with the minimax regret lower bound for $V_{T}>0$. ### III-A Lower Bound for Minimax Regret when $V_{T}=0$ In the setting of $V_{T}=0$, for each arm $k\in\mathcal{K}$, $\mu_{t}^{k}$ is identical for all $t\in\mathcal{T}$. In stationary stochastic MAB problems, the rewards from each arm $k\in\mathcal{K}$ are independent and identically distributed, so they belong to the environment set $\mathcal{E}(0,T,K)$. According to [32], if $V_{T}=0$, the minimax regret is no smaller than $1/20\sqrt{KT}$. This result is closely related to the standard logarithmic lower bound on regret for stationary stochastic MAB problems as discussed below. Consider a scenario in which there is a unique best arm and all other arms have identical mean rewards such that the gap between optimal and suboptimal mean rewards is $\Delta$. From [33], for such a stationary stochastic MAB problem $R_{T}^{\rho}\geq C_{1}\frac{K}{\Delta}\ln\Big{(}\frac{T\Delta^{2}}{K}\Big{)}+C_{2}\frac{K}{\Delta},$ (2) for any policy $\rho$, where $C_{1}$ and $C_{2}$ are some positive constants. It needs to be noted that for $\Delta=\sqrt{K/T}$, the above lower bound becomes $C_{2}\sqrt{KT}$, which matches with the lower bound $1/20\sqrt{KT}$. ### III-B Lower Bound for Minimax Regret when $V_{T}>0$ In the setting of $V_{T}>0$, we recall here the minimax regret lower bound for nonstationary stochastic MAB problems. ###### Lemma 1 (Minimax Lower Bound: $V_{T}>0$ [26]). For the non-stationary MAB problem with $K$ arms, time horizon $T$ and variation budget $V_{T}\in[1/K,T/K]$, $\inf_{\rho}\sup_{\mathcal{F}_{T}^{\mathcal{K}}\in\mathcal{E}(V_{T},T,K)}R_{T}^{\rho}\geq C(KV_{T})^{\frac{1}{3}}T^{\frac{2}{3}},$ where $C\in\mathbb{R}_{>0}$ is some constant. To understand this lower bound, consider the following non-stationary environment. The horizon $\mathcal{T}$ is partitioned into epochs of length $\tau=\big{\lceil}{K^{\frac{1}{3}}({T/V_{T}})^{\frac{2}{3}}}\big{\rceil}$. In each epoch, the reward distribution sequences are stationary and all the arms have identical mean rewards except for the unique best arm. Let the gap in the mean be $\Delta=\sqrt{K/\tau}$. The index of the best arm switches at the end of each epoch following some unknown rule. So, the total variation is no greater than $\Delta T/\tau$, which satisfies the variation budget $V_{T}$. Besides, for any policy $\rho$, we know from (2) that worst case regret in each epoch is no less than $C_{2}\sqrt{K\tau}$. Summing up the regret over all the epochs, minimax regret is lower bounded by $T/\tau\times C_{2}\sqrt{K\tau}$, which is consistent with Lemma 1. ### III-C UCB Algorithms in Stationary Environments The family of UCB algorithms uses the principle called optimism in the face of uncertainty. In these policies, at each time slot, a UCB index which is a statistical index composed of both mean reward estimate and the associated uncertainty measure is computed at each arm, and the arm with the maximum UCB is picked. Within the family of UCB algorithms, two state-of-the-art algorithms for the stationary stochastic MAB problems are UCB$1$ [15] and MOSS [31]. Let $n_{k}(t)$ be the number of times arm $k$ is sampled until time $t-1$, and $\hat{\mu}_{k,n_{k}(t)}$ be the associated empirical mean. Then, UCB$1$ computes the UCB index for each arm $k$ at time $t$ as $g_{k,t}^{\textup{UCB1}}=\hat{\mu}_{k,n_{k}(t)}+\sqrt{\frac{2\ln t}{n_{k}(t)}}.$ It has been proved in [15] that, for the stationary stochastic MAB problem, UCB1 satisfies $R_{T}^{\textup{UCB1}}\leq 8\sum_{k\mathrel{\mathop{\mathchar 58\relax}}\Delta_{k}>0}\frac{\ln T}{\Delta_{k}}+\left(1+\frac{\pi^{2}}{3}\right)\sum_{k=1}^{K}\Delta_{k},$ where $\Delta_{k}$ is the difference in the mean rewards from arm $k$ and the best arm. In [31], a simple variant of this result is given by selecting values for $\Delta_{k}$ to maximize the upper bound, resulting in $\sup_{\mathcal{F}_{T}^{\mathcal{K}}\in\mathcal{E}(0,T,K)}R_{T}^{\textup{UCB1}}\leq 10\sqrt{(K-1)T(\ln T)}.$ Comparing this result with the lower bound on the minimax regret discussed in Section III-A, there exists an extra factor $\sqrt{\ln T}$. This issue has been resolved by the MOSS algorithm. With prior knowledge of horizon length $T$, and the UCB index for MOSS is expressed as $g_{k,t}^{\textup{MOSS}}=\hat{\mu}_{k,n_{k}(t)}+\sqrt{\frac{\max\Big{(}\ln\Big{(}\frac{T}{Kn_{k}(t)}\Big{)},0\Big{)}}{n_{k}(t)}}.$ We now recall the worst-case regret upper bound for MOSS. ###### Lemma 2 (Worst-case regret upper bound for MOSS [31]). For the stationary stochastic MAB problem ($V_{T}=0$), the worst-case regret of the MOSS algorithm satisfies $\sup_{\mathcal{F}_{T}^{\mathcal{K}}\in\mathcal{E}(0,T,K)}R_{T}^{\text{MOSS}}\leq 49\sqrt{KT}.$ ## IV UCB Algorithms for Sub-Gaussian Nonstationary Stochastic MAB Problems In this section, we extend UCB$1$ and MOSS to design nonstationary UCB policies for scenarios with $V_{T}>0$. Three different techniques are employed, namely periodic resetting, sliding observation window and discount factor, to deal with the remembering-forgetting tradeoff. The proposed algorithms are analyzed to provide guarantees on the worst-case regret. We show their performances match closely with the lower bound in Lemma 1. The following notations are used in later discussions. Let $N=\left\lceil T/\tau\right\rceil$, for some $\tau\in\\{1,\dots,T\\}$, and let $\\{\mathcal{T}_{1},\ldots,\mathcal{T}_{N}\\}$ be a partition of time slots $\mathcal{T}$, where each epoch $\mathcal{T}_{i}$ has length $\tau$ except possibly $\mathcal{T}_{N}$. In particular, $\mathcal{T}_{i}=\Big{\\{}1+(i-1)\tau\>,\ldots,\>\min\left(i\tau,T\right)\Big{\\}},\;i\in\\{1,\dots,N\\}.$ Let the maximum mean reward within $\mathcal{T}_{i}$ be achieved at time $\tau_{i}\in\mathcal{T}_{i}$ and arm $\kappa_{i}$, i.e., $\mu^{\kappa_{i}}_{\tau_{i}}=\max_{t\in\mathcal{T}_{i}}\;\mu_{t}^{*}$. We define the variation within $\mathcal{T}_{i}$ as $v_{i}\mathrel{\mathop{\mathchar 58\relax}}=\sum_{t\in\mathcal{T}_{i}}\>\sup_{k\in\mathcal{K}}\>\mathinner{\\!\left\lvert\mu_{t+1}^{k}-\mu_{t}^{k}\right\rvert},$ where we trivially assign $\mu_{T+1}^{k}=\mu_{T}^{k}$ for all $k\in\mathcal{K}$. Let $\mathbf{1}\left\\{\cdot\right\\}$ denote the indicator function and $\mathinner{\\!\left\lvert\cdot\right\rvert}$ denote the cardinality of the set, if its argument is a set, and the absolute value if its argument is a real number. ### IV-A Resetting MOSS Algorithm Periodic resetting is an effective technique to preserve the freshness and authenticity of the information history. It has been employed in [26] to modify Exp$3$ to design Rexp$3$ policy for nonstationary stochastic MAB problems. We extend this approach to MOSS and propose nonstationary policy Resetting MOSS (R-MOSS). In R-MOSS, after every $\tau$ time slots, the sampling history is erased and MOSS is restarted. The pseudo-code is provided in Algorithm 1 and the performance in terms of the worst-case regret for is established below. Input : $V_{T}\in\mathbb{R}_{\geq 0}$ and $T\in\mathbb{N}$ Set : $\tau=\Big{\lceil}{K^{\frac{1}{3}}\left(T/V_{T}\right)^{\frac{2}{3}}}\Big{\rceil}$ Output : sequence of arm selection 1while _$t\leq T$_ do 2if _$\mod(t,\tau)=0$_ then 3Restart the MOSS policy; Algorithm 1 R-MOSS ###### Theorem 3. For the sub-Gaussian nonstationary MAB problem with $K$ arms, time horizon $T$, variation budget $V_{T}>0$, and $\tau=\Big{\lceil}{K^{\frac{1}{3}}\left(T/V_{T}\right)^{\frac{2}{3}}}\Big{\rceil}$, the worst case regret of R-MOSS satisfies $\sup_{\mathcal{F}_{T}^{\mathcal{K}}\in\mathcal{E}(V_{T},T,K)}R_{T}^{\textup{R-MOSS}}\in\mathcal{O}((KV_{T})^{\frac{1}{3}}T^{\frac{2}{3}}).$ ###### Sketch of the proof. Note that one run of MOSS takes place in each epoch. For epoch $\mathcal{T}_{i}$, define the set of _bad arms_ for R-MOSS by $\mathcal{B}_{i}^{\textup{R}}\mathrel{\mathop{\mathchar 58\relax}}=\left\\{k\in\mathcal{K}\;|\;\mu_{\tau_{i}}^{\kappa_{i}}-\mu_{\tau_{i}}^{k}\geq 2v_{i}\right\\}.$ (3) Notice that for any $t_{1},t_{2}\in\mathcal{T}_{i}$, $\mathinner{\\!\left\lvert\mu_{t_{1}}^{k}-\mu_{t_{2}}^{k}\right\rvert}\leq v_{i},\quad\forall k\in\mathcal{K}.$ (4) Therefore, for any $t\in\mathcal{T}_{i}$, we have $\displaystyle\mu_{t}^{*}-\mu_{t}^{\varphi_{t}}$ $\displaystyle\leq\mu_{\tau_{i}}^{\kappa_{i}}-\mu_{t}^{\varphi_{t}}\leq\mu_{\tau_{i}}^{\kappa_{i}}-\mu_{\tau_{i}}^{\varphi_{t}}+v_{i}.$ Then, the regret from $\mathcal{T}_{i}$ can be bounded as the following, $\displaystyle\mathbb{E}\bigg{[}\sum_{t\in\mathcal{T}_{i}}\mu_{t}^{*}-\mu_{t}^{\varphi_{t}}\bigg{]}$ $\displaystyle\leq\mathinner{\\!\left\lvert\mathcal{T}_{i}\right\rvert}v_{i}+\mathbb{E}\bigg{[}\sum_{t\in\mathcal{T}_{i}}\mu_{\tau_{i}}^{\kappa_{i}}-\mu_{\tau_{i}}^{\varphi_{t}}\bigg{]}$ $\displaystyle\leq 3\mathinner{\\!\left\lvert\mathcal{T}_{i}\right\rvert}v_{i}+S_{i},$ (5) where $\displaystyle S_{i}=\mathbb{E}\bigg{[}\sum_{t\in\mathcal{T}_{i}}\sum_{k\in\mathcal{B}_{i}^{\textup{R}}}\mathbf{1}\left\\{\varphi_{t}=k\right\\}\left(\mu_{\tau_{i}}^{\kappa_{i}}-\mu_{\tau_{i}}^{\varphi_{t}}-2v_{i}\right)\bigg{]}$. Now, we have decoupled the problem, enabling us to the generalize the analysis of MOSS in stationary environment [31] to bound $S_{i}$. We will only specify the generalization steps and skip the details for brevity. First notice inequality (4) indicates that for any $k\in\mathcal{B}_{i}^{\textup{R}}$ and any $t\in\mathcal{T}_{i}$, $\mu_{t}^{\kappa_{i}}\geq\mu_{\tau_{i}}^{\kappa_{i}}-v_{i}\text{ and }\mu_{t}^{k}\leq\mu_{\tau_{i}}^{k}+v_{i}.$ So, at any $t\in\mathcal{T}_{i}$, $\hat{\mu}_{{\kappa_{i}},n_{\kappa_{i}}(t)}$ concentrate around a value no smaller than $\mu_{\tau_{i}}^{\kappa_{i}}-v_{i}$, and $\hat{\mu}_{k,n_{k}(t)}$ concentrate around a value no greater than $\mu_{\tau_{i}}^{k}+v_{i}$ for any $k\in B_{i}^{\textup{R}}$. Also $\mu_{\tau_{i}}^{\kappa_{i}}-v_{i}\geq\mu_{\tau_{i}}^{k}+v_{i}$ due to the definition in (3). In the analysis of MOSS in stationary environment [31], the UCB of each suboptimal arm is compared with the best arm and each selection of suboptimal arm $k$ contribute $\Delta_{k}$ in regret. Here, we can apply a similar analysis by comparing the UCB of each arm $k\in B_{i}^{\textup{R}}$ with $\kappa_{i}$ and each selection of arm $k\in B_{i}^{\textup{R}}$ contributes $(\mu_{\tau_{i}}^{\kappa_{i}}-v_{i})-(\mu_{\tau_{i}}^{k}+v_{i})$ in $S_{i}$. Accordingly, we borrow the upper bound in Lemma 2 to get $S_{i}\leq 49\sqrt{K\mathinner{\\!\left\lvert\mathcal{T}_{i}\right\rvert}}$. Substituting the upper bound on $S_{i}$ into (5) and summarizing over all the epochs, we conclude that $\sup_{\mathcal{F}_{T}^{\mathcal{K}}\in\mathcal{E}(V_{T},T,K)}R_{T}^{\textup{R-MOSS}}\leq 3\tau V_{T}+\sum_{i=1}^{N}49\sqrt{K\tau},$ which implies the theorem. ∎ The upper bound in Theorem 3 is in the same order as the lower bound in Lemma 1. So, the worst-case regret for R-MOSS is order optimal. ### IV-B Sliding-Window MOSS Algorithm We have shown that periodic resetting coarsely adapts the stationary policy to a nonstationary setting. However, it is inefficient to entirely remove the sampling history at the restarting points and the regret accumulates quickly close to these points. In [19], a sliding observation window is used to erase the outdated information smoothly and more efficiently utilize the information history. The authors proposed the SW-UCB algorithm that intends to solve the MAB problem with piece-wise stationary mean rewards. We show that a similar approach can also deal with the general nonstationary environment with a variation budget. In contrast to SW-UCB, we integrate the sliding window technique with MOSS instead of UCB1 and achieve the order optimal worst-case regret. Let the sliding observation window at time $t$ be $\mathcal{W}_{t}\mathrel{\mathop{\mathchar 58\relax}}=\left\\{\min(1,t-\tau),\ldots,t-1\right\\}$. Then, the associated mean estimator is given by $\hat{\mu}_{n_{k}(t)}^{k}\\!=\\!\frac{1}{n_{k}(t)}\\!\sum_{s\in\mathcal{W}_{t}}\\!\\!X_{s}\mathbf{1}\\{\varphi_{s}=k\\},\,\,n_{k}(t)=\\!\sum_{s\in\mathcal{W}_{t}}\\!\\!\mathbf{1}{\\{\varphi_{s}=k\\}}.$ For each arm $k\in\mathcal{K}$, define the UCB index for SW-MOSS by $g_{t}^{k}=\hat{\mu}_{n_{k}(t)}^{k}+c_{n_{k}(k)},\,\,c_{n_{k}(t)}=\sqrt{\eta\frac{\max\Big{(}\ln\Big{(}\frac{\tau}{Kn_{k}(t)}\Big{)},0\Big{)}}{n_{k}(t)}},$ where $\eta>1/2$ is a tunable parameter. With these notations, SW-MOSS is defined in Algorithm 2. To analyze it, we will use the following concentration bound for sub-Gaussian random variables. Input : $V_{T}\in\mathbb{R}_{>0}$, $T\in\mathbb{N}$ and $\eta>1/2$ Set : $\tau=\Big{\lceil}{K^{\frac{1}{3}}\left(T/V_{T}\right)^{\frac{2}{3}}}\Big{\rceil}$ Output : sequence of arm selection 1Pick each arm once. 2while _$t\leq T$_ do Compute statistics within $\mathcal{W}_{t}=\left\\{\min(1,t-\tau),\ldots,t-1\right\\}$: $\hat{\mu}_{n_{k}(t)}^{k}\\!=\\!\frac{1}{n_{k}(t)}\\!\sum_{s\in\mathcal{W}_{t}}\\!\\!X_{s}\mathbf{1}\\{\varphi_{s}=k\\},\,\,n_{k}(t)=\\!\sum_{s\in\mathcal{W}_{t}}\\!\\!\mathbf{1}{\\{\varphi_{s}=k\\}}$ Pick arm $\displaystyle\varphi_{t}=\arg\max_{k\in\mathcal{K}}\,\hat{\mu}_{n_{k}(t)}^{k}+\sqrt{\eta\frac{\max\Big{(}\ln\Big{(}\frac{\tau}{Kn_{k}(t)}\Big{)},0\Big{)}}{n_{k}(t)}}$; Algorithm 2 SW-MOSS ###### Fact 1 (Maximal Hoeffding inequality[34]). Let $X_{1},\ldots,X_{n}$ be a sequence of independent $1/2$ sub-Gaussian random variables. Define $d_{i}\mathrel{\mathop{\mathchar 58\relax}}=X_{i}-\mu_{i}$, then for any $\delta>0$, $\displaystyle\mathbb{P}\bigg{(}\exists m\in\\{1,\dots,n\\}\mathrel{\mathop{\mathchar 58\relax}}\sum_{i=1}^{m}d_{i}\geq\delta\bigg{)}\leq\exp\left(-{2\delta^{2}}/{n}\right)$ and $\displaystyle\mathbb{P}\bigg{(}\exists m\in\\{1,\dots,n\\}\mathrel{\mathop{\mathchar 58\relax}}\sum_{i=1}^{m}d_{i}\leq-\delta\bigg{)}\leq\exp\left(-{2\delta^{2}}/{n}\right).$ At time $t$, for each arm $k\in\mathcal{K}$ define $M_{t}^{k}\mathrel{\mathop{\mathchar 58\relax}}=\frac{1}{n_{k}(t)}\sum_{s\in\mathcal{W}_{t}}\mu_{s}^{k}\mathbf{1}_{\\{\varphi_{s}=k\\}}.$ Now, we are ready to present concentration bounds for the sliding window empirical mean $\hat{\mu}_{n_{k}(t)}^{k}$. ###### Lemma 4. For any arm $k\in\mathcal{K}$ and any time $t\in\mathcal{T}$, if $\eta>1/2$, for any $x>0$ and $l\geq 1$, the probability of event $A\mathrel{\mathop{\mathchar 58\relax}}=\big{\\{}\hat{\mu}_{n_{k}(t)}^{k}+c_{n_{k}(t)}\leq M_{t}^{k}-x,n_{k}(t)\geq l\big{\\}}$ is no greater than $\frac{(2\eta)^{\frac{3}{2}}}{\ln(2\eta)}\frac{K}{\tau x^{2}}\exp\left(-{x^{2}l}/{\eta}\right).$ (6) The probability of event $B\mathrel{\mathop{\mathchar 58\relax}}=\big{\\{}\hat{\mu}_{n_{k}(t)}^{k}-c_{n_{k}(t)}\geq M_{t}^{k}+x,n_{k}(t)\geq l\big{\\}}$ is also upper bounded by (6). ###### Proof. For any $t\in\mathcal{T}$, let $u_{i}^{kt}$ be the $i$-th time slot when arm $k$ is selected within $\mathcal{W}_{t}$ and let $d_{i}^{kt}=X_{u_{i}^{kt}}^{k}-\mu_{u_{i}^{kt}}^{k}$. Note that $\mathbb{P}\left(A\right)\leq\mathbb{P}\bigg{(}\exists m\in\left\\{l,\ldots,\tau\right\\}\mathrel{\mathop{\mathchar 58\relax}}\frac{1}{m}\sum_{i=1}^{m}d_{i}^{kt}\leq-x-c_{m}\bigg{)},$ Let $a=\sqrt{2\eta}$ such that $a>1$. We now apply a peeling argument [35, Sec 2.2] with geometric grid $a^{s}l<m\leq a^{s+1}l$ over $\left\\{l,\ldots,\tau\right\\}$. Since $c_{m}$ is monotonically decreasing in $m$, $\displaystyle\mathbb{P}\bigg{(}\exists m\in\\{l,\ldots,\tau\\}\mathrel{\mathop{\mathchar 58\relax}}\frac{1}{m}\sum_{i=1}^{m}d_{i}^{kt}\leq-x-c_{m}\bigg{)}$ $\displaystyle\leq$ $\displaystyle\sum_{s\geq 0}\mathbb{P}\bigg{(}\exists m\in[a^{s}l,a^{s+1}l)\mathrel{\mathop{\mathchar 58\relax}}\sum_{i=1}^{m}d_{i}^{kt}\leq-a^{s}l\left(x+c_{a^{s+1}l}\right)\bigg{)}.$ According to Fact 1, the above summand is no greater than $\displaystyle\sum_{s\geq 0}\mathbb{P}\bigg{(}\exists m\in[1,a^{s+1}l)\mathrel{\mathop{\mathchar 58\relax}}\sum_{i=1}^{m}d_{i}^{kt}\leq-a^{s}l\left(x+c_{a^{s+1}l}\right)\bigg{)}$ $\displaystyle\leq$ $\displaystyle\sum_{s\geq 0}\exp\left(-2\frac{a^{2s}l^{2}}{\left\lfloor a^{s+1}l\right\rfloor}\left(x^{2}+c_{a^{s+1}l}^{2}\right)\right)$ $\displaystyle\leq$ $\displaystyle\sum_{s\geq 0}\exp\left(-2a^{s-1}lx^{2}-\frac{2\eta}{a^{2}}\ln\left(\frac{\tau}{Ka^{s+1}l}\right)\right)$ $\displaystyle=$ $\displaystyle\sum_{s\geq 1}\frac{Kla^{s}}{\tau}\exp\left(-2a^{s-2}lx^{2}\right).$ Let $b=2x^{2}l/a^{2}$. It follows that $\displaystyle\sum_{s\geq 1}\frac{Kla^{s}}{\tau}\exp\left(-ba^{s}\right)\leq\frac{Kl}{\tau}\int_{0}^{+\infty}a^{y+1}\exp\big{(}-ba^{y}\big{)}dy$ $\displaystyle=$ $\displaystyle\frac{Kla}{\tau\ln(a)}\int_{1}^{+\infty}\exp(-bz)dz\quad\left(\text{where we set }z=a^{y}\right)$ $\displaystyle=$ $\displaystyle\frac{Klae^{-b}}{\tau b\ln(a)},$ which concludes the bound for the probability of event $A$. By using upper tail bound, similar result exists for event $B$. ∎ We now leverage Lemma 4 to get an upper bound on the worst-case regret for SW- MOSS. ###### Theorem 5. For the nonstationary MAB problem with $K$ arms, time horizon $T$, variation budget $V_{T}>0$ and $\tau=\Big{\lceil}{K^{\frac{1}{3}}\left(T/V_{T}\right)^{\frac{2}{3}}}\Big{\rceil}$, the worst-case regret of SW-MOSS satisfies $\sup_{\mathcal{F}_{T}^{\mathcal{K}}\in\mathcal{E}(V_{T},T,K)}R_{T}^{\text{SW- MOSS}}\in\mathcal{O}((KV_{T})^{\frac{1}{3}}T^{\frac{2}{3}}).$ ###### Proof. The proof consists of the following five steps. Step 1: Recall that $v_{i}$ is the variation within $\mathcal{T}_{i}$. Here, we trivially assign $\mathcal{T}_{0}=\emptyset$ and $v_{0}=0$. Then, for each $i\in\\{1,\dots,N\\}$, let $\Delta_{i}^{k}\mathrel{\mathop{\mathchar 58\relax}}=\mu_{\tau_{i}}^{\kappa_{i}}-\mu_{\tau_{i}}^{k}-2v_{i-1}-2v_{i},\quad\forall k\in\mathcal{K}.$ Define the set of bad arms for SW-MOSS in $\mathcal{T}_{i}$ as $\mathcal{B}_{i}^{\textup{SW}}\mathrel{\mathop{\mathchar 58\relax}}=\left\\{k\in\mathcal{K}\;|\;\Delta_{i}^{k}\geq\epsilon\right\\},$ where we assign $\epsilon=4\sqrt{e\eta K/\tau}$. Step 2: We decouple the regret in this step. For any $t\in\mathcal{T}_{i}$, since $\mathinner{\\!\left\lvert\mu_{t}^{k}-\mu_{\tau_{i}}^{k}\right\rvert}\leq v_{i}$ for any $k\in\mathcal{K}$, it satisfies that $\displaystyle\mu_{t}^{*}-\mu_{t}^{\varphi_{t}}$ $\displaystyle\leq\mu_{\tau_{i}}^{\kappa_{i}}-\mu_{t}^{\varphi_{t}}$ $\displaystyle\leq\mu_{\tau_{i}}^{\kappa_{i}}-\mu_{\tau_{i}}^{\varphi_{t}}+v_{i}$ $\displaystyle\leq\mathbf{1}\left\\{\varphi_{t}\in\mathcal{B}_{i}^{\textup{SW}}\right\\}(\Delta_{i}^{\varphi_{t}}-\epsilon)+2v_{i-1}+3v_{i}+\epsilon.$ Then we get the following inequalities, $\displaystyle\sum_{t\in\mathcal{T}}\mu_{t}^{*}-\mu_{t}^{\varphi_{t}}$ $\displaystyle\leq$ $\displaystyle\sum_{i=1}^{N}\sum_{t\in\mathcal{T}_{i}}\mathbf{1}\left\\{\varphi_{t}\in\mathcal{B}_{i}^{\textup{SW}}\right\\}(\Delta_{i}^{\varphi_{t}}-\epsilon)+2v_{i-1}+3v_{i}+\epsilon$ $\displaystyle\leq$ $\displaystyle 5\tau V_{T}+T\epsilon+\sum_{i=1}^{N}\sum_{t\in\mathcal{T}_{i}}\mathbf{1}\left\\{\varphi_{t}\in\mathcal{B}_{i}^{\textup{SW}}\right\\}(\Delta_{i}^{\varphi_{t}}-\epsilon).$ (7) To continue, we take a decomposition inspired by the analysis of MOSS in [31] below, $\displaystyle\sum_{t\in\mathcal{T}_{i}}\mathbf{1}\left\\{\varphi_{t}\in\mathcal{B}_{i}^{\textup{SW}}\right\\}\left(\Delta_{i}^{\varphi_{t}}-\epsilon\right)$ $\displaystyle\leq$ $\displaystyle\sum_{t\in\mathcal{T}_{i}}\mathbf{1}\bigg{\\{}\varphi_{t}\in\mathcal{B}_{i}^{\textup{SW}},g_{t}^{\kappa_{i}}>M_{t}^{\kappa_{i}}-\frac{\Delta_{i}^{\varphi_{t}}}{4}\bigg{\\}}\Delta_{i}^{\varphi_{t}}$ (8) $\displaystyle+$ $\displaystyle\sum_{t\in\mathcal{T}_{i}}\mathbf{1}\bigg{\\{}\varphi_{t}\in\mathcal{B}_{i}^{\textup{SW}},g_{t}^{\kappa_{i}}\leq M_{t}^{\kappa_{i}}-\frac{\Delta_{i}^{\varphi_{t}}}{4}\bigg{\\}}\left(\Delta_{i}^{\varphi_{t}}-\epsilon\right),$ (9) where summands (8) describes the regret when arm $\kappa_{i}$ is fairly estimated and summand (9) quantifies the regret incurred by underestimating arm $\kappa_{i}$. Step 3: In this step, we bound $\mathbb{E}\left[\eqref{overestimate}\right]$. Since $g_{t}^{\varphi_{t}}\geq g_{t}^{\kappa_{i}}$, $\displaystyle\eqref{overestimate}\leq$ $\displaystyle\sum_{t\in\mathcal{T}_{i}}\mathbf{1}\bigg{\\{}\varphi_{t}\in\mathcal{B}_{i}^{\textup{SW}},g_{t}^{\varphi_{t}}>M_{t}^{\kappa_{i}}-\frac{\Delta_{i}^{\varphi_{t}}}{4}\bigg{\\}}\Delta_{i}^{\varphi_{t}}$ $\displaystyle=$ $\displaystyle\sum_{k\in\mathcal{B}_{i}^{\textup{SW}}}\sum_{t\in\mathcal{T}_{i}}\mathbf{1}\bigg{\\{}\varphi_{t}=k,g_{t}^{k}>M_{t}^{\kappa_{i}}-\frac{\Delta_{i}^{k}}{4}\bigg{\\}}\Delta_{i}^{k}.$ (10) Notice that for any $t\in\mathcal{T}_{i-1}\cup\mathcal{T}_{i}$, $\mathinner{\\!\left\lvert\mu_{t}^{k}-\mu_{\tau_{i}}^{k}\right\rvert}\leq v_{i-1}+v_{i},\quad\forall k\in\mathcal{K}.$ It indicates that an arm $k\in\mathcal{B}_{i}^{\textup{SW}}$ is at least $\Delta_{i}^{k}$ worse in mean reward than arm $\kappa_{i}$ at any time slot $t\in\mathcal{T}_{i-1}\cup\mathcal{T}_{i}$. Since $\mathcal{W}_{t}\subset\mathcal{T}_{i-1}\operatorname{\cup}\mathcal{T}_{i}$, for any $t\in\mathcal{T}_{i}$ $M_{t}^{\kappa_{i}}-M_{t}^{k}\geq\Delta_{i}^{k}\geq\epsilon,\quad\forall k\in\mathcal{B}_{i}^{\textup{SW}}.$ It follows that $\eqref{overestimate2}\leq\sum_{k\in\mathcal{B}_{i}^{\textup{SW}}}\sum_{t\in\mathcal{T}_{i}}\mathbf{1}\bigg{\\{}\varphi_{t}=k,g_{t}^{k}>M_{t}^{k}+\frac{3\Delta_{i}^{k}}{4}\bigg{\\}}\Delta_{i}^{k}.$ (11) Let $t_{s}^{ik}$ be the $s$-th time slot when arm $k$ is selected within $\mathcal{T}_{i}$. Then, for any $k\in\mathcal{B}_{i}^{\textup{SW}}$, $\displaystyle\sum_{t\in\mathcal{T}_{i}}\mathbf{1}{\bigg{\\{}\varphi_{t}=k,g_{t}^{k}>M_{t}^{k}+\frac{3\Delta_{i}^{k}}{4}\bigg{\\}}}$ $\displaystyle=$ $\displaystyle\sum_{s\geq 1}\mathbf{1}{\bigg{\\{}g_{t_{s}^{ik}}^{k}>M_{t_{s}^{ik}}^{k}+\frac{3\Delta_{i}^{k}}{4}\bigg{\\}}}$ $\displaystyle\leq$ $\displaystyle l_{i}^{k}+\sum_{s\geq l_{i}^{k}+1}\mathbf{1}{\bigg{\\{}g_{t_{s}^{ik}}^{k}>M_{t_{s}^{ik}}^{k}+\frac{3\Delta_{i}^{k}}{4}\bigg{\\}}},$ (12) where we set $l_{i}^{k}=\bigg{\lceil}{\eta\Big{(}\frac{4}{\Delta_{i}^{k}}\Big{)}^{2}\ln\left(\frac{\tau}{\eta K}\Big{(}\frac{\Delta_{i}^{k}}{4}\Big{)}^{2}\right)}\bigg{\rceil}$. Since $\Delta_{i}^{k}\geq\epsilon$, for $k\in\mathcal{B}_{i}^{\textup{SW}}$, we have $l_{i}^{k}\geq\Big{\lceil}{\eta\left({4}/{\Delta_{i}^{k}}\right)^{2}\ln\left(\frac{\tau}{\eta K}\left({\epsilon}/{4}\right)^{2}\right)}\Big{\rceil}\geq\eta\left({4}/{\Delta_{i}^{k}}\right)^{2},$ where the second inequality follows by substituting $\epsilon=4\sqrt{e\eta K/\tau}$. Additionally, since $t_{1}^{ik},\ldots,t_{s-1}^{ik}\in\mathcal{W}_{t_{s}^{ik}}$, we get $n_{k}(t_{s}^{ik})\geq s-1$. Furthermore, since $c_{m}$ is monotonically decreasing with $m$, $c_{n_{k}(t_{s}^{k})}\leq c_{l_{i}^{k}}\leq\sqrt{\frac{\eta}{l_{i}^{k}}\ln\left(\frac{\tau}{\eta K}\bigg{(}\frac{\Delta_{i}^{k}}{4}\bigg{)}^{2}\right)}\leq\frac{\Delta_{i}^{k}}{4},$ for $s\geq l_{i}^{k}+1$. Therefore, $\eqref{overestimate_k}\leq l_{i}^{k}+\sum_{s\geq l_{i}^{k}+1}\mathbf{1}{\bigg{\\{}g_{t_{s}^{ik}}^{k}-2c_{n_{k}(t_{s}^{ik})}>M_{t_{s}^{ik}}^{k}+\frac{\Delta_{i}^{k}}{4}\bigg{\\}}}.$ By applying Lemma 4, considering $n_{k}(t_{s}^{ik})\geq s-1$, $\displaystyle\sum_{s\geq l_{i}^{k}+1}\mathbb{P}{\bigg{\\{}g_{t_{s}^{ik}}^{k}-2c_{n_{k}(t_{s}^{ik})}>M_{t_{s}^{ik}}^{k}+\frac{\Delta_{i}^{k}}{4}\bigg{\\}}}$ $\displaystyle\leq$ $\displaystyle\sum_{s\geq l_{i}^{k}}\frac{(2\eta)^{\frac{3}{2}}}{\ln(2\eta)}\frac{K}{\tau}\bigg{(}\frac{4}{\Delta_{i}^{k}}\bigg{)}^{2}\exp\left(-\frac{s}{\eta}\bigg{(}\frac{\Delta_{i}^{k}}{4}\bigg{)}^{2}\right)$ $\displaystyle\leq$ $\displaystyle\int_{l_{i}^{k}-1}^{+\infty}\frac{(2\eta)^{\frac{3}{2}}}{\ln(2\eta)}\frac{K}{\tau}\bigg{(}\frac{4}{\Delta_{i}^{k}}\bigg{)}^{2}\exp\left(-\frac{y}{\eta}\bigg{(}\frac{\Delta_{i}^{k}}{4}\bigg{)}^{2}\right)\,dy$ $\displaystyle\leq$ $\displaystyle\frac{(2\eta)^{\frac{3}{2}}}{\ln(2\eta)}\frac{\eta K}{\tau}\bigg{(}\frac{4}{\Delta_{i}^{k}}\bigg{)}^{4}.$ (13) Let $h(x)=16\eta/x\ln\left({\tau x^{2}}/{16\eta K}\right)$ which achieves maximum at $4e\sqrt{\eta K/\tau}$. Combining (13), (12), (11), and (10), we obtain $\displaystyle\mathbb{E}[\eqref{overestimate}]\leq$ $\displaystyle\sum_{k\in\mathcal{B}_{i}}\frac{(2\eta)^{\frac{3}{2}}}{\ln(2\eta)}\frac{\eta K}{\tau}\frac{256}{\left(\Delta_{i}^{k}\right)^{3}}+l_{i}^{k}\Delta_{i}^{k}$ $\displaystyle\leq$ $\displaystyle\sum_{k\in\mathcal{B}_{i}}\frac{(2\eta)^{\frac{3}{2}}}{\ln(2\eta)}\frac{\eta K}{\tau}\frac{256}{\left(\Delta_{i}^{k}\right)^{3}}+h(\Delta_{i}^{k})+\Delta_{i}^{k}$ $\displaystyle\leq$ $\displaystyle\sum_{k\in\mathcal{B}_{i}}\frac{(2\eta)^{\frac{3}{2}}}{\ln(2\eta)}\frac{\eta K}{\tau}\frac{256}{\epsilon^{3}}+h\left(4e\sqrt{\eta K/\tau}\right)+b$ $\displaystyle\leq$ $\displaystyle\left(\frac{2.6\eta}{\ln(2\eta)}+3\sqrt{\eta}\right)\sqrt{K\tau}+Kb.$ Step 4: In this step, we bound $\mathbb{E}[\eqref{underestimate}]$. When event $\left\\{\varphi_{t}\in\mathcal{B}_{i}^{\textup{SW}},g_{t}^{\kappa_{i}}\leq M_{t}^{\kappa_{i}}-{\Delta_{i}^{\varphi_{t}}}/{4}\right\\}$ happens, we know $\Delta_{i}^{\varphi_{t}}\leq 4M_{t}^{\kappa_{i}}-4g_{t}^{\kappa_{i}}\text{ and }g_{t}^{\kappa_{i}}\leq M_{t}^{\kappa_{i}}-\frac{\epsilon}{4}.$ Thus, we have $\displaystyle\mathbf{1}\bigg{\\{}\varphi_{t}\in\mathcal{B}_{i}^{\textup{SW}},g_{t}^{\kappa_{i}}\leq M_{t}^{\kappa_{i}}-\frac{\Delta_{i}^{\varphi_{t}}}{4}\bigg{\\}}\left(\Delta_{i}^{\varphi_{t}}-\epsilon\right)$ $\displaystyle\leq$ $\displaystyle\mathbf{1}{\left\\{g_{t}^{\kappa_{i}}\leq M_{t}^{\kappa_{i}}-\frac{\epsilon}{4}\right\\}}\times\big{(}4M_{t}^{\kappa_{i}}-4g_{t}^{\kappa_{i}}-\epsilon\big{)}\mathrel{\mathop{\mathchar 58\relax}}=Y$ Since $Y$ is a nonnegative random variable, its expectation can be computed involving only its cumulative density function: $\displaystyle\mathbb{E}\left[Y\right]$ $\displaystyle=\int_{0}^{+\infty}\mathbb{P}\left(Y>x\right)dx$ $\displaystyle\leq\int_{0}^{+\infty}\mathbb{P}\Big{(}4M_{t}^{\kappa_{i}}-4g_{t}^{\kappa_{i}}-\epsilon\geq x\Big{)}dx$ $\displaystyle=\int_{\epsilon}^{+\infty}\mathbb{P}\Big{(}4M_{t}^{\kappa_{i}}-4g_{t}^{\kappa_{i}}>x\Big{)}dx$ $\displaystyle\leq\int_{\epsilon}^{+\infty}\frac{16(2\eta)^{\frac{3}{2}}}{\ln(2\eta)}\frac{K}{\tau x^{2}}dx=\frac{16(2\eta)^{\frac{3}{2}}}{\ln(2\eta)}\frac{K}{\tau\epsilon}.$ Hence, $\mathbb{E}[\eqref{underestimate}]\leq{16(2\eta)^{\frac{3}{2}}}K\mathinner{\\!\left\lvert\mathcal{T}_{i}\right\rvert}/\left(\ln(2\eta)\tau\epsilon\right).$ Step 5: With bounds on $\mathbb{E}\left[\eqref{overestimate}\right]$ and $\mathbb{E}[\eqref{underestimate}]$ from previous steps, $\displaystyle\mathbb{E}[\eqref{regret_sw}]\leq$ $\displaystyle 5\tau V_{T}+T\epsilon+N\left(\frac{2.6\eta}{\ln(2\eta)}+3\sqrt{\eta}\right)\sqrt{K\tau}$ $\displaystyle+NKb+\frac{16(2\eta)^{\frac{3}{2}}}{\ln(2\eta)}\frac{KT}{\tau\epsilon}\leq C(KV_{T})^{\frac{1}{3}}T^{\frac{2}{3}}$ for some constant $C$, which concludes the proof. ∎ We have shown that SW-MOSS also enjoys order optimal worst-case regret. One drawback of the sliding window method is that all sampling history within the observation window needs to be stored. Since window size is selected to be $\tau=\big{\lceil}{K^{\frac{1}{3}}({T}/{V_{T}})^{\frac{2}{3}}}\big{\rceil}$, large memory is needed for large horizon length $T$. The next policy resolves this problem. ### IV-C Discounted UCB Algorithm The discount factor is widely used in estimators to forget old information and put more attention on the recent information. In [19], such an estimation is used together with UCB$1$ to solve the piecewise stationary MAB problem, and the policy designed is called Discounted UCB (D-UCB). Here, we tune D-UCB to work in the nonsationary environment with variation budget $V_{T}$. Specifically, the mean estimator used is discounted empirical average given by $\displaystyle\hat{\mu}_{\gamma,t}^{k}$ $\displaystyle=\frac{1}{n_{\gamma,t}^{k}}\sum_{s=1}^{t-1}\gamma^{t-s}\mathbf{1}\\{\varphi_{s}=k\\}X_{s},$ $\displaystyle n_{\gamma,t}^{k}$ $\displaystyle=\sum_{s=1}^{t-1}\gamma^{t-s}\mathbf{1}\\{\varphi_{s}=k\\},$ where $\gamma=1-{K^{-\frac{1}{3}}({T}/{V_{T}})^{-\frac{2}{3}}}$ is the discount factor. Besides, the UCB is designed as $g_{t}^{k}=\hat{\mu}_{t}^{k}+2c_{t}^{k}$, where $c_{\gamma,t}^{k}=\sqrt{\xi\ln(\tau)/n_{\gamma,t}^{k}}$ for some constant $\xi>1/2$. The pseudo code for D-UCB is reproduced in Algorithm 3. It can be noticed that the memory size is only related to the number of arms, so D-UCB requires small memory. Input : $V_{T}\in\mathbb{R}_{>0}$, $T\in\mathbb{N}$ and $\xi>\frac{1}{2}$ Set : $\gamma=1-{K^{-\frac{1}{3}}({T}/{V_{T}})^{-\frac{2}{3}}}$ Output : sequence of arm selection 1for _$t\in\\{1,\dots,K\\}$ _ do Pick arm $\varphi_{t}=t$ and set $n^{t}\leftarrow\gamma^{K-t}$ and $\hat{\mu}^{t}\leftarrow X_{t}^{t}$; 2while _$t\leq T$_ do Pick arm $\displaystyle\varphi_{t}=\arg\max_{k\in\mathcal{K}}\hat{\mu}^{k}+2\sqrt{\frac{\xi\ln(\tau)}{n^{k}}}$; For each arm $k\in\mathcal{K}$, set $n^{k}\leftarrow\gamma n^{k}$; Set $n^{\varphi_{t}}\leftarrow n^{\varphi_{t}}+1\>\&\>\hat{\mu}^{\varphi_{t}}\leftarrow\hat{\mu}^{\varphi_{t}}+\frac{1}{n^{\varphi_{t}}}(X_{t}^{\varphi_{t}}-\bar{X}^{\varphi_{t}});$ Algorithm 3 D-UCB To proceed the analysis, we review the concentration inequality for discounted empirical average, which is an extension of Chernoff-Hoeffding bound. Let $M_{\gamma,t}^{k}\mathrel{\mathop{\mathchar 58\relax}}=\frac{1}{n_{\gamma,t}^{k}}\sum_{s=1}^{t-1}\gamma^{t-s}\mathbf{1}\\{\varphi_{s}=k\\}\mu_{s}^{k}.$ Then, the following fact is a corollary of [19, Theorem 18]. ###### Fact 2 (A Hoeffding-type inequality for discounted empirical average with a random number of summands). For any $t\in\mathcal{T}$ and for any $k\in\mathcal{K}$, the probability of event $A=\left\\{{\hat{\mu}_{\gamma,t}^{k}-M_{\gamma,t}^{k}}\geq\delta/\sqrt{n_{\gamma,t}^{k}}\right\\}$ is no greater than $\left\lceil\log_{1+\lambda}(\tau)\right\rceil\exp\left(-2\delta^{2}\big{(}1-{\lambda^{2}}/{16}\big{)}\right)$ (14) for any $\delta>0$ and $\lambda>0$. The probability of event $B=\left\\{\hat{\mu}_{\gamma,t}^{k}-M_{\gamma,t}^{k}\leq-\delta/\sqrt{n_{\gamma,t}^{k}}\right\\}$ is also upper bounded by (14). ###### Theorem 6. For the nonstationary MAB problem with $K$ arms, time horizon $T$, variation budget $V_{T}>0$, and $\gamma=1-{K^{-\frac{1}{3}}({T}/{V_{T}})^{-\frac{2}{3}}}$, if $\xi>1/2$, the worst case regret of D-UCB satisfies $\sup_{\mathcal{F}_{T}^{\mathcal{K}}\in\mathcal{E}(V_{T},T,K)}R_{T}^{\textup{D-UCB}}\leq C\ln(T)(KV_{T})^{\frac{1}{3}}T^{\frac{2}{3}}.$ ###### Proof. We establish the theorem in four steps. Step 1: In this step, we analyze $\big{|}{\mu_{\gamma,t}^{k}-M_{\gamma,t}^{k}}\big{|}$ at some time slot $t\in\mathcal{T}_{i}$. Let $\tau^{\prime}={\log_{\gamma}\big{(}(1-\gamma)\xi\ln(\tau)/b^{2}\big{)}}$ and take $t-\tau^{\prime}$ as a dividing point, then we obtain $\displaystyle\mathinner{\\!\left\lvert\mu_{\tau_{i}}^{k}-M_{\gamma,t}^{k}\right\rvert}\leq$ $\displaystyle\frac{1}{n_{\gamma,t}^{k}}\sum_{s=1}^{t-1}\gamma^{t-s}\mathbf{1}\\{\varphi_{s}=k\\}\mathinner{\\!\left\lvert\mu_{\tau_{i}}^{k}-\mu_{s}^{k}\right\rvert}$ $\displaystyle\leq$ $\displaystyle\frac{1}{n_{\gamma,t}^{k}}\sum_{s\leq t-\tau^{\prime}}\gamma^{t-s}\mathbf{1}\\{\varphi_{s}=k\\}\mathinner{\\!\left\lvert\mu_{\tau_{i}}^{k}-\mu_{s}^{k}\right\rvert}$ (15) $\displaystyle+$ $\displaystyle\frac{1}{n_{\gamma,t}^{k}}\sum_{s\geq t-\tau^{\prime}}^{t-1}\gamma^{t-s}\mathbf{1}\\{\varphi_{s}=k\\}\mathinner{\\!\left\lvert\mu_{\tau_{i}}^{k}-\mu_{s}^{k}\right\rvert}.$ (16) Since $\mu_{t}^{k}\in[a,a+b]$ for all $t\in\mathcal{T}$, we have $\eqref{bias: dis}\leq b$. Also, ${\eqref{bias: dis}}\leq\frac{1}{n_{\gamma,t}^{k}}\sum_{s\leq t-\tau^{\prime}}b\gamma^{t-s}\leq\frac{b\gamma^{\tau^{\prime}}}{(1-\gamma)n_{\gamma,t}^{k}}=\frac{\xi\ln(\tau)}{bn_{\gamma,t}^{k}}.$ Accordingly, we get ${\eqref{bias: dis}}\leq\min\left(b,\frac{\xi\ln(\tau)}{bn_{\gamma,t}^{k}}\right)\leq\sqrt{\frac{\xi\ln(\tau)}{n_{\gamma,t}^{k}}}.$ Furthermore, for any $t\in\mathcal{T}_{i}$, $\eqref{bias: dis2}\leq\max_{s\in[t-\tau^{\prime},t-1]}\mathinner{\\!\left\lvert\mu_{\tau_{i}}^{k}-\mu_{s}^{k}\right\rvert}\leq\sum_{j=i-n^{\prime}}^{i}v_{j},$ where $n^{\prime}=\lceil{\tau^{\prime}/\tau}\rceil$ and $v_{j}$ is the variation within $\mathcal{T}_{j}$. So we conclude that for any $t\in\mathcal{T}_{i}$, $\mathinner{\\!\left\lvert\mu_{\kappa_{i}}^{k}-M_{\gamma,t}^{k}\right\rvert}\leq c_{\gamma,t}^{k}+\sum_{j=i-n^{\prime}}^{i}v_{j},\quad\forall k\in\mathcal{K}.$ (17) Step 2: Within partition $\mathcal{T}_{i}$, let $\hat{\Delta}_{i}^{k}=\mu_{\tau_{i}}^{\kappa_{i}}-\mu_{\tau_{i}}^{k}-2\sum_{j=i-n^{\prime}}^{i}v_{j},$ and define a subset of bad arms as $\mathcal{B}_{i}^{\textup{D}}=\bigg{\\{}k\in\mathcal{K}\>|\>\hat{\Delta}_{i}^{k}\geq\epsilon^{\prime}\bigg{\\}},$ where we select $\epsilon^{\prime}=4\sqrt{\xi\gamma^{1-\tau}K\ln(\tau)/\tau}$. Since $\mathinner{\\!\left\lvert\mu_{t}^{k}-\mu_{\tau_{i}}^{k}\right\rvert}\leq v_{i}$ for any $t\in\mathcal{T}_{i}$ and for any $k\in\mathcal{K}$ $\displaystyle\sum_{t\in\mathcal{T}}\mu_{t}^{*}-\mu_{t}^{\varphi_{t}}\leq\sum_{i=1}^{N}\sum_{t\in\mathcal{T}_{i}}\mu_{\tau_{i}}^{\kappa_{i}}-\mu_{\tau_{i}}^{\varphi_{t}}+v_{i}$ $\displaystyle\leq$ $\displaystyle\tau V_{T}+\sum_{i=1}^{N}\sum_{t\in\mathcal{T}_{i}}\bigg{[}\mathbf{1}\left\\{\varphi_{t}\in\mathcal{B}_{i}^{\textup{D}}\right\\}\hat{\Delta}_{i}^{\varphi_{t}}+2\sum_{j=i-n^{\prime}}^{i}v_{j}+\epsilon^{\prime}\bigg{]}$ $\displaystyle\leq$ $\displaystyle(2n^{\prime}+3)\tau V_{T}+N\epsilon^{\prime}\tau\\!+\\!\sum_{i=1}^{N}\sum_{k\in\mathcal{B}_{i}^{\textup{D}}}\\!\\!\hat{\Delta}_{i}^{k}\sum_{t\in\mathcal{T}_{i}}\mathbf{1}\left\\{\varphi_{t}=k\right\\}.$ (18) Step 3: In this step, we bound $\mathbb{E}\big{[}\hat{\Delta}_{i}^{k}\sum_{t\in\mathcal{T}_{i}}\mathbf{1}\left\\{\varphi_{t}=k\right\\}\big{]}$ for an arm $k\in\mathcal{B}_{i}^{\textup{D}}$. Let $t_{i}^{k}(l)$ be the $l$-th time slot arm $k$ is selected within $\mathcal{T}_{i}$. From arm selection policy, we get $g_{t}^{\varphi_{t}}\geq g_{t}^{\kappa_{i}}$, which result in $\sum_{t\in\mathcal{T}_{i}}\mathbf{1}\left\\{\varphi_{t}=k\right\\}\leq l_{i}^{k}+\sum_{t\in\mathcal{T}_{i}}\mathbf{1}{\Big{\\{}g_{t}^{k}\geq g_{t}^{\kappa_{i}},t>t_{i}^{k}(l_{i}^{k})\Big{\\}}},$ (19) where we pick $l_{i}^{k}=\left\lceil{16\xi\gamma^{1-\tau}\ln(\tau)/{\big{(}\hat{\Delta}_{i}^{k}\big{)}^{2}}}\right\rceil$. Note that $g_{t}^{k}\geq g_{t}^{\kappa_{i}}$ is true means at least one of the followings holds, $\displaystyle\hat{\mu}_{\gamma,t}^{k}$ $\displaystyle\geq M_{\gamma,t}^{k}+c_{\gamma,t}^{k},$ (20) $\displaystyle\hat{\mu}_{\gamma,t}^{\kappa_{i}}$ $\displaystyle\leq M_{\gamma,t}^{\kappa_{i}}-c_{\gamma,t}^{\kappa_{i}},$ (21) $\displaystyle M_{\gamma,t}^{\kappa_{i}}+c_{\gamma,t}^{\kappa_{i}}$ $\displaystyle<M_{\gamma,t}^{k}+3c_{\gamma,t}^{k}.$ (22) For any $t\in\mathcal{T}_{i}$, since every sample before $t$ within $\mathcal{T}_{i}$ has a weight greater than $\gamma^{\tau-1}$, if $t>t_{i}^{k}(l_{i}^{k})$, $\displaystyle c_{\gamma,t}^{k}=\sqrt{\frac{\xi\ln(\tau)}{n_{\gamma,t}^{k}}}\leq\sqrt{\frac{\xi\ln(\tau)}{\gamma^{\tau-1}l_{i}^{k}}}\leq\frac{\hat{\Delta}_{i}^{k}}{4}.$ Combining it with (17) yields $\displaystyle M_{\gamma,t}^{\kappa_{i}}-M_{\gamma,t}^{k}$ $\displaystyle\geq\mu_{\tau_{i}}^{\kappa_{i}}-\mu_{\tau_{i}}^{k}-c_{\gamma,t}^{\kappa_{i}}-c_{\gamma,t}^{k}-2\sum_{j=i-n^{\prime}}^{i}v_{j}$ $\displaystyle\geq\hat{\Delta}_{i}^{k}-c_{\gamma,t}^{\kappa_{i}}-c_{\gamma,t}^{k}\geq 3c_{\gamma,t}^{k}-c_{\gamma,t}^{\kappa_{i}},$ which indicates (22) is false. As $\xi>1/2$, we select $\lambda=4\sqrt{1-1/(2\xi)}$ and apply Fact 2 to get $\mathbb{P}(\text{\eqref{h1} is true})\leq\left\lceil\log_{1+\lambda}(\tau)\right\rceil\tau^{-2\xi(1-{\lambda^{2}}/{16})}\leq\frac{\left\lceil\log_{1+\lambda}(\tau)\right\rceil}{\tau}.$ The probability of (21) to be true shares the same bound. Then, it follows from (19) that $\mathbb{E}\big{[}\hat{\Delta}_{i}^{k}\sum_{t\in\mathcal{T}_{i}}\mathbf{1}\left\\{\varphi_{t}=k\right\\}\big{]}$ is upper bounded by $\displaystyle\hat{\Delta}_{i}^{k}l_{i}^{k}+\hat{\Delta}_{i}^{k}\sum_{t\in\mathcal{T}_{i}}\mathbb{P}\left(\text{\eqref{h1} or~{}\eqref{h2} is true}\right)$ $\displaystyle\leq$ $\displaystyle\frac{16\xi\gamma^{1-\tau}\ln(\tau)}{\hat{\Delta}_{i}^{k}}+\hat{\Delta}_{i}^{k}+2\hat{\Delta}_{i}^{k}\left\lceil\log_{1+\lambda}\left(\tau\right)\right\rceil$ $\displaystyle\leq$ $\displaystyle\frac{16\xi\gamma^{1-\tau}\ln(\tau)}{\epsilon^{\prime}}+b+2b\left\lceil\log_{1+\lambda}\left(\tau\right)\right\rceil,$ (23) where we use $\epsilon^{\prime}\leq\hat{\Delta}_{i}^{k}\leq b$ in the last step. Step 4: From (18) and (23), and plugging in the value of $\epsilon^{\prime}$, an easy computation results in $\displaystyle R_{T}^{\textup{D-UCB}}\leq$ $\displaystyle(2n^{\prime}+3)\tau V_{T}+8N\sqrt{\xi\gamma^{1-\tau}K\tau\ln(\tau)}$ $\displaystyle+2Nb+2Nb\log_{1+\lambda}\left(\tau\right),$ where the dominating term is $(2n^{\prime}+3)\tau V_{T}$. Considering $\tau^{\prime}=\frac{\ln\big{(}(1-\gamma)\xi\ln(\tau)/b^{2}\big{)}}{\ln{\gamma}}\leq\frac{-\ln\big{(}(1-\gamma)\xi\ln(\tau)/b^{2}\big{)}}{1-\gamma},$ we get $n^{\prime}\leq C^{\prime}\ln(T)$ for some constant $C^{\prime}$. Hence there exists some absolute constant $C$ such that $R_{T}^{\textup{D-UCB}}\leq C\ln(T)(KV_{T})^{\frac{1}{3}}T^{\frac{2}{3}}.$ ∎ Although discount factor method requires less memory, there exists an extra factor $\ln(T)$ in the upper bound on the worst-case regret for D-UCB comparing with the minimax regret. This is due to the fact that discount factor method does not entirely cut off outdated sampling history like periodic resetting or sliding window techniques. ## V UCB Policies for Heavy-tailed Nonstationary Stochastic MAB Problems In this section, we propose and analyze UCB algorithms for non-stationary stochastic MAB problem with heavy-tailed rewards defined in Assumption 2. We first recall a minimax policy for the stationary heavy-tailed MAB problem called Robust MOSS [36]. We then extend it to nonstationary setting and design resetting robust MOSS algorithm and sliding-window robust MOSS algorithm. ### V-A Background on Robust MOSS algorithm for the stationary heavy-tailed MAB problem Robust MOSS algorithm handles stationary heavy-tailed MAB problems in which the rewards have finite moments of order $1+\epsilon$, for $\epsilon\in(0,1]$. For simplicity, as stated in Assumption 2, we restrict our discussion to $\epsilon=1$. Robust MOSS uses the saturated empirical mean instead of the empirical mean. Let $n_{k}(t)$ be the number of times that arm $k$ has been selected until time $t-1$. Pick $a>1$ and let $h(m)=a^{\left\lfloor\log_{a}\left(m\right)\right\rfloor+1}$. Let the saturation limit at time $t$ be defined by $B_{n_{k}(t)}\mathrel{\mathop{\mathchar 58\relax}}=\sqrt{\frac{h(n_{k}(t))}{\ln_{+}\left(\frac{T}{Kh(n_{k}(t))}\right)}},$ where $\ln_{+}(x)\mathrel{\mathop{\mathchar 58\relax}}=\max(\ln x,1)$. Then, the saturated empirical mean estimator is defined by $\bar{\mu}_{n_{k}(t)}\mathrel{\mathop{\mathchar 58\relax}}=\frac{1}{n_{k}(t)}\sum_{s=1}^{t-1}\mathbf{1}\\{\varphi_{s}=k\\}\operatorname{sat}(X_{s},B_{n_{k}(t)}),$ (24) where $\operatorname{sat}(X_{s},B_{m})\mathrel{\mathop{\mathchar 58\relax}}=\operatorname{sign}(X_{s})\min\big{\\{}\mathinner{\\!\left\lvert X_{s}\right\rvert},B_{m}\big{\\}}.$ The Robust MOSS algorithm initializes by selecting each arm once and subsequently, at each time $t$, selects the arm that maximizes the following upper confidence bound $g^{k}_{n_{k}(t)}=\bar{\mu}^{k}_{n_{k}(t)}+(1+\zeta)c_{n_{k}(t)},$ where $c_{n_{k}(t)}=\sqrt{{\ln_{+}\big{(}\frac{T}{Kn_{k}(t)}\big{)}}/{n_{k}(t)}}$, $\zeta$ is an positive constant such that $\psi(2\zeta/a)\geq 2a/\zeta$ and $\psi(x)=(1+1/x)\ln(1+x)-1$. Note that for $x\in(0,\infty)$, function $\psi(x)$ is monotonically increasing in $x$. ### V-B Resetting robust MOSS for the non-stationary heavy-tailed MAB problem Similarly to R-MOSS, Resetting Robust MOSS (R-RMOSS) restarts Robust MOSS after every $\tau$ time slots. For a stationary heavy-tailed MAB problem, it has been shown in [36] that the worst-case regret of Robust MOSS belongs to $\mathcal{O}(\sqrt{KT})$. This result along with an analysis similar to the analysis for R-MOSS in Theorem 3 yield the following theorem for R-RMOSS. For brevity, we skip the proof. ###### Theorem 7. For the nonstationary heavy-tailed MAB problem with $K$ arms, horizon $T$, variation budget $V_{T}>0$ and $\tau=\Big{\lceil}{K^{\frac{1}{3}}\left(T/V_{T}\right)^{\frac{2}{3}}}\Big{\rceil}$, if $\psi(2\zeta/a)\geq 2a/\zeta$, the worst-case regret of R-RMOSS satisfies $\sup_{\mathcal{F}_{T}^{\mathcal{K}}\in\mathcal{E}(V_{T},T,K)}R_{T}^{\textup{R-RMOSS}}\in\mathcal{O}((KV_{T})^{\frac{1}{3}}T^{\frac{2}{3}}).$ ### V-C Sliding-window robust MOSS for the non-stationary heavy-tailed MAB problem In Sliding-Window Robust MOSS (SW-RMOSS), $n_{k}(t)$ and $\bar{\mu}_{n_{k}(t)}$ are computed from the sampling history within $\mathcal{W}_{t}$, and $c_{n_{k}(t)}=\sqrt{{\ln_{+}\big{(}\frac{\tau}{Kn_{k}(t)}\big{)}}/{n_{k}(t)}}$. To analyze SW-RMOSS, we want to establish a similar property as Lemma 4 to bound the probability about an arm being under or over estimated. Toward this end, we need the following properties for truncated random variable. ###### Lemma 8. Let $X$ be a random variable with expected value $\mu$ and $\mathbb{E}[X^{2}]\leq 1$. Let $d\mathrel{\mathop{\mathchar 58\relax}}=\operatorname{sat}(X,B)-\mathbb{E}[\operatorname{sat}(X,B)]$. Then for any $B>0$, it satisfies (i) $\mathinner{\\!\left\lvert d\right\rvert}\leq 2B$ (ii) $\mathbb{E}[d^{2}]\leq 1$ (iii) $\mathinner{\\!\left\lvert\mathbb{E}[\operatorname{sat}(X,B)]-\mu\right\rvert}\leq 1/B$. ###### Proof. Property (i) follows immediately from definition of $d$ and property (ii) follows from $\mathbb{E}[d^{2}]\leq\mathbb{E}\big{[}\operatorname{sat}^{2}(X,B)\big{]}\leq\mathbb{E}\big{[}X^{2}\big{]}.$ To see property (iii), since $\mu=\mathbb{E}\big{[}X\big{(}\mathbf{1}{\left\\{\mathinner{\\!\left\lvert X\right\rvert}\leq B\right\\}}+\mathbf{1}{\left\\{\mathinner{\\!\left\lvert X\right\rvert}>B\right\\}}\big{)}\big{]},$ one have $\displaystyle\mathinner{\\!\left\lvert\mathbb{E}[\operatorname{sat}(X,B)]-\mu\right\rvert}$ $\displaystyle\leq\mathbb{E}\left[\left(\mathinner{\\!\left\lvert X\right\rvert}-B\right)\mathbf{1}{\left\\{\mathinner{\\!\left\lvert X\right\rvert}>B\right\\}}\right]$ $\displaystyle\leq\mathbb{E}\left[\mathinner{\\!\left\lvert X\right\rvert}\mathbf{1}{\left\\{\mathinner{\\!\left\lvert X\right\rvert}>B\right\\}}\right]\leq\mathbb{E}\left[{X^{2}}/{B}\right].$ ∎ Moreover, we will also use a maximal Bennett type inequality as shown in the following. ###### Lemma 9 (Maximal Bennett’s inequality [37]). Let $\left\\{X_{i}\right\\}_{i\in\\{1,\dots,n\\}}$ be a sequence of bounded random variables with support $[-B,B]$, where $B\geq 0$. Suppose that $\mathbb{E}[X_{i}|X_{1},\ldots,X_{i-1}]=\mu_{i}$ and $\operatorname{Var}[X_{i}|X_{1},\ldots,X_{i-1}]\leq v$. Let $S_{m}=\sum_{i=1}^{m}(X_{i}-\mu_{i})$ for any $m\in\\{1,\dots,n\\}$. Then, for any $\delta\geq 0$ $\displaystyle\mathbb{P}\left(\exists{m\in\\{1,\dots,n\\}}\mathrel{\mathop{\mathchar 58\relax}}S_{m}\geq\delta\right)\leq\exp\left(-\frac{\delta}{B}\psi\left(\frac{B\delta}{nv}\right)\right),$ $\displaystyle\mathbb{P}\left(\exists{m\in\\{1,\dots,n\\}}\mathrel{\mathop{\mathchar 58\relax}}S_{m}\leq-\delta\right)\leq\exp\left(-\frac{\delta}{B}\psi\left(\frac{B\delta}{nv}\right)\right).$ Now, we are ready to establish a concentration property for saturated sliding window empirical mean. ###### Lemma 10. For any arm $k\in\\{1,\dots,K\\}$ and any $t\in\left\\{K+1,\ldots,T\right\\}$, if $\psi(2\zeta/a)\geq 2a/\zeta$, the probability of either event $A=\big{\\{}g^{k}_{t}\leq M_{t}^{k}-x,n_{k}(t)\geq l\big{\\}}$ or event $B=\big{\\{}g^{k}_{t}-2c_{n_{k}(t)}\geq M_{t}^{k}+x,n_{k}(t)\geq l\big{\\}}$, for any $x>0$ and any $l\geq 1$, is no greater than $\frac{2a}{\beta^{2}\ln(a)}\frac{K}{\tau x^{2}}(\beta x\sqrt{h(l)/a}+1)\exp\left(-\beta x\sqrt{h(l)/a}\right),$ where $\beta=\psi\left(2\zeta/a\right)/(2a)$. ###### Proof. Recall that $u_{i}^{kt}$ is the $i$-th time slot when arm $k$ is selected within $\mathcal{W}_{t}$. Since $c_{m}$ is a monotonically decreasing in $m$, $1/B_{m}=c_{h(m)}\leq c_{m}$ due to $h(m)\geq m$. Then, it follows from property (iii) in Lemma 8 that $\displaystyle\mathbb{P}(A)\\!$ $\displaystyle\leq\mathbb{P}\bigg{(}\\!\exists m\\!\in\\!\\{l,\ldots,\tau\\}\\!\mathrel{\mathop{\mathchar 58\relax}}\\!\bar{\mu}^{k}_{m}\leq\sum_{i=1}^{m}\\!\frac{\mu_{u_{i}^{kt}}^{k}}{m}\\!\\!-(1+\zeta)c_{m}\\!-x\\!\bigg{)}$ $\displaystyle\leq\mathbb{P}\bigg{(}\\!\exists m\\!\in\\!\\{l,\ldots,\tau\\}\\!\mathrel{\mathop{\mathchar 58\relax}}\\!\sum_{i=1}^{m}\\!\frac{\bar{d}_{im}^{kt}}{m}\\!\leq\\!\frac{1}{B_{m}}\\!-(1+\zeta)c_{m}\\!-x\\!\bigg{)}$ $\displaystyle\leq\mathbb{P}\bigg{(}\\!\exists m\\!\in\\!\\{l,\ldots,\tau\\}\\!\mathrel{\mathop{\mathchar 58\relax}}\\!\frac{1}{m}\sum_{i=1}^{m}\bar{d}_{im}^{kt}\leq-x-\zeta c_{m}\bigg{)},\,$ (25) where $\bar{d}_{im}^{kt}=\operatorname{sat}\big{(}X_{u_{i}^{kt}}^{k},B_{m}\big{)}-\mathbb{E}\big{[}\operatorname{sat}\big{(}X_{u_{i}^{kt}}^{k},B_{m}\big{)}\big{]}$. Recall we select $a>1$. Again, we apply a peeling argument with geometric grid $a^{s}\leq m<a^{s+1}$ over time interval $\\{l,\ldots,\tau\\}$. Let $s_{0}=\left\lfloor\log_{a}(l)\right\rfloor$. Since $c_{m}$ is monotonically decreasing with $m$, $\eqref{prob:A}\leq\\!\\!\sum_{s\geq s_{0}}\\!\mathbb{P}\Bigg{(}\\!\exists m\in[a^{s},a^{s+1})\\!\mathrel{\mathop{\mathchar 58\relax}}\\!\sum_{i=1}^{m}\bar{d}_{im}^{kt}\leq\\!-a^{s}\left(x+\zeta c_{a^{s+1}}\right)\\!\\!\bigg{)}.$ For all $m\in[a^{s},a^{s+1})$, since $B_{m}=B_{a^{s}}$, from Lemma 8 we know $\mathinner{\\!\left\lvert\bar{d}_{im}^{kt}\right\rvert}\leq 2B_{a^{s}}$ and $\mathbf{Var}\left[\bar{d}_{im}^{kt}\right]\leq 1$. Continuing from previous step, we apply Lemma 9 to get $\displaystyle\eqref{prob:A}\leq$ $\displaystyle\sum_{s\geq s_{0}}\exp\left(-\frac{a^{s}\left(x+\zeta c_{a^{s+1}}\right)}{2B_{a^{s}}}\psi\left(\frac{2B_{a^{s}}}{a}\left(x+\zeta c_{a^{s+1}}\right)\right)\right)$ $\displaystyle\left(\text{since }\psi(x)\text{ is monotonically increasing}\right)$ $\displaystyle\leq$ $\displaystyle\sum_{s\geq s_{0}}\exp\left(-\frac{a^{s}\left(x+\zeta c_{a^{s+1}}\right)}{2B_{a^{s}}}\psi\left(\frac{2\zeta}{a}B_{a^{s}}c_{a^{s+1}}\right)\right)$ (substituting $c_{a^{s+1}}$, $B_{a^{s}}$ and using $h(a^{s})=a^{s+1}$) $\displaystyle=$ $\displaystyle\sum_{s\geq s_{0}+1}\exp\left(-a^{s}\left(\frac{x}{B_{a^{s-1}}}+\zeta c_{a^{s}}^{2}\right)\frac{\psi\left(2\zeta/a\right)}{2a}\right)$ $\displaystyle\left(\text{since }\zeta\psi(2\zeta/a)\geq 2a\right)$ $\displaystyle\leq$ $\displaystyle\frac{K}{\tau}\sum_{s\geq s_{0}+1}a^{s}\exp\left(-a^{s}\frac{x}{B_{a^{s-1}}}\frac{\psi\left(2\zeta/a\right)}{2a}\right).$ (26) Let $b={x\psi\left(2\zeta/a\right)}/(2a)$. Since $\ln_{+}(x)\geq 1$ for all $x>0$, $\displaystyle\eqref{sum:1}\leq$ $\displaystyle\frac{K}{\tau}\sum_{s\geq s_{0}+1}a^{s}\exp\left(-b\sqrt{a^{s}}\right)$ $\displaystyle\leq$ $\displaystyle\frac{K}{\tau}\int_{s_{0}+1}^{+\infty}a^{y}\exp\big{(}-b\sqrt{a^{y-1}}\big{)}dy$ $\displaystyle=$ $\displaystyle\frac{K}{\tau}a\int_{s_{0}}^{+\infty}a^{y}\exp\big{(}-b\sqrt{a^{y}}\big{)}dy$ $\displaystyle=$ $\displaystyle\frac{K}{\tau}\frac{2a}{\ln(a)b^{2}}\int_{b\sqrt{a^{s_{0}}}}^{+\infty}z\exp\big{(}-z\big{)}dz\,(\text{where }z=b\sqrt{a^{y}})$ $\displaystyle\leq$ $\displaystyle\frac{K}{\tau}\frac{2a}{\ln(a)b^{2}}(b\sqrt{a^{s_{0}}}+1)\exp(-b\sqrt{a^{s_{0}}}),$ which concludes the proof. ∎ With Lemma 10, the upper bound on the worst-case regret for SW-RMOSS in the nonstationary heavy-tailed MAB problem can be analyzed similarly as Theorem 5. ###### Theorem 11. For the nonstationary heavy-tailed MAB problem with $K$ arms, time horizon $T$, variation budget $V_{T}>0$ and $\tau=\Big{\lceil}{K^{\frac{1}{3}}\left(T/V_{T}\right)^{\frac{2}{3}}}\Big{\rceil}$, if $\psi(2\zeta/a)\geq 2a/\zeta$, the worst-case regret of SW-RMOSS satisfies $\sup_{\mathcal{F}_{T}^{\mathcal{K}}\in\mathcal{E}(V_{T},T,K)}R_{T}^{\textup{SW- RMOSS}}\leq C(KV_{T})^{\frac{1}{3}}T^{\frac{2}{3}}.$ ###### Sketch of the proof. The procedure is similar as the proof of Theorem 5. The key difference is due to the nuance between the concentration properties on mean estimator. Neglecting the leading constants, the probability upper bound in Lemma 4 has a factor $\exp(-x^{2}l/\eta)$ comparing with $(\beta x\sqrt{h(l)/a}+1)\exp\left(-\beta x\sqrt{h(l)/a}\right)$ in Lemma 10. Since both factors are no greater than $1$, by simply replacing $\eta$ with $(1+\zeta)^{2}$ and taking similar calculation in every step except inequality (13), comparable bounds that only differs in leading constants can be obtained. Applying Lemma 10, we revise the computation of (13) as the following, $\displaystyle\sum_{s\geq l_{i}^{k}+1}\mathbb{P}{\bigg{\\{}g_{t_{s}}^{k}-2c_{n_{k}(t_{s})}>M_{t_{s}}^{k}+\frac{\Delta_{i}^{k}}{4}\bigg{\\}}}$ $\displaystyle\leq$ $\displaystyle\sum_{s\geq l_{i}^{k}}C^{\prime}\left(\frac{\beta\Delta_{i}^{k}}{4}\sqrt{\frac{h(l)}{a}}+1\right)\exp\left(-\frac{\beta\Delta_{i}^{k}}{4}\sqrt{\frac{h(l)}{a}}\right)$ $\displaystyle\leq$ $\displaystyle\int_{l_{i}^{k}-1}^{+\infty}C^{\prime}\left(\frac{\beta\Delta_{i}^{k}}{4}\sqrt{\frac{y}{a}}+1\right)\exp\left(-\frac{\beta\Delta_{i}^{k}}{4}\sqrt{\frac{y}{a}}\right)\,dy$ $\displaystyle\leq$ $\displaystyle\frac{6a}{\beta^{2}}\frac{2a}{\beta^{2}\ln(a)}\frac{K}{\tau}\bigg{(}\frac{4}{\Delta_{i}^{k}}\bigg{)}^{4}.$ (27) where $C^{\prime}={2aK}\big{(}{4}/{\Delta_{i}^{k}}\big{)}^{2}/{\big{(}\beta^{2}\ln(a)\tau\big{)}}$.The second inequality is due to the fact that $(x+1)\exp(-x)$ is monotonically decreasing in $x$ for $x\in[0,\infty)$ and $h(l)>l$. In the last inequality, we change the lower limits of the integration from $l_{i}^{k}-1$ to $0$ since $l_{i}^{k}\geq 1$ and plug in the value of $C^{\prime}$. Comparing with (13), this upper bound only varies in constant multiplier. So is the worst-regret upper bound. ∎ ###### Remark 1. The benefit of discount factor method is that it is memory friendly. This advantage is lost if truncated empirical mean is used. As $n_{k}(t)$ could both increase and decrease with time, the truncated point could both grow and decline, so all sampling history needs to be recorded. It remains an open problem how to effectively using discount factor in a nonstationary heavy- tailed MAB problem. ## VI Numerical Experiments We complement the theoretical results in previous section with two Monte-Carlo experiments. For the light-tailed setting, we compare R-MOSS, SW-MOSS and D-UCB in this paper with other state-of-art policies. For the heavy-tailed setting, we test the robustness of R-RMOSS and SW-RMOSS against both heavy- tailed rewards and nonstationarity. Each result in this section is derived by running designated policies $500$ times. And parameter selections for compared policies are strictly coherent with referred literature. ### VI-A Bernoulli Nonstationay Stochastic MAB Experiment To evaluated the performance of different policies, we consider two nonstationary environment as shown in Figs. 1(a) and 1(b), which both have $3$ arms with nonstationary Bernoulli reward. The success probability sequence at each arm is a Brownian motion in environment $1$ and a sinusoidal function of time $t$ in environment $2$. And the variation budget $V_{T}$ is $8.09$ and $3$ respectively. (a) Environment $1$ (b) Environment $2$ (c) Regrets for environment $1$ (d) Regrets for environment $2$ Figure 1: Comparison of different policies. The growths of regret in Figs. 1(c) and 1(d) show that UCB based policies (R-MOSS, SW-MOSS, and D-UCB) maintain their superior performance against adversarial bandit based policies (Rexp$3$ and Exp$3$.S) for stochastic bandits even in nonstationary settings, especially for R-MOSS and SW-MOSS. Besides, DTS outperforms other polices when the best arm does not switch. While each switch of the best arm seems to incur larger regret accumulation for DTS, which results in a lager regret compared with SW-MOSS and R-MOSS. ### VI-B Heavy-tailed Nonstationay Stochastic MAB Experiment Again we consider the $3$-armed bandit problem with sinusoidal mean rewards. In particular, for each arm $k\in\\{1,2,3\\}$, $\mu_{t}^{k}=0.3\sin\left(0.001\pi t+2k\pi/3\right),\quad t\in\\{1,\dots,5000\\}.$ Thus, the variation budget is $3$. Besides, mean reward is contaminated by additive sampling noise $\nu$, where $\mathinner{\\!\left\lvert\nu\right\rvert}$ is a generalized Pareto random variable and the sign of $\nu$ has equal probability to be “$+$” and “$-$”. So the probability distribution for $X_{t}^{k}$ is $f_{t}^{k}(x)=\frac{1}{2\sigma}\left(1+\frac{\xi\mathinner{\\!\left\lvert x-\mu_{t}^{k}\right\rvert}}{\sigma}\right)^{-\frac{1}{\xi}-1}\,\text{for }x\in(-\infty,+\infty).$ We select $\xi=0.4$ and $\sigma=0.23$ such that Assumption 2 is satisfied. We select $a=1.1$ and $\zeta=2.2$ for both R-RMOSS and SW-RMOSS such that condition $\psi(2\zeta/a)\geq 2a/\zeta$ is met. (a) Regret (b) Histogram of $R_{T}$ Figure 2: Performances with heavy-tailed rewards. Fig. 2(a) show RMOSS based polices and slightly outperform MOSS based polices in heavy-tailed settings. While by comparing the estimated histogram of $R_{T}$ for different policies in Fig. 2(b), R-RMOSS and SW-RMOSS have a better consistency and a smaller possibility of a particular realization of the regret deviating significantly from the mean value. ## VII Conclusion We studied the general nonstationary stochastic MAB problem with variation budget and provided three UCB based policies for the problem. Our analysis showed that the proposed policies enjoy the worst-case regret that is within a constant factor of the minimax regret lower bound. Besides, the sub-Gaussian assumption on reward distributions is relaxed to define the nonstationary heavy-tailed MAB problem. We show the order optimal worst-case regret can be maintained by extending the previous policies to robust versions. There are several possible avenues for future research. In this paper, we relied on passive methods to balance the remembering-versus-forgetting tradeoff. The general idea is to keep taking in new information and removing out-dated information. Parameter-free active approaches that adaptively detect and react to environment changes are promising alternatives and may result in better experimental performance. Also extensions from the single decision- maker to distributed multiple decision-makers is of interest. Another possible direction is the nonstaionary version of rested and restless bandits. ## References * [1] A. B. H. Alaya-Feki, E. Moulines, and A. LeCornec, “Dynamic spectrum access with non-stationary multi-armed bandit,” in _IEEE Workshop on Signal Processing Advances in Wireless Communications_ , 2008, pp. 416–420. * [2] A. Anandkumar, N. Michael, A. K. Tang, and A. Swami, “Distributed algorithms for learning and cognitive medium access with logarithmic regret,” _IEEE Journal on Selected Areas in Communications_ , vol. 29, no. 4, pp. 731–745, 2011. * [3] Y. Li, Q. Hu, and N. Li, “A reliability-aware multi-armed bandit approach to learn and select users in demand response,” _Automatica_ , vol. 119, p. 109015, 2020. * [4] D. Kalathil and R. Rajagopal, “Online learning for demand response,” in _Annual Allerton Conference on Communication, Control, and Computing_ , 2015, pp. 218–222. * [5] J. R. Krebs, A. Kacelnik, and P. Taylor, “Test of optimal sampling by foraging great tits,” _Nature_ , vol. 275, no. 5675, pp. 27–31, 1978. * [6] V. Srivastava, P. Reverdy, and N. E. Leonard, “On optimal foraging and multi-armed bandits,” in _Annual Allerton Conference on Communication, Control, and Computing_ , Monticello, IL, USA, 2013, pp. 494–499. * [7] ——, “Surveillance in an abruptly changing world via multiarmed bandits,” in _IEEE Conference on Decision and Control_ , 2014, pp. 692–697. * [8] C. Baykal, G. Rosman, S. Claici, and D. Rus, “Persistent surveillance of events with unknown, time-varying statistics,” in _IEEE International Conference on Robotics and Automation_ , 2017, pp. 2682–2689. * [9] M. Y. Cheung, J. Leighton, and F. S. Hover, “Autonomous mobile acoustic relay positioning as a multi-armed bandit with switching costs,” in _IEEE/RSJ Int Conf on Intelligent Robots and Systems_ , Tokyo, Japan, Nov. 2013, pp. 3368–3373. * [10] D. Agarwal, B.-C. Chen, P. Elango, N. Motgi, S.-T. Park, R. Ramakrishnan, S. Roy, and J. Zachariah, “Online models for content optimization,” in _Advances in Neural Information Processing Systems_ , 2009, pp. 17–24. * [11] L. Li, W. Chu, J. Langford, and R. E. Schapire, “A contextual-bandit approach to personalized news article recommendation,” in _International Conference on World Wide Web_ , 2010, pp. 661–670. * [12] H. Robbins, “Some aspects of the sequential design of experiments,” _Bulletin of the American Mathematical Society_ , vol. 58, no. 5, pp. 527–535, 1952. * [13] T. L. Lai and H. Robbins, “Asymptotically efficient adaptive allocation rules,” _Advances in Applied Mathematics_ , vol. 6, no. 1, pp. 4–22, 1985\. * [14] A. N. Burnetas and M. N. Katehakis, “Optimal adaptive policies for sequential allocation problems,” _Advances in Applied Mathematics_ , vol. 17, no. 2, pp. 122–142, 1996. * [15] P. Auer, N. Cesa-Bianchi, and P. Fischer, “Finite-time analysis of the multiarmed bandit problem,” _Machine Learning_ , vol. 47, no. 2, pp. 235–256, 2002. * [16] A. Garivier and O. Cappé, “The KL-UCB algorithm for bounded stochastic bandits and beyond,” in _Annual Conference on Learning Theory_ , 2011, pp. 359–376. * [17] P. Auer, Y. F. N. Cesa-Bianchi, and R. Schapire, “The nonstochastic multiarmed bandit problem,” _SIAM Journal on Computing_ , vol. 32, no. 1, pp. 48–77, 2002. * [18] L. Kocsis and C. Szepesvári, “Discounted UCB,” in _2nd PASCAL Challenges Workshop_ , vol. 2, 2006. * [19] A. Garivier and E. Moulines, “On upper-confidence bound policies for switching bandit problems,” in _International Conference on Algorithmic Learning Theory_. Springer, 2011, pp. 174–188. * [20] L. Wei and V. Srivastava, “On abruptly-changing and slowly-varying multiarmed bandit problems,” in _American Control Conference_ , Milwaukee, WI, Jun. 2018, pp. 6291–6296. * [21] C. Hartland, N. Baskiotis, S. Gelly, M. Sebag, and O. Teytaud, “Change Point Detection and Meta-Bandits for Online Learning in Dynamic Environments,” in _Conférence Francophone Sur L’Apprentissage Automatique_ , Grenoble, France, Jul. 2007, pp. 237–250. * [22] F. Liu, J. Lee, and N. Shroff, “A change-detection based framework for piecewise-stationary multi-armed bandit problem,” in _Thirty-Second AAAI Conference on Artificial Intelligence_ , 2018. * [23] L. Besson and E. Kaufmann, “The generalized likelihood ratio test meets klucb: an improved algorithm for piece-wise non-stationary bandits,” _arXiv preprint arXiv:1902.01575_ , 2019. * [24] Y. Cao, Z. Wen, B. Kveton, and Y. Xie, “Nearly optimal adaptive procedure with change detection for piecewise-stationary bandit,” in _International Conference on Artificial Intelligence and Statistics_ , 2019, pp. 418–427. * [25] J. Mellor and J. Shapiro, “Thompson sampling in switching environments with bayesian online change detection,” in _Artificial Intelligence and Statistics_ , 2013, pp. 442–450. * [26] O. Besbes and Y. Gur, “Stochastic multi-armed-bandit problem with non-stationary rewards,” in _Advances in Neural Information Processing Systems_ , 2014, pp. 199–207. * [27] O. Besbes, Y. Gur, and A. Zeevi, “Optimal exploration–exploitation in a multi-armed bandit problem with non-stationary rewards,” _Stochastic Systems_ , vol. 9, no. 4, pp. 319–337, 2019. * [28] V. Raj and S. Kalyani, “Taming non-stationary bandits: A Bayesian approach,” _arXiv preprint arXiv:1707.09727_ , 2017. * [29] R. Albert and A.-L. Barabási, “Statistical mechanics of complex networks,” _Reviews of Modern Physics_ , vol. 74, no. 1, p. 47, 2002. * [30] M. Vidyasagar, “Law of large numbers, heavy-tailed distributions, and the recent financial crisis,” in _Perspectives in Mathematical System Theory, Control, and Signal Processing_. Springer, 2010, pp. 285–295. * [31] J. Audibert and S. Bubeck, “Minimax policies for adversarial and stochastic bandits,” in _Annual Conference on Learning Theory_ , Montreal, Canada, Jun. 2009, pp. 217–226. * [32] P. Auer, N. Cesa-Bianchi, Y. Freund, and R. E. Schapire, “Gambling in a rigged casino: The adversarial multi-armed bandit problem,” in _IEEE Annual Foundations of Computer Science_ , 1995, pp. 322–331. * [33] S. Mannor and J. N. Tsitsiklis, “The sample complexity of exploration in the multi-armed bandit problem,” _Journal of Machine Learning Research_ , vol. 5, no. Jun, pp. 623–648, 2004. * [34] W. Hoeffding, “Probability inequalities for sums of bounded random variables,” _Journal of the American Statistical Association_ , vol. 58, no. 301, pp. 13–30, 1963. * [35] S. Bubeck, “Bandits games and clustering foundations,” Theses, Université des Sciences et Technologie de Lille - Lille I, 2010. [Online]. Available: https://tel.archives-ouvertes.fr/tel-00845565 * [36] L. Wei and V. Srivastava, “Minimax policy for heavy-tailed bandits,” _IEEE Control Systems Letters_ , vol. 5, no. 4, pp. 1423–1428, 2021. * [37] X. Fan, I. Grama, and Q. Liu, “Hoeffding’s inequality for supermartingales,” _Stochastic Processes and their Applications_ , vol. 122, no. 10, pp. 3545–3559, 2012.
# CSI-Based Localization with CNNs Exploiting Phase Information Anastasios Foliadis12, Mario H. Castañeda Garcia1, Richard A. Stirling- Gallacher1, Reiner S. Thomä2 1Munich Research Center, Huawei Technologies Duesseldorf GmbH, Munich, Germany 2Electronic Measurements and Signal Processing, Technische Universität Ilmenau, Ilmenau, Germany {anastasios.foliadis, mario.castaneda<EMAIL_ADDRESS><EMAIL_ADDRESS> ###### Abstract In this paper we study the use of the Channel State Information (CSI) as fingerprint inputs of a Convolutional Neural Network (CNN) for localization. We examine whether the CSI can be used as a distinct fingerprint corresponding to a single position by considering the inconsistencies with its raw phase that cause the CSI to be unreliable. We propose two methods to produce reliable fingerprints including the phase information. Furthermore, we examine the structure of the CNN and more specifically the impact of pooling on the positioning performance, and show that pooling over the subcarriers can be more beneficial than over the antennas. ###### Index Terms: Localization, Positioning, Deep Learning, CSI, Fingerprint, Neural Network ## I Introduction Advances in mobile communications and the development of Internet of Things (IoT) has introduced a large variety of new applications in a number of different areas of modern life. One important requirement in several of these applications is the estimation of the user’s position. Although the ubiquity of Global Positioning System (GPS) provides a great solution for outdoor localization, other alternatives are needed indoors. Many different solutions have been proposed in the literature for indoor positioning, ranging from classical approaches, like angle of arrival (AoA) and time of arrival (ToA) based, to pattern recognition approaches. More specifically, the ability to store and transmit large amounts of data has directed the focus on using deep learning. Additionally, with the 5th Generation (5G) network being deployed, providing high data rates and bandwidth, the number of antennas on devices is increasing, enabled by the mm- Wave operation frequency. For coherent communication, the multi-antenna channel between a user and the base station (BS) is estimated using pilot symbols. The estimated channel referred as channel state information (CSI) can serve as a fingerprint for localization. For localization based on fingerprint inputs, a database for a given environment is created offline and during the online phase the UE’s position is estimated, by matching its signal to the fingerprint map. There exist a number of different approaches to implement the mapping. These range from conventional, like maximum likelihood and least squares, to machine learning, like k-nearest neighbors and neural networks. What is considered a fingerprint also differs depending on the application. In [1] the mapping is done using convolutional neural networks (CNNs), achieving a sub-meter accuracy with simulated and real measurements by utilizing the real and imaginary parts of the CSI. In [2], again a CNN was used with real measurements, but with its inputs consisting of a combination of raw features (real and imaginary), polar features and time-domain features. More complex neural network configurations were used in [3] using as input the time-domain channel impulse response. The authors of [4] were able to achieve sub-centimeter accuracy by employing a denoising technique and an ensemble of neural networks. They considered only the magnitude of the channel, since they identified that phase measurements at the same position can change over time. Phase spatial inconsistency arising because of implementation aspects is a common issue affecting the CSI. For this reason several prior works either ignore or completely reject the phase and focus primarily on the magnitude. However, the phase could embed important information of the underlying channel for localization purposes. In [5] and [6], a transformation per antenna is proposed to calibrate the phase of multi-antenna measurements used as inputs to a CNN. In this paper, we propose two techniques to obtain a processed phase which is consistent and reliable for localization. In contrast to [5] and [6], we propose the same transformation across all antennas to preserve valuable AoA information. By employing CNNs, we show that the use of the processed phase improves the localization accuracy compared to when using the raw phase. To optimize the CNN based on the structure of the CSI, we also examine the impact of the pooling layer on the localization performance. In the remainder of this paper we describe the system model in Section II. In Section III, we present the proposed methods to address the phase issues. In Section IV we introduce the machine leaning approach that we utilize and in Section V we present the results of our simulations. We present our conclusions in Section VI. ## II System Model and Database Description ### II-A System Model Due to its ubiquity in wireless communications and ease of deployment, we consider the use of orthogonal frequency division multiplexing (OFDM) waveform for localization. In addition to the degrees of freedom that the subcarriers in OFDM provide, we can exploit the multiple antennas at the transmitter and receiver. For simplicity we consider a static uplink setup with a single transmit antenna at a UE and multiple receive antennas at the BS, i.e. a single-input multiple-output (SIMO) system. The uplink SIMO channel estimated at the BS is given by: $\boldsymbol{H}=\left[\boldsymbol{h}_{0},\boldsymbol{h}_{1},...,\boldsymbol{h}_{N_{\text{C}}-1}\right]\in\mathbb{C}^{N_{\text{R}}\times N_{\text{C}}}$ (1) where $N_{\text{C}}$ is the number of subcarriers and $N_{\text{R}}$ the number of antennas at the receiver. $\boldsymbol{h}_{n}\in\mathbb{C}^{N_{\text{R}}}$ is the vector describing the CSI for the receive antenna array at the $n$-th subcarrier. Since $\boldsymbol{H}$ is based on the underlying transfer function between receiver and transmitter, it can be used to obtain a distinct fingerprint for each measured position of a UE. There exist many techniques to estimate the complex channels based on transmitted pilots. In reality $\boldsymbol{H}$ is the effective channel, i.e. it includes timing offsets and hardware imperfections. Such disturbances may hinder the ability of the raw CSI to provide a distinct fingerprint for each position. According to the analysis in[7] and [8], these timing offsets between the oscillators of the transmitter and receiver, influence the estimated channel $\boldsymbol{H}$ and thus its phase at the $k$-th antenna and $n$-th subcarrier, where $k\in[0,N_{\text{R}}-1]$ and $n\in[0,N_{\text{C}}-1]$, based on the actual channel $\tilde{\boldsymbol{H}}$ can be expressed as follows: $\angle{\boldsymbol{H}_{n,k}}=\angle{\tilde{\boldsymbol{H}}_{n,k}}+(\tau_{\text{p}}+\tau_{\text{s}})n+\tau_{\text{c}}+\beta+\epsilon_{n,k}$ (2) where the phase of the actual channel at the $k$-th antenna and $n$-th subcarrier is $\angle{\tilde{\boldsymbol{H}}_{n,k}}$. $\tau_{\text{p}}$ is the symbol time offset (STO), $\tau_{\text{s}}$ the sampling time offset, $\tau_{\text{c}}$ the carrier frequency offset, $\beta$ is the phase locked loop (PLL) phase offset and $\epsilon_{n,k}$ is random noise. Due to the continuous timing drift of transmitter and receiver, the estimated channel would not be constant even if the underlying channel does not change. This makes the raw phase practically unusable as a distinct fingerprint for positioning, as pointed out in [5] and [6]. ### II-B Database Description To evaluate our proposals we use channel measurements which were described in [2] using 3 different antenna configurations for the BS in a $2.5\times 2.5$ m indoor area shown in Fig. 1. The different antenna configurations for the BS consist of a Uniform Rectangular Array (URA) of $8\times 8$ antennas, a Uniform Linear Array (ULA) of 64 antennas and a distributed (DIS) configuration of 8 ULA arrays with 8 antennas each. For each configuration, the BS has $N_{\text{R}}=64$ receive antennas. The spacing between adjacent antenna elements in the ULAs and URA is 70 mm. The UE was equipped with a single antenna. Uplink SIMO channel measurements were performed for equidistantly spaced UE locations (5 mm apart), within the green area inf Fig. 1. In [2], the carrier frequency was 2.61 GHz with a bandwidth of 20 MHz and $N_{\text{C}}=100$ subcarriers. Figure 1: Database measurement scenarios (figure taken from [2]). Distances are indicated in millimeters. (a) Channel magnitude across subcarriers (b) Channel phase across subcarriers Figure 2: Channel Fingerprints based on raw CSI (a) Phase difference method across subcarriers (b) Phase alignment method across subcarriers Figure 3: Phase fingerprints based on proposed methods ## III Fingerprint based on Phase Information An important aspect for localization based on fingerprint inputs is that the fingerprint that corresponds to each position is desired to be unique and consistent, during both online and offline phases. Otherwise the correct mapping of a new measurement to the database may not be possible. In particular, the magnitude of the estimated channel can be expected to be spatially consistent, e.g. does not vary significantly (besides some additive noise) over measurements at the same position. The spatial consistency of the magnitude can also be expected for measurements corresponding to very nearby positions. This can be observed in Fig. 2(a), where the magnitude over the subcarriers for the first antenna in the ULA is depicted for three neighboring sampled positions of the database from [2]. On the other hand, the raw phase of the estimated channel is usually not spatially consistent as described before. This is shown in Fig. 2(b), where the phase over the subcarriers for the first antenna in the ULA is depicted for the same three neighboring positions considered in Fig. 2(a). However, for our purposes we are not interested in estimating the phase of the actual channel $\tilde{\boldsymbol{H}}$, as our goal is merely to acquire a distinct and consistent fingerprint for each position, which is to say for each channel between UE and BS. In the following, we present two methods which produce reliable fingerprints by considering both phase and magnitude, while also preserving the Angle of Arrival (AoA) information embedded in the relationship between the phases of adjacent antennas. The AoA information can be exploited for localization. ### III-A Phase Difference When the same oscillator is used for all antennas, the phase offsets in (2) are the same for each antenna. Thus, by taking the difference between phases of two adjacent antennas, we eliminate the phase offsets. The phase difference between adjacent antennas for each subcarrier $n$ can be used to produce a distinct fingerprint: $\phi_{n,k}=(\angle\boldsymbol{H}_{n,k}-\angle\boldsymbol{H}_{n,(k+1)\text{mod}N_{\text{R}}})+\epsilon$ (3) where we used the modulo operator to include the phase difference between the last and the first antennas. In this way, we do not lose any information when creating the fingerprints. Based on phase difference, the fingerprint associated with the $k$-th antenna and $n$-th subcarrier for each position would be: $\boldsymbol{H}_{D}[n,k]=|\boldsymbol{H}_{n,k}|e^{\text{j}\left(\angle\boldsymbol{H}_{n,k}-\angle\boldsymbol{H}_{n,(k+1)modN_{\text{R}}}\right)}.$ (4) Here we have managed to create a distinct fingerprint $\boldsymbol{H}_{D}$, for each position, whose phase at each element $[n,k]$ is the phase difference of antennas $k$ and $k+1$ of the original matrix $\boldsymbol{H}$ at subcarrier $n$ while its magnitude is the same as $\boldsymbol{H}$. Also, the difference of phases, takes into account the relationship between antennas. The fingerprint quality offered by the phase difference can be seen in Fig. 3(a), where the phase difference of the first and second antennas in the ULA over the subcarriers is shown for the same positions as in Fig. 2(b). ### III-B Phase Alignment (a) Phase wrapping of $\angle H_{D}[n,k]$ (b) Phase Sine (c) Phase Cosine Figure 4: Phase fingerprint considering phase wrapping By considering (3), we see that $\tau_{\text{p}}$ and $\tau_{\text{s}}$ influence the slope of the phases across subcarriers, while $\tau_{\text{c}}$ and $\beta$ simply add a constant offset. These offsets can be mitigated by rotating and shifting the channel across the subcarriers for each antenna as proposed in [6]. However, by applying such a transformation for each antenna separately as suggested in [5], the AoA information that is embedded in the relationship between the phases of adjacent antennas is lost, i.e., it can not be exploited for localization. To preserve this information, we propose to apply the same transformation across the subcarriers for all antennas. Thus, with this method we obtain the fingerprint $\boldsymbol{H}_{A}$, where the value associated with the $k$-th antenna and $n$-th subcarrier is calculated as: $\boldsymbol{H}_{A}[n,k]=\boldsymbol{H}[n,k]e^{-j(\lambda n+b)}$ (5) where $\lambda$ is the reference slope of the subcarrier phases, and $b$ is the reference offset, which are determined as follows. Firstly, we fit a linear regression model for the phase over the subcarriers of each antenna, resulting in $\angle\boldsymbol{H}[n,k]=\lambda_{k}n+b_{k}+\zeta_{n,k}$ (6) where $\zeta_{n,k}$ is the statistical error of the regression model which is minimized. Thus, in contrast to [5] and [6] where the slope $\lambda$ is calculated from the difference of the phases of first and last subcarriers and the offset $b$ as the mean value of the phases across subcarrier, the parameters in (5) are calculated as $\lambda=\frac{1}{N_{\text{R}}}\sum_{k=1}^{N_{\text{R}}}\lambda_{k},\quad\quad b=b_{1}$ (7) By calculating the slope $\lambda$ this way we avoid the possibility that it will be affected by outliers. As our first method, the proposed phase alignment results in more reliable fingerprints as compared to the raw phase. This can be seen in Fig. 3(b), where the phase of $\boldsymbol{H}_{A}$ over the subcarriers associated with the first antenna in the ULA is shown for the same positions considered in Fig. 2b. ### III-C Phase Wrapping Phase wrapping is another problem that can impair the fingerprint. As can be seen in Fig. 4(a) for a given antenna and two neighboring positions of the database in [2], phase measurements close to $-\pi$ may fluctuate across the subcarriers, and in some cases the phase wraps around to $\pi$ due to noise. Fig. 4(a) shows an example of the phase wrapping issue, where the phases across the subcarriers, associated with the first antenna of the ULA of [2], for two similar positions is shown. While the phase measurements of both positions fluctuate around $-\pi$, they do not wrap around to $\pi$ in the same way, creating two different fingerprints. The most common method to address this issue is to simply unwrap the phase [5]. This approach, however, is unreliable under noisy conditions as it could lead to large phase values, since the errors accumulate with unwrapping. Such large values then dominate the fingerprint and small variations in the range $[-\pi,\pi)$ will have less influence. We propose to leverage the fact that for any angle $\theta$ we have $\exp(\text{j}\theta)=\cos(\theta)+\text{j}\sin(\theta)$, such that the information provided by $\theta$ is encoded in $\sin(\theta)$ and $\cos(\theta)$, which are continuous everywhere from $\pi$ to $-\pi$. In Fig. 4(b) and 4(c) we plot the sine and cosine of the phase $\angle\boldsymbol{H}_{D}[n,k]$, in contrast to 4(a), we see that the fingerprint quality is preserved when using $\sin{(\angle\boldsymbol{H}_{D}[n,k])}$ and $\cos{(\angle\boldsymbol{H}_{D}[n,k])}$. This indicates that the use of real and imaginary parts of the complex valued matrix can solve the wrapping problem since they can be expressed by the magnitude and the cosine and sine of the phase respectively. Additionally, this enables the phase and magnitude to be processed separately, using different and suitable techniques for each one, by using the sine and cosine to represent only the phase information. The downside is that to fully represent the phase fingerprint one must use both those functions, increasing the amount of data to be processed. These techniques can also be used on channel estimates based on ray-tracing simulations, even though there are no timing offsets. In this way, the fingerprints from simulations match the fingerprint from measurements, and can be used as an extra layer of information as in [9]. ## IV Localization using CNNs Although there are several localization schemes based on fingerprint inputs, we focus on using CNNs. For a UE at position $\boldsymbol{r}\in\mathbb{R}^{2}$, we describe the channel with the function $f$, meaning $f(\boldsymbol{r})=\boldsymbol{H}_{\star}$, where $\boldsymbol{H}_{\star}\in\\{\boldsymbol{H},\boldsymbol{H}_{D},\boldsymbol{H}_{A},\boldsymbol{H}_{A}^{\prime}\\}$ as described in Section II, with $\boldsymbol{H}_{A}^{\prime}$ being the fingerprint based on the method proposed in [5]. We will attempt to approximate the inverse function, $f^{-1}(\boldsymbol{H}_{\star})=\boldsymbol{r}$, by using CNNs, which have shown promising results for positioning [1],[3]. The reason is that CNNs have some features that could be beneficial for the considered type of inputs. ### IV-A Convolutional Neural Networks In a CNN, the input is convoluted with a matrix of smaller dimension, called the kernel. It is almost certainly followed by the pooling operation which is used to reduce the data at the output. Usually, the term CNN is used to describe a NN that uses the convolution operation at some layer. In addition to the two dimensions (antennas and subcarriers) that our input matrix has, the CNN may also use a third dimension, meaning multiple matrices can be stacked at the input. In machine learning terminology, the input matrices are called channels (not to be confused with the wireless channel). This convolutional layer leverages the idea of sparse interactions [10]. A conventional fully connected layer is learning parameters that describe the interactions between each and every one element of the input, while the CNN makes use of the smaller kernel to learn only the interactions between neighboring elements of the input. As the wireless channel between neighboring antennas and subcarriers is usually more correlated than the channels from antennas or subcarriers which are farther apart, the use of CNNs is appealing. Figure 5: Neural Network Model ### IV-B Pooling As previously described, after every convolutional layer there is a pooling layer which downsamples the output of the previous layer. The pooling function replaces that output with a summary statistic of the nearby outputs [10]. For example, the max pooling (which we consider in this work) reports the maximum output within a rectangular region of size $p_{ant}\times p_{sc}$, where $p_{ant}$ and $p_{sc}$ are the configurable sizes of the pooling in the antenna and subcarrier dimension, respectively. Besides reducing the dimension of the data, the pooling layer’s purpose is to make the output invariant to small translations of the input [10]. In our case, the two dimensions of the input matrix are antennas and subcarriers. We expect that small translations in the antenna dimension are important to be detected, since that provides the AoA information. On the other hand adjacent subcarriers within the coherence bandwidth may not provide additional information, as these subcarriers can be correlated. Thus, it may be more beneficial to pool over subcarriers, as pooling over the antennas may lead to a reduction of the angular resolution. ## V Simulation Results ### V-A Neural Network Setup We consider the CNN depicted in Fig. 5 with the input being convoluted with 32 different kernels of dimensions $4\times 4$. The resulting matrices are pooled, which is followed again by a convolution and pooling layer. The outputs of that layer are vectorized and inputted into four dense layers. The last layer has only 2 neurons expressing the position estimate. The training set was 80% of the database of [2] and test set was 20%. The training procedure starts with a batch size of 32 samples and is increased to 128, 256 and 1024 with each transition set after 30 epochs. The loss function is defined as the Euclidean mean distance of the estimated position and the real position. All the activation functions are set as the Rectified linear unit (ReLU) [10], except the last one which is linear, and the input data is normalized from 0 to 1, with respect to all the data in the training set. Lastly, we consider as a metric the mean error (ME) given by the Euclidean distance between the estimated and actual position in the test set. ### V-B Different Fingerprint Inputs TABLE I: One-Channel Input Input | ME (m) ---|--- $|\boldsymbol{H}|$ | 0.03805 $\angle\boldsymbol{H}$ | 0.04251 $\angle\boldsymbol{H}_{D}$ | 0.04088 $\angle\boldsymbol{H}_{A}$ | 0.03246 $\angle\boldsymbol{H}_{A}^{\prime}$ | 0.08142 We first show the results for one, two and three number of input channels considering magnitude and phase information. For all the following configurations the pooling layers had a dimension of $4\times 4$ ($p_{sc}=p_{ant}=4$). We define $|\boldsymbol{H}_{\star}|\in\mathbb{R}^{N_{\text{R}}\times N_{\text{C}}}$ and $\angle\boldsymbol{H}_{\star}\in\mathbb{R}^{N_{\text{R}}\times N_{\text{C}}}$ as the element-wise absolute value and angle operator, respectively, of the fingerprint matrix $\boldsymbol{H}_{\star}\in\\{\boldsymbol{H},\boldsymbol{H}_{D},\boldsymbol{H}_{A},\boldsymbol{H}_{A}^{\prime}\\}$. For the following results, we consider only the ULA antenna configuration (see Fig. 1). For one input channel of the CNN, Table I lists the ME when using the magnitude of the channel, the raw phase and the two different processed phase inputs resulting from the phase difference and phase alignment methods, proposed in Section II. We see that the phase alignment method, not only outperforms using the raw phase, but actually achieves the best performance, even better than using only the magnitude. We also observe that performing a different phase alignment for each antenna as in [5], deteriorates the performance since the relationship of the phases between antennas (i.e., AoA information) is lost. Table II presents the results with two input channels, including magnitude and phase as well as the real and imaginary part of the different considered fingerprint matrices, to address the phase wrapping. In addition, we also considered the sine and cosine of $\angle\boldsymbol{H}_{D}$, thereby using only phase information. Similar to Table I, we see that properly processing the phase largely improves the results. We also see that using Re($\boldsymbol{H}_{D}$) and Im($\boldsymbol{H}_{D}$) outperforms using the $\sin(\angle\boldsymbol{H}_{D})$ and $\cos(\angle\boldsymbol{H}_{D})$, as the former includes also magnitude information. TABLE II: Two-Channel Input Input I | Input II | ME (m) ---|---|--- $|\boldsymbol{H}|$ | $\angle\boldsymbol{H}$ | 0.03126 $|\boldsymbol{H}_{D}|$ | $\angle\boldsymbol{H}_{D}$ | 0.03810 $|\boldsymbol{H}_{A}|$ | $\angle\boldsymbol{H}_{A}$ | 0.02792 $|\boldsymbol{H}_{A}^{\prime}|$ | $\angle\boldsymbol{H}_{A}^{\prime}$ | 0.03719 Re($\boldsymbol{H}$) | Im($\boldsymbol{H}$) | 0.01809 Re($\boldsymbol{H}_{A}$) | Im($\boldsymbol{H}_{A}$) | 0.01614 Re($\boldsymbol{H}_{D}$) | Im($\boldsymbol{H}_{D}$) | 0.01316 Re($\boldsymbol{H}_{A}^{\prime}$) | Im($\boldsymbol{H}_{A}^{\prime}$) | 0.03478 sin($\angle\boldsymbol{H}_{D}$) | cos($\angle\boldsymbol{H}_{D}$) | 0.01425 Lastly, Table III provides results with three channel inputs, where we can use the magnitude of the channel with the sine and cosine of the phase of the considered fingerprints. As in previous results, the use of properly processed phase information provides the best performance. From both Table II and III we see that using the fingerprints based on matrix $\boldsymbol{H}_{D}$ while also addressing phase wrapping achieved the best performance. The small improvement when using three channels can be attributed to the fact the the CNN is able to employ different processing for phase and magnitude, and extract the relevant information in each case. TABLE III: Three-Channel Input Input I | Input II | Input III | ME (m) ---|---|---|--- $|\boldsymbol{H}|$ | sin($\angle\boldsymbol{H}$) | cos($\angle\boldsymbol{H}$) | 0.01981 $|\boldsymbol{H}_{A}|$ | sin($\angle\boldsymbol{H}_{A}$) | cos($\angle\boldsymbol{H}_{A}$) | 0.01734 $|\boldsymbol{H}_{D}|$ | sin($\angle\boldsymbol{H}_{D}$) | cos($\angle\boldsymbol{H}_{D}$) | 0.01290 ### V-C Pooling In this subsection, we analyze the impact of different pooling options $[p_{ant},p_{sc}]$ on the positioning performance, considering the ULA, URA and DIS antenna configurations from [2]. For the evaluation, we use $|\boldsymbol{H}_{D}|$, $\sin(\angle\boldsymbol{H}_{D})$ and $\cos(\angle\boldsymbol{H}_{D})$ as the input channels of the CNN shown in Fig. 5. In Table IV, we show the ME for different pooling options with $p_{ant}\times p_{sc}=4$, such that that resulting CNNs have the same complexity. For each antenna configuration, pooling over the subcarriers, i.e. [1,4], leads to the smallest ME, while the largest ME is obtained when pooling over the antennas, i.e. [4,1]. For a given pooling option, the best performance is achieved with the distributed antenna configuration, as it collects CSI at distinct locations around a UE’s position. On the other hand, the worst performance results by using the URA, since its resolution on the horizontal plane where the UE lies, is smaller compared to the other antenna configurations. The results in Table IV suggest that it is more beneficial to pool over the subcarriers than over the antennas. Thus, in Fig. 6 we examine the ME with pooling option $[1,p_{sc}]$ for different values of $p_{sc}$. For each antenna configuration, we see there is an optimum pooling $p_{sc}$ over the subcarriers, which we posit that it depends on the coherence bandwidth of the channel. The lowest ME attained for each of the ULA, DIST and URA antenna configurations are 6.11 mm, 5.20 mm and 9.81 mm respectively, which is lower than the ones reported in [2]. This was achieved by using the optimal size of pooling $p_{sc}$, which is different for each antenna configuration, and the three channel input: $|\boldsymbol{H}_{D}|,\sin(\angle\boldsymbol{H}_{D}),\cos(\angle\boldsymbol{H}_{D})$. TABLE IV: Different Pooling Options Pooling [$p_{ant}$, $p_{sc}$] | Antenna Configuration | ME (m) ---|---|--- [1,4] | ULA | 0.00612 [2,2] | ULA | 0.01010 [4,1] | ULA | 0.01633 [1,4] | distributed | 0.00521 [2,2] | distributed | 0.00732 [4,1] | distributed | 0.00982 [1,4] | URA | 0.01096 [2,2] | URA | 0.01183 [4,1] | URA | 0.01908 Figure 6: ME for different pooling dimensions ## VI Conclusion We have examined the use of CSI over multiple antennas and subcarriers, as fingerprint inputs of a CNN for UE localization. As the raw phase of the estimated channel cannot be used as a consistent fingerprint, we have presented different methods for producing reliable fingerprints based on phase information. Although the proposed methods have been evaluated with CNNs, they can also be used for other localization schemes based on fingerprints. For different number of inputs of a CNN, simulation results have shown that UE localization can be improved with properly processed phase information. We have also investigated the impact of different pooling options on the positioning performance with CNNs, showing that it is more beneficial to pool over the subcarriers than over the antennas. Simulation results have shown there is an optimum pooling size over the subcarriers, whose dependency on the coherence bandwidth is part of future work. ## References * [1] M. Widmaier, M. Arnold, S. Dörner, S. Cammerer, and S. Brink, “Towards practical indoor positioning based on massive MIMO systems,” _2019 IEEE 90th Vehicular Technology Conference (VTC2019-Fall)_ , 2019. * [2] S. D. Bast, A. P. Guevara, and S. Pollin, “CSI-based positioning in massive MIMO systems using convolutional neural networks,” _2020 IEEE 91st Vehicular Technology Conference (VTC2020-Spring)_ , 2020. * [3] A. Niitsoo, T. Edelhäußer, E. Eberlein, N. Hadaschik, and C. Mutschler, “A deep learning approach to position estimation from channel impulse responses,” _Sensors (Basel, Switzerland)_ , vol. 19, 2019. * [4] A. Sobehy, É. Renault, and P. Mühlethaler, “CSI based indoor localization using ensemble neural networks,” in _MLN_ , 2019. * [5] X. Wang, L. Gao, and S. Mao, “CSI phase fingerprinting for indoor localization with a deep learning approach,” _IEEE Internet of Things Journal_ , vol. 3, no. 6, pp. 1113–1123, 2016. * [6] S. Sen, B. Radunovic, R. Choudhury, and T. Minka, “You are facing the Mona Lisa: spot localization using PHY layer information,” in _MobiSys ’12_ , 2012. * [7] M. Speth, S. A. Fechtel, G. Fock, and H. Meyr, “Optimum receiver design for wireless broad-band systems using OFDM. I,” _IEEE Transactions on Communications_ , vol. 47, no. 11, pp. 1668–1677, 1999. * [8] Y. Xie, Z. Li, and M. Li, “Precise power delay profiling with commodity Wi-Fi,” _IEEE Transactions on Mobile Computing_ , vol. 18, no. 6, pp. 1342–1355, 2019. * [9] M. N. de Sousa and R. S. Thomä, “Enhancement of localization systems in NLOS urban scenario with multipath ray tracing fingerprints and machine learning,” _Sensors (Basel, Switzerland)_ , vol. 18, 2018. * [10] I. Goodfellow, Y. Bengio, and A. Courville, _Deep Learning_. MIT Press, 2016, http://www.deeplearningbook.org.
# Artificial Intelligence Prediction of Stock Prices using Social Media Kavyashree Ranawat Durham University School of Engineering and Computing Sciences Durham University Lower Mountjoy, South Rd, Durham DH1 3LE, United Kingdom <EMAIL_ADDRESS> &Stefano Giani Durham University School of Engineering and Computing Sciences Durham University Lower Mountjoy, South Rd, Durham DH1 3LE, United Kingdom <EMAIL_ADDRESS> ###### Abstract The primary objective of this work is to develop a Neural Network based on LSTM to predict stock market movements using tweets. Word embeddings, used in the LSTM network, are initialised using Stanford’s GloVe embeddings, pretrained specifically on 2 billion tweets. To overcome the limited size of the dataset, an augmentation strategy is proposed to split each input sequence into 150 subsets. To achieve further improvements in the original configuration, hyperparameter optimisation is performed. The effects of variation in hyperparameters such as dropout rate, batch size, and LSTM hidden state output size are assessed individually. Furthermore, an exhaustive set of parameter combinations is examined to determine the optimal model configuration. The best performance on the validation dataset is achieved by hyperparameter combination 0.4,8,100 for the dropout, batch size, and hidden units respectively. The final testing accuracy of the model is 76.14%. _K_ eywords LSTM $\cdot$ Twitter $\cdot$ Stock Prediction $\cdot$ APPLE $\cdot$ Neural Networks $\cdot$ VADER ## Introduction Twitter is a microblogging and social media platform that allows users to communicate via short messages (280 characters) known as tweets [1, 2, 3]. It enables millions of users to express their opinions on a daily basis on a variety of different topics ranging from reviews on products and services to users’ political and religious views, making Twitter a potent tool for gauging public sentiment [4]. Thus, it manifestly follows that twitter data can be regarded as a corpus, forming the basis on which predictions can be made, and researchers have indeed exploited this fact to seek trends by performing numerous and varied analyses. A characteristic feature of the stock market is volatility and there is no general equation describing the prediction of stock prices, which is a complex function of a range of different factors. The methods of stock market prediction can be broadly classified into Technical Analysis and Fundamental Analysis [5]. The latter involves the consideration of macroeconomic factors as well as industry specific news and events to guide investment strategies [5]. The analysis of public sentiment via tweets performed in this project can be regarded as an aspect of Fundamental Analysis. Although the prediction of stock prices is highly nuanced, the Efficient Market Hypothesis (EMH), propounded by Eugene Farma in the 1960’s, suggested a relation between public opinion and stock prices [6]. The semi-strong form of the EMH implies that current events and new public information have a significant bearing on market trends [1, 6]. This view is supported by Nosfinger who draws upon evidence from several studies in the field of Behavioural Finance to reinforce that changes in aggregate stock price as well as the high degree of market volatility can be, in part, attributed to public emotion [7]. Numerous studies have been successful in unveiling and proving the perceived existence of a relationship between public mood gathered from social media and stock market trends [7, 8, 9, 10, 11, 12, 13, 14, 15]. The vast majority of ML (Machine Learning) techniques applied in this sector have integrated the characteristics of NLP (Natural Language Processing) to extract and quantify the sentiment of public opinion expressed via social media. SVM (Support Vector Machines) [8], Random Forest [9], and KNN (K-Nearest Neighbour) [11] classifiers have yielded impressive classification accuracies in this application. The only caveat is that most studies consider the compound effect of historic prices and public sentiment, thereby discounting the exclusive impact of sentiment. Some studies have gone beyond the classic ML approach, employing deep learning methods such as ANNs (Artificial Neural Networks) in one form or another for the purposes of making predictions [12, 13, 14, 15, 16, 17]. Bollen et al have established that collective public mood is predictive of DJIA (Dow Jones Industrial Average) closing values by making use of Granger Causality Analysis and SOFNN (Self-Organized Fuzzy Neural Networks). Many researchers have built on this work and others have explored alternate deep learning models such as MLP (Multi Layer Perceptron), CNN (Convolutional Neural Network) + LSTM (Long Short Term Memory) for market prediction. The primary focus of this work was on the development of a variant of RNN (Recursive Neural Network), known as LSTM, capable of predicting short-term price movements. Owing to the volatile and unpredictable nature of the stock market, it is plausible that the relationship between the societal mood and economic indicators perhaps is more complex and nuanced than linear. Deep learning methods are felicitous for this application in that hidden layers can exploit the inherent relational complexity and can potentially extract these implicit relationships. It is for this reason that an LSTM structure was selected as the principal model in this work. The popularity of RNNs in NLP and stock prediction tasks is attributed to the fact that they consider the temporal effect of events which is a significant advantage over other NNs (Neural Networks). With the aid of a popular sentiment analysis tool, known as VADER, the degree of correlation between the sentiment expressed via tweets and stock price direction was also investigated for the purposes of comparison with the results from the LSTM architecture. ## 1 VADER Implementation VADER is a state of the art technique employed by researchers in sentiment analysis tasks. One aspect of this work involves using VADER to explore the degree of correlation between public opinion and sentiment expressed via twitter and stock market direction. As aforementioned, VADER is a gold standard lexicon and rule-based tool for sentiment analysis [18, 19]. Developed and empirically validated by Hutto and Gilbert, the VADER lexicon is characteristically attuned to text segments in the social media domain [20, 21]. Unlike other lexicon approaches, VADER takes into account that microblog text often contains slang, emoticons, and abbreviated text [21]. It not only provides the semantic orientation of words but also quantifies sentiment intensity by considering generalisable heuristics such as word order, capitalisation, degree modifiers etc [21]. In this application, VADER was used to generate the polarity scores of tweets, including a compound score (normalised between 1 and -1) which reflects the combined effect of the degree of positivity, negativity, and neutrality expressed in a tweet. ### 1.1 Experimental Procedure The aim is to investigate the correlation between two variables; VADER scores and stock market trajectory. Firstly, tweets containing the APPLE stock ticker symbol were cleaned, using the algorithm described in the next section. Stock data associated with APPLE, among the Big Four technology companies, is deemed as a suitable choice upon which to perform analysis for several reasons; a detailed rationale is provided in a subsequent section. The compound score for each tweet was generated using VADER. Subsequently, the average of the scores for all tweets in a single day was taken. To obtain values for the second variable i.e. stock market movement, the direction of stock price movement was quantified. If the next-day close price of the security is greater than that of the current day, the value for that day is defined as 1, else it is defined as 0. After generating the values for both variables, a special case of the Pearson coefficient, known as Point biserial correlation coefficient, was applied on the data to determine the correlation. This metric is commonly used when one variable is continuous and the other is categorical, as is the case in this application, where the VADER scores are continuous whereas the stock price change data is dichotomous (binary) [22]. The biserial correlation method, however, requires the continuous variable to be normally distributed [22]. Therefore, the distribution of the VADER scores was plotted and a roughly normal distribution was obtained (as shown in figure 1), allowing the use of the biserial correlation method. Figure 1: Distribution of VADER scores (continuous variable). ### 1.2 Further Modifications Modifications were made to the stock price change data by redefining the classification strategy. Previously, it was based on a delay of 1 day. It is plausible, however, that the effect of information and opinions presented in tweets may take longer to manifest and reflect in the asset price. To explore this theory, a delay size of days in the range [1,7] was taken. For example, a delay size of 2 indicates that if the asset price at the close of the trading day 2 days hence is higher, it is classed as 1, else it is classed as 0. Figure 2: Correlation coefficient values for different delay periods. Another configuration was assessed after making modifications to the VADER scores. If a particular tweet has more retweets than others i.e. it has been frequently shared by users, it could indicate that the information contained in that tweet has a notable influence on other users. These tweets could potentially play a greater role in impacting an incremental change within the stock market. By taking a simple average of the compound scores of tweets to arrive at a singular value for a specific day, it is not possible to gauge the contribution of important tweets. Thus, based on the number of retweets per tweet, a weighted average of the compound scores was taken to ensure accurate representation of all tweets. The correlation was then determined using the modified scores. ### 1.3 Results The correlation obtained for variable delay configurations is shown graphically in figure 2. It is evident that the correlation for the original configuration using a delay of 1 day is 0.0812 (or 8.12%), suggesting a notably poor linear relationship between the sentiment expressed in tweets and market movement. Using delays higher than 1 generally tend to improve the degree of correlation, with a delay of 4 days yielding the highest correlation (30.37%). This result supports the assumption that there exists a lag between the release of public opinion and its consequent reflection in stock price. A decreasing trend is observed as the delay size is increased beyond 4 days. A potential implication of this could be that the information contained on a particular day is irrelevant after large delay periods and no longer has any bearing on stock market values. The retweets-based weighted average configuration results in a small negative correlation value of -8.87%, which is in complete contrast to the original configuration. There is ambiguity in what the underlying cause of the generated results could be. There is a possibility that the inclusion of retweets has no impact in moulding future stock values. Alternatively, it could also be that the lack in the number of data points is masking the true contribution of the retweets. It can be concluded from the results that VADER is not able to present any strong association between public sentiment and market trajectory. The low correlation values could be attributed to the use of an insufficient number of samples for both variables. The final VADER scores are derived by taking an average of all compound scores, which is a simplistic and naive approach. The inability of the lexicon-based tool to identify any notable relationships between the variables could also be due to the inadequacies inherent in the development of VADER sentiment. However, as mentioned earlier, there is a high possibility that the relationship between the variables is not linear and is perhaps more complex and nuanced. This could be one of the reasons why using a metric such as correlation, used to assess strength of a linear relationship, is not able to detect any significant associations. ## 2 Neural Network Model ANNs, inspired by the behaviour of neurons in biological systems, are a dense interconnection of nodes or pre-processing units connected in layers, which have the ability to discover complex relationships between inputs and outputs [23]. There are many varying implementations of ANNs, which differ in terms of network architecture, properties and complexity. Considering the sequential nature of tweets, it is essential to use a network which considers the temporal effect of an input sequence. RNNs satisfy this criterion and are indeed suitable for application in tasks of this nature. However, the main drawback of RNNs is their inability in capturing long term dependencies in an input sequence i.e. the Vanishing Gradient problem [24, 25]. For example, when an input sentence is fed into the network, the error must be backpropagated through the network in order to update the weights. If the input is long, the gradients diminish exponentially during backpropagation, resulting in virtually no contribution from the state in earlier time steps. It is particularly problematic when using the sigmoid activation function as its derivative lies in the range [0, 0.25], resulting in highly diminished gradient values after repeated multiplication. A variant of the vanilla RNN network, the LSTM, can overcome this limitation [24, 25], albeit not entirely, and is used in this application. Figure 3 shows a high level abstraction of the LSTM model architecture, consisting of one stacked LSTM layer. The input side shows the vectors, concatenated in the Embedding Layer, being input to the LSTM cells in the stacked layer. The output side depicts the hidden state outputs of the LSTM layer being taken as inputs by a sigmoid-activated node to make output predictions. Figure 3: General schematic of the LSTM Network. The input side shows the vectors, concatenated in the Embedding Layer, being input to the LSTM cells in the stacked layer. The output side depicts the hidden state outputs of the LSTM layer being taken as inputs by a sigmoid-activated node to make output predictions. ### 2.1 Preprocessing Raw tweets contain a great deal of noise which needs to be eliminated in order to extract relevant information from tweets and improve the predictive performance of the applied algorithm. Tweets contain twitter handles, URLs, numeric characters, and punctuation which do not contribute meaningfully to the analysis. A preprocessing algorithm was applied to remove these elements, convert words to lower case, and tokenize the words present in a tweet. The cleaned tweets were concatenated so as to form one input sequence for a given day. Ordinal or integer encoding was used to map all the words in the vocabulary to an integer value, resulting in an input vector $\\{w_{1},w_{2},w_{3},…,w_{n}\\}$, where $w_{t}$ corresponds to a unique integer index representing a feature (or word) in the vocabulary. Although LSTMs can take inputs of variable length [26], post-padding was applied as only vectors with homogeneous dimensionality can be used with the Keras Embedding Layer. For each input vector of length $t\in[1,n]$, $n-t$ zeros or dummy features must be appended, where $n$ is the length of the longest encoded vector. This produces vectors in an $m$-dimensional feature space, where $m$ is the total number of unique samples i.e. number of unique phrases or tweets fed to the network. ### 2.2 Embedding NLP tasks for textual representation and feature extraction commonly use BoW (Bag of Words) models owing to their flexibility and simplicity. The traditional BoW has two variants: N-gram BoW and TF-IDF (Term Frequency- Inverse Document Frequency). The former reduces dimensionality of the feature set by extracting phrases comprising $N$ words while the latter considers the frequency of words whilst considering the effect of rare words. The main drawback is that BoW fails to take into account word order and context and results in sparse representations. This work explores the use of GloVe (Global Vectors for Word Representation) which has gained momentum in text classification problems [27, 28]. GloVe overcomes the sparsity problems associated with the BoW model by generating dense vector representations and projecting the vectors to a markedly lower dimensional space. It has the ability of capturing the semantic and syntactic relationships that are present between words, where words with a similar meaning are locally clustered in the vector space. Stanford’s GloVe embeddings, trained specifically on 2 billion tweets, were used to project each feature as a 200-dimensional vector [27, 28]. The weights of the feature vectors were initialised using the pre-trained embeddings but adjusted with the progression of training to improve classification performance. Each integer encoded feature, $w_{t}$, corresponds to an embedding vector, $\textbf{x}_{t}$, where $\textbf{x}_{t}\in\mathbb{R}^{p}$. Owing to the fact that each feature is represented as a 200-dimensional vector, $p=200$. A padded input feature vector is thus represented in the embedding layer as $\textbf{X}\in\mathbb{R}^{np}$, formed as a result of concatenation of $m$ vector embeddings. In figure 3, although the embedding layer is not explicitly shown, the vector embeddings which constitute it can be seen as inputs to the LSTM layer at the corresponding time steps. ### 2.3 LSTM Layer The main distinction between a neuron in the vanilla RNN and an LSTM cell lies in the presence of a cell state vector, whose contents at each time state are maintained and modified via an LSTM gating mechanism [29, 30, 31]. The information flow in an LSTM memory cell is regulated by three primary gates viz. forget gate, input gate, and output gate [29, 30]. Figure 4 shows the schematic of an LSTM cell, including the gating mechanisms used to achieve its functionality. Figure 4: Structure of an LSTM cell, showing the role of the primary gates and flow of information within the cell to form the current memory state and hidden output state. At a certain time step, $k$, the vector embedding, $\textbf{x}_{k}$, along with the previous hidden output, $\textbf{h}_{k}$, will be used by the input and forget gates to update the internal state of the cell [29, 32]. The output gate combines the inputs and the current cell state, $\textbf{c}_{k}$, to determine the information to be carried over to the next cell in the repeating structure [29]. In this fashion, LSTMs can control the contribution of those words and word relationships that have a higher impact on prediction, whilst penalising those that are less significant. Equation 1 is a matrix representation [32] of the outputs of the gates. Equations 2 and 3 show the outputs of the cell, where the output vector is obtained by an element wise multiplication process [29, 32]. $\begin{pmatrix}i\\\ f\\\ o\\\ g\\\ \end{pmatrix}=\begin{pmatrix}\sigma\\\ \sigma\\\ \sigma\\\ \tanh\\\ \end{pmatrix}W\begin{pmatrix}x_{k}\\\ h_{k-1}\par\end{pmatrix}$ (1) $\displaystyle c_{k}$ $\displaystyle=f\odot c_{k-1}+i\odot g$ (2) $\displaystyle h_{k}$ $\displaystyle=o\odot\tanh c_{k}$ (3) where $\textbf{i},\textbf{f},$ and o are the outputs of the input gate, forget gate, and output gate respectively and g is the output of an additional gate which aids in updating cell memory. W is the weights matrix and $\sigma$ and $\tanh$ represent the sigmoid and $\tanh$ non linearities. Note that the system of equations in (1) also contains a bias term for each gate output. Dropout, a regularization method, is utilized to prevent the model from overfitting [25]. Overfitting occurs when the model learns the statistical noise present in the dataset, capturing unnecessary complex relationships and thus, resulting in decreased generalisability. During training, it is possible for neighbouring neurons to become co-dependent, inhibiting the effectiveness of individual neurons. Dropout causes a proportion of the nodes or outputs in the layer to become inactive, thereby forcing the model to become more robust. This results in an increase in the network weights, which must be scaled by the dropout rate after completion of training. ### 2.4 Output Layer A singular output node with a sigmoid activation function, presented in equation 4, was used for the purpose of classifying trend. $\sigma(z)=\frac{1}{1+e^{-z}}$ (4) where $z$ is the activation of the output node. The estimated probability returned by the node was compared against a threshold probability in order to perform binary classification. If the output probability for a given input sequence $\sigma(z)\geq 0.5$, the input was labelled as 1, predicting an increase in asset price for the following trading day. If the output probability did not exceed this threshold, the input was labelled as 0, indicating either no change or a decrease in next-day price. A binary cross entropy cost function, $J_{bce}$, was used as given by equation 5 [33]. $\displaystyle J_{bce}$ $\displaystyle=-\frac{1}{m}\sum_{j=1}^{m}[y_{j}\times log(\sigma(z_{j}))$ (5) $\displaystyle+(1-y_{j})\times log(1-\sigma(z_{j}))]$ where $y_{j}$ is the $j^{th}$ target variable or the actual class label from a set of $m$ training samples. The cost or error function is representative of how accurately the model predicts target values, using a given set of network parameters. The main aim is to optimise or minimise the cost function, updating the weights and biases of connections in the network as a result. A mini-batch Stochastic Gradient was used during backpropagation to allow the model to converge to a global cost minimum [25]. This optimum state represents a model configuration where the error between the actual and predicted values is at a minimum and the network can successfully detect patterns between word embeddings essential for classification. The learning rate determines the rate at which the tunable weights approach the global minimum and must be chosen judiciously. A very large value would risk overshooting the minimum and a learning rate that is too small will significantly delay convergence. ## 3 Experimental Procedure Twitter data was obtained from followthehashtag [34], an online resource containing a readily available corpus of tweets. Approximately 167,000 tweets mentioning or associated with APPLE stocks were used for analysis. APPLE is among the companies currently dominating the technological sector and is regarded as a suitable choice upon which to base analysis. Owing to its popularity and the fact that it has the largest market capitalisation out of all NASDAQ 100 companies, it is fair to assume that twitter contains sufficient information relating to its stocks. The stock price data was sourced from Yahoo Finance [35]. The granularity of stock data considered is 1 day i.e. daily changes in stock price were computed to capture the essence of short term price fluctuation. The input tweets were labelled according to the scheme described previously. Concatenation of tweets leads to an aggregate of 48 input samples, corresponding to 48 unique days. Due to the limited number of samples, each input was divided into 100 subsets, whilst keeping the labelling of the subsets consistent with that of the original day. Subsequently, 4800 input samples for the network were obtained. For the initial experimentation, a naive model configuration was used. This model forms the basis on which further improvements in performance can be achieved. The next section discusses the effects of using different network types, hyperparameter optimisation, and varying split values (for tuning the number of input samples) on model performance. For the initial configuration, consisting of 4800 samples, the training/testing/validation split was 70/20/10 i.e. training was performed on 3360 samples, testing was performed on 960 samples, and validation was performed on the remaining 480 samples. The validation set is used to configure the model so as to obtain the hyperparameters which give the best performance. The testing data is only used once after the network has been configured to give an unbiased evaluation of model performance. A single LSTM stacked layer was utilised with a dropout value of 0.2 and 100 hidden units (used for determining the dimension of the LSTM outputs). The gradients and weights are updated according to a batch size of 32. The performance of the initial configuration is reported in the next section. The primary metric used to assess performance is accuracy. Another commonly used metric is the F1 score, which is the harmonic mean of the precision and recall [36]. Precision refers to the percentage of instances correctly predicted as positive with respect to all instances classified as positive by the model, whereas recall refers to the percentage correctly classified as positive out of all positive classes [36]. The F1 score is also calculated for varying implementations discussed in the next section . However, it is only needed when there is a greater cost associated with either the false positives or false negatives. As this is not applicable for this task and class distribution is even (as shown in figure 5), it is only computed to ensure consistency in results and thus not reported for all configurations. Figure 5: Bar graphs representing an even class distribution ($\approx$ 58/42) of stock up or down in the original dataset. ## 4 Results The initial model configuration, described in the previous section, gives an impressive classification (testing) accuracy of 74.58%. A confusion matrix, displayed in figure 6, summarizes the classification performance of the model. Figure 6: Confusion Matrix showing classifier performance at the deep functional level. It indicates that the model is able to correctly identify the 0 class and 1 class with an accuracy of 76% and 73% respectively. The F1 score for this configuration is 71.76%, reinforcing model performance as reflected in the testing accuracy result. The achieved accuracy is far superior to the random guessing threshold of 50%. This result indicates the effectiveness of NNs in this task, which is in contrast with the results of the correlation analysis performed previously. This output reinforces the claim that NNs have the ability to detect nuanced patterns and produce complex mappings between the input and output. ### 4.1 Effect of Splitting Dataset In the original model, the concatenated tweets, resulting in 48 samples, were split into 100 subsets per input sample to augment the dataset. To investigate the effect of modifying the number of subsets per sample on overall performance, values for the input splits were selected in the range [25, 450]. Figure 7 shows the dependence of classification accuracy as a function of split size. In general, selecting large values of split size leads to performance degradation. Splitting up all tweets on a particular day into higher subsets can result in insufficient information contained within a unique sample, deteriorating prediction capability. It is reasonable to assume that on any given day, some tweets cause the price to go in the opposite direction to that observed in the stock market, however, the aggregate impact of other tweets outweigh this effect. As a result, higher subdivisions do not accurately capture the true nature of the task. To determine if this trend continues, an extreme case was considered i.e. the maximum logical split value was considered. Using all 60,233 filtered tweets as individual inputs to the network, the observed accuracy was 62.37%, validating the observed graphical results. On the other end of the spectrum, using the concatenated tweets in their unaltered form will reflect all the necessary information on a given day for the model to make predictions. However, in this work, this will entail using a considerably limited number of training instances, hindering the network’s ability to learn effectively and leading to erroneous outputs. The training results of this configuration further confirmed this intuition as it gave the worst performance in comparison with using other subset values. The training time for this configuration i.e. using no splits was also significantly higher than any other value tested. As the input sequence length is maximum in this case, a significant number of LSTM cells is required. This discernibly increases processing time and degrades performance due to the emergence of vanishing gradients. Therefore, there exists a trade-off between loss of information and creating a reasonably sized dataset. In light of this fact, a split size of 150 per day was selected for subsequent analyses as it is able to achieve a satisfactory balance of the aforementioned performance variables. Figure 7: Variation of classification accuracy with the split size per input sample. ### 4.2 Hyperparameter Optimisation To improve the results of the original model configuration, the hyperparameters employed in the NN were optimised. In particular, the dropout rate, batch size, and number of LSTM output hidden units were varied to investigate their impact on classification results. Each hyperparamter was considered in isolation, with all other model variables remaining unchanged, to determine its exclusive impact. The same experimental procedure was also applied to a different network configuration known as bidirectional LSTMs. In strict terms, the standard LSTM structure used in this work is called a unidirectional LSTM network which differs from the bidirectional LSTM structure. The bidirectional network is a variant of the classical LSTM, where information flows in both directions between LSTM cells such that the cell, at every time step, is able to maintain previous and future input information [31]. This is in contrast to the unidirectional structure, where each cell contains only past information. Table 1 shows the classification accuracies achieved by varying the network dropout percentage. It is apparent that the unidirectional structure tends to perform well on dropout values higher than 0.2. However, the bidirectional structure shows no notable trends and as such, no conclusive inference can be drawn. The best accuracy (78.12%) is achieved by the unidirectional architecture, using a dropout value of 0.3. The implications of utilising higher values of dropout are that a higher proportion of neurons become inactive, increasing the robustness of the model and resulting in better testing accuracies. Table 1: Effect of varying dropout rates Value | Unidirectional | Bidirectional ---|---|--- | Accuracy | Epoch | Accuracy | Epoch 0.2 | 74.58 | 6,9 | 75.63 | 7 0.3 | 78.12 | 4 | 74.79 | 3 0.4 | 77.29 | 7 | 72.71 | 8 0.5 | 76.25 | 6 | 75.63 | 3 0.6 | 75.00 | 5,10 | 76.46 | 7 Table 2: Effect of varying batch size Value | Unidirectional | Bidirectional ---|---|--- | Accuracy | Epoch | Accuracy | Epoch 8 | 78.33 | 2 | 77.92 | 8 16 | 78.12 | 2 | 75.00 | 1 32 | 74.58 | 6,9 | 75.63 | 7 64 | 74.79 | 6 | 72.29 | 6 128 | 73.54 | 8 | 73.12 | 3 Table 3: Effect of varying lstm output hidden units Value | Unidirectional | Bidirectional ---|---|--- | Accuracy | Epoch | Accuracy | Epoch 100 | 74.58 | 6,9 | 75.63 | 7 128 | 74.38 | 6 | 77.08 | 4 256 | 75.42 | 3 | 75.21 | 10 512 | 76.46 | 4 | 74.58 | 2 Table 2, which highlights the effect of variable batch sizes, indicates that there is a decline in the accuracy level with increasing batch size for the unidirectional structure. Mini-batch sizes of 8 and 16 yield comparable results, with batch size 8 achieving a remarkable accuracy of 78.33%. Similarly to the dropout case, there are no identifiable trends for the bidirectional structure, however the lowest batch size (8) performs the best for this variant of LSTM as well. The batch size determines the number of training instances after which the gradients and weights of the network are updated. The impressive performance of the lower batch sizes can be attributed to a more robust convergence as well as the network’s ability to circumvent local minima. The final parameter used for optimisation is the number of hidden state units of the LSTM cells. The hidden state units determine the dimensionality of the output space of the LSTM layer i.e. the dimensionality of the LSTM output vectors $\textbf{H}=\\{h_{1},h_{2},h_{3},...,h_{n}\\}$. As presented in Table 3, the accuracy shows an overall increase with an increase in the output dimensionality for the unidirectional framework. Although performance gains are observed upon using larger values of hidden state size, it is at the cost of an exponential increase in the number of parameters and processing time. Values greater than 512 are not used in this study as the resulting models will be prone to overfitting owing to their marked complexity [37]. The bidirectional variant performs better when 128 hidden state units are used however results in a performance degradation for higher values. Due to the inherent characteristics of the bidirectional model, the dimensionality of the LSTM output is double that of its unidirectional counterpart. This ultimately leads to an inordinately complex model that is more prone to overfitting. Some neural network structures are known to achieve satisfactory results by employing two hidden layers. Therefore, two identical LSTM stacked layers with the same hyperparameter values as the original configuration were integrated in the network. The classification accuracy achieved by this configuration was 75.10%, which is lower than the value (76.67%) achieved using a single hidden layer implementation. Hence, it is deemed apposite to forego such an architecture. A notable observation, based on the hyperparameter tuning results, is the performance of the bidirectional structure. It has the ability to preserve past and future values in each cell, thereby allowing the network to gain a fuller context of the information present in the input tweets. In theory, this should lead to improvements in the overall predictions made by the model. However, not only does the bidirectional implementation produce comparable results overall but it also occasionally generates lower accuracies than the unidirectional model. ### 4.3 Optimal Model In order to discover the optimal model configuration, experiments were conducted using an exhaustive set of combinations of the hyperparameters. Different combinations of the dropout rate, mini-batch size, and hidden state size were deployed, with the range of hyperparameter values limited to the those outlined in Table 1, Table 2, and Table 3. The combination 0.4 (dropout), 8 (batch size), and 100 (hidden units) produces the best results, obtaining a validation accuracy of 81.04%. Improvements in accuracy cannot be attained by merely using those hyperparameter values which give the highest accuracy when altered independently i.e. using the combination 0.3,8,512. Therefore, it can be argued that there exists some degree of interaction between the variables when varied simultaneously. The optimal model configuration is thus given by the combination 0.4,8,100. To perform an unbiased evaluation of the model, the testing data (960 samples) was used. The model produces a testing accuracy of 76.14% when presented with the unseen testing data. Although the model exhibits a remarkable performance in absolute terms, its results are specific to APPLE and are not generalisable to other technology companies. There could be different patterns and inherent complexities within the twitter datasets of other companies which could lead to similar or contrasting results to that observed in this analysis. ## 5 Conclusion The objective of this work is to develop a model capable of predicting the direction of next-day stock market fluctuations using twitter messages. Tweets associated with APPLE, regarded among the Big Four technology companies, is used as the basis for this analysis. The primary focus of this work is in the development, configuration, and deployment of an LSTM structure. A correlation analysis is briefly explored to determine the relationship between VADER scores, quantifying the sentiment of tweets, and stock market movement. Using Point biserial correlation coefficient as the measurement metric, a low correlation value of 0.0812 is obtained. Alternate configurations are considered based on the time taken for information contained in tweets to manifest in market movement and a retweet-weighted average configuration. A delay size of 4 days results in the highest correlation value (0.3037). However, it is apparent from these results that VADER is not able to extract any strong relations between societal sentiment and market direction. In order to initialise the weights of the LSTM network, GloVe embeddings, pre- trained on a sizeable twitter corpus, are utilised. Due to the limited number of training samples, the effect of splitting the dataset for augmentation is analysed. A value of 150 is selected for splitting the input sequence for each day into subsets as it provides a satisfactory trade-off between information loss and a suitable representation of dataset size. Hyperparameter tuning is performed using the validation set and independently varying the dropout rate, batch size, and hidden unit size to further optimise model performance. To determine the optimum configuration, an exhaustive set of varying combinations of selected parameters is tested. A combination of 0.4,8,100 performs the best on the validation set, achieving a testing accuracy of 76.14%. Despite the level of accuracy being an impressive standalone result, twitter datasets from other technological companies need to be analysed and contrasted with results from this study. This relative comparison will allow the LSTM network performance to be gauged more accurately and in a broader context, enabling the formation of generalisable results. The application of technical indicators such as historical price data can also be explored in conjunction with the components of Fundamental Analysis used in this task to provide more input information vital to classification. Word embeddings are effective in projecting words/features occurring in similar contexts within a neighbouring vector space. However, it is common for tweets to contain words expressing opposite sentiments that are collocated. This leads to erroneous representations of these fundamentally different words as similar vectors, mitigating their discriminative ability as required for classification [38]. Further work should be undertaken to incorporate linguistic lexicons such as SentiNet to capture the effect of word similarity and sentiment [27]. An alternative approach is to use SSWE (Sentiment-Specific Word Embeddings) which injects sentiment information into the loss function of neural networks [27]. This could potentially boost performance though the enhancement of the quality of word vectors. ## References * [1] Narendra Babu Anuradha Yenkikar, Manish Bali. Emp-sa: Ensemble model based market prediction using sentiment analysis. International Journal of Recent Technology and Engineering (IJRTE), 8(2), 2019. * [2] Selena Larson. Welcome to a world with 280-character tweets. https://money.cnn.com/2017/11/07/technology/twitter-280-character-limit/index.html. Accessed: 07 November 2019. * [3] Twitter. How to tweet. https://help.twitter.com/en/using-twitter/how-to-tweet. Accessed 03 March 2020. * [4] Alexander Pak and Patrick Paroubek. Twitter as a corpus for sentiment analysis and opinion mining. In Proceedings of LREC, volume 10, January 2010. * [5] N.B. GKumar and S. Mohapatra. The Use of Technical and Fundamental Analysis in the Stock Market in Emerging and Developed Economies, chapter Introduction. Emerald Group Publishing Limited, 2015. * [6] Eugene F. Fama. The behavior of stock-market prices. The Journal of Business, 38(1):34–105, 1965. * [7] John R. Nofsinger. Social mood and financial economics. Journal of Behavioral Finance, 6(3):144–160, 2005. * [8] John Kordonis, Symeon Symeonidis, and Avi Arampatzis. Stock price forecasting via sentiment analysis on twitter. In Proceedings of the 20th Pan-Hellenic Conference on Informatics, PCI ’16, New York, NY, USA, 2016. Association for Computing Machinery. * [9] V. S. Pagolu, K. N. Reddy, G. Panda, and B. Majhi. Sentiment analysis of twitter data for predicting stock market movements. In 2016 International Conference on Signal Processing, Communication, Power and Embedded System (SCOPES), pages 1345–1350, 2016. * [10] V. Kalyanaraman, S. Kazi, R. Tondulkar, and S. Oswal. Sentiment analysis on news articles for stocks. In 2014 8th Asia Modelling Symposium, pages 10–15, 2014. * [11] Ayman Khedr, S.E.Salama, and Nagwa Yaseen. Predicting stock market behavior using data mining technique and news sentiment analysis. International Journal of Intelligent Systems and Applications, 9:22–30, July 2017. * [12] Johan Bollen, Huina Mao, and Xiao-Jun Zeng. Twitter mood predicts the stock market. Journal of Computational Science, 2, October 2010. * [13] Franco Valencia, Alfonso Gómez-Espinosa, and Benjamin Valdes. Price movement prediction of cryptocurrencies using sentiment analysis and machine learning. Entropy, 21:1–12, June 2019. * [14] Zhigang Jin, Yang Yang, and Yuhong Liu. Stock closing price prediction based on sentiment analysis and lstm. Neural Computing and Applications, September 2019. * [15] Xiao Ding, Yue Zhang, Ting Liu, and Junwen Duan. Using structured events to predict stock price movement: An empirical investigation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1415–1425, Doha, Qatar, October 2014. Association for Computational Linguistics. * [16] Kyoung jae Kim, Kichun Lee, and Hyunchul Ahn. Predicting corporate financial sustainability using novel business analytics. Sustainability, 11(1):1–17, December 2018. * [17] Evita Stenqvist and Jacob Lönnö. Predicting bitcoin price fluctuation with twitter sentiment analysis. Master’s thesis, School of Computer Science and Communication, 2017. * [18] Sangeeta Oswal, Ravikumar Soni, Omkar Narvekar, and Abhijit Pradha. Named entity recognition and aspect based sentiment analysis. International Journal of Computer Applications, 178(46):18–23, September 2019. * [19] Venkateswarlu Bonta, Nandhini Kumaresh, and N. Janardhan. A comprehensive study on lexicon based approaches for sentiment analysis. Asian Journal of Computer Science and Technology, 8:1–6, 2019. * [20] C. W. Park and D. R. Seo. Sentiment analysis of twitter corpus related to artificial intelligence assistants. In 2018 5th International Conference on Industrial Engineering and Applications (ICIEA), pages 495–498, April 2018. * [21] C.J. Hutto and Eric Gilbert. Vader: A parsimonious rule-based model for sentiment analysis of social media text. In Proceedings of the 8th International Conference on Weblogs and Social Media, ICWSM 2014, January 2015. * [22] Diana Kornbrot. Point biserial correlation. In David Howell Brian S. Everitt, editor, The Encyclopedia of Statistics in Behavioral Science, volume 1, page 2352. Wiley, 1 edition, October 2005. * [23] Snezana Kustrin and Rosemary Beresford. Basic concepts of artificial neural network (ann) modeling and its application in pharmaceutical research. Journal of pharmaceutical and biomedical analysis, 22:717–27, June 2000. * [24] Rajalingappaa Shanmugamani. Deep Learning for Computer Vision. Packt Publishing Ltd., 2018. * [25] Aaron Courville Ian Goodfellow, Yoshua Bengio. Deep Learning. MIT Press, 2016. * [26] Mahidhar Dwarampudi and N. V. Subba Reddy. Effects of padding on lstms and cnns. ArXiv, abs/1903.07288, 2019. * [27] Erion Çano and Maurizio Morisio. Word embeddings for sentiment analysis: A comprehensive empirical survey. ArXiv, 2019. * [28] Jeffrey Pennington, Richard Socher, and Christoper Manning. Glove: Global vectors for word representation. In EMNLP, volume 14, pages 1532–1543, January 2014. * [29] Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9:1735–80, December 1997. * [30] Ralf C. Staudemeyer and Eric Rothstein Morris. Understanding lstm - a tutorial into long short-term memory recurrent neural networks. ArXiv, abs/1909.09586, 2019. * [31] Klaus Greff, Rupesh Srivastava, Jan Koutník, Bas Steunebrink, and Jürgen Schmidhuber. Lstm: A search space odyssey. IEEE transactions on neural networks and learning systems, 28, March 2015. * [32] Serena Yeung Fei-Fei Li, Justin Johnson. Recurrent neural network. http://cs231n.stanford.edu/slides/2017/cs231n_2017_lecture10.pdf. Accessed: 03 March 2020. * [33] Yaoshiang Ho and Samuel Wookey. The real-world-weight cross-entropy loss function: Modeling the costs of mislabeling. IEEE Access, PP:1–1, December 2019. * [34] Followthehashtag. One hundred nasdaq 100 companies – free twitter datasets. http://followthehashtag.com/datasets/nasdaq-100-companies-free-twitter-dataset/. Accessed: 18 October 2019. * [35] Yahoo Finance. Apple inc. (aapl), nasdaqgs real-time price. currency in usd. https://uk.finance.yahoo.com/quote/AAPL/history?p=AAPL. Accessed: 18 October 2019. * [36] Gavin Hackeling. Mastering Machine Learning with scikit-learn. Apply effective learning algorithms to real-world problems using scikit-learn. Packt publishing, August 2014. * [37] G.P. Zhang. Neural Networks in Business Forecasting. Idea Group Publishing, 2004. * [38] Duyu Tang, Furu Wei, Nan Yang, Ming Zhou, Ting Liu, and Bing Qin. Learning sentiment-specific word embedding for twitter sentiment classification. In 52nd Annual Meeting of the Association for Computational Linguistics, ACL 2014 - Proceedings of the Conference, volume 1, pages 1555–1565, June 2014.
# PROGRESSIVE IMAGE SUPER-RESOLUTION VIA NEURAL DIFFERENTIAL EQUATION ###### Abstract We propose a new approach for the image super-resolution (SR) task that progressively restores a high-resolution (HR) image from an input low- resolution (LR) image on the basis of a neural ordinary differential equation. In particular, we newly formulate the SR problem as an initial value problem, where the initial value is the input LR image. Unlike conventional progressive SR methods that perform gradual updates using straightforward iterative mechanisms, our SR process is formulated in a concrete manner based on explicit modeling with a much clearer understanding. Our method can be easily implemented using conventional neural networks for image restoration. Moreover, the proposed method can super-resolve an image with arbitrary scale factors on continuous domain, and achieves superior SR performance over state- of-the-art SR methods. ## 1 Introduction Image super-resolution (SR) is a classic low-level vision task that aims to recover a high-resolution (HR) image from a given low-resolution (LR) input image. For several decades, a large volume of literature documents the high demand of SR technique in various vision applications. However, SR problem still remains a challenge and is difficult to solve because it is a highly ill-posed inverse problem. With the recent development of deep learning technology, numerous deep- learning-based SR methods [1, 2, 3] have been presented, and they have shown plausible results. To further improve the SR performance, many researchers have attempted to restore the high-quality image by recovering the fine details of the LR input image progressively [4, 5]. Many previous works hinged on this progressive SR procedure are based on a variant of feedback network in the human visual system [6], and they show satisfactory SR results. However, owing to lack of theoretical clarity on the progressive system, these approaches need to develop a well-engineered method. For example, the number of iterations for the gradual refinements [7] and complicated learning strategies [5] as well as the network architectures [4, 8] are considered to improve the SR performance. Several researchers have conducted studies on differential equations to solve the image restoration problems [9, 10]. They also have developed progressive approaches, but these approaches are limited to modeling the prior and/or likelihood models. In this study, we introduce a neural ordinary differential equation (NODE [11]). formulation that describes an explicitly defined progressive SR procedure from the LR to HR images via a neural network. In particular, we reconstruct the HR image by numerically solving the initial value problem originated from the proposed ODE formulation, given the LR image as an initial condition. With the aid of the proposed ODE, our method eases implementation using conventional restoration networks and ODE solvers without any exertion to improve the performance. Furthermore, by simply changing the initial condition of our formulation at the test-time, ours can naturally handle a continuous-valued scale factor. Extensive experiments demonstrate the superiority of the proposed method over state-of-the-art SR approaches. ## 2 Proposed Method Fig. 1: (a) Overview of the proposed SR approach (NODE-SR). $\\{t_{i}\\}_{0\leq i\leq m}$ is a strictly decreasing sequence and $t_{m}=1$. Solid orange line represents our SR process that starts with the initial condition $\mathcal{I}(t_{0})$ until we reconstruct the final HR image $\mathcal{I}(1)$. (b) The neural network $f$ takes an input image $\mathcal{I}(t)$ with the scale factor $t$ and outputs the desired high- frequency detail. ### 2.1 Progressive Super-Resolution Formulation Existing SR methods utilizing progressive SR process [12, 4, 5] are based on iterative multi-stage approaches and can be viewed as variants of the following: $I_{n}=g_{n-1}(I_{n-1})\quad(n\leq N),$ (1) where $n$ denotes the iteration step, $I_{0}$ denotes the given initial input LR image, and $I_{n}$ is the iteratively refined image from its previous state $I_{n-1}$. These approaches typically produce multiple intermediate HR images during the refinement, and the rendered image at the last $N$-th iteration [7, 5] or a combined version of the multiple intermediate images ($\\{I_{n}\\}_{1\leq n\leq N}$) [13, 4] becomes the final SR result. Although these previous progressive methods show promising SR results, they still have some limitations. First, these methods need plenty of time and effort in determining the network configurations including the number of progressive updates $N$ and hyper-parameter settings, and designing cost functions to train the SR networks $g$. In addition, well-engineered and dedicated learning strategy, such as curriculum learning [5] and recursive supervision [13], is required for each method. This complication comes from the lack of clear understanding on their intermediate image states $\\{I_{n}\\}$. To alleviate these problems, we formulate the progressive SR process with a differential equation. This allows us to implement and train the SR networks in an established way while outperforming the performance of conventional progressive SR process. Assume that $(I_{HR})\downarrow_{t}$ is a downscaled version of a ground-truth clean image $I_{HR}$ using a traditional SR kernel (e.g., bicubic) with a scaling factor $\frac{1}{t}$. We then define $\mathcal{I}(t)$ by upscaling $(I_{HR})\downarrow_{t}$ using that SR kernel with a scaling factor ${t}$ so that $I_{HR}$ and $\mathcal{I}(t)$ have the same spatial resolution (see the illustration of Generating LR image in Figure 1(a)). Note that $t\geq 1$, and $\mathcal{I}(1)$ denotes the ground-truth clean image $I_{HR}$. To model a progressive SR process,we first estimate the high-frequency image residual with a neural network. Specifically, when $t$ is a conventional discrete- scaling factor (e.g., x2, x3, and x4), image residual between $\mathcal{I}(t)$ and $\mathcal{I}(t-1)$ can be modeled using a neural network $f_{\text{discrete}}$ as: $\mathcal{I}(t-1)-\mathcal{I}(t)=f_{\text{discrete}}(\mathcal{I}(t),t).$ (2) Notably, $\mathcal{I}(t-1)$ includes more high-frequency details than $\mathcal{I}(t)$ without loss of generality. In our method, we model the slightest image difference to formulate a continuously progressive SR process. Therefore, we take the scale factor $t$ to continuous domain, and reformulate (2) as an ODE with a neural network $f$ as: $\frac{d\mathcal{I}(t)}{dt}=f(\mathcal{I}(t),t,\theta),$ (3) where $\theta$ denotes the trainable parameter of the network $f$. Using this formulation, we can predict the high-frequency image detail required to slightly enhance $\mathcal{I}(t)$ with the network $f$. (Note that we can obtain $\mathcal{I}(t)$ with any rational number $t$ by adding padding to the border of image before resizing and then center cropping the image.) As existing SR neural networks have been proven to be successful at predicting the high-frequency residual image [2], we can use conventional SR architectures as our network $f$ in (3) without major changes. ### 2.2 Single Image Super-Resolution with Neural Ordinary Differential Equation In this section, we explain how to super-resolve a given LR image with a continuous scaling factor using our ODE-based SR formulation in (3). First, we obtain $\mathcal{I}(t_{0})$ by upscaling the given LR input image (Test time LR image in Figure 1(a)) using the bicubic SR kernel to a desired output resolution with a scaling factor $t_{0}$ . Next, we solve the ODE initial value problem in (3) with the initial condition $\mathcal{I}(t_{0})$ by integrating the neural network $f$ from $t_{0}$ to $1$ to acquire the high- quality image $\mathcal{I}(1)$ as follows: $\mathcal{I}(1)=\mathcal{I}(t_{0})+\int_{t_{0}}^{1}f(\mathcal{I}(t),t,\theta)dt.$ (4) Specifically, we approximate the high-quality image $\mathcal{I}(1)$ given a fully trained neural network $f$, network parameter $\theta$, initial condition $\mathcal{I}(t_{0})$, and integral interval $[t_{0},1]$ using an ODE solver ($ODESolve()$) as: $\mathcal{I}(1)\approx ODESolve(\mathcal{I}(t_{0}),f,\theta,[t_{0},1]).$ (5) Thus, our method does not need to consider the stop condition (i.e., the number of feedback iterations) of the progressive SR process unlike conventional approaches [7, 5]. Notably, during the training phase, we need to employ an ODE solver which allows end-to-end training using backpropagation with other components such as the neural network $f$. Unlike other progressive SR methods [13, 5], we do not require any other learning strategies like curriculum learning during the training phase. Fig. 2: Visual comparisons with conventional progressive SR methods (DRRN, SRFBN). For different scale factors (x2, and x4) intermediate HR images are visualized, and #it indicates the number of updates used to render results by DRRN and SRFBN. $\hat{I}()$ denotes the predicted results by our NODE-RDN. Fig. 3: Visual comparison of NODE-RDN (ours) with Meta-RDN on scale x2.5 and x4. In addition, our formulation is made upon a continuous context, allows a continuous scale factor $t_{0}$ where $t_{0}\geq 1$. This makes our method able to handle the arbitrary-scale SR problem. To train the deep neural network $f$, and learn the parameter $\theta$ in (5), we minimize the loss summed over scale factors $t$ using the L1 loss function as: $\mathcal{L}(\theta)=\sum_{t}\|I_{HR}-ODESolve(\mathcal{I}(t),f,\theta,[t,1])\|_{1}.$ (6) By minimizing the proposed loss function, our network parameter $\theta$ is trained to estimate the image detail to be added into the network input as in (3). ## 3 Experimental Results In this section, we carry out extensive experiments to demonstrate the superiority of the proposed method, and add various quantitative and qualitative comparison results. We also provide detailed analysis of our experimental results. ### 3.1 Implementation details We use VDSR [2] and RDN [3] as backbone CNN architectures for our network $f$ with slight modifications. For each CNN architecture, we change the first layer to feed the scale factor $t$ as an additional input. To be specific, we extend the input channel from 3 to 4, and the pixel locations of the newly concatenated channel (4-th channel) are filled with a scalar value $t$ as shown in Figure 1(b). In addition, for RDN, we remove the last upsampling layer so that input and output resolutions are the same in our work. Note that, no extra parameters are added except for the first layers of the networks. To train and infer the proposed SR process, we use Runge–Kutta (RK4) method as our ODE solver in (6). For simplicity, our approaches with VDSR and RDN backbones are called NODE-VDSR and NODE-RDN in the remaining parts of the experiments, respectively. We use the DIV2K [14] dataset to train our NODE- VDSR and NODE-RDN. We train the network by minimizing the L1 loss in (6) with the Adam optimizer ($\beta_{1}=0.9$, $\beta_{2}=0.999$, $\epsilon=10^{-8}$) [15]. The initial learning rate is set as $10^{-4}$, which is then decreased by half every 100k gradient update steps, and trained for 600k iterations in total. The mini-batch size of NODE-VDSR is 16 (200x200 patches), but our NODE- RDN takes 8 patches as a mini-batch (130x130 patches) owing to the memory limit of our graphic units. Similar to the training settings in Meta-SR [16], we train the network $f$ by randomly changing the scale factor $t$ in (6) from 1 to 4 with a stride of 0.1 (i.e., $t\in\\{1.1,1.2,1.3,...,4\\}$). Dataset | Scale | Bicubic | DRCN | LapSRN | DRRN | PRLSR | SRFBN | NODE-RDN (ours) | NODE-RDN+ (ours) ---|---|---|---|---|---|---|---|---|--- Set14 | x2 | 30.24/0.8688 | 33.04/0.9118 | 33.08/0.913 | 33.23/0.9136 | 33.69/0.9191 | 33.82/0.9196 | 33.90/0.9209 | 33.95/0.9214 x3 | 27.55/0.7742 | 29.76/0.8311 | 29.87/0.833 | 29.96/0.8349 | 30.43/0.8436 | 30.51/0.8461 | 30.53/0.8465 | 30.59/0.8473 x4 | 26.00/0.7027 | 28.02/0.7670 | 28.19/0.772 | 28.21/0.7721 | 28.71/0.7838 | 28.81/0.7868 | 28.76/0.7866 | 28.83/0.7877 B100 | x2 | 29.56/0.8431 | 31.85/0.8942 | 31.80/0.895 | 32.05/0.8973 | 32.25/0.9005 | 32.29/0.9010 | 32.34/0.9025 | 32.38/0.9028 x3 | 27.21/0.7385 | 28.80/0.7963 | 28.81/0.797 | 28.95/0.8004 | 29.14/0.8060 | 29.24/0.8084 | 29.25/0.8094 | 29.28/0.8100 x4 | 25.96/0.6675 | 27.23/0.7233 | 27.32/0.728 | 27.38/0.7284 | 27.64/0.7378 | 27.72/0.7409 | 27.72/0.7410 | 27.75/0.7417 Urban100 | x2 | 26.88/0.8403 | 30.75/0.9133 | 30.41/0.910 | 31.23/0.9188 | 32.35/0.9308 | 32.62/0.9328 | 32.81/0.9345 | 32.97/0.9355 x3 | 24.46/0.7349 | 27.15/0.8276 | 27.06/0.827 | 27.53/0.8378 | 28.27/0.8541 | 28.73/0.8641 | 28.81/0.8644 | 28.94/0.8662 x4 | 23.14/0.6577 | 25.14/0.7510 | 25.21/0.756 | 25.44/0.7638 | 26.22/0.7892 | 26.60/0.8015 | 26.56/0.7985 | 26.68/0.8010 Table 1: Comparison with progressive SR methods on the benchmark datsets (Set14 [17], B100 [18], and Urban100 [19]). We provide average PSNR/SSIM values for scaling factors x2, x3, and x4. Our NODE-RDN and NODE-RDN+ show the best performance. Red and blue colors denote the best and second best results, respectively. Methods Scale | x1.1 | x1.2 | x1.3 | x1.4 | x1.5 | x1.6 | x1.7 | x1.8 | x1.9 | x2.0 | x2.1 | x2.2 | x2.3 | x2.4 | x2.5 ---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|--- bicubic | 36.56 | 35.01 | 33.84 | 32.93 | 32.14 | 31.49 | 30.90 | 30.38 | 29.97 | 29.55 | 29.18 | 28.87 | 28.57 | 28.31 | 28.13 VDSR | - | - | - | - | - | - | - | - | - | 31.90 | - | - | - | - | - VDSR+t | 39.51 | 38.44 | 37.15 | 36.04 | 34.98 | 34.15 | 33.39 | 32.78 | 32.22 | 31.70 | 31.27 | 30.86 | 30.53 | 30.2 | 29.91 NODE-VDSR (ours) | 41.46 | 39.36 | 37.75 | 36.51 | 35.38 | 34.49 | 33.70 | 33.07 | 32.50 | 31.95 | 31.52 | 31.09 | 30.76 | 30.42 | 30.12 RDN | - | - | - | - | - | - | - | - | - | 32.34 | - | - | - | - | - RDN+t | 42.83 | 39.92 | 38.18 | 36.87 | 35.71 | 34.80 | 33.99 | 33.34 | 32.77 | 32.22 | 31.76 | 31.33 | 30.99 | 30.64 | 30.34 Meta-RDN | 42.82 | 40.04 | 38.28 | 36.95 | 35.86 | 34.90 | 34.13 | 33.45 | 32.86 | 32.35 | 31.82 | 31.41 | 31.06 | 30.62 | 30.45 NODE-RDN (ours) | 43.22 | 40.06 | 38.35 | 37.02 | 35.86 | 34.95 | 34.14 | 33.47 | 32.89 | 32.34 | 31.89 | 31.46 | 31.12 | 30.76 | 30.46 NODE-RDN+ (ours) | 43.33 | 40.13 | 38.40 | 37.07 | 35.90 | 34.99 | 34.17 | 33.50 | 32.93 | 32.38 | 31.93 | 31.50 | 31.16 | 30.80 | 30.50 Methods Scale | x2.6 | x2.7 | x2.8 | x2.9 | x3.0 | x3.1 | x3.2 | x3.3 | x3.4 | x3.5 | x3.6 | x3.7 | x3.8 | x3.9 | x4.0 ---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|--- bicubic | 27.89 | 27.66 | 27.51 | 27.31 | 27.19 | 26.98 | 26.89 | 26.59 | 26.60 | 26.42 | 26.35 | 26.15 | 26.07 | 26.01 | 25.96 VDSR | - | - | - | - | 28.83 | - | - | - | - | - | - | - | - | - | 27.29 VDSR+t | 29.64 | 29.39 | 29.15 | 28.93 | 28.74 | 28.55 | 28.38 | 28.22 | 28.05 | 27.89 | 27.76 | 27.58 | 27.47 | 27.34 | 27.20 NODE-VDSR (ours) | 29.85 | 29.61 | 29.36 | 29.14 | 28.94 | 28.75 | 28.58 | 28.41 | 28.25 | 28.08 | 27.96 | 27.79 | 27.66 | 27.54 | 27.40 RDN | - | - | - | - | 29.26 | - | - | - | - | - | - | - | - | - | 27.72 RDN+t | 30.06 | 29.80 | 29.55 | 29.33 | 29.12 | 28.92 | 28.76 | 28.59 | 28.43 | 28.26 | 28.13 | 27.95 | 27.84 | 27.71 | 27.58 Meta-RDN | 30.13 | 29.82 | 29.67 | 29.40 | 29.30 | 28.87 | 28.79 | 28.68 | 28.54 | 28.32 | 28.27 | 28.04 | 27.92 | 27.82 | 27.75 NODE-RDN (ours) | 30.18 | 29.93 | 29.67 | 29.45 | 29.25 | 29.05 | 28.88 | 28.71 | 28.54 | 28.37 | 28.24 | 28.07 | 27.96 | 27.81 | 27.72 NODE-RDN+ (ours) | 30.22 | 29.97 | 29.71 | 29.49 | 29.28 | 29.05 | 28.92 | 28.74 | 28.58 | 28.41 | 28.28 | 28.12 | 28.00 | 27.87 | 27.75 Table 2: Average PSNR values on the B100 [18] evaluated with different scale factors. The best performance is shown in bold number. ### 3.2 Comparison with Progressive SR Methods First, we compare our NODE-RDN with several state-of-the-art progressive SR methods: DRCN [13], LapSRN [12], DRRN [7], PRLSR [8], and SRFBN [5]. As in [20], self-ensemble method is used to further improve NODE-RDN (denoted as NODE-RDN+). Note that, our NODE-RDN and NODE-RDN+ can handle multiple scale factors $t$ including non-integer scale factors (e.g., x1.5) using the same network parameter. In contrast, other approaches are required to be trained for certain discrete integer scale factors (x2, x3, and x4) separately, resulting in a distinct parameter set for each scale factor. Nevertheless, quantitative restoration results in Table 1 show that our NODE-RDN, NODE-RDN+ consistently outperforms conventional progressive SR methods for the discrete integer scaling factors (x2, x3, and x4) in terms of PSNR. In Figure 2, we investigate intermediate images produced during the progressive SR process with the scale factors x2 and x4. Final results by DRRN are obtained after 25 iterations, and the final results by SRFBN are obtained with 4 iterations as in their original settings. We provide 4 intermediate HR images during the updates for visual comparisons. For our NODE-RDN, intermediate image states are represented as $\hat{\mathcal{I}}(t_{i})$ where ${1\leq t_{i}\leq t_{0}}$ and $\hat{\mathcal{I}}(t_{i})=ODESolve(\mathcal{I}(t_{0}),f,\theta,[t_{0},t_{i}])$. We observe that DRRN and SRFBN fail to progressively refine patches with high- frequency details, while our NODE-RDN can gradually improve the intermediate images and render promising results at the final states. ### 3.3 Comparison with Multi-scale SR Methods Our approach can handle a continuous scale factor for the SR task, thus we compare ours with existing multi-scale SR methods that can handle continuous scale factors: VDSR [2] and Meta-SR [16]. Notably, Meta-SR implemented using RDN (i.e., Meta-RDN) is the current state-of-the-art SR approach. In Table 2, we show quantitative results compared to existing SR methods (VDSR, RDN, and Meta-RDN). Note that, VDSR+t and RDN+t are modified versions of VDSR and RDN to take the scale factor $t$ as an additional input of the networks and have the same input and output resolutions as in our network $f$. We also compare our method with these new baselines (VDSR+t and RDN+t) for fair comparisons. We evaluate the SR performance on the B100 benchmark dataset by increasing the scaling factor from 1.1 to 4. Interestingly, we observe that NODE-VDSR outperforms VDSR and VDSR+t at every scale by a large margin although VDSR and VDSR+t have similar network architecture to our NODE-VDSR. Similarly, NODE-RDN shows better performance than Meta-RDN and RDN+t. We also provide qualitative comparison results with Meta-SR in Figure 3, and we see that our NODE-RDN recovers much clearer edges than Meta-RDN. ## 4 Conclusion In this work, we proposed a novel differential equation for the SR task to progressively enhance a given input LR image, and allow continuous-valued scale factor. Image difference between images over different scale factors is physically modeled with a neural network, and formulated as a NODE. To restore a high-quality image, we solve the ODE initial value problem with the initial condition given as an input LR image. The main difference with existing progressive SR methods is that our formulation is based on the physical modeling of the intermediate images, and adds fine high-frequency details gradually. Detailed experimental results show that our method achieves superior performance compared to state-of-the-art SR approaches. ## References * [1] Chao Dong, Chen Change Loy, Kaiming He, and Xiaoou Tang, “Image super-resolution using deep convolutional networks,” IEEE transactions on pattern analysis and machine intelligence, vol. 38, no. 2, 2015. * [2] Jiwon Kim, Jung Kwon Lee, and Kyoun Mu Lee, “Accurate image super-resolution using very deep convolutional networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016. * [3] Yulun Zhang, Yapeng Tian, Yu Kong, Bineng Zhong, and Yun Fu, “Residual dense network for image super-resolution,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018. * [4] Muhammad Haris, Gregory Shakhnarovich, and Norimichi Ukita, “Deep back-projection networks for super-resolution,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018. * [5] Zhen Li, Jinglei Yang, Zheng Liu, Xiaomin Yang, Gwanggil Jeon, and Wei Wu, “Feedback network for image super-resolution,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019. * [6] Amir R Zamir, Te-Lin Wu, Lin Sun, William B Shen, Bertram E Shi, Jitendra Malik, and Silvio Savarese, “Feedback networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 2017. * [7] Ying Tai, Jian Yang, and Xiaoming Liu, “Image super-resolution via deep recursive residual network,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017. * [8] Hong Liu, Zhisheng Lu, Wei Shi, and Juanhui Tu, “A fast and accurate super-resolution network using progressive residual learning,” in ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2020, pp. 1818–1822. * [9] Yunjin Chen and Thomas Pock, “Trainable nonlinear reaction diffusion: A flexible framework for fast and effective image restoration,” IEEE transactions on pattern analysis and machine intelligence, 2016\. * [10] Yuanxu Chen, Yupin Luo, and Dongcheng Hu, “A general approach to blind image super-resolution using a pde framework,” in Visual Communications and Image Processing 2005. International Society for Optics and Photonics, 2005. * [11] Ricky TQ Chen, Yulia Rubanova, Jesse Bettencourt, and David K Duvenaud, “Neural ordinary differential equations,” in Advances in neural information processing systems, 2018. * [12] Wei-Sheng Lai, Jia-Bin Huang, Narendra Ahuja, and Ming-Hsuan Yang, “Deep laplacian pyramid networks for fast and accurate super-resolution,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017. * [13] Jiwon Kim, Jung Kwon Lee, and Kyoung Mu Lee, “Deeply-recursive convolutional network for image super-resolution,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016. * [14] Eirikur Agustsson and Radu Timofte, “Ntire 2017 challenge on single image super-resolution: Dataset and study,” 2017\. * [15] Diederik P. Kingma and Jimmy Ba, “Adam: A method for stochastic optimization,” in International Conference on Learning Representations, 2015. * [16] Xuecai Hu, Haoyuan Mu, Xiangyu Zhang, Zilei Wang, Tieniu Tan, and Jian Sun, “Meta-sr: A magnification-arbitrary network for super-resolution,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019. * [17] Roman Zeyde, Michael Elad, and Matan Protter, “On single image scale-up using sparse-representations,” in International conference on curves and surfaces. Springer, 2010\. * [18] David Martin, Charless Fowlkes, Doron Tal, and Jitendra Malik, “A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics,” in Proceedings of the IEEE International Conference on Computer Vision. IEEE, 2001. * [19] Jia-Bin Huang, Abhishek Singh, and Narendra Ahuja, “Single image super-resolution from transformed self-exemplars,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015. * [20] Bee Lim, Sanghyun Son, Heewon Kim, Seungjun Nah, and Kyoung Mu Lee, “Enhanced deep residual networks for single image super-resolution,” in Proceedings of the IEEE conference on computer vision and pattern recognition workshops, 2017.
# Photoproduction $\gamma p\to K^{+}\Lambda(1520)$ in an effective Lagrangian approach Neng-Chang Wei School of Nuclear Science and Technology, University of Chinese Academy of Sciences, Beijing 101408, China Yu Zhang School of Nuclear Science and Technology, University of Chinese Academy of Sciences, Beijing 101408, China Fei Huang<EMAIL_ADDRESS>School of Nuclear Science and Technology, University of Chinese Academy of Sciences, Beijing 101408, China De-Min Li<EMAIL_ADDRESS>School of Physics and Microelectronics, Zhengzhou University, Zhengzhou, Henan 450001, China ###### Abstract The data on differential cross sections and photon-beam asymmetries for the $\gamma p\to K^{+}\Lambda(1520)$ reaction have been analyzed within a tree- level effective Lagrangian approach. In addition to the $t$-channel $K$ and $K^{\ast}$ exchanges, the $u$-channel $\Lambda$ exchange, the $s$-channel nucleon exchange, and the interaction current, a minimal number of nucleon resonances in the $s$ channel are introduced in constructing the reaction amplitudes to describe the data. The results show that the experimental data can be well reproduced by including either the $N(2060)5/2^{-}$ or the $N(2120)3/2^{-}$ resonance. In both cases, the contact term and the $K$ exchange are found to make significant contributions, while the contributions from the $K^{\ast}$ and $\Lambda$ exchanges are negligible in the former case and considerable in the latter case. Measurements of the data on target asymmetries are called on to further pin down the resonance contents and to clarify the roles of the $K^{\ast}$ and $\Lambda$ exchanges in this reaction. $K^{+}\Lambda(1520)$ photoproduction, effective Lagrangian approach, photon- beam asymmetries ###### pacs: 25.20.Lj, 13.60.Le, 14.20.Gk, 13.75.Jz ## I Introduction The traditional $\pi N$ elastic and inelastic scattering experiments have provided us with abundant knowledge of the mass spectrum and decay properties of the nucleon resonances ($N^{\ast}$’s). Nevertheless, both the quark model Isgur:1977ef ; Koniuk:1979vy and lattice QCD Edwards:2011jj ; Edwards:2012fx calculations predict more resonances than have been observed in the $\pi N$ scattering experiments. The resonances predicated by quark model or lattice QCD but not observed in experiments are called “missing resonances”, which are supposed to have small couplings to the $\pi N$ channel and, thus, escape from experimental detection. In the past few decades, intense efforts have been dedicated to search for the missing resonances in meson production reaction channels other than $\pi N$. In particular, the $\rho N$, $\phi N$, and $\omega N$ production reactions in the nonstrangeness sector and the $KY$, $K^{\ast}Y$ ($Y=\Lambda,\Sigma$) production reactions in the strangeness sector have been widely investigated both experimentally and theoretically. In the present paper, we focus on the $\gamma p\to K^{+}\Lambda(1520)$ reaction process. The threshold of the $K^{+}\Lambda(1520)$ photoproduction is about $2.01$ GeV, and, thus, this reaction provides a chance to study the $N^{\ast}$ resonances in the $W\sim 2.0$ GeV mass region in which we have infancy information as shown in the latest version of the Review of Particle Physics (RPP) Tanabashi:2018oca . Besides, the isoscalar nature of $\Lambda(1520)$ allows only the $I=1/2$ $N^{\ast}$ resonances exchanges in the $s$ channel, which simplifies the reaction mechanisms of the $K^{+}\Lambda(1520)$ photoproduction. Experimentally, the cross sections for the reaction $\gamma p\to K^{+}\Lambda(1520)$ have been measured at SLAC by Boyarski et al. in 1971 for photon energy $E_{\gamma}=11$ GeV Boyarski:1970yc , and by the LAMP2 group in 1980 at $E_{\gamma}=2.8$$-$$4.8$ GeV Barber:1980zv . In 2010, the LEPS Collaboration measured the differential cross sections and photon-beam asymmetries ($\Sigma$) at Spring-8 for $\gamma p\to K^{+}\Lambda(1520)$ at energies from threshold up to $E_{\gamma}=2.6$ GeV at forward $K^{+}$ angles Kohri:2009xe . In 2011, the SAPHIR Collaboration measured the cross sections at the Electron Stretcher Accelerator (ELSA) for the $K^{+}\Lambda(1520)$ photoproduction in the energy range from threshold up to $E_{\gamma}=2.65$ GeV Wieland:2010cq . Recently, the differential and total cross sections for the $K^{+}\Lambda(1520)$ photoproduction were reported by the CLAS Collaboration at energies from threshold up to the center-of-mass energy $W=2.86$ GeV over a large range of the $K^{+}$ production angle Moriya:2013hwg . Theoretically, the $K^{+}\Lambda(1520)$ photoproduction reaction has been extensively investigated based on effective Lagrangian approaches by four theory groups in $11$ publications Nam:2005uq ; Nam:2006cx ; Nam:2009cv ; Nam:2010au ; Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb ; He:2012ud ; He:2014gga ; Yu:2017kng . In Refs. Nam:2005uq ; Nam:2006cx ; Nam:2009cv ; Nam:2010au , Nam et al. found that the contact term and the $t$-channel $K$ exchange are important to the cross sections of $\gamma p\to K^{+}\Lambda(1520)$, while the contributions from the $t$-channel $K^{\ast}$ exchange and the $s$-channel nucleon resonance exchange are rather small. In Refs. Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb , Xie, Wang, and Nieves et al. found that apart from the contact term and the $t$-channel $K$ exchange, the $u$-channel $\Lambda$ exchange and the $s$-channel $N(2120)3/2^{-}$ [previously called $D_{13}(2080)$] exchange are also important in describing the cross-section data for $\gamma p\to K^{+}\Lambda(1520)$, while the contribution from the $t$-channel $K^{\ast}$ exchange is negligible in this reaction. In Refs. He:2012ud ; He:2014gga , He and Chen found that the contribution from the $t$-channel $K^{\ast}$ exchange in $\gamma p\to K^{+}\Lambda(1520)$ is also considerable besides the important contributions from the contact term, the $t$-channel $K$ exchange, the $u$-channel $\Lambda$ exchange, and the $s$-channel $N(2120)3/2^{-}$ exchange. In Ref. Yu:2017kng , Yu and Kong studied the $\gamma p\to K^{+}\Lambda(1520)$ reaction within a Reggeized model, and they claimed that the important contributions to this reaction are coming from the contact term, the $t$-channel $K$ exchange, and the $t$-channel $K^{\ast}_{2}$ exchange, while the contribution from the $t$-channel $K^{\ast}$ exchange is minor. One observes that the common feature reported in all the above-mentioned publications of Refs. Nam:2005uq ; Nam:2006cx ; Nam:2009cv ; Nam:2010au ; Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb ; He:2012ud ; He:2014gga ; Yu:2017kng is that the contributions from the contact term and the $t$-channel $K$ exchange are important to the $\gamma p\to K^{+}\Lambda(1520)$ reaction. Even so, the reaction mechanisms of $\gamma p\to K^{+}\Lambda(1520)$ claimed by those four theory groups are quite different. In particular, there are no conclusive answers which can be derived from Refs. Nam:2005uq ; Nam:2006cx ; Nam:2009cv ; Nam:2010au ; Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb ; He:2012ud ; He:2014gga ; Yu:2017kng for the following questions: Are the contributions from the $t$-channel $K^{\ast}$ exchange and $u$-channel $\Lambda$ exchange significant or not in this reaction, does one inevitably need to introduce nucleon resonances in the $s$ channel to describe the data, and if yes, is the $N(2120)3/2^{-}$ resonance the only candidate needed in this reaction and what are the parameters of it? Figure 1: Predictions of photon-beam asymmetries at $\cos\theta=0.8$ as a function of the photon laboratory energy for $\gamma p\to K^{+}\Lambda(1520)$ from Ref. Xie:2010yk (blue dashed line), the fit II of Ref. Xie:2013mua (red solid line), Ref. He:2012ud (green dotted line), and Ref. He:2014gga (black dot-dashed line). The data are located in $0.6<\cos\theta<1$ and taken from the LEPS Collaboration Kohri:2009xe (blue square). On the other hand, the data on photon-beam asymmetries for $\gamma p\to K^{+}\Lambda(1520)$ reported by the LEPS Collaboration in 2010 Kohri:2009xe have never been well reproduced in previous publications of Refs. Nam:2005uq ; Nam:2006cx ; Nam:2009cv ; Nam:2010au ; Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb ; He:2012ud ; He:2014gga . As an illustration, we show in Fig. 1 the theoretical results on photon-beam asymmetries from Refs. Xie:2010yk ; He:2012ud ; Xie:2013mua ; He:2014gga calculated at $\cos\theta=0.8$ and compared with the data located at $0.6<\cos\theta<1$. It is true that the data bins in scattering angles are wide; nevertheless, it has been checked that the averaged values of theoretical beam-asymmetry results in $0.6<\cos\theta<1$ are comparable with those calculated at $\cos\theta=0.8$. One sees that, in the energy region $E_{\gamma}>2$ GeV, even the signs of the photon-beam asymmetries predicated by these theoretical works are opposite to the data. In the Regge model analysis of Ref. Yu:2017kng , the photon-beam asymmetries have indeed been analyzed, but there the differential cross-section data have been only qualitatively described, and the structures of the angular distributions exhibited by the data were missing due to the lack of nucleon resonances in the $s$-channel interactions. The purpose of the present work is to perform a combined analysis of the available data on both the differential cross sections and the photon-beam asymmetries for $\gamma p\to K^{+}\Lambda(1520)$ within an effective Lagrangian approach, and, based on that, we try to get a clear understanding of the reaction mechanism of $\gamma p\to K^{+}\Lambda(1520)$. In particular, we aim to clarify whether the $t$-channel $K^{\ast}$ exchange and the $u$-channel $\Lambda$ exchange are important or not and what the resonance contents and their associated parameters are in this reaction. As discussed above, previous publications of Refs. Nam:2005uq ; Nam:2006cx ; Nam:2009cv ; Nam:2010au ; Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb ; He:2012ud ; He:2014gga can describe only the differential cross-section data, and they gave diverse answers to these questions. It is expected that more reliable results on the resonance contents and the roles of $K^{\ast}$ and $\Lambda$ exchanges in this reaction can be obtained from the theoretical analysis which can result in a satisfactory description of the data on both the differential cross sections and the photon-beam asymmetries. The present paper is organized as follows. In Sec. II, we briefly introduce the framework of our theoretical model, including the effective interaction Lagrangians, the resonance propagators, and the phenomenological form factors employed in this work. The results of our model calculations are shown and discussed in Sec. III. Finally, a brief summary and conclusions are given in Sec. IV. ## II Formalism (a) $s$ channel (b) $t$ channel (c) $u$ channel (d) Interaction current Figure 2: Generic structure of the amplitude for $\gamma p\to K^{+}\Lambda(1520)$. Time proceeds from left to right. The outgoing $\Lambda^{\ast}$ denotes $\Lambda(1520)$. The full amputated photoproduction amplitude for $\gamma N\to K\Lambda(1520)$ in our tree-level effective Lagrangian approach can be expressed as Wang:2017tpe ; Wang:2018vlv ; Wang:2020mdn ; Wei:2019imo $M^{\nu\mu}\equiv M^{\nu\mu}_{s}+M^{\nu\mu}_{t}+M^{\nu\mu}_{u}+M^{\nu\mu}_{\rm int},$ (1) with $\nu$ and $\mu$ being the Lorentz indices for outgoing $\Lambda(1520)$ and incoming photon, respectively. The first three terms $M^{\nu\mu}_{s}$, $M^{\nu\mu}_{t}$, and $M^{\nu\mu}_{u}$ stand for the amplitudes resulted from the $s$-channel $N$ and $N^{\ast}$ exchanges, the $t$-channel $K$ and $K^{\ast}$ exchanges, and the $u$-channel $\Lambda$ exchange, respectively, as diagrammatically depicted in Fig. 2. They can be calculated straightforward by using the effective Lagrangians, propagators, and form factors provided in the following part of this section. The last term in Eq. (1) represents the interaction current arising from the photon attaching to the internal structure of the $\Lambda(1520)NK$ vertex. In practical calculation, the interaction current $M^{\nu\mu}_{\rm int}$ is modeled by a generalized contact current Haberzettl:1997 ; Haberzettl:2006 ; Haberzettl:2011zr ; Huang:2012 ; Huang:2013 ; Wang:2017tpe ; Wang:2018vlv ; Wei:2019imo ; Wang:2020mdn ; Wei:2020fmh : $M^{\nu\mu}_{\rm int}=\Gamma_{\Lambda^{\ast}NK}^{\nu}(q)C^{\mu}+M^{\nu\mu}_{\rm KR}f_{t}.$ (2) Here $\Gamma_{\Lambda^{\ast}NK}^{\nu}(q)$ is the vertex function of $\Lambda(1520)NK$ coupling governed by the Lagrangian of Eq. (16): $\Gamma_{\Lambda^{\ast}NK}^{\nu}(q)=-\frac{g_{\Lambda^{\ast}NK}}{M_{K}}\gamma_{5}q^{\nu},$ (3) with $q$ being the four-momentum of the outgoing $K$ meson; $M^{\nu\mu}_{\rm KR}$ is the Kroll-Ruderman term governed by the Lagrangian of Eq. (15): $M^{\nu\mu}_{\rm KR}=\frac{g_{\Lambda^{\ast}NK}}{M_{K}}g^{\nu\mu}\gamma_{5}Q_{K}\tau,$ (4) with $Q_{K}$ being the electric charge of outgoing $K$ meson and $\tau$ being the isospin factor of the Kroll-Ruderman term; $f_{t}$ is the phenomenological form factor attached to the amplitude of $t$-channel $K$ exchange, which is given by Eq. (39); $C^{\mu}$ is an auxiliary current introduced to ensure the gauge invariance of the full photoproduction amplitude of Eq. (1). Note that the photoproduction amplitudes will automatically be gauge invariant in the cases that there are no form factors and the electromagnetic couplings are obtained by replacing the partial derivative by its covariant form in the corresponding hadronic vertices. In practical calculation, one has to introduce the form factors in hadronic vertices (cf. Sec. II.3) which violate the gauge invariance. The auxiliary current $C^{\mu}$ is then introduced to compensate the gauge violation caused by the form factors. Following Refs. Haberzettl:2006 ; Haberzettl:2011zr ; Huang:2012 , for the $\gamma N\to K\Lambda(1520)$ reaction, the auxiliary current $C^{\mu}$ is chosen to be $C^{\mu}=-Q_{K}\tau\frac{f_{t}-\hat{F}}{t-q^{2}}(2q-k)^{\mu}-\tau Q_{N}\frac{f_{s}-\hat{F}}{s-p^{2}}(2p+k)^{\mu},$ (5) with $\hat{F}=1-\hat{h}\left(1-f_{s}\right)\left(1-f_{t}\right).$ (6) Here $p$, $q$, and $k$ denote the four-momenta for incoming $N$, outgoing $K$, and incoming photon, respectively; $Q_{K}$ and $Q_{N}$ are electric charges of $K$ and $N$, respectively; $f_{s}$ and $f_{t}$ are phenomenological form factors for $s$-channel $N$ exchange and $t$-channel $K$ exchange, respectively; $\hat{h}$ is an arbitrary function going to unity in the high- energy limit and set to be $1$ in the present work for simplicity; $\tau$ depicts the isospin factor for the corresponding hadronic vertex. Alternatively, one can rewrite the auxiliary current $C^{\mu}$ in Eq. (5) as $\displaystyle C^{\mu}=$ $\displaystyle- Q_{K}\tau(2q-k)^{\mu}\frac{f_{t}-1}{t-q^{2}}\left[1-\hat{h}\left(1-f_{s}\right)\right]$ $\displaystyle-\tau Q_{N}(2p+k)^{\mu}\frac{f_{s}-1}{s-p^{2}}\left[1-\hat{h}\left(1-f_{t}\right)\right].$ (7) One sees clearly that if there are no form factors, i.e., $f_{t}=f_{s}=1$, one has $C^{\mu}\to 0$ and, consequently, $M^{\nu\mu}_{\rm int}\to M^{\nu\mu}_{KR}$. We mention that the auxiliary current $C^{\mu}$ in Eq. (5) works for both real and virtual photons; i.e., the amplitudes we constructed in Eq. (1) are gauge invariant for both photo- and electroproduction of $K^{+}\Lambda(1520)$. In Ref. Nam2013 , another prescription for keeping gauge invariance of the $K^{+}\Lambda(1520)$ electroproduction amplitudes was introduced, where additional terms are considered besides those for photoproduction reactions. In the rest of this section, we present the effective Lagrangians, the resonance propagators, the form factors, and the interpolated $t$-channel Regge amplitudes employed in the present work. ### II.1 Effective Lagrangians In this subsection, we list all the Lagrangians used in the present work. For further simplicity, we define the operators $\Gamma^{(+)}=\gamma_{5}\qquad{\rm and}\qquad\Gamma^{(-)}=1,$ (8) the field $\Lambda^{\ast}=\Lambda(1520),$ (9) and the field-strength tensor $F^{\mu\nu}=\partial^{\mu}A^{\nu}-\partial^{\nu}A^{\mu},$ (10) with $A^{\mu}$ denoting the electromagnetic field. The Lagrangians needed to calculate the amplitudes for nonresonant interacting diagrams are $\displaystyle{\cal L}_{\gamma NN}$ $\displaystyle=$ $\displaystyle-\,e\bar{N}\\!\left[\\!\left(\hat{e}\gamma^{\mu}-\frac{\hat{\kappa}_{N}}{2M_{N}}\sigma^{\mu\nu}\partial_{\nu}\\!\right)\\!A_{\mu}\right]\\!N,$ (11) $\displaystyle{\cal L}_{\gamma KK}$ $\displaystyle=$ $\displaystyle ie\\!\left[K^{+}\left(\partial_{\mu}K^{-}\right)-K^{-}\left(\partial_{\mu}K^{+}\right)\right]\\!A^{\mu},$ (12) $\displaystyle{\cal L}_{\gamma K{K^{\ast}}}$ $\displaystyle=$ $\displaystyle e\frac{g_{\gamma K{K^{\ast}}}}{M_{K}}\varepsilon^{\alpha\mu\lambda\nu}\left(\partial_{\alpha}A_{\mu}\right)\left(\partial_{\lambda}K\right)K^{\ast}_{\nu},$ (13) $\displaystyle{\cal L}_{\gamma\Lambda\Lambda^{\ast}}$ $\displaystyle=$ $\displaystyle-\,ie\frac{g^{(1)}_{\Lambda^{\ast}\Lambda\gamma}}{2M_{\Lambda}}{\bar{\Lambda}}^{\ast\mu}\gamma^{\nu}F_{\mu\nu}\Lambda$ (14) $\displaystyle+\,e\frac{g^{(2)}_{\Lambda^{\ast}\Lambda\gamma}}{(2M_{\Lambda})^{2}}{\bar{\Lambda}}^{\ast\mu}F_{\mu\nu}\partial^{\nu}\Lambda+\text{H.\,c.},$ $\displaystyle{\cal L}_{\gamma\Lambda^{\ast}NK}$ $\displaystyle=$ $\displaystyle- iQ_{K}\frac{g_{\Lambda^{\ast}NK}}{M_{K}}\bar{\Lambda}^{{}^{\ast}\mu}A_{\mu}K\gamma_{5}N+\text{H.\,c.},$ (15) $\displaystyle{\cal L}_{\Lambda^{\ast}NK}$ $\displaystyle=$ $\displaystyle\frac{g_{\Lambda^{\ast}NK}}{M_{K}}{\bar{\Lambda}}^{\ast\mu}\left(\partial_{\mu}K\right)\gamma_{5}N+\text{H.\,c.},$ (16) $\displaystyle{\cal L}_{\Lambda^{\ast}NK^{\ast}}$ $\displaystyle=$ $\displaystyle-\frac{ig_{\Lambda^{\ast}NK^{\ast}}}{M_{K^{\ast}}}\bar{\Lambda}^{{}^{\ast}\mu}\gamma^{\nu}\left(\partial_{\mu}K^{\ast}_{\nu}-\partial_{\nu}K^{\ast}_{\mu}\right)N$ (17) $\displaystyle+\,\text{H.\,c.},$ $\displaystyle{\cal L}_{\Lambda NK}$ $\displaystyle=$ $\displaystyle-ig_{\Lambda NK}\bar{\Lambda}\gamma_{5}KN+\text{H.\,c.},$ (18) where $M_{K^{\ast}}$, $M_{K}$, $M_{N}$, and $M_{\Lambda}$ denote the masses of $K^{\ast}$, $K$, $N$, and $\Lambda$, respectively; $\hat{e}$ stands for the charge operator and $\hat{\kappa}_{N}=\kappa_{p}\left(1+\tau_{3}\right)/2+\kappa_{n}\left(1-\tau_{3}\right)/2$ with the anomalous magnetic moments $\kappa_{p}=1.793$ and $\kappa_{n}=-1.913$. The coupling constant $g_{\gamma KK^{\ast}}=0.413$ is calculated by the radiative decay width of $K^{\ast}\to K\gamma$ given by RPP Tanabashi:2018oca with the sign inferred from $g_{\gamma\pi\rho}$ Garcilazo:1993av via the flavor SU(3) symmetry considerations in conjunction with the vector-meson dominance assumption. The coupling constants $g^{(1)}_{\Lambda^{\ast}\Lambda\gamma}$ and $g^{(2)}_{\Lambda^{\ast}\Lambda\gamma}$ are fit parameters, but only one of them is free since they are constrained by the $\Lambda(1520)$ radiative decay width $\Gamma_{\Lambda(1520)\to\Lambda\gamma}=0.133$ MeV as given by RPP Tanabashi:2018oca . The value of $g_{\Lambda^{\ast}NK}=10.5$ is determined by the decay width of $\Lambda(1520)\to NK$, $\Gamma_{\Lambda(1520)\to NK}=7.079$ MeV, as advocated by RPP Tanabashi:2018oca . The coupling constant $g_{\Lambda^{\ast}NK^{\ast}}$ is a parameter to be determined by fitting the data. The coupling constant $g_{\Lambda NK}\approx-14$ is determined by the flavor SU(3) symmetry, $g_{\Lambda NK}=\left(-3\sqrt{3}/5\right)g_{NN\pi}$ with $g_{NN\pi}=13.46$. For nucleon resonances in the $s$ channel, the Lagrangians for electromagnetic couplings read Wang:2017tpe ; Wang:2018vlv ; Wei:2019imo ; Wang:2020mdn $\displaystyle{\cal L}_{RN\gamma}^{1/2\pm}$ $\displaystyle=$ $\displaystyle e\frac{g_{RN\gamma}^{(1)}}{2M_{N}}\bar{R}\Gamma^{(\mp)}\sigma_{\mu\nu}\left(\partial^{\nu}A^{\mu}\right)N+\text{H.\,c.},$ (19) $\displaystyle{\cal L}_{RN\gamma}^{3/2\pm}$ $\displaystyle=$ $\displaystyle-\,ie\frac{g_{RN\gamma}^{(1)}}{2M_{N}}\bar{R}_{\mu}\gamma_{\nu}\Gamma^{(\pm)}F^{\mu\nu}N$ (20) $\displaystyle+\,e\frac{g_{RN\gamma}^{(2)}}{\left(2M_{N}\right)^{2}}\bar{R}_{\mu}\Gamma^{(\pm)}F^{\mu\nu}\partial_{\nu}N+\text{H.\,c.},$ $\displaystyle{\cal L}_{RN\gamma}^{5/2\pm}$ $\displaystyle=$ $\displaystyle e\frac{g_{RN\gamma}^{(1)}}{\left(2M_{N}\right)^{2}}\bar{R}_{\mu\alpha}\gamma_{\nu}\Gamma^{(\mp)}\left(\partial^{\alpha}F^{\mu\nu}\right)N$ (21) $\displaystyle\pm\,ie\frac{g_{RN\gamma}^{(2)}}{\left(2M_{N}\right)^{3}}\bar{R}_{\mu\alpha}\Gamma^{(\mp)}\left(\partial^{\alpha}F^{\mu\nu}\right)\partial_{\nu}N$ $\displaystyle+\,\text{H.\,c.},$ $\displaystyle{\cal L}_{RN\gamma}^{7/2\pm}$ $\displaystyle=$ $\displaystyle ie\frac{g_{RN\gamma}^{(1)}}{\left(2M_{N}\right)^{3}}\bar{R}_{\mu\alpha\beta}\gamma_{\nu}\Gamma^{(\pm)}\left(\partial^{\alpha}\partial^{\beta}F^{\mu\nu}\right)N$ (22) $\displaystyle-\,e\frac{g_{RN\gamma}^{(2)}}{\left(2M_{N}\right)^{4}}\bar{R}_{\mu\alpha\beta}\Gamma^{(\pm)}\left(\partial^{\alpha}\partial^{\beta}F^{\mu\nu}\right)\partial_{\nu}N$ $\displaystyle+\,\text{H.\,c.},$ and the Lagrangians for hadronic couplings to $\Lambda(1520)K$ read $\displaystyle{\cal L}_{R\Lambda^{\ast}K}^{1/2\pm}$ $\displaystyle=$ $\displaystyle\frac{g^{(1)}_{R\Lambda^{\ast}K}}{M_{K}}\bar{\Lambda}^{\ast\mu}\Gamma^{(\pm)}\left(\partial_{\mu}K\right)R+\text{H.\,c.},$ (23) $\displaystyle{\cal L}_{R\Lambda^{\ast}K}^{3/2\pm}$ $\displaystyle=$ $\displaystyle\frac{g^{(1)}_{R\Lambda^{\ast}K}}{M_{K}}\bar{\Lambda}^{\ast\mu}\gamma_{\nu}\Gamma^{(\mp)}\left(\partial^{\nu}K\right)R_{\mu}$ (24) $\displaystyle+\,i\frac{g^{(2)}_{R\Lambda^{\ast}K}}{M_{K}^{2}}\bar{\Lambda}^{\ast}_{\alpha}\Gamma^{(\mp)}\left(\partial^{\mu}\partial^{\alpha}K\right)R_{\mu}+\text{H.\,c.},$ $\displaystyle{\cal L}_{R\Lambda^{\ast}K}^{5/2\pm}$ $\displaystyle=$ $\displaystyle i\frac{g^{(1)}_{R\Lambda^{\ast}K}}{M_{K}^{2}}\bar{\Lambda}^{\ast\alpha}\gamma_{\mu}\Gamma^{(\pm)}\left(\partial^{\mu}\partial^{\beta}K\right)R_{\alpha\beta}$ (25) $\displaystyle-\,\frac{g^{(2)}_{R\Lambda^{\ast}K}}{M_{K}^{3}}\bar{\Lambda}^{\ast}_{\mu}\Gamma^{(\pm)}\left(\partial^{\mu}\partial^{\alpha}\partial^{\beta}K\right)R_{\alpha\beta}$ $\displaystyle+\,\text{H.\,c.},$ $\displaystyle{\cal L}_{R\Lambda^{\ast}K}^{7/2\pm}$ $\displaystyle=$ $\displaystyle-\frac{g^{(1)}_{R\Lambda^{\ast}K}}{M_{K}^{3}}\bar{\Lambda}^{\ast\alpha}\gamma_{\mu}\Gamma^{(\mp)}\left(\partial^{\mu}\partial^{\beta}\partial^{\lambda}K\right)R_{\alpha\beta\lambda}$ (26) $\displaystyle-\,i\frac{g^{(2)}_{R\Lambda^{\ast}K}}{M_{K}^{4}}\bar{\Lambda}^{\ast}_{\mu}\Gamma^{(\mp)}\left(\partial^{\mu}\partial^{\alpha}\partial^{\beta}\partial^{\lambda}K\right)R_{\alpha\beta\lambda}$ $\displaystyle+\,\text{H.\,c.},$ where $R$ designates the $N^{\ast}$ resonance and the superscript of ${\cal L}_{RN\gamma}$ and ${\cal L}_{R\Lambda^{\ast}K}$ denotes the spin and parity of the resonance $R$. The coupling constants $g_{RN\gamma}^{(i)}$ and $g^{(i)}_{R\Lambda^{\ast}K}$ $(i=1,2)$ are fit parameters. Actually, only the products of $g_{RN\gamma}^{(i)}g^{(j)}_{R\Lambda^{\ast}K}$ $(i,j=1,2)$ are relevant to the reaction amplitudes, and they are what we really fit in practice. In Ref. Nam:2005uq , the off-shell effects for spin-$3/2$ resonances in $\gamma p\to K^{+}\Lambda(1520)$ have been tested. It was found that the off- shell effects are small and the off-shell parameter $X$ can be set to zero. In the present work, we simply ignore the off-shell terms in the interaction Lagrangians for high spin resonances and leave this issue for future work. ### II.2 Resonance propagators We follow Ref. Wang:2017tpe to use the following prescriptions for the propagators of resonances with spin $1/2$, $3/2$, $5/2$, and $7/2$: $\displaystyle S_{1/2}(p)$ $\displaystyle=$ $\displaystyle\frac{i}{{p\\!\\!\\!/}-M_{R}+i\Gamma_{R}/2},$ (27) $\displaystyle S_{3/2}(p)$ $\displaystyle=$ $\displaystyle\frac{i}{{p\\!\\!\\!/}-M_{R}+i\Gamma_{R}/2}\left(\tilde{g}_{\mu\nu}+\frac{1}{3}\tilde{\gamma}_{\mu}\tilde{\gamma}_{\nu}\right),$ (28) $\displaystyle S_{5/2}(p)$ $\displaystyle=$ $\displaystyle\frac{i}{{p\\!\\!\\!/}-M_{R}+i\Gamma_{R}/2}\,\bigg{[}\,\frac{1}{2}\big{(}\tilde{g}_{\mu\alpha}\tilde{g}_{\nu\beta}+\tilde{g}_{\mu\beta}\tilde{g}_{\nu\alpha}\big{)}$ (29) $\displaystyle-\,\frac{1}{5}\tilde{g}_{\mu\nu}\tilde{g}_{\alpha\beta}+\frac{1}{10}\big{(}\tilde{g}_{\mu\alpha}\tilde{\gamma}_{\nu}\tilde{\gamma}_{\beta}+\tilde{g}_{\mu\beta}\tilde{\gamma}_{\nu}\tilde{\gamma}_{\alpha}$ $\displaystyle+\,\tilde{g}_{\nu\alpha}\tilde{\gamma}_{\mu}\tilde{\gamma}_{\beta}+\tilde{g}_{\nu\beta}\tilde{\gamma}_{\mu}\tilde{\gamma}_{\alpha}\big{)}\bigg{]},$ $\displaystyle S_{7/2}(p)$ $\displaystyle=$ $\displaystyle\frac{i}{{p\\!\\!\\!/}-M_{R}+i\Gamma_{R}/2}\,\frac{1}{36}\sum_{P_{\mu}P_{\nu}}\bigg{(}\tilde{g}_{\mu_{1}\nu_{1}}\tilde{g}_{\mu_{2}\nu_{2}}\tilde{g}_{\mu_{3}\nu_{3}}$ (30) $\displaystyle-\,\frac{3}{7}\tilde{g}_{\mu_{1}\mu_{2}}\tilde{g}_{\nu_{1}\nu_{2}}\tilde{g}_{\mu_{3}\nu_{3}}+\frac{3}{7}\tilde{\gamma}_{\mu_{1}}\tilde{\gamma}_{\nu_{1}}\tilde{g}_{\mu_{2}\nu_{2}}\tilde{g}_{\mu_{3}\nu_{3}}$ $\displaystyle-\,\frac{3}{35}\tilde{\gamma}_{\mu_{1}}\tilde{\gamma}_{\nu_{1}}\tilde{g}_{\mu_{2}\mu_{3}}\tilde{g}_{\nu_{2}\nu_{3}}\bigg{)},$ where $\displaystyle\tilde{g}_{\mu\nu}$ $\displaystyle=$ $\displaystyle-\,g_{\mu\nu}+\frac{p_{\mu}p_{\nu}}{M_{R}^{2}},$ (31) $\displaystyle\tilde{\gamma}_{\mu}$ $\displaystyle=$ $\displaystyle\gamma^{\nu}\tilde{g}_{\nu\mu}=-\gamma_{\mu}+\frac{p_{\mu}{p\\!\\!\\!/}}{M_{R}^{2}},$ (32) and the summation over $P_{\mu}$ $\left(P_{\nu}\right)$ in Eq. (30) goes over the $3!=6$ possible permutations of the indices $\mu_{1}\mu_{2}\mu_{3}$ $\left(\nu_{1}\nu_{2}\nu_{3}\right)$. In Eqs. (27)$-$(32), $M_{R}$ and $\Gamma_{R}$ are the mass and width of resonance $R$ with four-momentum $p$, respectively. ### II.3 Form factors In practical calculation of the reaction amplitudes, a phenomenological form factor is introduced in each hadronic vertex. For the $t$-channel meson exchanges, we adopt the following form factor Wang:2017tpe ; Wang:2020mdn ; Wang:2018vlv ; Wei:2019imo : $\displaystyle f_{M}(q^{2}_{M})=\left(\frac{\Lambda_{M}^{2}-M_{M}^{2}}{\Lambda_{M}^{2}-q^{2}_{M}}\right)^{2},$ (33) and for the $s$-channel and $u$-channel baryon exchanges, we use Wang:2017tpe ; Wang:2020mdn ; Wang:2018vlv ; Wei:2019imo $\displaystyle f_{B}(p^{2}_{x})=\left(\frac{\Lambda_{B}^{4}}{\Lambda_{B}^{4}+\left(p_{x}^{2}-M_{B}^{2}\right)^{2}}\right)^{2}.$ (34) Here, $q_{M}$ denotes the four-momentum of the intermediate meson in the $t$ channel, and $p_{x}$ stands for the four-momentum of the intermediate baryon in $s$ and $u$ channels with $x=$ $s$ and $u$, respectively. $\Lambda_{M(B)}$ is the corresponding cutoff parameter. In the present work, in order to reduce the number of adjustable parameters, we use the same cutoff parameter $\Lambda_{B}$ for all the nonresonant diagrams, i.e., $\Lambda_{B}\equiv\Lambda_{K}=\Lambda_{K^{\ast}}=\Lambda_{\Lambda}=\Lambda_{N}$. The parameter $\Lambda_{B}$ and the cutoff parameter $\Lambda_{R}$ for $N^{\ast}$ resonances are determined by fitting the experimental data. ### II.4 Interpolated $t$-channel Regge amplitudes A Reggeized treatment of the $t$-channel $K$ and $K^{\ast}$ exchanges is usually employed to economically describe the high-energy data, which corresponds to the following replacement of the form factors in Feynman amplitudes: $\displaystyle f_{K}(q_{K}^{2})\to{\cal F}_{K}(q_{K}^{2})=$ $\displaystyle\left(\frac{s}{s_{0}}\right)^{\alpha_{K}(t)}\frac{\pi\alpha^{\prime}_{K}}{\sin[\pi\alpha_{K}(t)]}$ $\displaystyle\times\frac{t-M^{2}_{K}}{\Gamma[1+\alpha_{K}(t)]},$ (35) $\displaystyle f_{K^{\ast}}(q_{K^{\ast}}^{2})\to{\cal F}_{K^{\ast}}(q_{K^{\ast}}^{2})=$ $\displaystyle\left(\frac{s}{s_{0}}\right)^{\alpha_{K^{\ast}}(t)-1}\frac{\pi\alpha^{\prime}_{K^{\ast}}}{\sin[\pi\alpha_{K^{\ast}}(t)]}$ $\displaystyle\times\frac{t-M^{2}_{K^{\ast}}}{\Gamma[\alpha_{K^{\ast}}(t)]}.$ (36) Here $s_{0}$ is a mass scale which is conventionally taken as $s_{0}=1$ GeV2, and $\alpha^{\prime}_{M}$ is the slope of the Regge trajectory $\alpha_{M}(t)$. For $M=K$ and $K^{\ast}$, the trajectories are parameterized as Wang:2019mid $\displaystyle\alpha_{K}(t)$ $\displaystyle=0.7~{}{\rm GeV}^{-2}\left(t-m_{K}^{2}\right),$ (37) $\displaystyle\alpha_{K^{\ast}}(t)$ $\displaystyle=1+0.85~{}{\rm GeV}^{-2}\left(t-m_{K^{\ast}}^{2}\right).$ (38) Note that, in Eqs. (35) and (36), degenerate trajectories are employed for $K$ and $K^{\ast}$ exchanges; thus, the signature factors reduce to $1$. In the present work, we use the so-called interpolated Regge amplitudes for the $t$-channel $K$ and $K^{\ast}$ exchanges. The idea of this prescription is that at high energies and small angles one uses Regge amplitudes, and at low energies one uses Feynman amplitudes, while in the intermediate energy region an interpolating form factor is introduced to ensure a smooth transition from the low-energy Feynman amplitudes to the high-energy Regge amplitudes. This hybrid Regge approach has been applied to study the $\gamma p\to K^{+}\Lambda(1520)$ reaction in Refs. Nam:2010au ; Wang:2014jxb ; He:2014gga ; Yu:2017kng and the other reactions in Refs. Wang:2015hfm ; Wang:2017plf ; Wang:2017qcw ; Wang:2019mid . Instead of making the replacements of Eqs. (35) and (36) in a pure Reggeized treatment, in this hybrid Regge model the amplitudes for $t$-channel $K$ and $K^{\ast}$ exchanges are constructed by making the following replacements of the form factors in the corresponding Feynman amplitudes: $f_{M}(q_{M}^{2})\to{\cal F}_{R,M}={\cal F}_{M}(q_{M}^{2})R+f_{M}(q_{M}^{2})\left(1-R\right),$ (39) where ${\cal F}_{M}(q_{M}^{2})$ $(M=K,K^{\ast})$ is defined in Eqs. (35) and (36) and $R=R_{s}R_{t}$ with $\displaystyle R_{s}=$ $\displaystyle\frac{1}{1+e^{-(s-s_{R})/s_{0}}},$ $\displaystyle R_{t}=$ $\displaystyle\frac{1}{1+e^{-(t+t_{R})/t_{0}}}.$ (40) Here $s_{R}$, $t_{R}$, $s_{0}$, and $t_{0}$ are parameters to be determined by fitting the experimental data. The auxiliary current $C^{\mu}$ introduced in Eq. (5) and the interaction current $M^{\nu\mu}_{\rm int}$ given in Eq. (2) ensures that the full photoproduction amplitude of Eq. (1) satisfies the generalized Ward-Takahashi identity and, thus, is fully gauge invariant Haberzettl:2006 ; Haberzettl:2011zr ; Huang:2012 . Note that our prescription for $C^{\mu}$ and $M^{\nu\mu}_{\rm int}$ is independent of any particular form of the $t$-channel form factor $f_{K}(q_{K}^{2})$, provided that it is normalized as $f_{K}(q_{K}^{2}=M_{K}^{2})=1$. One sees that, when the interpolated Regge amplitude is employed for $t$-channel $K$ exchange, the replacement of Eq. (39) still keeps the the normalization condition of the form factor: $\lim_{q^{2}_{K}\to M^{2}_{K}}{\cal F}_{R,K}=1.$ (41) Therefore, as soon as we do the same replacement of Eq. (39) for the form factor of $t$-channel $K$ exchange everywhere in $C^{\mu}$ and $M^{\nu\mu}_{\rm int}$, the full photoproduction amplitude still satisfies the generalized Ward-Takahashi identity and, thus, is fully gauge invariant. ## III Results and discussion As discussed in the introduction section of this paper, the reaction $\gamma p\to K^{+}\Lambda(1520)$ has been theoretically investigated based on effective Lagrangian approaches by four theory groups in $11$ publications Nam:2005uq ; Nam:2006cx ; Nam:2009cv ; Nam:2010au ; Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb ; He:2012ud ; He:2014gga ; Yu:2017kng . The common feature of the results from these theoretical works is that the contributions from the contact term and the $t$-channel $K$ exchange are important for the $\gamma p\to K^{+}\Lambda(1520)$ reaction. Apart from that, no common ground has been found by these theoretical works for the reaction mechanisms of $\gamma p\to K^{+}\Lambda(1520)$. In particular, different groups gave quite different answers for the following questions: Are the contributions from the $t$-channel $K^{\ast}$ exchange and $u$-channel $\Lambda$ exchange significant or not in this reaction, are the nucleon resonances introduced in the $s$ channel indispensable or not to describe the available data, and, if yes, what are the resonance contents and their associated parameters in this reaction? On the other hand, we notice that even though the data on photon-beam asymmetries for $\gamma p\to K^{+}\Lambda(1520)$ have been reported by the LEPS Collaboration in 2010, they have never been well reproduced in previous theoretical publications of Refs. Nam:2005uq ; Nam:2006cx ; Nam:2009cv ; Nam:2010au ; Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb ; He:2012ud ; He:2014gga . One believes that these photon-beam-asymmetry data will definitely put further constraints on the reaction amplitudes. In Ref. Yu:2017kng , the photon-beam-asymmetry data have indeed been analyzed, but there, the structures of the angular distributions exhibited by the data are missed due to the lack of nucleon resonances. In a word, all previous theoretical publications in regards to $\gamma p\to K^{+}\Lambda(1520)$ are divided over the reaction mechanism and the resonance contents and parameters of this reaction. A simultaneous description of the differential cross-section data and the photon-beam-asymmetry data still remains to be accomplished. The purpose of the present work is to get a clear understanding of the reaction mechanism of $\gamma p\to K^{+}\Lambda(1520)$ based on a combined analysis of the available data on both the differential cross sections and the photon-beam asymmetries within an effective Lagrangian approach. As the differential cross-section data exhibit clear bump structures in the near- threshold region, apart from the $N$, $K$, $K^{\ast}$, and $\Lambda$ exchanges and the interaction current in the nonresonant background, we introduce as few as possible near-threshold nucleon resonances in the $s$ channel in constructing the $\gamma p\to K^{+}\Lambda(1520)$ reaction amplitudes to reproduce the data. Figure 3: Differential cross sections for $\gamma p\to K^{+}\Lambda(1520)$ at a few selected scattering angles as a function of the photon incident energy. The black solid lines, red dot-double-dashed lines, blue dashed lines, and green dot-dashed lines denote the results obtained by including the $N(2000)5/2^{+}$, $N(2040)3/2^{+}$, $N(2100)1/2^{+}$, and $N(2190)7/2^{-}$ resonances in the $s$ channel, respectively. Data are taken from the CLAS Collaboration Moriya:2013hwg (red circles) and the LEPS Collaboration Kohri:2009xe (blue squares). For $\cos\theta=0.85$, the CLAS data at $\cos\theta=0.84$ ($E_{\gamma}<3.25$ GeV) and $\cos\theta=0.83$ ($E_{\gamma}>3.25$ GeV) are shown. In the most recent version of RPP Tanabashi:2018oca , there are six nucleon resonances near the $K^{+}\Lambda(1520)$ threshold, namely, the $N(2000)5/2^{+}$, $N(2040)3/2^{+}$, $N(2060)5/2^{-}$, $N(2100)1/2^{+}$, $N(2120)3/2^{-}$, and $N(2190)7/2^{-}$ resonances. If none of these nucleon resonances are introduced in the construction of the $s$-channel reaction amplitudes, we find that it is not possible to achieve a simultaneous description of both the differential cross-section data and the photon-beam- asymmetry data in our model. We then try to reproduce the data by including one of these six near-threshold resonances. If we include one of the $N(2000)5/2^{+}$, $N(2040)3/2^{+}$, $N(2100)1/2^{+}$, and $N(2190)7/2^{-}$ resonances, we find that the obtained theoretical results for differential cross sections and photon-beam asymmetries have rather poor fitting qualities. As an illustration, we show in Fig. 3 the differential cross sections at a few selected scattering angles as a function of the incident photon energy which are obtained by including one of the $N(2000)5/2^{+}$ (black solid lines), $N(2040)3/2^{+}$ (red dot-double-dashed lines), $N(2100)1/2^{+}$ (blue dashed lines), and $N(2190)7/2^{-}$ (green dot-dashed lines) resonances and compared with the corresponding data Kohri:2009xe ; Moriya:2013hwg . One sees clearly from Fig. 3 that the fits with one of the $N(2040)3/2^{+}$, $N(2100)1/2^{+}$, and $N(2190)7/2^{-}$ resonances fail to describe the differential cross sections at $\cos\theta=0.95$, and the fit with the $N(2000)5/2^{+}$ resonance fails to reproduce the differential cross-section data at the other three selected scattering angles. In a word, none of these four fits that includes one of the $N(2000)5/2^{+}$, $N(2040)3/2^{+}$, $N(2100)1/2^{+}$, and $N(2190)7/2^{-}$ resonances can well describe the differential cross-section data. Thus, they are excluded to be acceptable fits. On the other hand, if either the resonance $N(2060)5/2^{-}$ or the resonance $N(2120)3/2^{-}$ is considered, a simultaneous description of both the differential cross-section data and the photon-beam-asymmetry data can be satisfactorily obtained, which will be discussed below in detail. Consequently, these two fits, i.e., the ones including the $N(2060)5/2^{-}$ or the $N(2120)3/2^{-}$ resonance, are treated as acceptable. When an additional resonance is further included, the fit quality will be improved a little bit, since one has more adjustable model parameters. But, in this case, one would obtain too many solutions with similar fitting qualities, and meanwhile the fitted error bars of adjustable parameters are also relatively large. As a consequence, no conclusive conclusion can be drawn about the resonance contents and parameters extracted from the available data for the considered reaction. We thus conclude that the available differential cross-section data and the photon-beam-asymmetry data for $\gamma p\to K^{+}\Lambda(1520)$ can be described by including one of the $N(2060)5/2^{-}$ and $N(2120)3/2^{-}$ resonances and postpone the analysis of these available data with two or more nucleon resonances until more data for this reaction become available in the future. Table 1: Fitted values of model parameters. The asterisks below resonance names represent the overall status of these resonances evaluated by RPP Tanabashi:2018oca . The numbers in the brackets below the resonance masses and widths denote the corresponding values advocated by RPP Tanabashi:2018oca . $\sqrt{\beta_{\Lambda^{\ast}K}}A_{j}$ represents the reduced helicity amplitude for resonance with $\beta_{\Lambda^{\ast}K}$ denoting the branching ratio of resonance decay to $\Lambda(1520)K$ and $A_{j}$ standing for the helicity amplitude with spin $j$ for resonance radiative decay to $\gamma p$. | Fit A | Fit B ---|---|--- $s_{R}$ $[{\rm GeV}^{2}]$ | $5.17\pm 0.02$ | $3.80\pm 0.12$ $s_{0}$ $[{\rm GeV}^{2}]$ | $0.81\pm 0.02$ | $8.00\pm 0.05$ $t_{R}$ $[{\rm GeV}^{2}]$ | $0.80\pm 0.03$ | $1.16\pm 0.07$ $t_{0}$ $[{\rm GeV}^{2}]$ | $1.60\pm 0.07$ | $0.96\pm 0.07$ $\Lambda_{B}$ $[{\rm MeV}]$ | $748\pm 2$ | $770\pm 5$ $g^{(1)}_{\Lambda^{\ast}\Lambda\gamma}$ | $0.00\pm 0.01$ | $8.99\pm 0.51$ $g_{\Lambda^{\ast}NK^{\ast}}$ | $-22.48\pm 0.91$ | $-54.22\pm 3.72$ | $N(2060){5/2}^{-}$ | $N(2120){3/2}^{-}$ | $\ast$$\ast$$\ast$ | $\ast$$\ast$$\ast$ $M_{R}$ $[{\rm MeV}]$ | $2020\pm 1$ | $2184\pm 2$ | $[2030$$-$$2200]$ | $[2060$$-$$2160]$ $\Gamma_{R}$ $[{\rm MeV}]$ | $200\pm 30$ | $83\pm 4$ | $[300$$-$$450]$ | $[260$$-$$360]$ $\Lambda_{R}$ $[{\rm MeV}]$ | $1086\pm 3$ | $2000\pm 52$ $\sqrt{\beta_{\Lambda^{\ast}K}}A_{1/2}$ $[10^{-3}\,{\rm GeV}^{-1/2}]$ | $3.07\pm 0.02$ | $3.04\pm 0.11$ $\sqrt{\beta_{\Lambda^{\ast}K}}A_{3/2}$ $[10^{-3}\,{\rm GeV}^{-1/2}]$ | $0.54\pm 0.02$ | $5.27\pm 0.19$ $g^{(2)}_{R\Lambda^{\ast}K}/g^{(1)}_{R\Lambda^{\ast}K}$ | $-1.26\pm 0.01$ | $-4.06\pm 0.28$ As discussed above, we introduce nucleon resonances as few as possible in constructing the reaction amplitudes to describe the available data for $\gamma p\to K^{+}\Lambda(1520)$. It is found that a simultaneous description of both the differential cross-section data and the photon-beam-asymmetry data can be achieved by including either the $N(2060)5/2^{-}$ resonance or the $N(2120)3/2^{-}$ resonance. We thus get two acceptable fits named as “fit A,” which includes the $N(2060)5/2^{-}$ resonance, and “fit B,” which includes the $N(2120)3/2^{-}$ resonance. The fitted values of the adjustable model parameters in these two fits are listed in Table 1, and the corresponding results on differential cross sections and photon-beam asymmetries are shown in Figs. 4$-$6. In Table 1, for $u$-channel $\Lambda$ exchange, only the value of the coupling constant $g^{(1)}_{\Lambda^{\ast}\Lambda\gamma}$ is listed. The other coupling constant $g^{(2)}_{\Lambda^{\ast}\Lambda\gamma}$ is not treated as a free parameter, since it is constrained by the $\Lambda(1520)$ radiative decay width $\Gamma_{\Lambda(1520)\to\Lambda\gamma}=0.133$ MeV as given by RPP Tanabashi:2018oca , which results in $g^{(2)}_{\Lambda^{\ast}\Lambda\gamma}=2.13$ in fit A and $-13.01$ in fit B, respectively. The asterisks below the resonance names represent the overall status of these resonances evaluated in the most recent RPP Tanabashi:2018oca . One sees that both the $N(2060)5/2^{-}$ and the $N(2120)3/2^{-}$ resonances are evaluated as three-star resonances. The symbols $M_{R}$, $\Gamma_{R}$, and $\Lambda_{R}$ denote the resonance mass, width, and cutoff parameter, respectively. The numbers in brackets below the resonance mass and width are the corresponding values estimated by RPP. It is seen that the fitted masses of the $N(2060)5/2^{-}$ and $N(2120)3/2^{-}$ resonances are comparable with their values quoted by RPP, while the fitted widths for these two resonances are smaller than the corresponding RPP values. For resonance couplings, since in the tree-level calculation only the products of the resonance hadronic and electromagnetic coupling constants are relevant to the reaction amplitudes, we list the reduced helicity amplitudes $\sqrt{\beta_{\Lambda^{\ast}K}}A_{j}$ for each resonance instead of showing their hadronic and electromagnetic coupling constants separately Huang:2013 ; Wang:2017tpe ; Wang:2018vlv ; Wang:2019mid . Here $\beta_{\Lambda^{\ast}K}$ is the branching ratio for resonance decay to $\Lambda(1520)K$, and $A_{j}$ is the helicity amplitude with spin $j$ ($j=1/2,3/2$) for resonance radiative decay to $\gamma p$. We have, in total, as shown in Figs. 4$-$6, $220$ data points in the fits. Fit A has a global $\chi^{2}/N=2.10$, and fit B has a global $\chi^{2}/N=2.63$. Note that, in the fitting procedure, $11.6\%$ and $5.92\%$ systematic errors for the data from the CLAS Collaboration and the LEPS Collaboration, respectively, have been added in quadrature to the statistical errors Moriya:2013hwg ; Kohri:2009xe . Overall, one sees that both the differential cross-section data and the photon-beam-asymmetry data have been well described simultaneously in both fit A and fit B. Figure 4: Differential cross sections for $\gamma p\to K^{+}\Lambda(1520)$ as a function of $\cos\theta$ from fit A (left panel) and fit B (right panel). The symbols $W$ and $E_{\gamma}$ denote the center-of-mass energy of the whole system and the photon laboratory energy, respectively, both in MeV. The black solid lines represent the results calculated from the full amplitudes. The red dotted lines, blue dashed lines, green dot-dashed lines, cyan double-dot- dashed lines, and magenta dot-double-dashed lines denote the individual contributions from the interaction current, the $t$-channel $K$ exchange, the $t$-channel $K^{\ast}$ exchange, the $s$-channel $N^{\ast}$ resonance exchange, and the $u$-channel $\Lambda$ exchange, respectively. The scattered symbols are data from the CLAS Collaboration Moriya:2013hwg . Figure 5: Differential cross sections for $\gamma p\to K^{+}\Lambda(1520)$ at a few selected scattering angles as a function of the photon incident energy from fit A (left panel) and fit B (right panel). The notations for the lines are the same as in Fig. 4. Data are taken from the CLAS Collaboration Moriya:2013hwg (red circles) and the LEPS Collaboration Kohri:2009xe (blue squares). For $\cos\theta=0.85$, the CLAS data at $\cos\theta=0.84$ ($E_{\gamma}<3.25$ GeV) and $\cos\theta=0.83$ ($E_{\gamma}>3.25$ GeV) are shown. Figures 4 and 5 show the differential cross sections for $\gamma p\to K^{+}\Lambda(1520)$ resulted from fit A (left panels), which includes the $N(2060)5/2^{-}$ resonance, and fit B (right panels), which includes the $N(2120)3/2^{-}$ resonance. There, the black solid lines represent the results calculated from the full reaction amplitudes. The red dotted lines, blue dashed lines, green dot-dashed lines, cyan double-dot-dashed lines, and magenta dot-double-dashed lines denote the individual contributions from the interaction current, the $t$-channel $K$ exchange, the $t$-channel $K^{\ast}$ exchange, the $s$-channel $N^{\ast}$ resonance exchange, and the $u$-channel $\Lambda$ exchange, respectively. The individual contributions from the $s$-channel nucleon exchange are too small to be clearly shown in these figures. One sees from Figs. 4 and 5 that the differential cross-section data are well reproduced in both fit A (left panels) and fit B (right panels). Note that in Fig. 5, for $\cos\theta=0.85$, the CLAS data at $\cos\theta=0.84$ ($E_{\gamma}<3.25$ GeV) and $\cos\theta=0.83$ ($E_{\gamma}>3.25$ GeV) are shown. That explains why in Fig. 4 the theoretical results agree with the CLAS data at high-energy forward angles but in Fig. 5 the theoretical differential cross sections at $\cos\theta=0.85$ overestimate the CLAS data at the last two energy points. From Figs. 4 and 5, one sees that, in fit A, the contribution from the interaction current [cf. Eq. (2)] plays a rather important role in the whole energy region. In the near-threshold region, the differential cross sections are dominated by the interaction current and the $N(2060)5/2^{-}$ resonance exchange. Actually, the contributions from these two terms are responsible for the sharp rise of differential cross sections near the $K^{+}\Lambda(1520)$ threshold, in particular, the bump structure near $E_{\gamma}\approx 2$ GeV at forward angles as exhibited by the LEPS data in Fig. 5. The $t$-channel $K$ exchange is seen to contribute significantly at higher energies and forward angles. The $t$-channel $K^{\ast}$ exchange has tiny contributions at high- energy forward angles, while the contributions from the $u$-channel $\Lambda$ exchange are negligible. In fit B, the interaction current plays a dominant role in the whole energy region and is also responsible for the sharp rise of the differential cross sections at forward angles near the $K^{+}\Lambda(1520)$ threshold. The bump structure near $E_{\gamma}\approx 2$ GeV at forward angles as exhibited by the LEPS data in Fig. 5 is caused by the $N(2120)3/2^{-}$ resonance on the base of the background dominated by the interaction current. The $t$-channel $K$ exchange and the $u$-channel $\Lambda$ exchange have significant contributions at forward and backward angles, respectively, mostly at higher energies. Considerable contributions are also seen from the $t$-channel $K^{\ast}$ exchange at high-energy forward angles. Figure 6: Photon-beam asymmetries for $\gamma p\to K^{+}\Lambda(1520)$ at $\cos\theta=0.8$ as a function of the photon incident energy from fit A (left panel) and fit B (right panel). The black solid lines represent the results calculated from the full amplitudes. The red dotted lines, blue dashed lines, green dot-dashed lines, cyan double-dot-dashed lines, and magenta dot-double- dashed lines denote the results obtained by switching off the contributions of the interaction current, the $t$-channel $K$ exchange, the $t$-channel $K^{\ast}$ exchange, the $s$-channel $N^{\ast}$ resonance exchange, and the $u$-channel $\Lambda$ exchange, respectively, from the full model. Data are in the bin $0.6<\cos\theta<1$ and taken from the LEPS Collaboration Kohri:2009xe . The results of photon-beam asymmetries for $\gamma p\to K^{+}\Lambda(1520)$ from fit A and fit B are shown, respectively, in the left and right panels in Fig. 6. There, the black solid lines represent the results calculated from the full amplitudes. The red dotted lines, blue dashed lines, green dot-dashed lines, cyan double-dot-dashed lines, and magenta dot-double-dashed lines denote the results obtained by switching off the contributions of the interaction current, the $t$-channel $K$ exchange, the $t$-channel $K^{\ast}$ exchange, the $s$-channel $N^{\ast}$ resonance exchange, and the $u$-channel $\Lambda$ exchange, respectively, from the full model. One sees that the photon-beam-asymmetry data are well reproduced in both fits. In fit A, when the contributions of the $N(2060){5/2}^{-}$ resonance exchange are switched off from the full model, one gets almost zero beam asymmetries. We have checked and found that the $N(2060){5/2}^{-}$ resonance exchange alone results in negligible beam asymmetries. This means that it is the interference between the $N(2060){5/2}^{-}$ resonance exchange and the other interaction terms that is crucial for reproducing the experimental values of the beam asymmetries. A similar observation also holds for the interaction current [cf. Eq. (2)]. The interaction current alone results in almost zero beam asymmetries, but one gets rather negative beam asymmetries when the contributions from the interaction current are switched off from the full model. This means that the interference between the interaction current and the other interaction terms is very important for reproducing the beam asymmetries. Switching off the contributions of the individual terms other than the $N(2060){5/2}^{-}$ resonance exchange and the interaction current from the full model does not affect too much the theoretical beam asymmetries. In fit B, the interaction current alone is found to result in almost zero beam asymmetries, the same as in fit A. Nevertheless, it is seen from Fig. 6 that one gets rather negative beam asymmetries when the contributions of the interaction current are switched off from the full model, showing the importance of the interference of the interaction current and the other interacting terms in photon-beam asymmetries for $\gamma p\to K^{+}\Lambda(1520)$. Switching off the contributions of the individual terms other than the interaction current from the full model would not affect the theoretical beam asymmetries too much. In Ref. Kohri:2009xe , it is expected that the positive values of the $K^{+}\Lambda(1520)$ asymmetries indicate a much larger contribution from the $K^{\ast}$ exchange. In both fit A and fit B of the present work, we have checked and found that the $K^{\ast}$ exchange alone does result in positive beam asymmetries, but, when the contributions of the $K^{\ast}$ exchange are switched off from the full model, the calculated beam asymmetries do not change significantly. In particular, the theoretical beam asymmetries are still positive and close to the experimental values when the contributions of the $K^{\ast}$ exchange are switched off from the full model. Figure 7: Total cross sections for $\gamma p\to K^{+}\Lambda(1520)$ predicated by fit A (left panel) and fit B (right panel). Notations for the lines are the same as in Fig. 4. Data are taken from the CLAS Collaboration Moriya:2013hwg but not included in the fits. Figure 7 shows the total cross sections for $\gamma p\to K^{+}\Lambda(1520)$ predicated from fit A (left panel) and fit B (right panel), which are obtained by integrating the corresponding differential cross sections calculated in these two fits. In Fig. 7, the black solid lines represent the results calculated from the full reaction amplitudes. The red dotted lines, blue dashed lines, green dot-dashed lines, cyan double-dot-dashed lines, and magenta dot-double-dashed lines denote the individual contributions from the interaction current, the $t$-channel $K$ exchange, the $t$-channel $K^{\ast}$ exchange, the $s$-channel $N^{\ast}$ resonance exchange, and the $u$-channel $\Lambda$ exchange, respectively. The individual contributions from the $s$-channel nucleon exchange are too small to be clearly shown in these figures. Note that the data for the total cross sections of $\gamma p\to K^{+}\Lambda(1520)$ are not included in the fits. Even so, one sees that, in both fit A and fit B, the theoretical total cross sections are in good agreement with the data. In fit A, the $s$-channel $N(2060){5/2}^{-}$ exchange, the interaction current, and the $t$-channel $K$ exchange provide the most important contributions to the total cross sections, while the contributions from the $u$-channel $\Lambda$ exchange, the $s$-channel $N$ exchange, and the $t$-channel $K^{\ast}$ exchange are negligible. The bump structure near $E_{\gamma}\approx 2$ GeV is caused mainly by the $N(2060){5/2}^{-}$ resonance exchange and the interaction current. The sharp rise of the total cross sections near the $K^{+}\Lambda(1520)$ threshold is dominated by the $s$-channel $N(2060){5/2}^{-}$ exchange. In fit B, the dominant contributions to the total cross sections come from the interaction current, which is also responsible for the sharp rise of the total cross sections near the $K^{+}\Lambda(1520)$ threshold. The individual contributions from the $s$-channel $N(2120){3/2}^{-}$ exchange, the $t$-channel $K$ and $K^{\ast}$ exchanges, and the $u$-channel $\Lambda$ exchange are considerable, while those from the $s$-channel $N$ exchange are negligible to the total cross sections. Comparing the individual contributions in fit A and fit B, one sees that the contributions from the resonance exchange are rather important in fit A, but they are much smaller in fit B. The contributions from the $t$-channel $K^{\ast}$ exchange and the $u$-channel $\Lambda$ exchange are negligible in fit A, but they are considerable in fit B. In both fits, the interaction current provides dominant contributions, and the $t$-channel $K$ exchange results in considerable contributions to the cross sections. As mentioned in the introduction section, the $K^{+}\Lambda(1520)$ photoproduction reaction has been theoretically investigated based on effective Lagrangian approaches by four theory groups in $11$ publications Nam:2005uq ; Nam:2006cx ; Nam:2009cv ; Nam:2010au ; Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb ; He:2012ud ; He:2014gga ; Yu:2017kng . In these previous publications, the photon-beam-asymmetry data reported by the LEPS Collaboration in 2010 Kohri:2009xe have never been well reproduced except in Ref. Yu:2017kng . But in Ref. Yu:2017kng , the structures of the angular distributions exhibited by the data are missed due to the lack of nucleon resonances in the employed Reggeized model. As shown in Figs. 4$-$6, the present work for the first time presents a simultaneous description of the data on both differential cross sections and photon-beam asymmetries within an effective Lagrangian approach. The common feature of the results from the previous works Nam:2005uq ; Nam:2006cx ; Nam:2009cv ; Nam:2010au ; Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb ; He:2012ud ; He:2014gga ; Yu:2017kng is that the contributions from the contact term and the $t$-channel $K$ exchange are important for the $\gamma p\to K^{+}\Lambda(1520)$ reaction. This feature has also been observed in the present work, as illustrated in Fig. 7. The contributions of nucleon resonance exchanges are reported to be small in Refs. Nam:2005uq ; Nam:2006cx ; Nam:2009cv ; Nam:2010au , while the $N(2120)3/2^{-}$ exchange is found to be important to the cross sections of $\gamma p\to K^{+}\Lambda(1520)$ in Refs. Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb ; He:2012ud ; He:2014gga . In the present work, we found that, to get a satisfactory description of the data on both differential cross sections and photon-beam asymmetries of $\gamma p\to K^{+}\Lambda(1520)$, the exchange of at least one nucleon resonance in the $s$ channel needs to be introduced in constructing the reaction amplitudes. The required nucleon resonance could be either the $N(2060){5/2}^{-}$ or the $N(2120){3/2}^{-}$, both evaluated as three-star resonances in the most recent version of RPP Tanabashi:2018oca . In the fit with the $N(2060){5/2}^{-}$ resonance, the contributions of the resonance exchange are found to be rather important to the cross sections, and, in particular, they are responsible for the sharp rise of the cross sections near the $K^{+}\Lambda(1520)$ threshold, as can be seen in Fig. 7. In the fit with the $N(2120){3/2}^{-}$ resonance, although much smaller than those of the interaction current, the contributions of the resonance exchange are still considerable to the cross sections. In Refs. Nam:2005uq ; Nam:2006cx ; Nam:2009cv ; Nam:2010au ; Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb ; Yu:2017kng , the $t$-channel $K^{\ast}$ exchange is found to provide negligible contributions. In Refs. He:2012ud ; He:2014gga , it is reported that the contributions of the $t$-channel $K^{\ast}$ exchange are considerable to the cross sections. In our present work, the contributions of the $t$-channel $K^{\ast}$ exchange are negligible in the fit with the $N(2060){5/2}^{-}$ resonance, and are considerable in the fit with the $N(2120){3/2}^{-}$ resonance. As for the $u$-channel $\Lambda$ exchange, important contributions are reported in Refs. Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb ; He:2012ud ; He:2014gga , while in the present work, considerable contributions of this term are seen only in the fit with the $N(2120){3/2}^{-}$ resonance. Figure 8: Predictions of target nucleon asymmetries for $\gamma p\to K^{+}\Lambda(1520)$ from fit A (black solid lines) and fit B (blue dashed lines) at two selected center-of-mass energies. From Figs. 4$-$7 one sees that the fit with the $N(2060){5/2}^{-}$ resonance (fit A) and the fit with the $N(2120){3/2}^{-}$ resonance (fit B) describe the data on differential cross sections and photon-beam asymmetries for $\gamma p\to K^{+}\Lambda(1520)$ almost equally well. In Fig. 8, we show the predictions of the target nucleon asymmetries ($T$) from fit A (black solid lines) and fit B (blue dashed lines) at two selected center-of-mass energies. One sees that unlike the differential cross sections and the photon-beam asymmetries, the target nucleon asymmetries predicted by fit A and fit B are quite different. Future experimental data on target nucleon asymmetries are expected to be able to distinguish the fit A and fit B of the present work and to further clarify the resonance content, the resonance parameters, and the reaction mechanism for the $\gamma p\to K^{+}\Lambda(1520)$ reaction. ## IV Summary and conclusion The photoproduction reaction $\gamma p\to K^{+}\Lambda(1520)$ is of interest since the $K^{+}\Lambda(1520)$ has isospin $1/2$, excluding the contributions of the $\Delta$ resonances from the reaction mechanisms, and the threshold of $K^{+}\Lambda(1520)$ is at $2.01$ GeV, making this reaction more suitable than $\pi$ production reactions to study the nucleon resonances in a less-explored higher resonance mass region. Experimentally, the data for $\gamma p\to K^{+}\Lambda(1520)$ on differential cross sections, total cross sections, and photon-beam asymmetries are available from several experimental groups Boyarski:1970yc ; Barber:1980zv ; Kohri:2009xe ; Wieland:2010cq ; Moriya:2013hwg , with the photon-beam- asymmetry data coming from the LEPS Collaboration Kohri:2009xe and the most recent differential and total cross-section data coming from the CLAS Collaboration Moriya:2013hwg . Theoretically, the cross-section data for $\gamma p\to K^{+}\Lambda(1520)$ have been analyzed by several theoretical groups Nam:2005uq ; Nam:2006cx ; Nam:2009cv ; Nam:2010au ; Toki:2007ab ; Xie:2010yk ; Xie:2013mua ; Wang:2014jxb ; He:2012ud ; He:2014gga within effective Lagrangian approaches, and the photon-beam-asymmetry data Kohri:2009xe have been reproduced only in Ref. Yu:2017kng within a Reggeized framework. In the latter, the apparent structures of the angular distributions exhibited by the data are missing due to the lack of nucleon resonances in $s$-channel interactions in the Regge model. In these publications, the reported common feature for the $\gamma p\to K^{+}\Lambda(1520)$ reaction is that the contributions from the contact term and the $t$-channel $K$ exchange are important to the cross sections of this reaction. Nevertheless, the reaction mechanisms of $\gamma p\to K^{+}\Lambda(1520)$ claimed by different theoretical groups are quite different. In particular, there are no conclusive answers for the questions of whether the contributions from the $t$-channel $K^{\ast}$ exchange and $u$-channel $\Lambda$ exchange are significant or not, whether the introduction of nucleon resonances in the $s$ channel is inevitable or not for describing the data, and if yes, what resonance contents and parameters are needed in this reaction. In the present work, we performed a combined analysis of the data on both the differential cross sections and photon-beam asymmetries for $\gamma p\to K^{+}\Lambda(1520)$ within an effective Lagrangian approach. We considered the $t$-channel $K$ and $K^{\ast}$ exchange, the $u$-channel $\Lambda$ exchange, the $s$-channel nucleon and nucleon resonance exchanges, and the interaction current, with the last one being constructed in such a way that the full photoproduction amplitudes satisfy the generalized Ward-Takahashi identity and, thus, are fully gauge invariant. The strategy for introducing the nucleon resonances in the $s$ channel used in the present work was that we introduce nucleon resonances as few as possible to describe the data. For the first time, we achieved a satisfactory description of the data on both the differential cross sections and the photon-beam asymmetries for $\gamma p\to K^{+}\Lambda(1520)$. We found that either the $N(2060){5/2}^{-}$ or the $N(2120){3/2}^{-}$ resonance needs to be introduced in constructing the $s$-channel reaction amplitudes in order to get a simultaneous description of the data on differential cross sections and photon-beam asymmetries for $\gamma p\to K^{+}\Lambda(1520)$. In both cases, the contributions of the interaction current and the $t$-channel $K$ exchange are found to dominate the background contributions. The $s$-channel resonance exchange is found to be rather important in the fit with the $N(2060){5/2}^{-}$ resonance and to be much smaller but still considerable in the fit with the $N(2120){3/2}^{-}$ resonance. The contributions of the $t$-channel $K^{\ast}$ exchange and the $u$-channel $\Lambda$ exchange are negligible in the fit with the $N(2060){5/2}^{-}$ resonance and are significant in the fit with the $N(2120){3/2}^{-}$ resonance. The target nucleon asymmetries for $\gamma p\to K^{+}\Lambda(1520)$ are predicted, on which the future experimental data are expected to verify our theoretical models, to distinguish the two fits with either the $N(2060){5/2}^{-}$ or the $N(2120){3/2}^{-}$ resonance, and to further clarify the reaction mechanisms of the $K^{+}\Lambda(1520)$ photoproduction reaction. ###### Acknowledgements. This work is partially supported by the National Natural Science Foundation of China under Grants No. 11475181 and No. 11635009, the Fundamental Research Funds for the Central Universities, and the Key Research Program of Frontier Sciences of the Chinese Academy of Sciences under Grant No. Y7292610K1. ## References * (1) N. Isgur and G. Karl, Phys. Lett. 72B, 109 (1977). * (2) R. Koniuk and N. Isgur, Phys. Rev. D 21, 1868 (1980); 23, 818(E) (1981). * (3) R. G. Edwards, J. J. Dudek, D. G. Richards, and S. J. Wallace, Phys. Rev. D 84, 074508 (2011). * (4) R. G. Edwards et al. (Hadron Spectrum Collaboration), Phys. Rev. D 87, 054506 (2013). * (5) M. Tanabashi et al. (Particle Data Group), Phys. Rev. D 98, 030001 (2018). * (6) A. Boyarski, R. E. Diebold, S. D. Ecklund, G. E. Fischer, Y. Murata, B. Richter, and M. Sands, Phys. Lett. 34B, 547 (1971). * (7) D. P. Barber et al., Z. Phys. C 7, 17 (1980). * (8) H. Kohri et al. (LEPS Collaboration), Phys. Rev. Lett. 104, 172001 (2010). * (9) F. W. Wieland et al., Eur. Phys. J. A 47, 47 (2011); 47, 133(E) (2011). * (10) K. Moriya et al. (CLAS Collaboration), Phys. Rev. C 88, 045201 (2013); 88, A049902 (2013). * (11) S. I. Nam, A. Hosaka, and H. C. Kim, Phys. Rev. D 71, 114012 (2005). * (12) S. I. Nam, K. S. Choi, A. Hosaka, and H. C. Kim, Phys. Rev. D 75, 014027 (2007). * (13) S. I. Nam, Phys. Rev. C 81, 015201 (2010). * (14) S. I. Nam and C. W. Kao, Phys. Rev. C 81, 055206 (2010). * (15) H. Toki, C. Garcia-Recio, and J. Nieves, Phys. Rev. D 77, 034001 (2008). * (16) J. J. Xie and J. Nieves, Phys. Rev. C 82, 045205 (2010). * (17) J. J. Xie, E. Wang, and J. Nieves, Phys. Rev. C 89, 015203 (2014). * (18) E. Wang, J. J. Xie, and J. Nieves, Phys. Rev. C 90, 065203 (2014). * (19) J. He and X. R. Chen, Phys. Rev. C 86, 035204 (2012). * (20) J. He, Nucl. Phys. A927, 24 (2014). * (21) B. G. Yu and K. J. Kong, Phys. Rev. C 96, 025208 (2017). * (22) A. C. Wang, W. L. Wang, F. Huang, H. Haberzettl, and K. Nakayama, Phys. Rev. C 96, 035206 (2017). * (23) A. C. Wang, W. L. Wang, and F. Huang, Phys. Rev. C 98, 045209 (2018). * (24) A. C. Wang, W. L. Wang, and F. Huang, Phys. Rev. D 101, 074025 (2020). * (25) N. C. Wei, F. Huang, K. Nakayama, and D. M. Li, Phys. Rev. D 100, 114026 (2019). * (26) H. Haberzettl, Phys. Rec. C 56, 2041 (1997). * (27) H. Haberzettl, K. Nakayama, and S. Krewald, Phys. Rev. C 74, 045202 (2006). * (28) H. Haberzettl, F. Huang, and K. Nakayama, Phys. Rev. C 83, 065502 (2011). * (29) F. Huang, M. Döring, H. Haberzettl, J. Haidenbauer, C. Hanhart, S. Krewald, U.-G. Meißner, and K. Nakayama, Phys. Rev. C 85, 054003 (2012). * (30) F. Huang, H. Haberzettl, and K. Nakayama, Phys. Rev. C 87, 054004 (2013). * (31) N. C. Wei, A. C. Wang, F. Huang, and D. M. Li, Phys. Rev. C 101, 014003 (2020). * (32) Seung-il Nam, J. Phys. G 40, 115001 (2013). * (33) H. Garcilazo and E. Moya de Guerra, Nucl. Phys. A562, 521 (1993). * (34) A. C. Wang, F. Huang, W. L. Wang, and G. X. Peng, Phys. Rev. C 102, 015203 (2020). * (35) X. Y. Wang, J. He, and H. Haberzettl, Phys. Rev. C 93, 045204 (2016). * (36) X. Y. Wang and J. He, Phys. Rev. D 95, 094005 (2017). * (37) X. Y. Wang and J. He, Phys. Rev. D 96, 034017 (2017).
# Self-similar analysis of the time-dependent compressible and incompressible boundary layers including heat conduction Imre Ferenc Barna1, Krisztián Hriczó2, Gabriella Bognár2, and László Mátyás3 1 Wigner Research Center for Physics, Konkoly-Thege Miklós út 29 - 33, 1121 Budapest, Hungary 2 University of Miskolc, Miskolc-Egyetemváros 3515, Hungary, 3Department of Bioengineering, Faculty of Economics, Socio-Human Sciences and Engineering, Sapientia Hungarian University of Transylvania, Libertătii sq. 1, 530104 Miercurea Ciuc, Romania ###### Abstract We investigate the incompressible and compressible heat conducting boundary layer with applying the two-dimensional self-similar Ansatz. Analytic solutions can be found for the incompressible case which can be expressed with special functions. The parameter dependencies are studied and discussed in details. In the last part of our study we present the ordinary differential equation (ODE) system which is obtained for compressible boundary layers. ###### pacs: 47.10.-g,47.10.ab,47.10.ad ## I Introduction It is evident that the study of hydrodynamical equations has a crucial role in engineering and science as well. It is also clear that numerous classifications exist for various flow systems. One class of fluid flows is the field of boundary layer. The development of this scientific field started with the pioneering work of Prandtl prandt who used scaling arguments and derived that half of the terms of the Naiver-Stokes equations are negligible in boundary layer flows. In 1908 Blasius blasius gave the solutions of the steady-state incompressible two-dimensional laminar boundary layer equation forms on a semi-infinite plate which is held parallel to a constant unidirectional flow. Later Falkner and Skan falkner ; falkner1 generalized the solutions for steady two-dimensional laminar boundary layer that forms on a wedge, i.e. flows in which the plate is not parallel to the flow. An exhaustive description of the hydrodynamics of boundary layers can be found in the classical textbook of Schlichting sch recent applications in engineering is discussed by Hori hori . The mathematical properties of the corresponding partial differential equations (PDEs) attracted remarkable interest as well. Without completeness we mention some of the available mathematical results. Libby and Fox libby derived some solutions using perturbation method. Ma and Hui ma gave similarity solution to the boundary layer problems. Burde burde1 ; burde2 ; burde3 gave additional numerous explicit analytic solutions in the nineties. Weidman weid presented solutions for boundary layers with additional cross flows. Ludlow and coworkers lud evaluated and analyzed solutions with similarity methods as well. Vereshchagina ver investigated the spatial unsteady boundary layer equations with group fibering. Polyanin in his papers poy1 ; poy2 presents numerous independent solutions derived with various methods like general variable separation. Makinde mak investigated the laminar falling liquid film with variable viscosity along an inclined heated plate problem using perturbation technique together with a special type of Hermite – Padé approximation. In nanofluids the importance of buoyancy AnSa2016 , aspects on bioconvection MaAn2016 , and possible modified viscosity SaKoAn2016 are also discussed. One may find exact solutions for the oscillatory shear flow in SaGiKo2017 ; SaGi2018 . Bognár bogn applied the steady-state boundary layer flow equations for non- Newtonian fluids and presented self-similar results. Later it was generalized bognhri , and the steady-state heat conduction mechanism was included in the calculations as well. Certain parameters of the nanofluid can be tuned by varying the amount of nanoparticles in the fluid MaEbTe1993 ; Ch1995 ; Ng2007 . In our former studies we investigated three different kind of Rayleigh-Bénard heat conduction problems imre1 ; imre2 ; imre3 which are full two-dimensional viscous flows coupled to the heat conduction equation. We might say that the heated boundary layer equations - from the mathematical point of view - show some similarities to the Rayleigh-Bénard problem. These last five publications of us bogn ; bognhri ; imre1 ; imre2 ; imre3 led us to the decision that it would be worst examining heated boundary layers with the self-similar Ansatz. In the following we apply the Sedov type self-simiar Ansatz sedov ; zeldovich to the original partial differential equation (PDE) systems of incompressible and compressible boundary layers with heat conduction and reduce them to coupled non-linear ordinary differential equation (ODE) system. For the incompressible case the ODE system can be solved with quadrature giving analytic solutions for the velocity, pressure and temperature fields. Due, to our knowledge there are no self-similar solutions known and analyzed for any type of time-dependent boundary layer equations including heat conduction. ## II Theory ### II.1 The incompressible case We start with the PDE system of $\displaystyle\frac{\partial u}{\partial x}+\frac{\partial v}{\partial y}$ $\displaystyle=$ $\displaystyle 0,$ (1) $\displaystyle\frac{\partial p}{\partial y}$ $\displaystyle=$ $\displaystyle 0,$ (2) $\displaystyle\rho_{\infty}\frac{\partial u}{\partial t}+\rho_{\infty}\left(u\frac{\partial u}{\partial x}+v\frac{\partial u}{\partial y}\right)$ $\displaystyle=$ $\displaystyle\mu\frac{\partial^{2}u}{\partial y^{2}}-\frac{\partial p}{\partial x},$ (3) $\displaystyle\rho_{\infty}c_{p}\frac{\partial T}{\partial t}+\rho_{\infty}c_{p}\left(u\frac{\partial T}{\partial x}+v\frac{\partial T}{\partial y}\right)$ $\displaystyle=$ $\displaystyle\kappa\frac{\partial^{2}T}{\partial y^{2}},$ (4) where the dynamical variables are the two velocities components $u(x,y,t),v(x,y,t)$ of the fluid the pressure $p(x,y,t)$ and the temperature $T(x,y,t)$. The additional physical parameters are $\rho_{\infty},c_{p},\mu,\kappa,$ the fluid density at asymptotic distances and times, the heat capacity at fixed pressure, the kinematic viscosity and the thermal diffusivity, respectively. It is important to emphasize at this point, that this description for the heated boundary layer is valid for small velocities in laminar flow, only. More information can be found in the classical book of Schlichting sch ($8^{th}$ addition page 211). Outside the laminar flow regime a viscous heating term should be added to the final temperature equation with the form of $\mu(u_{y})^{2}$. (A similar analysis for that system is already in progress and will be the topic of our next distinct study.) There is no general fundamental theory for nonlinear PDEs, but over time, some intuitive methods have evolved, most of them can be derived from symmetry considerations. Numerous (almost arbitrary) functions can be constructed which couple the temporal and spatial variables to a new reduced variable from intuitive reasons. Our long term experience shows that two of them are superior to all others and have direct physical meanings. These are the traveling wave and the self-similar Ansätze. The first is more or less well known from the community of physicists and engineers and, has the form of $G(x,t)=f(x\mp ct)$ and we may call $\eta=x\mp ct$ as the new reduced variable, where c is the propagation speed of the corresponding wave. Here $G(x,t)$ is the investigated dynamical variable in the PDE. $G(x,t)$ could be any physically relevant property, like temperature, electric field or the like. This Ansatz can be applied to any kind of PDE and will mimic the general wave property of the investigated physical system. The second (and not so well known) is the self-similar Ansatz with the from of $G(\eta)=t^{-\alpha}f(x/t^{\beta})$. There $\alpha$ and $\beta$ are two free real parameters, it can be shown that this Ansatz automatically gives the Gaussian or fundamental solution of the diffusion (or heat conduction) equation. In general, and this is the key point here, this trial functions helps us to get a deeper insight into the dispersive and decaying behavior of the investigated physical system. This is the main reason why we use it in this form. Viscous fluid dynamic equations automatically fulfill this condition, therefore it is highly probable, that this Ansatz leads to physically rational solutions. It is easy to modify the original form of the Ansatz to two (or even three) spacial dimensions and generalize it to multiple dynamical variables, hereupon we apply the following form of: $\displaystyle u(x,y,t)$ $\displaystyle=$ $\displaystyle t^{-\alpha}f(\eta),\hskip 28.45274ptv(x,y,t)=t^{-\delta}g(\eta),$ $\displaystyle T(x,y,t)$ $\displaystyle=$ $\displaystyle t^{-\gamma}h(\eta),\hskip 28.45274ptp(x,y,t)=t^{-\epsilon}i(\eta),$ (5) with the new argument $\eta=\frac{x+y}{t^{\beta}}$ of the shape functions. (To avoid later physical interpretation problems of negative values we define temperature as a temperature difference relative to the average $T=\tilde{T}-T_{av}$.) All the exponents $\alpha,\beta,\gamma,\delta$ are real numbers. (Solutions with integer exponents are called self-similar solutions of the first kind, non-integer exponents generate self-similar solutions of the second kind.) It is important to emphasize that the obtained results fulfill well-defined initial and boundary problems of the original PDE system via fixing their integration constants of the derived ODE system. The shape functions $f,g,h$ and $i$ could be any continuous functions with existing first and second continuous derivatives and will be evaluated later on. The logic, the physical and geometrical interpretation of the Ansatz were exhaustively analyzed in all our former publications imre1 ; imre2 ; imre3 therefore we skip it here. The general scheme of the calculation, how the self-similar exponents can derived is given in imre4 in details. The main idea is the following: after having done the spatial and temporal derivatives of the Ansatz the obtained terms should be replaced into the original PDE system. Due to the derivations all terms pick up an extra time dependent factor like $t^{-\alpha-1}$ or $t^{-2\beta}$ because of the reduction mechanism the new variable of the shape functions is now $\eta$ therefore all kind of extra time dependences have to be canceled. Therefore all the exponents of the time dependences eg. $\alpha+1$ or $2\beta$ should cancel each other which dictates a relation among the self-similar variables. In our very first paper we gave all the details of this kind of a calculation for the non-compressible newtonian three dimensional Navier-Stokes equation imre4 . The main points are, that $\alpha,\delta,\gamma,\epsilon$ are responsible for the rate of decay and $\beta$ is for the rate of spreading of the corresponding dynamical variable for positive exponents. Negative self-similar exponents (except for some extreme cases) mean unphysical, exploding and contracting solutions. The numerical values of the exponents are now the following: $\alpha=\beta=\delta=1/2,\hskip 28.45274pt\epsilon=1,\hskip 28.45274pt\gamma=\textrm{arbitrary real number}.$ (6) Exponents with numerical values of one half mean the regular Fourier heat conduction (or Fick’s diffusion) process. One half values for the exponent of the velocity components and unit value exponent for the pressure decay are usual for the incompressible Navier-Stokes equation imre4 . The obtained ODE system reads $\displaystyle f^{\prime}+g^{\prime}$ $\displaystyle=$ $\displaystyle 0,$ (7) $\displaystyle i^{\prime}$ $\displaystyle=$ $\displaystyle 0,$ (8) $\displaystyle\rho_{\infty}\left(-\frac{f}{2}-\frac{f^{\prime}\eta}{2}\right)+\rho_{\infty}(ff^{\prime}+gf^{\prime})$ $\displaystyle=$ $\displaystyle\mu f^{\prime\prime}-i^{\prime},$ (9) $\displaystyle\rho_{\infty}c_{p}\left(-\gamma h-\frac{h^{\prime}\eta}{2}\right)+\rho_{\infty}c_{p}(fh^{\prime}+gh^{\prime})$ $\displaystyle=$ $\displaystyle\kappa h^{\prime\prime},$ (10) where prime means derivation in respect to the variable $\eta$. The first two equations are total derivatives and can be integrated directly yielding: $f+g=c_{1}$ and $i=c_{2}$. Having total derivatives in a dynamical systems automatically mean conserved quantities, (the first of them is now mass conservation). After some straightforward algebraic manipulation we arrive to a separate second order ODE for the velocity shape which is also a total derivative and can be integrated leading to: $\mu f^{\prime}+\rho_{\infty}f\left(\frac{\eta}{2}-c_{1}\right)-c_{2}=0,$ (11) with the analytic solution of $\displaystyle f=$ $\displaystyle\left(\frac{c_{2}\sqrt{\pi}e^{-\frac{\rho_{\infty}c_{1}^{2}}{\mu}}\cdot\emph{erf}\left[\frac{1}{2}\sqrt{-\frac{\rho_{\infty}}{\mu}}\eta+\frac{\rho_{\infty c_{1}}}{\sqrt{-\mu\rho_{\infty}}}\right]}{\sqrt{-\mu\rho_{\infty}}}+c3\right)\cdot\emph{e}^{\frac{\eta(-\eta+4c_{1})\rho_{\infty}}{4\mu}}$ (12) where erf means the usual error function NIST . Note, that for the positive real constants $\rho_{\infty},\mu$ the complex quantity $\sqrt{-\rho_{\infty}\mu}$ appears in the argument of the error functions and as a complex multiplicative prefactor simultaneously making the final result a pure real function. The second important thing is to note, that for the $c_{1}=c_{2}=0$ trivial integration constants the solution is simplified to the Gaussian function of $f=c_{4}e^{-\frac{\rho_{\infty\eta^{2}}}{4\mu}}.$ (13) This means that the velocity flow process shows similarity to the regular diffusion of heat conduction phenomena. Similar solutions (containing exponential and error functions) were found for the stationary velocity field by Weyburne in 2006 with probability distribution function methodology wey . Figure (1) shows the general velocity shape function (12) for various parameter sets. The choice of these parameters are arbitrary, we are not limited to real fluid parameters, however we try to create the most general and most informative figures, which mimic the general features of the solution function. The functions are the modification of the error function. The crucial parameter is the ratio $\rho_{\infty}/\mu$, if this is larger than unity then the function tends to a sharp Gaussian. Figure 1: The graphs of the velocity shape function $f(\eta)$ in Eq. (12) for three different parameter sets ($c_{1},c_{2},c_{3},\mu,\rho_{\infty}$). The solid, dashed and dotted lines are for $(1,0,1,4.1,0.9)$, $(2,-1,0.5,2.5,1)$ and $(2,2,0.3,10,1)$, respectively. Figure 2: The velocity distribution function $u(x,y=0,t)=\frac{1}{t^{1/2}}f(\eta)$ for the third parameter set presented on the previous figure. Figure (1) presents the velocity distribution function. Note, the very sharp peak in the origin and the extreme quick time decay along the time axis. There is a separate ODE for the temperature distribution as well $\frac{\kappa}{\rho_{\infty}c_{p}}h^{\prime\prime}-h^{\prime}\left(c_{1}-\frac{\eta}{2}\right)+\gamma h=0.$ (14) For the most general case (when $\gamma$ is an arbitrary real number,) and $c_{1}\neq 0$ the solutions of Eq. (14) can be expressed with the Kummer M and Kummer U functions NIST $h=c_{2}M\left(\gamma,\frac{1}{2};-\frac{c_{p}\rho_{\infty}[\eta-2c_{1}]^{2}}{4\kappa}\right)+c_{3}U\left(\gamma,\frac{1}{2};-\frac{c_{p}\rho_{\infty}[\eta-2c_{1}]^{2}}{4\kappa}\right).$ (15) M is regular in the origin and U is irregular, therefore we investigate only the properties of M which means $(c_{3}=0)$. The M and U functions form a complete orthogonal function system if the argument is linear. Now, the argument is quadratic, - in our former studies we found numerous such solutions, for incompressible imre4 or for compressible imre5 multidimensional Navier-Stokes or Euler equations – however, we still do not know the physical message of this property. It can be easily proven with the definition of the Kummer functions using the Pochhammer symbols NIST , that for negative integer $\gamma$ values our results can be expanded into finite order polynomials, which are divergent for large arguments $\eta$. For non-integer $\gamma<0$ values, we get infinite divergent polynomials as well. The most relevant parameter of the solutions is evidently $\gamma$. The integral constant $c_{1}$ just shifts the solutions parallel to the $x$ axis, $c_{2}$ scales the solutions, and $c_{p}\rho_{\infty}/\kappa$ parameter just scales the width of the solution. Figure (3) presents three different solutions for various positive $\gamma$ values. (All negative $\gamma$ values mean divergent shape functions for large $\eta$s which are unphysical and outside of our scope.) Note, larger $\gamma$s mean more oscillations. For a better understanding we present the projection of the total solution of the temperature field $T(t,x,y)$ on Figure (4) for the $y=0$ coordinates. Figure 3: The graphs of the temperature shape function Eq. (15) for three different parameter sets ($\gamma,c_{2},c_{3},c_{p},\rho_{\infty},\kappa$). The solid, dashed and dotted lines are for $(0.8,4,0,1,0.9,0.3)$, $(3.4,4,0,1,1,0.6)$ and $(6.3,4,0,1,3,10)$, respectively. Figure 4: The temperature distribution function $T(x,y=0,t)=\frac{1}{t^{1}}h(\eta)$ for the first parameter set presented on the previous figure. For some special values of $\gamma$ the temperature shape function can be expressed with other simpler special functions. For values of $\gamma=\pm\frac{1}{2}$ and $0$ the shape functions all contain the error function. Negative integer $\gamma$s result even order polynomials. (E.g. $\gamma=-1$ defines the shape function of $f=(c_{2}+c_{3})\cdot(2\kappa+c_{p}\rho_{\infty}[\eta-2c_{1}]^{2})$. ) Polynomials are divergent in infinity therefore are out of our physical interest. For the sake of completeness we present the solutions for the pressure as well. The ODE of the shape function is trivial with the solution of: $i^{\prime}=0,\hskip 28.45274pti=c_{4}.$ (16) Therefore, the final pressure distribution reads: $p(x,y,t)=t^{-\epsilon}\cdot i(x,y,t)=\frac{c_{4}}{t},$ (17) which means that the pressure is constant in the entire space at a given time point, but has a quicker time decay than the velocity field. ### II.2 The compressible case In the last part of our study we investigate the compressible boundary layer equations. The starting PDE system is now changed to the following: $\displaystyle\frac{\partial\rho}{\partial t}+\frac{\partial\rho}{\partial x}u+\rho\frac{\partial u}{\partial x}+\frac{\partial\rho}{\partial y}v+\rho\frac{\partial v}{\partial y}$ $\displaystyle=$ $\displaystyle 0,$ (18) $\displaystyle\frac{\partial p}{\partial y}$ $\displaystyle=$ $\displaystyle 0,$ (19) $\displaystyle\rho\frac{\partial u}{\partial t}+\rho\left(u\frac{\partial u}{\partial x}+v\frac{\partial u}{\partial y}\right)$ $\displaystyle=$ $\displaystyle\mu\frac{\partial^{2}u}{\partial y^{2}}-\frac{\partial p}{\partial x},$ (20) $\displaystyle c_{p}\rho\frac{\partial T}{\partial t}+c_{p}\rho\left(u\frac{\partial T}{\partial x}+v\frac{\partial T}{\partial y}\right)$ $\displaystyle=$ $\displaystyle k\frac{\partial^{2}T}{\partial y^{2}},$ (21) the notation of all the variables are the same as for the incompressible case. For closing constitutive equation (or with other name ”equation of state” (EOS)) we apply the ideal gas $p=R\rho T$ where $R$ is the universal gas constant. (Of course, there are numerous EOS available for physically relevant materials, and each gives us an additional new system to investigate, but that lies outside the scope of our present study.) For the dynamical variables we apply the next self-similar Ansatz of: $\displaystyle\rho(x,y,t)$ $\displaystyle=$ $\displaystyle t^{-\alpha}f(\eta),\hskip 28.45274ptu(x,y,t)=t^{-\gamma}g(\eta),$ (22) $\displaystyle v(x,y,t)$ $\displaystyle=$ $\displaystyle t^{-\delta}h(\eta),\hskip 28.45274ptT(x,y,t)=t^{-\epsilon}i(\eta),$ (23) with the usual new variable of $\eta=\frac{x+y}{t^{\beta}}$. To obtain a closed ODE system the following relations must held for the similarity exponents $\alpha=0,\hskip 28.45274pt\beta=\delta=\gamma=\epsilon=1/2.$ (24) Note, that now all the exponents have fixed numerical values. The $\alpha=0$ means two things, first the density as dynamical variable has no spreading property (just decay $\beta>0$), second, the first continuity ODE is not a total derivative and cannot be integrated directly. This system has an interesting peculiarity, our experience showed, that the incompressible Navier-Stokes (NS) equation imre4 has all fixed self-similar exponents and the compressible one imre5 has one free exponent. It is obvious that an extra free exponent makes the mathematical structure richer leaving more room to additional solutions. (As we mentioned above, self-similar exponents with the value of one half has a close connection to regular Fourier type heat conduction mechanism.) Parallel, the obtained ODE system reads $\displaystyle-\frac{1}{2}\eta f^{\prime}+fg^{\prime}+f^{\prime}g+f^{\prime}h+fh^{\prime}$ $\displaystyle=$ $\displaystyle 0,$ (25) $\displaystyle R(f^{\prime}i+fi^{\prime})$ $\displaystyle=$ $\displaystyle 0,$ (26) $\displaystyle f\left(-\frac{g}{2}-\frac{g^{\prime}\eta}{2}\right)+f(gg^{\prime}+gh^{\prime})$ $\displaystyle=$ $\displaystyle\mu g^{\prime\prime}-R(f^{\prime}i+fi^{\prime}),$ (27) $\displaystyle c_{p}f\left(-\frac{i}{2}-\frac{i^{\prime}\eta}{2}\right)+c_{p}f(gi^{\prime}+hi^{\prime})$ $\displaystyle=$ $\displaystyle\kappa i^{\prime\prime},$ (28) where prime means derivation in respect to $\eta$. Having done some non-trivial algebraic steps a decoupled ODE can be derived for the density field. First, the pressure equation (26) can be integrated, then $i(\eta)$ can be expressed, after the derivatives $i^{\prime}$ and $i^{\prime\prime}$ can be evaluated, then plugging it into (28) the $(g+h)$ quantity can be expressed with $f,f^{\prime}$ and $f^{\prime\prime}$. Finally, calculating the derivatives of $(f+g)$ and substituting them into (25) an independent ODE can be deduced for the density shape function. These algebraic manipulations are more compound and contain many more steps what we had in the past for various flow systems like imre3 ; imre4 . With the conditions $f(\eta)\neq 0$ and $f^{\prime}(\eta)\neq 0$, the next highly non-linear ODE can be derived $-\kappa f^{\prime}f^{2}f^{\prime\prime\prime}+f^{\prime\prime}\left(\kappa f^{2}f^{\prime\prime}+2\kappa ff^{\prime 2}+\frac{1}{2}c_{p}f^{4}\right)+f^{\prime 2}\left(-2\kappa f^{\prime 2}-c_{p}f^{\prime}f^{2}\cdot\eta-\frac{3}{2}c_{p}f^{3}\right)=0.$ (29) Such ODEs have no analytic solutions for any kind of parameter set (of course $\kappa\neq 0$ and $c_{p}\neq 0$). Therefore, pure numerical integration processes have to be applied. We have to mention, that an analogous fourth- order non-linear ODE was derived in the viscous heated Bénard system imre3 and was analyzed with numerical means. The shape function of the temperature field can be easily derived from (26) without any additional derivation $i=\frac{c_{1}}{Rf}.$ (30) We have to note two things here. First, the condition of $f\neq 0$ should hold. Second, the numerical value $c_{1}$ of the integration constant fixes the absolute magnitude of the temperature. The final physical field quantity which has to be determined is the velocity shape function and distribution. Note, that due to our original Ansatz the two velocity components cannot be determined separately from each other, only the $g+h$ is possible to evaluate. This can be easily done from (25) if we introduce the variable $L:=g+h$. Now the ODE is $L^{\prime}f+Lf^{\prime}-\frac{\eta f^{\prime}}{2}=0.$ (31) The formal solution now became trivial, namely $L=g+h=\frac{\int_{0}^{\eta}\omega f(\omega)d\omega+c_{2}}{2f(\eta)}.$ (32) This means that our Ansatz is not unique for the velocity field because the $x$ and $y$ coordinates are handled on the same footing. The in-depth numerical analysis of the density (29) and the velocity (32) shape functions lies outside the scope of the present study. Here, we just wanted to present that incompressible and compressible flow systems having initially comparable PDE systems, which describe similar processes, but behave completely differently during a self-similar analysis. Such derivations always give a glimpse into the deep mathematical layers of non-linear PDE systems. ## III Summary and Outlook We analyzed the incompressible and compressible time-dependent boundary flow equations with additional heat conduction mechanism with the self-similar Ansatz. Analytic solutions were derived for the incompressible flow. The velocity fields can be expressed with the error functions (in some special cases with Gaussian functions) and the temperature with the Kummer functions. The last one has the most complex mathematical structure including some oscillations. It is often asked what are analytic results are good for, we may say that our analytic solution could help to test complex numerical fluid dynamics program packages, new numerical routines endre or PDE solvers. For a $t=t_{0}$ starting time point the time propagation is exactly given by the analitic formula and can be compared to the results of any numerical scheme. In the second part of our treatise we investigated the compressible time- dependent boundary flow equations with additional heat conduction again with the self-similar Ansatz. For closing constitutive equation, the ideal gas EOS was used. It is impossible to derive analytic solutions for the dynamical variables from the coupled ODE system. However, highly non-linear independent ODEs exist for each dynamical variables which can be integrated numerically. An in-depth analysis could be the subject of a next publication. Work is in progress to apply our self-similar method to more realistic complex boundary layer flows containing viscous heating or other mechanisms. ## IV Authors Contributions The corresponding author (Imre Ferenc Barna) had the original idea of the study, performed all the calculations, created the figures and wrote large part of the manuscript. The second and third authors (Krisztián Hriczó and Gabriella Bognár) checked the written manuscript, improved the language of the final text and gave some general instructions. The third author (Gabriella Bognár) organized the financial support and the general founding. The last author (László Mátyás) checked the literature of the investigated scientific field, corrected the manuscript and had an everyday contact with the first author. ## V Acknowledgments One of us (I.F. Barna) was supported by the NKFIH, the Hungarian National Research Development and Innovation Office. This study was supported by project no. 129257 implemented with the support provided from the National Research, Development and Innovation Fund of Hungary, financed under the $K\\_18$ funding scheme. ## VI Conflicts of Interest The authors declare no conflict of interest. ## VII Data Availability The data that supports the findings of this study are all available within the article. ## References * (1) L. Prandtl, Verhandlungen 3. Int. Math. Kongr. Heidelberg 3, 484 (1904). * (2) H. Blasius, Z. Angew. Math. Phys. 56, 1 (1908). * (3) V. M. Falkner and S. W. Skan, Aero. Res. Coun. Rep. and Mem. No. 1314 (1930). * (4) V.M. Falkner and S.W. Skan, Phil. Mag., 12, 865 (1931). * (5) H. Schlichting and K. Gersten Boundary-Layer Theory, Springer, 2017. * (6) Y. Hori, Hydrodynamic Lubrication, Springer, 2006. * (7) P.A. Libby and Fox, J. Fluid Mech. 17, 3 (1963). * (8) P.K.H Ma and W.H. Hui, J. Fluid Mech. 216, 537 (1990). * (9) G.I. Burde, Quart. J. Mech.Appl. Math. 47, 247 (1994). * (10) G.I. Burde, Quart. J. Mech. Appl. Math. 48, 611 (1995). * (11) G.I. Burde, J. Physica A: Math. Gen. 29, 1665 (1996). * (12) P.D. Weidman, Z. Angew. Math. Phys. 48, 341 (1997). * (13) D.K. Ludlow, P.A. Clarkson and A.P. Bassom, Quart. J. Mech. and Appl. Math., 53, 175 (2000). * (14) L.I. Vereshchagina, [in Russian], Vestnik LGU, 13, 82 (1973). * (15) A.D. Polyanin, Theor. Found. Chem. Eng. 35, 319 (2001). * (16) P. D. Polyanin, DokladyPhysics 46, 526 (2001). * (17) O.D. Makinde, Applied Mathematics and Computation, 175, 80 (2006). * (18) I.L. Animasaun and N. Sandeep, Powder Technology 301, 858 (2016). * (19) O.D. Makinde and I.L. Animasaun, Journal of Molecular Liquids 221, 733 (2016). * (20) N. Sandeep, O.K. Koriko and I.L. Animasaun, Journal of Molecular Liquids 221, 1197 (2016). * (21) C. Saengov, A.J. Giacomin, and C. Kolitawong, Physics of Fluids 29, 043101 (2017). * (22) C. Saengov and A.J. Giacomin, Physics of Fluids 30, 030703 (2018). * (23) G. Bognár, Int. J. Nonlin. Science and Num. Simulation 10, 1555 (2009). * (24) G. Bognár and K. Hriczó, Acta Polytechnica Hungarica 8, 131 (2011). * (25) H. Masuda, A. Ebata, K. Teramae and N. Hishinuma, Netsu Bussei 7, 227 (1993). * (26) S.U.S. Choi, ASME Publ. 66, 99 (1995). * (27) C.T. Nguyen, H.A. Mintsa and G. Roy, Thermal Engineering and Environment 290, 25 (2007). * (28) I.F. Barna and L. Mátyás, Chaos Solitons and Fractals 78, 249 (2015). * (29) I.F. Barna, M.A. Pocsai, S. Lökös and L. Mátyás, Chaos Solitons and Fractals 103, 336 (2017). * (30) I.F. Barna, L. Mátyás and M.A. Pocsai, Fluid. Dyn. Res. 52, 015515 (2020). * (31) L. Sedov, Similarity and Dimensional Methods in Mechanics CRC Press, 1993. * (32) Ya. B. Zel’dovich and Yu. P. Raizer Physics of Shock Waves and High Temperature Hydrodynamic Phenomena Academic Press, New York, 1966. * (33) I.F. Barna, Commun. Theor. Phys. 56, 745 (2011). * (34) D. W. Weyburne, Applied Mathematics and Computation 175, 1675 (2006). * (35) F. W. J. Olver, D. W. Lozier, R. F. Boisvert and C. W. Clark NIST Handbook of Mathematical Functions Cambridge University Press, 2010. * (36) I.F. Barna and L. Mátyás, Fluid Dyn. Res. 46, 055508 (2014). * (37) E. Kovács, Numer. Methods Partial Differ., 37, 2469 (2020).
# Cross Chest Graph for Disease Diagnosis with Structural Relational Reasoning Gangming Zhao Baolian Qi Jinpeng Li Gangming Zhao and Baolian Qi contributed equally to this work.Gangming Zhao is with the Department of Computer Science, The University of Hong Kong, Hong Kong.Qibao Lian and Jinpeng Li are with University of Chinese Academy of Sciences, Beijing, China ###### Abstract Locating lesions is important in the computer-aided diagnosis of X-ray images. However, box-level annotation is time-consuming and laborious. How to locate lesions accurately with few, or even without careful annotations is an urgent problem. Although several works have approached this problem with weakly- supervised methods, the performance needs to be improved. One obstacle is that general weakly-supervised methods have failed to consider the characteristics of X-ray images, such as the highly-structural attribute. We therefore propose the Cross-chest Graph (CCG), which improves the performance of automatic lesion detection by imitating doctor’s training and decision-making process. CCG models the intra-image relationship between different anatomical areas by leveraging the structural information to simulate the doctor’s habit of observing different areas. Meanwhile, the relationship between any pair of images is modeled by a knowledge-reasoning module to simulate the doctor’s habit of comparing multiple images. We integrate intra-image and inter-image information into a unified end-to-end framework. Experimental results on the NIH Chest-14 database (112,120 frontal-view X-ray images with 14 diseases) demonstrate that the proposed method achieves state-of-the-art performance in weakly-supervised localization of lesions by absorbing professional knowledge in the medical field. ## 1 Introduction Chest radiographs are a type of medical images that can be conveniently acquired for disease diagnosis. With the rapid development of deep learning, automatic disease detection in chest X-ray images has become an important task in the computer-aided diagnosis. Figure 1: CCG network models the intra-image relationship between different anatomical areas by leveraging the structural information to simulate the doctor’s habit of observing different areas. Meanwhile, the relationship between any pair of images is modeled by a knowledge-reasoning module to simulate the doctor’s habit of comparing multiple images. Deep convolutional neural networks (DCNN) have been widely applied in many computer vision tasks, such as image classification [6, 18] , object detection [3, 5, 16, 15, 10] and semantic segmentation [12, 17]. To achieve good performance in these tasks, substantial images with careful annotations are needed. Encouraged by the success of DCNN in computer vision, some researches have directly applied DCNN models to analyze the medical images but cannot achieve the same performance as in the natural images. The reasons lie in two folds: 1. it is expensive to acquire accurate localization or classification labels in chest X-ray images. 2. there exists much professional knowledge in medical images that DCNN cannot exploit well. Therefore, how to exploit the professional knowledge into DCNN models for solving these two questions still opens a fully challenging problem. Our work transfers the knowledge into DCNN models to reduce the problem of shortage of carefully annotated images. Recent work paid much attention to utilize professional knowledge of chest X-ray images into DCNN frameworks. However, they just proposed a simple fused strategy to embed low-level information of chest X-ray into models, such as Liu et al. [9] utilized contrastive learning to provide more localization information with the help of healthy images. Zhao et al. [22] proposed to exploit the contralateral information of chest X-ray via a simple fusion module. These methods only exploit the apparent information of chest-Xray images. They all overlooked the inner structure information of chest X-rays. Therefore, they cannot apply their methods into real applications. In this paper, we propose a Cross Chest Graph Network (CCG-Net) as shown in Fig 1, which firstly utilizes deep expert knowledge to automatical detect disease in chest X-ray images. We have known that medical experts have much experience in finding out disease and how to treat patients. In fact, the actions of medical experts consist of two phases: training and decision-making processes. They pay much time to learn distinguish disease and embed their experience into the decision process. During the training process, experts would like to observe different areas and find out the relationship between any pair of images. Our CCG-Net aims to model the observation way by a knowledge-reasoning module to simulate the doctor’s habit of comparing multiple images. Then we integrate intra-image and inter-image information into a unified end-to-end framework. Inspired from the experience of medical experts, our proposed CCG-Net consists of four modules, 1. an end-to-end framework for deciding where and what is a disease, 2. a inter-image relation module, which formulates the training process of medical experts, to compare multiple images, 3. a intra-image knowledge learning module, which builds the local relation graph for different patches of chest X-ray images. Due to their highly structured property, every chest X-ray image can be divided into several patches, we build a patch-wise relation graph on them, 4. a knowledge reasoning module, which excavates the inner knowledge from cross-image structural features. The last three operations (2, 3, and 4) are similar to medical experts’ training process, which learn intra-image and inter-image information to gain professional knowledge. The first operation embeds the learned knowledge into DCNN frameworks leading to better disease diagnosis models. Above all, our contribution consists of three folds: * • We propose CCG-Net, which is the first to formulate the medical experts’ training process by building relation graphs in the intra-image and inter- image information of chest X-ray images. More generally, it provides inspiration to address medical vision tasks with much professional knowledge like in chest X-ray images. * • We divide the experts’ professional actions into two stages including training and decision-making processes. In addition, we utilize intra-image and inter- image relation to learn much professional knowledge that would be embedded in an end-to-end detection framework. * • We achieve state-of-the-art results on the localization of NIH ChestX-ray14. ## 2 Related Work ### 2.1 Disease Detection Object detection is one of the most important computer vision tasks, aiming to localize and classify. Due to their strong feature representation ability, DCNN achieved much progress in object detection tasks. For detection tasks, DCNN methods consist of two style framework: 1. two-stage models, such as RCNN series [16], 2. one-stage models, such as YOLO [15] and SSD [10]. However, for disease detection, because of the shortage in careful annotations, traditional detection framework cannot directly be applied in chest X-ray images. Besides, since there is much distortion caused by other chest X-ray tissues, such low contrast also causes the difficulty of disease finding. Weakly supervised object detection (WSOD) can be considered as an effective method to solve these problems. Based on CAM [23], researchers proposed many techniques to use only image-level labels to detect objects. Although there is no enough detection supervision, WSOD still achieved much progress. However, researchers still face a big challenge when it comes to disease detection in medical images. the existence of much professional knowledge greatly limits the development of the applications of DCNN in medical fields. Therefore, in this paper, we are inspired by the experts’ learning and decision processes to propose CCG-Net, which not only exploits a larger amount of knowledge in chest X-ray images but also builds a unified framework to detect disease in an end- to-end style. ### 2.2 Knowledge-based Disease Diagnosis Automatical disease diagnosis is a key problem in medical fields. However, due to the shortage of careful annotations and the existence of much professional knowledge, DCNN methods cannot achieve a good performance in medical tasks, especially such a tough problem: disease detection in chest X-ray images. To exploit medical knowledge and embed it into DCNN frameworks, researchers paid much effort to utilize medical experts’ experience for disease diagnosis. Wang et al. [20] firstly proposed a carefully annotated chest X-ray dataset and led to a series of work that focuses on using image-level labels to localize the disease. Li et al. [8] integrated classification and localization in a whole framework with two multi instance-level losses and performed better. Liu et al. [9] improved their work to propose contrastive learning of paired samples, which utilizes healthy images to provide more localization information for disease detection. Zhao et al. [22] proposed to utilize the symmetry information in a chest X-ray to improve the disease localization performance. Figure 2: The network consists of four modules: 1. an end-to-end framework for disease detection under weakly-supervised settings, 2. the inter-image relation module among different samples, 3. the intra-image knowledge learning module based on the thoracic spatial structure, 4. the knowledge reasoning module mining cross-image structural features. Our four modules are tightly related and can be easily integrated into an end-to-end framework. Besides, many works applied relation knowledge models to chest X-ray diagnosis. Ypsilantis et al. [21], Pesce et al. [14], and Guan et al. [4] proposed to build a relation attention model fusing DCNN models achieved much progress. Li et al. [7] proposed a knowledge-graph based on medical reports and images to determine the dependencies among chest X-ray images. Cheng et al. [11] also proposed a new total strongly supervised dataset for tuberculosis detection. However, they all overlooked the structural relation among chest X-ray images. In this paper, we propose to build a structural, relational graph for disease detection under weakly supervised scenarios in chest X-ray images. Specifically, we build the global and local graph in chest X-ray via three modules: 1. a inter-image relation module, 2. a intra-image knowledge learning module, 3. knowledge reasoning module. Furthermore, we integrate three modules into an end-to-end framework to jointly train our network. Our proposed three relational modules provide better supervision since we exploit the local structural knowledge and global relation among different samples. ## 3 Method ### 3.1 Overview Given the images $X=\\{x_{1},x_{2},...,x_{n}\\}$. Our proposed framework consists of four modules: * • The End to End Framework is to localize and classify the disease in chest X-ray images. In our paper, we utilize the same multi-instance level losses used in [9] and [8]. * • Inter-image Relation Module, which includes a learnable matrix $G\in R^{n\times n}$. We also use a contrast-constrained loss to share similar information of $X$ and exploit their contrasted structural knowledge. We build a cross-sample graph for them to exploit the dependencies among different samples. The graph $G\in R^{n\times n}$ is to build the inter-image relation among sampled samples, which is a learnable matrix, and every element is initialized by $\frac{1}{n}$. The element $g_{ij}$ of $G$, $i,j\in\\{1,2,...,n\\}$, represents the similarity wight of images $x_{i}$ and $x_{j}$. * • Intra-image Knowledge Learning, which firstly acquires patch-wise features of different images. Then the network can achieve a new image graph via building a structural knowledge-based module. We denote this graph as $G_{k}\in R^{n\times n}$. Assumed that the number of patches are $|p_{i}|$ and $|p_{j}|$ of images $x_{i}$ and $x_{j}$. The graph $G_{k}$ would be calculated on using the graph $G_{l}\in R^{|p_{i}|\times|p_{j}|}$, which learns the relationship between different paired patches of images. * • Knowledge Reasoning Module, which is based on cross image structural knowledge. When we get the whole structural information of different images, we will utilize it to reason the inner structural dependencies among different patches in different images. ### 3.2 End to End Framework The end to end framework is to localize and classify the disease in chest X-ray images in a coarse-grained style. More specifically, the input images $X=\\{x_{1},x_{2},...,x_{n}\\}$ of the module are resized to $512\times 512$. ResNet-50 pre-trained from the ImageNet dataset is adopted as the backbone for this module. We use the feature map $F$ after C5 (last convolutional output of 5th-stage), which is 32 times down-sampled from the input image, and of size $2048\times 16\times 16$. Each grid in the feature map denotes the existent probability of disease. We pass $F$ through two $1\times 1$ convolutional layers and a sigmoid layer to obtain the class-aware feature map P of size $C\times H\times W$, where $C$ is the number of classes. Then we follow the paradigm used in [9], computing losses and making predictions in each channel for the corresponding class. For images with box-level annotations, if the grid in the feature map overlaps with the projected ground truth box, we assign label 1 to the grid. Otherwise, we assign 0 to it. Therefore, we use the binary cross-entropy loss as used in [9] for each grid: $L^{k}_{i}(\emph{P})=\sum_{j}-y_{ij}^{k}\log(p_{ij}^{k})-\sum_{j}(1-y_{ij}^{k})\log(1-p_{ij}^{k})$ (1) where $k$, $i$, and $j$ are the index of classes, samples, and grids respectively. $y^{k}_{ij}$ denotes the target label of the grid and $p^{k}_{ij}$ denotes the predicted probability of the grid. For images with only image-level annotations, we use the MIL loss used in [8]. $\begin{split}L^{k}_{i}(\emph{P})=-&y^{k}_{i}\log(1-\prod_{j}(1-p^{k}_{ij}))\\\ -&(1-y^{k}_{i})\log(\prod_{j}(1-p^{k}_{ij}))\end{split}$ (2) where $y^{k}_{i}$ denotes the target label of the image. For this end to end framework, the whole loss $L_{base}$ as shown in Fig. 2, is formulated as follows. $\begin{split}L_{base}=\sum_{i}\sum_{k}\lambda^{k}_{i}\beta_{B}L^{k}_{i}(\emph{P})+(1-\lambda^{k}_{i})L^{k}_{i}(\emph{P})\end{split}$ (3) where $\lambda^{k}_{i}\in{0,1}$ denotes if the $k_{th}$ class in the $i_{th}$ sample has box annotation, and $\beta_{B}$ is the balance weight of the two losses and is set to 4. ### 3.3 Inter-image Relation Module Inter-image relation is formulated as a learnable matrix $G\in R^{n\times n}$. A contrast-constrained loss is used to share similar information of $X$ and exploit their contrasted structural knowledge, as following equation. $\begin{split}L_{IR}=\frac{\sum_{(u,v)\in G}G(u,v)D(F_{u},F_{v})}{n\times n}\end{split}$ (4) $D(\cdot)$ is the distance metric function, where it is a Euclidean distance. $F_{u}$ and $F_{v}$ means the feature map after C5 of the image $x_{u}$ and $x_{v}$. We build a cross-sample graph for them to exploit the dependencies among different samples. The graph $G\in R^{n\times n}$ is to build the inter- image relation among sampled samples, which is a learnable matrix, and every element is initialized by $\frac{1}{n}$. The element $g_{ij}$ of $G$, $i,j\in\\{1,2,...,n\\}$, represents the similarity wight of images $x_{i}$ and $x_{j}$. G is adaptively adjusted during training processes and changes with diverse inputs to exploit the relationship fully. ### 3.4 Intra-image Knowledge Learning Intra-image Knowledge Learning, which firstly utilizes Simple linear iterative clustering (SLIC) [1], a super-pixel method to generate the patches for different images. Assumed that the patches of the image $x_{i}$ is $p_{i}={p^{i}_{1},p^{i}_{2},...,p^{i}_{m}}$. Then the network can achieve a new image graph via building a structural knowledge-based module with the help of $p_{i},i\in{1,2,...,n}$. We denote the graph as $G_{k}\in R^{n\times n}$, which is the intra-image graph between paired images $x_{i}$ and $x_{j}$. The graph $G_{k}$ is calculated on using the graph $G_{l}\in R^{|p_{i}|\times|p_{j}|}$, which learns the dependencies among different paired patches of images. Then the same contrast-constrained loss using this graph to provide more structural knowledge for the whole framework. $\begin{split}L_{IK}=\frac{\sum_{(u,v)\in G_{k}}G_{k}(u,v)D(F_{u},F_{v})}{n\times n}\end{split}$ (5) $\begin{split}G_{k}=W_{l}(G_{l})\end{split}$ (6) Where, $W_{l}$ is a fully connected layer and $\begin{split}G_{l}(l,p)=D^{{}^{\prime}}(H_{l},H^{{}^{\prime}}_{p}),l\in{1,2,...,|p_{i}|},p\in{1,2,...,|p_{j}|}\end{split}$ (7) $H_{l}$ is the hash code [19] of the patch $p^{i}_{l}$ in the image $x_{i}$ and $H^{{}^{\prime}}_{p}$ is the hash code of the patch $p^{j}_{p}$ in the image $x_{j}$. $D^{{}^{\prime}}(\cdot)$ is the Hamming distance. ### 3.5 Knowledge Reasoning Module In addition to previous efforts to focus on information in a whole image, we also explored the value of cross-image semantic relations in the medical object. The correlations between patches across images are emphasized, especially, the correlations between corresponding patches in two images. Knowledge Reasoning Module focuses on the correlations of two images. After getting the feature map $F_{u}$ and $F_{v}$ of the images, the affinity matrix $P$ is firstly calculated between $F_{u}$ and $F_{v}$. $P=F^{\mathrm{T}}_{u}W_{P}F_{v}\in\mathbb{R}^{HW\times HW}$ where the feature map $F_{u}\in\mathbb{R}^{C\times HW}$ and $F_{v}\in\mathbb{R}^{C\times HW}$ are flattened into matrix formats, and $W_{P}\in\mathbb{R}^{C\times C}$ is a learnable matrix. The affinity matrix $P$ represents the similarity of all pairs of patches in $F_{u}$ and $F_{v}$. Then $P$ is normalized column-wise to get the attention map of $F_{u}$ for each patch in $F_{v}$ and row-wise to get the attention map of $F_{v}$ for each patch in $F_{u}$. $F^{{}^{\prime}}_{u}=F_{u}softmax(P)\in\mathbb{R}^{C\times HW}$ $F^{{}^{\prime}}_{v}=F_{v}softmax(P^{\mathrm{T}})\in\mathbb{R}^{C\times HW}$ where $softmax(P)$ and $softmax(P^{\mathrm{T}})$ pay attention to the similar patches of the feature map $F_{u}$ and $F_{v}$ respectively. Therefore, they can be used to enhance $F_{u}$ and $F_{v}$ respectively, so that similar patches in $F_{u}$ and $F_{v}$ are highlighted. The cross-image method can extract more contextual information between images than using a single image. This module exploits the context of other related images to improve the reasoning ability of the feature map, which is beneficial to the localization and classification of disease in chest X-ray images. Furthermore, we exploit the enhanced feature map to calculate the new similarity between the paired images to gain a more strong supervisor. $\begin{split}L_{KR}=\frac{\sum_{(u,v)\in G_{k}^{{}^{\prime}}}G_{k}^{{}^{\prime}}(u,v)D(F^{{}^{\prime}}_{u},F^{{}^{\prime}}_{v})}{n\times n}\end{split}$ (8) T (IoU) | Models | Atelectasis | Cardiomegaly | Effusion | Infiltration | Mass | Nodule | Pneumonia | Pneumothorax | Mean ---|---|---|---|---|---|---|---|---|---|--- 0.3 | X, Wang [20] | 0.24 | 0.46 | 0.30 | 0.28 | 0.15 | 0.04 | 0.17 | 0.13 | 0.22 Z, Li [8] | 0.36 | 0.94 | 0.56 | 0.66 | 0.45 | 0.17 | 0.39 | 0.44 | 0.49 J, Liu [9] | 0.53 | 0.88 | 0.57 | 0.73 | 0.48 | 0.10 | 0.49 | 0.40 | 0.53 | Ours | 0.44 | 0.86 | 0.68 | 0.84 | 0.47 | 0.29 | 0.67 | 0.40 | 0.60 0.5 | X, Wang [20] | 0.05 | 0.18 | 0.11 | 0.07 | 0.01 | 0.01 | 0.03 | 0.03 | 0.06 Z, Li [8] | 0.14 | 0.84 | 0.22 | 0.30 | 0.22 | 0.07 | 0.17 | 0.19 | 0.27 J, Liu [9] | 0.32 | 0.78 | 0.40 | 0.61 | 0.33 | 0.05 | 0.37 | 0.23 | 0.39 | Ours | 0.27 | 0.86 | 0.48 | 0.72 | 0.53 | 0.14 | 0.58 | 0.35 | 0.49 0.7 | X, Wang [20] | 0.01 | 0.03 | 0.02 | 0.00 | 0.00 | 0.00 | 0.01 | 0.02 | 0.01 Z, Li [8] | 0.04 | 0.52 | 0.07 | 0.09 | 0.11 | 0.01 | 0.05 | 0.05 | 0.12 J, Liu [9] | 0.18 | 0.70 | 0.28 | 0.41 | 0.27 | 0.04 | 0.25 | 0.18 | 0.29 | Ours | 0.20 | 0.86 | 0.48 | 0.68 | 0.32 | 0.14 | 0.54 | 0.30 | 0.44 Table 1: The comparison results of disease localization among the models using 50% unannotated images and 80% annotated images. For each disease, the best results are bolded. T (IoU) | Models | Atelectasis | Cardiomegaly | Effusion | Infiltration | Mass | Nodule | Pneumonia | Pneumothorax | Mean ---|---|---|---|---|---|---|---|---|---|--- 0.1 | Z, Li [8] | 0.59 | 0.81 | 0.72 | 0.84 | 0.68 | 0.28 | 0.22 | 0.37 | 0.57 J, Liu [9] | 0.39 | 0.90 | 0.65 | 0.85 | 0.69 | 0.38 | 0.30 | 0.39 | 0.60 Ours | 0.66 | 0.88 | 0.79 | 0.85 | 0.69 | 0.28 | 0.40 | 0.47 | 0.63 0.3 | J, Liu [9] | 0.34 | 0.71 | 0.39 | 0.65 | 0.48 | 0.09 | 0.16 | 0.20 | 0.38 Baseline | 0.36 | 0.69 | 0.35 | 0.64 | 0.44 | 0.08 | 0.02 | 0.23 | 0.35 Ours | 0.31 | 0.79 | 0.37 | 0.75 | 0.40 | 0.06 | 0.24 | 0.27 | 0.40 0.5 | J, Liu [9] | 0.19 | 0.53 | 0.19 | 0.47 | 0.33 | 0.03 | 0.08 | 0.11 | 0.24 Baseline | 0.18 | 0.51 | 0.14 | 0.47 | 0.27 | 0.03 | 0.01 | 0.12 | 0.22 Ours | 0.19 | 0.71 | 0.14 | 0.52 | 0.31 | 0.08 | 0.05 | 0.13 | 0.27 0.7 | J, Liu [9] | 0.08 | 0.30 | 0.09 | 0.25 | 0.19 | 0.01 | 0.04 | 0.07 | 0.13 Baseline | 0.11 | 0.34 | 0.06 | 0.32 | 0.20 | 0.01 | 0.00 | 0.06 | 0.14 Ours | 0.06 | 0.64 | 0.08 | 0.38 | 0.19 | 0.01 | 0.08 | 0.09 | 0.19 Table 2: The comparison results of disease localization among the models using 100% unannotated images and no any annotated images. For each disease, the best results are bolded. T (IoU) | Models | Atelectasis | Cardiomegaly | Effusion | Infiltration | Mass | Nodule | Pneumonia | Pneumothorax | Mean ---|---|---|---|---|---|---|---|---|---|--- 0.3 | J, Liu [9] | 0.55 | 0.73 | 0.55 | 0.76 | 0.48 | 0.22 | 0.39 | 0.30 | 0.50 Baseline | 0.47 | 0.84 | 0.65 | 0.82 | 0.33 | 0.04 | 0.57 | 0.29 | 0.50 Ours | 0.49 | 0.87 | 0.66 | 0.88 | 0.48 | 0.10 | 0.51 | 0.20 | 0.52 0.5 | J, Liu [9] | 0.36 | 0.57 | 0.37 | 0.62 | 0.34 | 0.13 | 0.23 | 0.17 | 0.35 Baseline | 0.27 | 0.76 | 0.39 | 0.58 | 0.24 | 0.02 | 0.39 | 0.21 | 0.36 Ours | 0.26 | 0.80 | 0.41 | 0.67 | 0.15 | 0.06 | 0.42 | 0.18 | 0.37 0.7 | J, Liu [9] | 0.19 | 0.47 | 0.20 | 0.41 | 0.22 | 0.06 | 0.12 | 0.11 | 0.22 Baseline | 0.14 | 0.62 | 0.20 | 0.42 | 0.07 | 0.00 | 0.23 | 0.08 | 0.22 Ours | 0.18 | 0.71 | 0.20 | 0.50 | 0.20 | 0.02 | 0.29 | 0.06 | 0.27 Table 3: The comparison results of disease localization among the models using 100% unannotated images and 40% annotated images. For each disease, the best results are bolded. Figure 3: Visualization of the predicted results on both the baseline model and our method. The first column shows the original images, the second and third columns show baseline and our method. The green bounding box and red area mean the the ground truth and prediction. The graph $G_{k}^{{}^{\prime}}$ is calculated on using the graph $G^{{}^{\prime}}_{l}\in R^{|p_{i}|\times|p_{j}|}$. $\begin{split}G_{k}^{{}^{\prime}}=W^{{}^{\prime}}_{l}(G^{{}^{\prime}}_{l})\end{split}$ (9) where $W^{{}^{\prime}}_{l}$ is a fully connected layer and $\begin{split}G_{l}^{{}^{\prime}}(l,p)=&D^{{}^{\prime}}(P_{l},P_{p}),\\\ &l\in\\{1,2,...,|p_{i}|\\},p\in\\{1,2,...,|p_{j}|\\}\end{split}$ (10) $P_{l}$ is the $l$-th feature patch of $F^{{}^{\prime}}_{u}$ and $P_{p}$ is the $p$-th feature patch of $F^{{}^{\prime}}_{v}$, respectively. ### 3.6 Training Loss The overall loss function during the training is a weighted combination of four loss functions, $\displaystyle L_{all}=w_{1}L_{base}+w_{2}L_{IR}+w_{3}L_{IK}+w_{4}L_{KR}$ (11) where $\sum^{4}_{i=1}w_{i}=1$. In our experiments, we always set $w_{i}=0.25,i\in{1,2,..,4}$. ### 3.7 Training and Test Figure 4: Visualization of the generated heatmap and ground truth of our method, where the green bounding box means the ground truth. Data | Models | Atelectasis | Cardiomegaly | Effusion | Infiltration | Mass | Nodule | Pneumonia | Pneumothorax | Mean ---|---|---|---|---|---|---|---|---|---|--- 0.5_0.8 | X, Wang [20] | 0.01 | 0.03 | 0.02 | 0.00 | 0.00 | 0.00 | 0.01 | 0.02 | 0.01 Z, Li [8] | 0.04 | 0.52 | 0.07 | 0.09 | 0.11 | 0.01 | 0.05 | 0.05 | 0.12 J, Liu [9] | 0.18 | 0.70 | 0.28 | 0.41 | 0.27 | 0.04 | 0.25 | 0.18 | 0.29 Baseline | 0.34 | 1.00 | 0.40 | 0.68 | 0.11 | 0.14 | 0.65 | 0.00 | 0.41 IK | 0.22 | 0.82 | 0.36 | 0.56 | 0.32 | 0.14 | 0.25 | 0.35 | 0.38 IR | 0.24 | 0.82 | 0.40 | 0.56 | 0.32 | 0.07 | 0.38 | 0.30 | 0.39 KR | 0.24 | 0.89 | 0.32 | 0.68 | 0.26 | 0.14 | 0.21 | 0.30 | 0.38 (IR+IK) | 0.20 | 0.86 | 0.48 | 0.68 | 0.32 | 0.14 | 0.54 | 0.30 | 0.44 IR+IK+KR | 0.27 | 0.86 | 0.40 | 0.56 | 0.37 | 0.14 | 0.13 | 0.30 | 0.38 1.0_0.0 | J, Liu [9] | 0.08 | 0.30 | 0.09 | 0.25 | 0.19 | 0.01 | 0.04 | 0.07 | 0.13 Baseline | 0.11 | 0.34 | 0.06 | 0.32 | 0.20 | 0.01 | 0.00 | 0.06 | 0.14 IK | 0.10 | 0.59 | 0.07 | 0.37 | 0.20 | 0.00 | 0.13 | 0.06 | 0.19 GR | 0.06 | 0.61 | 0.07 | 0.28 | 0.14 | 0.00 | 0.05 | 0.08 | 0.16 IK | 0.09 | 0.63 | 0.06 | 0.36 | 0.22 | 0.00 | 0.09 | 0.07 | 0.19 IR+IK | 0.06 | 0.64 | 0.08 | 0.38 | 0.19 | 0.01 | 0.08 | 0.09 | 0.19 IR+IK+KR | 0.12 | 0.51 | 0.07 | 0.36 | 0.22 | 0.03 | 0.02 | 0.07 | 0.17 1.0_0.4 | J, Liu [9] | 0.19 | 0.47 | 0.20 | 0.41 | 0.22 | 0.06 | 0.12 | 0.11 | 0.22 Baseline | 0.14 | 0.62 | 0.20 | 0.42 | 0.07 | 0.00 | 0.23 | 0.08 | 0.22 IK | 0.14 | 0.66 | 0.09 | 0.47 | 0.15 | 0.00 | 0.30 | 0.06 | 0.23 GR | 0.14 | 0.75 | 0.24 | 0.42 | 0.11 | 0.00 | 0.26 | 0.12 | 0.25 KR | 0.13 | 0.68 | 0.20 | 0.47 | 0.19 | 0.06 | 0.17 | 0.08 | 0.25 IR+IK | 0.13 | 0.72 | 0.13 | 0.43 | 0.20 | 0.00 | 0.23 | 0.06 | 0.24 IR+IK+KR | 0.18 | 0.71 | 0.20 | 0.50 | 0.20 | 0.02 | 0.29 | 0.06 | 0.27 Table 4: The comparison results of disease localization among the models using three sets of data at T(IoU)=0.7, including 50% unannotated and 80% annotated images (0.5_0.8), 100% unannotated and no any annotated images (1.0_0.0), and 100% unannotated and 40% unannotated images (1.0_0.4). For each disease, the best results are bolded. Training All the models are trained on NIH chest X-ray dataset using the SGD algorithm with the Nesterov momentum. With a total of 9 epochs, the learning rate starts from 0.001 and decreases by 10 times after every 4 epochs. Additionally, the weight decay and the momentum is 0.0001 and 0.9, respectively. All the weights are initialized by pre-trained ResNet [6] models on ImageNet [2]. The mini batch size is set to 2 with the NVIDIA 1080Ti GPU. All models proposed in this paper are implemented based on PyTorch [13]. Testing We also use the threshold of 0.5 to distinguish positive grids from negative grids in the class-wise feature map as described in [8] and [9]. All test setting is same as [9], we also up-sampled the feature map before two last fully convolutional layers to gain a more accurate localization result. ## 4 Experiments ### 4.1 Dataset and Evaluation Metrics Dataset. NIH chest X-ray dataset [20] include 112,120 frontal-view X-ray images of 14 classes of diseases. There are different diseases in each image. Furthermore, the dataset contains 880 images with 984 labeled bounding boxes. We follow the terms in [8] and [9] to call 880 images as “annotated” and the remaining 111,240 images as “unannotated”. Following the setting in [9], we also resize the original 3-channel images from resolution of $1024\times 1024$ to $512\times 512$ without any data augmentation techniques. Evaluation Metrics. We follow the metrics used in [8]. The localization accuracy is calculated by the IoU (Intersection over Union) between predictions and ground truths. Since it is a coarse-grained task, our localization predictions are discrete small rectangles. The eight diseases with ground truth boxes is reported in our paper. The localization result is regarded as correct when $IoU>T(IoU)$, where T(*) is the threshold. ### 4.2 Comparison with the State-of-the-art In order to evaluate the effectiveness of our models for weakly supervised disease detection, we design the experiments on three sets of data and conduct a 5-fold cross-validation. In the first experiment, we use the 50% unannotated images and 80% annotated images for training, and test the models with the remaining 20% annotated images. In the second experiment, we use the 100% unannotated images and no any annotated images for training, and test the models with all annotated images. In the third experiment, we use the 100% unannotated images and 40% annotated images for training, and test the models with remaining 60% annotated images. Additionally, our experimental results are mainly compared with four methods. The first method is X, Wang [20], which proposes a carefully annotated chest X-ray dataset and a unified weakly supervised multi-label image classification and disease localization framework. The second method is Z, Li [8], which uses fully convolutional neural network to localize and classify the disease in chest X-ray images. The third method is J, Liu [9], which proposes contrastive learning of paired samples to provide more localization information for disease detection. The last method is our baseline model, which is a unified end-to-end framework that doesn’t use our approach to locate and classify the disease. In the first experiment, we compare the localization results of our model with [20], [8] and [9]. We can observe that our model outperforms existing methods in most cases, as shown in Table 1. Particularly, with the increase of T(IoU), our model has greater advantages over the reference models. For example, when T(IoU) is 0.3, the mean accuracy of our model is 0.60, and outperforms [20], [8] and [9] by 0.38, 0.11 and 0.07 respectively. However, when T(IoU) is 0.7, the mean accuracy of our model is 0.44, and outperforms [20], [8] and [9] by 0.43, 0.32 and 0.15 respectively. Overall, the experimental results shown in Table 1 demonstrate that our method is more accurate for disease localization and classification, which provides a great role for clinical practices. In the second experiment, we train our model without any annotated images comparing the first experiment. Since [8] only provides the results when T(IoU) = 0.1, in order to better show the performance of our model, we add an evaluation method of T(IoU) = 0.1. It can be seen that our model outperforms [8] and [9] in most cases, as shown in Table 2. For example, when T(IoU) is 0.1, the mean accuracy of our models is 0.63, which is 0.06 higher than [8], and 0.03 higher than [9]. Furthermore, when T(IoU) is 0.7, the mean localization result of our model is 0.19, which is 0.06 higher than [8] and 0.05 higher than [9]. Compared with the baseline model, our approach performs better in most classes except for “Atelectasis” and “Nodule”. The trend stays the same that at higher T(IoU), our approach demonstrates more advantages over baseline methods. The added unannotated training samples contribute more than the removed annotated ones in those classes, which implies that our approach can better utilize the unannotated samples. The overall results show that even without annotated data used for training, our approach can achieve decent localization results. In the third experiment, we use more annotated images comparing the second experiment. We compare the localization results of our model with [9] in same data setting. It can be seen that our model outperforms [9] in most cases, as shown in Table 3. With T(IoU) = 0.3 and 0.7, our model outperforms [9] by 0.02 and 0.05 respectively. Similar improvements are achieved comparing the second experiment. Overall, the experimental results demonstrate that our method can improve the performance of models with limited annotated images. To better demonstrate the final effect of our approach on disease localization and classification, we visualize some of typical predictions of both the baseline model and our method, as shown in Figure 3. The first column shows the original images, the second and third columns show baseline model and our method. The green bounding box and red area mean the ground truth and prediction. It can be seen that our models can predict more accurate in most cases comparing the baseline model. For example, the class “Atelectasis” and “Nodule”, the localization reslut of the baseline model is completely inconsistent with the ground truth, but the localization reslut of our method is consistent with the ground truth. It shows that using the structural information of intra-image and inter-image can improve the performance of automatic lesion detection. Additionally, we also visualize the generated heatmap and ground truth of our model, as shown in Figure 4. It can be seen that the proposed method can effectively locate and classify medical images. ### 4.3 Ablation Studies In this section, we explore the influence of different modules on our method for ablation studies. To evaluate our method more comprehensively, we build 6 models, including the model of the end to end framework (Baseline), the model with the intra-image knowledge learning (IK), the model with the inter-image relation module (IR), the model with the knowledge reasoning (KR), the model combining the inter-image relation module and the intra-image knowledge learning (IR+IK), the model combining the inter-image relation module, the intra-image knowledge learning and the knowledge reasoning module (IR+IK+KR). Table 4 shows the results of the three experiments mentioned in section 4.2 at T(IOU)=0.7. It can be seen that our method performs better in most classes except for “Atelectasis”, “Effusion” and “Mass” comparing [20], [8] and [9]. Furthermore, comparing the baseline model, it can be observed that the performance of our other models are improved in most cases, which shows that our method is effective for improving model performance. However, a model does not always maintain the advantage in the three experiments, for example, the model (IR+IK) achieves the best performance in the data (0.5_0.8), the model (IK), the model (KR) and the model (IR+IK) achieve the best performance in the data (1.0_0.0), and the model (IR+IK+KR) achieves the best performance in the data (1.0_0.4). Overall, the experimental results demonstrate that using structural relational information can improve the performance of models. For different experimental data, our models can achieve different results. It is difficult for us to determine which model is the best, but we can be sure that our method is effective, because no matter what kind of data we use, our models achieve great improvement. Particularly, the method can achieve good localization results even without any annotation images for training. ## 5 Conclusion By imitating doctor’s training and decision-making process, we propose the Cross-chest Graph (CCG) to improve the performance of automatic lesion detection under limited supervision. CCG models the intra-image relationship between different anatomical areas by leveraging the structural information to simulate the doctor’s habit of observing different areas. Meanwhile, the relationship between any pair of images is modeled by a knowledge-reasoning module to simulate the doctor’s habit of comparing multiple images. We integrate intra-image and inter-image information into a unified end-to-end framework. Experimental results on the NIH Chest-14 dataset demonstrate that the proposed method achieves state-of-the-art performance in diverse situations. ## References * [1] Radhakrishna Achanta, Appu Shaji, Kevin Smith, Aurelien Lucchi, Pascal Fua, and Sabine Süsstrunk. Slic superpixels. Technical report, 2010. * [2] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009. * [3] Ross Girshick. Fast r-cnn. In Proceedings of the IEEE international conference on computer vision, pages 1440–1448, 2015. * [4] Qingji Guan, Yaping Huang, Zhun Zhong, Zhedong Zheng, Liang Zheng, and Yi Yang. Diagnose like a radiologist: Attention guided convolutional neural network for thorax disease classification. arXiv preprint arXiv:1801.09927, 2018. * [5] K. He, X. Zhang, S. Ren, and J. Sun. Spatial pyramid poolling in deep convolutional networks for visual recognition. In European Conference of Computer Vision (ECCV), 2014. * [6] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016. * [7] Christy Y Li, Xiaodan Liang, Zhiting Hu, and Eric P Xing. Knowledge-driven encode, retrieve, paraphrase for medical image report generation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 6666–6673, 2019. * [8] Zhe Li, Chong Wang, Mei Han, Yuan Xue, Wei Wei, Li-Jia Li, and Li Fei-Fei. Thoracic disease identification and localization with limited supervision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 8290–8299, 2018. * [9] J. Liu, G. Zhao, Y. Fei, M. Zhang, Y. Wang, and Y. Yu. Align, attend and locate: Chest x-ray diagnosis via contrast induced attention network with limited supervision. In 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pages 10631–10640, 2019. * [10] Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott Reed, Cheng-Yang Fu, and Alexander C Berg. Ssd: Single shot multibox detector. In European conference on computer vision, pages 21–37. Springer, 2016. * [11] Yun Liu, Yu-Huan Wu, Yunfeng Ban, Huifang Wang, and Ming-Ming Cheng. Rethinking computer-aided tuberculosis diagnosis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2646–2655, 2020. * [12] Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3431–3440, 2015. * [13] Adam Paszke, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang, Zachary DeVito, Zeming Lin, Alban Desmaison, Luca Antiga, and Adam Lerer. Automatic differentiation in pytorch. 2017\. * [14] Emanuele Pesce, Petros-Pavlos Ypsilantis, Samuel Withey, Robert Bakewell, Vicky Goh, and Giovanni Montana. Learning to detect chest radiographs containing lung nodules using visual attention networks. arXiv preprint arXiv:1712.00996, 2017. * [15] Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi. You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 779–788, 2016. * [16] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster r-cnn: Towards real-time object detection with region proposal networks. In C. Cortes, N. D. Lawrence, D. D. Lee, M. Sugiyama, and R. Garnett, editors, Advances in Neural Information Processing Systems 28, pages 91–99. Curran Associates, Inc., 2015. * [17] Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical image computing and computer-assisted intervention, pages 234–241. Springer, 2015. * [18] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, et al. Imagenet large scale visual recognition challenge. International journal of computer vision, 115(3):211–252, 2015\. * [19] R. Venkatesan, S. M. Koon, M. H. Jakubowski, and P. Moulin. Robust image hashing. In International Conference on Image Processing, 2000. * [20] Xiaosong Wang, Yifan Peng, Le Lu, Zhiyong Lu, Mohammadhadi Bagheri, and Ronald M Summers. Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2097–2106, 2017. * [21] Petros-Pavlos Ypsilantis and Giovanni Montana. Learning what to look in chest x-rays with a recurrent visual attention model. arXiv preprint arXiv:1701.06452, 2017. * [22] Gangming Zhao, Chaowei Fang, Guanbin Li, Licheng Jiao, and Yizhou Yu. Contralaterally enhanced networks for thoracic disease detection, 2020\. * [23] Bolei Zhou, Aditya Khosla, Agata Lapedriza, Aude Oliva, and Antonio Torralba. Learning deep features for discriminative localization. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2921–2929, 2016.
# Everything You Wanted to Know About Noninvasive Glucose Measurement and Control Prateek Jain Amit M. Joshi Saraju P. Mohanty Dept. of ECE Dept. of ECE Computer Science and Engineering MNIT, Jaipur, India MNIT, Jaipur, India University of North Texas, USA<EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS> ###### Abstract Diabetes is a chronicle disease where the body of a human is irregular to dissolve the blood glucose properly. The diabetes is due to lack of insulin in human body. The continuous monitoring of blood glucose is main important aspect for health care. Most of the successful glucose monitoring devices is based on methodology of pricking of blood. However, such kind of approach may not be advisable for frequent measurement. The paper presents the extensive review of glucose measurement techniques. The paper covers various non- invasive glucose methods and its control with smart healthcare technology. To fulfill the imperatives for non-invasive blood glucose monitoring system, there is a need to configure an accurate measurement device. Noninvasive glucose-level monitoring device for clinical test overcomes the problem of frequent pricking for blood samples. There is requirement to develop the Internet-Medical-Things (IoMT) integrated Healthcare Cyber-Physical System (H-CPS) based Smart Healthcare framework for glucose measurement with purpose of continuous health monitoring. The paper also covers selective consumer products along with selected state of art glucose measurement approaches. The paper has also listed several challenges and open problems for glucose measurement. ###### Index Terms: Smart Healthcare, Internet-of-Medical-Things (IoMT), Healthcare Cyber-Physical System (H-CPS), Diabetes, Glucose measurement, Non invasive measurement, Spectroscopy and calibration ## I Introduction The glucose is considered as important source of energy for the human body. The body requires blood glucose of normal range (80 to 150 mg/dl) in order to perform the daily activities [1]. However, the higher or lower value of glucose would lead to various complication inside the body. At the same time, insulin is also crucial hormone generated inside the body from the food intake. The glucose is produced from the food digestion which enters the blood cell to supply the energy and also helps in the growth. In case, the insulin is not properly generated then blood would accumulate the high glucose concentration. Fig. 1 illustrates the closed-loop of glucose generation and consumption in human body [2]. A consistently high blood glucose concentration is possible if the generation of $\alpha$ cells is larger as compared to that of the $\beta$ cells. Because of this condition, enough insulin is not secreted in the body for glucose consumption. This condition refers to as the Diabetes Mellitus. Diabetes is termed as chronic disease which defines high blood glucose levels inside the human body. The unbalanced glycemic profile is main reason for the cause of diabetic condition. The rate of prevalence for Non Communicable Diseases (NCD)/Chronic Disease has increased with many fold from last several years. There are around 20 million death reported yearly through cardiovascular disease, for which high blood glucose is significant predisposing factors. Moreover, people with diabetes are more affected during the viral pandemic outbreaks [3, 4, 5]. Figure 1: Illustration of the closed loop form of glucose generation and consumption [2]. There has been exponential growth of diabetes patients over past few years because of obesity, unhealthy diet plan, old-age population, and inactive lifestyle. Diabetes is considered as one of the fastest growing health challenges, with the number of adults living with diabetes having more than tripled over the past 2 decades (Refer Fig. 2) [6]. The prevalence of diabetes around the world was 9.3% during 2019 with approximate 463 million people. It is expected to rise to 578 million by 2030 with 10.2% prevalence rate and the same would be 10.9% with 700 million population by 2045. It has been observed that prevalence is quite higher in urban to 10.8% whereas 7.2% in rural region. Almost half of the diabetes patients unaware about their situation due to lack of knowledge. The diabetes has indeed global outbreak which has affected presently almost 1 in 10 people around the world. It is projected that more than 0.5 billion adults would suffer from the diabetes in the next decade [7]. As per the report from International Diabetes Federation (IDF), the death from diabetes has large number than combined death from Malaria (0.6mio), HIV/AIDS (1.5mio) and tuberculosis (1.5mio) [8]. There are around 8 million new patients are being added to diabetic community every year. This has grown the demands immensely for the effective diabetic management. It is important to monitor the blood glucose over time to time for avoiding late- stage complication from diabetes. This has necessitate the design of various reliable and robust solutions for efficient diabetes management. The market of diabetes devices has also grown rapidly with significant requisite for frequent glucose measurement for better glycemic profile control. Figure 2: Global trend of Diabetes, Adopted from [6]. Diabetes is one of the major chronic disease which has long-term impact of the well-being life of a person. Diabetes Mellitus (DM) is considered as physiological dysfunctions with high blood glucose because of insufficient insulin, insulin resistance, or excess generation of glucagon [9]. It is the critical health issue of $21^{st}$ century. Type 2 Diabetes (T2DM) has shown rapid growth around the world from past few years. Any form of diabetes may lead to complications in various body parts which increase the possibility of premature death. The higher value of blood glucose known as hyperglycemia, would lead to thickening of blood vessels which could resulted in kidneys damage and loss of sight and some times even to these organs failure. Diabetes is also associated with limb amputation, peripheral vascular diseases and myocardial. Contrary, the low blood glucose defined as hypoglycemia may occur in Type 1 Diabetes Patients (T1DM) for excessive insulin dosage [10]. The most common symptoms for hypoglycemia pateints are dizziness, sweating and fatigue and in the worst case it can lead to coma and death. The diabetic patients would have several common symptoms such as thirsty, tiredness, changes in vision, consistently hungriness, unexpected weight loss and the excretion of urine within short durations [11]. If the diabetes remain untreated over the period of time, it may cause blindness, heart stroke, kidney disease, lower limb amputation and blindness. It would lead to increase the probability of death almost 50% higher in comparison of the patients without diabetes. The diabetes also brings the additional financial burden for the treatment and point of care. The diabetic patients could also result in loss of productivity at workplace and may lead to disability. There are several health issues which may also arise from diabetes like depression, digestive problem, anxiety disorders, mood disorder and eating habits change. The diabetes could be controlled with proper diet plan, through some physical exercise, insulin dosage and medicines. The early stage of diabetes is possible to control with oral medicines. The diabetes control also helps to reduce the associated risk of high blood pressure, cardiovascular and amputation. The rest of the article is organized in the following manner: Section II briefly presents different types of diabetes while making case for the need of glucose level monitoring. Section III presents overview of various types of glucose-level measurement mechanisms. Section IV provides details of available approaches for noninvasive glucose-level monitoring. Section V has discussions on various post-processing and calibration techniques for noninvasive glucose- level monitoring. Section VI briefly discusses various consumer products for noninvasive glucose level measurement. Section VII presents the approaches for glucose-level control and corresponding consumer products. Section VIII provides the Internet-of-Medical-Things (IoMT) perspectives of glucose level measurements and control in healthcare Cyber-Physical Systems (H-CPS) that makes smart healthcare possible. Section IX outlines the shortcomings and open problems of glucose-level measurements and control. Section X summarizes the learning of this comprehensive review work. ## II The Health Issue of Diabetes and Need for Glucose-Level Measurement This Section presents details of different types of diabetes, the health issues arise due to diabetes, while making case for the need of glucose level monitoring. ### II-A Types of Diabetes The diabetes occurs because of insufficient insulin with respect to glucose generated inside the body. The insulin from body is either insufficient or not any which is generated from beta cells of the pancreas. In case of diabetes, the cells of liver, muscles and fat unable to balance glucose insulin effectively. The diabetes are classified mainly in three categories: Type 1 diabetes, Type 2 diabetes and gestational diabetes (Refer Fig. 3) [12]. Figure 3: Different types of diabetes and their symptoms. For diabetes of type-1, the pancreas does not produce insulin inside the body which is resulted in a weak immune system. This results in a person who is unable to generate insulin naturally [2, 13]. In case of type 2 diabetes, the amount of insulin from pancreas is not sufficient to maintain glycemic profile of the body. Gestational diabetes usually occurs in a pregnant woman at later stage of the delivery. in the year 2020, total 2 billion adults around the globe suffers from overweight, and 300 million of them are obese. In addition, a minimum of 155 million children in the world is overweight or obese. It is projected that the prevalence of hyperglycemia is 8.0% and expected to increase to 10% by 2025 [7]. There has been concern for diabetic people specially in developing countries due to increase in Type 2 Diabetes cases rapidly at earlier age which have overweight children even before puberty. Whereas for developed countries, most of people have high blood glucose at age around 60 years. Most frequently affected are at middle aged between 35 and 64 in developed countries [6]. In 2019, 69.2 millions population in India had Type-2 diabetes. Approximately 2.35 million adults have Type-1 diabetes. In general, there are around 5% adults have been considered for Type-1 diabetic patients while the others 90-95% are of Type-2 diabetic patients. Type-1 diabetic patient must have insulin to control the blood glucose level. Type-2 diabetic patients can control their glucose level by following an optimized diet with medication and a regular physical exercise schedule. ### II-B The Health Crisis due to Diabetes The diabetes mainly occurs due to unbalanced glucose insulin level of the body where insulin is demolished and muscles and cells are not able to generate insulin properly [14, 13]. The probability of death would also increase upto 50% in comparison to non-diabetes case. The control action of the diabetes would be possible using proper precautionary measure after frequent glucose measurements. Therefore, there is a real need for smart healthcare solution which would provide instant self measurement of blood glucose with high accuracy. Figure 4: Diseases in human body due to diabetes Hyperglycemia is the major issue which has been considered by several health organizations at worldwide level [15, 16]. There are several attempts which have been used for glucose measurement [17]. There have been substitutional work using various techniques to make the device more familiar with clinicians and patients [18]. Diabetes is possible in the age group 18 to 80 years usually [19, 20]. The normal range of glucose is in the range of 70-150 mg/dL and pathophysiological would be from 40 mg/dL to 550 mg/dL [21]. One of the emerging issues is to design the glucose measurement device for continuous health care monitoring [22]. The devices for monitoring the glucose level are available for last two decades [23]. ### II-C Glucose Measurement: A Brief History The glucose meter (aka glucometer) is a portable medical device for predicting the glucose level concentration in the blood [24, 25]. It may also be a strip based dipped into any substance and determined the glucose profile. It is a prime device for blood glucose measurement by people with diabetes mellitus or hypoglycemia. With the objective of glucose monitoring device advancement, the concept of the biosensor has been proposed earlier in 1962 by Lyons and Clark from Cincinnati. Clark is known as the “father of biosensors”, and modern-day glucose sensor which is used daily by millions of diabetics. This glucose biosensor had been composed with an inner oxygen semipermeable membrane, a thin layer of GOx, an outer dialysis membrane and an oxygen electrode. Enzymes could be gravitated at an electrochemical detector to form an enzyme electrode [26]. However, the main disadvantage of first-generation glucose biosensors was that there was the requirement of high operation potential of hydrogen peroxide amperometric measurement for high selectivity. The first-generation glucose biosensors were replaced by mediated glucose biosensors (second- generation glucose sensors). The proposed biosensors till present scenario represent the advancements in terms of portability of device and precision in measurement. But, due to some environmental and measurement limitations; these biosensors were not taken for real-time diagnosis. The history of glucose measurement is shown in Fig. 5 [27]. Figure 5: History of Glucose Measurement. Figure 6: Invasive versus Noninavsive Glucose Measurement. ### II-D Glucose Measurement Technique Presently, the glucose monitoring is carried out either laboratory based technique or home based monitoring. These both approaches are invasive in nature which provides discomfort by blood pricking and it only helps to measure the glucose measurement at that point of time. It is also not very convenient for the user to take out blood samples multiple times in a day and many patients are reluctant to opt such type of solution. Therefore, significant changes of glycemic profile may go unnoticed because of unanticipated side effects and low compliance from the patients. This could impact on improper insulin dosage and unknown food ingredient. However, they are reliable solution due to their good sensitivity and higher accuracy for glucose measurement [28, 29]. The novel approach for glucose measurement has been explored from past several years which is based on the principle of physical detection than conventional chemical based principle. Such non-invasive based method does not require the blood sample but uses the interstitial fluid (ISF) for glucose molecule detection. There are several attempts in the same direction for glucose measurement through sweat, saliva, tears and skin surface [30]. However, the main challenge is to have precise measurement, good sensitivity and reliability from such measurement. Such approach could be suitable for Continuous Glucose Measurement (CGM) and self monitoring purpose. Such CGM techniques would provide the frequent measurement in a day which would helpful for better glucose control and also for the necessary preventive actions for hyperglycemia and hypoglycemia patients. Such kind of techniques would also support for the dietician and healthcare provider to prepare proper diet plan according to glucose fluctuation for the patient. ### II-E The Need for Continuous Glucose Measurement (CGM) The measurement of glucose could be done through non-invasive, semi (or minimal) invasive and invasive approach. The frequent measurement may not be possible using invasive method which can cause trauma. The semi-invasive and non-invasive could be useful for Continuous Glucose Measurement (CGM) without any pricking of the blood. However, the non-invasive glucose measurement is most suitable technique which helps to measure the blood glucose painlessly [31]. CGM assist to have proper blood glucose level analysis at each prandial mode. It helps to measure glucose insulin level after insulin secretion,hysical exercise or subsequent to medication. The frequent glucose reading also helpful to endocrinologist for providing the proper prescription. It mainly helps for type 1 diabetic patients to take care of their insulin dosage over the period of time. The proper diet management could be possible with help of recurrent glucose monitoring and flow diagram of CGM is shown as Fig. 7 [28, 29]. The CGM is useful for the patients for frequent glucose measurement over the period of time. This would helpful to identify the average blood glucose value for the last 90 days, by which glycated haemoglobin (HbA1c) can be determined. Figure 7: The objectives of continuous glucose monitoring. ## III Approaches for Glucose-Level Measurement: A Broad Overview This Section discusses an overview of various types of glucose-level measurement mechanisms. In the past, many works has done for the glucose measurement. They can be invasive, non-invasive, or minimally invasive. A lot of works has been completed based on the non-invasive technique. They are technically based on optical and non-optical methods. Some of the optical techniques used methods based on Raman Spectroscopy, NIR spectroscopy, and PPG method. A taxonomy of the different methods is provided in Fig. 8 [25, 28, 29, 32]. Figure 8: An overview of the Glucose Measurement Options [32, 28, 29, 25]. ### III-A Invasive Methods Many commercial continuous blood glucose measurement devices use cost- effective electrochemical sensors [33]. They are available to respond quickly for glucose detection in blood [34]. Lancets (for pricking the blood) is used at the primary stage for blood glucose monitoring for various commercial devices available in the market [35]. The frequent measurement through the process is so much panic due to picking the blood sample from the fingertip more than 3-4 times in a day for frequent monitoring[36]. The low invasive biosensor for glucose monitoring has been developed with glucose oxidase that require around 1mm penetration inside the skin for measurement [37]. The technique of photometric was attempted to detect glucose with help of small blood volumes [38]. ### III-B Minimally Invasive Methods The minimally invasive method using prototype sensor was developed to have frequent monitoring of glucose tissue [39]. The sensor is wearable and is implanted on membrane which contains the immobilized glucose oxidase. The glucose monitoring through implantable devices were developed [40]. The semi or minimal invasive method using biosensors designed for diabetes patient [41]. The wearable micro system explored for frequent measurement of glucose [42]. Similarly, there was an attempt of continuous glucose monitoring with help of microfabricated biosensor through transponder chip [43]. The signal coming out of transponder chip was used for the calibration for semi invasive approach of Dexcom sensor [44]. The diabetes control explored by glucose sensor with artificial pancreas system [45]. The minimal invasive approaches have limitations mainly accuracy and may have shorter life span for monitoring. This is a wearable microsystem for the continuous monitoring of the blood glucose. It’s a minimally invasive method for the glucose monitoring. The main idea behind this is that it uses micro-actuator which consists the shape memory alloy (SMA) for the extraction of the blood sample from human skin [46]. An upgraded version of SMA is used for the implementation of PCB. Because of it’s feasibility and performance, it can be considered as the first wearable device for the glucose monitoring but it is large in size which makes it inconvenient. ### III-C Non-invasive Methods Non-invasive measurement would mitigate all the previous issues and would provide painless and accurate solutions [47, 48]. The non-invasive glucose measurement solution for smart healthcare had developed through portable measurement [31]. A lot of approaches have been introduced for glucose measurement [49]. The non-invasive measurement are more convenient for continuous glucose measurement in comparison to invasive method and semi invasive [47], [48]. The glucose measurement with help of optical method has observed more reliable and precise in the literature [50]. The popular optical methods include non-invasive measurement such as Raman spectroscopy, near infer-red spectroscopy,polarimetric,scattering spectroscopy [51], photoacoustic spectroscopy [52] etc. For the development of a non-invasive measurement device, it is considered by the researcher that the device would be much convenient for the user’s perspective [53, 54]. Calibration of the blood glucose to interstitial glucose dynamics have been considered for the accuracy of continuous glucose monitoring system [55, 56]. Several calibration algorithms have been developed and implemented for portable setup [57]. There has been several concious efforts towards the development of the self- monitoring system [58]. Figure 9: NIR Spectroscopy Mechanism of Serum Glucose Measurement. ### III-D Invasive Versus Non-invasive Glucose Measurements: The Trade-Offs Recent glucose measurement methods for the ever-increasing the diabetic patients over the world are invasive, time-consuming, painful and a bunch of the disposable items which constantly burden for the household budget. The non-invasive glucose measurement technique overcomes such limitations, for which this has become significantly researched era. Although, there is tradeoff between these two methods which is represented in Fig. 10. Figure 10: Representation of Tradeoffs between Invasive and Non-invasive Glucose Measurement. ### III-E Capillary Glucose versus Serum Glucose for Noninvasive Measurement The serum glucose value is precise which is always close to actual blood glucose measurement with compare to capillary glucose level. Traditional approaches able to measure capillary glucose instantly but the serum glucose measurement identification is difficult. It is observed that the glucose level of capillary is always higher than serum glucose. The accurate measurement of blood glucose would help for appropriate control actions. Therefore, it is really important to measure the serum glucose than the capillary glucose which is more reliable for medication. Capillary blood glucose measurement has been used widely than serum glucose estimation for medication purpose. The serum glucose is not possible for continuous glucose measurement or frequent measurement for diabetes. The blood glucose is controlled in much better way if one can measure serum glucose at regular interval. Laboratory analysis of glycosylated haemoglobin (HbA1c) which provides 6-8 weeks blood glucose measurement is also being done through the serum blood only. For the non- invasive measurement point of view, serum and capillary glucose are being measured through the optical spectroscopy. The mechanism of blood glucose measurement is based on received IR light after absorptions and scattering from glucose molecules which flow in blood vessels. The methodology is quite similar for both types of glucose measurement except the post-processing computation models which are necessary for blood glucose estimation. ### III-F Non-invasive Method for Glucose Level Estimation by Saliva As the most convenient method to estimate glucose level is via saliva [59] and is used for children and adults. This saliva has specific type of parts which can be defined as: (1) gland-specific saliva and (2) whole saliva. The collection of the Gland-specific saliva is done by individual glands like parotid, Sub mandibular, sublingual, and minor salivary glands. This diagnosis is done by the history of the patient in terms of associated risk factors, family history, age, sex, duration of diabetes,and any associated illness. Other Glucose measuring methods consist of measurement using photo-metric glucometers requiring very small sample volumes [60]. The basic approach is based on the reaction of the chemical test strip that reacts with sample. Measurement is done by capturing the reflections of the test area and then glucose level is estimated. It requires validation in large number of patients. ## IV Approaches for Noninvasive Glucose-Level Measurement This Section presents detailed discussions of various available approaches for noninvasive glucose-level monitoring. There have been several efforts for noninvasive glucose measurement using optical techniques [27, 29, 61, 62]. These techniques are mainly based on various spectroscopy based methods. For the development of a non-invasive measurement device, it is considered by the researcher that the device would be much convenient for the users perspective. Fig. 11 presents summary of various types of noninvasive glucose measurement techniques, whereas their comparative perspectives are presented in Fig. 12. A qualitative comparative perspective of various noninvasive methods is summarized in Table IV. Figure 11: Various spectroscopy techniques for noninvasive glucose measurement. Figure 12: Comparative Perspective of Various popular spectroscopy techniques for noninvasive glucose measurement. Table I: Qualitative comparison of various noninvasive glucose-level monitoring methods. Technique | Advantages | Disadvantages ---|---|--- Near Infra-Red (NIR) | • The signal intensity is directly proportional to glucose molecule • The glucose detection concept would work with other interfacing substance such as plastic or glass | • The glucose signal weak comparatively so complex machine learning model is required for interpretation • High scattering level Mid Infra-Red (MIR) | • The glucose molecule absorption stronger • Low scattering | • The light has limited penetration with tissue • Noise is present in the signal so water and other non-glucose metabolites would be detected. Far Infra-Red (FIR)/Thermal emission spectroscopy | • Frequent Calibration is not required • Least sensitive towards scattering | • The radiation intensity depends on temprature and substance thickness • Strong absorption with water so it is difficult to have precise glucose measurement Raman Spectroscopy | • Less sensitivity towards temperature and water • High specificity | • Requirement of the laser radiation source hence it can dangerous cell for CGM • Susceptible towards noise interference so low SNR Photo acoustic | • Simple and compact sensor design • Optical radiation will not harmful for the tissue | • Signal is vulnerable towards acoustic noise, temperature,motion etc. • It carries some noise from some non-glucose blood components Polarimetry | • The laser intensity variation will not change much the glucose prediction | • Requirement external laser source and requires proper alignment with eye • sensitive for the change in PH and temperature Reverse Iontophoresis | • Based on simple enzyme based electrode system • Highly accurate as it measure glucose from interstitial fluid | • Difficult to have proper calibration • Not so user-friendly approach due to passing of the current through skin Fluorescence | • Highly sensitive for glucose molecule detection due to immune for light scattering • Good sensitivity because of distinctive optical properties | • Very much sensitive for local pH and/or oxygen, • Suffers from foreign body reaction Bio impedance spectroscopy | • Comparatively less extensive • Easy for CGM | • Prone towards sweating, motion and temperature • Require large calibration period Millimetre and Microwave sensing | • Deep penetration depth for precise glucose measurement • No risk for ionization | • Poor selectivity • Very much sensitive for physiological parameters such as sweating, breathing and cardiac activity Optical Coherence Tomography | • High resolution and good SNR • Not vulnerable for blood pressure and cardiac activity | • Glucose value may change as per skin and motion • Suffers from tissue inhomogeneity Surface Plasma Resonance | • Small glucose molecule can be detected due to high sensitivity | • Long calibration process and size is bulky • Glucose value changes with variation in temperature,sweat and motion Time of flight and THz Time domain Spectroscopy | • strong absorption and dispersion for glucose molecule | • Lesser depth resolution and longer time for measurement Metabolic Heat Conformation | • Uses the concept of well-known various physiological parameters for glucose prediction | • Sensitive towards variation in temperature and sweat Electromagnetic sensing | • low-cost and can be easily miniaturized • No risk of ionization | • Lack of selectivity due to dielectric constant is mainly affected with other blood components • More sensitive for the slight change of temperature Ultrasound Technology | • Well established technology with not much harm to tissue cell • Long penetration below the skin or tissue | • Limited accuracy with ultrasound only hence mostly used with NIR as multi-model • costly technology for measurement and not useful for CGM Sonophoresis | • Favourable technology as there is no side-effect to skin • Based on well known enzymatic method | • Error prone due to environmental parameters ### IV-A Near-Infrared (NIR) Spectroscopy It is well known as Infrared spectroscopy (IR spectroscopy) or vibration spectroscopy where radiation of infrared type are incident on the matter [63, 64]. Various types of IR spectroscopy is shown in Fig. 13. In general, IR spectroscopy includes reflection, scattering and absorption spectroscopy [65]. The wave from IR absorption cause the molecular vibration and generate the spectrum band with wavelength number in $cm^{-1}$ [66]. Figure 13: Classification of vibrational spectroscopy [67]. In this case, the light in the wavelength range of 700nm to 2500nm for Near- infrared region is applied at the object (may be finger or ear lobe) [68]. The light may interact with blood components and it may scattered, absorbed and reflected [69, 70]. The intensity of received light varies as per glucose concentration as per Beer-Lambert law [71, 72] The receiver would help to measure the presence glucose molecule from the blood vessel [73]. Figure 14: Block diagram representation of IR spectroscopy. #### IV-A1 Long-Wave versus Short-Wave NIR Spectroscopy The optical detection is useful approach to have precise glucose measurement. FIR (Far infra-red) based optical technique help to get the resonance between OH and CH for first overtone. However, long wave NIR has good performance in vitro testing. In similar way, the fiber-optic sensor is used along with laser based mid-infrared spectroscopy for vitro based glucose measurement. The continuous glucose measurement has been achieved with multivariate calibration model for error analysis [74]. The FIR approach has limitation of shallow penetration in comparison with short wave NIR. The short NIR would help to detect the glucose molecule more accurately [75]. The concept of NIR spectroscopy for glucose detection is shown in Fig. 15. The specific wavelength of NIR spectroscopy has already been applied earlier for precise glucose measurement using non-invasive measurement [76]. Some specific wavelength such as 940 nm has been considered for the detection of glucose [77]. The vibration of CH molecule has been observed at 920 nm with NIR spectroscopy [76]. In some other works, the glucose absorption has been validated for the range 1300 to 1350 nm and stretching of glucose has been identified in NIR region [78, 79]. The presence of glucose component has been measured at 1300 nm in the work [80]. Figure 15: Penetration depth various Infrared Signals in Human Skin [25, 32]. #### IV-A2 NIR Spectroscopy Based Methods A method to estimate the non-invasive blood glucose with NIR spectroscopy using PPG has been proposed in literature [81]. This method is performed using NIR LED and photo detector with an optode pair. At NIR wavelengths(935nm, 950nm, 1070nm), PPG signal is obtained by implementation of analog front end system. The glucose levels has been estimated using Artificial Neural Network (ANN) running in FPGA. A microcontroller is used, for the painless and autonomous blood extraction [82]. The ideal system Blood Glucose Measurement (BGM) in which the microcontroller is used to display the blood glucose and for the transmission of blood glucose. A remote device is used for the tracking of the insulin pump which is needed for diabetes management. This type of measurement [83] method uses change in the pressure of the sensitive body part, because it generates the sound waves. The response of the photo acoustic signals will be stronger when glucose concentration is higher. In order to improve SNR and for the reduction of noise to transfer the signal to the computer for further processing,the signal is then amplified. Feature extraction and glucose estimation is estimated by photo acoustic amplitude. In order to gather the photo acoustic signals, two pulsed laser diodes and piezoelectric transducer is used. Utilization of the LASER makes the setup costly and bulky. #### IV-A3 Non-invasive Blood Glucose Measurement Device (iGLU) In this approach, “Intelligent Glucose Meter (iGLU)” [84] has been utilized for the acquisition of data. This device works on a combination of NIR spectroscopy and machine learning. This device has been implemented using three channels. It uses an Internet-of-Medical-Things platform for storage and remote monitoring of data. In the proposed device, an NIR Spectroscopy is used with multiple short wavelengths [85]. It uses three channels for data collection. Each channel has its own emitter and detector for optical detection. Then the data collection processed by a 16 bit ADC with the sampling rate of 128 samples/second. Regression techniques is used to calibrate and validate the data and analyse the optimized model. The data that is stored on cloud can be used and monitored by the patients and the doctors. Treatment can be given based on the stored data values. This is a low cost device with more than 90 % accuracy but it does not give real time results. #### IV-A4 Why NIR is Preferred Over other Noninvasive Approaches? Glucose measurement has been done using various non-invasive approaches such as impedance spectroscopy, NIR light spectroscopy, PPG signal analysis and so on. But, apart from optical detection, other techniques have not be able to provide the precise measurement. PPG is one of the promising alternative but the PPG signal varies according to blood concentration [86, 87]. It may not be useful to have precise prediction of the blood glucose. The saliva and sweat properties vary from one person to another person. Therefore, it could not be reliable glucose measurement method. The other spectroscopy have been also applied for glucose measurement. However, they are not able to provide portable, cost effective and accurate prediction of body glucose.The glucose measurement using optical detection using long NIR wave which is not capable to detect the glucose molecules beneath the skin as it has shallow penetration [75]. Therefore, small NIR wave has been considered as potential solution for real-time glucose detection [77, 88]. ### IV-B Mid Infrared (MIR) Spectroscopy The bending and stretching of glucose molecules would be observed very well with Mid Infrared (MIR) spectroscopy [89]. The depth of skin penetration is very less because it tends to have larger absorption of water. This technique helps to have ISF glucose value in vivo measurement. There are some attempts for precise glucose measurement through saliva and palm samples. ### IV-C Blood Glucose Level Measurement using PPG The change of blood volume with absorption of the light from tissue has been detected with PPG signal [87]. The change of the blood volume has been measured using pressure pulse with help of light detector [86]. The change in volume of blood would result as the change of light intensity hence it may not be occur due to glucose molecule. This may result as inaccurate glucose measurement. The difference of NIR against PPG has been shown in Fig 16. The intelligent glucose measurement device iGLU is mainly based on principle of NIR spectroscopy which helps to have precise glucose measurement. There have been several work for glucose detection based on PPG signal [90]. The data from patient body has logged to estimate the presence of glucose using PPG. Subsequently, various machine leaning models have been used for prediction of body glucose value [91]. The different parameters from total 70 subjects of healthy and diabetes have been considered for the prediction using Auto- Regressive Moving Average (ARMA) models [92]. There have been also several other smart solutions for glucose estimation using PPG signal with intelligent algorithms [93, 94, 95]. One of the optical based techniques is Photo-plethysmography (PPG) which is used in advanced health care. It is non-invasive glucose measurement technique. In NIR spectrum a sensor similar to a pulse oximeter is used to record the PPG signal [87]. Photo transmitter and receiver is used to build the sensor which will operate in near infrared region at 920nm. At wavelength 920nm, by measuring changes in the absorption of light, a PPG signal can be obtained. The veins in the finger grow and contract with every heartbeat. A method of measuring blood glucose using pulse oximeter and transmission of the PPG glucose monitoring system is available [90]. As the glucose concentration increases, there is decrease in the light absorbance in the blood. The obtained signal is in the form of photo current, and for the filtering of this signal is then changed into the measurable voltage values. For the processing of filtered signal, lab view is used to estimate the blood glucose level. A system using machine learning techniques and PPG system for the measurement of blood glucose level non-invasively has been prototyped [86]. In this model, a PPG sensor, an activity detector, and a signal processing module is used to extract the features of PPG waveform. It finds the shape of the PPG waveform and the blood pressure glucose levels, the functional relationship between these two can be obtained then. In PPG, the change in light intensity will be varied according to changes in blood volume. PPG signal analysis is not based on the principle of glucose molecule detection. Hence, the system has limited accuracy [25, 32]. Fig. 16 illustrates the differences. Figure 16: PPG Versus NIR for Non-invasive Glucose Measurement [25, 32]. ### IV-D Impedance Spectroscopy Impedance spectroscopy (IMPS) refers to the dielectric spectroscopy [96]. The steps of impedance spectroscopy (IMPS) is shown in Fig. 17. This technique finds the dielectric properties of skin [97]. The current is directed through the skin [98]. Due to directed small current at multiple wavelengths, the impedance range is obtained [99]. The range lies between 100 Hz to 100 MHz [100, 101]. Change in glucose concentration will reflect the change in sodium ions and potassium ions concentration [102]. So, the cell membrane potential difference will be changed [103]. Thus, the dielectric value will be changed which predicts the glucose value of human body [104]. Figure 17: The Steps of Impedance Spectroscopy (IMPS). An enzyme sensor in a flow cell has been explored for glucose measurement in saliva [105]. Polypyrrole (PPy) supported with copper (Cu) nanoparticles on alkali anodized steel (AS) electrode for glucose detection in human saliva is available in [106]. The high precision level cannot be possible through these methods as sweat and saliva properties vary according to person. Hence, this approach is not suitable for glucose measurement in smart healthcare. ### IV-E Raman Spectroscopy Due to the interaction of light with a glucose molecule, the polarization of the detected molecule will change [107]. In this technique, oscillation and rotation of molecules of the solution are possible through the incident of LASER light [108]. The vibration of the molecule affects the emission of scattered light [109]. Due to this principle, blood glucose concentration can be predicted as [110]. This technique provides more accuracy with compared to infra-red spectroscopy technique [111]. There has been several research based on Raman spectroscopy to have precise glucose measurement. The validation has been also carried out on using vivo testing. Fig. 18 presents basic framework of Raman spectroscopy, whereas Fig. 19 presents its usage for noninvasive glucose measurement. Figure 18: Building blocks of Raman spectroscopy. Figure 19: Noninvasive glucose measurement using Raman spectroscopy. ### IV-F Time of Flight and THz Domain The blood glucose estimation is adopted though Time of Flight (TOF) measurements for vitro testing [112]. The short pulse of laser light is inserted in the sample for photon migration measurement. This photon will experience scattering and absorption phenomenon while traveling from the sample. The optical analysis of the photons would be useful for precise glucose measurement. ### IV-G Photo Acoustic Spectroscopy Photoacoustic spectroscopy refers to the photoacoustic effect for the generation of the acoustic pressure wave from an object (refer Fig. 20) [113]. In this spectroscopy technique, the absorption of modulated optical input provides the estimation of blood glucose detection [114]. High intense optical light is absorbed by an object according to its optical conditions [115]. This process provided excitations of particular molecules according to its resonant frequency [116]. The absorbed light is considered as heat which provides rising in localized temperature and thermal expansion of the sample [117]. The expansion in volume generates pressure in acoustic form [118]. The generated photoacoustic wave can be used to predict the glucose concentration through specific excited wavelengths which are resonant for the vibration of glucose molecules [119]. At the specific resonance frequency, the glucose molecule changes own characteristic. This change is in the acoustic waveform [120]. In previous work, 905 nm wavelength optical light is used for excitation [121, 122]. Figure 20: Photo acoustic spectroscopy. ### IV-H Capacitance Spectroscopy In the capacitance spectroscopy technique, inductor stray capacitance varies according to body capacitance (Fig. 21) [123]. The body capacitance is used to estimate body glucose concentration [124]. Flexible inductor based sensor follows the coupling capacitance principle for body glucose detection. In this technique, there is not any interaction between the inductive sensor and body skin through the current [125]. This is the advantage of the impedance spectroscopy technique. The stray capacitance of the inductive sensor will vary according to body glucose. In this technique, the effect of fat and muscles will be negligible with respect to body glucose [126]. Figure 21: The typical steps of capacitance spectroscopy. ### IV-I Surface Plasmon Resonance (SPR) The Surface Plasmon Resonance (SPR) utilizes electron oscillation approach at dielectric and metal interface for glucose sensing [127]. It detects mainly the change in refractive index before as well as after the analytes interaction. The optical fiber based SPR has been used for point of care measurement for glucose due to its portability. ### IV-J Radio Frequency (RF) Technique and Microwave Sensing In the RF technique, the variation in the s-parameters response reflects the change in blood glucose [128, 129]. Fig. 22 shows typical steps of this technique. The response is determined through the antenna or resonator [130, 131]. They follow the changes in dielectric constant value through the transmission [132]. The change in dielectric constant can be found as the change in resonance frequency spectrum through the antenna or resonator [133, 134]. The dielectric of blood varies according to blood glucose concentration. The human finger is an appropriate measurement object but there are many factors that play a cardinal and dominant role in the accuracy of measurement and repeatability. These are; the skin thickness, fingerprints, the applied pressure by the fingertip during measurement and positioning of a finger on the sensor [135]. Figure 22: Glucose measurement using RF sensing technique. ### IV-K Ocular Spectroscopy In the Ocular Spectroscopy technique, glucose concentration is measured through the tears. A specific lens is used to predict the body glucose concentration [136]. A hydrogel wafer is deposited to the lens. This wafer is prepared by boronic acid with 7 $\mu m$ thickness. The wafer is deposited on lens and then optical rays are inserted on the lens. Then reflected light will change its wavelength. Change in wavelength will refer to a change in glucose concentration in tears. ### IV-L Iontophoresis In the Iontophoresis or Ionization technique, a small electric current passes through the skin diffusively. Three electrodes are used for the same [137]. A small potential is applied through the electrodes to the different behaviour electrodes. During this process, glucose is transferred towards the cathode. The working electrode can have the bio-sensing function by the generation of current during applied potential through electrodes. This biosensor determines passively body glucose. The measurement is possible through wrist frequently [138]. ### IV-M Optical Coherence Tomography The Optical Coherence Tomography technique is based on the principle followed by reflectance spectroscopy. In this technique, low coherent light is excited through the sample (sample is placed in an interferometer). In an interferometer, a moving mirror is placed in reference arc. A photodetector is placed on another side and it detects the interferometric signal. This signal contains backscattered and reflected light. Due to this process, we could get high-quality 2-D images. The glucose concentration increases with the increment of the refractive index in interstitial fluids. Change in the refractive index indicates the change in the scattering coefficient [107]. So, the scattering coefficient relates to glucose concentration indirectly. ### IV-N Polarimetry The Polarimetry technique is commonly used in a clinical laboratory with more accuracy. The optical linear polarization-based technique is used for glucose monitoring [139]. This technique is usually based on the rotation of vector due to thickness, temperature and concentration of blood glucose. Due to the process of prediction of glucose, the polarized light is transmitted through the medium containing glucose molecule. Due to high scattering through the skin, the depolarization of beam is possible. To overcome this drawback, a polarimetric test has been done through the eye. The light passes through the cornea. This technique is totally unaffected due to rotation of temperature and pH value of blood [140]. Figure 23: Non-invasive glucose measurement using Polarimetry. ### IV-O Thermal Emission Spectroscopy The Thermal Emission Spectroscopy based technique is based on the naturally generated IR wave from the body. The emitted IR waves will vary according to body glucose concentration. The usual mid-IR emission from tympanic membrane of human body is modulated with tissue emitting. The selectivity of this technique is same as the absorption spectroscopy. Due to this technique, glucose can be determined through the skin, fingers and earlobe. This technique is highly precise and accurate for glucose measurement [141]. It could provide the useful solution which is precise and acceptable at clinical with measurement of thermal emission from tympanic membrane. ### IV-P Ultrasound The Ultrasound method is based on low frequency components to extract the molecules from skin similar as reverse iontophoresis method [142]. It is also alike sonophoresis and has larger skin permeability than reverse iontophoresis. Few or several tens of minutes of ultrasound exposure are required to pull glucose outward through the skin. There are few attempts for such technology and there is not any commercial device with such type of technology. ### IV-Q Metabolic Heat Conformation (MHC) The Metabolic Heat Conformation (MHC) method helps to measure the glucose value with metabolic heat and oxygen level along various physiological parameters considerations [143]. The mathematical model for metabolic energy conservation has been modified by several physiological parameters consideration such as pulse rate, oxyhemoglobin saturation, heat metabolic rate and the blood flow volume. This method has shown good reproducibility and decent accuracy in humans. ### IV-R Fluorescence The Fluorescence technique is based on the excitation of blood vessels by UV rays at particular specified frequency ranges [144]. This is followed through the detection of fluorescence at a specified wavelength. The sensing of glucose using fluorescence through tear has been done by the diffraction of visible light. At 380 nm, an ultraviolet LASER was taken for excitation through the glucose solution medium. Fluorescence was estimated which is directly related to glucose concentration. In this technique, the signal is not affected by variation in light intensity through the environment. ### IV-S Kromoscopy The Kromoscopy technique uses the response from various spectroscopic of NIR light with four different detectors over different wavelength [145]. It employs the multi-channel approach with overlapped band-pass series filters to determine the glucose molecule. In this method, the radiation of IR are imparted on the sample and this will be divided among four detectors with band-pass filter. Each detector will detect the light of the similar structure of the tissue. Subsequently, the complex vector analysis has been utilised to measure the glucose concentration. ### IV-T Electromagnetic Sensing In the Electromagnetic Sensing method, the variations in blood sample conductivity is observed by change in blood glucose concentration [146]. The alternation of electric field would be measured by electromagnetic sensor whenever there will be change in blood glucose concentration. This method utilizes the dielectric parameter of the blood samples. The frequency range for electromagnetic sensing is in the range of 2.4 to 2.9 MHz. The glucose molecule has maximum sensitivity at particular optimal frequency for given temperature of the medium. ### IV-U Bioimpedance Spectroscopy and Dielectric Spectroscopy It is useful to measure the variation of the blood glucose with help of conductivity and permittivity from red blood cells membrane [147]. The spectrum of bioimpedance spectrum is measured from 0.1 to 100 MHz frequency range. It help to find the resistance with passing through electric current which is flowing from human biological tissue. The change of plasma glucose would allow the changes in potassium and sodium to have the change in conductivity of the membrane of the red blood cell. The multisensor approach is usually incorporated with this spectroscopy in order to measure sweat, moisture, movement and temperature for precise glucose measurement. ### IV-V Reverse Ionospheresis The small DC current is passed from anode to cathode on the skin surface to have small interstitial fluid (ISF). Iontophoresis is employed for ionized molecules penetration at skin surface by such low current [148]. The electric potential is passed from anode and cathode to electroosmotic flow across the skin. This would allow to extract the molecules through skin whereas the the molecules of glucose are moved towards the cathode. The enzyme method helps to sense the concentration of glucose molecules through oxidation process. the method has very widely accepted and has good potential to measure accurate glucose value. ### IV-W Sonophoresis The Sonophoresis technique is based on the cutaneous permittivity of the interstitial fluid (ISF) [149]. It also uses enzyme method for glucose measurement. The low frequency ultrasound wave has been applied in order to have glucose molcules at the skin surface. The cutaneous permittivity of the ISF is increased to enable glucose at the epidermis surface. The contraction and expansion occurs in stratum corneum that subsequently opens the ISF pathway. There has been some attempts with this method for glucose detection but it has been observed that it could be helpful in drug delivery in stead of glucose measurement. ### IV-X Occlusion Spectroscopy The Occlusion spectroscopy based methods depend on the concept of light scattering which is of inverse proportion of glucose concentration [150]. The flow is ceased for few seconds by applying pressure with pneumatic cuff. The volume of blood would change due to pulse generated from the pressure excursion. The light is transmitted through the sample and the variation of the intensity of in a received light defines the glucose concentration. The momentary blood flow cessation helps to get higher SNR value of the received signal. Hence, the sensitivity for glucose detection would be increases with good robustness for accurate glucose measurement. ### IV-Y Skin Suction Bluster Technique The Skin Suction Bluster technique uses the concept of blister generation through vacuum suction over limited skin area [151]. The glucose measurement is performed on fluid which is collected from the blister. It has lower glucose molecules than plasma but it is well enough to have the glucose measurement. This method has low risk of infection, painless and well- tolerated. It is actually useful to measure HbA1c value which represents three month average glucose value. ### IV-Z Multimodal approach based measurement A two modal spectroscopy combining IMPS and mNIR spectroscopy is explored for high-level reproducibility of non-invasive blood glucose measurement [152]. These two techniques are combined to overcome the limitation of individual employed technique [153]. Impedance spectroscopy based circuit measures the dielectric constant value of skin or tissue through RLC resonant frequency and impedance to predict glucose level [154]. To improve the accuracy of NIR spectroscopy, mNIR spectroscopy technique is used. In this technique, three wavelengths 850 nm, 950 nm and 1300 nm are used [155]. For precise and accurate measurement, IMPS and mNIR are joined by an ANN (Artificial Neural Network) through DSP processor [80]. Therefore multimodel approaches have been explored for precise glucose measurement in the literature [joo2020vivo, feng2019multi]. Figure 24: Multimodal IC based non invasive glucose measurement. Table II: Approaches Comparison with Noninvasive Works [25, 32]. Works | Spectroscopy | Spectra | Specific | Measurement | Linearity ---|---|---|---|---|--- | technique | | wavelength | range | (%) Singh, et al. [160] | Optical | - | - | 32-516 mg/dl | 80 Song, et al. [80] | Impedance and Reflectance | NIR | 850-1300 nm | 80-180 mg/dl | - Pai, et al. [161] | Photoacoustic | NIR | 905 nm | upto 500 mg/dl | - Dai, et al. [162] | Bioimpedance | - | - | - | - Beach, et al. [163] | Biosensing | - | - | - | - Ali, et al. [88] | Transmittance and Refraction | NIR | 650 nm | upto 450 mg/dl | - Haxha, et al. [77] | Transmission | NIR | 940 nm | 70-120 mg/dl | 96 Jain, et al. [164] | Absorption and Reflectance | NIR | 940 nm | 80-350 mg/dl | 90 Jain, et al. (iGLU 1.0) [32, 28] | Absorption and Reflectance | NIR | 940 and 1300 nm | 80-420 mg/dl | 95 Jain et al. (iGLU 2.0) [29, 25] | Absorption and Reflectance | NIR | 940 and 1300 nm | 80-420 mg/dl | 97 Table III: Statistical and Parametrical Comparison with Noninvasive Works [25, 32]. Works | R | MARD | AvgE | MAD | RMSE | Samples | Used | Measurement | Device ---|---|---|---|---|---|---|---|---|--- | value | (%) | (%) | (mg/dl) | (mg/dl) | (100%) | model | sample | cost Singh, et al. [160] | 0.80 | - | - | - | - | A&B | Human | Saliva | Cheaper Song, et al. [80] | - | 8.3 | 19 | - | - | A&B | Human | Blood | Cheaper Pai, et al. [161] | - | 7.01 | - | 5.23 | 7.64 | A&B | in-vitro | Blood | Costly Dai, et al. [162] | - | 5.99 | 5.58 | - | - | - | in-vivo | Blood | Cheaper Beach, et al. [163] | - | - | 7.33 | - | - | - | in-vitro | Solution | - Ali, et al. [88] | - | 8.0 | - | - | - | A&B | Human | Blood | Cheaper Haxha, et al. [77] | 0.96 | - | - | - | 33.49 | A&B | Human | Blood | Cheaper Jain, et al. [164] | 0.90 | 5.20 | 5.14 | 5.82 | 7.5 | A&B | Human | Blood | Cheaper Jain, et al. (iGLU 1.0) [32] | 0.95 | 6.65 | 7.30 | 12.67 | 21.95 | A&B | Human | Blood | Cheaper Jain et al. (iGLU 2.0) [29] | 0.97 | 4.86 | 4.88 | 9.42 | 13.57 | Zone A | Human | Serum | Cheaper ## V Post-processing and calibration techniques for non-invasive Glucose- Level measurement This Section presents various post-processing and calibration techniques which are deployed in various systems or frameworks for noninvasive glucose-level monitoring. ### V-A Post processing and calibration techniques Various calibration processes have been applied for a high level of accuracy and noise reduction from received signal. These post-processing techniques are used to design the model for errorless continuous monitoring [165, 166]. #### V-A1 Noise Minimization and Signal Conditioning The coherent averaging technique has been adapted to minimize the variance of random noise [167]. The impact of noise is minimized with averaging of N number of individual samples coming from the continuous frames [168]. Frames in the maximum count have been chosen for averaging to have SNR improvement [169]. This proposed coherent averaging has been used frequently through MATLAB and coherent averaged signal acquired. Golay code has been proposed as calibration of measured data. The filtering or cancellation of unusual measured data has been achieved through the implementation of Golay code [170, 171, 172]. #### V-A2 Computation Models for glucose Estimation The regression model of regularized least square is proposed by several researchers for measurement [173]. The estimated value is computed from photoacoustic signals. These photoacoustic signals are used to calibrate for estimation of glucose concentration [174]. This can be possible through multi- variable linear regression model [175]. With the objective of high-level accuracy, a post-processing SVM technique is proposed [176]. Support vector machine is a better option of correct measurement in glucose monitoring system [177]. Artificial Neural Network (ANN) has also been proposed for data combining [178]. The measured data from multiple techniques are combined through the proposed neural network model [179]. This artificial neural network has been implemented in DSP processor [180, pancholi2018portable]. This proposed data interpretation model has been used for combining and calibrating of data for final estimated glucose concentration [181]. ### V-B Metrics for Model Validation The calibration method is used to have precise blood glucose estimation for measurements [182]. The obtained glucose concentration values are used to compare with conventionally measured glucose concentrations [183]. The Clarke error grid analysis has been considered maximum measurement for analysis which is used to check the performance of any device for accuracy measurement [184]. The process flow is represented in Fig. 25. Figure 25: Representation of Metrics for model Validation. ### V-C Clinical Accuracy Evaluations using Clarke error grid analysis The Clarke Error Grid has been analysed as benchmark tool to examine the clinical precision for biomedical application. It has prediction of point as well as rate accuracy, and it amends for physiologic time lags inherent for measurement of body glucose. The exploitation of the Clarke error grid modelling will significantly make easy the development and refinement of a precise biomedical device. In 1970, this technique was developed by C.G. Clark to identify the accuracy of the clinical trials which helps to find the precision of estimated blood glucose with blood glucose value through the conventional method. A description of the Error Grid Analysis came into view in diabetes care in 1987. The grid is divided with five different zones mainly as zone A, zone B, zone C, zone D, and zone E. If the values residing in either zone A or zone B then it signifies satisfactory or accurate prediction of glucose results according to Beckman analyzer. The zone C values may prompt gratuitous corrections which may lead to a poor outcome. If the values are under zone D which actually defines a hazardous failure to sense. Zone E reflects the “erroneous treatment” [185]. Figure 26: Clarke error grid analysis. ## VI Consumer Products for Glucose-Level Measurement There have been several non-invasive glucometer at market(such as Freestyle Libre sensor, SugarBEAT from Nemaura medical) which used for proper medication. They would be like skin-patch with daily disposable feature and adhesive to have the continuous glucose monitoring. Most of the consumer products fail to provide precise glucose measurement and hence they are much popular for diabetes management. There are some products as DiaMon Tech, glucowise, glucotrack, glutrac and CNOGA medical device. Glutrac is smart healthcare device but it has accuracy issues for the blood glucose measurement. It has higher cost while precision is still not acceptable. The non-invasive stripless device known as Omelon B-2 has been used for the CGM. The fluorescent technique based Glucosense has been made for contonuous monitoring of the glucose value. The flexible textile-based biosensor has developed from Texas University to measure the glucose level. All the available device have accuracy issues and considerable higher cost. ### VI-A Wearable versus Non-wearable for Glucose Monitoring The glucose monitoring have been attempted using non-wearable and wearable solutions in the literature. Most of the non-wearable approaches are based on various spectroscopy such as photoacoustic spectroscopy, Raman spectroscopy etc. The implantable devices are of semi-invasive type and are mainly of biosensors in nature. Sweat patches, Glucowatch and Smart contact lenses are of wearable devices category. LifePlus has developed non-invasive and wearable device for CGM purpose and it is under consideration for commercialization. Most of non-invasive device are wearable and helpful for frequent glucose measurement. The continuous glucose monitoring would be more acceptable if they could measure the blood glucose values in day to day life. Therefore, the wearable devices are more state of art solutions then non wearable devices. ### VI-B Noninvasive Glucose Measurement Consumer Products There are variety of products such as GlucoTrack®, glucometer from Labiotech [186], and similar available solutions have accuracy issues and cost is also high. The glucowise is another non-invasive device for continuous glucose measurement from Medical Training Initiative (MTI) ™. The Raman scattering spectroscopy based non-invasive solution is also developed by 2M Engineering [187]. These devices are not much popular because of their cost and precision. Further for the high level of accuracy of glucose measurement, Glucotrack™ has been developed by integrity applications Ltd. [188]. This non-invasive glucose monitoring device employed three consecutive ultrasonic spectroscopy, thermal emission and electromagnetic techniques. This device is highly precise and accurate because of a combination of three techniques [189]. A comparative perspective of various consumer products for noninvasive glucose measurement has been summarized in Table VI-B. Table IV: A Comparative perspective of a selected consumer products for noninvasive glucose measurement. Company | Device | Technology | Object | Summary | Snapshot ---|---|---|---|---|--- Cygus Inc. (USA) | GlucoWatch G2 Biographer | Reverse iontophoresis | Wrist skin | It would be worn as watch is used with disposable component, autosensor which is to be attached at back of biographer that contact with the skin to provide frequent glucose monitoring | CNOGA (Israel) | Combo glucometer | Tissue photography analysis | Finger | On basis of tissue photography analysis from fingertip capillaries, this device can analyze various bio parameters in very short time | Pendragon Medical (Switzerland) | Pendra | Impedance Spectroscopy | Wrist Skin | It helps to measure the glucose with sodium transport of erythrocyte membrane, The change of fluxes of transmembraneous sodium occur due to impedance field which is detected by device to generate final glucose value | OrSense Ltd. (Israel) | OrSense NBM-200G | Occlusion Spectroscopy | Fingertip skin | It is based on optical concept on finger which is attached to a ring-shaped sensor probe. The probe has red/near-infrared RNIR spectral region light source as well as detector. It has pneumatic cuffs which generates systolic pressure to produce optical signal for glucose monitoring | C8 Medisensors (USA) | C8 Medisensor Glucose detector | Raman Spectroscopy | Fingertip skin | This technique is based on monochromatic light source passes through skin where scattered light is detected. The generated colors from Raman spectra helps to exact chemical structure of glucose molecule | Integrity Applications (Israel) | Glucotrack | Combination of Electromagnetic, ultrasonic and Thermal | Ear lobe tissue | In this device, three different techniques are used concurrently to increase the accuracy and precision | Tech4Life Enterprises (USA) | Non invasive glucometer | Infra red Spectroscopy | Finger | It is helpful for Hyperglycemia or Pre-Diabetic patients which allow for regular monitoring of precise blood glucose measurement at every 30 seconds | MediWise Ltd. (United Kingdom) | Glucowise | Radio Wave Spectroscopy | Forefinger skin/Earlobe | This non invasive wireless device can measure glucose concentration in very short time. It is based on electromagnetic waves of specific frequencies for blood glucose detection. It uses a thin-film layer of metamaterial which increases the penetration for precise glucose measurement | Nemaura Medical (united Kingdom) | SugarBeat | Reverse iontophoresis | Arm,Leg and adbomen | This has been proved accurate device, pain-free continuous blood glucose monitoring. SugarBEAT® provides real-time, needle-free glucose measurement. Generally, it needs one time finger-prick test for calibration. One time finger prick is used when new patch is required to insert | Abott Ltd. (USA) | Free Style Libre | Glucose oxidase method | Fore-arm skin | It uses enzyme glucose sensing technology for the detection of glucose levels through interstitial fluid Glucose oxidase method is applied through sensor where electrical current proportional to the glucose concentration and glucose can be measured. | C8 Medisensor | Non invasive glucose monitor | Raman Spectroscopy | Fore arm skin | Raman spectroscopy technique based this device can detect glucose in blood through returning spectrum from the skin | ## VII Glucose Level Controls Approaches and Consumer Products Various models have been developed for diet control using various parameters for glucose-insulin balance. The parameters are mainly includes net hepatic glucose balance, renal excretion rate, glucose absorption rate and peripheral glucose utilization for the glucose consumption prediction for the diabetic patients. These are useful parameters to calculate the glucose level by proper insulin dosage along with scheduled diet plan. Therefore, the glucose-insulin control model was designed to balance glucose insulin level in the body for diabetes persons using proper medication. ### VII-A Glucose Controls Approaches The mathematical models for insulin delivery have been presented to determine the coefficients of blood regulation. The model has been proposed for insulin secretion with glycemic profile for type 2 diabetic person [190, 191]. The non-linear model is developed using differential equation with delay model with help of non-diabetic subjects [192]. Most poplar “Uva/Padova Simulator” was also explored which was approved from FDA to have the proper clinical trials. The parameter are extracted with type 1 diabetic virtual patients [193]. The intravenous test for glucose tolerance with Hovorka maximal model has explored for non-diabetic subjects [194]. The samples from type 1 diabetic persons were collected to explain the model with help of time monitoring. The model is proposed mathematically for blood glucose value prediction in the postprandial mode for type 1 diabetes patients [195, 196]. The mathematical model for glucose-insulin balance for longer period is explored using two days clinical information [197]. A algorithm was developed for T1DM patient meal detection for the purpose of frequent glucose measurement. The work has integrated bolus meal mathematical model for glucose-insulin delivery model [198]. Diabetic and healthy people were considered to acquire the values for the variable state dimension algorithm. The diet plan was examined in the absence of meal profile to have the glycemic profile balance, an intelligent PID controller (iPID) was developed to type 1 diabetic person [199, 200]. ### VII-B Glucose Controls Consumer Products Type-1 diabetic patients aren’t able to produce insulin. Insulin is a hormone that can balance body sugar (glucose) which is a prime source of energy that obtains from carbohydrates. If anybody has type 1 diabetes, it is necessary to be ready for insulin therapy. Insulin may be injected by self-injection, patients who take multiple injections daily of insulin may also think about use of an insulin pump. An insulin pump gives short-acting insulin all day long continuously. The insulin pump replaces the requirement of long-acting insulin. A pump also substitutes the requirement of multiple injections per day along with continuous insulin infusion and also serves to improve the glucose levels. Various types of insulin pumps are already available in the market as consumable product mainly as Animas, Medtronic, Roche, Tandem and Omnipod insulin pump are consumables. These insulin pumps are advanced to each other in terms of their upgraded features. A comparative perspective of a selected state of art approach for glucose measurement to have better glyncenic profile control is presented in Table VII-B. Table V: A comparative perspective of a selected state of art approaches for glucose measurement Work | Technology | Object | Findings | Observation ---|---|---|---|--- [86] | photoplethy- -smography (PPG) | Finger | It helps to extract the features of PPG signal through machine learning models to estimate Systolic and diastolic blood pressure and blood glucose values | machine learning models applied where random forest technique has best prediction results as $R^{2}_{SBP}$ = 0.91, $R^{2}_{DBP}$ = 0.89 and $R^{2}_{BGL}$ = 0.90. CEG has 87.7% observation in Zone A, 10.3 % in Zone B, and 1.9% in Zone D [201] | mid-infrared attenuated total reflection (ATR) spectroscopy and trapezoidal multi-reflection ATR prism | oral mucosa inner lips | Using a multi-reflection prism brought about higher sensitivity, and the flat and wide contact surface of the prism resulted in higher measurement reproducibility & spectra around 1155 cm$-1$ for different blood glucose levels for fasting and before fasting | the coefficient of determination $R^{2}$ is 0.75. The standard error is 12 mg/dl, and all the measured values are in Region A [202] | Optical Coherence Tomography | Fingertip | It measures the optical rotation angle and depolarization index of aqueous glucose solutions with low and high scattering, respectively. The value of angle increases while depolarization index decreases with glucose value increases | The correlation factor has a value of $R^{2}$ 0.9101, the average deviation is found around 0.027. [203] | Contactlenses fluoresence | Tears | The fabrication of a soft, smart contact lens in which glucose sensors, wireless power transfer circuits, and display pixels to visualize sensing signals in real time are fully integrated using transparent and stretchable nanostructures | The usage of smart and soft lens would provide the wireless operation at real-time for glucose monitoring in tears [204] | transmission spectroscopy | Sliva | After completely absorbing the sufficient amount of saliva on the strip, the sample would reach detection zone via paper microfluidic movement and enzymatic reaction between GOx and salivary glucose would initiate a pH change, resulting in a change in strip color that was recorded by using RGB detector on the handheld instrument which helps for glucose detection | The developed biosensor had a wide detection range of detection between 32- and 516-mg/dL glucose concentration while the sensitivity of detection was 1.0 mg/dL/count at a limit of detection (LOD) of 32 mg/dL within a response time of 15 s [205] | impedance spectroscopy (IMPS) and multi-wavelength near-infrared spectroscopy (mNIRS) | Left Handand wrist Hand | IMPS and mNIRS use the indirect dielectric characteristics of the surrounding tissue around blood and the optical scattering characteristics of glucose itself in blood, respectively, the proposed IC can remove various systemic noises to enhance the glucose level estimation accuracy | mean absolute relative differences (mARD) to 8.3% from 15.0% of the IMPS and 15.0–20.0% of the mNIRS in the blood glucose level range of 80–180 mg/dL. From the Clarke grid error (CGE) analysis, all of the measurement results are clinically acceptable and 90% of total samples can be used for clinical treatment [84] | NIR Spectroscopy | Fingertip | short NIR waves with absorption and reflectance of light using specific wavelengths (940 and 1,300 nm) has been introduced | The Pearson’s correlation coefficient (R) is 0.953 and MAD is 09.89 which is RMSE 11.56 [206] | Microwave Detection | earlobe | The absorption spectrum of microwave signal helps to measure using two antenna. The sine wave of 500 MHz is for blood glucose measurement. | It can measure blood glucose from 0 to 500 mg/dl with step size of 200 mg/dl used for the experiment for testing the resolution. It obtained 0.5226 mean standard deviation while the minimum value of standard deviation is 0.04119. [94] | PPG | Finger | The prediction of blood glucose was with machine-learning using a smartphone camera. First the invalid data was separated and The system did not require any type of calibration | The device was able to measure glucose only 70-130 mg/dl range. The results show accuracy of 98.2% for invalid single-period classification and and the overall accuracy is 86.2%. [46] | MEMS | Finger | It is minimally invasive technique known as e-Mosquito which extracts blood sample with shape memory alloy (SMA)-based microactuator. It considered as first ever wearable device which performs the automatic situ blood extraction and performs the glucose analysis. | The method provided linear correlation ($R^{2}$ = 0.9733) between standard measurements and the e-Mosquito prototype. [207] | Visible NIR | Wrist | The paper developed biosensor which helps to exploit pulsation of arterial blood volume from the wrist tissue. The visible NIR spectroscopy was used for reflected optical signal to estimate blood glucose. | The correlation coefficient (Rp) value after averaging all observation is 0.86, whereas the standard prediction error is around 6.16 mg/dl. [201] | mid-infrared attenuated total reflection (ATR) spectroscopy | inner lip mucosa | Novel optical fiber probe was introduced using multireflection prism with ATR spectroscopy. The sensitivity increases with the number of reflections while measurement reproducibility was higher due to prism’s wide and flat & wide contact surface. | The experimental results reveals the glucose signature at various spectra between fasting state and after the glucose injection. The plot for calibration defines peak for absorption at 1155 $cm^{-1}$ which has glucose measurement error less than 20% [chowdhury2016noninvasive] | modulated ultrasound and infrared technique | Finger | the MATLAB toolbox is used with Fast Fourier Transform (FFT) for blood glucose extraction. The random blood glucose level test and oral glucose tolerance test was done for the human subjects for performance measurement | The RMSE value of noninvasive and invasive measurement from both tests 28.20 mg/dl and 23.76 mg/dl. The pearson correlation coefficient was 0.85 and 0.76, respectively. At the same time MSE was 17.76 mg/dl and 15.92 mg/dl. ## VIII Glucose-Level Measurement and Controls - IoMT Perspectives The practical and sustainable mechanisms are the prime factors of smart and automated healthcare system. These are being optimized to support the population migration and quality of life in smart cities and smart villages [208, 209]. The features of smart healthcare system are continuous monitoring for critical care, ambient intelligence and quality of service for proper point of care mechanism [210, 211]. The non-invasive and precise glucose measurement is requirement for diabetic person and would also needed to store the information using IoMT for proper treatment [26]. The traditional method for glucose measurement has limited capability and is not able to assist the remotely located healthcare provider. The diabetic person would like to monitor their glycemic profile frequently in a day with support of storing at cloud server. The smart health care system would allow the point of care treatment for diabetes person with frequent monitoring. The internet of Medical Things (IoMT) has allowed to connect the patients with doctors remotely for rapid treatment and special assistance using smart healthcare [208]. The continuous monitoring of vital parameters have provided to awareness about the diet plan and routine activity management with contemporary healthcare consumers devices. With the active support of remote healthcare solution, the smart healthcare has potential to ameliorate the quality of service at reduced cost. The smart sensors would capture the patient data continuously and help to store the data on cloud data centre. It is also useful for the analysing the data and easy exchange of the information through mobile applications to doctors as well as patients. The healthcare Cyber-Physical System (H-CPS) has been used successfully to address the various challenges of healthcare sector with intelligent algorithms. The continuous glucose monitoring would certainly help the diabetic patients to plan their diet for the purpose of glucose control. The solution should be precise, low cost and easy to operate for rapid diagnosis [32, 28]. The serum glucose would always consider as accurate than capillary measurement. Therefore, the rapid serum glucose measurement solution with continuous monitoring is desired for the smart healthcare. The novel serum glucometer is portable device and is also integrated with IoMT to store the glucose values continuously at cloud. It would be useful for the healthcare provider to track the health records of remote located diabetes person. The smart healthcare management of continuous glucose measurement is defined in Fig. 27. Figure 27: Blood glucose diagnosis and Control in smart health care system. A detailed example of a closed-loop system that presents glucose-level monitoring and insulin release to control it is illustrated in Fig. 28 [2]. This IoMT framework can provide a better solution for evaluation of insulin doses through the closed-loop automated insulin secretion diabetes control. Such an integrated IoMT framework can be implemented to diagnose and for the treatment of diabetic patients in terms of controlling their blood glucose level in smart healthcare and be effective in smart village and smart cities for healthcare with limited medical personnel. Figure 28: A closed-loop automated insulin secretion diabetes control system in an IoMT framework [2]. The security and privacy issues of the medical devices are paramount aspect in any IoT network. The hardware security of wearable device is very crucial because control actions mainly occur in wireless media. The security vulnerabilities are defined for glucose measurement device and its control are shown in Fig. 29. The devices security are important due to connected health system in an insecure and unreliable IoMT framework [212]. The integrity of useful medical information is also crucial security aspect of smart healthcare. All patients medical records are stored over the server therefore the security of such data are also really important. The controlled access with proper authentication is required to have secure monitoring with proper patient treatment. Figure 29: Our Long-Term Vision of Security-Assured Non-invasive Glucose-Level Measurement and Control through our Proposed iGLU. ## IX Short-Comings of Existing Works and Open Problems This Section outlines the shortcomings and discusses some open problems of glucose level measurements and control. ### IX-A Limitations of the Existing Approaches and Products 1. 1. Photoacoustic spectroscopy has been implemented for glucose measurement. Real- time testing and validation have not been done from human blood. The artificial solution was prepared in the laboratory for glucose measurement. The prototype module with LASER and corresponding detector is costly and at the same time requires considerable bigger area and does not provide portable solution. Therefore, it is not much popular solution for continuous glucose monitoring. 2. 2. Raman spectroscopy is a nonlinear scattering which occurs when monochromatic light interacts with a certain sample. Raman spectroscopy based solution is applicable for a laboratory test and also occupies the significant larger area. Hence, the system based on this approach will not be applicable for frequent glucose measurement. 3. 3. The retina based glucose measurement is also one of the alternate non-invasive glucose detection approach, data has also been collected through retina for glucose measurement. Such technique is not useful for the glucose measurement all the time. 4. 4. In case of bio-capacitance spectroscopy, the slight difference in placing the sensor at the same location might affect the output of the sensor. Effect of pressure on the sensor, body temperature and sweat on the skin may also affect the output of the sensor. 5. 5. Glucose detection is performed with the impedance spectroscopy (IMPS) by electrodes connection to the skin which is affected with skin. The accuracy is always an issue as the saliva and sweat could change for each individual and that may reflect to the precision of glucose. Therefore, this technique is not best for reliable glucose measurement in smart healthcare. 6. 6. PPG signal has been used to extract features for blood glucose level prediction. But the PPG may be precise blood glucose measurement technique where the output value would vary according the blood volume only. Therefore, the glucose molecule has not been detected precisely in the blood sample using this technique. ### IX-B The Open Problems in Non-invasive Glucose Measurement There are lots of challenges for commercialization of non-invasive glucose measurement device. But, some open problems have been discussed which are prime challenges for precise non-invasive glucose measurement. These challenges have been represented in Fig. 30. The precise glucose measurement of hypoglycemic patient and long-time continuous glucose measurement without instantaneous error are the open problems which are focussed by the researchers recently. Figure 30: Open Challenges in Noninvasive Glucose-Level Measurement. * • The effect of blood pressure, body temperature and humidity have not been considered in the literature which affect the values of glucose measurement. * • The cost effective and portable solution of continuous glucose measurement device has also not been addressed properly. * • The accurate glucose measurement has been also been open challenge for full rage from 40 mg/dl to 450 mg/dl. * • The effective integration of glucometer with internet of medical things for continuously data logging to the cloud has still not potentially resolved. * • The mathematical model for automatic insulin secretion according to measured glucose value has to be address in better manner with internet framework. * • The privacy and security issues of insulin and blood glucose measurement system is still not resolved yet. * • The efficient power management mechanism has to be developed for continuous glucose measurement with insulin delivery system. ## X Conclusions and Future Research The paper presents survey of glucose measurement approaches along with overview of glucose control mechanism. Many techniques available in literature are only a proof of concept, showing good correlation between device estimated result and reference value of blood glucose. However, they are neither accurate nor cost effective solutions and not available for commercial purpose. The optical detection using short NIR has been potential solution to mitigate the drawbacks of all previous methods. In future, the multi-model approaches could be considered for precise glucose estimation. The device or prototype model should be more effective in different zones to support the continuous health monitoring. It should be implemented as a portable device for real time application with more frequently. This device should be developed as continuous health monitoring with minimum cost. The future research for upcoming noninvasive glucose monitoring device is mentioned in Fig. 31. The device is required to be integrated with advanced IoMT framework. This advanced IoMT framework will alow to connect the device with all nearest diabetic centers to get best treatment. Unification of glucose-level measurement and automatic diet quantification can have strong impact on smart healthcare domain [213]. The durability, portability and user- friendly device is also the future vision in this era. The device should have the feature of border-line cross indication. Because of this feature, any person will be aware to take own blood glucose level. A secured device with end to end users control and authentication is also necessary for future advancement. Physical Unclonable Function (PUF) based security of IoMT-devices can be effective for IoMT-devices which are intrinsically resource and battery constrained [212, 214]. Unified healthcare Cyber-Physical System (H-CPS) with blockchain based data and device management can be effective and needs research [215, 216]. Figure 31: Our Future Vision for Non-invasive Glucose-Level Measurement. ## References * [1] Diabetestalk, “What is the normal fasting blood sugar range for adults,” 2018, last Accessed on 18 Jan 2021. [Online]. Available: https://diabetestalk.net/blood-sugar/what-is-the-normal-fasting-blood-sugar-range-for-adults * [2] P. Jain, A. M. Joshi, and S. P. Mohanty, “iGLU 1.1: Towards a glucose-insulin model based closed loop iomt framework for automatic insulin control of diabetic patients,” in _2020 IEEE 6th World Forum on Internet of Things (WF-IoT)_. IEEE, 2020, pp. 1–6. * [3] A. M. Joshi, U. P. Shukla, and S. P. Mohanty, “Smart healthcare for diabetes during COVID-19,” _IEEE Consumer Electronics Magazine_ , vol. 10, no. 1, pp. 66–71, January 2020. * [4] Amit M Joshi, U. P. Shukla, and S. P. Mohanty, “Smart healthcare for diabetes: A COVID-19 perspective,” _arXiv preprint arXiv:2008.11153_ , 2020. * [5] A. M. Alsamman and H. Zayed, “The transcriptomic profiling of COVID-19 compared to SARS, MERS, Ebola, and H1N1,” _bioRxiv_ , 2020. * [6] I. D. Federation, “IDF Diabetes Atlas - Diabetes is rising worldwide… and is set to rise even further,” 2019, last Accessed on 21 March 2020. [Online]. Available: https://diabetesatlas.org/en/sections/worldwide-toll-of-diabetes.html * [7] N. H. Cho, J. E. Shaw, S. Karuranga, Y. Huang, J. D. da Rocha Fernandes, A. W. Ohlrogge, and B. Malanda, “IDF diabetes atlas: Global estimates of diabetes prevalence for 2017 and projections for 2045,” _Diabetes Research and Clinical Practice_ , vol. 138, pp. 271–281, April 2018. * [8] P. Saeedi, I. Petersohn, P. Salpea, B. Malanda, S. Karuranga, N. Unwin, S. Colagiuri, L. Guariguata, A. A. Motala, K. Ogurtsova _et al._ , “Global and regional diabetes prevalence estimates for 2019 and projections for 2030 and 2045: Results from the international diabetes federation diabetes atlas,” _Diabetes research and clinical practice_ , vol. 157, p. 107843, 2019. * [9] Clevelandclinic, “Diabetes Mellitus: An Overview,” 2020, last Accessed on 18 Jan 2021. [Online]. Available: https://my.clevelandclinic.org/health/diseases/7104-diabetes-mellitus-an-overview * [10] Drugs, “Type 1 diabetes mellitus,” 2020, last Accessed on 18 Jan 2021. [Online]. Available: https://www.drugs.com/health-guide/type-1-diabetes-mellitus.html * [11] L. M. Leontis and A. Hess-Fischl, “Diabetes mellitus: An overview,” 2019, last Accessed on 18 Jan 2021. [Online]. Available: https://www.endocrineweb.com/conditions/type-2-diabetes/type-2-diabetes-symptoms * [12] A. Pietrangelo, “What are the different types of diabetes?” 2018, last Accessed on 18 Jan 2021. [Online]. Available: https://www.healthline.com/health/diabetes/types-of-diabetes * [13] M. J. Fowler, “Diabetes: Magnitude and mechanisms,” _Clinical Diabetes_ , vol. 28, no. 1, pp. 42–46, 2010. [Online]. Available: https://clinical.diabetesjournals.org/content/28/1/42 * [14] H. Yin, B. Mukadam, X. Dai, and N. Jha, “DiabDeep: Pervasive Diabetes Diagnosis based on Wearable Medical Sensors and Efficient Neural Networks,” _IEEE Transactions on Emerging Topics in Computing_ , pp. 1–1, 2019. * [15] P. Zhang, “Global healthcare expenditure on diabetes for 2010 and 2030.” _Diabetes Research and Clinical Practice_ , 2011. * [16] J. Venkataraman and B. Freer, “Feasibility of non-invasive blood glucose monitoring: In-vitro measurements and phantom models,” in _2011 IEEE International Symposium on Antennas and Propagation (APSURSI)_ , July 2011, pp. 603–606. * [17] S. H. Wild, G. Roglic, A. Green, R. Sicree, and H. King, “Global prevalence of diabetes: Estimates for the year 2000 and projections for 2030,” _Diabetes Care_ , vol. 27, no. 10, pp. 2569–2569, 2004. [Online]. Available: http://care.diabetesjournals.org/content/27/10/2569.2 * [18] D. R. Whiting, L. Guariguata, C. Weil, and J. Shaw, “Idf diabetes atlas: Global estimates of the prevalence of diabetes for 2011 and 2030,” _Diabetes Research and Clinical Practice_ , vol. 94, no. 3, pp. 311 – 321, 2011. * [19] P. H. Siegel, A. Tang, G. Virbila, Y. Kim, M. C. F. Chang, and V. Pikov, “Compact non-invasive millimeter-wave glucose sensor,” in _2015 40th International Conference on Infrared, Millimeter, and Terahertz waves (IRMMW-THz)_ , Aug 2015, pp. 1–3. * [20] S. M. Alavi, M. Gourzi, A. Rouane, and M. Nadi, “An original method for non-invasive glucose measurement: preliminary results,” in _2001 Conference Proceedings of the 23rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society_ , vol. 4, 2001, pp. 3318–3320 vol.4. * [21] X. Li and C. Li, “Study on the application of wavelet transform to non-invasive glucose concentration measurement by nirs,” in _2015 Fifth International Conference on Instrumentation and Measurement, Computer, Communication and Control (IMCCC)_ , Sept 2015, pp. 1294–1297. * [22] P. P. Pai, P. K. Sanki, S. K. Sahoo, A. De, S. Bhattacharya, and S. Banerjee, “Cloud computing-based non-invasive glucose monitoring for diabetic care,” _IEEE Transactions on Circuits and Systems I: Regular Papers_ , vol. PP, no. 99, pp. 1–14, 2017. * [23] P. S. Reddy and K. Jyostna, “Development of smart insulin device for non invasive blood glucose level monitoring,” in _2017 IEEE 7th International Advance Computing Conference (IACC)_ , Jan 2017, pp. 516–519. * [24] P. Jain, A. M. Joshi, and S. P. Mohanty, “iGLU: An Intelligent Device for Accurate Non-Invasive Blood Glucose-Level Monitoring in Smart Healthcare,” _IEEE Consumer Electronics Magazine_ , vol. 9, no. 1, p. Accepted, January 2020. * [25] A. M. Joshi, P. Jain, S. P. Mohanty, and N. Agrawal, “iGLU 2.0: A new wearable for accurate non-invasive continuous serum glucose measurement in IoMT framework,” _IEEE Transactions on Consumer Electronics_ , no. 10.1109/TCE.2020.3011966, p. in Press, 2020. * [26] S. P. Mohanty and E. Kougianos, “Biosensors: A Tutorial Review,” _IEEE Potentials_ , vol. 25, no. 2, pp. 35–40, March 2006. * [27] N. A. Salam, W. H. M. Saad, Z. Manap, and F. Salehuddin, “The evolution of non-invasive blood glucose monitoring system for personal application,” _Journal of Telecommunication, Electronic and Computer Engineering_ , vol. 8, pp. 59–65, 2016. * [28] P. Jain, A. M. Joshi, and S. P. Mohanty, “iGLU 1.0: An Accurate Non-Invasive Near-Infrared Dual Short Wavelengths Spectroscopy based Glucometer for Smart Healthcare,” _arXiv Electrical Engineering and Systems Science_ , no. arXiv:1911.04471, November 2019. * [29] P. Jain, A. M. Joshi, N. Agrawal, and S. P. Mohanty, “iGLU 2.0: A New Non-invasive, Accurate Serum Glucometer for Smart Healthcare,” _arXiv Electrical Engineering and Systems Science_ , vol. abs/2001.09182, 2020\. [Online]. Available: http://arxiv.org/abs/2001.09182 * [30] Q. Liu, Y. Liu, F. Wu, X. Cao, Z. Li, M. Alharbi, A. N. Abbas, M. R. Amer, and C. Zhou, “Highly Sensitive and Wearable In2O3 Nanoribbon Transistor Biosensors with Integrated On-Chip Gate for Glucose Monitoring in Body Fluids,” _ACS Nano_ , vol. 12, no. 2, pp. 1170–1178, 2018, pMID: 29338249\. * [31] P. H. Siegel, W. Dai, R. A. Kloner, M. Csete, and V. Pikov, “First millimeter-wave animal in vivo measurements of l-glucose and d-glucose: Further steps towards a non-invasive glucometer,” in _2016 41st International Conference on Infrared, Millimeter, and Terahertz waves (IRMMW-THz)_ , Sept 2016, pp. 1–3. * [32] P. Jain, A. M. Joshi, and S. P. Mohanty, “iGLU: An Intelligent Device for Accurate Non-Invasive Blood Glucose-Level Monitoring in Smart Healthcare,” _IEEE Consumer Electronics Magazine_ , vol. 9, no. 1, pp. 35–42, January 2020\. * [33] N. M. Zhilo, P. A. Rudenko, and A. N. Zhigaylo, “Development of hardware-software test bench for optical non-invasive glucometer improvement,” in _2017 IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (EIConRus)_ , Feb 2017, pp. 89–90. * [34] S. I. Gusev, A. A. Simonova, P. S. Demchenko, M. K. Khodzitsky, and O. P. Cherkasova, “Blood glucose concentration sensing using biological molecules relaxation times determination,” in _2017 IEEE International Symposium on Medical Measurements and Applications (MeMeA)_ , May 2017, pp. 458–463. * [35] M. W. Sari and M. Luthfi, “Design and analysis of non-invasive blood glucose levels monitoring,” in _2016 International Seminar on Application for Technology of Information and Communication (ISemantic)_ , Aug 2016, pp. 134–137. * [36] S. Lekha and M. Suchetha, “Non- invasive diabetes detection and classification using breath analysis,” in _2015 International Conference on Communications and Signal Processing (ICCSP)_ , April 2015, pp. 0955–0958. * [37] J. Li, P. Koinkar, Y. Fuchiwaki, and M. Yasuzawa, “A fine pointed glucose oxidase immobilized electrode for low-invasive amperometric glucose monitoring,” _Biosensors and Bioelectronics_ , vol. 86, pp. 90–94, 2016\. * [38] N. Demitri and A. M. Zoubir, “Measuring blood glucose concentrations in photometric glucometers requiring very small sample volumes,” _IEEE Transactions on Biomedical Engineering_ , vol. 64, no. 1, pp. 28–39, 2017. * [39] J. Y. Lucisano, T. L. Routh, J. T. Lin, and D. A. Gough, “Glucose monitoring in individuals with diabetes using a long-term implanted sensor/telemetry system and model,” _IEEE Transactions on Biomedical Engineering_ , vol. 64, no. 9, pp. 1982–1993, 2017. * [40] A. Sun, A. G. Venkatesh, and D. A. Hall, “A multi-technique reconfigurable electrochemical biosensor: Enabling personal health monitoring in mobile devices,” _IEEE Transactions on Biomedical Circuits and Systems_ , vol. 10, no. 5, pp. 945–954, Oct 2016. * [41] A. Gani, A. V. Gribok, Y. Lu, W. K. Ward, R. A. Vigersky, and J. Reifman, “Universal glucose models for predicting subcutaneous glucose concentration in humans,” _IEEE Transactions on Information Technology in Biomedicine_ , vol. 14, no. 1, pp. 157–165, Jan 2010. * [42] G. Wang, M. D. Poscente, S. S. Park, C. N. Andrews, O. Yadid-Pecht, and M. P. Mintchev, “Wearable microsystem for minimally invasive, pseudo-continuous blood glucose monitoring: The e-mosquito,” _IEEE Transactions on Biomedical Circuits and Systems_ , vol. 11, no. 5, pp. 979–987, Oct 2017. * [43] M. M. Ahmadi and G. A. Jullien, “A wireless-implantable microsystem for continuous blood glucose monitoring,” _IEEE Transactions on Biomedical Circuits and Systems_ , vol. 3, no. 3, pp. 169–180, June 2009. * [44] G. Acciaroli, M. Vettoretti, A. Facchinetti, G. Sparacino, and C. Cobelli, “Reduction of blood glucose measurements to calibrate subcutaneous glucose sensors: A bayesian multiday framework,” _IEEE Transactions on Biomedical Engineering_ , vol. 65, no. 3, pp. 587–595, 2018. * [45] I. Pagkalos, P. Herrero, C. Toumazou, and P. Georgiou, “Bio-inspired glucose control in diabetes based on an analogue implementation of a $\beta$-cell model,” _IEEE Transactions on Biomedical Circuits and Systems_ , vol. 8, no. 2, pp. 186–195, April 2014. * [46] G. Wang, M. D. Poscente, S. S. Park, C. N. Andrews, O. Yadid-Pecht, and M. P. Mintchev, “Wearable microsystem for minimally invasive, pseudo-continuous blood glucose monitoring: The e-mosquito,” _IEEE Transactions on Biomedical Circuits and Systems_ , vol. 11, no. 5, pp. 979–987, 2017. * [47] T. Kossowski and R. Stasinski, “Robust ir attenuation measurement for non-invasive glucose level analysis,” in _2016 International Conference on Systems, Signals and Image Processing (IWSSIP)_ , May 2016, pp. 1–4. * [48] L. P. Pavlovich and D. Y. Mynziak, “Noninvasive method for blood glucose measuring and monitoring,” in _2013 IEEE XXXIII International Scientific Conference Electronics and Nanotechnology (ELNANO)_ , April 2013, pp. 255–257. * [49] Y. Liu, W. Li, T. Zheng, and W. K. Ling, “Overviews the methods of non-invasive blood glucose measurement,” in _2016 IEEE International Conference on Consumer Electronics-China (ICCE-China)_ , Dec 2016, pp. 1–2. * [50] N. K. Sharma and S. Singh, “Designing a non invasive blood glucose measurement sensor,” in _2012 IEEE 7th International Conference on Industrial and Information Systems (ICIIS)_ , Aug 2012, pp. 1–3. * [51] X. Zhao, Q. Zheng, and Z. M. Yang, “Two types of photonic crystals applied to glucose sensor,” in _2016 IEEE International Nanoelectronics Conference (INEC)_ , May 2016, pp. 1–2. * [52] Y. Tanaka, C. Purtill, T. Tajima, M. Seyama, and H. Koizumi, “Sensitivity improvement on cw dual-wavelength photoacoustic spectroscopy using acoustic resonant mode for noninvasive glucose monitor,” in _2016 IEEE SENSORS_ , Oct 2016, pp. 1–3. * [53] I. Gouzouasis, H. Cano-Garcia, I. Sotiriou, S. Saha, G. Palikaras, P. Kosmas, and E. Kallos, “Detection of varying glucose concentrations in water solutions using a prototype biomedical device for millimeter-wave non-invasive glucose sensing,” in _2016 10th European Conference on Antennas and Propagation (EuCAP)_ , April 2016, pp. 1–4. * [54] Y. Nikawa and D. Someya, “Non-invasive measurement of blood sugar level by millimeter waves,” in _2001 IEEE MTT-S International Microwave Sympsoium Digest (Cat. No.01CH37157)_ , vol. 1, May 2001, pp. 171–174 vol.1. * [55] J. Shao, F. Yang, F. Xia, Q. Zhang, and Y. Chen, “A novel miniature spiral sensor for non-invasive blood glucose monitoring,” in _2016 10th European Conference on Antennas and Propagation (EuCAP)_ , April 2016, pp. 1–2. * [56] P. H. Siegel, Y. Lee, and V. Pikov, “Millimeter-wave non-invasive monitoring of glucose in anesthetized rats,” in _2014 39th International Conference on Infrared, Millimeter, and Terahertz waves (IRMMW-THz)_ , Sept 2014, pp. 1–2. * [57] D. Wang, “An improved integration sensor of non-invasive blood glucose,” in _The 7th IEEE/International Conference on Advanced Infocomm Technology_ , Nov 2014, pp. 70–75. * [58] N. Bayasi, H. Saleh, B. Mohammad, and M. Ismail, “The revolution of glucose monitoring methods and systems: A survey,” in _2013 IEEE 20th International Conference on Electronics, Circuits, and Systems (ICECS)_ , Dec 2013, pp. 92–93. * [59] R. Agrawal, N. Sharma, M. Rathore, V. Gupta, S. Jain, V. Agarwal, and S. Goyal, “Noninvasive method for glucose level estimation by saliva,” _J Diabetes Metab_ , vol. 4, no. 5, pp. 2–5, 2013. * [60] N. Demitri and A. M. Zoubir, “Measuring blood glucose concentrations in photometric glucometers requiring very small sample volumes,” _IEEE Transactions on Biomedical Engineering_ , vol. 64, no. 1, pp. 28–39, 2016. * [61] M. Shokrekhodaei and S. Quinones, “Review of Non-invasive Glucose Sensing Techniques: Optical, Electrical and Breath Acetone,” _Sensors_ , vol. 20, no. 5, p. 1251, 2020. * [62] S. Delbeck, T. Vahlsing, S. Leonhardt, G. S. G, and H. M. Heise, “Non-invasive monitoring of blood glucose using optical methods for skin spectroscopy-opportunities and recent advances,” _Anal Bioanal Chem._ , vol. 411, no. 1, pp. 63–77, 2019. * [63] N. K. Madzhi, S. A. Shamsuddin, and M. F. Abdullah, “Comparative investigation using gaas(950nm), gaaias (940nm) and ingaasp (1450nm) sensors for development of non-invasive optical blood glucose measurement system,” in _2014 IEEE International Conference on Smart Instrumentation, Measurement and Applications (ICSIMA)_ , Nov 2014, pp. 1–6. * [64] N. A. M. Aziz, N. Arsad, P. S. Menon, A. R. Laili, M. H. Laili, and A. A. A. Halim, “Analysis of difference light sources for non-invasive aqueous glucose detection,” in _2014 IEEE 5th International Conference on Photonics (ICP)_ , Sept 2014, pp. 150–152. * [65] S. Tommasone, “Infrared spectroscopy: An overview,” 2018, last Accessed on 18 Jan 2021. [Online]. Available: https://www.azolifesciences.com/article/Infrared-Spectroscopy-An-Overview.aspx * [66] A. A. Muley and R. B. Ghongade, “Design and simulate an antenna for aqueous glucose measurement,” in _2014 Annual IEEE India Conference (INDICON)_ , Dec 2014, pp. 1–6. * [67] K. U. Menon, D. Hemachandran, and A. T. Kunnath, “Voltage intensity based non-invasive blood glucose monitoring,” in _2013 Fourth International Conference on Computing, Communications and Networking Technologies (ICCCNT)_. IEEE, 2013, pp. 1–5. * [68] J. L. Lai, S. Y. Huang, R. S. Lin, and S. C. Tsai, “Design a non-invasive near-infrared led blood glucose sensor,” in _2016 International Conference on Applied System Innovation (ICASI)_ , May 2016, pp. 1–4. * [69] M. Tamilselvi and G. Ramkumar, “Non-invasive tracking and monitoring glucose content using near infrared spectroscopy,” in _2015 IEEE International Conference on Computational Intelligence and Computing Research (ICCIC)_ , Dec 2015, pp. 1–3. * [70] K. Lawand, M. Parihar, and S. N. Patil, “Design and development of infrared led based non invasive blood glucometer,” in _2015 Annual IEEE India Conference (INDICON)_ , Dec 2015, pp. 1–6. * [71] J. Yadav, A. Rani, V. Singh, and B. M. Murari, “Near-infrared led based non-invasive blood glucose sensor,” in _2014 International Conference on Signal Processing and Integrated Networks (SPIN)_ , Feb 2014, pp. 591–594. * [72] M. T. B. Z. Abidin, M. K. R. Rosli, S. A. Shamsuddin, N. K. Madzhi, and M. F. Abdullah, “Initial quantitative comparison of 940nm and 950nm infrared sensor performance for measuring glucose non-invasively,” in _2013 IEEE International Conference on Smart Instrumentation, Measurement and Applications (ICSIMA)_ , Nov 2013, pp. 1–6. * [73] Y. Nikawa and T. Michiyama, “Non-invasive measurement of blood-sugar level by reflection of millimeter-waves,” in _2006 Asia-Pacific Microwave Conference_ , Dec 2006, pp. 47–50. * [74] M. Goodarzi and W. Saeys, “Selection of the most informative near infrared spectroscopy wavebands for continuous glucose monitoring in human serum,” _Talanta_ , vol. 146, pp. 155–165, 2016. * [75] S. Sharma, M. Goodarzi, L. Wynants, H. Ramon, and W. Saeys, “Efficient use of pure component and interferent spectra in multivariate calibration,” _Analytica chimica acta_ , vol. 778, pp. 15–23, 2013. * [76] Y. Uwadaira, N. Adachi, A. Ikehata, and S. Kawano, “Factors affecting the accuracy of non-invasive blood glucose measurement by short-wavelength near infrared spectroscopy in the determination of the glycaemic index of foods,” _Journal of Near Infrared Spectroscopy_ , vol. 18, no. 5, pp. 291–300, 2010\. * [77] S. Haxha and J. Jhoja, “Optical based noninvasive glucose monitoring sensor prototype,” _IEEE Photonics Journal_ , vol. 8, no. 6, pp. 1–11, 2016. * [78] W. Zhang, R. Liu, W. Zhang, H. Jia, and K. Xu, “Discussion on the validity of nir spectral data in non-invasive blood glucose sensing,” _Biomedical optics express_ , vol. 4, no. 6, pp. 789–802, 2013. * [79] M. Golic, K. Walsh, and P. Lawson, “Short-wavelength near-infrared spectra of sucrose, glucose, and fructose with respect to sugar concentration and temperature,” _Applied spectroscopy_ , vol. 57, no. 2, pp. 139–145, 2003\. * [80] K. Song, U. Ha, S. Park, J. Bae, and H. J. Yoo, “An impedance and multi-wavelength near-infrared spectroscopy ic for non-invasive blood glucose estimation,” _IEEE Journal of Solid-State Circuits_ , vol. 50, no. 4, pp. 1025–1037, April 2015. * [81] S. Ramasahayam, K. S. Haindavi, and S. R. Chowdhury, “Noninvasive estimation of blood glucose concentration using near infrared optodes,” in _Sensing Technology: Current Status and Future Trends IV_. Springer, 2015, pp. 67–82. * [82] A. Heller, “Integrated medical feedback systems for drug delivery,” _AIChE journal_ , vol. 51, no. 4, pp. 1054–1066, 2005. * [83] P. P. Pai, P. K. Sanki, A. De, and S. Banerjee, “NIR photoacoustic spectroscopy for non-invasive glucose measurement,” in _2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)_. IEEE, 2015, pp. 7978–7981. * [84] P. Jain, A. M. Joshi, and S. P. Mohanty, “iGLU: An intelligent device for accurate noninvasive blood glucose-level monitoring in smart healthcare,” _IEEE Consumer Electronics Magazine_ , vol. 9, no. 1, pp. 35–42, 2019. * [85] P. Jain, R. Maddila, and A. M. Joshi, “A precise non-invasive blood glucose measurement system using NIR spectroscopy and Huber’s regression model,” _Optical and Quantum Electronics_ , vol. 51, no. 2, p. 51, 2019. * [86] E. Monte-Moreno, “Non-invasive estimate of blood glucose and blood pressure from a photoplethysmograph by means of machine learning techniques,” _Artificial intelligence in medicine_ , vol. 53, no. 2, pp. 127–138, 2011\. * [87] S. Habbu, M. Dale, and R. Ghongade, “Estimation of blood glucose by non-invasive method using photoplethysmography,” _Sādhanā_ , vol. 44, no. 6, p. 135, 2019. * [88] H. Ali, F. Bensaali, and F. Jaber, “Novel approach to non-invasive blood glucose monitoring based on transmittance and refraction of visible laser light,” _IEEE Access_ , vol. 5, pp. 9163–9174, 2017. * [89] C. Vrančić, A. Fomichova, N. Gretz, C. Herrmann, S. Neudecker, A. Pucci, and W. Petrich, “Continuous glucose monitoring by means of mid-infrared transmission laser spectroscopyin vitro,” _Analyst_ , vol. 136, pp. 1192–1198, 2011. [Online]. Available: http://dx.doi.org/10.1039/C0AN00537A * [90] B. Paul, M. P. Manuel, and Z. C. Alex, “Design and development of non invasive glucose measurement system,” in _2012 1st International Symposium on Physics and Technology of Sensors (ISPTS-1)_. IEEE, 2012, pp. 43–46. * [91] L. A. Philip, K. Rajasekaran, and E. S. J. Jothi, “Continous monitoring of blood glucose using photophlythesmograph signal,” in _2017 International Conference on Innovations in Electrical, Electronics, Instrumentation and Media Technology (ICEEIMT)_. IEEE, 2017, pp. 187–191. * [92] H. Karimipour, H. T. Shandiz, and E. Zahedi, “Diabetic diagnose test based on ppg signal and identification system,” _Journal of Biomedical Science and Engineering_ , vol. 2, no. 06, p. 465, 2009. * [93] F. R. G. Cruz, C. C. Paglinawan, C. N. V. Catindig, J. C. B. Lamchek, D. D. C. Almiranez, and A. F. Sanchez, “Application of reflectance mode photoplethysmography for non-invasive monitoring of blood glucose level with moving average filter,” in _Proceedings of the 2019 9th International Conference on Biomedical Engineering and Technology_. ACM, 2019, pp. 22–26. * [94] Y. Zhang, Y. Zhang, S. A. Siddiqui, and A. Kos, “Non-invasive blood-glucose estimation using smartphone ppg signals and subspace knn classifier,” _Elektrotehniski Vestnik_ , vol. 86, no. 1/2, pp. 68–74, 2019. * [95] Y. Yamakoshi, K. Matsumura, T. Yamakoshi, J. Lee, P. Rolfe, Y. Kato, K. Shimizu, and K.-i. Yamakoshi, “Side-scattered finger-photoplethysmography: experimental investigations toward practical noninvasive measurement of blood glucose,” _Journal of biomedical optics_ , vol. 22, no. 6, p. 067001, 2017. * [96] O. Olarte, W. V. Moer, K. Barbé, Y. V. Ingelgem, and A. Hubin, “Influence of the type and position of the sensor on the precision of impedance glucose measurements,” in _2013 IEEE International Instrumentation and Measurement Technology Conference (I2MTC)_ , May 2013, pp. 1750–1754. * [97] S. K. Dhar, P. Biswas, and S. Chakraborty, “Dc impedance of human blood using eis: An appraoch to non-invasive blood glucose measurement,” in _2013 International Conference on Informatics, Electronics and Vision (ICIEV)_ , May 2013, pp. 1–6. * [98] Y. Khawam, M. Ali, H. Shazada, S. Kanan, and H. Nashash, “Non-invasive blood glucose measurement using transmission spectroscopy,” in _2013 1st International Conference on Communications, Signal Processing, and their Applications (ICCSPA)_ , Feb 2013, pp. 1–4. * [99] M. N. Anas and P. K. Lim, “A bio-impedance approach,” in _2013 IEEE International Conference on Smart Instrumentation, Measurement and Applications (ICSIMA)_ , Nov 2013, pp. 1–5. * [100] C. E. F. Amaral and B. Wolf, “Effects of glucose in blood and skin impedance spectroscopy,” in _AFRICON 2007_ , Sept 2007, pp. 1–7. * [101] P. Jain and A. M. Joshi, “Low leakage and high cmrr cmos differential amplifier for biomedical application,” _Analog Integrated Circuits and Signal Processing_ , pp. 1–15, 2017. * [102] B. Paul, M. P. Manuel, and Z. C. Alex, “Design and development of non invasive glucose measurement system,” in _2012 1st International Symposium on Physics and Technology of Sensors (ISPTS-1)_ , March 2012, pp. 43–46. * [103] M. Hofmann, M. Bloss, R. Weigel, G. Fischer, and D. Kissinger, “Non-invasive glucose monitoring using open electromagnetic waveguides,” in _2012 42nd European Microwave Conference_ , Oct 2012, pp. 546–549. * [104] Y. Liu, M. Xia, Z. Nie, J. Li, Y. Zeng, and L. Wang, “In vivo wearable non-invasive glucose monitoring based on dielectric spectroscopy,” in _2016 IEEE 13th International Conference on Signal Processing (ICSP)_ , Nov 2016, pp. 1388–1391. * [105] M. Yamaguchi, M. Mitsumori, and Y. Kano, “Noninvasively measuring blood glucose using saliva,” _IEEE Engineering in Medicine and Biology Magazine_ , vol. 17, no. 3, pp. 59–63, May 1998. * [106] M. S. Prasad, R. Chen, Y. Li, D. Rekha, D. Li, H. Ni, and N. Y. Sreedhar, “Polypyrrole supported with copper nanoparticles modified alkali anodized steel electrode for probing of glucose in real samples,” _IEEE Sensors Journal_ , vol. 18, no. 13, pp. 5203–5212, July 2018. * [107] G. Yoon, K. J. Jeon, A. K. Amerov, Y.-J. Kim, D. Y. Hwang, J. B. Kim, and H. S. Kim, “Non-invasive monitoring of blood glucose,” in _Lasers and Electro-Optics, 1999. CLEO/Pacific Rim ’99. The Pacific Rim Conference on_ , vol. 4, Aug 1999, pp. 1233–1234 vol.4. * [108] Y. Yamakoshi, M. Ogawa, T. Yamakoshi, M. Satoh, M. Nogawa, S. Tanaka, T. Tamura, P. Rolfe, and K. Yamakoshi, “A new non-invasive method for measuring blood glucose using instantaneous differential near infrared spectrophotometry,” in _2007 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society_ , Aug 2007, pp. 2964–2967. * [109] G. Yoon, A. K. Amerov, K. J. Jeon, J. B. Kim, and Y.-J. Kim, “Optical measurement of glucose levels in scattering media,” in _Proceedings of the 20th Annual International Conference of the IEEE Engineering in Medicine and Biology Society. Vol.20 Biomedical Engineering Towards the Year 2000 and Beyond (Cat. No.98CH36286)_ , vol. 4, Oct 1998, pp. 1897–1899 vol.4. * [110] H. Ishizawa, A. Muro, T. Takano, K. Honda, and H. Kanai, “Non-invasive blood glucose measurement based on atr infrared spectroscopy,” in _2008 SICE Annual Conference_ , Aug 2008, pp. 321–324. * [111] T. Harada, K. Yamamoto, M. Kondo, K. Gesho, and I. Ishimaru, “Spectroscpy optical coherence tomography of biomedical tissue,” in _SICE Annual Conference 2007_ , Sept 2007, pp. 3056–3059. * [112] A. Popov, A. Bykov, S. Toppari, M. Kinnunen, A. Priezzhev, and R. Myllylä, “Glucose sensing in flowing blood and intralipid by laser pulse time-of-flight and optical coherence tomography techniques,” _IEEE Journal of Selected Topics in Quantum Electronics_ , vol. 18, pp. 1335–1342, 07 2012. * [113] J. Y. Sim, C. G. Ahn, E. Jeong, and B. K. Kim, “Photoacoustic spectroscopy that uses a resonant characteristic of a microphone for in vitro measurements of glucose concentration,” in _2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)_ , Aug 2016, pp. 4861–4864. * [114] P. P. Pai, P. K. Sanki, and S. Banerjee, “A photoacoustics based continuous non-invasive blood glucose monitoring system,” in _2015 IEEE International Symposium on Medical Measurements and Applications (MeMeA) Proceedings_ , May 2015, pp. 106–111. * [115] P. P. Pai, P. K. Sanki, A. De, and S. Banerjee, “Nir photoacoustic spectroscopy for non-invasive glucose measurement,” in _2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)_ , Aug 2015, pp. 7978–7981. * [116] Y. Tanaka, Y. Higuchi, and S. Camou, “Noninvasive measurement of aqueous glucose solution at physiologically relevant blood concentration levels with differential continuous-wave laser photoacoustic technique,” in _2015 IEEE SENSORS_ , Nov 2015, pp. 1–4. * [117] L. Xiaoli and L. Chengwei, “Research on glucose concentration sensing with single wavelength laser,” in _2015 12th IEEE International Conference on Electronic Measurement Instruments (ICEMI)_ , vol. 03, July 2015, pp. 1547–1551. * [118] H. A. A. Naam, M. O. Idrees, A. Awad, O. S. Abdalsalam, and F. Mohamed, “Non invasive blood glucose measurement based on photo-acoustic spectroscopy,” in _2015 International Conference on Computing, Control, Networking, Electronics and Embedded Systems Engineering (ICCNEEE)_ , Sept 2015, pp. 1–4. * [119] S. Camou, Y. Ueno, and E. Tamechika, “New cw-photoacoustic-based protocol for noninvasive and selective determination of aqueous glucose level: A potential alternative towards noninvasive blood sugar sensing,” in _2011 IEEE SENSORS Proceedings_ , Oct 2011, pp. 798–801. * [120] N. Wadamori, R. Shinohara, and Y. Ishihara, “Photoacoustic depth profiling of a skin model for non-invasive glucose measurement,” in _2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society_ , Aug 2008, pp. 5644–5647. * [121] S. Koyama, Y. Miyauchi, T. Horiguchi, and H. Ishizawa, “Non-invasive measurement of blood glucose of diabetic based on ir spectroscopy,” in _Proceedings of SICE Annual Conference 2010_ , Aug 2010, pp. 3425–3426. * [122] P. Domachuk, M. Hunter, R. Batorsky, M. Cronin-Golomb, F. G. Omenetto, A. Wang, A. K. George, and J. C. Knight, “A path for non-invasive glucose detection using mid-ir supercontinuum,” in _2008 Conference on Lasers and Electro-Optics and 2008 Conference on Quantum Electronics and Laser Science_ , May 2008, pp. 1–2. * [123] R. Periyasamy and S. Anand, “A study on non-invasive blood glucose estimation- an approach using capacitance measurement technique,” in _2016 International Conference on Signal Processing, Communication, Power and Embedded System (SCOPES)_ , Oct 2016, pp. 847–850. * [124] T. Yilmaz, R. Foster, and Y. Hao, “Towards accurate dielectric property retrieval of biological tissues for blood glucose monitoring,” _IEEE Transactions on Microwave Theory and Techniques_ , vol. 62, no. 12, pp. 3193–3204, Dec 2014. * [125] M. Gourzi, A. Rouane, M. B. McHugh, R. Guelaz, and M. Nadi, “New biosensor for non-invasive glucose concentration measurement,” in _Proceedings of IEEE Sensors 2003 (IEEE Cat. No.03CH37498)_ , vol. 2, Oct 2003, pp. 1343–1347 Vol.2. * [126] V. Turgul and I. Kale, “On the accuracy of complex permittivity model of glucose/water solutions for non-invasive microwave blood glucose sensing,” in _2015 E-Health and Bioengineering Conference (EHB)_ , Nov 2015, pp. 1–4. * [127] D. Li, D. Yang, J. Yang, Y. Lin, Y. Sun, H. Yu, and K. Xu, “Glucose affinity measurement by surface plasmon resonance with borate polymer binding,” _Sensors and Actuators A: Physical_ , vol. 222, pp. 58 – 66, 2015. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0924424714004737 * [128] R. Kaul and U. P. Khot, “Design of microstrip antennas for glucometer application,” in _2016 IEEE International Conference on Advances in Electronics, Communication and Computer Technology (ICAECCT)_ , Dec 2016, pp. 352–357. * [129] V. Turgul and I. Kale, “Influence of fingerprints and finger positioning on accuracy of rf blood glucose measurement from fingertips,” _Electronics Letters_ , vol. 53, no. 4, pp. 218–220, 2017. * [130] H. Cano-Garcia, I. Gouzouasis, I. Sotiriou, S. Saha, G. Palikaras, P. Kosmas, and E. Kallos, “Reflection and transmission measurements using 60 ghz patch antennas in the presence of animal tissue for non-invasive glucose sensing,” in _2016 10th European Conference on Antennas and Propagation (EuCAP)_ , April 2016, pp. 1–3. * [131] S. Saha, I. Sotiriou, I. Gouzouasis, H. Cano-Garcia, G. Palikaras, P. Kosmas, and E. Kallos, “Evaluation of the sensitivity of transmission measurements at millimeter waves using patch antennas for non-invasive glucose sensing,” in _2016 10th European Conference on Antennas and Propagation (EuCAP)_ , April 2016, pp. 1–4. * [132] V. Turgul and I. Kale, “A novel pressure sensing circuit for non-invasive rf/microwave blood glucose sensors,” in _2016 16th Mediterranean Microwave Symposium (MMS)_ , Nov 2016, pp. 1–4. * [133] M. S. Ali, N. J. Shoumy, S. Khatun, L. M. Kamarudin, and V. Vijayasarveswari, “Non-invasive blood glucose measurement performance analysis through uwb imaging,” in _2016 3rd International Conference on Electronic Design (ICED)_ , Aug 2016, pp. 513–516. * [134] H. Choi, J. Nylon, S. Luzio, J. Beutler, and A. Porch, “Design of continuous non-invasive blood glucose monitoring sensor based on a microwave split ring resonator,” in _2014 IEEE MTT-S International Microwave Workshop Series on RF and Wireless Technologies for Biomedical and Healthcare Applications (IMWS-Bio2014)_ , Dec 2014, pp. 1–3. * [135] V. Turgul and I. Kale, “Simulating the effects of skin thickness and fingerprints to highlight problems with non-invasive rf blood glucose sensing from fingertips,” _IEEE Sensors Journal_ , vol. 17, no. 22, pp. 7553–7560, Nov 2017. * [136] Y. Miyauchi, T. Horiguchi, H. Ishizawa, S. i. Tezuka, and H. Hara, “Blood glucose level measurement by confocal reflection photodetection system,” in _SICE Annual Conference 2011_ , Sept 2011, pp. 2686–2689. * [137] D. W. Kim, H. S. Kim, D. H. Lee, and H. C. Kim, “Importance of skin resistance in the reverse iontophoresis-based noninvasive glucose monitoring system,” in _The 26th Annual International Conference of the IEEE Engineering in Medicine and Biology Society_ , vol. 1, Sept 2004, pp. 2434–2437. * [138] M. Hofmann, T. Fersch, R. Weigel, G. Fischer, and D. Kissinger, “A novel approach to non-invasive blood glucose measurement based on rf transmission,” in _2011 IEEE International Symposium on Medical Measurements and Applications_ , May 2011, pp. 39–42. * [139] K. Mitsubayashi, “Novel biosensing devices for medical applications soft contact-lens sensors for monitoring tear sugar,” in _2014 International Conference on Simulation of Semiconductor Processes and Devices (SISPAD)_ , Sept 2014, pp. 349–352. * [140] B. D. Cameron and G. L. Cote, “Polarimetric glucose sensing in aqueous humor utilizing digital closed-loop control,” in _Proceedings of 18th Annual International Conference of the IEEE Engineering in Medicine and Biology Society_ , vol. 1, Oct 1996, pp. 204–205 vol.1. * [141] R. J. Buford, E. C. Green, and M. J. McClung, “A microwave frequency sensor for non-invasive blood-glucose measurement,” in _2008 IEEE Sensors Applications Symposium_ , Feb 2008, pp. 4–7. * [142] S. Lee, V. Nayak, J. Dodds, M. Pishkou, and N. B. Smith, “Glucose measurements with sensors and ultrasound,” _Ultrasound in Medicine and Biology_ , vol. 31, no. 7, pp. 971–977, 2005. * [143] O. K. Cho, Y. O. Kim, H. Mitsumaki, and K. Kuwa, “Noninvasive Measurement of Glucose by Metabolic Heat Conformation Method,” _Clinical Chemistry_ , vol. 50, no. 10, pp. 1894–1898, 10 2004. [Online]. Available: https://doi.org/10.1373/clinchem.2004.036954 * [144] A. B. Blodgett, R. K. Kothinti, I. Kamyshko, D. H. Petering, S. Kumar, and N. M. Tabatabai, “A fluorescence method for measurement of glucose transport in kidney cells,” _Diabetes Technology & Therapeutics_, vol. 13, no. 7, pp. 743–751, 2011, pMID: 21510766. * [145] A. K. Amerov, Y. Sun, G. W. Small, and M. A. Arnold, “Kromoscopic measurement of glucose in the first overtone region of the near-infrared spectrum,” in _Optical Diagnostics and Sensing of Biological Fluids and Glucose and Cholesterol Monitoring II_ , A. V. Priezzhev and G. L. Cote, Eds., vol. 4624, International Society for Optics and Photonics. SPIE, 2002, pp. 11 – 19. [Online]. Available: https://doi.org/10.1117/12.468318 * [146] R. Zhang, S. Liu, H. Jin, Y. Luo, Z. Zheng, F. Gao, and Y. Zheng, “Noninvasive electromagnetic wave sensing of glucose,” _Sensors_ , vol. 19, no. 5, p. 1151, 2019. * [147] P. Bertemes-Filho, R. Weinert, T. Barato, and T. Baratto de Albuquerque, “Detection of glucose by using impedance spectroscopy,” in _Proc. International Conference on Electrical Bio-Impedance_ , 06 2016. * [148] R. O. Potts, J. A. Tamada, and M. J. Tierney, “Glucose monitoring by reverse iontophoresis,” _Diabetes/Metabolism Research and Reviews_ , vol. 18, no. S1, pp. S49–S53, 2002. [Online]. Available: https://onlinelibrary.wiley.com/doi/abs/10.1002/dmrr.210 * [149] R. Rao and S. Nanda, “Sonophoresis: recent advancements and future trends,” _Journal of Pharmacy and Pharmacology_ , vol. 61, no. 6, pp. 689–705, 2009\. [Online]. Available: https://onlinelibrary.wiley.com/doi/abs/10.1211/jpp.61.06.0001 * [150] O. Amir, D. Weinstein, S. Zilberman, M. Less, D. Perl-Treves, H. Primack, A. Weinstein, E. Gabis, B. Fikhte, and A. Karasik, “Continuous noninvasive glucose monitoring technology based on “occlusion spectroscopy”,” _Journal of Diabetes Science and Technology_ , vol. 1, no. 4, pp. 463–469, 2007, pMID: 19885108. * [151] B. M. Jensen, P. Bjerring, J. S. Christiansen, and H. ørskov, “Glucose content in human skin: relationship with blood glucose levels,” _Scandinavian Journal of Clinical and Laboratory Investigation_ , vol. 55, no. 5, pp. 427–432, 1995. * [152] T. Kossowski and R. Stasiński, “Multi-wavelength analysis of substances levels in human blood,” in _2017 International Conference on Systems, Signals and Image Processing (IWSSIP)_ , May 2017, pp. 1–4. * [153] A. Ficorella, A. D’Amico, M. Santonico, G. Pennazza, S. Grasso, and A. Zompanti, “A multi-frequency system for glucose detection with optical sensors,” in _2015 XVIII AISEM Annual Conference_ , Feb 2015, pp. 1–3. * [154] K. Song, U. Ha, S. Park, and H.-J. Yoo, “An impedance and multi-wavelength near-infrared spectroscopy ic for non-invasive blood glucose estimation,” in _2014 Symposium on VLSI Circuits Digest of Technical Papers_ , June 2014, pp. 1–2. * [155] E. L. Litinskaia, N. A. Bazaev, K. V. Pozhar, and V. M. Grinvald, “Methods for improving accuracy of non-invasive blood glucose detection via optical glucometer,” in _2017 IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (EIConRus)_ , Feb 2017, pp. 47–49. * [156] P. Jain and S. Akashe, “An innovative design: Mos based full-wave centre-tapped rectifier,” _Wireless Personal Communications_ , vol. 90, no. 4, pp. 1673–1693, 2016. * [157] P. JAIN and S. AKASHE, “Performance analysis of analog to digital converter with augmented voltage swing boost logic cum schmitt trigger mos switches configuration.” _Journal of Active & Passive Electronic Devices_, vol. 11, 2016. * [158] P. Jain and S. Akashe, “Analyzing the impact of bootstrapped adc with augmented nmos sleep transistors configuration on performance parameters,” _Circuits, Systems, and Signal Processing_ , vol. 33, no. 7, pp. 2009–2025, 2014. * [159] P.Jain and S. Akashe, “Design and optimization of flash type analog to digital converter using augmented sleep transistors with current mode logic,” _Radioelectronics and Communications Systems_ , vol. 56, no. 10, pp. 472–480, 2013. * [160] A. K. Singh and S. K. Jha, “Fabrication and validation of a handheld non-invasive, optical biosensor for self-monitoring of glucose using saliva,” _IEEE Sensors Journal_ , vol. 19, no. 18, pp. 8332–8339, Sep. 2019\. * [161] P. P. Pai, A. De, and S. Banerjee, “Accuracy enhancement for noninvasive glucose estimation using dual-wavelength photoacoustic measurements and kernel-based calibration,” _IEEE Transactions on Instrumentation and Measurement_ , vol. 67, no. 1, pp. 126–136, 2018. * [162] T. Dai and A. Adler, “In vivo blood characterization from bioimpedance spectroscopy of blood pooling,” _IEEE Transactions on Instrumentation and Measurement_ , vol. 58, no. 11, p. 3831, 2009. * [163] R. D. Beach, R. W. Conlan, M. C. Godwin, and F. Moussy, “Towards a miniature implantable in vivo telemetry monitoring system dynamically configurable as a potentiostat or galvanostat for two- and three-electrode biosensors,” _IEEE Transactions on Instrumentation and Measurement_ , vol. 54, no. 1, pp. 61–72, Feb 2005. * [164] P. Jain, S. Pancholi, and A. M. Joshi, “An iomt based non-invasive precise blood glucose measurement system,” in _2019 IEEE International Symposium on Smart Electronic Systems (iSES)(Formerly iNiS)_. IEEE, 2019, pp. 111–116. * [165] B. A. Malik, A. Naqash, and G. M. Bhat, “Backpropagation artificial neural network for determination of glucose concentration from near-infrared spectra,” in _2016 International Conference on Advances in Computing, Communications and Informatics (ICACCI)_ , Sept 2016, pp. 2688–2691. * [166] D. Yotha, C. Pidthalek, S. Yimman, and S. Niramitmahapanya, “Design and construction of the hypoglycemia monito wireless system for diabetic,” in _2016 9th Biomedical Engineering International Conference (BMEiCON)_ , Dec 2016, pp. 1–4. * [167] S. Sarangi, P. P. Pai, P. K. Sanki, and S. Banerjee, “Comparative analysis of golay code based excitation and coherent averaging for non-invasive glucose monitoring system,” in _2014 IEEE 27th International Symposium on Computer-Based Medical Systems_ , May 2014, pp. 485–486. * [168] S. R. Naqvi, N. Z. Azeemi, A. Hameed, R. Baddar, and T. Rasool, “Improving accuracy of non-invasive glucose monitoring through non-local data denoising,” in _2008 Cairo International Biomedical Engineering Conference_ , Dec 2008, pp. 1–4. * [169] Y. Yamakoshi, M. Ogawa, T. Yamakoshi, T. Tamura, and K. i. Yamakoshi, “Multivariate regression and discreminant calibration models for a novel optical non-invasive blood glucose measurement method named pulse glucometry,” in _2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society_ , Sept 2009, pp. 126–129. * [170] C. Z. Ming, P. Raveendran, and P. S. Chew, “A comparison analysis between partial least squares and neural network in non-invasive blood glucose concentration monitoring system,” in _2009 International Conference on Biomedical and Pharmaceutical Engineering_ , Dec 2009, pp. 1–4. * [171] S. Sarangi, P. P. Pai, P. K. Sanki, and S. Banerjee, “Comparative analysis of golay code based excitation and coherent averaging for non-invasive glucose monitoring system,” in _2014 IEEE 27th International Symposium on Computer-Based Medical Systems_ , May 2014, pp. 485–486. * [172] H. M. Heise, “Technology for non-invasive monitoring of glucose,” in _Proceedings of 18th Annual International Conference of the IEEE Engineering in Medicine and Biology Society_ , vol. 5, Oct 1996, pp. 2159–2161 vol.5. * [173] P. P. Pai, S. Bhattacharya, and S. Banerjee, “Regularized least squares regression for calibration of a photoacoustic spectroscopy based non-invasive glucose monitoring system,” in _2015 IEEE International Ultrasonics Symposium (IUS)_ , Oct 2015, pp. 1–4. * [174] M. Stemmann, F. Ståhl, J. Lallemand, E. Renard, and R. Johansson, “Sensor calibration models for a non-invasive blood glucose measurement sensor,” in _2010 Annual International Conference of the IEEE Engineering in Medicine and Biology_ , Aug 2010, pp. 4979–4982. * [175] D. K. Rollins, K. Kotz, and C. Stiehl, “Non-invasive glucose monitoring from measured inputs,” in _UKACC International Conference on Control 2010_ , Sept 2010, pp. 1–5. * [176] B. A. Malik, “Determination of glucose concentration from near infrared spectra using least square support vector machine,” in _2015 International Conference on Industrial Instrumentation and Control (ICIC)_ , May 2015, pp. 475–478. * [177] M. Ogawa, Y. Yamakoshi, M. Satoh, M. Nogawa, T. Yamakoshi, S. Tanaka, P. Rolfe, T. Tamura, and K. i. Yamakoshi, “Support vector machines as multivariate calibration model for prediction of blood glucose concentration using a new non-invasive optical method named pulse glucometry,” in _2007 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society_ , Aug 2007, pp. 4561–4563. * [178] Z. T. Dag, E. Koklukaya, F. Temurtas, H. M. Saraoglu, and S. Altikat, “Detection of the blood glucose and haemoglobin a1c with palm perspiration by using artificial neural networks,” in _2011 7th International Conference on Electrical and Electronics Engineering (ELECO)_ , Dec 2011, pp. II–302–II–305. * [179] O. Olarte, W. V. Moer, K. Barbé, Y. V. Ingelgem, and A. Hubin, “Using random phase multisines to perform non-invasive glucose measurements,” in _2011 IEEE International Symposium on Medical Measurements and Applications_ , May 2011, pp. 300–304. * [180] M. B. Savage, S. Kun, H. Harjunmaa, and R. A. Peura, “Development of a non-invasive blood glucose monitor: application of artificial neural networks for signal processing,” in _Proceedings of the IEEE 26th Annual Northeast Bioengineering Conference (Cat. No.00CH37114)_ , 2000, pp. 29–30. * [181] R. Baghbani, M. A. Rad, and A. Pourziad, “Microwave sensor for non-invasive glucose measurements design and implementation of a novel linear,” _IET Wireless Sensor Systems_ , vol. 5, no. 2, pp. 51–57, 2015. * [182] J. S. Parab, R. S. Gad, and G. M. Naik, “Influence of pca components on glucose prediction using non-invasive technique,” in _2016 International Conference on Advances in Electrical, Electronic and Systems Engineering (ICAEES)_ , Nov 2016, pp. 473–476. * [183] T. R. J. C. Lekha and C. S. Kumar, “Nir spectroscopic algorithm development for glucose detection,” in _2015 International Conference on Innovations in Information, Embedded and Communication Systems (ICIIECS)_ , March 2015, pp. 1–6. * [184] O. Olarte, W. V. Moer, K. Barbé, S. Verguts, Y. V. Ingelgem, and A. Hubin, “Using the best linear approximation as a first step to a new non-invasive glucose measurement,” in _2012 IEEE International Instrumentation and Measurement Technology Conference Proceedings_ , May 2012, pp. 2747–2751. * [185] W. L. Clarke, “The original Clarke error grid analysis (EGA),” _Diabetes Technology & Therapeutics_, vol. 7, no. 5, pp. 776–779, 2005\. * [186] C. Fernandez, “Needle-free diabetes care: 7 devices that painlessly measure blood glucose,” _Labiotech_ , vol. 23, 2018. * [187] M. F. Schemmann and T. O’brien, “Blood glucose sensor,” Apr. 11 2013, uS Patent App. 13/646,721. * [188] A. Gal, I. Harman-Boehm, E. Naidis, Y. Mayzel, and L. Trieman, “Validity of glucotrack®, a non-invasive glucose monitor, for variety of diabetics,” _Age_ , vol. 1, no. 295, p. 61, 2011. * [189] D. Huber, L. Falco-Jonasson, M. Talary, F. Dewarrat, A. Caduff, W. Stahel, and N. Stadler, “Multi-sensor data fusion for non-invasive continuous glucose monitoring,” in _2007 10th International Conference on Information Fusion_ , July 2007, pp. 1–10. * [190] V. W. Bolie, “Coefficients of normal blood glucose regulation,” _Journal of applied physiology_ , vol. 16, no. 5, pp. 783–788, 1961. * [191] R. N. Bergman, L. S. Phillips, and C. Cobelli, “Physiologic evaluation of factors controlling glucose tolerance in man: measurement of insulin sensitivity and beta-cell glucose sensitivity from the response to intravenous glucose.” _The Journal of clinical investigation_ , vol. 68, no. 6, pp. 1456–1467, 1981. * [192] A. De Gaetano and O. Arino, “Mathematical modelling of the intravenous glucose tolerance test,” _Journal of mathematical biology_ , vol. 40, no. 2, pp. 136–168, 2000. * [193] C. Cobelli, C. Dalla Man, G. Sparacino, L. Magni, G. De Nicolao, and B. P. Kovatchev, “Diabetes: models, signals, and control,” _IEEE reviews in biomedical engineering_ , vol. 2, pp. 54–96, 2009. * [194] R. Hovorka, F. Shojaee-Moradie, P. V. Carroll, L. J. Chassin, I. J. Gowrie, N. C. Jackson, R. S. Tudor, A. M. Umpleby, and R. H. Jones, “Partitioning glucose distribution/transport, disposal, and endogenous production during ivgtt,” _American Journal of Physiology-Endocrinology and Metabolism_ , vol. 282, no. 5, pp. E992–E1007, 2002. * [195] A. Haidar, M. E. Wilinska, J. A. Graveston, and R. Hovorka, “Stochastic virtual population of subjects with type 1 diabetes for the assessment of closed-loop glucose controllers,” _IEEE Transactions on Biomedical Engineering_ , vol. 60, no. 12, pp. 3524–3533, 2013. * [196] H. Kirchsteiger, G. C. Estrada, S. Pölzer, E. Renard, and L. del Re, “Estimating interval process models for type 1 diabetes for robust control design,” _IFAC Proc. Volumes_ , vol. 44, no. 1, pp. 11 761–11 766, 2011\. * [197] N. Magdelaine, L. Chaillous, I. Guilhem, J.-Y. Poirier, M. Krempf, C. H. Moog, and E. Le Carpentier, “A long-term model of the glucose–insulin dynamics of type 1 diabetes,” _IEEE Transactions on Biomedical Engineering_ , vol. 62, no. 6, pp. 1546–1552, 2015. * [198] K. Turksoy, S. Samadi, J. Feng, E. Littlejohn, L. Quinn, and A. Cinar, “Meal detection in patients with type 1 diabetes: a new module for the multivariable adaptive artificial pancreas control system,” _IEEE journal of biomedical and health informatics_ , vol. 20, no. 1, pp. 47–54, 2015\. * [199] J. Xie and Q. Wang, “A variable state dimension approach to meal detection and meal size estimation: in silico evaluation through basal-bolus insulin therapy for type 1 diabetes,” _IEEE Transactions on Biomedical Engineering_ , vol. 64, no. 6, pp. 1249–1260, 2016. * [200] T. MohammadRidha, M. Aït-Ahmed, L. Chaillous, M. Krempf, I. Guilhem, J.-Y. Poirier, and C. H. Moog, “Model free ipid control for glycemia regulation of type-1 diabetes,” _IEEE Transactions on Biomedical Engineering_ , vol. 65, no. 1, pp. 199–206, 2017. * [201] S. Kino, S. Omori, T. Katagiri, and Y. Matsuura, “Hollow optical-fiber based infrared spectroscopy for measurement of blood glucose level by using multi-reflection prism,” _Biomedical optics express_ , vol. 7, no. 2, pp. 701–708, 2016. * [202] T.-L. Chen, Y.-L. Lo, C.-C. Liao, and Q.-H. Phan, “Noninvasive measurement of glucose concentration on human fingertip by optical coherence tomography,” _Journal of biomedical optics_ , vol. 23, no. 4, p. 047001, 2018. * [203] J. Park, J. Kim, S.-Y. Kim, W. H. Cheong, J. Jang, Y.-G. Park, K. Na, Y.-T. Kim, J. H. Heo, C. Y. Lee _et al._ , “Soft, smart contact lenses with integrations of wireless circuits, glucose sensors, and displays,” _Science Advances_ , vol. 4, no. 1, p. eaap9841, 2018. * [204] A. K. Singh and S. K. Jha, “Fabrication and validation of a handheld non-invasive, optical biosensor for self-monitoring of glucose using saliva,” _IEEE Sensors Journal_ , vol. 19, no. 18, pp. 8332–8339, 2019. * [205] K. Song, U. Ha, S. Park, J. Bae, and H.-J. Yoo, “An impedance and multi-wavelength near-infrared spectroscopy ic for non-invasive blood glucose estimation,” _IEEE Journal of solid-state circuits_ , vol. 50, no. 4, pp. 1025–1037, 2015. * [206] Q. Li, X. Xiao, and T. Kikkawa, “Absorption spectrum for non-invasive blood glucose concentration detection by microwave signals,” _Journal of Electromagnetic Waves and Applications_ , vol. 33, no. 9, pp. 1093–1106, 2019\. * [207] V. P. Rachim and W.-Y. Chung, “Wearable-band type visible-near infrared optical biosensor for non-invasive blood glucose monitoring,” _Sensors and Actuators B: Chemical_ , vol. 286, pp. 173–180, 2019. * [208] S. P. Mohanty, U. Choppali, and E. Kougianos, “Everything you wanted to know about smart cities: The Internet of things is the backbone,” _IEEE Consumer Electronics Magazine_ , vol. 5, no. 3, pp. 60–70, July 2016\. * [209] P. Chanak and I. Banerjee, “Internet of things-enabled smart villages: Recent advances and challenges,” _IEEE Consumer Electronics Magazine_ , pp. 1–1, 2020. * [210] C. P. Antonopoulos, G. Keramidas, N. S. Voros, M. Huebner, F. Schwiegelshohn, D. Goehringer, M. Dagioglou, G. Stavrinos, S. Konstantopoulos, and V. Karkaletsis, “Toward an ict-based service oriented health care paradigm,” _IEEE Consumer Electronics Magazine_ , vol. 9, no. 4, pp. 77–82, 2020. * [211] M. Aazam, S. Zeadally, and K. A. Harras, “Health fog for smart healthcare,” _IEEE Consumer Electronics Magazine_ , vol. 9, no. 2, pp. 96–102, 2020. * [212] A. M. Joshi, P. Jain, and S. P. Mohanty, “Secure-iGLU: A secure device for noninvasive glucose measurement and automatic insulin delivery in iomt framework,” in _2020 IEEE Computer Society Annual Symposium on VLSI (ISVLSI)_. IEEE, 2020, pp. 440–445. * [213] L. Rachakonda, S. P. Mohanty, and E. Kougianos, “iLog: An intelligent device for automatic food intake monitoring and stress detection in the iomt,” _IEEE Transactions on Consumer Electronics_ , vol. 66, no. 2, pp. 115–124, 2020. * [214] V. P. Yanambaka, S. P. Mohanty, E. Kougianos, and D. Puthal, “PMsec: Physical unclonable function-based robust and lightweight authentication in the internet of medical things,” _IEEE Transactions on Consumer Electronics_ , vol. 65, no. 3, pp. 388–397, 2019. * [215] L. Rachakonda, A. K. Bapatla, S. P. Mohanty, and E. Kougianos, “SaYoPillow: Blockchain-integrated privacy-assured iomt framework for stress management considering sleeping habits,” _IEEE Transactions on Consumer Electronics_ , pp. 1–1, 2020. * [216] S. P. Mohanty, V. P. Yanambaka, E. Kougianos, and D. Puthal, “PUFchain: A hardware-assisted blockchain for sustainable simultaneous device and data security in the internet of everything (IoE),” _IEEE Consumer Electronics Magazine_ , vol. 9, no. 2, pp. 8–16, 2020. ## About the Authors | Prateek Jain (Member, IEEE) earned his B.E. degree in Electronics Engineering from Jiwaji University, India in 2010 and Master degree from ITM University Gwalior. Currently, he is an Assistant Professor in SENSE, VIT University, Amaravati (A.P.). His current research interest includes VLSI design, Biomedical Systems and Instrumentation. He is an author of 14 peer- reviewed publications. He is a regular reviewer of 12 journals and 10 conferences. ---|--- | Amit M. Joshi (Member, IEEE) received the Ph.D. degree from the NIT, Surat, India. He is currently an Assistant Professor at National Institute of Technology, Jaipur. His area of specialization is Biomedical signal processing, Smart healthcare, VLSI DSP Systems and embedded system design. He has also published papers in international peer reviewed journals with high impact factors. He has published six book chapters and also published more than 70 research articles in excellent peer reviewed international journals/conferences. He has worked as a reviewer of technical journals such as IEEE Transactions/ IEEE Access, Springer, Elsevier and also served as Technical Programme Committee member for IEEE conferences which are related to biomedical field. He also received honour of UGC Travel fellowship, the award of SERB DST Travel grant and CSIR fellowship to attend well known IEEE Conferences TENCON, ISCAS, MENACOMM etc across the world. He has served as session chair at various IEEE Conferences like TENCON -2016, iSES-2018, iSES-2019, ICCIC-14 etc. He has also supervised 19 PG Dissertations and 16 UG projects. He has completed supervision of 4 Ph.D thesis and six more research scholars are also working. ---|--- | Saraju P. Mohanty (Senior Member, IEEE) received the bachelor’s degree (Honors) in electrical engineering from the Orissa University of Agriculture and Technology, Bhubaneswar, in 1995, the master’s degree in Systems Science and Automation from the Indian Institute of Science, Bengaluru, in 1999, and the Ph.D. degree in Computer Science and Engineering from the University of South Florida, Tampa, in 2003. He is a Professor with the University of North Texas. His research is in “Smart Electronic Systems” which has been funded by National Science Foundations (NSF), Semiconductor Research Corporation (SRC), U.S. Air Force, IUSSTF, and Mission Innovation. He has authored 350 research articles, 4 books, and invented 4 granted and 1 pending patents. His Google Scholar h-index is 39 and i10-index is 149 with 6600 citations. He is regarded as a visionary researcher on Smart Cities technology in which his research deals with security and energy aware, and AI/ML-integrated smart components. He introduced the Secure Digital Camera (SDC) in 2004 with built-in security features designed using Hardware-Assisted Security (HAS) or Security by Design (SbD) principle. He is widely credited as the designer for the first digital watermarking chip in 2004 and first the low-power digital watermarking chip in 2006. He is a recipient of 12 best paper awards, Fulbright Specialist Award in 2020, IEEE Consumer Technology Society Outstanding Service Award in 2020, the IEEE-CS-TCVLSI Distinguished Leadership Award in 2018, and the PROSE Award for Best Textbook in Physical Sciences and Mathematics category in 2016. He has delivered 10 keynotes and served on 11 panels at various International Conferences. He has been serving on the editorial board of several peer- reviewed international journals, including IEEE Transactions on Consumer Electronics (TCE), and IEEE Transactions on Big Data (TBD). He is the Editor- in-Chief (EiC) of the IEEE Consumer Electronics Magazine (MCE). He has been serving on the Board of Governors (BoG) of the IEEE Consumer Technology Society, and has served as the Chair of Technical Committee on Very Large Scale Integration (TCVLSI), IEEE Computer Society (IEEE-CS) during 2014-2018. He is the founding steering committee chair for the IEEE International Symposium on Smart Electronic Systems (iSES), steering committee vice-chair of the IEEE-CS Symposium on VLSI (ISVLSI), and steering committee vice-chair of the OITS International Conference on Information Technology (ICIT). He has mentored 2 post-doctoral researchers, and supervised 12 Ph.D. dissertations, 26 M.S. theses, and 12 undergraduate projects. ---|---
# Wet to dry self-transitions in dense emulsions: from order to disorder and back Andrea Montessori<EMAIL_ADDRESS>Istituto per le Applicazioni del Calcolo CNR, via dei Taurini 19, 00185, Rome, Italy Marco Lauricella Istituto per le Applicazioni del Calcolo CNR, via dei Taurini 19, 00185, Rome, Italy Adriano Tiribocchi Center for Life Nanoscience at la Sapienza, Istituto Italiano di Tecnologia, viale Regina Elena 295, 00161, Rome, Italy Istituto per le Applicazioni del Calcolo CNR, via dei Taurini 19, 00185, Rome, Italy Fabio Bonaccorso Center for Life Nanoscience at la Sapienza, Istituto Italiano di Tecnologia, viale Regina Elena 295, 00161, Rome, Italy Istituto per le Applicazioni del Calcolo CNR, via dei Taurini 19, 00185, Rome, Italy Università degi studi Roma ”Tor Vergata”, Via Cracovia, 50, 00133, Rome, Italy Sauro Succi Center for Life Nanoscience at la Sapienza, Istituto Italiano di Tecnologia, viale Regina Elena 295, 00161, Rome, Italy Istituto per le Applicazioni del Calcolo CNR, via dei Taurini 19, 00185, Rome, Italy Institute for Applied Computational Science, Harvard John A. Paulson School of Engineering and Applied Sciences, Cambridge, MA 02138, United States ###### Abstract One of the most distinctive hallmarks of many-body systems far from equilibrium is the spontaneous emergence of order under conditions where disorder would be plausibly expected. Here, we report on the self-transition between ordered and disordered emulsions in divergent microfluidic channels, i.e. from monodisperse assemblies to heterogeneous polydisperse foam-like structures, and back again to ordered ones. The transition is driven by the nonlinear competition between viscous dissipation and surface tension forces as controlled by the device geometry, particularly the aperture angle of the divergent microfluidic channel. An unexpected route back to order is observed in the regime of large opening angles, where a trend towards increasing disorder would be intuitively expected. ## I Introduction Self-organization can be broadly defined as the complex of processes which drives the emergence of spontaneous order in a given system, due to the action of local interactions between its elementary constituents [30]. This concept has provided a major paradigm to gain a deeper insight into a number of phenomena across a broad variety of complex systems in physics, engineering, biology and society [33, 2, 15, 21, 1]. Self-organization is usually triggered and sustained by competing processes far from equilibrium, as they occur in a gamut of different scientific endeavors, from natural sciences and biology to economics and anthropology [14, 35, 39, 40, 3, 10]. and often efficiently exploited to find innovative design solutions in a number of engineering applications [38, 5, 12]. From this standpoint, droplet-based microfluidics, namely the science of generating and manipulating large quantities of micron- sized droplets, offers a literal Pandora’s box of possibilities to investigate the physics of many-body systems out of equilibrium. In particular, the self- assembly between droplets and bubbles which results from the subtle multiscale competition between different forces and interactions, such as the external drive, interfacial (attractive) forces, near-contact (repulsive) interactions, viscous dissipation and inertia [26, 29]. Most importantly, in many instances, such competition is highly sensitive to geometrical factors, primarily the presence of confining boundaries. Among others, the ability to manipulate and control tiny volumes of droplets allows the generation of highly ordered porous matrices with finely tunable structural parameters [4]. This opens up the possibility of designing novel families of materials with potential use in a wide range of advanced applications, such as catalyst supports, ion-exchange modules, separation media and scaffolds in tissue engineering [20, 13, 25, 27]. Recently, Gai et al.[8], reported an unexpected ordering in the flow of a quasi-2D concentrated emulsion in a convergent microfluidic channel, and showed that confinement of the 2D soft crystal in the extrusion flow causes the reorganization of the crystal internal structure in a highly ordered pattern [8, 7]. The self-reorganization of the crystal is expected to bear major implications for the realization of confined low-dimensional materials, crucial for applications ranging from optoelectronics to energy conversion, which might be easier to control than previously thought, thus leading to novel flow control and mixing strategies in droplet microfluidics. In this paper, we report on the self-transition between wet and dry emulsions [32, 9], namely from ordered monodisperse assemblies to heterogeneous and polydisperse foam-like structures, in divergent microfluidic channel. Following the common terminology [24, 6], foams are wet when their droplets appears nearly round and the structures they form are organized according to ordered hexagonal patterns which flow basically deformation-free. In the dry regime instead, the droplets come closer and deform, assuming typical polyhedral shapes and giving rise to typical disordered foam-like structures. Such transition is driven by the capillary number, i.e. the competition between viscous dissipation and surface tension, which is in turn modulated by the device geometry. In particular, we observe a return to an ordered state in a parameter regime where a transition towards disorder would be intuitively expected. ## II Method In this section, we briefly describe the numerical model employed, namely an extended color-gradient lattice Boltzmann approach with repulsive near-contact interactions, previously introduced in [28]. In the multicomponent LB model, two sets of distribution functions evolve, according to the usual streaming- collision algorithm (see [36, 16]), to track the evolution of the two fluid components: $f_{i}^{k}\left(\vec{x}+\vec{c}_{i}\Delta t,\,t+\Delta t\right)=f_{i}^{k}\left(\vec{x},\,t\right)+\Omega_{i}^{k}[f_{i}^{k}\left(\vec{x},\,t\right)]+S_{i}(\vec{x},t),$ (1) where $f_{i}^{k}$ is the discrete distribution function, representing the probability of finding a particle of the $k-th$ component at position $\vec{x}$, time $t$ with discrete velocity $\vec{c}_{i}$, and $S_{i}$ is a source term coding for the effect of external forces (such as gravity, near- contact interactions, etc). In equation 1 the time step is taken equal to $1$, and the index $i$ spans over the discrete lattice directions $i=1,...,b$, being $b=9$ for a two dimensional nine speed lattice (D2Q9). The density $\rho^{k}$ of the $k-th$ component and the total linear momentum of the mixture $\rho\vec{u}=\sum_{k}\rho^{k}\vec{u^{k}}$ are obtained, respectively, via the zeroth and the first order moment of the lattice distributions $\rho^{k}\left(\vec{x},\,t\right)=\sum_{i}f_{i}^{k}\left(\vec{x},\,t\right)$ and $\rho\vec{u}=\sum_{i}\sum_{k}f_{i}^{k}\left(\vec{x},\,t\right)\vec{c}_{i}$. The collision operator splits into three components [11, 19, 18]: $\Omega_{i}^{k}=\left(\Omega_{i}^{k}\right)^{(3)}\left[\left(\Omega_{i}^{k}\right)^{(1)}+\left(\Omega_{i}^{k}\right)^{(2)}\right].$ (2) where $\left(\Omega_{i}^{k}\right)^{(1)}$, stands for the standard collisional relaxation [36], $\left(\Omega_{i}^{k}\right)^{(2)}$ code for the perturbation step [11], contributing to the buildup of the interfacial tension while $\left(\Omega_{i}^{k}\right)^{(3)}$ is the recoloring step [11, 17], which promotes the segregation between the two species, minimising their mutual diffusion. A Chapman-Enskog multiscale expansion can be employed to show that the hydrodynamic limit of Eq.1 is a set of equations for the conservation of mass and linear momentum (i.e. the Navier-Stokes equations), with a capillary stress tensor of the form: $\bm{\Sigma}=-\tau\sum_{i}\sum_{k}\left(\Omega_{i}^{k}\right)^{(2)}\vec{c}_{i}\vec{c_{i}}=\frac{\sigma}{2|\nabla\rho|}(|\nabla\rho|^{2}\mathbf{I}-\nabla\rho\otimes\nabla\rho)$ (3) being $\tau$ the collision relaxation time, controlling the kinematic viscosity via the relation $\nu=c_{s}^{2}(\tau-1/2)$ ( $c_{s}=1/\sqrt{3}$ the sound speed of the model) and $\sigma$ is the surface tension [36, 16]. In eq. 3, the symbol $\otimes$ denotes a dyadic tensor product. The stress-jump condition across a fluid interface is then augmented with an intra-component repulsive term aimed at condensing the effect of all the repulsive near- contact forces (i.e., Van der Waals, electrostatic, steric and hydration repulsion) acting on much smaller scales ($\sim O(1\;nm)$) than those resolved on the lattice (typically well above hundreds of nanometers). It takes the following form: $\mathbf{T}^{1}\cdot\vec{n}-\mathbf{T}^{2}\cdot\vec{n}=-\nabla(\sigma\mathbf{I}-\sigma(\vec{n}\otimes\vec{n}))-\pi\vec{n}$ (4) where $\pi[h(\vec{x})]$ is responsible for the repulsion between neighboring fluid interfaces, $h(\vec{x})$ being the distance between interacting interfaces along the normal $\vec{n}$. The additional, repulsive term can be readily included within the LB framework, by adding a forcing term acting only on the fluid interfaces in near-contact, namely: $\vec{F}_{rep}=\nabla\pi:=-A_{h}[h(\vec{x})]\vec{n}\delta_{I}$ (5) In the above, $A_{h}[h(\vec{x})]$ is the parameter controlling the strength (force per unit volume) of the near-contact interactions, $h(\vec{x})$ is the distance between the interfaces, $\vec{n}$ is a unit vector normal to the interface and $\delta_{I}\propto\nabla\phi$ is a function, proportional to the phase field $\phi=\frac{\rho^{1}-\rho^{2}}{\rho^{1}+\rho^{2}}$, employed to localize the force at the interface. The addition of the repulsive force (added to the right hand side of Eq. 1) naturally leads to the following (extended) conservation law for the momentum equation: $\frac{\partial\rho\vec{u}}{\partial t}+\nabla\cdot{\rho\vec{u}\vec{u}}=-\nabla p+\nabla\cdot[\rho\nu(\nabla\vec{u}+\nabla\vec{u}^{T})]+\nabla\cdot(\bm{\Sigma}+\pi\mathbf{I})$ (6) namely the Navier-Stokes equation for a multicomponent system, augmented with a surface-localized repulsive term, expressed through the gradient of the potential function $\pi$. ## III Results and Discussion The simulation set-up (see figure 1) consists of a microfluidic device composed by an inlet channel ($h_{c}$), a divergent channel with opening angle $\alpha$ and a main channel connected with the divergent ($h=10h_{c}$). The droplets are continuously generated in a buffer channel, placed upstream the inlet channel whose height is $h_{in}\sim 1.7h_{c}$, equal to the droplets diameter. This way, the droplets are forced to deform as they enter the narrower inlet channel, taking a typical oblate shape. The fluid motion is driven by a body force which mimics the effect of a pressure gradient across the device, which is set in such a way as to guarantee laminar flow conditions within both inlet and main channels. The main parameters employed (expressed in simulation (lattice) units) are the following. The microfluidic device is composed by a thin inlet channel (height $h_{c}=30$, length $l_{c}=220$) within which droplets are produced and a main, or self-assembly channel, height $h=10h_{c}$, length $l=900$, where the droplets are transported downstream by the main flow and self-assemble in clusters during their motion. The droplet diameter is set to $D=50$ lattice units, more than sufficient to capture the complex interfacial phenomena occurring in droplet microfluidics (50 lattice points per diameter means a Cahn number of the order $0.08$, typical in resolved diffuse interface simulations of complex interfaces (see [23])). The motion of the droplet is realized by imposing a constant body force $g=10^{-5}$. The viscosity of the two fluids has been set to $\nu=0.167$ while the near-contact force has been set to $A_{h}=0.1$. The choice of magnitude of the body force, along with the kinematic viscosity of the fluids is such to determine a droplet Reynolds number within the inlet channel $Re\sim 2.5$, small enough to guarantee laminar flow conditions. The surface tension has been varied in the range $\sigma=0.007\div 0.02$. Finally, the droplet generation is performed by implementing an internal periodic boundary condition whose short explanation is reported in appendix. All the simulations were performed in two-dimensions, being this a reasonable approximation for the simulation of droplets’ phenomena in shallow microfluidic channels. We wish to point out that the only parameter which has been varied throughout the simulation is the surface tension between the two components which, in turn, allowed us to tune the capillary number. The rate of injection of both the dispersed and the continuous phase were kept constant in all the simulations. In figure 1, we report two different assemblies of droplets within a microfluidic channel with a divergent opening angle $\alpha=45^{\circ}$. This figure shows that the tuning of the inlet Capillary number, ($Ca=U_{d}\mu_{d}/\sigma$, the $d$ subscript standing for droplet, $U_{d}$ is the average droplet velocity within the inlet channel, $\mu_{d}$ the dynamic viscosity and $\sigma$ the surface tension of the mixture), allows to switch between a closely packed, ordered, monodisperse emulsion (1(a)) characterized by regular hexagonal assemblies of droplets, traveling along the micro- channel, to a foam-like structures, formed by polyhedral-shaped droplets (see fig. 1 (c) (d)). The resulting structures appear to be irregular and polydispersed, as indicated by the distortion of the Delaunay triangulation and its dual Voronoi tessellation [22]. We wish to highlight that, both the dispersed and continuous phases’ discharges are kept constant in all the simulations. Thus, the observed transition is likely to by due to (i) the breakup processes promoting the formation of liquid films and (ii) the increased deformability of the droplets interface, due to the lower values of surface tension employed. Typically, droplet breakups increase the amount of interface, leading to an augment of the total length of the thin-film and to a redistribution of the dispersed phase in the system. Such dynamics is controlled by the Capillary number whose increase (within a quite wide range of aperture angles) leads to a spontaneous transition from an ordered state to a disordered one displaying the typical features of a dense foam, namely (i) polydispersity, (ii) formation of an interconnected web of plateaus, (iii) departure of the droplets shapes from the circular or spherical one and (iv) formation of droplets assemblies which are not regular as in the wet case. The transition between different droplets’ structures depends not only on the inlet capillary number but also on the geometrical details of the device, the latter being responsible for a counter-intuitive behavior, to be detailed shortly. Figure 1: Droplet assemblies within a microfluidic channel with a divergent opening angle $\alpha=45^{\circ}$ for two different Inlet channel Capillary number (a) $Ca=0.04$ (c) $Ca=0.16$. Panel (b) clearly shows the ordered, hexagonal packing typical of wet-state emulsions, while (d) the foam-like structure which results in a neat distortion of the Delauney triangulation (blue solid lines connecting the centers of neighboring droplets) and the associated Voronoi tesselation (dotted polygons enclosing the droplets) as well. The red lines are isocontour lines drawn for $\phi_{min}<\phi<\phi_{max}$ being $\phi$ the local phase field $\phi=(\rho_{1}-\rho_{2})/(\rho_{1}+\rho_{2})$ ($\rho_{i}$ the density of the $i$-th phase). The lines are superimposed to a density field. The thickness of the red isocontour line has been widening in order to better visualized the droplet contours. We begin with a phenomenological description of the droplets injection within the diverging channel (for $\alpha=45^{\circ}$), as influenced by the Capillary number. As shown in figure 2 (a-d), below a given value of the Capillary number at the inlet, $Ca\lesssim{0.05}$, every new droplet emerging in the divergent channel pushes away another immediately downstream, taking its place in the process. Indeed, as clearly sequenced in the figure, the yellow-triangle droplet comes out of the channel, pushes the orange-dotted one which, in turn, takes the place of its nearest-neighbor droplet (the red-star one). This process is metronomic i.e. it does not involve any breakup event and this rythmic push- and-slide mechanism reflects into the regular hexagonal crystal which forms downstream the main channel. Figure 2: (a-d) ($Ca\sim 0.04$) Push and slide mechanism of the outcoming droplet. The yellow-triangle droplet comes out of the channel, pushes the orange-dotted one which, in turn, takes the place of its nearest-neighbor droplet (the red-star one). (e-h) ($Ca\sim 0.16$) Droplet pinch-off process. The dotted-orange droplet undergoes a transversal stretching due to the squeezing between the outcoming droplet and the red-star drop. The stretched droplet finally reaches a critical elongation and thinning under the confinement of the neighbor drops before pinch-off. (i-n) Experimental sequence of the breakup mechanism at $Ca\sim 0.08$ (see [37]). The experimental and numerical critical capillary numbers above which droplets pinch-off can be observed are $Ca\geq\sim 0.04$ and $Ca\geq\sim 0.05$ respectively. As stated before, an increase of the Capillary number above a critical value, around $Ca\sim 0.1$, determines the transition to a heterogeneous, foam-like structure, as shown in fig. 1(b). This latter is due to the subsequent breakup events taking place immediately downstream the injection channel, a process highlighted in figure 2 (e-h). The dotted-orange droplet undergoes a transversal stretching due to the squeezing between the outcoming droplet (i.e. the hammer droplet ) and the red-star drop (i.e. the wall droplet). The stretched droplet finally reaches a critical elongation and thinning under the confinement of the neighbor drops before pinch-off. In the meantime, the yellow-triangle droplet, due to the rapid slow down determined by the channel expansion, gradually takes on a crescent shape, fills the area left free by the splitting of the orange drop and becomes a wall droplet in turn. The splitting mechanism just described is responsible for the formation of smaller droplets, which assemble in such a way as to form a heterogeneous foam-like structure within the main channel (fig.1(b)). Briefly, what we observe from the simulations is that, frequent and precise pinch-off requires sufficiently high capillary numbers to occur ($Ca>0.1$ for $\alpha=45^{\circ}$). This suggests that the ratio between the viscous forces (extensional force) and surface tension (retraction/restoring force), namely the Capillary number, is likely to govern the behavior of the droplet-droplet pinch-off process. Indeed, as the viscous force retard the expansion of the impinging droplet, the central one stretches and breaks at the midpoint due to the deformation arising from the normal stresses exerted by the impinging and wall droplet. The surface tension then acts so to contrast the effect of the normal stresses, since both the hammer and the wall droplets tend to retract to their undeformed circular state. It is worth noting that a similar pinching mechanism has been recently observed experimentally in [37] in the same range of capillary numbers as in our simulations.Incidentally, the transitional Capillary number of the experiments (i.e. $Ca$ above which the pinching mechanism is observed) was found to be in satisfactory agreement with the one predicted by the simulations (see caption fig. 2). Figure 3: Equivalent droplet diameter distributions for each pair of Capillary number and opening angle of the divergent channel ($Ca=0.04$(Dashed line), $Ca=0.1$ (Dotted line) and $Ca=0.16$ (full line)). The insets report snapshots of the droplet fields for different values of the Capillary number ($Ca$ increases from top to bottom). The equivalent droplet diameter is the diameter of the circular droplet with the same area of the deformed droplet and can be computed as $D_{e}=\sqrt{4(A_{d}/\pi)}$ being $A_{d}$ the area of the droplet. At this stage, a question naturally arises as to the role of geometrical details of the divergent channel on the wet to dry self-transition. To address this question, we performed a series of simulations by varying both the Capillary number and the opening angle of the divergent channel, so to systematically assess their combined effect on the final shape of the assemblies of droplets within the microfluidic channel. The results of this investigation are summarized in the histograms reported in figure 3. Each histogram shows the distribution of the equivalent droplet diameters within the microfluidic channel (i.e. the diameter of the circular droplet with the same area of the deformed droplet, computed as $\sqrt{4(A_{d}/\pi)}$ being $A_{d}$ the area of the droplet) for a given pair $Ca$ and $\alpha$. A number of comments is in order : i) Below $Ca\sim 0.05$, no breakup event is observed, regardless of the opening angle: the outcoming soft structures are monodisperse assemblies of droplets, as clearly suggested by the dashed-line histograms of fig. 3. ii) Upon raising the Capillary number, it is possible to trigger the breakup events which lead to the transition between ordered and disordered emulsions. By inspecting the histograms ($Ca\sim 0.1$ (dotted line) and $Ca\sim 0.16$ (solid line) ), it is evident that, for a given $\alpha$, the number of breakup events, and in turn, the structure of the resulting emulsion, depend on the inlet capillary number. Indeed, by increasing the Capillary number, the droplets structure increasingly takes the hallmarks of a dense-foam or highly packed dense emulsion (HIPE). For $\alpha$ in range $30^{\circ}-60^{\circ}$, $Ca\sim 0.1$ can be regarded as a critical value of the Capillary number, around which the emergent soft structure is a hybrid between a monodisperse (ordered) and polydisperse (disordered) emulsion, as also evidenced by the the central droplet fields reported in the insets of the histograms. iii) By further increasing the Capillary number, the assemblies of droplets take a typical foam-like structure, completely loosing memory of the structural hexagonal-ordering obtained at lower $Ca$. A more complex structure is found, due to the (a) higher degree of deformability of the droplets, an emergent effect due to the higher values of the Capillary numbers and (b) the presence of smaller droplets which fill the voids between groups of neighbor droplets. iv) Focusing on the highest value of the Capillary number, $Ca=0.16$, we note that the polydispersity, revealed by bimodal histograms, increases as $\alpha$ increases from $22.5^{\circ}$ to $45^{\circ}$. By further increasing the opening angle, the polydispersity starts to recede, nearly vanishing at $\alpha=90^{\circ}$. Figure 4: (a) Equivalent droplet diameter distributions for two couples of Capillary numbers and opening angles of the divergent channel, namely $Ca=0.04$, $Ca=0.16$ and $\alpha=45^{\circ}$, $\alpha=60^{\circ}$, $\alpha=90^{\circ}$. (b) Droplets’ field within the main channel for the case $Ca=0.16$ and $\alpha=90^{\circ}$ To better highlight the aforementioned return towards monodispersity for increasing values of $\alpha$, we directly compare the histograms for six cases, namely, (i) $Ca=0.04$ and $\alpha=45^{\circ}$, (ii) $Ca=0.16$ and $\alpha=45^{\circ}$, (iii)$Ca=0.04$ and $\alpha=60^{\circ}$,(iv)$Ca=0.16$ and $\alpha=60^{\circ}$, (v)$Ca=0.04$ and $\alpha=90^{\circ}$ and (iv)$Ca=0.16$ and $\alpha=90^{\circ}$ (panel (a)). A close inspection of the histograms leave no doubt as to the return to monodispersity for the case $Ca=0.16$ and $\alpha=90$. Indeed, at $Ca=0.16$ and $\alpha=45^{\circ}$ and $\alpha=60^{\circ}$ the emulsion is roughly bidisperse as evidenced by the two peaks at $D_{e}\sim 50$ and $D_{e}\sim 36$ (dashed circles) displayed in the two histograms, the latter one absent in the case $Ca=0.16$ and $\alpha=90^{\circ}$. The decreasing trend of the ratio between the peaks in the histograms at $Ca=0.16$ ($\sim 1.2$ for $\alpha=45^{\circ}$ and $\sim 2.5$ for $\alpha=60^{\circ}$) further points to a gradual return to an ordered structure as the aperture angle increases. This is also apparent from a visual inspection of the droplets’ field reported in panel (b) ($Ca=0.16$, $\alpha=90^{\circ}$) which shows an ensemble of flowing circular droplets of (approximately) the same size.The rare breakup events occurring at the outlet of the injection channel produces a limited number of smaller droplets, an effect evidenced by the small peaks in the upper right histogram. It is worth noting here that, by a plain argument of mass conservation, polydispersity can arise only as a result of droplet breakup via the droplet hammer mechanism since coalescence is frustrated due to the effect of the near contact forces. The counterintuitive behaviour described above, can be intuitively explained as follows: at high opening angles (approaching to $90^{\circ}$), each droplet exiting from the narrow channel experiences a sudden expansion, responsible for a fast recovery of their circular shape, just after their emergence within the main channel. The fast expansion, in turn, determines a strong deceleration (see plot in fig. 5), which forces the next outcoming droplet to loose its droplet hammer action, as the opening angle of the divergent increases above a critical value between $\alpha=45-60^{\circ}$. Further, the sharp deceleration favors the crossflow, transversal displacement of the outcoming droplets, which slide preferentially on the downstream, neighbour droplets rather than squeezing them. The process described above is reported in figure 5. Figure 5: (a-g) Droplets’ field at the outlet of the injection nozzle for $\alpha=90^{\circ}$ and $Ca=0.16$ with the normalized vector field superimposed. Even at high capillary numbers, the sharp deceleration, clearly evidenced by the velocity profiles reported in panel (h) taken in two distinct sections, within the nozzle(open circles) and at a downstream section, favors the transversal displacement of the outcoming droplets, which preferentially slide on the neighboring droplets rather than squeezing them. The velocity field is scaled with the maximum flow velocity at the inlet channel. The two sections, at which the velocity profiles are evaluated are at $x=190lu$ (inside the injection channel) and $x=340lu$ (within the main channel). The arrow within the plot indicates the average velocity drop between the inlet and the main channel. The axis of the plot are, $u/u_{M}$ (normalized magnitude of the velocity) versus $y$(crossflow coordinate). It is to note that, by varying the surface tension, and by keeping the other parameters fixed (so that the Reynolds number can be kept fixed), it is possible to vary viscous dissipation over surface forces independently of the inertia over viscous dissipation ratio. Indeed, the viscous vs surface tension forces ratio is responsible for the frequency of breakup events, hence, in turn, for the degree of order/disorder observed in the system. This competition strictly depends on the geometrical features of the microfluidic environment. To be noted that, the importance of Weber and Capillary number over the droplet breakup frequency in microchannels has been also highlighted in a very recent experimental work of Salari et al. [34] reporting power-law scaling of the dropplets’ breakup frequencies as a function of the product $WeCa^{2}$. To sum up, the simulations suggest that, the dependence of the crystal order on the geometrical feature of the device is not one-way since, once the monodispersity and the hexagonal order are lost, they can be reclaimed back by either decreasing or increasing the opening angle of the divergent channel below/above a critical angle. In other words, either ways, the system looses memory of the disordered configuration. As a note, we wish to stress out that, the detection of the specific regime of capillarity in which the transition occurs required an extensive set of simulations on a broad range of Capillary numbers, as such transition was found to occur indeed in a very narrow window of capillarity space. Thus, even though many more $Ca-\alpha$ combinations have been explored, it was found that the cases reported capture the essence of the phenomenon in point. To gain a quantitative insight into the order to disorder transition, we introduce a dispersity number, $\delta$, defined as the ratio between the number of droplets with a diameter below a critical value, $D_{crit}$, and the total number of droplets. This parameter has been evaluated for each pair of Capillary number $Ca$ and opening angle $\alpha$. The plot in figure 6 reports these data, made non-dimensional by the maximum opening angle $\tilde{\alpha}=\frac{\alpha}{\pi/2}$ and the maximum value of dispersity $\tilde{\delta}=\frac{\delta}{\delta_{M}}$, respectively. Figure 6: Non dimensional dispersity ($\tilde{\delta}$) as a function of the opening angle $\tilde{\alpha}$ for different values of $Ca$. Each dispersity set, $\delta(Ca,\tilde{\alpha})$, follows a Gaussian trend, with mean and variance depending on the Capillary number. Fitting function: $\tilde{\delta}(\tilde{\alpha})=e^{-\left((\tilde{\alpha}-\alpha_{M})/(2Ca)\right)^{2}}$ A few comments are in order: The first observation is that each dispersity set, $\delta(Ca,\tilde{\alpha})$, follows a Gaussian trend, with mean and variance depending on the Capillary number: $\tilde{\delta}(\tilde{\alpha})=e^{-\left((\tilde{\alpha}-\alpha_{M}(Ca))/(2Ca)\right)^{2}}$ (7) . In other words, the dispersity of the system features a ”temperature” which scales linearly with the inlet Capillary number, $T=2Ca$, and a mean value of the opening angle, slightly depending on the Capillary number and ranging between $45-60^{\circ}$. The analysis carried out in this paper should be of direct use for experimental research. Indeed, for each inlet Capillary number, which can be readily determined by evaluating the droplet velocity within the inlet channel, one can single out the channel geometry which allows to obtain the desired degree of polydispersity of the soft structure, by simply querying the gaussian curves. Reciprocally, given the channel geometry, the capillary number can be tuned in such a way as to modify the morphology of the droplet assembly, according again to the gaussian relation provided in this paper. The present findings are expected to help in defining experimental protocols for the development of novel, optimized, low-dimensional, soft porous matrices with tunable properties. We refer in particular to the so-called functionally graded materials [31], namely composite materials characterized by a controlled spatial variation of their microstructure, which are capturing mounting interest for a variety of material science, biology and medical applications. ## IV Conclusions In summary, we reported on order to disorder self-transition in dense emulsions in divergent microfluidic channels, as originated by a geometry- controlled competition between viscous dissipation and interfacial forces. We unveiled a counterintuitive mechanism, namely the spontaneous reordering of the emulsion at high Capillary numbers, obtained by increasing of the opening angle of the divergent channel. Such comeback of order is interpreted as the result of a subtle balance between viscous dissipation and interfacial forces, straight downstream the inlet channel. Moreover, We found that the dispersity of the droplet system follows a simple Gaussian law, whose temperature is directly proportional to the inlet Capillary number. The present findings are expected to offer valuable guidance for the future development of optimised functional materials with locally tunable properties. ## Acknowledgments A. M., M. L., A. T. and S. S. acknowledge funding from the European Research Council under the European Union’s Horizon 2020 Framework Programme (No. FP/2014-2020) ERC Grant Agreement No.739964 (COPMAT). A.M. acknowledges the ISCRA award SDROMOL (HP10CZXK6R) under the ISCRA initiative, for the availability of high performance computing resources and support. ## Appendix ### Droplets’ generation The droplet generation is performed by implementing an internal periodic boundary condition which is sketched in figure 7, for simplicity. Figure 7: Droplet generation via the internal periodic boundary conditions. Populations of the dispersed and the continuous phase are copied from section (b) back to section (a) and, at the same time, a standard streaming and collision process occurs within the bulk domain. As shown in figure, in order to generate a droplets inflow in the inlet thin channel, the generating region is employed as a source of new droplets. When a droplet passes through section (b), it simultaneosly i) enters into the downstream region and ii) is copied back to the inlet section by applying periodic boundary conditions from $(b)$ to $(a)$ (see figure 7). ## References * [1] M. Ballerini, N. Cabibbo, R. Candelier, A. Cavagna, E. Cisbani, I. Giardina, V. Lecomte, A. Orlandi, G. Parisi, A. Procaccini, et al. Interaction ruling animal collective behavior depends on topological rather than metric distance: Evidence from a field study. Proceedings of the national academy of sciences, 105(4):1232–1237, 2008. * [2] S. Camazine, J.-L. Deneubourg, N. R. Franks, J. Sneyd, E. Bonabeau, and G. Theraula. Self-organization in biological systems, volume 7. Princeton university press, 2003. * [3] B. Chopard and M. Droz. Cellular automata, volume 1. Springer, 1998. * [4] M. Costantini, C. Colosi, J. Guzowski, A. Barbetta, J. Jaroszewicz, W. Swieszkowski, M. Dentini, and P. Garstecki. Highly ordered and tunable polyhipes by using microfluidics. J. Mater. Chem. B, 2:2290–2300, 2014. * [5] A. Dinsmore, M. F. Hsu, M. Nikolaides, M. Marquez, A. Bausch, and D. Weitz. Colloidosomes: selectively permeable capsules composed of colloidal particles. Science, 298(5595):1006–1009, 2002. * [6] Y. Furuta, N. Oikawa, and R. Kurita. Close relationship between a dry-wet transition and a bubble rearrangement in two-dimensional foam. Scientific reports, 6:37506, 2016. * [7] Y. Gai, A. Bick, and S. K. Tang. Timescale and spatial distribution of local plastic events in a two-dimensional microfluidic crystal. Physical Review Fluids, 4(1):014201, 2019. * [8] Y. Gai, C. M. Leong, W. Cai, and S. K. Tang. Spatiotemporal periodicity of dislocation dynamics in a two-dimensional microfluidic crystal flowing in a tapered channel. Proceedings of the National Academy of Sciences, 113(43):12082–12087, 2016. * [9] P. Garstecki and G. M. Whitesides. Flowing crystals: nonequilibrium structure of foam. Phys. Rev. Lett., 97:024503, 2006. * [10] N. Goldenfeld and L. P. Kadanoff. Simple lessons from complexity. science, 284(5411):87–89, 1999. * [11] A. K. Gunstensen, D. H. Rothman, S. Zaleski, and G. Zanetti. Lattice boltzmann model of immiscible fluids. Physical Review A, 43(8):4320, 1991. * [12] J. Guzowski and P. Garstecki. Droplet clusters: Exploring the phase space of soft mesoscale atoms. Physical review letters, 114(18):188302, 2015. * [13] S. J. Hollister. Porous scaffold design for tissue engineering. Nature materials, 4(7):518–524, 2005. * [14] E. Karsenti. Self-organization in cell biology: a brief history. Nature reviews Molecular cell biology, 9(3):255–262, 2008. * [15] B. S. Kerner. Experimental features of self-organization in traffic flow. Physical review letters, 81(17):3797, 1998. * [16] T. Kruger, H. Kusumaatmaja, A. Kuzmin, O. Shardt, G. Silva, and E. M. Viggen. The lattice boltzmann method. Springer International Publishing, 10:978–3, 2017. * [17] M. Latva-Kokko and D. H. Rothman. Diffusion properties of gradient-based lattice boltzmann models of immiscible fluids. Physical Review E, 71(5):056702, 2005. * [18] S. Leclaire, A. Parmigiani, O. Malaspinas, B. Chopard, and J. Latt. Generalized three-dimensional lattice boltzmann color-gradient method for immiscible two-phase pore-scale imbibition and drainage in porous media. Physical Review E, 95(3):033306, 2017. * [19] S. Leclaire, M. Reggio, and J.-Y. Trepanier. Numerical evaluation of two recoloring operators for an immiscible two-phase flow lattice boltzmann model. Applied Mathematical Modelling, 36(5):2237–2252, 2012. * [20] M. Lee, J. C. Dunn, and B. M. Wu. Scaffold fabrication by indirect three-dimensional printing. Biomaterials, 26(20):4281–4289, 2005. * [21] W. Lee, H. Amini, H. A. Stone, and D. Di Carlo. Dynamic self-assembly and control of microfluidic particle crystals. Proceedings of the National Academy of Sciences, 107(52):22413–22418, 2010. * [22] M. Lulli, R. Benzi, and M. Sbragaglia. Metastability at the yield-stress transition in soft glasses. Physical Review X, 8(2):021031, 2018. * [23] F. Magaletti, F. Picano, M. Chinappi, L. Marino, and C. M. Casciola. The sharp-interface limit of the cahn-hilliard/navier-stokes model for binary fluids. Journal of Fluid Mechanics, 714:95, 2013. * [24] P. Marmottant and J. P. Raven. Microfluidics with foams. Soft Matter, 5:3385–3388, 2009. * [25] M. M. Montemore, A. Montessori, S. Succi, C. Barroo, G. Falcucci, D. C. Bell, and E. Kaxiras. Effect of nanoscale flows on the surface structure of nanoporous catalysts. The Journal of chemical physics, 146(21):214703, 2017. * [26] A. Montessori, M. Lauricella, and S. Succi. Mesoscale modelling of soft flowing crystals. Phil. Trans. Roy. Soc., Ser. A, 377:20180149, 2019. * [27] A. Montessori, M. Lauricella, S. Succi, E. Stolovicki, and D. Weitz. Elucidating the mechanism of step emulsification. Phys. Rev. F., 3:072202, 2018. * [28] A. Montessori, M. Lauricella, N. Tirelli, and S. Succi. Mesoscale modeling of near-contact interactions for complex flowing interfaces. Jour. Fluid Mech., 872:327–347, 2019. * [29] A. Montessori, M. Lauricella, A. Tiribocchi, and S. Succi. Modeling pattern formation in soft flowing crystals. Physical Review Fluids, 4(7):072201, 2019. * [30] I. Prigogine and I. Stengers. Order out of chaos: Man’s new dialogue with nature. Verso Books, 2018. * [31] B. Rabin and I. Shiota. Functionally gradient materials. MRS bulletin, 20(1):14–18, 1995. * [32] J. P. Raven and P. Marmottant. Microfluidic crystals: Dynamic interplay between rearrangement waves and flow. Phys. Rev. Lett., 102:084501, 2009. * [33] B. M. Rosen, C. J. Wilson, D. A. Wilson, M. Peterca, M. R. Imam, and V. Percec. Dendron-mediated self-assembly, disassembly, and self-organization of complex systems. Chemical reviews, 109(11):6275–6540, 2009. * [34] A. Salari, J. Xu, M. C. Kolios, and S. S. Tsai. Expansion-mediated breakup of bubbles and droplets in microfluidics. Physical Review Fluids, 5(1):013602, 2020. * [35] H. Stanley, L. Amaral, S. V. Buldyrev, P. Gopikrishnan, V. Plerou, and M. Salinger. Self-organized complexity in economics and finance. Proceedings of the National Academy of Sciences, 99(suppl 1):2561–2565, 2002. * [36] S. Succi. The Lattice Boltzmann Equation: For Complex States of Flowing Matter. Oxford University Press, 2018. * [37] D. Vecchiolla, V. Giri, and S. L. Biswal. Bubble–bubble pinch-off in symmetric and asymmetric microfluidic expansion channels for ordered foam generation. Soft matter, 14(46):9312–9325, 2018. * [38] N. Vogel, S. Utech, G. T. England, T. Shirman, K. R. Phillips, N. Koay, I. B. Burgess, M. Kolle, D. A. Weitz, and J. Aizenberg. Color from hierarchy: Diverse optical properties of micron-sized spherical colloidal assemblies. Proceedings of the National Academy of Sciences, 112(35):10845–10850, 2015. * [39] S. Wolfram. Cellular automata as models of complexity. Nature, 311(5985):419–424, 1984. * [40] S. Wolfram. Cellular automata and complexity: collected papers. CRC Press, 2018.
Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). FIRE 2020: Forum for Information Retrieval Evaluation, December 16-20, 2020, Hyderabad, India [orcid=0000-0001-8336-195X, ]<EMAIL_ADDRESS> <EMAIL_ADDRESS> # CMSAOne@Dravidian-CodeMix-FIRE2020: A Meta Embedding and Transformer model for Code-Mixed Sentiment Analysis on Social Media Text Suman Dowlagar [ International Institute of Information Technology - Hyderabad (IIIT-Hyderabad), Gachibowli, Hyderabad, Telangana, India, 500032 Radhika Mamidi [ (2020) ###### Abstract Code-mixing(CM) is a frequently observed phenomenon that uses multiple languages in an utterance or sentence. CM is mostly practiced on various social media platforms and in informal conversations. Sentiment analysis (SA) is a fundamental step in NLP and is well studied in the monolingual text. Code-mixing adds a challenge to sentiment analysis due to its non-standard representations. This paper proposes a meta embedding with a transformer method for sentiment analysis on the Dravidian code-mixed dataset. In our method, we used meta embeddings to capture rich text representations. We used the proposed method for the Task: “Sentiment Analysis for Dravidian Languages in Code-Mixed Text”, and it achieved an F1 score of $0.58$ and $0.66$ for the given Dravidian code mixed data sets. The code is provided in the Github https://github.com/suman101112/fire-2020-Dravidian-CodeMix. ###### keywords: social media code-mixed sentiment analysis meta embedding Transformer GRU ## 1 Introduction Code-mixing(CM) of text is prevalent among social media users, where words of multiple languages are used in the sentence. Code-mixing occurs when conversant uses both languages together to the extent that they change from one language to another in the course of a single utterance [1]. The computational modeling of code-mixed text is challenging due to the linguistic complexity, nature of mixing, the presence of non-standard variations in spellings, grammar, and transliteration [2]. Because of such non-standard variations, CM poses several unseen difficulties in fundamental fields of natural language processing (NLP) tasks such as language identification, part- of-speech tagging, shallow parsing, Natural language understanding, sentiment analysis. Gysels [3] defined the Code-mixing as “the embedding of linguistic units of one language into an utterance of another language”. Code-mixing is broadly classified into two types, intra-sentential and inter-sentential. Intra- sentential code-mixing happens after every few words. Whereas, in inter- sentential code-mixing, one part of the sentence consists of Hindi words, and the other part is entirely English. The code-mixing helps people to express their emotions or opinions emphatically, thus leading to a phenomenal increase of use in code-mixed messages on social media platforms. With the increase in code-mixed data, the analysis of CMSM text has become an essential research challenge from the perspectives of both Natural Language Processing (NLP) and Information Retrieval (IR) communities. There have been some research works in this direction, such as GLUECoS, an evaluation benchmark in code-mixed text [4], automatic word-level language identification for CMSM text [5, 6], parsing pipeline for Hindi-English CMSM text [7, 8], and POS tagging for CMSM text [9]. To encourage research on code-mixing, the NLP community organizes several tasks and workshops such as Task9: SentiMix, SemEval 2020111https://competitions.codalab.org/competitions/20654, and 4th Workshop on Computational Approaches for Linguistic Code- Switching222https://www.aclweb.org/portal/content/fourth-workshop- computational-approaches-linguistic-code-switching. Similarly, the FIRE 2020’s Dravidian-CodeMix task333https://dravidian-codemix.github.io/2020/ was devoted to code-mixed sentiment analysis on Tamil and Malayalam languages. This task aims to classify the given CM youtube comments into one of the five predefined categories: positive, negative, mixed_feelings, not_<language>444the language might be Tamil and Malayalam, unknown_state. In this paper, we present a meta-embedding with a transformer model for Dravidian Code-Mixed Sentiment Analysis. Our work is similar to the meta embedding approach used for named entity recognition on code-mixed text [10]. The paper is organized as follows. Section 2 provides related work on code- mixed sentiment analysis. Section 3 describes the proposed work. Section 4 presents the experimental setup and the performance of the model. Section 5 concludes our work. ## 2 Related Work Sentiment analysis is one of the essential tasks in the field of NLP. Sentiment analysis is the process of understanding the polarity of the sentence. Sentiment analysis helps to attain the public’s attitude and mood, which can help us gather insightful information to make future decisions on large datasets [11]. Initially, sentiment analysis was used on government campaigns and news articles [12, 13]. Recently, due to social media prevalence, the research turned towards capturing the sentiment on social media texts in code-mixing scenarios [14]. The earlier approaches used syntactic rules and lexicons to extract features followed by traditional machine learning classifiers for sentiment analysis on code-mixed text. The process of rule extraction and defining lexicons is a time consuming, laborious process, and is domain-dependent. The recent work in the field of CMSA uses embeddings with deep learning and traditional classifiers [15]. The paper [16] used sub-word information for sentiment analysis on code-mixed text. The recent SentiMix 2020 task used BERT-like models and ensemble methods to capture the code-mixed texts’ sentiment [14]. We used meta embeddings with state of the art transformer model for this task. ## 3 Proposed Model This section presents our proposed code-mixed sentiment analysis framework. It has three main components: a sub-word level tokenizer, a text representation layer, and a transformer model. ### 3.1 Sub-word Level Tokenizer To deal with the non-standard variations in spellings, we used the SentencePiece [17]. SentencePiece is an unsupervised text tokenizer and de- tokenizer mainly used for neural network models. SentencePiece treats the sentences just as sequences of Unicode characters. It implements subword units by using byte-pair-encoding (BPE) [18] and unigram language model [19] . The byte pair encoding initializes the vocabulary to every character present in the corpus and progressively learn a given number of merge rules. The unigram language model trains the model with multiple subword segmentations probabilistically sampled during training. ### 3.2 Text Representation Layer Pre-trained embedding models do not perform well on the code-mixed corpus as they consider all the code-mixed words as OOV words [20]. Thus, we have to train word representations from the code-mixed corpus. Given the complexity of the code-mixed data, it is not easy to determine which embedding model to be used for better performance. Hence, we chose the combination of fastText [21], ELMO [22], and TF-IDF [23] embeddings. fastText captures efficient text representations and local dependencies at the word and sub-word level. ELMO captures contextual representations at the sentence level. TF-IDF captures the distribution of the words in the corpus. The use of TF-IDF for sentiment analysis helps in extracting a better correlation between words and their polarity. All these diverse text representations, when combined, proved beneficial in obtaining better embeddings for the downstream tasks. ### 3.3 Transformer model From [14], For the task of sentiment analysis, we saw that the attention mechanism works better in deciding which part of the sentence is essential for capturing the sentiment. Thus we chose the transformer model for our code- mixed sentiment analysis task. As the data is a classification type, we used only the encoder side from the Transformer. The encoder encodes the entire source sentence into a sequence of context vectors. First, the tokens are passed through a standard embedding layer, and the positional embeddings are concatenated with each source sequence. The embeddings are then passed through a series of encoder layers to get an encoded sequence. The encoder layers is an essential module where all the processing of the input sequence happens. We first pass the source sentence and its mask into the multi-head attention layer, then perform dropout, apply a residual connection, and pass it through a normalization layer. We later pass it through a position-wise feedforward layer and then, again, apply dropout, a residual connection, and layer normalization to get encoded output sequence. The output of this layer is fed into the next encoder layer. The Transformer model uses scaled dot-product attention given in equation 1, where the query $Q$ and key $K$ are combined by taking the dot product between them, then applying the softmax operation and scaled by a scaling factor $d_{k}$ then multiplied by the value $V$. Attention is a critical unit in the Transformer model as it helps in deciding which parts of the sequence are important. $Attention(Q,K,V)=Softmax(\frac{QK^{T}}{\sqrt[]{d_{k}}})V$ (1) The other main block inside the encoder layer is the position-wise feedforward layer. The input is transformed from hid_dim to pf_dim, where pf_dim is usually a lot larger than hid_dim. The ReLU activation function and dropout are applied before it is transformed back into a hid_dim representation. The intuition borrows from infinitely wide neural networks. The wide neural network grants more approximation power and helps to optimize the model faster. ### 3.4 Our Approach Figure 1: Meta Embedding with transformer and GRU model Initially, we tokenized the sentence using the SentencePiece model. After tokenization, we extracted local dependencies between embeddings at the subword level using the fastText model. The fastText model gave embeddings at the word level. We then applied the transformer model to obtain the encoded representations. We got the encoded representations at the word level. A GRU unit is used to get the encoded representation of all the words. We considered the representation of the last hidden layer of the GRU as the final encoded representation. We then obtained the ELMO contextual and TF-IDF representations at the sentence level. We concatenated the representations of the last hidden GRU layer, ELMO, and TF-IDF giving us the meta-embeddings. The meta embeddings are then passed to the output feed-forward network to predict the polarity of the sentence. ## 4 Experimental Setup ### 4.1 Data For Dravidian code-mixed sentiment analysis, we used the dataset provided by the organizers of Dravidian Code-mixed FIRE-2020. The training dataset consists of 15,744 Tamil CM and 6,739 Malayalam CM youtube video comments. The details of the dataset and the initial benchmarks on the corpus are given in [24, 25, 26, 27, 28] ### 4.2 Hyperparameters #### For Embedding models Embeddings play a vital role in improving the model’s performance. As mentioned above, we used fastText and ELMO embeddings. The dimensionality was set to 300 in-case of fastText embeddings. The embeddings are trained on training data using the parameters: learning rate = 0.05, context window = 5, epochs = 20. The ELMO model is obtained from tensorflow_hub555https://tfhub.dev/google/elmo/2, and the pre-set dimensionality of 1024 is used. #### For Transformer and GRU model After evaluating the model performance on the validation data, the optimal values of the hyper-parameters were set. We used the following list of hyper- parameters: learning rate = 0.0005, transformer encoder layer = 1, dropout rate = 0.1, optimizer = Adam, loss function = Cross-Entropy Loss, and batch size = 32, point wise feed forward dimension (pf_dim) = 2048. ### 4.3 Performance Table 1: Accuracy and weighted F1 score on Tamil Code-Mixed Text Method | Accuracy | weighted F1 ---|---|--- Fine Tuned BERT | 0.65 | 0.53 fastText + Tranformer | 0.66 | 0.57 fastText + ELMO + Transformer | 0.66 | 0.57 fastText + ELMO + TF-IDF + Transformer | 0.67 | 0.58 Table 2: Accuracy and weighted F1 score on Malayalam Code-Mixed Text Method | Accuracy | weighted F1 ---|---|--- Fine Tuned BERT | 0.51 | 0.46 fastText + Tranformer | 0.47 | 0.45 fastText + ELMO + Transformer | 0.50 | 0.47 fastText + ELMO + TF-IDF + Transformer | 0.67 | 0.66 We evaluated the performance of the method using weighted F1. The model performed well in classifying positive and not-language comments. The results are given in table 1 and 2 The positive comments had a lot of corpora to train. It made the classification of positive comments an easier task. The not_Malayalam and not_Tamil tweets had another language words in the data, as these language words had higher TF-IDF scores w.r.t the non-language label, their classification was straight-forward. We observed that the system could not identify the sentiment when sarcasm is used in the negative polarity comments. The words in the sarcasm are similar to those of positive comments. It made the sentiment analysis a difficult task. Mixed feelings had both positive and negative sentences. As the classifier was trained on a lot of positive corpora, it could not deduce the negative polarity sentences with sarcasm and irony imbibed in them. Thus the classifier labeled them as positive. It affected the performance of the classifier. More training data could help resolve such issues. ## 5 Conclusion This paper describes the approach we proposed for the Dravidian Code-Mixed FIRE-2020 task: Sentiment Analysis for Davidian Languages in Code-Mixed Text. We proposed meta embeddings with the transformer and GRU model for the sentiment analysis of Dravidian code mixed data set given in the shared task. Our model obtained 0.58 and 0.66 average-F1 for Tamil and Malayalam code-mixed datasets, respectively. We observed that the proposed model did a good job distinguishing positive and not_Malayalam and not_Tamil youtube comments. For future work, we will explore our model’s performance with larger corpora. As we observed sarcasm and irony in negative polarity sentences, we feel that it would be interesting to focus on techniques to detect irony and sarcasm in a code-mixed scenario. ## References * Wardhaugh [2011] R. Wardhaugh, An introduction to sociolinguistics, volume 28, John Wiley & Sons, 2011. * Bali et al. [2014] K. Bali, J. Sharma, M. Choudhury, Y. Vyas, “i am borrowing ya mixing?" An Analysis of English-Hindi Code Mixing in Facebook, in: Proceedings of the First Workshop on Computational Approaches to Code Switching, 2014, pp. 116–126. * Gysels [1992] M. Gysels, French in urban lubumbashi swahili: Codeswitching, borrowing, or both?, Journal of Multilingual & Multicultural Development 13 (1992) 41–55. * Khanuja et al. [2020] S. Khanuja, S. Dandapat, A. Srinivasan, S. Sitaram, M. Choudhury, Gluecos: An evaluation benchmark for code-switched nlp, arXiv preprint arXiv:2004.12376 (2020). * Barman et al. [2014] U. Barman, A. Das, J. Wagner, J. Foster, Code mixing: A challenge for language identification in the language of social media, in: Proceedings of the first workshop on computational approaches to code switching, 2014, pp. 13–23. * Gella et al. [2014] S. Gella, K. Bali, M. Choudhury, “ye word kis lang ka hai bhai?” testing the limits of word level language identification, in: Proceedings of the 11th International Conference on Natural Language Processing, 2014, pp. 368–377. * Sharma et al. [2016] A. Sharma, S. Gupta, R. Motlani, P. Bansal, M. Srivastava, R. Mamidi, D. M. Sharma, Shallow parsing pipeline for hindi-english code-mixed social media text, arXiv preprint arXiv:1604.03136 (2016). * Nelakuditi et al. [2016] K. Nelakuditi, D. S. Jitta, R. Mamidi, Part-of-speech tagging for code mixed english-telugu social media data, in: International Conference on Intelligent Text Processing and Computational Linguistics, Springer, 2016, pp. 332–342. * Vyas et al. [2014] Y. Vyas, S. Gella, J. Sharma, K. Bali, M. Choudhury, Pos tagging of english-hindi code-mixed social media content, in: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2014, pp. 974–979. * Priyadharshini et al. [2020] R. Priyadharshini, B. R. Chakravarthi, M. Vegupatti, J. P. McCrae, Named entity recognition for code-mixed indian corpus using meta embedding, in: 2020 6th International Conference on Advanced Computing and Communication Systems (ICACCS), IEEE, 2020, pp. 68–72. * Liu [2020] B. Liu, Sentiment analysis: Mining opinions, sentiments, and emotions, Cambridge university press, 2020\. * Tayal and Yadav [2017] D. K. Tayal, S. K. Yadav, Sentiment analysis on social campaign “swachh bharat abhiyan” using unigram method, AI & SOCIETY 32 (2017) 633–645. * Godbole et al. [2007] N. Godbole, M. Srinivasaiah, S. Skiena, Large-scale sentiment analysis for news and blogs., Icwsm 7 (2007) 219–222. * Patwa et al. [2020] P. Patwa, G. Aguilar, S. Kar, S. Pandey, S. PYKL, B. Gambäck, T. Chakraborty, T. Solorio, A. Das, Semeval-2020 task 9: Overview of sentiment analysis of code-mixed tweets, arXiv preprint arXiv:2008.04277 (2020). * Mishra et al. [2018] P. Mishra, P. Danda, P. Dhakras, Code-mixed sentiment analysis using machine learning and neural network approaches, arXiv preprint arXiv:1808.03299 (2018). * Prabhu et al. [2016] A. Prabhu, A. Joshi, M. Shrivastava, V. Varma, Towards sub-word level compositions for sentiment analysis of hindi-english code mixed text, arXiv preprint arXiv:1611.00472 (2016). * Kudo and Richardson [2018] T. Kudo, J. Richardson, Sentencepiece: A simple and language independent subword tokenizer and detokenizer for neural text processing, arXiv preprint arXiv:1808.06226 (2018). * Sennrich et al. [2015] R. Sennrich, B. Haddow, A. Birch, Neural machine translation of rare words with subword units, arXiv preprint arXiv:1508.07909 (2015). * Kudo [2018] T. Kudo, Subword regularization: Improving neural network translation models with multiple subword candidates, arXiv preprint arXiv:1804.10959 (2018). * Pratapa et al. [2018] A. Pratapa, M. Choudhury, S. Sitaram, Word embeddings for code-mixed language processing, in: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018, pp. 3067–3072. * Bojanowski et al. [2017] P. Bojanowski, E. Grave, A. Joulin, T. Mikolov, Enriching word vectors with subword information, Transactions of the Association for Computational Linguistics 5 (2017) 135–146. * Peters et al. [2018] M. E. Peters, M. Neumann, M. Iyyer, M. Gardner, C. Clark, K. Lee, L. Zettlemoyer, Deep contextualized word representations, arXiv preprint arXiv:1802.05365 (2018). * Aizawa [2003] A. Aizawa, An information-theoretic perspective of tf–idf measures, Information Processing & Management 39 (2003) 45–65. * Chakravarthi et al. [2020a] B. R. Chakravarthi, R. Priyadharshini, V. Muralidaran, S. Suryawanshi, N. Jose, E. Sherly, J. P. McCrae, Overview of the track on Sentiment Analysis for Dravidian Languages in Code-Mixed Text, in: Working Notes of the Forum for Information Retrieval Evaluation (FIRE 2020). CEUR Workshop Proceedings. In: CEUR-WS. org, Hyderabad, India, 2020a. * Chakravarthi et al. [2020b] B. R. Chakravarthi, R. Priyadharshini, V. Muralidaran, S. Suryawanshi, N. Jose, E. Sherly, J. P. McCrae, Overview of the track on Sentiment Analysis for Dravidian Languages in Code-Mixed Text, in: Proceedings of the 12th Forum for Information Retrieval Evaluation, FIRE ’20, 2020b. * Chakravarthi et al. [2020c] B. R. Chakravarthi, N. Jose, S. Suryawanshi, E. Sherly, J. P. McCrae, A sentiment analysis dataset for code-mixed Malayalam-English, in: Proceedings of the 1st Joint Workshop on Spoken Language Technologies for Under-resourced languages (SLTU) and Collaboration and Computing for Under-Resourced Languages (CCURL), European Language Resources association, Marseille, France, 2020c, pp. 177–184. URL: https://www.aclweb.org/anthology/2020.sltu-1.25. * Chakravarthi et al. [2020d] B. R. Chakravarthi, V. Muralidaran, R. Priyadharshini, J. P. McCrae, Corpus creation for sentiment analysis in code-mixed Tamil-English text, in: Proceedings of the 1st Joint Workshop on Spoken Language Technologies for Under-resourced languages (SLTU) and Collaboration and Computing for Under-Resourced Languages (CCURL), European Language Resources association, Marseille, France, 2020d, pp. 202–210. URL: https://www.aclweb.org/anthology/2020.sltu-1.28. * Chakravarthi [2020] B. R. Chakravarthi, Leveraging orthographic information to improve machine translation of under-resourced languages, Ph.D. thesis, NUI Galway, 2020.
# High-efficient two-step entanglement purification using hyperentanglement Lan Zhou,1 Yu-Bo<EMAIL_ADDRESS>1 School of Science, Nanjing University of Posts and Telecommunications, Nanjing, 210003, China 2Institute of Quantum Information and Technology, Nanjing University of Posts and Telecommunications, Nanjing, 210003, China ###### Abstract Entanglement purification is a powerful method to distill the high-quality entanglement from low-quality entanglement. In the paper, we propose an efficient two-step entanglement purification protocol (EPP) for the polarization entanglement by using only one copy of two-photon hyperentangled state in polarization, spatial-mode, and time-bin DOFs. We suppose that the entanglement in all DOFs suffer from channel noise. In two purification steps, the parties can reduce the bit-flip error and phase-flip error in polarization DOF by consuming the imperfect entanglement in the spatial-mode and time-bin DOFs, respectively. This EPP effectively reduces the consumption of entanglement pairs and the experimental difficulty. Moreover, if consider the practical photon transmission and detector efficiencies, our EPP has much higher purification efficiency than previous recurrence EPPs. Meanwhile, when one or two purification steps fail, the distilled mixed state may have residual entanglement. Taking use of the residual entanglement, the parties may still distill higher-quality polarization entanglement. Even if not, they can still reuse the residual entanglement in the next purification round. The existence of residual entanglement benefits for increasing the yield of the EPP. All the above advantages make our EPP have potential application in future quantum information processing. ###### pacs: 03.67.Pp, 03.67.Hk, 03.65.Ud ## I Introduction Entanglement is an indispensable resource which is widely applied in quantum communication field, such as quantum teleportationteleportation1 ; teleportation2 ; teleportation3 , quantum repeater repeater1 ; repeater2 ; repeater3 , quantum key distribution (QKD) qkd , quantum secret sharing (QSS) qss , and quantum secure direct communication (QSDC) qsdc1 ; qsdc2 ; qsdc3 ; qsdc4 . The above applications often require the maximal entanglement. However, entanglement is generally fragile due to the channel noise. During the practical applications, the degraded entanglement may decrease the quantum communication efficiency or even make quantum communication insecure. Entanglement purification which was first proposed by Bennett _et al._ in 1996 EPP0 is an efficient method to distill high quality entanglement from low quality entanglement with local operation and classical communications (LOCC). Recurrence entanglement purification is the most common entanglement purification form, which has been well developed in both theory and experiment EPP0 ; Deustch ; Murao ; Pan1 ; Pan2 ; Pan3 ; graph ; atom ; Cheong ; sheng1 ; sheng2 ; Zhou1 ; Wang ; Zhou2 ; Zhang ; Dur ; Rozpeedek ; Krastanov ; Wu ; zhouap ; zhouoe ; hu ; Du ; nest ; network ; shenghyper1 ; shenghyper2 . The recurrence entanglement purification protocols (EPPs) require two or more copies of low-quality entangled states from the same enables. After two communication parties in distant locations performing the controlled-not (CNOT) or other similar operations, one pair of low-quality entangled state is measured. If the purification is successful, the fidelity of left photon pair can be increased. For example, in 2001, Pan _et al._ presented an EPP of general mixed entangled states with linear optical elements Pan1 , and later they improved their EPP by adopting available parametric down conversion sources Pan2 . In 2003, Pan _et al._ demonstrated the experiment of the entanglement purification for general mixed states of polarization-entangled photons Pan3 . In 2008, Sheng _et al._ proposed an EPP based on nondestructive quantum nondemolition detectors sheng1 . In 2017, Pan _et al._ experimentally realized the nested purification for a linear optical quantum repeater nest . In addition, the experimental purification between two-atom entanglement and solid state quantum network nodes were also demonstrated atom ; network . In 2010, Sheng _et al._ proposed the deterministic EPPs (DEPPs) by adopting the hyperentanglement shenghyper1 ; shenghyper2 . Although the recurrence entanglement purification has been well studied, existing recurrence EPPs often have relatively low yield. The reason is that in each purification round, at least one pair of low-quality entangled states should be consumed. In practical applications, entanglement purification process often has to be iterated for many rounds to obtain high-fidelity entangled pairs, so that a large amount of low-quality entangled pairs have to be consumed. It is a big waste of the precious entanglement resources. In 2021, Hu _et al._ proposed and experimentally demonstrated the first long- distance polarization entanglement recurrence purification using only one polarization-spatial-mode hyperentangled photon pair hu . They supposed that the entanglement in both polarization and spatial-mode DOFs suffer from one kind of error, say, the bit-flip error or phase-flip error. After performing the EPP, they obtained a significant improvement in the fidelity of polarization entanglement. This protocol effectively reduces consumption of copies of entanglement pairs, especially in purification consisting of many rounds. Actually, in practical entanglement distribution process, the bit-flip error and phase-flip error may be occurred simultaneously. In the paper, we consider a more general recurrence EPP which can simultaneously reduce the bit-flip error and phase-flip error of the polarization entanglement by using only one pair of polarization-spatial-time-bin hyperentangled photon pair. We choose the spatial mode and time-bin entanglement for the entanglement in both two DOFs, especially in time-bin DOF being highly robust to the channel noise. The entanglement in time-bin DOF has been successfully used in the transmission of qubits over hundreds of kilometers Pan2 ; time1 ; time2 ; time3 ; time4 and in teleportation using real-world fiber networks time5 ; time6 . In 2005, Barreiro _et al._ experimentally demonstrated the generation of hyperentanglement in polarization, spatial-mode and time-energy DOFs of photon systems using pairs of photons produced in spontaneous parametric down- conversion generation5 . In our protocol, we suppose that the entanglement in all DOFs suffer from channel noise and degrade to mixed states. As the entanglement in above three DOFs have different noise robustness, after the photon transmission, the fidelities in three DOFs are naturally different. After performing our EPP, we can efficiently reduce both the bit-flip error and phase-flip error rate in polarization DOF by consuming the imperfect entanglement in the spatial-mode and time-bin DOFs. Moreover, we will prove that if a purification step fails, there may exist residual entanglement in the corresponding distilled mixed state. By using the residual entanglement, we may also increase the fidelity of the polarization entanglement after the whole purification process. The paper is organized as follows. In Sec. II, we describe our EPP in a simple case where the entanglement in spatial-mode and time-bin DOFs only suffer from a bit-flip error. In Sec. III, we extend our EPP to a general case where both the entanglement in spatial-mode and time-bin DOFs suffer from both bit-flip error and phase-flip error. In Sec. IV, we make a discussion. In Sec. V, we make a conclusion. ## II Entanglement purification principle In this section, we propose our long-distance EPP using a hyperentangled photon pair. Suppose the photon hyperentanglement source S generates a two- photon hyperentanglement in polarization, spatial-mode, and time-bin DOFs, which can be described as $\displaystyle|\Phi^{+}_{p}\rangle|\Phi^{+}_{s}\rangle|\Phi^{+}_{t}\rangle=\frac{1}{\sqrt{2}}(|HH\rangle+|VV\rangle)$ (1) $\displaystyle\otimes$ $\displaystyle\frac{1}{\sqrt{2}}(|a_{1}^{\prime}b_{1}^{\prime}\rangle+|a_{2}^{\prime}b_{2}^{\prime}\rangle)\otimes\frac{1}{\sqrt{2}}(|LL\rangle+|SS\rangle).$ Here, $H$ ($V$) represents the horizontal (vertical) polarization, $a_{1}^{\prime}$, $a_{2}^{\prime}$, $b_{1}^{\prime}$ and $b_{2}^{\prime}$ are four different spatial modes, and $L$ ($S$) represents the long (short) time- bin. The photon in $a_{1}^{\prime}$ and $a_{2}^{\prime}$ are sent to Alice, while the photon in $b_{1}^{\prime}$ and $b_{2}^{\prime}$ are sent to Bob. After long distance transmission, the channel noise may degrade the entanglement in all DOFs. Here, we suppose that the entanglement in polarization DOF degrade to a Werner state with the form of $\displaystyle\rho_{p}$ $\displaystyle=$ $\displaystyle p_{p}|\Phi^{+}_{p}\rangle\langle\Phi^{+}_{p}|+\frac{1-p_{p}}{3}(|\Psi^{+}_{p}\rangle\langle\Psi^{+}_{p}|$ (2) $\displaystyle+$ $\displaystyle|\Phi^{-}_{p}\rangle\langle\Phi^{-}_{p}|+|\Psi^{-}_{p}\rangle\langle\Psi^{-}_{p}|),$ where $\displaystyle|\Phi^{\pm}_{p}\rangle$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|HH\rangle\pm|VV\rangle),$ $\displaystyle|\Psi^{\pm}_{p}\rangle$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|HV\rangle\pm|VH\rangle).$ (3) We notice that the state $|\Phi^{+}_{p}\rangle$ becomes $|\Psi^{+}_{p}\rangle$ when a bit-flip error occurs, $|\Phi^{+}_{p}\rangle$ becomes $|\Phi^{-}_{p}\rangle$ when a phase-flip error occurs, and $|\Phi^{+}_{p}\rangle$ becomes $|\Psi^{-}_{p}\rangle$ when bit-flip error and phase-flip error both occur. Considering the noise robustness of the entanglement in spatial-mode and time- bin DOFs are higher than that in the polarization DOF, we first focus on a simple case where the entanglement in spatial-mode and time-bin DOFs only suffer from bit-flip error. In this case, the entanglement in both DOFs degrade to $\displaystyle\rho_{s}$ $\displaystyle=$ $\displaystyle p_{s}|\Phi^{+}_{s}\rangle\langle\Phi^{+}_{s}|+(1-p_{s})|\Psi^{+}_{s}\rangle\langle\Psi^{+}_{s}|,$ (4) $\displaystyle\rho_{t}$ $\displaystyle=$ $\displaystyle p_{t}|\Phi^{+}_{t}\rangle\langle\Phi^{+}_{t}|+(1-p_{t})|\Psi^{+}_{t}\rangle\langle\Psi^{+}_{t}|,$ (5) where $\displaystyle|\Phi^{\pm}_{s}\rangle$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|a_{1}b_{1}\rangle\pm|a_{2}b_{2}\rangle),$ $\displaystyle|\Psi^{\pm}_{s}\rangle$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|a_{1}b_{2}\rangle\pm|a_{2}b_{1}\rangle),$ $\displaystyle|\Phi^{\pm}_{t}\rangle$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|LL\rangle\pm|SS\rangle),$ $\displaystyle|\Psi^{\pm}_{t}\rangle$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(|LS\rangle\pm|SL\rangle).$ (6) Therefore, the initial state in Eq. (1) degrades to $\rho_{p}\otimes\rho_{s}\otimes\rho_{t}$. Here, we suppose that all the fidelity $p_{p}$, $p_{s}$, and $p_{t}$ are higher than $\frac{1}{2}$. The schematic principle of our EPP is shown in Fig. 1. The protocol includes two steps. In the first step, Alice and Bob correct the bit-flip error in polarization DOF by consuming the entanglement in spatial-mode DOF. In the second step, they correct the phase-flip error in polarization DOF with the help of the entanglement in time-bin DOF. Figure 1: The basic principle of our two-step EPP. Suppose a photon pair hyperentangled in polarization, spatial-mode, and time-bin DOFs suffer from channel noise, and the entanglement in all DOFs degrade to mixed states. Here, PBS means polarization beam splitter, which can totally transmit the photon in $|H\rangle$ and reflect the photon in $|V\rangle$. QWP means the $\lambda/4$-wave plate, which can make the Hadamard (H) operation in the polarization DOF. $PC_{l(s)}$ presents the Pockels cell, which can revers the polarization of a photon with the time-bin L (S). $D_{1}-D_{8}$ are the single photon detectors. In the first step, we only use the states in polarization and spatial-mode DOF and leave the state in time-bin DOF unchanged, so that we first neglect the state in time-bin DOF and only consider the mixed state as $\rho_{1}=\rho_{p}\otimes\rho_{s}$ for simplicity. There are totally eight possible cases. The photon system may be in $|\Phi^{+}_{p}\rangle|\Phi^{+}_{s}\rangle$ with the probability of $p_{p}p_{s}$, while it may be in $|\Phi^{+}_{p}\rangle|\Psi^{+}_{s}\rangle$ with the probability of $p_{p}(1-p_{s})$. Meanwhile, the photon system may be in $|\Psi^{+}_{p}\rangle|\Phi^{+}_{s}\rangle$, $|\Phi^{-}_{p}\rangle|\Phi^{+}_{s}\rangle$, or $|\Psi^{-}_{p}\rangle|\Phi^{+}_{s}\rangle$ with the equal probability of $\frac{(1-p_{p})p_{s}}{3}$, and it may be in the state $|\Psi^{+}_{p}\rangle|\Psi^{+}_{s}\rangle$, $|\Phi^{-}_{p}\rangle|\Psi^{+}_{s}\rangle$, or $|\Psi^{-}_{p}\rangle|\Psi^{+}_{s}\rangle$ with the probability of $\frac{(1-p_{p})(1-p_{s})}{3}$. Alice and Bob pass the photons in $a_{1}$ and $a_{2}$, $b_{1}$ and $b_{2}$ spatial modes through two polarization beam splitters (PBSs), which can totally transmit the photon in $|H\rangle$ and reflect the photon in $|V\rangle$. If the initial photon state is $|\Phi^{\pm}_{p}\rangle|\Phi^{+}_{s}\rangle$, after the PBS, the state will evolve to $\displaystyle|\Phi^{\pm}_{p}\rangle|\Phi^{+}_{s}\rangle=\frac{1}{\sqrt{2}}(|HH\rangle\pm|VV\rangle)\otimes\frac{1}{\sqrt{2}}(|a_{1}b_{1}\rangle+|a_{2}b_{2}\rangle)$ (7) $\displaystyle\rightarrow$ $\displaystyle\frac{1}{\sqrt{2}}(|HH\rangle\pm|VV\rangle)\otimes\frac{1}{\sqrt{2}}(|a_{3}b_{3}\rangle+|a_{4}b_{4}\rangle).$ When the initial state is $|\Psi^{\pm}_{p}\rangle|\Psi^{+}_{s}\rangle$, it will evolve to $\displaystyle|\Psi^{\pm}_{p}\rangle|\Psi^{+}_{s}\rangle=\frac{1}{\sqrt{2}}(|HV\rangle\pm|VH\rangle)\otimes\frac{1}{\sqrt{2}}(|a_{1}b_{2}\rangle+|a_{2}b_{1}\rangle)$ (8) $\displaystyle\rightarrow$ $\displaystyle\frac{1}{\sqrt{2}}(|HV\rangle\pm|VH\rangle)\otimes\frac{1}{\sqrt{2}}(|a_{3}b_{3}\rangle+|a_{4}b_{4}\rangle).$ All the four above states make the spatial modes $a_{3}b_{3}$ or $a_{4}b_{4}$ each have one photon. In these cases, the first step is successful. On the other hand, if the initial state is one of the other four cases, say, $|\Phi^{\pm}_{p}\rangle|\Psi^{+}_{s}\rangle$ and $|\Psi^{\pm}_{p}\rangle|\Phi^{+}_{s}\rangle$, after the PBSs, we can obtain the spatial-modes $a_{3}b_{4}$ or $a_{4}b_{3}$ each has one photon and the first step fails. As a result, when the first step is successful, we can distill a new mixed state as $\displaystyle\rho_{1p}$ $\displaystyle=$ $\displaystyle F_{1}|\Phi^{+}_{p}\rangle\langle\Phi^{+}_{p}|+F_{2}|\Phi^{-}_{p}\rangle\langle\Phi^{-}_{p}|+F_{3}(|\Psi^{+}_{p}\rangle\langle\Psi^{+}_{p}|$ (9) $\displaystyle+$ $\displaystyle|\Psi^{-}_{p}\rangle\langle\Psi^{-}_{p}|),$ in the spatial modes $a_{3}b_{3}$ or $a_{4}b_{4}$ with the success probability of $\displaystyle P_{1}$ $\displaystyle=$ $\displaystyle p_{p}p_{s}+\frac{1-p_{p}}{3}[p_{s}+2(1-p_{s})]$ (10) $\displaystyle=$ $\displaystyle\frac{1}{3}(4p_{p}p_{s}-p_{s}-2p_{p}+2).$ The four coefficients in Eq. (9) can be written as $\displaystyle F_{1}$ $\displaystyle=$ $\displaystyle\frac{p_{p}p_{s}}{P_{1}}=\frac{3p_{p}p_{s}}{4p_{p}p_{s}-p_{s}-2p_{p}+2},$ $\displaystyle F_{2}$ $\displaystyle=$ $\displaystyle\frac{(1-p_{p})p_{s}}{3P_{1}}=\frac{(1-p_{p})p_{s}}{4p_{p}p_{s}-p_{s}-2p_{p}+2},$ $\displaystyle F_{3}$ $\displaystyle=$ $\displaystyle\frac{(1-p_{p})(1-p_{s})}{3P_{1}}=\frac{(1-p_{p})(1-p_{s})}{4p_{p}p_{s}-p_{s}-2p_{p}+2}.$ (11) It is obvious that when $p_{p}>\frac{1}{2}$ and $p_{s}>\frac{1}{2}$, the rate of $|\Psi^{\pm}_{p}\rangle$ ($F_{3}$) is smaller than their original rate $\frac{(1-p_{p})}{3}$. As a result, the first step can reduce the rate of $|\Psi^{\pm}_{p}\rangle$. Moreover, the reduction of bit-flip error directly increases the fidelity of $|\Phi^{+}_{p}\rangle$. We can obtain $F_{1}>p_{p}$ and $F_{1}>p_{s}$ when $p_{p}>\frac{1}{2}$ and $\frac{5p_{p}-2}{4p_{p}-1}>p_{s}>\frac{1}{2}$. However, the first step cannot deal with the phase-flip error ($|\Phi^{-}_{p}\rangle$) and the rate of $|\Phi^{-}_{p}\rangle$ is still in a relatively high level. Next, we try to correct the phase-flip error. In the second step, we require to consume the entanglement in the time-bin DOF. After the first step, the whole photon system collapse to $\rho_{1p}\otimes\rho_{t}$ in $a_{3}b_{3}$ or $a_{4}b_{4}$ modes. We first consider that the photons are in $a_{3}b_{3}$. Alice and Bob first pass the photons in $a_{3}b_{3}$ modes through two $\lambda/4$-wave plates (QWPs), respectively. The QWP performs a Hadamard (H) operation in the polarization DOF, which makes $|H\rangle\rightarrow\frac{1}{\sqrt{2}}(|H\rangle+|V\rangle)$ and $|V\rangle\rightarrow\frac{1}{\sqrt{2}}(|H\rangle-|V\rangle)$. After the H operation, $|\Phi^{-}_{p}\rangle\leftrightarrow|\Psi^{+}_{p}\rangle$, while $|\Phi^{+}_{p}\rangle$ and $|\Psi^{-}_{p}\rangle$ keep unchanged. In this way, they can transform $\rho_{1p}$ to $\displaystyle\rho_{2p}$ $\displaystyle=$ $\displaystyle F_{1}|\Phi^{+}_{p}\rangle\langle\Phi^{+}_{p}|+F_{2}|\Psi^{+}_{p}\rangle\langle\Psi^{+}_{p}|+F_{3}(|\Phi^{-}_{p}\rangle\langle\Phi^{-}_{p}|$ (12) $\displaystyle+$ $\displaystyle|\Psi^{-}_{p}\rangle\langle\Psi^{-}_{p}|)$ in the spatial modes $a_{5}b_{5}$. The whole photon system $\rho_{2p}\otimes\rho_{t}$ can be described as follows. With probability of $F_{1}p_{t}$ and $F_{1}(1-p_{t})$ the photon pair is in $|\Phi^{+}_{p}\rangle|\Phi^{+}_{t}\rangle$ and $|\Phi^{+}_{p}\rangle|\Psi^{+}_{t}\rangle$, respectively. With a probability of $F_{2}p_{t}$ and $F_{2}(1-p_{t})$, it is in $|\Psi^{+}_{p}\rangle|\Phi^{+}_{t}\rangle$ and $|\Psi^{+}_{p}\rangle|\Psi^{+}_{t}\rangle$, respectively. On the other hand, the whole system is in $|\Phi^{-}_{p}\rangle|\Phi^{+}_{t}\rangle$ or $|\Psi^{-}_{p}\rangle|\Phi^{+}_{t}\rangle$ with an equal probability of $F_{3}p_{t}$, and in $|\Phi^{-}_{p}\rangle|\Psi^{+}_{t}\rangle$ or $|\Psi^{-}_{p}\rangle|\Psi^{+}_{t}\rangle$ with an equal probability of $F_{3}(1-p_{t})$. Suppose that the photon pair is in $|\Phi^{+}_{p}\rangle_{a_{5}b_{5}}|\Phi^{+}_{t}\rangle$. As shown in Fig. 1, Alice (Bob) passes the photon in $a_{5}$ $(b_{5})$ through a Pockels cell ($PC_{S}$), which can flip the polarization feature of the incoming photon under the temporal mode $S$. After the $PC_{s}$, Alice (Bob) passes the photon through a PBS, which makes the photon in $|H\rangle$ be in $a_{7}$ $(b_{7})$ and enter a $PC_{S}$ and the photon in $|V\rangle$ be in $a_{8}$ $(b_{8})$ and enter a $PC_{L}$. As a result, the states $|\Phi^{\pm}_{p}\rangle|\Phi^{+}_{t}\rangle$ will finally evolve to $\displaystyle|\Phi^{\pm}_{p}\rangle_{a_{3}b_{3}}|\Phi^{+}_{t}\rangle$ (13) $\displaystyle\rightarrow$ $\displaystyle\frac{1}{2}(|H^{L}H^{L}\rangle_{a_{9}b_{9}}+|V^{S}V^{S}\rangle_{a_{10}b_{10}}$ $\displaystyle\pm$ $\displaystyle|H^{L}H^{L}\rangle_{a_{10}b_{10}}\pm|V^{S}V^{S}\rangle_{a_{9}b_{9}}).$ Then, with the help of two PBSs, Alice and Bob can make the photon in $|H\rangle$ pass through the short (S) arm and photon in $|V\rangle$ pass through the long (L) arm. By precisely controlling the length of long and short arms, they can adjust the time-bin feature of the photons in $|H\rangle$ and $|V\rangle$ to be the same. In this way, we can neglect the time-bin features of the photons, and the state in Eq. (13) evolves to $\displaystyle|\Phi^{\pm}_{p}\rangle_{a_{3}b_{3}}|\Phi^{+}_{t}\rangle$ (14) $\displaystyle\rightarrow$ $\displaystyle\frac{1}{\sqrt{2}}(|HH\rangle\pm|VV\rangle)\otimes\frac{1}{\sqrt{2}}(|d_{2}d_{6}\rangle\pm|d_{1}d_{5}\rangle),$ and will be detected by the single photon detector $D_{2}D_{6}$ or $D_{1}D_{5}$. In this case, the second purification step is successful. The polarization feature of the photon pair remains to be $|\Phi^{\pm}_{p}\rangle$. If the initial state is $|\Psi^{\pm}_{p}\rangle|\Psi^{+}_{t}\rangle$, after the above operations, it will evolve to $\displaystyle|\Psi^{\pm}_{p}\rangle|\Psi^{+}_{t}\rangle$ (15) $\displaystyle\rightarrow$ $\displaystyle\frac{1}{\sqrt{2}}(|HV\rangle\pm|VH\rangle)\otimes\frac{1}{\sqrt{2}}(|d_{2}d_{6}\rangle\pm|d_{1}d_{5}\rangle),$ which will also lead to the successful detection results. The state in Eq. (15) will finally collapse to $|\Psi^{\pm}_{p}\rangle$. If the initial state is $|\Phi^{\pm}_{p}\rangle|\Psi^{+}_{t}\rangle$ or $|\Psi^{\pm}_{p}\rangle|\Phi^{+}_{t}\rangle$, after the above operations, Alice and Bob would never obtain the successful measurement results. In detail, after above operations, we can obtain $\displaystyle|\Phi^{\pm}_{p}\rangle|\Psi^{+}_{t}\rangle$ (16) $\displaystyle\rightarrow$ $\displaystyle\frac{1}{\sqrt{2}}(|HV\rangle\pm|VH\rangle)\otimes\frac{1}{\sqrt{2}}(|d_{2}d_{5}\rangle\pm|d_{1}d_{6}\rangle),$ $\displaystyle|\Psi^{\pm}_{p}\rangle|\Phi^{+}_{t}\rangle$ $\displaystyle\rightarrow$ $\displaystyle\frac{1}{\sqrt{2}}(|HH\rangle\pm|VV\rangle)\otimes\frac{1}{\sqrt{2}}(|d_{2}d_{5}\rangle\pm|d_{1}d_{6}\rangle),$ which makes the single photon detectors $D_{1}D_{6}$ or $D_{2}D_{5}$ each register a single photon. In this case, the second purification step fails. On the other hand, if the photon state is in $a_{4}b_{4}$ modes, we can obtain when the photon detectors $D_{3}D_{7}$ or $D_{4}D_{8}$ each register one photon, the second purification step is successful. Therefore, when the second purification step is successful, we can distill a new mixed state in polarization DOF as $\displaystyle\rho_{3p}$ $\displaystyle=$ $\displaystyle F^{\prime}_{1}|\Phi^{+}_{p}\rangle\langle\Phi^{+}_{p}|+F^{\prime}_{2}|\Phi^{-}_{p}\rangle\langle\Phi^{-}_{p}|$ (17) $\displaystyle+$ $\displaystyle F^{\prime}_{3}|\Psi^{+}_{p}\rangle\langle\Psi^{+}_{p}|+F^{\prime}_{4}|\Psi^{-}_{p}\rangle\langle\Psi^{-}_{p}|,$ with a success probability of $\displaystyle P_{2}$ $\displaystyle=$ $\displaystyle(F_{1}+F_{3})p_{t}+(F_{2}+F_{3})(1-p_{t})$ (18) $\displaystyle=$ $\displaystyle\frac{1-p_{p}-p_{s}p_{t}+4p_{p}p_{s}p_{t}}{2-2p_{p}-p_{s}+4p_{p}p_{s}}.$ The four coefficients in Eq. (17) can be written as $\displaystyle F^{\prime}_{1}$ $\displaystyle=$ $\displaystyle\frac{F_{1}p_{t}}{P_{2}}=\frac{3p_{p}p_{s}p_{t}}{1-p_{t}p_{s}+p_{p}(4p_{t}p_{s}-1)},$ $\displaystyle F^{\prime}_{2}$ $\displaystyle=$ $\displaystyle\frac{F_{3}p_{t}}{P_{2}}=\frac{(1-p_{p})(1-p_{s})p_{t}}{1-p_{t}p_{s}+p_{p}(4p_{t}p_{s}-1)},$ $\displaystyle F^{\prime}_{3}$ $\displaystyle=$ $\displaystyle\frac{F_{2}(1-p_{t})}{P_{2}}=\frac{(1-p_{p})(1-p_{t})p_{s}}{1-p_{t}p_{s}+p_{p}(4p_{t}p_{s}-1)},$ $\displaystyle F^{\prime}_{4}$ $\displaystyle=$ $\displaystyle\frac{F_{3}(1-p_{t})}{P_{2}}=\frac{(1-p_{p})(1-p_{s})(1-p_{t})}{1-p_{t}p_{s}+p_{p}(4p_{t}p_{s}-1)}.$ (19) Similarly as the first step, the second step can reduce the rate of $|\Psi^{\pm}_{p}\rangle$ ($F^{\prime}_{3}<F_{2}$ and $F^{\prime}_{4}<F_{3}$). It can be calculated that $F^{\prime}_{1}>F_{1}$ and $F^{\prime}_{1}>p_{t}$ when $\frac{3p_{p}p_{s}}{+}p_{p}-1{p_{s}(4p_{p}-1)}>p_{t}>\frac{1}{2}$. Comparing with original mixed state $\rho_{p}$ in Eq. (2), after two steps of purification, the rates of $|\Psi^{\pm}\rangle$ and $|\Phi^{-}\rangle$ can be all reduced, so that the fidelity of $|\Phi^{+}_{p}\rangle$ can be efficiently increased. ## III General entanglement purification In this section, we consider a general case that after the long-distance transmission in noisy channel, the entanglement in the spatial-mode and time- bin DOFs also degrade to Werner states. In this way, the mixed states in above two DOFs can be written as $\displaystyle\rho_{sn}$ $\displaystyle=$ $\displaystyle p_{s}|\Phi^{+}_{s}\rangle\langle\Phi^{+}_{s}|+\frac{1-p_{s}}{3}(|\Psi^{+}_{s}\rangle\langle\Psi^{+}_{s}|$ $\displaystyle+$ $\displaystyle|\Phi^{-}_{s}\rangle\langle\Phi^{-}_{s}|+|\Psi^{-}_{s}\rangle\langle\Psi^{-}_{s}|),$ $\displaystyle\rho_{tn}$ $\displaystyle=$ $\displaystyle p_{t}|\Phi^{+}_{t}\rangle\langle\Phi^{+}_{t}|+\frac{1-p_{t}}{3}(|\Psi^{+}_{t}\rangle\langle\Psi^{+}_{t}|$ (20) $\displaystyle+$ $\displaystyle|\Phi^{-}_{t}\rangle\langle\Phi^{-}_{t}|+|\Psi^{-}_{t}\rangle\langle\Psi^{-}_{t}|).$ In Eq. (2) and Eq. (20), we also suppose that $p_{p(t,s)}>\frac{1}{2}$. In the first step, we only consider $\rho_{p}\otimes\rho_{sn}$, which has 16 possible cases, say $|\Phi^{\pm}_{p}\rangle|\Phi^{\pm}_{s}\rangle$, $|\Phi^{\pm}_{p}\rangle|\Psi^{\pm}_{s}\rangle$, $|\Psi^{\pm}_{p}\rangle|\Phi^{\pm}_{s}\rangle$, and $|\Psi^{\pm}_{p}\rangle|\Psi^{\pm}_{s}\rangle$. By passing the photons in $a_{1}a_{2}$ and $b_{1}b_{2}$ modes through the PBSs, we also select the items which make the spatial modes $a_{3}b_{3}$ or $a_{4}b_{4}$ each have one photon. All the 8 initial states $|\Phi^{\pm}_{p}\rangle|\Phi^{\pm}_{s}\rangle$ and $|\Psi^{\pm}_{p}\rangle|\Psi^{\pm}_{s}\rangle$ can lead to the successful cases. As a result, when the first step is successful, the parties can distill a new mixed state in polarization DOF as $\displaystyle\rho_{1pn}$ $\displaystyle=$ $\displaystyle F_{1n}|\Phi^{+}_{p}\rangle\langle\Phi^{+}_{p}|+F_{2n}|\Phi^{-}_{p}\rangle\langle\Phi^{-}_{p}|$ (21) $\displaystyle+$ $\displaystyle F_{3n}(|\Psi^{+}_{p}\rangle\langle\Psi^{+}_{p}|+|\Psi^{-}_{p}\rangle\langle\Psi^{-}_{p}|).$ with a probability of $\displaystyle P_{1n}$ $\displaystyle=$ $\displaystyle p_{p}p_{s}+\frac{(1-p_{p})(1-p_{s})}{9}$ (22) $\displaystyle+$ $\displaystyle\frac{p_{p}(1-p_{s})+p_{s}(1-p_{p})}{3}+\frac{4(1-p_{p})(1-p_{s})}{9}$ $\displaystyle=$ $\displaystyle\frac{8p_{p}p_{s}-2p_{p}-2p_{s}+5}{9}.$ In $\rho_{1pn}$, the rates of $|\Phi^{\pm}_{p}\rangle$ and $|\Psi^{\pm}_{p}\rangle$ can be written as $\displaystyle F_{1n}$ $\displaystyle=$ $\displaystyle\frac{p_{p}p_{s}+\frac{(1-p_{p})(1-p_{s})}{9}}{P_{1n}}$ $\displaystyle=$ $\displaystyle\frac{10p_{p}p_{s}-p_{p}-p_{s}+1}{8p_{p}p_{s}-2p_{p}-2p_{s}+5},$ $\displaystyle F_{2n}$ $\displaystyle=$ $\displaystyle\frac{p_{p}(1-p_{s})+(1-p_{p})p_{s}}{3P_{1n}}$ $\displaystyle=$ $\displaystyle\frac{3p_{p}+3p_{s}-6p_{p}p_{s}}{8p_{p}p_{s}-2p_{p}-2p_{s}+5},$ $\displaystyle F_{3n}$ $\displaystyle=$ $\displaystyle\frac{2(1-p_{p})(1-p_{s})}{9P_{1n}}$ (23) $\displaystyle=$ $\displaystyle\frac{2(1-p_{p})(1-p_{s})}{8p_{p}p_{s}-2p_{p}-2p_{s}+5}.$ Figure 2: The value of $F_{1n}$, $F_{2n}$, and $F_{3n}$ as a function of $p_{s}$. Here, we control $p_{p}=0.6$ and adjust $p_{s}$ from 0.505 to 1. In Fig. 2, we show the values of $F_{1n}$, $F_{2n}$, and $F_{3n}$ as a function of $p_{s}$. Here, we control $p_{p}=0.6$ and adjust $p_{s}$ from 0.505 to 1. It is obvious that both $F_{2n}$ and $F_{3n}$ reduce with the growth of $p_{s}$, which makes $F_{1n}$ increase with the growth of $p_{s}$. We also obtain that $F_{3n}<\frac{1-p_{p}}{3}$ when $p_{s}>\frac{1}{2}$, so that we can reduce the rate of bit-flip error. However, in this general case, as the rate of $|\Phi^{-}_{p}\rangle$ ($F_{2n}$) may be relatively high, and we cannot simply obtain $F_{1n}>p_{p}$ or $F_{1n}>p_{s}$ when $p_{s}>\frac{1}{2}$ and $p_{p}>\frac{1}{2}$. Here, we provide the criterion of $F_{1n}>p_{p}$ as $\displaystyle p_{s}>\frac{6p_{p}-2p_{p}^{2}-1}{12p_{p}-8p_{p}^{2}-1},$ (24) and the criterion of $F_{1n}>p_{s}$ as $\displaystyle(8p_{p}-2)p_{s}^{2}+6(1-2p_{p})p_{s}+p_{p}-1<0.$ (25) Figure 3: The minimum threshold of $p_{s}$ corresponding to $F_{1n}>p_{p}$, and the maximum threshold of $p_{s}$ corresponding to $F_{1n}>p_{s}$ as a function of $p_{p}$. Here, we control $p_{p}\in[0.505,0.95]$. Fig. 3 provides the minimum threshold of $p_{s}$ corresponding to $F_{1n}>p_{p}$, and the maximum threshold of $p_{s}$ corresponding to $F_{1n}>p_{s}$ under different values of $p_{p}$. It can be found that both the minimum and maximum thresholds of $p_{s}$ increase with the growth of $p_{p}$. Combined with Fig. 2 and Fig. 3, we can obtain that a high $p_{s}$ can lead to a small $F_{3n}$ and relatively high $F_{1n}$. However, when the value of $p_{s}$ is too high, we cannot ensure that $F_{1n}>p_{s}$. In this way, for satisfying both the criterions in Eq. (24) and Eq. (30), the practical value of $p_{s}$ should be between two thresholds. On the other hand, even when $p_{s}$ is relatively high, $F_{2n}$ can be still in a relatively high level, so that we need to perform the second purification step to further reduce $F_{2n}$ and increase the fidelity of $|\Phi^{+}_{p}\rangle$. In the second purification step, with the help of H operation, we can transform $\rho_{1pn}$ to $\rho_{2pn}$ with the form of $\displaystyle\rho_{2pn}$ $\displaystyle=$ $\displaystyle F_{1n}|\Phi^{+}_{p}\rangle\langle\Phi^{+}_{p}|+F_{3n}|\Phi^{-}_{p}\rangle\langle\Phi^{-}_{p}|$ (26) $\displaystyle+$ $\displaystyle F_{2n}|\Psi^{+}_{p}\rangle\langle\Psi^{+}_{p}|+F_{3n}|\Psi^{-}_{p}\rangle\langle\Psi^{-}_{p}|,$ and $F_{2n}$ transforms to the rate of $|\Psi^{+}_{p}\rangle$. In this way, the whole photon system is $\rho_{2pn}\otimes\rho_{tn}$, which also has 16 possible cases. Alice and Bob pass the photons in $a_{3}b_{3}$ and $a_{4}b_{4}$ modes through the purification units. When the detection result is $D_{2}D_{6}$, $D_{1}D_{5}$, $D_{3}D_{7}$ or $D_{4}D_{8}$ each registering one photon, the second purification step will be successful. According to the description in Sec. II, all the states $|\Phi^{\pm}_{p}\rangle|\Phi^{\pm}_{t}\rangle$ and $|\Psi^{\pm}_{p}\rangle|\Psi^{\pm}_{t}\rangle$ can lead to above successful detection. In this way, the success probability of the second step is $\displaystyle P_{2n}$ $\displaystyle=$ $\displaystyle F_{1n}p_{t}+F_{1n}\frac{1-p_{t}}{3}+F_{3n}p_{t}$ $\displaystyle+$ $\displaystyle 3F_{3n}\frac{1-p_{t}}{3}+2F_{2n}\frac{1-p_{t}}{3}$ $\displaystyle=$ $\displaystyle\frac{7-p_{s}+p_{p}(4p_{s}-1)+2p_{t}(4p_{p}-1)(4p_{s}-1)}{3(8p_{p}p_{s}-2p_{p}-2p_{s}+5)}.$ When the successful detection is obtained, we can distill a new mixed state as $\displaystyle\rho_{3pn}$ $\displaystyle=$ $\displaystyle F^{\prime}_{1n}|\Phi^{+}_{p}\rangle\langle\Phi^{+}_{p}|+F^{\prime}_{2n}|\Phi^{-}_{p}\rangle\langle\Phi^{-}_{p}|$ (28) $\displaystyle+$ $\displaystyle F^{\prime}_{3n}(|\Psi^{+}_{p}\rangle\langle\Psi^{+}_{p}|+|\Psi^{-}_{p}\rangle\langle\Psi^{-}_{p}|),$ where $\displaystyle F^{\prime}_{1n}$ $\displaystyle=$ $\displaystyle\frac{F_{1n}p_{t}+F_{3n}\frac{1-p_{t}}{3}}{P_{2n}}$ $\displaystyle=$ $\displaystyle\frac{2(1-p_{p})(1-p_{s})+p_{t}(1-p_{s}-p_{p}+28p_{p}p_{s})}{7-p_{s}-p_{p}+4p_{p}p_{s}+2p_{t}(4p_{p}-1)(4p_{s}-1)},$ $\displaystyle F^{\prime}_{2n}$ $\displaystyle=$ $\displaystyle\frac{F_{1n}\frac{1-p_{t}}{3}+F_{3n}p_{t}}{P_{2n}}$ $\displaystyle=$ $\displaystyle\frac{p_{s}+p_{p}-10p_{p}p_{s}-1+p_{t}(5p_{s}+5p_{p}+4p_{p}p_{s}-5)}{7-p_{s}-p_{p}+4p_{p}p_{s}+2p_{t}(4p_{p}-1)(4p_{s}-1)},$ $\displaystyle F^{\prime}_{3n}$ $\displaystyle=$ $\displaystyle\frac{(F_{2n}+F_{3n})\frac{1-p_{t}}{3}}{P_{2n}}$ (29) $\displaystyle=$ $\displaystyle\frac{(1-p_{t})(p_{s}+p_{p}+2-4p_{p}p_{s})}{7-p_{s}-p_{p}+4p_{p}p_{s}+2p_{t}(4p_{p}-1)(4p_{s}-1)}.$ Figure 4: The value of $F^{\prime}_{1n}$, $F^{\prime}_{2n}$, and $F^{\prime}_{3n}$ as a function of $p_{t}$. Here, we control $p_{p}=0.6$ and $p_{s}=0.8$, and adjust $p_{t}$ from 0.505 to 1. In Fig. 4, we control $p_{p}=0.6$ and $p_{s}=0.8$ (In this case, we can obtain $F_{1n}\approx 0.7285$), and show the values of $F^{\prime}_{1n}$, $F^{\prime}_{2n}$, and $F^{\prime}_{3n}$ as a function of $p_{t}$. It can be found that both $F^{\prime}_{2n}$ and $F^{\prime}_{3n}$ reduce with the growth of $p_{t}$, which makes $F^{\prime}_{1n}$ increase. The higher value of $p_{t}$ leads to higher $F^{\prime}_{1n}$ and lower $F^{\prime}_{2n}$ and $F^{\prime}_{3n}$. It is important to compare $F^{\prime}_{1n}$ with $F_{1n}$ and $p_{t}$. We can also calculate the criterion for $F^{\prime}_{1n}>F_{1n}$ as $\displaystyle\frac{3(p_{t}-\frac{1}{2})}{1-p_{t}}>\frac{(F_{1n}-\frac{1}{2})(F_{1n}+F_{3n})}{F_{1n}[1-(F_{1n}+F_{3n})]}.$ (30) and the criterion for $F^{\prime}_{1n}>p_{t}$ as $\displaystyle p_{t}^{2}(4p_{p}-1)(4p_{s}-1)+3p_{t}(1-4p_{p}p_{s})<(1-p_{p})(1-p_{s}).$ (31) Figure 5: The minimum threshold of $p_{t}$ corresponding to $F^{\prime}_{1n}>F_{1n}$, and the maximum threshold of $p_{t}$ corresponding to $F^{\prime}_{1n}>p_{t}$ as a function of $p_{s}$. Here, we control $p_{p}=0.65$, and change $p_{s}$ from 0.61 to 0.71. Similar as the fist step, Eq. (30) and Eq. (31) provide the minimum threshold and maximum threshold of $p_{t}$. The practical value of $p_{t}$ should be between the two thresholds. For example, we suppose that $p_{p}=0.65$, where the suitable value of $p_{s}$ should be in the scale $(0.601,0.715)$. Under this case, we control the $p_{s}$ in the scale of $[0.61,0.71]$ and provide the minimum value and maximal value of $p_{t}$ altered with the value of $p_{s}$ in Fig. 5. ## IV Discussion In the paper, we demonstrate an efficient and simple two-step EPP assisted with hyperentanglement. This EPP requires only one copy of photon pair hyperentangled in polarization, spatial-mode, and time-bin DOFs. We consider a general degradation model that the entanglement in all DOFs suffer from channel noise. By consuming the imperfect entanglement in spatial-mode and time-bin DOFs, we can reduce both the bit-flip error and phase-flip error in polarization DOF and increase the fidelity of the target polarization state. Our two-step EPP has some attractive advantages. First, comparing with previous recurrence EPPs which require two or more copies of low-quality entangled pairs, our EPP uses the spatial and time-bin information to complete the measurement pointer, which avoids consuming an entangled photon copy. In this way, our EPP protocol effectively reduces consumption of entanglement pairs, especially in purification consisting of many rounds. Second, using only one pair of hyperentangled state also reduces the experimental difficulty, for it is hard to generate two pairs of hyperentangled states simultaneously. Third, all the devices in our EPP are available under current experimental condition, so that our EPP is feasible for experiment. Figure 6: The value of $LgR$ as a function of the photon distribution length. We set $\eta_{d}=0.9$ and $\eta_{c}=0.95$ L , and consider the low initial fidelity case ($p_{p}=0.52$, $p_{s}=0.56$, $p_{t}=0.60$) and high initial fidelity case ($p_{p}=0.8$, $p_{s}=0.82$, $p_{t}=0.85$), respectively. It is important to compare the purification efficiency of our two-step EPP with previous recurrence EPPs Pan1 in linear optics in a practical environment. In previous EPPs, for reducing both the bit-flip and phase-flip error in polarization DOF, four pairs of identical low-quality mixed states should be distributed to Alice and Bob (Suppose that the low quality mixed states are the Werner states with the form of Eq. (2)). The transmission efficiency of each photon is $\eta_{t}=e^{-\frac{d}{d_{0}}}$, where $d_{0}$ is the attenuation length of the channel (25 km for commercial fibre L ) and $d$ is the practical photon transmission distance. They first perform the purification operation on each two identical pairs to reduce the bit-flip error. Only when both the purification operations are successful, they perform the H operation on the distilled two photon pairs and further correct the phase-flip error. We suppose that $\eta_{d}$ and $\eta_{c}$ are the detection efficiency of the practical photon detector and the coupling efficiency of a photon to the photon detector, respectively. The total purification efficiency of previous EPP Pan1 can be calculated as $\displaystyle E_{o}=\frac{1}{4}P_{1t}^{2}P_{2t}\eta_{t}^{8}\eta_{d}^{8}\eta_{c}^{8}.$ (32) Here, $P_{1t}$ and $P_{2t}$ represent the success probability of the first and second purification rounds, respectively, which can be calculated as $\displaystyle P_{1t}$ $\displaystyle=$ $\displaystyle\frac{1}{2}[p_{p}^{2}+\frac{2p_{p}(1-p_{p})}{3}+\frac{5(1-p_{p})^{2}}{9}],$ $\displaystyle P_{2t}$ $\displaystyle=$ $\displaystyle\frac{1}{8P_{1t}^{2}}\\{[p_{p}^{2}+\frac{(1-p_{p})^{2}}{9}]^{2}+\frac{8}{81}(1-p_{p})^{4}$ (33) $\displaystyle+$ $\displaystyle\frac{4}{9}[p_{p}^{2}+\frac{(1-p_{p})^{2}}{9}](1-p_{p})^{2}+\frac{4}{9}p_{p}^{2}(1-p_{p})^{2}$ $\displaystyle+$ $\displaystyle\frac{8}{27}p_{p}(1-p_{p})^{3}\\}.$ On the other hand, according to above description, the total purification efficiency of our two-step EPP is $\displaystyle E_{n}=P_{1n}P_{2n}\eta_{t}^{2}\eta_{d}^{2}\eta_{c}^{2}.$ (34) In this way, the ratio of $E_{n}$ and $E_{o}$ can be defined as $\displaystyle R=\frac{E_{n}}{E_{o}}=\frac{4P_{1n}P_{2n}}{P_{1t}^{2}P_{2t}\eta_{t}^{6}\eta_{d}^{6}\eta_{c}^{6}}$ (35) Fig. 6 shows the value of $LgR$ as a function of the photon transmission length $d$. Here, we set $d_{0}=25$ $km$, $\eta_{d}=0.9$ and $\eta_{c}=0.95$ L , and change the distance $d$ from 0 to 100 $km$. Here, we select the low initial fidelity case ($p_{p}=0.52$, $p_{s}=0.56$, $p_{t}=0.60$) and high initial fidelity case ($p_{p}=0.8$, $p_{s}=0.82$, $p_{t}=0.85$), respectively. It can be found that the influences from the initial fidelities in three DOFs on $LgR$ are slight, and $LgR$ increases linearly with the growth of $d$. In this way, our EPP is extremely useful in the long-distance entanglement distribution. Next, we discuss the residual entanglement when the purification steps fail. Here, we consider the case in Sec. II for simplicity. We first consider the case that the first purification step fails, but the second step is successful. When the first step fails, say, the spatial modes $a_{3}b_{4}$ or $a_{4}b_{3}$ each have a photon, the parties can distill a new mixed state with the form of $\displaystyle\rho_{fail1}$ $\displaystyle=$ $\displaystyle F_{fail1}|\Phi^{+}_{p}\rangle\langle\Phi^{+}_{p}|+A_{fail1}|\Phi^{-}_{p}\rangle\langle\Phi^{-}_{p}|$ (36) $\displaystyle+$ $\displaystyle B_{fail1}(|\Psi^{+}_{p}\rangle\langle\Psi^{+}_{p}|+|\Psi^{-}_{p}\rangle\langle\Psi^{-}_{p}|),$ with the probability of $\displaystyle P_{fail1}$ $\displaystyle=$ $\displaystyle p_{p}(1-p_{s})+\frac{(1-p_{p})(1-p_{s})}{3}+\frac{2(1-p_{p})p_{s}}{3}$ (37) $\displaystyle=$ $\displaystyle\frac{1+p_{s}+2p_{p}-4p_{p}p_{s}}{3}.$ The three coefficients can be calculated as $\displaystyle F_{fail1}$ $\displaystyle=$ $\displaystyle\frac{p_{p}(1-p_{s})}{P_{fail1}}=\frac{3p_{p}(1-p_{s})}{1+p_{s}+2p_{p}-4p_{p}p_{s}},$ $\displaystyle A_{fail1}$ $\displaystyle=$ $\displaystyle\frac{\frac{(1-p_{p})(1-p_{s})}{3}}{P_{fail1}}=\frac{(1-p_{p})(1-p_{s})}{1+p_{s}+2p_{p}-4p_{p}p_{s}},$ $\displaystyle B_{fail1}$ $\displaystyle=$ $\displaystyle\frac{\frac{(1-p_{p})p_{s}}{3}}{P_{fail1}}=\frac{(1-p_{p})p_{s}}{1+p_{s}+2p_{p}-4p_{p}p_{s}}.$ (38) In order to make the distilled mixed state have residual entanglement, we require $F_{fail1}>\frac{1}{2}$. This requirement can be satisfied when $p_{s}<\frac{4p_{p}-1}{1+2p_{p}}$. Under this case, when the second purification step is successful, say, the photon detectors $D_{2}D_{7}$, $D_{1}D_{8}$, $D_{3}D_{6}$, or $D_{4}D_{5}$ each registering a single photon, the parties can obtain a new mixed state as $\displaystyle\rho^{\prime}_{fail1}$ $\displaystyle=$ $\displaystyle F^{\prime}_{fail1}|\Phi^{+}_{p}\rangle\langle\Phi^{+}_{p}|+A^{\prime}_{fail1}|\Phi^{-}_{p}\rangle\langle\Phi^{-}_{p}|$ (39) $\displaystyle+$ $\displaystyle B^{\prime}_{fail1}|\Psi^{+}_{p}\rangle\langle\Psi^{+}_{p}|+C^{\prime}_{fail1}|\Psi^{-}_{p}\rangle\langle\Psi^{-}_{p}|,$ with the probability of $\displaystyle P^{\prime}_{fail1}$ $\displaystyle=$ $\displaystyle(F_{fail1}+B_{fail1})p_{t}+(A_{fail1}+B_{fail1})(1-p_{t})$ (40) $\displaystyle=$ $\displaystyle\frac{1-p_{p}+p_{t}(4p_{p}-1)(1-p_{s})}{1+p_{s}-2p_{p}(2p_{s}-1)}.$ The coefficients in Eq. (39) can be written as $\displaystyle F^{\prime}_{fail1}$ $\displaystyle=$ $\displaystyle\frac{F_{fail1}p_{t}}{P^{\prime}_{fail1}}$ $\displaystyle=$ $\displaystyle\frac{3p_{t}p_{p}(1-p_{s})}{1-p_{p}+p_{t}(4p_{p}-1)(1-p_{s})},$ $\displaystyle A^{\prime}_{fail1}$ $\displaystyle=$ $\displaystyle\frac{B_{fail1}p_{t}}{P^{\prime}_{fail1}}$ $\displaystyle=$ $\displaystyle\frac{p_{t}p_{s}(1-p_{p})}{1-p_{p}+p_{t}(4p_{p}-1)(1-p_{s})},$ $\displaystyle B^{\prime}_{fail1}$ $\displaystyle=$ $\displaystyle\frac{A_{fail1}(1-p_{t})}{P^{\prime}_{fail1}}$ $\displaystyle=$ $\displaystyle\frac{(1-p_{t})(1-p_{s})(1-p_{p})}{1-p_{p}+p_{t}(4p_{p}-1)(1-p_{s})},$ $\displaystyle C^{\prime}_{fail1}$ $\displaystyle=$ $\displaystyle\frac{B_{fail1}(1-p_{t})}{P^{\prime}_{fail1}}$ (41) $\displaystyle=$ $\displaystyle\frac{p_{s}(1-p_{t})(1-p_{p})}{1-p_{p}+p_{t}(4p_{p}-1)(1-p_{s})}.$ Figure 7: The value of $F^{\prime}_{fail}$ as a function of $p_{t}$. Here, we control $p_{p}=0.65$, so that maximal value of $p_{s}$ for ensuring $F_{fail1}>\frac{1}{2}$ is 0.70. In this way, we control $p_{s}=$ 0.52, 0.62, 0.68, respectively. Fig. 7 shows the value of $F^{\prime}_{fail1}$ as a function of $p_{t}$ when $p_{p}=0.65$. For ensuring $F_{fail1}>\frac{1}{2}$, we require $p_{s}<0.7$. In this way, we control $p_{s}=$ 0.52, 0.62, 0.68, respectively. It can be found that $F^{\prime}_{fail1}$ increases with the growth of $p_{t}$, but reduces with the growth of $p_{s}$. With suitable value of $p_{t}$, we can obtain $F^{\prime}_{fail1}>p_{p}$ and $F^{\prime}_{fail1}>p_{t}$. Second, we consider the case that the first purification is successful, but the second purification fails. After the first step, the parties share a mixed state as $\rho_{1p}$ in Eq. (9). When the second purification step fails, say, the photon detectors $D_{1}D_{6}$, $D_{2}D_{5}$, $D_{3}D_{8}$, or $D_{4}D_{7}$ each registering a single photon, they can distill a new mixed state with the form of $\displaystyle\rho_{fail2}$ $\displaystyle=$ $\displaystyle F_{fail2}|\Phi^{+}_{p}\rangle\langle\Phi^{+}_{p}|+A_{fail2}|\Phi^{-}_{p}\rangle\langle\Phi^{-}_{p}|$ (42) $\displaystyle+$ $\displaystyle B_{fail2}|\Psi^{+}_{p}\rangle\langle\Psi^{+}_{p}|+C_{fail2}|\Psi^{-}_{p}\rangle\langle\Psi^{-}_{p}|,$ with the probability of $\displaystyle P_{fail2}$ $\displaystyle=$ $\displaystyle(F_{1}+F_{3})(1-p_{t})+(F_{2}+F_{3})p_{t}$ (43) $\displaystyle=$ $\displaystyle\frac{1-(1-p_{t})p_{s}-p_{p}+4p_{p}p_{s}(1-p_{t})}{2-p_{s}-2p_{p}(1-2p_{s})}.$ The coefficients in Eq. (42) can be written as $\displaystyle F_{fail2}$ $\displaystyle=$ $\displaystyle\frac{F_{1}(1-p_{t})}{P_{fail2}}$ $\displaystyle=$ $\displaystyle\frac{3(1-p_{t})p_{p}p_{s}}{1-p_{s}+p_{t}p_{s}-p_{p}+4p_{p}p_{s}(1-p_{t})},$ $\displaystyle A_{fail2}$ $\displaystyle=$ $\displaystyle\frac{F_{3}(1-p_{t})}{P_{fail2}}$ $\displaystyle=$ $\displaystyle\frac{(1-p_{p})(1-p_{s})(1-p_{t})}{1-p_{s}+p_{t}p_{s}-p_{p}+4p_{p}p_{s}(1-p_{t})},$ $\displaystyle B_{fail2}$ $\displaystyle=$ $\displaystyle\frac{F_{2}p_{t}}{P_{fail2}},$ $\displaystyle=$ $\displaystyle\frac{(1-p_{p})p_{s}p_{t}}{1-p_{s}+p_{t}p_{s}-p_{p}+4p_{p}p_{s}(1-p_{t})},$ $\displaystyle C_{fail2}$ $\displaystyle=$ $\displaystyle\frac{F_{3}p_{t}}{P_{fail2}},$ (44) $\displaystyle=$ $\displaystyle\frac{(1-p_{p})(1-p_{s})p_{t}}{1-p_{s}+p_{t}p_{s}-p_{p}+4p_{p}p_{s}(1-p_{t})}.$ Figure 8: The value of $F_{fail2}$ as a function of $p_{t}$. Here, we control $p_{p}=0.65$ and $p_{s}=$ 0.52, 0.62, 0.7, 0.8, respectively, and adjust $p_{t}$ from 0.51 to 1. In Fig. 8, we show the value of $F_{fail2}$ as a function of $p_{t}$ by controlling $p_{p}=0.65$ and $p_{s}=$ 0.52, 0.62, 0.7, 0.8. It can be found that $F_{fail2}$ increases with the growth of $p_{s}$ but reduces with the growth of $p_{t}$. In this way, for obtaining $F_{fail2}>p_{p}$, we require $p_{t}$ to be relatively low. With the growth of $p_{s}$, the maximal values of $p_{t}$ which make $F_{fail2}>p_{p}$ increase. In detail, when $p_{p}=0.65$, and $p_{s}=$ 0.52, 0.62, 0.7, 0.8, the maximal values of $p_{t}$ are 0.519, 0.596, 0.642, and 0.687, respectively. Figure 9: The values of $F_{fail3}$ as a function of $p_{t}$. Here, we control $p_{p}=0.65$ and $p_{s}=$ 0.52, 0.62, 0.68, respectively, and adjust $p_{t}$ from 0.51 to 1. Finally, we will discuss the case that both two purification steps fail. This case corresponds to the photon detector $D_{2}D_{8}$, $D_{1}D_{7}$, $D_{3}D_{5}$, or $D_{4}D_{6}$ each registering one photon. After the first purification step, the parties share a new mixed state with the form of $\rho_{fail1}$ in Eq. (36). Then, when the second purification step fails, they can finally obtain a new mixed state as $\displaystyle\rho_{fail3}$ $\displaystyle=$ $\displaystyle F_{fail3}|\Phi^{+}_{p}\rangle\langle\Phi^{+}_{p}|+A_{fail3}|\Phi^{-}_{p}\rangle\langle\Phi^{-}_{p}|$ (45) $\displaystyle+$ $\displaystyle B_{fail3}|\Psi^{+}_{p}\rangle\langle\Psi^{+}_{p}|+C_{fail3}|\Psi^{-}_{p}\rangle\langle\Psi^{-}_{p}|,$ with the probability of $\displaystyle P_{fail3}$ $\displaystyle=$ $\displaystyle(F_{fail1}+B_{fail1})(1-p_{t})+(A_{fail1}+B_{fail1})p_{t},$ (46) $\displaystyle=$ $\displaystyle\frac{p_{p}(3-4p_{s})-p_{t}(4p_{p}-1)(1-p_{s})+p_{s}}{1-2p_{p}(2p_{s}-1)+p_{s}}.$ The four coefficients can be calculated as $\displaystyle F_{fail3}$ $\displaystyle=$ $\displaystyle\frac{F_{fail1}(1-p_{t})}{P_{fail3}}$ $\displaystyle=$ $\displaystyle\frac{3p_{p}(1-p_{t})(1-p_{s})}{p_{p}(3-4p_{s})-p_{t}(4p_{p}-1)(1-p_{s})+p_{s}},$ $\displaystyle A_{fail3}$ $\displaystyle=$ $\displaystyle\frac{B_{fail1}(1-p_{t})}{P_{fail3}}$ $\displaystyle=$ $\displaystyle\frac{(1-p_{p})(1-p_{t})p_{s}}{p_{p}(3-4p_{s})-p_{t}(4p_{p}-1)(1-p_{s})+p_{s}},$ $\displaystyle B_{fail3}$ $\displaystyle=$ $\displaystyle\frac{A_{fail1}p_{t}}{P_{fail3}}$ $\displaystyle=$ $\displaystyle\frac{(1-p_{p})(1-p_{s})p_{t}}{p_{p}(3-4p_{s})-p_{t}(4p_{p}-1)(1-p_{s})+p_{s}},$ $\displaystyle C_{fail3}$ $\displaystyle=$ $\displaystyle\frac{B_{fail1}p_{t}}{P_{fail3}}$ (47) $\displaystyle=$ $\displaystyle\frac{(1-p_{p})p_{s}p_{t}}{p_{p}(3-4p_{s})-p_{t}(4p_{p}-1)(1-p_{s})+p_{s}}.$ In Fig. 9, we show the value of $F_{fail3}$ as a function of $p_{t}$ with $p_{p}=0.65$. For ensuring the existence of residual entanglement after the first purification step, we require $p_{s}<0.7$. In this way, we control $p_{s}=$ 0.52, 0.62, and 0.68, respectively. It can be found that $F_{fail3}$ reduces with the growth of $p_{s}$ and $p_{t}$. $F_{fail3}$ is lower than $p_{p}$ under arbitrary value of $p_{s}$ and $p_{t}$. However, with relatively low $p_{s}$ and $p_{t}$, there may still exist residual entanglement ($F_{fail3}>\frac{1}{2}$) in $\rho_{fail3}$. For example, for ensuring the existence of residual entanglement in $\rho_{fail3}$, we can calculate the maximal values of $p_{t}$ to be 0.682, 0.599, and 0.523, corresponding to $p_{s}=$ 0.52, 0.62, and 0.68, respectively. From above discussion, there may residual entanglement exist when the purification steps fail. When only one purification step fails, the parties may still distill high-quality entanglement with suitable $p_{s}$ and $p_{t}$ by using of the residual entanglement. Even if the parties can not directly obtain high-quality polarized mixed state, they can still reuse the residual entanglement to distill high-quality entanglement in the next purification round. The existence of residual entanglement provides us a possibility to increase the fidelity of polarization entanglement, thus increase the yield of our EPP. Finally, it is interesting to compare out two-step EPP with the deterministic EPPs (DEPPs), which also adopt the hyperentanglement to realize the purification shenghyper1 ; shenghyper2 . The DEPPs require one pair of hyperentangled state, i.e., polarization-spatial-mode and polarization- spatial-mode-frequency hyperentanglement, respectively. Actually, the DEPPs can completely transform the entanglement in the other DOF to the target DOF, and they do not require the initial target DOF to be entangled. The upper bound of the fidelity in the target DOF is the initial fidelity of the consumed entanglement. On the other hand, our current two-step EPP belongs to recurrence EPP. In our two-step EPP, we suppose that the initial fidelities in three DOFs are larger than $\frac{1}{2}$. After two purification steps, the fidelity of the target polarization state can be higher than $p_{p}$, $p_{s}$, and even $p_{t}$. Actually, in the first purification step, if $p_{p}<\frac{1}{2}$, we can obtain the fidelity $p_{p}<F_{1n}<p_{s}$. Under this case, when the $p_{p}$ and $p_{s}$ satisfy $p_{p}p_{s}>\frac{1}{4}$, we can obtain $F_{1n}>\frac{1}{2}$, say, the distilled new mixed state in polarization DOF has entanglement. In the second purification step, the fidelity of the target polarization DOF can be increased to be higher than $F_{1n}$ and even $p_{t}$. If $p_{p}$ is so low that $p_{p}p_{s}>\frac{1}{4}$ can not be satisfied, our two-step EPP can not work. Based on above comparison, if the initial fidelity in the polarization DOF is relatively high, the current two-step EPP may be more advantageous, and while if that of the polarization DOF is low, the DEPP may be more advantageous. ## V Conclusion In conclusion, we present an efficient two-step recurrence EPP for purifying the entanglement in polarization DOF. In the protocol, we only require one copy of two-photon pair, which is hyperentangled in polarization, spatial- mode, and time-bin DOFs. We suppose that after the photon transmission, the entanglement in all DOFs suffer from the channel noise and degrade to mixed states. As the entanglement in different DOFs have different noisy robustness, the initial mixed states in three DOFs have different fidelities. In the first purification step, the bit-flip error in polarization DOF can be reduced by consuming the imperfect spatial-mode entanglement, while in the second step, the phase-flip error in polarization DOF can be reduced by consuming the imperfect time-bin entanglement. As a result, the fidelity of the target polarization state can be efficiently increased. Our EPP has some attractive advantages. First, comparing with previous two-step recurrence EPPs, which require two or more same copies of nonlocal entangled pairs, our EPP largely reduces the consumption of entanglement pairs. Second, using only one pair of hyperentangled state also reduces the experimental difficulty, for it is hard to generate two pairs of hyperentangled states simultaneously. Third, if we consider the practical photon transmission and detector efficiency, our EPP has much higher purification efficiency. Forth, all the devices in our EPP are available under current experimental condition, so that our EPP is feasible for experiment. Moreover, in traditional two-step recurrent EPP, the parties can distill a high-quality entanglement only when both two steps are successful. If any one step fails, there are no residual entanglement in the distilled mixed state and the distilled photon states have to be discarded. However, when a purification step of our EPP fails, there may exist residual entanglement in the distilled mixed state. The existence of residual entanglement may make the parties distill higher-fidelity polarization entanglement. Even not, the residual entanglement may be reused in the next purification round. In this way, the existence of residual entanglement benefits for further increasing the yield of our EPP. All the above features make our EPP protocol have potential application in future quantum information processing field. ## ACKNOWLEDGEMENTS This work was supported by the National Natural Science Foundation of China (No. 11974189). ## References * (1) C. H. Bennett, G. Brassard, C. Crépeau, R. Jozsa, A. Peres and W. K. Wootters, Teleporting an unknown quantum state via dual classical and Einstein-Podolsky-Rosen channels, Phys. Rev. Lett. 70, 1895 (1993). * (2) D. Bouwmeester, J. W. Pan, K. Mattle, M. Eibl, H. Weinfurter, and A. Zeilinger, Experimental quantum teleportation, Nature 390, 575-579 (1997). * (3) X. M. Hu, C. Zhang, C. J. Zhang, B. H. Liu, Y. F. Huang, Y. J. Han, C. F. Li, and G. C. Guo, Experimental certification for nonclassical teleportation, Quan. Engin. 1, e3 (2019). * (4) L. M. Duan, M. D. Lukin, J. I. Cirac, and P. Zoller, Long-distance quantum communication with atomic ensembles and linear optics, Nature 414, 413 (2001). * (5) C. Simon, H. de Riedmatten, M. Afzelius, N. Sangouard, H. Zbinden, and N. Gisin, Quantum repeaters with photon pair sources and multimode memories, Phys. Rev. Lett. 98, 190503 (2007). * (6) N. Sangouard, C. Simon, H. de Riedmatten, and N. Gisin, Quantum repeaters based on atomic ensembles and linear optics, Rev. Mod. Phys. 83, 33 (2011). * (7) A. K. Ekert, Quantum cryptography based on Bell’s theorem, Phys. Rev. Lett. 67, 661 (1991). * (8) M. Hillery, V. Buz̆ek, and A. Berthiaume, Quantum secret sharing, Phys. Rev. A 59, 1829 (1999). * (9) G. L. Long and X. S. Liu, Theoretically efficient high-capacity quantum-key-distribution scheme, Phys. Rev. A 65, 032302 (2002). * (10) F. G. Deng, G. L. Long, and X. S. Liu, Two-step quantum direct communication protocol using the Einstein- Podolsky-Rosen pair block, Phys. Rev. A 68, 042317 (2003). * (11) W. Zhang, D. S. Ding, Y. B. Sheng, L. Zhou, B. S. Shi, and G. C. Guo, Quantum secure direct communication with quantum memory, Phys. Rev. Lett. 118, 220501 (2017). * (12) L. Zhou, Y. B. Sheng, and G. L. Long, Device-independent quantum secure direct communication against collective attacks, Sci. Bull. 65, 12-20 (2020). * (13) C. H. Bennett, G. Brassard, S. Popescu, et al. Purification of noisy entanglement and faithful teleportation via noisy channels. Phys. Rev. Lett. 76, 722-725 (1996). * (14) D. Deutsch, A. Ekert, R. Jozsa, C. Macchiavello, S. Popescu, and A. Sanpera, Quantum privacy amplification and the security of quantum cryptography over noisy channels, Phys. Rev. Lett. 77, 2818-2821 (1996). * (15) M. Murao, M. B. Plenio, S. Popescu, V. Vedral, and P. L. Knight, Multiparticle entanglement purification protocols, Phys. Rev. A 57, R4075-R4078 (1998). * (16) J. W. Pan, C. Simon, and A. Zellinger, Entanglement purification for quantum communication. Nature (London) 410, 1067 (2001). * (17) C. Simon and J. W. Pan, Polarization entanglement purification using spatial entanglement. Phys. Rev. Lett. 89, 257901 (2002). * (18) J. W. Pan, S. Gasparoni, R. Ursin, G. Weihs, A. Zeilinger, Experimental entanglement purification of arbitrary unknown states, Nature, 423, 417-422 (2003). * (19) W. Dür, H. Aschauer, and H. J. Briegel, Multiparticle entanglement purification for graph states, Phys. Rev. Lett. 91, 107903 (2003). * (20) R. Reichle, D. Leibfried, E. Knill, et al. Experimental purification of two-atom entanglement, Nature 443, 838-841 (2006). * (21) Y. W. Cheong, S. W. Lee, J. Lee, and H. W. Lee, Entanglement purification for high-dimensional multipartite systems, Phys. Rev. A 76, 042314 (2007). * (22) Y. B. Sheng, F. G. Deng, and H. Y. Zhou, Efficient polarization-entanglement purification based on parametric down-conversion sources with cross-kerr nonlinearity, Phys. Rev. A 77, 042308 (2008). * (23) Y. B. Sheng and F. G. Deng, Deterministic entanglement purification and complete nonlocal Bell-state analysis with hyperentanglement, Phys. Rev. A 81, 032307 (2010). * (24) Y. B. Sheng and F. G. Deng, One-step deterministic polarization-entanglement purification using spatial entanglement, Phys. Rev. A 82, 044305 (2010). * (25) Y. B. Sheng, L. Zhou, and G. L. Long, Hybrid entanglement purification for quantum repeaters, Phys. Rev. A 88, 022302 (2013). * (26) L. Zhou and Y. B. Sheng, Purification of logic-qubit entanglement, Sci. Rep. 6, 28813 (2016). * (27) T. J. Wang, S. C. Mi, and C. Wang, Hyperentanglement purification using imperfect spatial entanglement, Opt. Express 25, 2969 (2017). * (28) L. Zhou and Y. B. Sheng, Polarization entanglement purification for concatenated Geenberger-Horne-Zeilinger state, Ann. Phys. 385, 10-35 (2017). * (29) H. Zhang, Q. Liu, X. S. Xu, J. Xiong, A. Alsaedi, T. Hayat, and F. G. Deng, Polarization entanglement purification of nonlocal microwave photons based on the cross-kerr effect in circuit qed, Phys. Rev. A 96, 052330 (2017). * (30) L. K. Chen, H. L. Yong, and P. Xu et al., Experimental nested purification for a linear optical quantum repeater, Nat. Photon. 11, 695-699 (2017). * (31) Kalb, N., et al. Entanglement distillation between solid state quantum network nodes. Science 356, 928-932 (2017). * (32) J. Miguel-Ramiro and W. Dür, Efficient entanglement purification protocols for d-level systems, Phys. Rev. A 98, 042309 (2018). * (33) F. Rozpeedek, T. Schiet, D. Elkouss, A. C. Doherty, and S. Wehner, Optimizing practical entanglement distillation, Phys. Rev. A 97, 062333 (2018). * (34) S. Krastanov, V. V. Albert, and L. Jiang, Optimized entanglement purification, Quantum 3, 123123 (2019). * (35) X. D. Wu, L. Zhou, W. Zhong, Y. B. Sheng, Purification of the concatenated Greenberger-Horne- Zeilinger state with linear optics, Quan. Inform. Process. 17, UNSP255 (2018). * (36) F. F. Du, Y. T. Liu, Z. R. Shi, Y. X. Liang, J. Tang, and J. Liu, Efficient hyperentanglement purification for three-photon systems with the fidelity-robust quantum gates and hyperentanglement link, Opt. Express 27, 27046 (2019). * (37) L. Zhou, S. S. Zhang, W. Zhong, and Y. B. Sheng, Multi-copy nested entanglement purification for quantum repeaters, Ann. Phys. 412, 168042 (2020). * (38) L. Zhou, W. Zhong, and Y. B. Sheng, Purification of the residual entanglement, Opt. Express 28, 2291-2301 (2020). * (39) X. M. Hu, C. X. Huang, Y. B. Sheng, et al. Long-distance entanglement purification for quantum communication, Phys. Rev. Lett. 126, 010503 (2021). * (40) I. Marcikic, H. de Riedmatten, W. Tittel, H. Zbinden, and N. Gisin, Long-distance teleportation of qubits at telecommunication wavelengths, Nature 421, 509 (2003). * (41) R. T. Thew, S. Tanzilli, W. Tittel, H. Zbinden, and N. Gisin, Experimental investigation of the robustness of partially entangled qubits over 11 km, Phys. Rev. A 66, 062304 (2002). * (42) I. Marcikic, H. de Riedmatten, W. Tittel, H. Zbinden, M. Legré and N. Gisin, Distribution of time-bin entangled qubits over 50 km of optical fiber, Phys. Rev. Lett. 93, 180502 (2004). * (43) T. Inagaki, N. Matsuda, O. Tadanaga, and H. Takesue, Entanglement distribution over 300 km of fiber, Opt. Express 21, 23241 (2013). * (44) R. Valivarthi, M. G. Puigibert, Q. Zhou, G. H. Aguilar, V. B. Verma, F. Marsili, M. D. Shaw, S. W. Nam, D. Oblak, and W. Tittel, Quantum teleportation across a metropolitan fibre network, Nat. Photon. 10, 676 (2016). * (45) Q. C. Sun, Y. L. Mao, S. J. Chen, et al. Quantum teleportation with independent sources and prior entanglement distribution over a network, Nat. Photon. 10, 671 (2016). * (46) J. T. Barreiro, N. K.Langford, N. A. Peters, and P.G. Kwiat: Generation of hyperentangled photon pairs. Phys. Rev. Lett. 95, 260501 (2005). * (47) W. K. Wootters, Entanglement of formation of an arbitrary state of two qubits, Phys. Rev. Lett. 80, 2245-2248 (1998). * (48) C. H. Bennett, H. J. Bernstein, S. Popescu, and S. Benjamin, Concentrating partial entanglement by local operations, Phys. Rev. A 53, 2046-2052 (1996). * (49) S. Hill and W. K. Wootters, Entanglement of a pair of quantum bits, Phys. Rev. Lett. 78, 5022-5025 (1997). * (50) W. J. Munro, A. M. Stephens, S. J. Devitt, K. A. Harrison, and K. Nemoto, Quantum communication without the necessity of quantum memories, Nat. Photon. 6, 777 (2012).
Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). Hate Speech and Offensive Content Identification in Indo-European Languages (HASOC) at Forum for Information Retrieval and Evaluation (FIRE), 16th-20th December, 2020, Hyderabad, IN [orcid=0000-0001-8336-195X, ]<EMAIL_ADDRESS> <EMAIL_ADDRESS> # HASOCOne@FIRE-HASOC2020: Using BERT and Multilingual BERT models for Hate Speech Detection Suman Dowlagar [ International Institute of Information Technology - Hyderabad (IIIT-Hyderabad), Gachibowli, Hyderabad, Telangana, India, 500032 Radhika Mamidi [ (2020) ###### Abstract Hateful and Toxic content has become a significant concern in today’s world due to an exponential rise in social media. The increase in hate speech and harmful content motivated researchers to dedicate substantial efforts to the challenging direction of hateful content identification. In this task, we propose an approach to automatically classify hate speech and offensive content. We have used the datasets obtained from FIRE 2019 and 2020 shared tasks. We perform experiments by taking advantage of transfer learning models. We observed that the pre-trained BERT model and the multilingual-BERT model gave the best results. The code is made publically available at https://github.com/suman101112/hasoc-fire-2020 ###### keywords: Hate speech offensive content label classification transfer learning BERT ## 1 Introduction Nowadays, people are frequently using social media platforms to communicate their opinions and share information. Although the communication among users can lead to constructive conversations, the people have been increasingly hit by hateful and offensive content due to these platforms’ anonymity features. It has become a significant issue. The threat of abuse and harassment made many people stop expressing themselves. According to the Cambridge dictionary, Hate speech and offensive content is defined as, * • To harass and cause lasting pain by attacking something uniquely dear to the target. * • To use words that are considered insulting by most people. The main obstacle with hate speech is, it is difficult to classify based on a single sentence because most of the hate speech has context attached to it, and it can morph into many different shapes depending on the context. Another obstacle is that humans cannot always agree on what can be classified as hate speech. Hence it is not very easy to create a universal machine learning algorithm that would detect it. Also, the datasets used to train models tend to "reflect the majority view of the people who collected or labeled the data". To deal with the above scenarios and to encourage research on hate speech and offensive content, the NLP community organized several tasks and workshops such as Task 12: OffensEval 2: Multilingual Offensive content identification in Social Media text 111https://sites.google.com/site/offensevalsharedtask/, OSATC4 shared task on offensive content detection 222http://edinburghnlp.inf.ed.ac.uk/workshops/OSACT4/. Similarly, the FIRE 2020’s HASOC shared task was devoted to the Hate Speech and Offensive Content Identification in Indo-European Languages. This task aims to classify the given annotated tweets. This paper presents the state-of-the-art BERT transfer learning models for automated detection of hate speech and offensive content. The paper is organized as follows. Section 2 provides related work on hate speech and offensive content detection. Section 3 describes the methodology used for this task. Section 4 presents the experimental setup and the performance of the model. Section 5 concludes our work. ## 2 Related Work Machine learning and natural language processing approaches have made a breakthrough in detecting hate speech on web platforms. Many scientific studies have been dedicated to using Machine Learning (ML) [1, 2] and Deep Learning (DL) [3, 4] methods for automated hate speech and offensive content detection. The features used in traditional machine learning approaches are word-level and character-level n-grams, etc. Although supervised machine learning-based approaches have used different text mining-based features such as surface features, sentiment analysis, lexical resources, linguistic features, knowledge-based features, or user-based and platform-based metadata, they necessitate a well-defined feature extraction approach. Nowadays, the neural network models apply text representation and deep learning approaches such as Convolutional Neural Networks (CNNs) [5], Bi-directional Long Short- Term Memory Networks (LSTMs) [6], and BERT [7] to improve the performance of hate speech and offensive content detection models. ## 3 Methodology Figure 1: BERT model for sequence classification on Hate Speech Data. Here, we use the pre-trained BERT transformer model for hate speech and offensive content detection. Figure 1 depicts the abstract view of BERT model that is used for hate speech detection and offensive language identification. Bidirectional Encoder Representations from Transformers (BERT) is a transformer Encoder stack trained on the large English corpus. It has 2 models, $BERT_{base}$ and $BERT_{large}$. These model sizes have a large number of transformer layers. The $BERT_{base}$ version has 12 transformer layers and the $BERT_{large}$ has 24. These also have larger feed-forward networks with 768 and 1024 hidden representations, and attention heads are 12 and 16 for the respective models. Like the vanilla transformer model [8], BERT takes a sequence of words as input. Each layer applies self-attention, passes its results through a feed-forward network, and then hands it off to the next encoder. Embeddings from $BERT_{base}$ have 768 hidden units. The BERT configuration model takes a sequence of words/tokens at a maximum length of 512 and produces an encoded representation of dimensionality 768. The pre-trained BERT models have a better word representation as they are trained on a large Wikipedia and book corpus. As the pre-trained BERT model is trained on generic corpora, we need to fine-tune the model for the downstream tasks. During fine-tuning, the pre-trained BERT model parameters are updated when trained on the labeled hate speech and offensive content dataset. When fine-tuned on the downstream sentence classification task, a very few changes are applied to the $BERT_{base}$ configuration. In this architecture, only the [CLS] (classification) token output provided by BERT is used. The [CLS] output is the output of the 12th transformer encoder with a dimensionality of 768. It is given as input to a fully connected neural network, and the softmax activation function is applied to the neural network to classify the given sentence. Thus, BERT learns to predict whether a tweet can be classified as a hate speech or offensive content. Apart from $BERT_{base}$ model, we used the pre-trained multilingual $BERT_{base}$ model, as our data consisted of German and Hindi multilingual languages. The multilingual BERT and vanilla BERT models’ architecture is the same, but the pre-trained multilingual BERT model is trained on multilingual Wikipedia language sources. ## 4 Experiment Initially, we introduce datasets used, the task description, and then review the BERT model’s performance on hate speech and offensive content detection. We also include our implementation details and error analysis in the subsequent sections. ### 4.1 Dataset Table 1: Data Statistics Language | Train Sentences | Test Sentences ---|---|--- English (HASOC 2019) | 5852 | 1153 German (HASOC 2019) | 3819 | 850 Hindi (HASOC 2019) | 4665 | 1318 English (HASOC 2020) | 3708 | 814 German (HASOC 2020) | 2373 | 526 Hindi (HASOC 2020) | 2963 | 663 We used the dataset provided by the organizers of HASOC FIRE-2020 [hasoc2020overview] and FIRE-2019 [9]. The HASOC dataset was subsequently sampled from Twitter and partially from Facebook for English, German, and Hindi languages. The tweets were acquired using hashtags and keywords that contained offensive content. The statistics of FIRE 2020 and 2019 datasets are given in the Table 1. ### 4.2 Task description The following tasks are in HASOC 2020. Sub-task A focuses on coarse-grained Hate speech detection in all three languages. The task is to classify tweets into two classes: * • (NOT) Non Hate-Offensive - Post does not contain any Hate speech, profane, offensive content. * • (HOF) Hate and Offensive - Post contains Hate, offensive, and profane content. Sub-task B represents a fine-grained classification. Hate-speech and offensive posts from the sub-task A are further classified into three categories. The task is to classify the tweets into three classes: * • (HATE) Hate speech - Post contains Hate speech content. * • (OFFN) Offenive - Post contains offensive content such as insulting, degrading, dehumanizing and threatening. * • (PRFN) Profane - Post contains profane words. This typically concerns the usage of swearwords and cursing. ### 4.3 Implementation For the implementation, we used the transformers library provided by HuggingFace [10]. The HuggingFace transformers package is a python library providing pre-trained and configurable transformer models useful for a variety of NLP tasks. It contains the pre-trained BERT and multilingual BERT, and other models suitable for downstream tasks. As the implementation environment, we use the PyTorch library that supports GPU processing. The BERT models were run on NVIDIA RTX 2070 graphics card with an 8 GB graphics card. We trained our classifier with a batch size of 64 for 5 to 10 epochs based on our experiments. The dropout is set to 0.1, and the Adam optimizer is used with a learning rate of 2e-5. We used the hugging face transformers pre-trained BERT tokenizer for tokenization. We used the BertForSequenceClassification module provided by the HuggingFace library during finetuning and sequence classification. ### 4.4 Baseline models Here, we compared the BERT model with other machine learning algorithms. #### 4.4.1 SVM with TF_IDF text representation We chose Support Vector Machines (SVM) for hate speech and offensive content detection. The tokenizer used is SentencePiece [11]. SentencePiece is a commonly used technique to segment words into a subword-level. In both cases, the vocabulary is initialized with all the individual characters in the language, and then the most frequent or likely combinations of the symbols are iteratively added to the vocabulary. #### 4.4.2 ELMO embeddings with SVM model ELMO(Embeddings from Language Models) [12] deals with contextual embeddings. Contextual word-embeddings are born to capture the word meaning in its context. Instead of using a fixed embedding for each word, ELMO looks at the word’s context, i.e., the word’s entire sentence, before assigning embedding to the word. It uses a bi-LSTM trained on a specific task to be able to create those embeddings. We used the ELMO model present on tensorflow hub (https://tfhub.dev/google/elmo/2) to obtain the ELMO embeddings on the hate speech data for all the languages. After obtaining the embeddings, we take the mean of embeddings and apply an SVM classifier to classify the given sentence into hate speech or offensive content. We used the SentencePiece tokenizer. ## 5 Results Table 2: macro F1 and Accuracy on English Subtasks A and B | Hate speech Detection | Offensive Content Identification ---|---|--- Model | macro F1 | Accuracy | macro F1 | Accuracy SVM | 81.56% | 81.57% | 47.49% | 76.78% ELMO + SVM | 82.43% | 83.78% | 49.62% | 79.54% BERT | 88.33% | 88.33% | 54.44% | 81.57% Table 3: macro F1 and Accuracy on German Subtasks A and B | Hate speech Detection | Offensive Content Identification ---|---|--- Model | macro F1 | Accuracy | macro F1 | Accuracy SVM | 73.29% | 79.27% | 45.54% | 77.94% ELMO + SVM | 71.73% | 80.42% | 45.94% | 78.21% multilingual-BERT | 77.91% | 82.51% | 47.78% | 80.42% Table 4: macro F1 and Accuracy on Hindi Subtasks A and B | Hate speech Detection | Offensive Content Identification ---|---|--- Model | macro F1 | Accuracy | macro F1 | Accuracy SVM | 59.73% | 70.13% | 36.78% | 72.39% ELMO + SVM | 60.91% | 71.47% | 39.89% | 72.76% multilingual-BERT | 63.54% | 74.96% | 49.71% | 73.15% The results are tabulated in Tables 2, 3 and 4. We evaluated the performance of the method using macro F1 and accuracy. The BERT model performed well when compared to the other SVM with TF-IDF and ELMO text representations. Given all the languages and both the subtasks A and B, we have observed an increase of 1-2% in classification metrics for ELMO embeddings + SVM classifier compared to the baseline SVM classifier. However, BERT showed an increase of 5-7% in classification metrics compared to ELMO and SVM models. It shows the pre- trained BERT model’s capability, which learnt better text representations from the generic data. The state of the art transformer architecture used in the BERT model helped the model learn better parameter weights in hate speech and offensive content detection. Figure 2: Confusion matrix on the given test data for the English, German and Hindi languages given subtask A: Hate Speech Detection and subtask B: Offensive Content Identification ## 6 Error Analysis The confusion matrix of BERT model for subtasks A and B for the english, german and hindi datasets is given in the Figure 2. For the binary classification, the best-performed model was for English subtask A. The binary classification for the Hindi model is not helpful. The model misclassified most of the hate-speech labels. It can be seen in subfigure 2. For offensive content evaluation, the model performed better on English subtask B. It correctly classified "NONE (not offensive)" and "PROF (profane)" but was unable to classify "HATE (hate speech)" and "OFFN (offensive)" and misunderstood most of them as "PROF". The multilingual-BERT model misclassified most of the hate speech and offensive content labels for the German and Hindi languages as "NONE" and didn’t perform well on those datasets. ## 7 Conclusion and Future work We used pre-trained bi-directional encoder representations using transformers (BERT) and multilingual-BERT for hate speech and offensive content detection for English, German, and Hindi languages. We compared the BERT with other machine learning and neural network classification methods. Our analysis showed that using the pre-trained BERT and multilingual BERT models and finetuning it for downstream hate-speech text classification tasks showed an increase in macro F1 score and accuracy metrics compared to traditional word- based machine learning approaches. The given data has both hate speech and offensive content labeled for a given same sentence. It implies that both tasks are related. In such a scenario, we can use joint learning models to help obtain a strong relationship between the two tasks. Which, in turn, helps a deep joint classification model to understand the given datasets better. ## References * Davidson et al. [2017] T. Davidson, D. Warmsley, M. Macy, I. Weber, Automated hate speech detection and the problem of offensive language, arXiv preprint arXiv:1703.04009 (2017). * Gaydhani et al. [2018] A. Gaydhani, V. Doma, S. Kendre, L. Bhagwat, Detecting hate speech and offensive language on twitter using machine learning: An n-gram and tfidf based approach, arXiv preprint arXiv:1809.08651 (2018). * Gambäck and Sikdar [2017] B. Gambäck, U. K. Sikdar, Using convolutional neural networks to classify hate-speech, in: Proceedings of the first workshop on abusive language online, 2017, pp. 85–90. * Badjatiya et al. [2017] P. Badjatiya, S. Gupta, M. Gupta, V. Varma, Deep learning for hate speech detection in tweets, in: Proceedings of the 26th International Conference on World Wide Web Companion, 2017, pp. 759–760. * Kim [2014] Y. Kim, Convolutional neural networks for sentence classification, arXiv preprint arXiv:1408.5882 (2014). * Hochreiter and Schmidhuber [1997] S. Hochreiter, J. Schmidhuber, Long short-term memory, Neural computation 9 (1997) 1735–1780. * Devlin et al. [2018] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, Bert: Pre-training of deep bidirectional transformers for language understanding, arXiv preprint arXiv:1810.04805 (2018). * Vaswani et al. [2017] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, I. Polosukhin, Attention is all you need, in: Advances in neural information processing systems, 2017, pp. 5998–6008. * Mandl et al. [2019] T. Mandl, S. Modha, P. Majumder, D. Patel, M. Dave, C. Mandlia, A. Patel, Overview of the hasoc track at fire 2019: Hate speech and offensive content identification in indo-european languages, in: Proceedings of the 11th Forum for Information Retrieval Evaluation, 2019, pp. 14–17. * Wolf et al. [2019] T. Wolf, L. Debut, V. Sanh, J. Chaumond, C. Delangue, A. Moi, P. Cistac, T. Rault, R. Louf, M. Funtowicz, et al., Huggingface’s transformers: State-of-the-art natural language processing, ArXiv (2019) arXiv–1910. * Kudo and Richardson [2018] T. Kudo, J. Richardson, Sentencepiece: A simple and language independent subword tokenizer and detokenizer for neural text processing, arXiv preprint arXiv:1808.06226 (2018). * Peters et al. [2018] M. E. Peters, M. Neumann, M. Iyyer, M. Gardner, C. Clark, K. Lee, L. Zettlemoyer, Deep contextualized word representations, arXiv preprint arXiv:1802.05365 (2018).
# Strong gravitational lensing by Kerr and Kerr-Newman black holes Tien Hsieh Da-Shin Lee<EMAIL_ADDRESS>Chi-Yong Lin <EMAIL_ADDRESS>Department of Physics, National Dong Hwa University, Hualien 97401, Taiwan, Republic of China ###### Abstract We study the strong gravitational lensing due to the Kerr black holes with angular momentum $a$ and the Kerr-Newman black holes with additional charge $Q$. We first derive the analytical expressions of the deflection angles of light rays that particularly diverge as they travel near the photon sphere. In this strong deflection limit, the light rays can circle around the black hole multiple times before reaching the observer, giving relativistic images. The obtained analytical expressions are then applied to compute the angular positions of relativistic images due to the supermassive galactic black holes. In this work, we focus on the outermost image with reference to the optical axis. We find that its angular separation from the one closest to the optical axis increases with the increase of angular momentum $a$ of the black holes for light rays in direct orbits. Additionally, the effects of the charge $Q$ of black holes also increase the angular separation of the outermost image from the others for both direct and retrograde orbits. The potentially increasing observability of the relativistic images from the effects of angular momentum and charge of the black holes will be discussed. ###### pacs: 04.70.-s, 04.70.Bw, 04.80.Cc ## I Introduction Gravitational lensing is one of the powerful tools to test general relativity (GR) MIS ; HAR . Weak lensing has been fully studied in the formalism of weak field approximations, which can be used to successfully explain various lensing phenomena in a broad array of astrophysical contexts SEF1992 . Nevertheless, in recent years, there have been significant theoretical studies looking into lensing phenomena from strong field perspectives Vir ; Fri ; Bozza1 ; Bozza2 ; Bozza_2003 ; Bozza3 ; Eiroa ; Iyer1 ; Tsuka1 ; Tsuka2 ; Vir3 ; Sha . Through the gravitational lensing in the vicinity of the compact massive objects such as a black hole would provide another avenue to test GR. So far, observational evidence has shown that almost every large galaxy has a supermassive black hole at the galaxy’s center Ric . The Milky Way has a supermassive black hole in its Galactic Center with the location of Sagittarius A* Ghez ; Sch . Together with the first image of the black hole captured by the Event Horizon Telescope EHT1 ; EHT2 ; EHT3 , gravitational lensing will also become an important probe to study the isolated dim black hole. Recently, Virbhadra and Ellis have developed a new gravitational lens equation, which allows us to study large deflection of light rays, resulting in the strong gravitational lensing Vir . This lens equation is then applied to analyze the lensing by a Schwarzschild black hole in the center of the galaxy using numerical methods. Later, Frittelli et al. propose the definition of an exact lens equation without reference to the background spacetime, and construct the exact lens equation explicitly in the Schwarzschild spacetime Fri . Strong field lensing in the general spherically symmetric and static spacetime is first studied analytically by Bozza in Bozza1 ; Bozza2 ; Bozza3 and later by Tsukamoto in Tsuka1 ; Tsuka2 . These works show that the deflection angle $\hat{\alpha}(b)$ of light rays for a given impact parameter $b$, which in the strong deflection limit (SDL) as $b\to b_{c}$, can be approximated in the form $\hat{\alpha}(b)\approx-\bar{a}\log{\left(\frac{b}{b_{c}}-1\right)}+\bar{b}+O((b-b_{c})\log(b-b_{c}))$ (1) with two parameters $\bar{a}$ and $\bar{b}$ as a function of the black hole’s parameters. Then, in Bozza_2003 , the Kerr black hole of the nonspherically symmetric black holes is considered, exploring $\bar{a}$ and $\bar{b}$ numerically. In this paper, we extend the works of Bozza1 and Tsuka1 ; Tsuka2 and find the analytic form of $\bar{a}$ and $\bar{b}$ for nonspherically symmetric Kerr and Kerr-Newman black holes, respectively, using the analytical closed-form expressions of the deflection angles in Iyer2 and Hsiao . Although one might not expect that astrophysical black holes have large residue electric charge, some accretion scenarios are proposed to investigate the possibility of the spinning charged back holes Wilson_1975 ; Dam . It is then still of great interest to extend the studies to the Kerr-Newman black holes Liu ; Jiang_2018 ; Kraniotis_2014 . The analytical expressions can be applied to examine the lensing effects due to the supermassive galactic black holes as illustrated in Fig.(1). The light rays are emitted from the source, and circle around the black hole multiple times in the SDL along a direct orbit (red line) or a retrograde orbit (blue line), giving two sets of the relativistic images. Following the approach of Bozza2 enables us to study the observational consequences. Figure 1: Gravitational lens about relativistic images. Considering the Kerr or the Kerr-Newman black hole with angular momentum of the clockwise rotation, the light rays are emitted from the source, and circle around the black hole multiple times in the SDL along a direct orbit (red line) or a retrograde orbit (blue line). The graph illustrates two sets of the relativistic images. The layout of the paper is as follows. In Sec.II, we first review the closed- form expression of the deflection angle due to the Kerr and/or the Kerr-Newman black holes. In particular, we discuss the results of the radius of the innermost circular motion of light rays as well as the associated critical impact parameters as a function of the black hole’s parameters. These will serve as the important inputs to find the values of the coefficients $\bar{a}$ and $\bar{b}$ in the SDL deflection angle. Then we derive the analytic form of $\bar{a}$ and $\bar{b}$ in the cases of Kerr and Kerr-Newman black holes, respectively, and check the consistency with the known results from taking the proper limits of the black holes’s parameters. In Sec. III, the analytical expressions on the equatorial gravitational lensing are then applied to compute the angular positions of relativistic images due to the supermassive galactic black holes. When the light rays travel on the quasiequatorial plane, the obtained results for $\theta=\frac{\pi}{2}$ can also be used to estimate the magnification of relativistic images, as the light sources are near one of the caustic points with the additional inputs from the dynamics of the light rays in the angle $\theta$. The potentially increasing observability of the relativistic images from the effects of angular momentum and charge of the black holes will be summarized in the closing section. ## II Deflection angle due to black holes in the strong deflection limit We consider nonspherically symmetric spacetimes of the Kerr and Kerr-Newman metrics respectively to obtain the deflection angle $\hat{\alpha}(b)$ of light rays for a given impact parameter $b$. In the SDL, as $b\to b_{c}$, $\hat{\alpha}(b)$ can be approximated in the form of (1). In what follows, we will consider the above two types of the black holes separately. ### II.1 Kerr black holes The line element of the Kerr black hole in which spacetime outside a black hole with the gravitational mass $M$ and angular momentum per unit mass $a=J/M$ is described by $\displaystyle{ds}^{2}$ $\displaystyle=$ $\displaystyle g_{\mu\nu}dx^{\mu}dx^{\nu}$ (2) $\displaystyle=$ $\displaystyle-\frac{\left(\Delta-a^{2}\sin^{2}\theta\right)}{\Sigma}{dt}^{2}-\frac{a\sin^{2}\theta\left(2Mr\right)}{\Sigma}({dt}{d\phi+d\phi dt)}$ $\displaystyle+\frac{\Sigma}{\Delta}dr^{2}+\Sigma{\,d\theta}^{2}+\frac{\sin^{2}\theta}{\Sigma}\left((r^{2}+a^{2})^{2}-a^{2}\Delta\sin^{2}\theta\right){d\phi}^{2}\,$ with $\Sigma=r^{2}+a^{2}\cos^{2}\theta\,,\quad\Delta=r^{2}+a^{2}-2Mr\,.$ (3) The outer (inner) event horizon $r_{+}$ ($r_{-}$) can be found by solving $\Delta(r)=0$, and is given by $r_{\pm}=M\pm\sqrt{M^{2}-a^{2}}\,$ (4) with the condition $M^{2}>a^{2}$. Notice that we just adopt the notation of $r_{-}$ where the light rays traveling outside the horizon are considered. The Lagrangian of a particle is then $\displaystyle\mathcal{L}=\frac{1}{2}g_{\mu\nu}u^{\mu}u^{\nu}\,$ (5) with the 4-velocity $u^{\mu}=dx^{\mu}/d\lambda$ defined in terms of an affine parameter $\lambda$. The metric of the Kerr black hole, which is independent on $t$ and $\phi$, gives the associated Killing vectors $\xi_{(t)}^{\mu}$ and $\xi_{(\phi)}^{\mu}$ $\displaystyle\xi_{(t)}^{\mu}=\delta_{t}^{\mu}\;,\quad\xi_{\phi}^{\mu}=\delta_{\phi}^{\mu}\,.$ (6) Then, together with 4-velocity of light rays, the conserved quantities along a geodesic, can be constructed by the above Killing vectors $\varepsilon\equiv-\xi_{(t)}^{\mu}u_{\mu}$ and $\ell\equiv\xi_{\phi}^{\mu}u_{\mu}$, where $\varepsilon$ and $\ell$ are the light ray’s energy and azimuthal angular momentum at spatial infinity. Light rays traveling along null world lines obey the condition $u^{\mu}u_{\mu}=0$. To indicate whether the light rays are traversing along the direction of frame dragging or opposite to it, we define the following impact parameter : $\displaystyle b_{s}=s\left|\frac{\ell}{\varepsilon}\right|\equiv s\,b\;,$ (7) where $s=\text{Sign$(\ell/\varepsilon)$}$ and $b$ is the positive magnitude. The parameter $s=+1$ for $b_{s}>0$ will be referred to as direct orbits, and those with $s=-1$ for $b_{s}<0$ as retrograde orbits (see Fig.(1) for the sign convention). Here we restrict the light rays traveling on the equatorial plane of the black hole by choosing $\theta={\pi}/{2}$, and $\dot{\theta}=0$. The equation of motion along the radial direction can be cast in the form Hsiao $\displaystyle\frac{1}{b^{2}}$ $\displaystyle=\frac{\dot{r}^{2}}{\ell^{2}}+W_{\text{eff}}(r)\;,$ (8) from which we define the function $W_{\text{eff}}$ as $\displaystyle W_{\text{eff}}(r)=\frac{1}{r^{2}}\left[1-\frac{a^{2}}{b^{2}}-\frac{2M}{r}\left(1-\frac{a}{b_{s}}\right)^{2}\right]\,.$ (9) The above equation is analogous to that of particle motion in the effective potential $W_{\text{eff}}(r)$ with the kinetic energy ${\dot{r}}^{2}/\ell^{2}$ and constant total energy $1/b^{2}$. Let us consider that a light ray starts in the asymptotic region to approach the black hole, and then turns back to the asymptotic region to reach the observer. Such light rays have a turning point, the closest approach distance to a black hole $r_{0}$, which crucially depends on the impact parameter $b$, determined by $\displaystyle\left.\frac{\dot{r}^{2}}{\ell^{2}}\right|_{r=r_{0}}=\frac{1}{b^{2}}-W_{\text{eff}}(r_{0})=0\,.$ (10) From (10), also shown in Hsiao ; Iyer2 , one can find the impact parameter $b$ for a given $r_{0}$, which becomes the important input for the analytical expressions of the deflection angle in the SDL, as $b(r_{0})=\frac{2sMa-r_{0}\sqrt{a^{2}-2r_{0}M+r_{0}^{2}}}{2M-r_{0}}\,.$ (11) The behavior of the light ray trajectories depends on whether $1/b^{2}$ is greater or less than the maximum height of $W_{\text{eff}}(r)$. The innermost trajectories of light rays have a direct consequence on the apparent shape of the black hole. The smallest radius ${r_{sc}}$, when the turning point $r_{0}$ is located at the maximum of $W_{\text{eff}}(r)$, with the critical impact parameter ${b_{sc}}$, obeys $\displaystyle\left.\frac{d\,W_{\text{eff}}(r)}{dr}\right|_{r=r_{sc}}$ $\displaystyle=0\,.$ (12) Then the radius of the circular motion forming the photon sphere is given by (See Hsiao ; Iyer2 ). $\displaystyle r_{sc}=2M\bigg{\\{}1+\cos\bigg{[}\frac{2}{3}\cos^{-1}\bigg{(}\frac{-sa}{M}\bigg{)}\bigg{]}\bigg{\\}}\;$ (13) with the corresponding impact parameter $\displaystyle b_{sc}$ $\displaystyle=-a+s6M\cos\bigg{[}\frac{1}{3}\cos^{-1}\bigg{(}\frac{-sa}{M}\bigg{)}\bigg{]}\,.$ (14) In the case of a Kerr black hole, the nonzero spin of the black hole is found to give more repulsive effects to the light rays in the direct orbits than those in the retrograde orbits due to the $1/r^{3}$ term in the effective potential. The repulsive effects in turn affect light rays in the direct orbits in a way to prevent them from collapsing into the event horizon. As a result, this shifts the innermost circular trajectories of the light rays toward the black hole with the smaller critical impact parameter $b_{+c}$ than $b_{-c}$ in the retrograde orbits as shown in Fig.(2). As such, when $a$ increases, the impact parameter $b_{+c}$ decreases whereas $|b_{-c}|$ increases instead Iyer2 ; Hsiao . It will be shown in the next section that the value of $b_{sc}$ is a key quantity to determine the features of the angular position of the induced images of the distant light sources due to the strong gravitational lensing effects. Also, the presence of black hole’s spin is to give the smaller deflection angle in the direct orbits as compared with the retrograde orbits with the same impact parameter $b$ Hsiao ; Iyer2 . We proceed by introducing the variable $z\equiv 1-\frac{r_{0}}{r}\,.$ (15) The geodesic equations for $r$ and $\phi$ found in Hsiao can be rewritten in terms of $z$ as Tsuka1 $\frac{dz}{d\phi}=\frac{1}{r_{0}}\frac{1-\frac{2{M}}{r_{0}}(1-z)+\frac{a^{2}}{r_{0}^{2}}(1-z)^{2}}{1-\frac{2{M}}{r_{0}}(1-z)(1-\frac{a}{b_{s}})}\sqrt{B(z,r_{0})}\;,$ (16) where the function $B(z,r_{0})$ has the trinomial form in $z$ $B(z,r_{0})=c_{1}(r_{0})z+c_{2}(r_{0})z^{2}+c_{3}(r_{0})z^{3}$ (17) with the coefficients $\begin{split}c_{1}(r_{0})=&-6Mr_{0}\left(1-\frac{a}{b_{s}}\right)^{2}+2r_{0}^{2}\left(1-\frac{a^{2}}{b^{2}}\right)\,,\\\ c_{2}(r_{0})=&6Mr_{0}\left(1-\frac{a}{b_{s}}\right)^{2}-r_{0}^{2}\left(1-\frac{a^{2}}{b^{2}}\right)\,,\\\ c_{3}(r_{0})=&-2Mr_{0}\left(1-\frac{a}{b_{s}}\right)^{2}\,.\end{split}$ (18) Next we rewrite $\frac{1-\frac{2{M}}{r_{0}}(1-\frac{a}{b_{s}})+\frac{2{M}}{r_{0}}(1-\frac{a}{b_{s}})z}{1-\frac{2M}{r_{0}}+\frac{a^{2}}{r_{0}^{2}}+(\frac{2{M}}{r_{0}}-\frac{2a^{2}}{r_{0}^{2}})z+\frac{a^{2}}{r_{0}^{2}}z^{2}}=\frac{r_{0}^{2}}{a^{2}}\left(\frac{C_{-}}{z-z_{-}}+\frac{C_{+}}{z-z_{+}}\right)\;,$ (19) where the roots $z_{-}$, $z_{+}$, and the coefficients $C_{-}$, $C_{+}$ are $\begin{split}z_{-}=&1-\frac{r_{0}r_{-}}{a^{2}}\,,\\\ z_{+}=&1-\frac{r_{0}r_{+}}{a^{2}}\,,\end{split}$ (20) $\begin{split}C_{-}=&\frac{a^{2}-2Mr_{-}(1-\frac{a}{b_{s}})}{2r_{0}\sqrt{M^{2}-a^{2}}}\,,\\\ C_{+}=&\frac{-a^{2}+2Mr_{+}(1-\frac{a}{b_{s}})}{2r_{0}\sqrt{M^{2}-a^{2}}}\end{split}$ (21) with $r_{+}$ ($r_{-}$) being the outer (inner) horizon of a Kerr black hole defined in (4). Also note that $z_{-}$, $z_{+}\leq 0$, for all spin $a$. Then the deflection angle can be calculated as a function of the closest approach distance $r_{0}$ from (16) giving $\hat{\alpha}(r_{0})=I(r_{0})-\pi\,,\quad I(r_{0})=\int_{0}^{1}f(z,r_{0})dz\,,$ (22) where the integrand becomes $f(z,r_{0})=\frac{r_{0}^{2}}{a^{2}}\left(\frac{C_{-}}{z-z_{-}}+\frac{C_{+}}{z-z_{+}}\right)\frac{2{r_{0}}}{\sqrt{c_{1}(r_{0})z+c_{2}(r_{0})z^{2}+c_{3}(r_{0})z^{3}}}\,.$ (23) In the SDL of our interest, when the closest approach distance reaches its critical limit, namely $r_{0}\to r_{sc}$, and $c_{1}(r_{0})\to 0$ in (18) obtained from (12), the integrand $f(z,r_{0})\rightarrow\frac{1}{z}$ for small $z$ leads to the logarithmic divergence as $r_{0}\to r_{sc}$. Let us now define a new function $f_{D}(z,r_{0})$ $f_{D}(z,r_{0})=\frac{r_{0}^{2}}{a^{2}}\left(\frac{C_{-}}{z-z_{-}}+\frac{C_{+}}{z-z_{+}}\right)\frac{2{r_{0}}}{\sqrt{c_{1}(r_{0})z+c_{2}(r_{0})z^{2}}}\;,$ (24) that separates the divergent part from the regular part given by $f_{R}(z,r_{0})=f(z,r_{0})-f_{D}(z,r_{0})$. The integral of $f_{R}$ is thus finite. The divergent part comes from an integral of the function $f_{D}(z,r_{0})$, which contributes not only to $\bar{a}$ for the logarithmic term but also $\bar{b}$ for the regular part in (1), giving $\begin{split}I_{D}(r_{0})=&\int_{0}^{1}f_{D}(z,r_{0})dz\\\ =&\frac{2r_{0}^{3}}{a^{2}}\frac{C_{-}}{\sqrt{c_{1}(r_{0})z_{-}+c_{2}(r_{0})z_{-}^{2}}}\log{\left(\frac{\sqrt{c_{1}(r_{0})z_{-}+c_{2}(r_{0})z_{-}}+\sqrt{c_{1}(r_{0})+c_{2}(r_{0})z_{-}}}{\sqrt{c_{1}(r_{0})z_{-}+c_{2}(r_{0})z_{-}}-\sqrt{c_{1}(r_{0})+c_{2}(r_{0})z_{-}}}\right)}\\\ +&\frac{2r_{0}^{3}}{a^{2}}\frac{C_{+}}{\sqrt{c_{1}(r_{0})z_{+}+c_{2}(r_{0})z_{+}^{2}}}\log{\left(\frac{\sqrt{c_{1}(r_{0})z_{+}+c_{2}(r_{0})z_{+}}+\sqrt{c_{1}(r_{0})+c_{2}(r_{0})z_{+}}}{\sqrt{c_{1}(r_{0})z_{+}+c_{2}(r_{0})z_{+}}-\sqrt{c_{1}(r_{0})+c_{2}(r_{0})z_{+}}}\right)}\,.\end{split}$ (25) In the SDL, the expansions of the coefficient $c_{1}(r_{0})$ (18) and the impact parameter $b(r_{0})$ in powers of small $r_{0}-r_{sc}$ read $c_{1}(r_{0})=c_{1sc}^{\prime}(r_{0}-r_{sc})+{O}(r_{0}-r_{sc})^{2}\,,$ (26) $b(r_{0})=b_{sc}+\frac{b_{sc}^{\prime\prime}}{2!}(r_{0}-r_{sc})^{2}+{O}(r_{0}-r_{sc})^{3}\,,$ (27) where $c_{1}(r_{sc})\equiv c_{1sc}=0$ and $b(r_{sc})\equiv b_{sc}$ is the critical impact parameter given by (14). The subscript $sc$ denotes evaluating the function at $r=r_{sc}$. The prime means the derivative with respect to $r_{0}$. Notice that using $c_{1sc}=0$ in (18), one finds $c_{3sc}=-\frac{2}{3}c_{2sc}\;.$ (28) Combining (26) with (27), we can write $c_{1}(r_{0})$ in terms of small $b-b_{sc}$ as $\lim_{r_{0}\to r_{sc}}c_{1}(r_{0})=\lim_{b\to b_{sc}}c_{1sc}^{\prime}\sqrt{\frac{2b_{sc}}{b_{sc}^{\prime\prime}}}\left(\frac{b}{b_{sc}}-1\right)^{1/2}\;.$ (29) In the SDL, substituting (29) into (25), $I_{D}$ becomes $\begin{split}I_{D}(b)\simeq&-\left(\frac{r_{sc}^{3}}{a^{2}}\frac{C_{-sc}}{\sqrt{c_{2sc}\,z_{-sc}^{2}}}+\frac{r_{sc}^{3}}{a^{2}}\frac{C_{+sc}}{\sqrt{c_{2sc}\,z_{+sc}^{2}}}\right)\log{\left(\frac{b}{b_{sc}}-1\right)}\\\ &+\frac{r_{sc}^{3}}{a^{2}}\frac{C_{-sc}}{\sqrt{c_{2sc}\,z_{-sc}^{2}}}\log{\left(\frac{16\,c^{2}_{2sc}\,z_{-sc}^{2}b_{sc}^{\prime\prime}}{c_{1sc}^{\prime 2}2b_{sc}(z_{-sc}-1)^{2}}\right)}+\frac{r_{sc}^{3}}{a^{2}}\frac{C_{+sc}}{\sqrt{c_{2sc}z_{+sc}^{2}}}\log{\left(\frac{16\,c^{2}_{2sc}\,z_{+sc}^{2}\,b_{sc}^{\prime\prime}}{c_{1sc}^{\prime 2}2b_{sc}(z_{+sc}-1)^{2}}\right)}\,.\end{split}$ (30) Finally, the coefficients $\bar{a}$ and the contribution from $I_{D}(b)$ to $\bar{b}$ denoted by $b_{D}$ in (1) are $\begin{split}\bar{a}=&\frac{r_{sc}^{3}}{\sqrt{c_{2sc}}}\left[\frac{C_{-sc}}{r_{sc}r_{-}-a^{2}}+\frac{C_{+sc}}{r_{sc}r_{+}-a^{2}}\right]\end{split}$ (31) and $\begin{split}b_{D}=\bar{a}\log{\left[\frac{8c_{2sc}^{2}b_{sc}^{\prime\prime}}{c_{1sc}^{\prime 2}b_{sc}}\right]}+\frac{2r_{sc}^{3}}{\sqrt{c_{2sc}}}\left[\frac{C_{-sc}}{r_{sc}r_{-}-a^{2}}\log{\left(1-\frac{a^{2}}{r_{sc}r_{-}}\right)}+\frac{C_{+sc}}{r_{sc}r_{+}-a^{2}}\log{\left(1-\frac{a^{2}}{r_{sc}r_{+}}\right)}\right]\,,\end{split}$ (32) where $z_{\pm}$ are replaced by $r_{\pm}$ through (20). The leading order result in the SDL from the integration of $f_{R}(z,r_{sc})$, which contributes the coefficient $\bar{b}$, is denoted by $b_{R}$, and is obtained as $\begin{split}b_{R}&=I_{R}(r_{sc})=\int_{0}^{1}f_{R}(z,r_{sc})dz\\\ =&\frac{2r_{0}^{3}}{a^{2}}\frac{C_{-}}{\sqrt{c_{2}}z_{-}}\log{\left(\frac{z_{-}}{z_{-}-1}\frac{\sqrt{c_{2}+c_{3}}+\sqrt{c_{2}}}{\sqrt{c_{2}+c_{3}}-\sqrt{c_{2}}}\frac{c_{3}}{4c_{2}}\right)}\\\ &+\frac{2r_{0}^{3}}{a^{2}}\frac{C_{-}}{\sqrt{c_{2}+c_{3}z_{-}}z_{-}}\log{\left(\frac{\sqrt{c_{2}+c_{3}z_{-}}-\sqrt{c_{2}+c_{3}}}{\sqrt{c_{2}+c_{3}z_{-}}+\sqrt{c_{2}+c_{3}}}\frac{\sqrt{c_{2}+c_{3}z_{-}}+\sqrt{c_{2}}}{\sqrt{c_{2}+c_{3}z_{-}}-\sqrt{c_{2}}}\right)}\\\ &+\frac{2r_{0}^{3}}{a^{2}}\frac{C_{+}}{\sqrt{c_{2}}z_{+}}\log{\left(\frac{z_{+}}{z_{+}-1}\frac{\sqrt{c_{2}+c_{3}}+\sqrt{c_{2}}}{\sqrt{c_{2}+c_{3}}-\sqrt{c_{2}}}\frac{c_{3}}{4c_{2}}\right)}\\\ &+\frac{2r_{0}^{3}}{a^{2}}\frac{C_{+}}{\sqrt{c_{2}+c_{3}z_{+}}z_{+}}\log{\left(\frac{\sqrt{c_{2}+c_{3}z_{+}}-\sqrt{c_{2}+c_{3}}}{\sqrt{c_{2}+c_{3}z_{+}}+\sqrt{c_{2}+c_{3}}}\frac{\sqrt{c_{2}+c_{3}z_{+}}+\sqrt{c_{2}}}{\sqrt{c_{2}+c_{3}z_{+}}-\sqrt{c_{2}}}\right)}\Big{|}_{r_{0}=r_{sc}}\;.\end{split}$ (33) Thus, the coefficient $\bar{b}$ can be computed from the sum of $b_{D}$ and $b_{R}$ $\bar{b}=-\pi+b_{D}+b_{R}\,$ (34) with the help of (32) and (33). In (33) we again use (28) and (20) to replace $c_{3sc}$ by $c_{2sc}=-\frac{2}{3}c_{3sc}$ and $z_{\pm}$ by $r_{\pm}$. After some straightforward algebra we find $\begin{split}\bar{b}=&-\pi+\bar{a}\log{\left(\frac{36}{7+4\sqrt{3}}\frac{8c_{2sc}^{2}b_{sc}^{\prime\prime}}{c_{1sc}^{\prime 2}b_{sc}}\right)}\\\ &+\frac{r_{sc}^{3}}{\sqrt{c_{2sc}}}\frac{2aC_{-sc}}{a^{2}-r_{sc}r_{-}}\frac{\sqrt{3}}{\sqrt{a^{2}+2r_{sc}r_{-}}}\log{\left(\frac{\sqrt{a^{2}+2r_{sc}r_{-}}-a}{\sqrt{a^{2}+2r_{sc}r_{-}}+a}\frac{\sqrt{a^{2}+2r_{sc}r_{-}}+\sqrt{3}a}{\sqrt{a^{2}+2r_{sc}r_{-}}-\sqrt{3}a}\right)}\\\ &+\frac{r_{sc}^{3}}{\sqrt{c_{2sc}}}\frac{2aC_{+sc}}{a^{2}-r_{sc}r_{+}}\frac{\sqrt{3}}{\sqrt{a^{2}+2r_{sc}r_{+}}}\log{\left(\frac{\sqrt{a^{2}+2r_{sc}r_{+}}-a}{\sqrt{a^{2}+2r_{sc}r_{+}}+a}\frac{\sqrt{a^{2}+2r_{sc}r_{+}}+\sqrt{3}a}{\sqrt{a^{2}+2r_{sc}r_{+}}-\sqrt{3}a}\right)}\,.\end{split}$ (35) Using the results of $r_{sc}$ (13), $b_{sc}$ (14) and the expression of $b(r_{0})$ (11), together with the definitions of $C_{\pm}$ and $c_{2}$ in (21) and (18) respectively, one can compute the coefficients $\bar{a}$ and $\bar{b}$ given by (31) and (35) in the form of (1). Notice that with the parameters under investigation $\bar{a}>0$, but $\bar{b}<0$. Our results are shown in Fig.(3), where both $\bar{a}$ and $|\bar{b}|$ increase (decrease) in $a$ in direct (retrograde) orbits, giving the fact that the deflection angle $\hat{\alpha}$ decreases (increases) with the increase of the black hole’s spin for a given impact parameter. Later in Sec. III we will compare with the full numerical computations from (22) in the SDL. The results of $\bar{a}$ and $\bar{b}$ due to the Schwarzschild black hole in Bozza2 ; Tsuka1 can be reproduced by sending $a\to 0$ where $r_{+}\rightarrow 2M$, $r_{-}\to a^{2}/2M$, $C_{+sc}\to 2M/r_{sc}$, $C_{-sc}\to a^{3}/2b_{sc}Mr_{sc}$, and $c_{2sc}\to r_{sc}^{2}$ using $c_{1sc}=0$ in (4) and (21). We can check that $\bar{a}=1$ in (31) and $\bar{b}$ in (35) reduces to the expression proportional to $\bar{a}$ given by $\begin{split}\bar{b}=&-\pi+\bar{a}\log{\left(36(7-4\sqrt{3})\frac{8c_{2sc}^{2}b_{sc}^{\prime\prime}}{c_{1sc}^{\prime 2}b_{sc}}\right)}\\\ =&-\pi+\log{\left(216(7-4\sqrt{3})\right)}\,.\end{split}$ (36) In the second equality above we have further used substitutions $b_{sc}\to 3\sqrt{3}M$, $b_{sc}^{\prime\prime}\to\sqrt{3}/M$, $c_{1sc}^{\prime}\to 6M$, and $c_{2sc}\to 9M^{2}$ obtained from $r_{sc}=3M$ in the Schwarzschild black hole. In Fig.(5), we compare the approximate results of the deflection angle in the SDL with the exact ones in Iyer2 and Hsiao , and find that they are in good agreement when $b\to b_{sc}$. The analytical expressions of the coefficient $\bar{a}$ and $\bar{b}$ in the form of the SDL deflection angle due to the Kerr black hole are successfully achieved. They are an extension of the works in Bozza3 and Tsuka1 where the spherically symmetric black holes are considered. This is one of the main results in this work. ### II.2 Kerr-Newman black holes We now consider another example with the nonspherically symmetric metric of a charged spinning black hole. With an addition of charge $Q$ comparing with the Kerr case, the line element associated with the Kerr-Newman metric is $\displaystyle{ds}^{2}$ $\displaystyle=$ $\displaystyle g_{\mu\nu}dx^{\mu}dx^{\nu}$ (37) $\displaystyle=$ $\displaystyle-\frac{\left(\Delta-a^{2}\sin^{2}\theta\right)}{\Sigma}{dt}^{2}+\frac{a\sin^{2}\theta\left(Q^{2}-2Mr\right)}{\Sigma}({dt}{d\phi+d\phi dt)}$ $\displaystyle+\frac{\Sigma}{\Delta}dr^{2}+\Sigma{\,d\theta}^{2}+\frac{\sin^{2}\theta}{\Sigma}\left((r^{2}+a^{2})^{2}-a^{2}\Delta\sin^{2}\theta\right){d\phi}^{2}\,,$ where $\Sigma=r^{2}+a^{2}\cos^{2}\theta\,,\quad\Delta=r^{2}+a^{2}+Q^{2}-2Mr\,.$ (38) The outer (inner) event horizon $r_{+}$ ($r_{-}$) is $r_{\pm}=M\pm\sqrt{M^{2}-(Q^{2}+a^{2})}\,$ (39) with $M^{2}>Q^{2}+a^{2}$. The light rays traveling on the equatorial plane of the black hole have been studied analytically in our previous work in Hsiao , in which the function $W_{\text{eff}}$ from the equation of motion along the radial direction in (8) can be regarded as an effective potential given by $\displaystyle W_{\text{eff}}(r)=\frac{1}{r^{2}}\left[1-\frac{a^{2}}{b^{2}}+\left(-\frac{2M}{r}+\frac{Q^{2}}{r^{2}}\right)\left(1-\frac{a}{b_{s}}\right)^{2}\right]\,.$ (40) For the Kerr-Newman black hole, the nonzero charge of the black hole is found to give repulsive effects to light rays as seen from its contributions to the function $W_{\text{eff}}$ of the $1/r^{4}$ term, which shifts the innermost circular trajectories of the light rays toward the black holes with the smaller critical impact parameter $b_{sc}$ for both direct and retrograde orbits, as illustrated in Fig.(2). Also, the presence of black hole’s charge is to decrease the deflection angle due to the additional repulsive effects on the light rays, as compared with the Kerr case with the same impact parameter $b$ Hsiao . As we will discuss in the next section, the angular positions of the relativistic images of the distant light sources due to the gravitational lensing of the black holes critically depends on the critical impact parameter $b_{sc}$. The impact parameter $b$ as a function of the radius of the circular motion $r_{0}$ is obtained as $\displaystyle b(r_{0})=\frac{s(2aM-a\frac{Q^{2}}{r_{0}})-r_{0}\sqrt{(a\frac{Q^{2}}{r_{0}^{2}}-a\frac{2M}{r_{0}})^{2}+(1-\frac{2M}{r_{0}}+\frac{Q^{2}}{r_{0}^{2}})[a^{2}(1+\frac{2M}{r_{0}}-\frac{Q^{2}}{r_{0}^{2}})+r_{0}^{2}]}}{2M-r_{0}-\frac{Q^{2}}{r_{0}}}\,.$ (41) The solution of $r_{sc}$ of the radius of the innermost circular motion has been found in Hsiao as $\displaystyle r_{sc}$ $\displaystyle=\frac{3M}{2}+\frac{1}{2\sqrt{3}}\sqrt{9M^{2}-8Q^{2}+U_{c}+\frac{P_{c}}{U_{c}}}$ $\displaystyle\quad-\frac{s}{2}\sqrt{6M^{2}-\frac{16Q^{2}}{3}-\frac{1}{3}\left(U_{c}+\frac{P_{c}}{U_{c}}\right)+\frac{8\sqrt{3}Ma^{2}}{\sqrt{9M^{2}-8Q^{2}+U_{c}+\frac{P_{c}}{U_{c}}}}}\;\;,$ (42) where $\displaystyle P_{c}$ $\displaystyle=(9M^{2}-8Q^{2})^{2}-24a^{2}(3M^{2}-2Q^{2})\,,$ $\displaystyle U_{c}$ $\displaystyle=\bigg{\\{}(9M^{2}-8Q^{2})^{3}-36a^{2}(9M^{2}-8Q^{2})(3M^{2}-2Q^{2})+216M^{2}a^{4}$ $\displaystyle\quad\quad+24\sqrt{3}a^{2}\sqrt{(M^{2}-a^{2}-Q^{2})\left[Q^{2}(9M^{2}-8Q^{2})^{2}-27M^{4}a^{2}\right]}\bigg{\\}}^{\frac{1}{3}}\,.$ (43) The analytical expression of the critical value of the impact parameter ${b_{sc}}$ can be written as a function of black hole’s parameters Hsiao , $\displaystyle b_{sc}$ $\displaystyle=-a+\frac{M^{2}a}{2(M^{2}-Q^{2})}+\frac{s}{2\sqrt{3}(M^{2}-Q^{2})}\Bigg{[}\sqrt{V+(M^{2}-Q^{2})\left(U+\frac{P}{U}\right)}$ $\displaystyle+\sqrt{2V-(M^{2}-Q^{2})\left(U+\frac{P}{U}\right)-\frac{s6\sqrt{3}M^{2}a\left[(M^{2}-Q^{2})(9M^{2}-8Q^{2})^{2}-M^{4}a^{2}\right]}{\sqrt{V+(M^{2}-Q^{2})\left(U+\frac{P}{U}\right)}}}\Bigg{]}\;,$ (44) where $\displaystyle P$ $\displaystyle=(3M^{2}-4Q^{2})\left[9(3M^{2}-4Q^{2})^{3}+8Q^{2}(9M^{2}-8Q^{2})^{2}-216M^{4}a^{2}\right]\,,$ $\displaystyle U$ $\displaystyle=\bigg{\\{}-\left[3(3M^{2}-2Q^{2})^{2}-4Q^{4}\right]\left[9M^{2}(9M^{2}-8Q^{2})^{3}-8\left[3(3M^{2}-2Q^{2})^{2}-4Q^{4}\right]^{2}\right]\,$ $\displaystyle\qquad+108M^{4}a^{2}\left[9(3M^{2}-4Q^{2})^{3}+4Q^{2}(9M^{2}-8Q^{2})^{2}-54M^{4}a^{2}\right]\,$ $\displaystyle\qquad+24\sqrt{3}M^{2}\sqrt{(M^{2}-a^{2}-Q^{2})\left[Q^{2}(9M^{2}-8Q^{2})^{2}-27M^{4}a^{2}\right]^{3}}\bigg{\\}}^{\frac{1}{3}}\,,$ $\displaystyle V$ $\displaystyle=3M^{4}a^{2}+(M^{2}-Q^{2})\left[6(3M^{2}-2Q^{2})^{2}-8Q^{4}\right]\,.$ (45) These will serve as the important inputs for the analytical expressions of the coefficients $\bar{a}$ and $\bar{b}$ in (1). Figure 2: The critical impact parameter $b_{sc}/M$ as a function of the spin parameter $a/M$ for (a) $Q/M=0.3$, (b) $Q/M=0.6$. Also, the critical impact parameter $b_{sc}/M$ as a function of charge $Q/M$ for (c) ${a/M}=0.3$, (d) ${a/M}=0.6$. The plots show the Schwarzschild, Reissner-Nordström, Kerr and Kerr-Newman black holes for comparison. The plot convention used henceforth: Kerr-Newman direct (solid red line), Kerr-Newman retrograde (solid blue line), Kerr direct (black dashed line with $Q=0$), Kerr retrograde (black dotted line, with $Q=0$), Reissner-Nordström (solid purple line, with $a=0$), and Schwarzschild (solid black line, with $Q=0,a=0$). The counterpart of (16) for the Kerr-Newman case as a function of $z$ in (15) can be easily derived giving $\frac{dz}{d\phi}=\frac{1}{r_{0}}\frac{1-\frac{2{M}}{r_{0}}(1-z)+\frac{a^{2}+Q^{2}}{r_{0}^{2}}(1-z)^{2}}{1-\frac{2{M}}{r_{0}}(1-\frac{a}{b_{s}})(1-z)+\frac{Q^{2}}{r_{0}^{2}}(1-\frac{a}{b_{s}})(1-z)^{2}}\sqrt{B(z,r_{0})}\;,$ (46) where $B(z,r_{0})=c_{1}(r_{0})z+c_{2}(r_{0})z^{2}+c_{3}(r_{0})z^{3}+c_{4}(r_{0})z^{4}\,.$ (47) The function $B(z,r_{0})$ is then the quartic polynomial in $z$ with the coefficients $\begin{split}c_{1}(r_{0})=&4Q^{2}\left(1-\frac{a}{b_{s}}\right)^{2}-6Mr_{0}\left(1-\frac{a}{b_{s}}\right)^{2}+2r_{0}^{2}\left(1-\frac{a^{2}}{b^{2}}\right)\,,\\\ c_{2}(r_{0})=&-6Q^{2}\left(1-\frac{a}{b_{s}}\right)^{2}+6Mr_{0}\left(1-\frac{a}{b_{s}}\right)^{2}-r_{0}^{2}\left(1-\frac{a^{2}}{b^{2}}\right)\,,\\\ c_{3}(r_{0})=&4Q^{2}\left(1-\frac{a}{b_{s}}\right)^{2}-2Mr_{0}\left(1-\frac{a}{b_{s}}\right)^{2}\,,\\\ c_{4}(r_{0})=&-Q^{2}\left(1-\frac{a}{b_{s}}\right)^{2}\,.\end{split}$ (48) All coefficients have the additional contributions from the charge $Q$. In particular, the presence of the $z^{4}$ term with the coefficient $c_{4}(r_{0})$ in $B$, which vanishes in the Kerr case, makes the calculations of $\bar{a}$ and $\bar{b}$ more involved. The integrant function $f(z,r_{0})$ in (22) now takes the form $f(z,r_{0})=\frac{r_{0}^{2}}{a^{2}+Q^{2}}\left(\frac{C_{-}}{z-z_{-}}+\frac{C_{Q}z+C_{+}}{z-z_{+}}\right)\frac{2{r_{0}}}{\sqrt{c_{1}(r_{0})z+c_{2}(r_{0})z^{2}+c_{3}(r_{0})z^{3}+c_{4}(r_{0})z^{4}}}\,.$ (49) The corresponding coefficients $C_{-}$, $C_{Q}$, and $C_{+}$ in the Kerr- Newman case are $\begin{split}C_{-}=&\frac{a^{2}+Q^{2}-2Mr_{-}(1-\frac{a}{b_{s}})+\frac{Q^{2}r_{-}^{2}}{a^{2}+Q^{2}}(1-\frac{a}{b_{s}})}{2r_{0}\sqrt{M^{2}-a^{2}-Q^{2}}}\,,\\\ C_{Q}=&\frac{Q^{2}}{r_{0}^{2}}\left(1-\frac{a}{b_{s}}\right)\,,\\\ C_{+}=&\frac{a^{2}+Q^{2}-2Mr_{-}(1-\frac{a}{b_{s}})+\frac{Q^{2}}{r_{0}}(r_{+}-r_{-})(1-\frac{a}{b_{s}})+Q^{2}(1-\frac{a}{b_{s}})}{-2r_{0}\sqrt{M^{2}-a^{2}-Q^{2}}}\,,\end{split}$ (50) where $z_{+}$, $z_{-}$ then become $\begin{split}z_{-}=&1-\frac{r_{0}r_{-}}{a^{2}+Q^{2}}\;,\\\ z_{+}=&1-\frac{r_{0}r_{+}}{a^{2}+Q^{2}}\;,\end{split}$ (51) defined in terms of the outer(inner) black hole horizon $r_{+}$ ($r_{-}$). Again, $z_{\pm}\leq 0$ for all $a$ and $Q$ with the nonzero $r_{+}$. Note that, for charge $Q\to 0$, $C_{Q}$ vanishes. Analogous to the previous subsection of the Kerr case, we define the function $f_{D}(z,r_{0})$ as $f_{D}(z,r_{0})=\frac{r_{0}^{2}}{a^{2}+Q^{2}}\left(\frac{C_{-}}{z-z_{-}}+\frac{C_{Q}z+C_{+}}{z-z_{+}}\right)\frac{2{r_{0}}}{\sqrt{c_{1}(r_{0})z+c_{2}(r_{0})z^{2}}}\,.$ (52) As $z\to 0$, $f_{D}(z,r_{0})\to 1/z$. Its integration over $z$ gives the divergent part of $I_{D}(r_{0})$ when $b\to b_{c}$. Here we find $\begin{split}I_{D}(r_{0})=&\int_{0}^{1}f_{D}(z,r_{0})dz\\\ =&\frac{2r_{0}^{3}}{a^{2}+Q^{2}}\frac{C_{-}}{\sqrt{c_{1}(r_{0})z_{-}+c_{2}(r_{0})z_{-}^{2}}}\log{\left(\frac{\sqrt{c_{1}(r_{0})z_{-}+c_{2}(r_{0})z_{-}}+\sqrt{c_{1}(r_{0})+c_{2}(r_{0})z_{-}}}{\sqrt{c_{1}(r_{0})z_{-}+c_{2}(r_{0})z_{-}}-\sqrt{c_{1}(r_{0})+c_{2}(r_{0})z_{-}}}\right)}\\\ &+\frac{2r_{0}^{3}}{a^{2}+Q^{2}}\frac{C_{+}+C_{Q}z_{+}}{\sqrt{c_{1}(r_{0})z_{+}+c_{2}(r_{0})z_{+}^{2}}}\log{\left(\frac{\sqrt{c_{1}(r_{0})z_{+}+c_{2}(r_{0})z_{+}}+\sqrt{c_{1}(r_{0})+c_{2}(r_{0})z_{+}}}{\sqrt{c_{1}(r_{0})z_{+}+c_{2}(r_{0})z_{+}}-\sqrt{c_{1}(r_{0})+c_{2}(r_{0})z_{+}}}\right)}\\\ &-\frac{2r_{0}^{3}}{a^{2}+Q^{2}}\frac{2C_{Q}}{\sqrt{c_{2}(r_{0})}}\log{\left(\sqrt{c_{1}(r_{0})}\sqrt{c_{2}(r_{0})}\right)}\\\ &+\frac{2r_{0}^{3}}{a^{2}+Q^{2}}\frac{2C_{Q}}{\sqrt{c_{2}(r_{0})}}\log{\left(c_{2}(r_{0})+\sqrt{c_{2}(r_{0})}\sqrt{c_{1}(r_{0})+c_{2}(r_{0})}\right)}\,.\end{split}$ (53) In the SDL, by substituting (29), $I_{D}(b)$ becomes $\begin{split}I_{D}(b)\approx&-\frac{r_{sc}^{3}}{a^{2}+Q^{2}}\left(\frac{C_{-sc}}{\sqrt{c_{2sc}z_{-sc}^{2}}}+\frac{C_{+sc}+C_{Qsc}z_{+sc}}{\sqrt{c_{2sc}z_{+sc}^{2}}}+\frac{C_{Qsc}}{\sqrt{c_{2sc}}}\right)\log{\left(\frac{b}{b_{sc}}-1\right)}\\\ &+\frac{r_{sc}^{3}}{a^{2}+Q^{2}}\frac{C_{-sc}}{\sqrt{c_{2sc}z_{-sc}^{2}}}\log{\left[\frac{16c_{2sc}^{2}z_{-sc}^{2}b_{sc}^{\prime\prime}}{c_{1sc}^{\prime 2}2b_{sc}(z_{-sc}-1)^{2}}\right]}\\\ &+\frac{r_{sc}^{3}}{a^{2}+Q^{2}}\frac{C_{+sc}+C_{Qsc}z_{+sc}}{\sqrt{c_{2}z_{+}^{2}}}\log{\left[\frac{16c_{2sc}^{2}z_{+sc}^{2}b_{sc}^{\prime\prime}}{c_{1}^{\prime 2}2b_{sc}(z_{+sc}-1)^{2}}\right]}+\frac{r_{sc}^{3}}{a^{2}+Q^{2}}\frac{C_{Qsc}}{\sqrt{c_{2sc}}}\log{\left[\frac{16c_{2sc}^{2}b_{sc}^{\prime\prime})}{c_{1sc}^{\prime 2}2b_{sc}}\right]}\,,\end{split}$ (54) from which we can read off the coefficients $\bar{a}$ and $b_{D}$ as follows $\begin{split}\bar{a}=&\frac{r_{sc}^{3}}{\sqrt{c_{2sc}}}\left[\frac{C_{-sc}}{r_{sc}r_{-}-(a^{2}+Q^{2})}+\frac{C_{+sc}}{r_{sc}r_{+}-(a^{2}+Q^{2})}\right]\end{split}$ (55) $\begin{split}b_{D}=&\bar{a}\log{\left[\frac{8c_{2sc}^{2}b_{sc}^{\prime\prime}}{c_{1sc}^{\prime 2}b_{sc}}\right]}\\\ &+\frac{2r_{sc}^{3}}{\sqrt{c_{2sc}}}\left[\frac{C_{-sc}}{r_{sc}r_{-}-(a^{2}+Q^{2})}\log{\left(1-\frac{a^{2}+Q^{2}}{r_{sc}r_{-}}\right)}+\frac{C_{+sc}+C_{Qsc}z_{+sc}}{r_{sc}r_{+}-(a^{2}+Q^{2})}\log{\left(1-\frac{a^{2}+Q^{2}}{r_{sc}r_{+}}\right)}\right]\,.\end{split}$ (56) They reduce to their counterparts in (31) and (32) respectively as $Q\to 0$. As for the remaining contributions to the regular part, and in the SDL, we have $\begin{split}&b_{R}=I_{R}(r_{sc})=\int_{0}^{1}f_{R}(z,r_{sc})dz\\\ &=\frac{2r_{0}^{3}}{a^{2}+Q^{2}}\frac{C_{-}}{\sqrt{c_{2}}z_{-}}\log{\left(\frac{z_{-}}{z_{-}-1}\frac{2c_{2}+c_{3}+2\sqrt{c_{2}+c_{3}+c_{4}}\sqrt{c_{2}}}{4c_{2}}\right)}\\\ &+\frac{2r_{0}^{3}}{a^{2}+Q^{2}}\frac{C_{-}}{z_{-}\sqrt{c_{2}+c_{3}z_{-}+c_{4}z^{2}_{-}}}\log{\left(\frac{z_{-}-1}{z_{-}}\frac{\left(\sqrt{c_{2}+c_{3}z_{-}+c_{4}z_{-}^{2}}+\sqrt{c_{2}}\right)^{2}-c_{4}z_{-}^{2}}{\left(\sqrt{c_{2}+c_{3}z_{-}+c_{4}z_{-}^{2}}+\sqrt{c_{2}+c_{3}+c_{4}}\right)^{2}-c_{4}(z_{-}-1)^{2}}\right)}\\\ &+\frac{2r_{0}^{3}}{a^{2}+Q^{2}}\frac{C_{+}}{\sqrt{c_{2}}z_{+}}\log{\left(\frac{z_{+}}{z_{+}-1}\frac{2c_{2}+c_{3}+2\sqrt{c_{2}+c_{3}+c_{4}}\sqrt{c_{2}}}{4c_{2}}\right)}\\\ &+\frac{2r_{0}^{3}}{a^{2}+Q^{2}}\frac{C_{+}+C_{Q}z_{+}}{z_{+}\sqrt{{c_{2}+c_{3}z_{-}+c_{4}z^{2}_{-}}}}\log{\left(\frac{z_{+}-1}{z_{-}}\frac{\left(\sqrt{c_{2}+c_{3}z_{+}+c_{4}z_{+}^{2}}+\sqrt{c_{2}}\right)^{2}-c_{4}z_{+}^{2}}{\left(\sqrt{c_{2}+c_{3}z_{+}+c_{4}z_{+}^{2}}+\sqrt{c_{2}+c_{3}+c_{4}}\right)^{2}-c_{4}(z_{+}-1)^{2}}\right)}\\\ &+\frac{2r_{0}^{3}}{a^{2}+Q^{2}}\frac{C_{Q}}{\sqrt{c_{2}}}\log{\left(\frac{z_{+}}{z_{+}-1}\right)}\Big{|}_{r_{0}=r_{sc}}\end{split}$ (57) In the limit of $Q\to 0$, where $C_{Q}$ and $c_{4}$ go to zero, the above expression of $b_{R}$ reduces to (33) in the Kerr case after implementing straightforward algebra. The coefficient $\bar{b}$ is obtained using (56) and (57) as $\begin{split}\bar{b}=&-\pi+\bar{a}\log{\left[\frac{36}{4(1-c_{4sc}/c_{2sc})^{2}+4\sqrt{3}(1-c_{4sc}/c_{2sc})^{3/2}+3(1-c_{4sc}/c_{2sc})}\frac{8c_{2sc}^{2}b_{sc}^{\prime\prime}}{c_{1sc}^{\prime 2}b_{sc}}\right]}\\\ &+\frac{r_{sc}^{3}}{\sqrt{c_{2sc}}}\frac{2(a^{2}+Q^{2})C_{-sc}}{(a^{2}+Q^{2}-r_{sc}r_{-})}\frac{\sqrt{3}}{{P_{-}}}\\\ &\quad\quad\times\log\left[\frac{-r_{sc}r_{-}}{a^{2}+Q^{2}-r_{sc}r_{-}}\frac{\left({P_{-}}+\sqrt{3}({a^{2}+Q^{2}})\right)^{2}-3\left({a^{2}+Q^{2}}-{r_{sc}r_{-}}\right)^{2}({c_{4sc}}/{c_{2sc}})}{\left(P_{-}+(a^{2}+Q^{2})(1-c_{4sc}/c_{2sc})^{1/2}\right)^{2}-3{r^{2}_{sc}r^{2}_{-}}({c_{4sc}}/{c_{2sc}})}\right]\\\ &+\frac{r_{sc}^{3}}{\sqrt{c_{2sc}}}\frac{2[(a^{2}+Q^{2})C_{+sc}+(a^{2}+Q^{2}-r_{sc}r_{+})C_{Qsc}]}{(a^{2}+Q^{2}-r_{sc}r_{+})}\frac{\sqrt{3}}{{P_{+}}}\\\ &\quad\quad\times\log{\left[\frac{-r_{sc}r_{+}}{a^{2}+Q^{2}-r_{sc}r_{+}}\frac{\left({P_{+}}+\sqrt{3}({a^{2}+Q^{2}})\right)^{2}-3\left({a^{2}+Q^{2}}-{r_{sc}r_{+}}\right)^{2}({c_{4sc}}/{c_{2sc}})}{\left(P_{+}+(a^{2}+Q^{2})(1-c_{4sc}/c_{2sc})^{1/2}\right)^{2}-3{r^{2}_{sc}r^{2}_{+}}({c_{4sc}}/{c_{2sc}})}\right]}\,.\end{split}$ (58) In the equation above, we have replaced $c_{3sc}$ by the linear combination of $c_{2sc}$ and $c_{4sc}$ in (48), given by $\begin{split}c_{3sc}=&-\frac{2}{3}c_{2sc}-\frac{4}{3}c_{4sc}\,.\end{split}$ (59) We also have $\begin{split}P^{2}_{\pm}&=(a^{2}+Q^{2}+2r_{sc}r_{\pm})(a^{2}+Q^{2})-(a^{2}+Q^{2}+r_{sc}r_{\pm})(a^{2}+Q^{2}-r_{sc}r_{\pm})(c_{4sc}/c_{2sc})\,.\end{split}$ (60) Combining (41),(II.2) and (II.2), the coefficients $\bar{a}$ and $\bar{b}$ in (55) and (58) can be analytically expressed as a function of the black hole’s parameters in the SDL. Our results are ploted in Fig.(4). Again, notice that $\bar{a}>0$ but $\bar{b}<0$ with the parameters in the figure. Due to the fact that the bending angle for light rays resulting from the charged black hole is suppressed as compared with the neutral black hole with the same impact parameter $b$, both $\bar{a}$ and $|\bar{b}|$ are found to increase with the charge $Q$. Figure 3: The coefficients $\bar{a}$ and $\bar{b}$ as a function of the spin parameter $a/M$ for the Kerr black hole with $Q/M=0$ and the Kerr-Newman black hole with $Q/M=0.6$: (a) the coefficient $\bar{a}$, (b) the coefficient $\bar{b}$. The display of the plot follows the convention in Fig.(2). Figure 4: The coefficients of $\bar{a}$ and $\bar{b}$ as a function of charge $Q/M$ for the Schwarzschild black hole with ${a/M}=0$, ${Q/M}=0$, the the Reissner- Nordström blck hole with ${a/M}=0$ and the Kerr-Newman black hole with ${a/M}=0.6$: (a) the coefficient $\bar{a}$, (b) the coefficient $\bar{b}$. The display of the plot follows the convention in Fig.(2). Figure 5: The SDL deflection angle (dotted lines) and the exact one (solid lines): (a)${a/M}=0.5$ and ${Q/M}=0.6$, (b) Error between them defined by $(\hat{\alpha}_{exact}-\hat{\alpha})/\hat{\alpha}_{exact}\times 100\%$; (c) ${a/M}=0.5$ and ${Q/M}=0.3$,(d) Error; (e) ${a/M}=0.9$ and ${Q/M}=0.3$, (f) Error; (g) ${a/M}=0.5$ and ${Q/M}=0.8$, (h) Error. It is then quite straightforward to check that the coefficients $\bar{a}$ and $\bar{b}$ in the Kerr-Newmann case can reduce to those in (31) and (35) in the Kerr case by setting $c_{4}\to 0$ in the limit of $Q\to 0$, also leading to $P_{\pm}\to a\sqrt{a^{2}+2r_{sc}r_{\pm}}$. To compare with the Reissner- Nordström black hole in Tsuka1 ; Tsuka2 , it is known that the impact parameter $b$ as a function of $r_{0}$ is $b(r_{0})=\frac{r_{0}^{2}}{\sqrt{Q^{2}-2Mr_{0}+r_{0}^{2}}}$ (61) and the circular motion of light rays forms the photon sphere with the radius $r_{c}=\frac{3M+\sqrt{9M^{2}-8Q^{2}}}{2}\;.$ (62) The critical impact parameter as a function of $r_{c}$ is given by $b_{c}=\frac{r_{c}^{2}}{\sqrt{Mr_{c}-Q^{2}}}\,.$ (63) Notice the subscript is changed from $sc$ to $c$ since the same critical impact parameters are obtained for light rays in direct orbits and retrograde orbits in the case of the nonspinning black holes. In the limit of $a\to 0$, we have $C_{-sc}\to 0$ in (50) using the definition of $r_{-}$ in (39). Thus, the coefficient $\bar{a}$ in (55) can be further simplified using (50), (39) and $c_{1sc}=0$ giving $\bar{a}=\frac{r_{c}}{\sqrt{3Mr_{c}-4Q^{2}}}\,,$ (64) which reproduces the expression in Tsuka1 ; Tsuka2 . As for the coefficient $\bar{b}$, in the limit of $a\to 0$, apart from $C_{-sc}\to 0$, $(a^{2}+Q^{2})C_{+sc}+(a^{2}+Q^{2}-r_{sc}r_{+})C_{Qsc}\to 0$ as well. So, the coefficient $\bar{b}$ in (58) has the contribution only from the term proportional to $\bar{a}$. After substituting (61) and (48) in the limit of $a\to 0$ to (58) and going through nontrivial algebra, we indeed recover the compact analytical expression in Tsuka1 ; Tsuka2 : $\bar{b}=-\pi+\bar{a}\log{\left[\frac{8(3Mr_{c}-4Q^{2})^{3}}{M^{2}r_{c}^{2}(Mr_{c}-Q^{2})^{2}}\left(2\sqrt{Mr_{c}-Q^{2}}-\sqrt{3Mr_{c}-4Q^{2}}\right)^{2}\right]}\,.$ (65) Figure 5 shows good agreement between the obtained SDL expression and the exact one in Hsiao computed numerically when $b$ approaches $b_{c}$ for some values of $a$ and $Q$. In conclusion, we have successfully achieved the analytical expression of the coefficient $\bar{a}$ and $\bar{b}$ in the form (1) of the SDL deflection angle due to the spherically nonsymmetric black holes, although they look not as simple as in the cases of the spherically symmetric black holes. Additionally, the obtained expressions can reproduce the respective ones due to the Kerr, Reissner-Nordström black holes and also due to the Schwarzschild black hole by taking the appropriate limits of the black hole’s parameters. ## III Relativistic Images of Gravitational lens and applications to supermassive galactic black holes We consider the cases of the planar light rays with the lens diagram shown in Fig.(1), where $d_{L}$ and $d_{S}$ are the distances of the lens (black hole) and the light source from the observer, and also $d_{LS}$ represents the distance between the lens and the source. The line joining the observer and the lens is considered as a reference optical axis. The angular positions of the source and the image are measured from the optical axis, and are denoted by $\beta$ and $\theta$, respectively. The lens equation is given by $\tan s\beta=\tan\theta-\frac{d_{LS}}{d_{S}}[\tan\theta+\tan(\hat{\alpha}-\theta)]\;,$ (66) where $\hat{\alpha}$ is the deflection angle of light rays obtained from (22) that can be expressed in terms of the impact parameter $b$ as the light rays approach to the black holes. In Eiroa , it is mentioned that the lens equations are applied for the observer and the source immersed in the asymptotically flat spacetime, where the Kerr and Kerr-Newman black holes have the asymptotically flat metric. Also, in the small $\beta$ and $\theta$ limits, we will see that the approximate lens equations to be found later are the same ones in Bozza_2003 , in which the Kerr black holes are considered. In the SDL of our interest, when the light rays wind around the black hole $n$ times, the deflection angle $\hat{\alpha}$ can be approximately by (1). The angle appearing in the lens equation should be within $2\pi$ and can be obtained from the deflection angle $\hat{\alpha}$ subtracting $2n\pi$. Together with the relation between the impact parameter $b$ and the angular position of the image given by $b=d_{L}\sin\theta\;,$ (67) in Fig.(1), we can solve the lens equation (66) with a given angular position of the source $\beta$ for the angular position of the observed image $\theta$. In the SDL, when the angular position of the source is small, $\theta$ is expectedly small with the small impact parameter $b$. Then the lens equation (66) can be further simplified by $s\beta\simeq\theta-\frac{d_{LS}}{d_{S}}[\hat{\alpha}(\theta)-2n\pi]\,$ (68) and (67) can be approximated by $b\simeq d_{L}\theta$. This can reduce to the lens equations in Bozza_2003 , in which the small angle limits are considered. According to Bozza3 , the zeroth order solution $\theta_{sn}^{0}$ is obtained from $\hat{\alpha}(\theta_{sn}^{0})=2n\pi$. Using the SDL deflection angle in (1) we have then $\theta_{sn}^{0}=\frac{|b_{sc}|}{d_{L}}\left(1+e^{\frac{\bar{b}-2n\pi}{\bar{a}}}\right)$ (69) for $n=1,2,\cdots$. The angular position $\theta_{sn}$ decrease in $n$ and reaches the asymptotic angular position given by $\theta_{s\infty}=|b_{sc}|/{d_{L}}$ as $n\to\infty$. With the zeroth order solution (69), the expansion of $\hat{\alpha}(\theta)$ around $\theta=\theta_{sn}^{0}$ is written explicitly as $\hat{\alpha}(\theta)=\hat{\alpha}(\theta_{sn}^{0})-\frac{\bar{a}}{e^{(\bar{b}-2n\pi)/\bar{a}}}\frac{d_{L}d_{LS}}{|b_{sc}|d_{S}}(\theta-\theta_{sn}^{0})+{O}(\theta-\theta_{sn}^{0})^{2}\,.$ (70) Then the approximate lens equation (68) to the order $(\theta-\theta_{sn}^{0})$ becomes $s\beta\simeq\theta_{sn}^{0}+\left(1+\frac{\bar{a}}{e^{(\bar{b}-2n\pi)/\bar{a}}}\frac{d_{L}d_{LS}}{|b_{sc}|d_{S}}\right)(\theta-\theta_{sn}^{0})\,.$ (71) Solving for $\theta$, by keeping the lowest order term in $|b_{sc}|/d_{L}\ll 1$, we find the angular position of the image as Bozza2 $\begin{split}\theta_{sn}\simeq\theta_{sn}^{0}+\frac{e^{(\bar{b}-2n\pi)/\bar{a}}}{\bar{a}}\frac{|b_{sc}|d_{S}}{d_{LS}d_{L}}(s\beta-\theta_{sn}^{0})\;.\end{split}$ (72) We assume that either Kerr or Kerr-Newman black holes have the clockwise rotation shown in Fig.(1). The light rays emitted from the source circle around the black hole multiple times in the SDL along a direct orbit (red line) with $s=+1$, where both the image and the source end up in the same sides of the optical axis with the angular position $\theta_{+n}$ and/or along a retrograde orbit (blue line) with $s=-1$, where the image and the source are in the opposite sides with the angular position $\theta_{-n}$. We also define the angular position difference between the outermost image $\theta_{1\pm}$ and the asymptotic one near the optical axis as $\Delta\theta_{s}=\theta_{s1}-\theta_{s\infty}\,\,,$ (73) which is the value to compare with the resolution of the observation that allows to distinguish among a set of the relativistic images. We now compute the angular positions of the relativistic images of the sources for $n=1$ (the outermost image) due to either Kerr or Kerr-Newman black holes with the mass $M=4.1\times 10^{6}M_{\odot}$ and the distance $d_{L}=26000\;{\rm ly}$ of the supermassive black hole Sagittarius A* at the center of our Galaxy as an example. We also take the ratio to be $d_{LS}/d_{S}=1/2$. In Table 1 (2), we consider both the image and the source are in the same (opposite) sides of the optical axis, where the light rays travel along the direct (retrograde) orbits, and choose $\beta\sim\theta_{\pm 1}$. The angular positions of the relativistic images are computed by (72). In the case of $|b_{sc}|\ll d_{L}$, $\theta_{sn}$ is not sensitive to $\beta$ but mainly determined by $\theta_{sn}^{0}$ in (69). Given $\bar{a}$ and $|\bar{b}|$ of the magnitudes shown in Fig.(3) and (4), $e^{-\frac{|\bar{b}|+2n\pi}{\bar{a}}}\ll 1$. The behavior of $\theta_{sn}$ thus depends mainly on $|b_{sc}|$ as a function of angular momentum $a$ and charge $Q$ of the black holes. As discussed in the previous section, since the effects from the angular momentum of the black hole for direct orbits effectively induces more repulsive effects compared with the retrograde orbits, clearly shown in their effective potential $W_{\text{eff}}$ (9), the resulting $b_{+c}<|b_{-c}|$ yields asymmetric values of $\theta_{+1}<\theta_{-1}$ for the same $a$ and $Q$. These features are shown in the Tables 1 and 2. Additionally, we notice that $\theta_{+1}$ ($\theta_{-1}$) decreases (increase) in $a$ for fixed $Q$ resulting from the decrease (increase) of $b_{+c}$ ($|b_{-c}|$) as $a$ increases. As for $\Delta\theta$, for the same $Q$, $\Delta\theta_{+}$ increases with $a$ whereas $\Delta\theta_{-}$ decreases with $a$. In particular, $\Delta\theta_{+}$ can be increased from about $10^{-2}\mu\rm{as}$ with $a/M\sim 10^{-3}$ and $Q/M=10^{-3}$ to the value as high as $0.6\mu{\rm as}$ with $a/M=0.9$ and $Q/M=10^{-3}$, which certainly increases their observability by the current very long baseline interferometry (VLBI) Ulv ; Johnson_2020 . As for the finite $Q$ effects, also showing the repulsion to the light rays seen in the effective potential (40), both $\theta_{+1}$ and $\theta_{-1}$ decrease in $Q$ for fixed $a$, resulting in the slightly increase of $\Delta\theta_{\pm}$ as $Q$ increases. $a/{M}$ | $Q/{M}$ | $\theta_{+1}$ ($\mu$as) | $\hat{\alpha}$ | $b/M$ | $\theta_{+\infty}$ ($\mu$as) | $\Delta\theta_{+}$ ($\mu$as) ---|---|---|---|---|---|--- $10^{\tiny-3}$ | $10^{\tiny-3}$ | 26.4231 | $2\pi+32.8135$ ($\mu$as) | $5.2007$ | 26.3900 | 0.0331 | $0.3$ | 26.0217 | $2\pi+32.0563$ ($\mu$as) | $5.1217$ | 25.9866 | 0.0351 | $0.6$ | 24.7179 | $2\pi+29.4336$ ($\mu$as) | $4.8651$ | 24.6747 | 0.0432 | $0.8$ | 23.1445 | $2\pi+26.2837$ ($\mu$as) | $4.5554$ | 23.0849 | 0.0596 $0.5$ | $10^{\tiny-3}$ | 20.9290 | $2\pi+21.8561$ ($\mu$as) | $4.1193$ | 20.8119 | 0.1171 | $0.3$ | 20.4085 | $2\pi+20.8203$ ($\mu$as) | $4.0169$ | 20.2758 | 0.1327 | $0.6$ | 18.6189 | $2\pi+17.2398$ ($\mu$as) | $3.6646$ | 18.4049 | 0.2140 | $0.8$ | 16.0922 | $2\pi+12.1835$ ($\mu$as) | $3.1673$ | 15.5372 | 0.5550 $0.9$ | $10^{\tiny-3}$ | 15.1170 | $2\pi+10.2354$ ($\mu$as) | $2.9754$ | 14.4517 | 0.6653 | $0.3$ | 14.1818 | $2\pi+8.36638$ ($\mu$as) | $2.7913$ | 13.2701 | 0.9117 | $0.6$ | $\cdots$ | $\cdots$ | $\cdots$ | $\cdots$ | $\cdots$ | $0.8$ | $\cdots$ | $\cdots$ | $\cdots$ | $\cdots$ | $\cdots$ Table 1: Relativistic images on the same side of the source with the angular position $\beta=10$ ($\mu$as) where the light rays are along direct orbits seen in Fig.(1). $a/{M}$ | $Q/{M}$ | $\theta_{-1}$ ($\mu$as) | $\hat{\alpha}$ | $b/M$ | $\theta_{-\infty}$ ($\mu$as) | $\Delta\theta_{-}$ ($\mu$as) ---|---|---|---|---|---|--- $10^{\tiny-3}$ | $10^{\tiny-3}$ | 26.4433 | $2\pi+72.8286$ ($\mu$as) | $5.20464$ | 26.4103 | 0.0330 | $0.3$ | 26.0422 | $2\pi+72.0654$ ($\mu$as) | $5.12569$ | 26.0073 | 0.0349 | $0.6$ | 24.7395 | $2\pi+69.4844$ ($\mu$as) | $4.86931$ | 24.6966 | 0.0429 | $0.8$ | 23.1680 | $2\pi+66.3165$ ($\mu$as) | $4.56000$ | 23.1088 | 0.0592 $0.5$ | $10^{\tiny-3}$ | 31.1994 | $2\pi+82.4458$ ($\mu$as) | $6.14075$ | 31.1862 | 0.0132 | $0.3$ | 30.8561 | $2\pi+81.6482$ ($\mu$as) | $6.07318$ | 30.8422 | 0.0139 | $0.6$ | 29.7638 | $2\pi+79.5352$ ($\mu$as) | $5.85820$ | 29.7479 | 0.0159 | $0.8$ | 28.5058 | $2\pi+76.9916$ ($\mu$as) | $5.61059$ | 28.4866 | 0.0192 $0.9$ | $10^{\tiny-3}$ | 34.7203 | $2\pi+89.4397$($\mu$as) | $6.83374$ | 34.7130 | 0.0073 | $0.3$ | 34.4063 | $2\pi+88.5984$ ($\mu$as) | $6.77195$ | 34.3988 | 0.0075 | $0.6$ | $\cdots$ | $\cdots$ | $\cdots$ | $\cdots$ | $\cdots$ | $0.8$ | $\cdots$ | $\cdots$ | $\cdots$ | $\cdots$ | $\cdots$ Table 2: Relativistic images on the opposite side of the source with the angular position $\beta=10$ ($\mu$as) where the light rays are in retrograde orbits seen in Fig.(1). Another application of the analytical expression of the deflection angle on the equatorial plane is to consider the quasiequatorial gravitational lensing based upon the works of Bozza_2003 ; Gyu_2007 . In this situation, the polar angle $\theta$ is set to be slightly away from $\theta=\frac{\pi}{2}$ and now becomes time dependent. In the SDL, the deflection angle of light rays with the additional initial declination can also be cast into the form of (1) where the coefficients are replaced by $\hat{a}$ and $\hat{b}$. In particular, the coefficient $\hat{a}$ obtained from the slightly off the equatorial plane can be related by the coefficient $\bar{a}$ on the equatorial plane through the $\omega$ function as $\hat{a}=\omega(r_{sc})\,\bar{a}\;,$ (74) where $\omega$ depends on $r$, and in turn depends on the deflection angle $\phi(r)$. Notice that the above relation (74) involves $\omega$, which is evaluated at $r=r_{sc}$. In the case of the Kerr black hole, it is found that Bozza_2003 $\omega(r_{sc})=\frac{(r_{sc}^{2}+a^{2}-2Mr_{sc})\sqrt{b_{sc}^{2}-a^{2}}}{2Mar_{sc}+b_{sc}(r_{sc}^{2}-2Mr_{sc})}\,,$ (75) and thus for the Schwarzschild case we have $a\rightarrow 0$, $\omega\rightarrow 1$. Then, substituting (13) and (14) into (75), together with the expression of $\bar{a}$ in (31), through (74) gives $\hat{a}=1$ for the Kerr case. However, in the Kerr-Newman black hole, the straightforward calculations show that the above relation (74) still holds true. The detailed derivations will appear in our future publication. Thus, the coefficient $\hat{a}$ can be analytically given by the coefficient $\bar{a}$ in (55), together with the $\omega$ function in the Kerr-Newman case in below $\omega(r_{sc})=\frac{(r_{sc}^{2}+a^{2}+Q^{2}-2Mr_{sc})\sqrt{b_{sc}^{2}-a^{2}}}{-a(Q^{2}-2Mr_{sc})+b_{sc}(r_{sc}^{2}+Q^{2}-2Mr_{sc})}\,.$ (76) As $Q\rightarrow 0$, (76) reduces to (75) in the Kerr case. The behavior of $\hat{a}$ as a function of the charge $Q$ with the choices of the angular momentum $a$ for direct and retrograde orbits is displayed in Fig.(6). The value of $\hat{a}$ ($\hat{a}>1$) increases with $Q$ for both direct and retrograde orbits. According to Bozza_2003 ; Gyu_2007 , the magnification of relativistic images might formally diverge when the angular positions of the sources are at caustic points. The corresponding magnifying power close to caustic points due to the light rays winding around the black hole $n$ times is given by $\bar{\mu}_{n}$ with the ratio between two neighboring caustic points $\frac{\bar{\mu}_{n+1}}{\bar{\mu}_{n}}\propto e^{-\pi/{\hat{a}}}\,$ (77) depending only on $\hat{a}$. In the Kerr case with $\hat{a}=1$, this ratio is independent of the black hole angular momentum $a$, whereas in the Kerr-Newman case with $\hat{a}>1$ shown in the Fig.(6), the ratio decreases with $Q$ for both direct and retrograde orbits Gyu_2007 . Here we just sketch some of the effects from the charge $Q$ of the black hole on the magnification of relativistic images. To have the full pictures of the caustic points and find the magnification of relativistic images, it in fact deserves the extensive study to compute not only $\hat{a}$ but also $\hat{b}$ by following Bozza_2003 ; Gyu_2007 . The further extension from quasiequatorial plane to the full sky is also of great interest Gralla_2020a ; Gralla_2020b ; Johnson_2020 . Figure 6: The coefficient $\hat{a}$ as a function of the black hole charge $Q/M$ for the direct (retrograde) orbits with (a) $a/M=0.3$, (b) $a/M=0.6$. The display of the plot follows the convention in Fig.(2). ## IV Summary and outlook In summary, the dynamics of light rays traveling around the Kerr black hole and the Kerr-Newman black hole, respectively, is studied with the detailed derivations on achieving analytical expressions of $\bar{a}$ and $\bar{b}$ in the approximate form of the deflection angle in the SDL. Various known results are checked by taking the proper limits of the black hole’s parameters. The analytical expressions are then applied to compute the angular positions of relativistic images due to the supermassive galactic black holes. We find that the effects from the angular momentum $a$ for direct orbits of light rays and the charge $Q$ for both direct and retrograde orbits increase the angular separation of the outermost images from the others. Although the observation of relativistic images is a very difficult task Ulv , our studies show potentially increasing observability of the relativistic images from the effects of angular momentum and charge of the black holes. Hopefully, relativistic images will be observed in the near future. Through the analytical results we present in this work, one can reconstruct the black hole’s parameters that give strong lensing effects. As light rays travel on the quasiequatorial plane, our analytical results on the equatorial plane can also be applied to roughly estimate the relative magnifications of relativistic images with the sources near one of the caustic points by taking account of the dynamics of the light rays in the polar angle. The work of investigating the structure of the caustic points from the effects of the charge $Q$ of the Kerr-Newman black holes and the magnification of relativistic images is in progress. Also, inspired by the recent advent of horizon-scale observations of astrophysical black holes, the properties of null geodesics become of great relevance to astronomy. The recent work of Gralla_2020a ; Gralla_2020b ; Johnson_2020 provides an extensive analysis on Kerr black holes. We also plan to extend the analysis of null geodesic to Kerr-Newman black holes, focusing on the effects from the charge of black holes. ###### Acknowledgements. This work was supported in part by the Ministry of Science and Technology, Taiwan, under Grant No.109-2112-M-259-003. ## References * (1) C. W. Misner, K. S. Thorne, and J. A. Wheeler, Gravitation (W. H. Freeman and Company, San Francisco, 1973). * (2) J. B. Hartle, Gravity: An Introduction to Einstein’s General Relativity (Addison-Wesley, Reading, MA. 2003). * (3) P. Schneider, J. Ehlers, and E. E. Falco, Gravitational Lenses (Springer-Verlag, New York, Berlin, Heidelberg, 1992). * (4) K. S. Virbhadra and G. F. R. Ellis, Schwarzschild black hole lensing, Phys. Rev. D 62, 084003 (2000). * (5) S. Frittelli, T. P. Kling, and E. T. Newman, Spacetime perspective of Schwarzschild lensing, Phys. Rev. D 61, 064021 (2000). * (6) V. Bozza, S. Capozziello, G. Iovane, and G. Scarptta, Strong field limit of black hole gravitational lensing, Gen. Relativ. Gravit. 33, 1535 (2001). * (7) V. Bozza, Gravitational lensing in the strong field limit, Phys. Rev. D 66, 103001 (2002). * (8) V. Bozza, Quasiequatorial gravitational lensing by spinning black holes in the strong field limit, Phys. Rev. D 67, 103006 (2003). * (9) V. Bozza, Gravitational lensing by black holes, Gen. Relativ. Gravit. 42, 2269 (2010). * (10) E. F. Eiroa, G. E. Romero, and D. F. Torres, Reissner-Nordstrom black hole lensing, Phys. Rev. D 66, 024010 (2002). * (11) S. V. Iyer and A. O. Petters, Light’s bending angle due to black holes: From the photon sphere to infinity, Gen. Relativ. Gravit. 39, 1563 (2007). * (12) N. Tsukamoto, Deflection angle in the strong deflection limit in a general asymptotically flat, static, spherically symmetric space-time, Phys. Rev. D 95, 064035 (2017). * (13) N. Tsukamoto and Y. Gong, Retrolensing by a charged black hole, Phys. Rev. D 95, 064034 (2017). * (14) K. S. Virbhadra and C. R. Keeton, Time delay and magnification centroid due to gravitational lensing by black holes and naked singularities, Phys. Rev. D 77, 124014 (2008). * (15) R. Shaikh, P. Banerjee, S. Paul, and T. Sarkar, Strong gravitational lensing by wormholes, J. Cosmol. Astropart. Phys. 07, 028 (2019). * (16) D. Richstone et al., Supermassive black holes and the evolution of galaxies, Nature (London) 14, 395 (1998). * (17) A. Ghez, et al., High proper-motion stars in the vicinity of Sagittarius A*: Evidence for a supermassive black hole at the center of our galaxy, Astrophys. J. 509, 678 (1998). * (18) R. Schodel, et al. A star in a 15.2-year orbit around the black hole at the centre of the Milky Way, Nature (London) 419, 694 (2002). * (19) K. Akiyama et al., Event Horizon Telescope Collaboration, First M87 event horizon telescope results. I. The shadow of the supermassive black hole, Astrophys. J. 875, L1 (2019). * (20) K. Akiyama et al., Event Horizon Telescope Collaboration, First M87 event horizon telescope results. V. Physical origin of the asymmetric ring, Astrophys. J. 875, L5 (2019). * (21) K. Akiyama et al., Event horizon telescope collaboration, First M87 event horizon telescope results. VI. The shadow and mass of the central black hole, Astrophys. J. 875, L6 (2019). * (22) S. V. Iyer and E. C. Hansen, Light’s bending angle in the equatorial plane of a Kerr black hole, Phys. Rev. D 80, 124023 (2009). * (23) Y.-W. Hsiao, D.-S. Lee, and C.-Y. Lin, Equatorial light bending around Kerr-Newman black holes, Phys. Rev. D 101, 064070 (2020). * (24) J. R. Wilson, Some magnetic effects in stellar collapse and accretion, Ann. N.Y. Acad. Sci. 262, 123 (1975). * (25) T. Damour, R. Hanni, R. Ruffini, and J.Wilson, Regions of magnetic support of a plasma around a black hole, Phys. Rev. D 17, 1518 (1978). * (26) C.-Y. Liu, D.-S. Lee, and C.-Y. Lin, Geodesic motion of neutral particles around a Kerr-Newman black hole, Classical Quantum Gravity 34, 235008 (2017). * (27) C. Jiang and W. Lin, Post-Newtonian light propagation in Kerr-Newman spacetime, Phys. Rev. D 97, 024045 (2018). * (28) G. V. Kraniotis, Gravitational lensing and frame dragging of light in the Kerr-Newman and the Kerr-Newman (anti) de Sitter black hole spacetimes, Gen. Relativ. Gravit. 46, 1818 (2014). * (29) J. S. Ulvestad, Goals of the ARISE Space VLBI Mission, New Astron.Rev. 43, 531 (1999). * (30) M. D. Johnson, et al. Universal interferometric signatures of a black hole’s photon ring, Sci. Adv. 6, eaaz1310 (2020). * (31) G. N. Gyulchev and S. S. Yazadjiev, Kerr-Sen dilaton-axion black hole lensing in the strong deflection limit, Phys. Rev. D 75, 023006 (2007). * (32) S. E. Gralla and A. Lupsasca, Null geodesics of the Kerr exterior, Phys. Rev. D 101, 044032 (2020). * (33) S. E. Gralla and A. Lupsasca, Lensing by Kerr black holes, Phys. Rev. D 101, 044031 (2020).
# Does a Hybrid Neural Network based Feature Selection Model Improve Text Classification? Suman Dowlagar LTRC IIIT-Hyderabad suman.dowlagar@ research.iiit.ac.in &Radhika Mamidi LTRC IIIT-Hyderabad radhika.mamidi@ iiit.ac.in ###### Abstract Text classification is a fundamental problem in the field of natural language processing. Text classification mainly focuses on giving more importance to all the relevant features that help classify the textual data. Apart from these, the text can have redundant or highly correlated features. These features increase the complexity of the classification algorithm. Thus, many dimensionality reduction methods were proposed with the traditional machine learning classifiers. The use of dimensionality reduction methods with machine learning classifiers has achieved good results. In this paper, we propose a hybrid feature selection method for obtaining relevant features by combining various filter-based feature selection methods and fastText classifier. We then present three ways of implementing a feature selection and neural network pipeline. We observed a reduction in training time when feature selection methods are used along with neural networks. We also observed a slight increase in accuracy on some datasets. ## 1 Introduction Text classification assigns one or more class labels from a predefined set to a document based on its content. Text classification has broad applications in real-world scenarios such as document categorization, news filtering, spam detection, Optical character recognition (OCR), and intent recognition. Giving high weights to relevant features is the objective of text classification. The field of text classification has gained more interest during the machine learning (ML) era. Many discriminative and generative machine learning classifiers have achieved excellent results in the field of text classification Deng et al. (2019). Feature selection and feature extraction methods are often used to reduce high dimensionality Bharti and Singh (2015). Feature extraction generates features from text Agarwal and Mittal (2014). Feature selection (FS) selects the most prominent features Saleh and El- Sonbaty (2007). These feature selection and extraction methods are used along with traditional classification algorithms. These methods reduced the curse of dimensionality and increased the classification accuracy Deng et al. (2019). Recently, deep learning models are used to learn better text representations and to classify the text Minaee et al. (2020). Such models include convolutional neural networks (CNN) Kim (2014), recurrent neural networks (RNN) Hochreiter and Schmidhuber (1997), Transformer models Adhikari et al. (2019), and graph convolutional networks (GCN) Yao et al. (2019). These NN models capture semantic and syntactic information in local and global word sequences. Even though the neural networks capture a complex and dense representation of data, the set of words introducing noise in the classifier is still present. Such words add the burden of increased vocabulary, which results in increased textual representation and an increase in the training time of the classifiers Song et al. (2011). Similar to the traditional approaches, we want to understand the effects of using statistical feature selection algorithms beforehand to calculate the features’ relevance and then train a fastText text-classification algorithm on those relevant features. Using this feature selection and neural network pipeline, we assume that the complexity of dealing with larger vocabulary decreases. Including feature selection with fastText text-classification helps reduce the classifier’s training time and helps the classifier reach better local optima, showing a significant increase in classification accuracy. In this work, we analyzed a feature selection and neural network pipeline for text classification. We used a hybrid feature selection method to get a score on relevant features. Using this score, we formulated three methods. The first and second methods deal with modifying the original text by extracting the relevant features. The third method deals with using the feature selection scores and pass it along with the word embeddings. We then observed the effect of feature selection on various neural networks. The rest of the paper is organized as follows. Section 2 gives a brief review of previous works in the field of feature selection and text classification. Section 3 presents a detailed procedure of the proposed pipeline and presents the experiments and datasets used for our study. Section 4 reports the performance of text classifiers with and without feature selection methods. Section 5 concludes the paper. ## 2 Literature Survey This section presents a brief description of the neural network (NN) classification algorithms and various feature selection methods. ### 2.1 Deep Learning for Text Classification Nowadays, various NNs such as CNN, RNN, BERT, and Text GCN achieve state-of- the-art results on text classification. CNN uses 1d convolutions Zhang et al. (2015) and character level convolutions Conneau et al. (2016) to learn the semantic similarity of words or characters, which helps in classifying the text. RNN models such as GRU, LSTM, and BiLSTM Liu et al. (2016) take word to word sequences to learn a better textual representation of a document that helps in text classification. Attention mechanisms have been introduced in these LSTM models, which increased the representativeness of the text for better classification Yang et al. (2016). Transformer models such as BERT Devlin et al. (2018) uses the attention mechanism that learns contextual relations between words or sub-words in a text Adhikari et al. (2019). Text GCN Yao et al. (2019) uses a graph-convolutional network to learn a heterogeneous word document graph on the whole corpus. Text GCN can capture global word co-occurrence information and use graph convolutions to learn a global representation, which helps classify the documents. ### 2.2 Feature Selection on Text Data The text classification often involves extensive data with thousands of features. Although tens of thousands of words are in a typical text collection, most of them contain little or no information to predict the text label. These features introduce complexity and increase the training time of an ML classifier. Feature selection is one method for giving high scores to relevant features (Deng et al., 2019). The goal of feature selection is to select highly-relevant features with minimum redundancy. The relevance of a feature indicates that the feature is always necessary to predict the class label. There are various text feature selection methods in the literature, each being filter, wrapper, hybrid, and embedded methods. The filter method evaluates the quality of a feature using a scoring function. Some filter methods evaluate the goodness of a term based on how frequently it appears in a text corpus. Document Frequency (DF) Lam and Lee (1999) and Term Frequency - Inverse Document Frequency (TFIDF) Rajaraman and Ullman (2011) comes under this category. Other filter methods that originate from information theory are, Mutual Information (MI) Taira and Haruno (1999); Tang et al. (2019), Information Gain (IG) Yang and Pedersen (1997), CHI Rogati and Yang (2002), ANOVA F measure Elssied et al. (2014), Bi-Normal Separation (BNS) Forman (2003) and the GINI method Shang et al. (2013). They use hypothesis testing, contingency tables, mean and variance scores, conditional and posterior probabilities for selecting the features. The wrapper method Maldonado and Weber (2009) use a search strategy to construct each possible subset, feeds each subset to the chosen classifier, and then evaluates the classifier’s performance. These two steps are repeated until the desired quality of the feature subset is reached. The wrapper approach achieves better classification accuracy than filter methods. However, the time taken by the wrapper method is very high when compared to filter methods. Embedded methods complete the FS process within the construction of the machine learning algorithm itself. In other words, they perform feature selection during the model training. An embedded method is Decision Tree (DT) Quinlan (1986). In DT, while constructing the classifier, DT selects the best features/attributes that may give the best discriminative power. Hybrid methods are robust and take less time when compared to the wrapper and embedded methods. They combine a filter method with a wrapper method during the feature selection process. The HYBRID model Günal (2012) employs a combination of filter methods to select to rank the features and then a wrapper method to the obtained final features set. Our FS method is similar to the HYBRID model. A detailed report on the benefits of using the feature selection methods in the pipeline with traditional classifiers is presented in Deng et al. (2019); Forman (2003). Apart from using traditional classification methods, deep feature selection using neural networks were also proposed. These models use deep neural network autoencoders for the feature set reduction and text generation Mirzaei et al. (2019); Han et al. (2018). Lam and Lee (1999) studies the effect of feature set reduction before applying the neural network classifiers. The paper uses a multi-layer perceptron (MLP) classifier in combination with filter-based FS method. Alkhatib et al. (2017) proposes the use of neural network-based feature selection and text classification. Our work comes under this category. ## 3 Proposed Pipeline In this section, we present our feature selection and neural network pipeline. The feature selection and neural network pipeline start with selecting a good tokenizer to tokenize the data and create a feature set. The tokenizer used for our feature selection is the Sentencepiece tokenizer Kudo and Richardson (2018). Sentencepiece tokenizer implements subword units by using byte-pair- encoding (BPE) Sennrich et al. (2015) and unigram language model Kudo (2018). In the feature subset generation, we considered a hybrid feature selection method known as HYBRID Günal (2012). It has proved that a combination of the features selected by various methods is more effective and computationally faster than the features selected by individual filter and wrapper methods. Similar to the HYBRID model, we used three filters to obtain the relevancy score. The filters we considered were CHI2, ANOVA-F, and MI. These filters calculate the relevancy between the word and the class labels. Before feature selection, we used the Bag-of-Words(BoW) model to vectorize the data. In the BoW model, each feature vector is represented by $TF\\_IDF$ scores. Then we used statistical measures such as $\chi^{2}$, ANOVA-F, and MI for obtaining feature scores. $\chi^{2}$ 111A detailed explanation and a simple example of $\chi^{2}$ is given at https://www.mathsisfun.com/data/chi-square-test.html is a statistic to measure a relationship between feature vectors and a label vector. Analysis of Variance (ANOVA222A detailed explanation for ANOVA is given in https://towardsdatascience.com/anova-for-feature-selection-in-machine- learning-d9305e228476.) is a statistical method used to check the means of two or more groups that are significantly different from each other. Mutual Information (MI333A simple explanation and working python example of MI is available at https://machinelearningmastery.com/information-gain-and- mutual-information/) is frequently used to measure the mutual dependency between two variables. Using different statistical methods, we measured the relevance of each feature. We then aggregate the relevance scores of all satistical methods for each feature. The relevance of a feature $x_{i}$ is given by, $Relevancy(x_{i})=\\\ \chi^{2}(x_{i})+ANOVA(x_{i})+MI(x_{i})$ (1) Figure 1: Modifying text by masking the low ranked words Figure 2: Meta-Embeddings, including feature scores along with word embeddings Instead of an LR classifier given in the HYBRID model, we used the fastText classifier Joulin et al. (2016) for the feature selection. We used the fastText classifier as it is often on par with deep learning classifiers in terms of accuracy and performs faster computations. The fastText classifier treats the average of word embeddings as document embeddings, then feeds document embeddings into a feed-forward NN or a multinomial LR classifier. We used pre-trained fastText word embeddings Grave et al. (2018) while training a classifier. To get the final features list, we sorted the normalized, aggregated value in descending order and divided the entire feature space into k sets. In our model, we divided the sorted feature space into 20 sets. The value of k is fixed to 20 using a trial and error basis. We take the first set as the vocabulary of the classifier. We then trained the classifier and noted its accuracy. In the second iteration, we considered the vocabulary as the combination of first and second sets. Similarly, the third set has the vocabulary of the first three sets combined. We repeated the process until all the lists are exhausted. The set of features that resulted in a better classification metric is considered as the final feature set. According to the proposed FS method, the final feature set is considered relevant, and they are necessary to perform the text classification. In contrast, the other features have little to no effect on the text classification or might degrade the classifier’s performance. After feature subset generation, we propose three methods for including the feature selection information before training the neural network classifiers. 1. 1. Method 1 (Selecting only the relevant features)444This method is already used while selecting the final features set by the fastText classifier.: Like traditional classification algorithms, we select only the relevant features that are estimated to be important by the feature selection method before training a neural network classifier. 2. 2. Method 2 (Masking the features that were given low importance by our FS method): We felt that removing the features given low rank by our FS method might disturb the original data’s grammatical structure, thus disturbing the word to word dependencies. We masked the low ranked words with the help of $<MASK>+POS(word)$ tag. $<MASK>$ word masks the low ranked word, and POS preserves the word’s part of speech. The visual representation of method 2 is shown in figure 1 3. 3. Method 3 (Meta Embeddings): As shown in figure 2, we pass the relevancy and feature selection information along with embeddings in this method. Each slot holds the filter scores, i.e., CHI, ANOVA, MI scores of each feature. The last slot holds a 1 or 0 value. 1 is used for the selected features, and 0 is used for low ranked features that were not selected by our hybrid feature selection approach. We analyzed and evaluated the above methods with various state-of-the-art NN classifiers on the benchmark datasets. ### 3.1 Experiment In this section, we evaluated our feature selection and neural network pipeline on two tasks. We wanted to determine: * • If the pipeline decreases the training time of the classifier * • If it helps in obtaining better local optima, thus improving the classification accuracy. We tested our pipeline across multiple state-of-the-art text classification algorithms. 1. 1. CNN: Kim (2014) This convolutional neural network-based text classifier is trained by considering pre-trained word vectors. 2. 2. Bi-LSTM: Liu et al. (2016) A two-layer, bi-directional LSTM text classifier with pre-trained word embeddings as input was considered for the task of text classification. 3. 3. fastText: Joulin et al. (2016) This is a simple, efficient, and the fastest text classification method. It treats the average of word/n-grams embeddings as document embeddings, then feeds document embeddings into a linear classifier. 4. 4. Text GCN: Yao et al. (2019) Builds a heterogeneous word document graph for a whole corpus and turns document classification into a node classification problem. It uses GCN Kipf and Welling (2017) to learn word and document embeddings. 5. 5. DocBERT: Adhikari et al. (2019) A fine-tuned BERT model for document classification. The BERT model Devlin et al. (2018) uses a series of multiheaded attention and feedforward networks for various NLP tasks. ### 3.2 Datasets We ran our experiments on three widely used benchmark corpora and multilingual corpora. They are 20Newsgroups(20NG), R8, and R52 of Reuters 21578 and MLMRD. * • The 20NG dataset contains 18,846 documents divided into 20 different categories. 11,314 documents were used for training, and 7,532 documents were used for testing. * • R52 and R8 are two subsets of the Reuters 21578 dataset. R8 has 8 categories of the top eight document classes. It was split into 5,485 training and 2,189 test documents. R52 has 52 categories and was split into 6,532 training and 2,568 test documents. * • MLMRD is a Multilingual Movie Review Dataset. It consists of the genre and synopsis of movies across multiple languages, namely Hindi, Telugu, Tamil, Malayalam, Korean, French, and Japanese. The data set is minimal and unbalanced. It has 9 classes and a total of 14,997 documents. The data was split into 10,493 training and 4,504 test documents. We first preprocessed all the datasets by cleaning and tokenizing. The tokenizer used is the fastText tokenizer. For baseline 1 models, we used multilingual fastText embeddings Grave et al. (2018) of dimensionality 300, and baseline 2 models had the dimensionality of 304. We used default parameter settings as in their original papers for implementations. For calculating TFIDF, CHI2, ANOVA-F, MI scores, we used the scikit-learn library Pedregosa et al. (2011). For POS tagging, we used the NLTK Bird et al. (2009) pos tagger. All the neural network models were run on the GPU processor on the Windows platform with NVIDIA RTX 2070 graphics card. ## 4 Performance Datasets | Our FS | HYBRID FS ---|---|--- 20Newsgroups | 81.27% | 77.34% R8 | 96.94% | 93.79% R52 | 92.72% | 86.43% MLMRD | 47.09% | 42.98% Table 1: The classification accuracy of our FS model when compared to the HYBRID model. In our work, we modified the HYBRID Günal (2012) feature selection model by changing the LR classifier to the fastText classifier. We selected the fastText classifier in the feature selection process because of its fast learning ability of a NN model compared to the traditional ML classifiers and other neural network classifiers Joulin et al. (2016) without any decrease in classification accuracy. The neural network classifiers such as MLP, CNN, RNN, transformer, and GCN models achieve better classification accuracy when compared to traditional ML classifiers, but their training time is very high. Using a fastText classifier during feature selection, we observed that our model performed better on all the benchmark datasets than the HYBRID model. The results are shown in table 1. The fastText classifier’s use helped the model obtain better relevant features, increasing the current feature selection model’s accuracy compared to the HYBRID model. Datasets | 20Newsgroups | R8 | R52 | MLMRD ---|---|---|---|--- Baseline 1 & 2 | 1,01,631 (V) | 19,956 (V) | 26287 (V) | 94073 (V) Method 1 | 25732 (0.25V) | 17364 (0.87V) | 22372 (0.85V) | 52015 (0.55V) Method 2 | 25732+30 (0.25V) | 17364+30 (0.87V) | 22372+30 (0.85V) | 52015+143 (0.55V) Method 3 | 1,01,631 (V) | 19,956 (V) | 26287 (V) | 94073 (V) Table 2: The vocabulary size in all the FS inclusion methods when compared to the baselines. “V” is denoted as the vocabulary size of the actual data. Baselines 1,2, and method 3 have no change in vocabulary. However, using our FS method, the vocabulary is reduced to a maximum of 75% (for 20Newsgroups data). Other datasets have seen a 13% to 45% decrease in vocabulary size. We can see an increase in vocabulary from method 1 to method 2. It is due to the additional vocabulary resulted from the mask words when they are accompanied by pos tags. Here Penn Treebank POS tagset is used. Datasets | Method | Classifier(s) ---|---|--- | | CNN | Bi-LSTM | fastText | DocBERT | Text GCN 20Newsgroups | Baseline 1 | 79.31% | 73.60% | 81.04% | 90.19% | 86.13% | Baseline 2 | 79.46% | 74.25% | 82.44% | NA | 86.23% | Method 1 | 78.27% | 73.44% | 81.27% | 89.37% | 86.25% | Method 2 | 77.29% | 70.48% | 80.14% | 88.43% | 85.65% | Method 3 | 80.59% | 76.57% | 84.48% | NA | 86.15% R8 | Baseline 1 | 97.24% | 92.70% | 96.13% | 97.62% | 96.80% | Baseline 2 | 97.37% | 93.82% | 96.50% | NA | 96.94% | Method 1 | 97.39% | 93.74% | 96.94% | 97.44% | 96.28% | Method 2 | 96.57% | 94.34% | 96.07% | 97.44% | 96.85% | Method 3 | 97.39% | 96.74% | 97.18% | NA | 96.94% R52 | Baseline 1 | 94.78% | 87.53% | 92.02% | 92.95% | 93.56% | Baseline 2 | 94.84% | 90.79% | 92.76% | NA | 93.64% | Method 1 | 94.29% | 87.47% | 92.72% | 93.10% | 92.97% | Method 2 | 91.71% | 91.90% | 90.30% | 92.10% | 93.19% | Method 3 | 94.84% | 91.48% | 92.83% | NA | 93.74% MLMRD | Baseline 1 | 47.63% | 46.43% | 46.92% | 53.11% | 47.62% | Baseline 2 | 47.79% | 47.43% | 48.92% | NA | 49.62% | Method 1 | 44.98% | 44.82% | 47.09% | 51.90% | 46.58% | Method 2 | 44.63% | 44.05% | 46.61% | 50.90% | 46.98% | Method 3 | 48.44% | 49.13% | 49.55% | NA | 51.50% Table 3: Test accuracy on various neural network classifiers for the task of document classification. As the BERT model used is a fine-tuned one, we did not modify the model. As mentioned above, we used the training time-taken and test accuracy as the metrics to evaluate our approach. The accuracy and training time are recorded by running the model 10 times, and the average of the metrics was presented. ### 4.1 Effects of our methods on classification accuracy Table 3 demonstrates the accuracy of feature selection methods on NN classifiers. When methods 1 and 2 were used, there is a slight decrease in classification accuracy because the first two methods lost semantic connection among words. Thus, the classification performance is degraded. Also, some words which were relevant to the classifier were masked out during the FS method. Whereas in method 3, including the feature selection scores with word-embeddings, has shown a significant improvement in accuracy on all the datasets. Compared to the other datasets, the 20NG dataset has seen a significant decrease in vocabulary size. The vocabulary was decreased by 75%. However, eliminating those features did not affect the accuracy of the classifier for methods 1 and 2. Introducing the masked features in method 2 shown an increase in accuracy only in the Bi-LSTM method as this method considers word dependencies while training a classifier. Including the feature selection scores along with the word-embeddings improved the classification accuracy on all the datasets. The feature selection metadata helped the neural network classifier learn a better relationship between the words and classes and improve the classifier’s accuracy by reaching better local optima. In R8 and R52 datasets, we have seen an increase in accuracy using method 1 because our hybrid FS method worked better on these datasets by removing the noisy words without disturbing the relevant words. The maximum improvement in accuracy is shown in the R8 dataset, with a +4% increase in classification accuracy. Our approach did not show any better results on MLMRD datasets as this dataset has a limited number of documents to train and test the data for some languages (Telugu, Tamil, Malayalam, Korean). Reducing vocabulary size by the FS method decreased the classification accuracy. ### 4.2 Effects of our methods on training time The pictorial representation of time taken by the classifiers for all the datasets is given in appendix B of the supplementary material. The time taken by method 1 is lower than in all baseline models. In method 1, as the text is modified by considering only relevant features, the vocabulary size is reduced, and the sentence length is reduced. It resulted in the more accelerated training of the neural network. The time taken by baseline 2 and method 3 is similar because of the same embedding dimensionality of 304, but method 3 has achieved local optima a few epochs before compared to baseline 2, resulting in a time decrease of a few seconds. This phenomenon is attributed to the use of feature selection scores along with word embeddings. Method 2 has shown an increase in training time even though the vocabulary is decreased because of 2 factors. 1. 1. The masking of features created unknown words in the data, and the classifier has to be trained to learn the representation of masked words, whereas the other words had pre-trained embeddings. 2. 2. Apart from vocabulary, the neural network training time also depends on the input batch size given to the network and the length of the sentence in each batch. Because of the masked words, there is no decrease in either batch size or the sentence length. So the masking of data did not decrease the training time of the classifier. On the contrary, the Text GCN model has shown a decrease in training time because the classifier computes heterogeneous graph embeddings of each word based on the textual data before classification. It did not use any pre- trained embeddings. In method 3, there is a slight increase in training time because of increased vocabulary size due to the inclusion of feature selection metadata. Of all the NN classifiers, the Text GCN model had shown a maximum decrease in training time by 488 sec when method 1 was used on 20NG data. As the Text GCN operates on building a graph on the complete vocabulary of data, the time taken by the method to build the graph is reduced significantly by reducing the vocabulary size. It is followed by the DocBERT and Bi-LSTM on 20NG data with a decrease in training time by 480 and 394 sec. Text GCN and Bi-LSTM have shown a significant decrease in training time on all the datasets. On the contrary, fastText and CNN are very fast while training the NN model. The training time of such models was unchanged when our method 1 was used. When compared to all the classifiers, DocBERT achieved better results because of its evolutionary multi-headed attention and transformer models. As the Text GCN captures both local and word embeddings by constructing a heterogeneous graph, their results were better than those of the CNN and Bi-LSTM models, which work only on local word dependencies. As we increased the size of the embedding in FS method 2, this increased the dimensionality of vocabulary, resulting in the classifier’s increased training time. ## 5 Conclusion In our work, ”Does a Hybrid NN FS Model Improve Text Classification?”, we used the NN based hybrid FS method to extract relevant features and used NN classifiers for text classification. We extracted the relevant or high ranked features using filter-based methods and a fastText classifier. We then proposed three methods on how the feature selection can be included in the NN classification process. First, modifying the corpus by considering only relevant features. Second, modifying the data by masking the low ranked features, and the third method introduces feature selection information along with word embeddings. We observed that method 1 had shown a significant reduction in training time when large datasets or slower models are used, accompanied by a minimal change in classification accuracy. By introducing $MASK+P0S(word)$, we inferred that the masked word was a burden to the classifier, and it always tried to adjust the word embeddings, which resulted in increased epoch time during training and a slightly negative effect on classification accuracy. Whereas method 3 has shown no effect on decreasing the training time, it has shown a maximum of $4\%$ increase in the classification accuracy compared to baseline. It proved that introducing feature scores along with pre-trained word embeddings while training the NN classifier is beneficial. Instead of opting for random naive vocabulary reduction techniques such as using min_df and max_df (minimum and maximum document frequency) for selecting features, by using FS methods, we can calculate the relevance of the word beforehand and use that as metadata to the NN classifier. When the datasets are huge, these methods are of more significance. We can use the modified data while tuning the hyperparameters. Then we can use the real data to train and evaluate the model. Even in the critical domain datasets such as “medical”, we cannot rely on removing a word based on min_df and max_df scores. Each word in those datasets should be treated with utmost significance. FS methods help in such scenarios by calculating the word’s relevance and helps maintain better vocabulary before training neural network classifiers. ## References * Adhikari et al. (2019) Ashutosh Adhikari, Achyudh Ram, Raphael Tang, and Jimmy Lin. 2019. Docbert: Bert for document classification. _arXiv preprint arXiv:1904.08398_. * Agarwal and Mittal (2014) Basant Agarwal and Namita Mittal. 2014. Text classification using machine learning methods-a survey. In _Proceedings of the Second International Conference on Soft Computing for Problem Solving (SocProS 2012), December 28-30, 2012_ , pages 701–709, New Delhi. Springer India. * Alkhatib et al. (2017) Wael Alkhatib, Christoph Rensing, and Johannes Silberbauer. 2017. Multi-label text classification using semantic features and dimensionality reduction with autoencoders. In _International Conference on Language, Data and Knowledge_ , pages 380–394. Springer. * Bharti and Singh (2015) Kusum Kumari Bharti and Pramod Kumar Singh. 2015. Hybrid dimension reduction by integrating feature selection with feature extraction method for text clustering. _Expert Systems with Applications_ , 42(6):3105–3114. * Bird et al. (2009) Steven Bird, Ewan Klein, and Edward Loper. 2009. _Natural language processing with Python: analyzing text with the natural language toolkit_. ” O’Reilly Media, Inc.”. * Conneau et al. (2016) Alexis Conneau, Holger Schwenk, Loïc Barrault, and Yann Lecun. 2016. Very deep convolutional networks for text classification. _arXiv preprint arXiv:1606.01781_. * Deng et al. (2019) Xuelian Deng, Yuqing Li, Jian Weng, and Jilian Zhang. 2019. Feature selection for text classification: A review. _Multimedia Tools and Applications_ , 78(3):3797–3816. * Devlin et al. (2018) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. _arXiv preprint arXiv:1810.04805_. * Elssied et al. (2014) Nadir Omer Fadl Elssied, Othman Ibrahim, and Ahmed Hamza Osman. 2014. A novel feature selection based on one-way anova f-test for e-mail spam classification. _Research Journal of Applied Sciences, Engineering and Technology_ , 7(3):625–638. * Forman (2003) George Forman. 2003. An extensive empirical study of feature selection metrics for text classification. _Journal of machine learning research_ , 3(Mar):1289–1305. * Grave et al. (2018) Edouard Grave, Piotr Bojanowski, Prakhar Gupta, Armand Joulin, and Tomas Mikolov. 2018. Learning word vectors for 157 languages. _arXiv preprint arXiv:1802.06893_. * Günal (2012) Serkan Günal. 2012. Hybrid feature selection for text classification. _Turkish Journal of Electrical Engineering and Computer Science_ , 20(Sup. 2):1296–1311. * Han et al. (2018) Kai Han, Yunhe Wang, Chao Zhang, Chao Li, and Chao Xu. 2018. Autoencoder inspired unsupervised feature selection. In _2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_ , pages 2941–2945. IEEE. * Hochreiter and Schmidhuber (1997) Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. _Neural computation_ , 9(8):1735–1780. * Joulin et al. (2016) Armand Joulin, Edouard Grave, Piotr Bojanowski, and Tomas Mikolov. 2016. Bag of tricks for efficient text classification. _arXiv preprint arXiv:1607.01759_. * Kim (2014) Yoon Kim. 2014. Convolutional neural networks for sentence classification. _arXiv preprint arXiv:1408.5882_. * Kipf and Welling (2017) TN Kipf and M Welling. 2017. Semi-supervised classification with graph convolutional networks iclr. * Kudo (2018) Taku Kudo. 2018. Subword regularization: Improving neural network translation models with multiple subword candidates. _arXiv preprint arXiv:1804.10959_. * Kudo and Richardson (2018) Taku Kudo and John Richardson. 2018. Sentencepiece: A simple and language independent subword tokenizer and detokenizer for neural text processing. _arXiv preprint arXiv:1808.06226_. * Lam and Lee (1999) Savio LY Lam and Dik Lun Lee. 1999. Feature reduction for neural network based text categorization. In _Proceedings. 6th international conference on advanced systems for advanced applications_ , pages 195–202. IEEE. * Liu et al. (2016) Pengfei Liu, Xipeng Qiu, and Xuanjing Huang. 2016. Recurrent neural network for text classification with multi-task learning. _arXiv preprint arXiv:1605.05101_. * Maldonado and Weber (2009) Sebastián Maldonado and Richard Weber. 2009. A wrapper method for feature selection using support vector machines. _Information Sciences_ , 179(13):2208–2217. * Minaee et al. (2020) Shervin Minaee, Nal Kalchbrenner, Erik Cambria, Narjes Nikzad, Meysam Chenaghlu, and Jianfeng Gao. 2020. Deep learning based text classification: A comprehensive review. _arXiv preprint arXiv:2004.03705_. * Mirzaei et al. (2019) Ali Mirzaei, Vahid Pourahmadi, Mehran Soltani, and Hamid Sheikhzadeh. 2019. Deep feature selection using a teacher-student network. _Neurocomputing_. * Pedregosa et al. (2011) F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. 2011. Scikit-learn: Machine learning in Python. _Journal of Machine Learning Research_ , 12:2825–2830. * Quinlan (1986) J. Ross Quinlan. 1986. Induction of decision trees. _Machine learning_ , 1(1):81–106. * Rajaraman and Ullman (2011) Anand Rajaraman and Jeffrey David Ullman. 2011. _Mining of massive datasets_. Cambridge University Press. * Rogati and Yang (2002) Monica Rogati and Yiming Yang. 2002. High-performing feature selection for text classification. In _Proceedings of the Eleventh International Conference on Information and Knowledge Management_ , CIKM ’02, page 659–661, New York, NY, USA. Association for Computing Machinery. * Saleh and El-Sonbaty (2007) S. N. Saleh and Y. El-Sonbaty. 2007. A feature selection algorithm with redundancy reduction for text classification. In _2007 22nd international symposium on computer and information sciences_ , pages 1–6. * Sennrich et al. (2015) Rico Sennrich, Barry Haddow, and Alexandra Birch. 2015. Neural machine translation of rare words with subword units. _arXiv preprint arXiv:1508.07909_. * Shang et al. (2013) Changxing Shang, Min Li, Shengzhong Feng, Qingshan Jiang, and Jianping Fan. 2013\. Feature selection via maximizing global information gain for text classification. _Knowledge-Based Systems_ , 54:298–309. * Song et al. (2011) Qinbao Song, Jingjie Ni, and Guangtao Wang. 2011. A fast clustering-based feature subset selection algorithm for high-dimensional data. _IEEE transactions on knowledge and data engineering_ , 25(1):1–14. * Taira and Haruno (1999) Hirotoshi Taira and Masahiko Haruno. 1999. Feature selection in svm text categorization. In _AAAI/IAAI_ , pages 480–486. * Tang et al. (2019) Xiaochuan Tang, Yuanshun Dai, and Yanping Xiang. 2019. Feature selection based on feature interactions with application to text categorization. _Expert Systems with Applications_ , 120:207–216. * Yang and Pedersen (1997) Yiming Yang and Jan O Pedersen. 1997. A comparative study on feature selection in text categorization. In _Icml_ , volume 97, page 35. * Yang et al. (2016) Zichao Yang, Diyi Yang, Chris Dyer, Xiaodong He, Alex Smola, and Eduard Hovy. 2016\. Hierarchical attention networks for document classification. In _Proceedings of the 2016 conference of the North American chapter of the association for computational linguistics: human language technologies_ , pages 1480–1489. * Yao et al. (2019) Liang Yao, Chengsheng Mao, and Yuan Luo. 2019. Graph convolutional networks for text classification. In _Proceedings of the AAAI Conference on Artificial Intelligence_ , volume 33, pages 7370–7377. * Zhang et al. (2015) Xiang Zhang, Junbo Zhao, and Yann LeCun. 2015. Character-level convolutional networks for text classification. In _Advances in neural information processing systems_ , pages 649–657.
# Multilingual Pre-Trained Transformers and Convolutional NN Classification Models for Technical Domain Identification Suman Dowlagar LTRC IIIT-Hyderabad suman.dowlagar@ research.iiit.ac.in &Radhika Mamidi LTRC IIIT-Hyderabad radhika.mamidi@ iiit.ac.in ###### Abstract In this paper, we present a transfer learning system to perform technical domain identification on multilingual text data. We have submitted two runs, one uses the transformer model BERT, and the other uses XLM-ROBERTa with the CNN model for text classification. These models allowed us to identify the domain of the given sentences for the ICON 2020 shared Task, TechDOfication: Technical Domain Identification. Our system ranked the best for the subtasks 1d, 1g for the given TechDOfication dataset. ## 1 Introduction Automated technical domain identification is a categorization/classification task where the given text is categorized into a set of predefined domains. It is employed in tasks like Machine Translation, Information Retrieval, Question Answering, Summarization, and so on. In Machine Translation, Summarization, Question Answering, and Information Retrieval, the domain classification model will help leverage the contents of technical documents, select the appropriate domain-dependent resources, and provide personalized processing of the given text. Technical domain identification comes under text classification or categorization. Text classification is one of the fundamental tasks in the field of NLP. Text classification is the process of identifying the category where the given text belongs. Automated text classification helps to organize unstructured data, which can help us gather insightful information to make future decisions on downstream tasks. Traditional text classification approaches mainly focus on feature engineering techniques such as bag-of-words and classification algorithms Yang (1999). Nowadays, the sate-of-the-art results on text classification are achieved by various NNs such as CNN Kim (2014), LSTM Hochreiter and Schmidhuber (1997), BERT Adhikari et al. (2019), and Text GCN Adhikari et al. (2019). Attention mechanisms Vaswani et al. (2017) have been introduced in these models, which increased the representativeness of the text for better classification. Transformer models such as BERT Devlin et al. (2018) uses the attention mechanism that learns contextual relations between words or sub-words in a text. Text GCN Yao et al. (2019) uses a graph-convolutional network to learn a heterogeneous word document graph on the whole corpus, which helped classify the text. However, of all the deep learning approaches, transformer models provided SOTA results in text classification. In this paper, We present two approaches for technical domain identification. One approach uses the pre-trained Multilingual BERT model, and the other uses XLM-ROBERTa with CNN model. The rest of the paper is structured as follows. Section 2 describes our approach in detail. In Section 3, we provide the analysis and evaluation of results for our system, and Section 4 concludes our work. ## 2 Our Approach Here we present two approaches for the TechDOfication task. ### 2.1 BERT for TechDOfication Figure 1: The architecture of the BERT model for sentence classification. In the first approach, we use the pre-trained multilingual BERT model for domain identification of the given text. Bidirectional Encoder Representations from Transformers (BERT) is a transformer encoder stack trained on the large corpora. Like the vanilla transformer model Vaswani et al. (2017), BERT takes a sequence of words as input. Each layer applies self-attention, passes its results through a feed-forward network, and then hands it off to the next encoder. The BERT configuration model takes a sequence of words/tokens at a maximum length of 512 and produces an encoded representation of dimensionality 768. The pre-trained multilingual BERT models have a better word representation as they are trained on a large multilingual Wikipedia and book corpus. As the pre-trained BERT model is trained on generic corpora, we need to finetune the model for the given domain identification tasks. During finetuning, the pre- trained BERT model parameters are updated. In this architecture, only the [CLS] (classification) token output provided by BERT is used. The [CLS] output is the output of the 12th transformer encoder with a dimensionality of 768. It is given as input to a fully connected neural network, and the softmax activation function is applied to the neural network to classify the given sentence. ### 2.2 XLM-ROBERTa with CNN for TechDOfication Figure 2: The architecture of the XLM-ROBERTa with CNN for sentence classification. XLM-ROBERTa Conneau et al. (2019) is a transformer-based multilingual masked language model pre-trained on the text in 100 languages, which obtains state- of-the-art performance on cross-lingual classification, sequence labeling, and question answering. XLM-ROBERTa improves upon BERT by adding a few changes to the BERT model such as training on a larger dataset, dynamically masking out tokens compared to the original static masking, and uses a known pre- processing technique (Byte-Pair-Encoding) and a dual-language training mechanism with BERT in order to learn better relations between words in different languages. The given model is trained for the language modeling task, and the output is of dimensionality 768. It is given as input to a CNN Kim (2014) because convolution layers can extract better data representations than Feed Forward layers, which indirectly helps in better domain identification. ## 3 Experiment This section presents the datasets used, the task description, and two models’ performance on technical domain identification. We also include our implementation details and error analysis in the subsequent sections. ### 3.1 Dataset We used the dataset provided by the organizers of TechDOfication ICON-2020. There are two subtasks, one is coarse-grained, and the other is fine-grained. The coarse-grained TechDOfication dataset contains sentences about Chemistry, Communication Technology, Computer Science, Law, Math, and Physics domains in different languages such as English, Bengali, Gujarati, Hindi, Malayalam, Marathi, Tamil, and Telugu. Whereas the fine-grained English dataset focuses on the Computer-Science domain with sub-domain labels as Artificial Intelligence, Algorithm, Computer Architecture, Computer Networks, Database Management system, Programming, and Software Engineering. ### 3.2 Implementation For the implementation, we used the transformers library provided by HuggingFace111https://huggingface.co/. The HuggingFace contains the pre- trained multilingual BERT, XLM-ROBERTa, and other models suitable for downstream tasks. The pre-trained multilingual BERT model used is _“bert-base- multilingual-cased”_ and pre-trained XLM-R model used is _“xlm-roberta-base”_. We programmed the CNN architecture as given in the paper Kim (2014). We used the PyTorch library that supports GPU processing for implementing deep neural nets. The BERT models were run on the Google Colab and Kaggle GPU notebooks. We trained our classifier with a batch size of 128 for 10 to 30 epochs based on our experiments. The dropout is set to 0.1, and the Adam optimizer is used with a learning rate of 2e-5. We used the hugging face transformers pre- trained BERT tokenizer for tokenization. We used the BertForSequenceClassification module provided by the HuggingFace library during finetuning and sequence classification for the multilingual-BERT based approach. ### 3.3 Baseline models Here, we compared the BERT model with other machine learning algorithms. #### SVM with TF_IDF text representation We chose Support Vector Machines (SVM) with TF_IDF text representation for technical domain identification. SVM classifier and TF_IDF vector representation is obtained from the scikit-learn library Pedregosa et al. (2011). #### CNN: Convolutional Neural Network Kim (2014). We explored CNN-non-static, which uses pre-trained word embeddings. ### 3.4 Results The results are tabulated in Table 1. We evaluated the performance of the method using macro F1. The multilingual-BERT model performed well when compared to the other SVM with TF-IDF and CNN models. Given all the languages, we have observed an increase of 7 to 25% in classification metrics for BERT compared to the baseline SVM classifier, it showed a 2 to 5% increase in classification metrics compared to the CNN classifier on the validation data. On the test data, multilingual BERT showed better performance in subtasks 1a, 1b, 1c, 1h and 2a whereas XLM-ROBERTa with CNN showed better performance in the subtasks 1d, 1e, 1f, 1g. This increase in classification metrics is due to the transformer model’s and convolutional NN’s capability, which learned better text representations from the generic data than other models. | Classifier Models ---|--- Dataset | Validation | Test | SVM | CNN | M-Bert | XLM-R+CNN | M-Bert | XLM-R+CNN English subtask-1a | 81.48 | 83.05 | 88.87 | 87.09 | 79.84 | 73.57 Bengali subtask-1b | 66.35 | 85.78 | 86.81 | 85.71 | 80.35 | 78.17 Gujarati subtask-1c | 69.63 | 86.27 | 87.21 | 86.89 | 68.67 | 66.73 Hindi subtask-1d | 58.21 | 81.03 | 83.40 | 82.13 | 59.89 | 60.44 Malayalam subtask-1e | 80.60 | 92.51 | 94.72 | 93.40 | 34.47 | 34.86 Marathi subtask-1f | 73.32 | 86.89 | 87.42 | 86.37 | 59.52 | 59.89 Tamil subtask-1g | 65.95 | 85.75 | 87.50 | 86.54 | 49.24 | 51.34 Telugu subtask-1h | 71.98 | 88.07 | 90.28 | 89.43 | 67.17 | 62.26 English subtask-2a | 70.24 | 72.53 | 77.36 | 76.77 | 78.98 | 78.07 Table 1: macro F1 on validation and test data for all the subtasks ## 4 Error Analysis The multilingual-BERT model’s confusion matrix is compared with the poorly performed model for languages, Hindi, and Tamil languages are shown in Figure 3. We chose Hindi and Tamil languages because, here, the difference in performance is more significant. For the Hindi subtask, the SVM classifier confused between “cse”, “com_tech”, and “mgmt” labels, whereas the BERT model performed better. For the Tamil subtask, the SVM classifier confused between “com_tech” and “mgmt” labels, whereas the BERT model performed better than the other models. This is because both the approaches (pre-trained multilingual- BERT and pre-trained XLM-ROBERTa with CNN) learned better representation of the above data than the other models that helped in technical document identification. (a) (b) (c) (d) Figure 3: Confusion matrix on the given validation data for the Hindi and Tamil languages ## 5 Conclusion and Future work We used pre-trained bi-directional encoder representations using multilingual- BERT and XLM-ROBERTa with CNN technical domain identification for English, Bengali, Gujarati, Hindi, Malayalam, Marathi, Tamil, and Telugu languages. We compared the approaches with the baseline methods. Our analysis showed that pre-trained multilingual BERT and XLM-ROBERTa with CNN models and finetuning it for text classification tasks showed an increase in macro F1 score and accuracy metrics compared to baseline approaches. Some datasets are large, like for the Hindi, Tamil, and Telugu, we can train the BERT and XLM-ROBERTa models from scratch and consider its hidden layer representation, and concatenate this with the representation of the pre- trained model. It might help to classify the datasets even better. ## References * Adhikari et al. (2019) Ashutosh Adhikari, Achyudh Ram, Raphael Tang, and Jimmy Lin. 2019. Docbert: Bert for document classification. _arXiv preprint arXiv:1904.08398_. * Conneau et al. (2019) Alexis Conneau, Kartikay Khandelwal, Naman Goyal, Vishrav Chaudhary, Guillaume Wenzek, Francisco Guzmán, Edouard Grave, Myle Ott, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Unsupervised cross-lingual representation learning at scale. _arXiv preprint arXiv:1911.02116_. * Devlin et al. (2018) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. _arXiv preprint arXiv:1810.04805_. * Hochreiter and Schmidhuber (1997) Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. _Neural computation_ , 9(8):1735–1780. * Kim (2014) Yoon Kim. 2014. Convolutional neural networks for sentence classification. _arXiv preprint arXiv:1408.5882_. * Pedregosa et al. (2011) F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. 2011. Scikit-learn: Machine learning in Python. _Journal of Machine Learning Research_ , 12:2825–2830. * Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In _Advances in neural information processing systems_ , pages 5998–6008. * Yang (1999) Yiming Yang. 1999. An evaluation of statistical approaches to text categorization. _Information retrieval_ , 1(1-2):69–90. * Yao et al. (2019) Liang Yao, Chengsheng Mao, and Yuan Luo. 2019. Graph convolutional networks for text classification. In _Proceedings of the AAAI Conference on Artificial Intelligence_ , volume 33, pages 7370–7377.
# Personal Fixations-Based Object Segmentation with Object Localization and Boundary Preservation Gongyang Li, Zhi Liu, , Ran Shi, Zheng Hu, Weijie Wei, Yong Wu, Mengke Huang, and Haibin Ling Gongyang Li, Zhi Liu, Zheng Hu, Weijie Wei, Yong Wu, and Mengke Huang are with Shanghai Institute for Advanced Communication and Data Science, Shanghai University, Shanghai 200444, China, and School of Communication and Information Engineering, Shanghai University, Shanghai 200444, China (email<EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS>huangmengke@shu.edu.cn).Ran Shi is with School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210094, China (email: rshi@njust.edu.cn).Haibin Ling is with the Department of Computer Science, Stony Brook University, Stony Brook, NY 11794, USA (email: hling@cs.stonybrook.edu).Corresponding author: Zhi Liu ###### Abstract As a natural way for human-computer interaction, fixation provides a promising solution for interactive image segmentation. In this paper, we focus on Personal Fixations-based Object Segmentation (PFOS) to address issues in previous studies, such as the lack of appropriate dataset and the ambiguity in fixations-based interaction. In particular, we first construct a new PFOS dataset by carefully collecting pixel-level binary annotation data over an existing fixation prediction dataset, such dataset is expected to greatly facilitate the study along the line. Then, considering characteristics of personal fixations, we propose a novel network based on Object Localization and Boundary Preservation (OLBP) to segment the gazed objects. Specifically, the OLBP network utilizes an Object Localization Module (OLM) to analyze personal fixations and locates the gazed objects based on the interpretation. Then, a Boundary Preservation Module (BPM) is designed to introduce additional boundary information to guard the completeness of the gazed objects. Moreover, OLBP is organized in the mixed bottom-up and top-down manner with multiple types of deep supervision. Extensive experiments on the constructed PFOS dataset show the superiority of the proposed OLBP network over 17 state-of- the-art methods, and demonstrate the effectiveness of the proposed OLM and BPM components. The constructed PFOS dataset and the proposed OLBP network are available at https://github.com/MathLee/OLBPNet4PFOS. ###### Index Terms: Personal fixations, interactive image segmentation, object localization, boundary preservation. ## I Introduction Fixation is a flexible interaction mechanism of the human visual system. Compared with scribble, click and bounding box, fixation provides the most convenient interaction for patients with hand disability, amyotrophic lateral sclerosis (ALS) and polio. This kind of eye control interaction, i.e. fixation, can greatly improve the interaction efficiency of these patients. In addition, fixation is closely related to personal information such as age [1, 2] and gender [3, 4]. This means that different individuals may have different perceptions and preferences of a scene [5, 6]. Thus motivated, in this paper, we pay close attention to personal fixations-based object segmentation, which is a more natural manner for interactive image segmentation. \begin{overpic}[width=433.62pt]{Figs/Fig1_background.png} \end{overpic} Figure 1: Examples of image with ambiguous fixations. Green dots in each image indicate fixations. Some fixations fall in the background. The typical manners of interaction, such as scribbles [7, 8, 9, 10, 11], clicks [12, 13, 14, 15, 16, 17] and bounding boxes [18, 19, 20, 21] for interactive image segmentation, are explicit behaviors without interference. By contrast, fixations are implicit [22, 23, 24, 25], and their convenience comes with interaction ambiguity. Concretely, the positive and negative labels of scribbles and clicks are deterministic. However, fixations are unlabeled when collected. They do not distinguish between positive labels and negative labels (i.e. some fixations may fall in the background as shown in Fig. 1), resulting in a few noise in the fixations. Such ambiguous interaction makes the fixations-based object segmentation task difficult. Recently, with the rise of convolutional neural networks (CNNs), the clicks-based interactive image segmentation has been greatly developed. Even though fixation points and clicking points are similar to some extent, clicks-based methods [12, 13, 14, 16, 17] cannot be directly applied to fixations-based object segmentation. The above observations suggest that there are two main reasons that limit the development of fixations-based object segmentation. First, there is not a suitable dataset for the fixations-based object segmentation task, let alone dataset based on the personal fixations. Second, as aforementioned, the ambiguous representation of fixations makes this type of interaction difficult to handle by other methods which are based on clicks and scribbles. To address the first crucial issue, we construct a Personal Fixations-based Object Segmentation (PFOS) dataset, which is extended from the fixation prediction dataset OSIE [26]. The PFOS dataset contains 700 images, and each image has 15 personal fixation maps collected from 15 subjects with corresponding pixel-level annotations of objects. To overcome the ambiguity of fixations, we propose an effective network based on Object Localization and Boundary Preservation (OLBP). The key idea of OLBP is to locate the gazed objects based on the analysis of fixations, and then the boundary information is introduced to guard the completeness of the gazed objects and to filter the background. In particular, the overall structure of OLBP network is a mixture of bottom-up and top-down architectures. To narrow the gap between fixations and objects, we propose the Object Localization Module (OLM) to analyze personal fixations in detail and grasp location information of the gazed objects of different individuals. Based on the interpretation of location information, OLM modulates CNN features of image in a bottom-up way. Moreover, considering that the object location information may involve confusing noise, we propose a Boundary Preservation Module (BPM) to exploit boundary information to enforce object completeness and filter the background of erroneous localization. BPM is integrated into the top-down prediction. Both OLMs and BPMs employ deep supervision to further improve the capabilities of feature representation. In this way, the scheme of object localization and boundary preservation is successfully applied to the bottom-up and top-down structure, and the proposed OLBP network greatly promotes the performance of the personal fixations-based object segmentation task. Experimental results on the challenging PFOS dataset demonstrate that OLBP outperforms 17 state-of-the-art methods under various evaluation metrics. The contributions of this work are summarized as follows: * • We construct a new dataset for Personal Fixations-based Object Segmentation (PFOS), which focuses on the natural interaction (i.e. fixation). This dataset contains free-view personal fixations without any constraints, expanding its applicability. We believe that the PFOS dataset will boost the research of fixations-based human-computer interaction. * • We propose a novel Object Localization and Boundary Preservation (OLBP) network to segment the gazed objects based on personal fixations. The OLBP network, equipped with the Object Localization Module and the Boundary Preservation Module, effectively overcomes the difficulties from ambiguous fixations. * • We conduct extensive experiments to evaluate our OLBP network and other state- of-the-art methods on the PFOS dataset. Comprehensive results demonstrate the superiority of our OLBP network, and also reveal the difficulties and challenges of the constructed PFOS dataset. The rest of the paper is organized as follows: Sec. II reviews related previous works. Then, we formulate the PFOS task in Sec. III. After that, in Sec. IV, we construct the PFOS dataset. Sec. V presents the proposed OLBP network in detail. In Sec. VI, we evaluate the performance of the proposed OLBP network and other methods on the constructed PFOS dataset. Finally, the conclusion is drawn in Sec. VII. ## II Related Work In this section, we first give an overview of previous works of interactive image segmentation in Sec. II-A. Then, we introduce related works of fixations-based object segmentation in Sec. II-B. Finally, we review some related works on boundary-aware segmentation in Sec. II-C. ### II-A Interactive Image Segmentation 1) Scribbles-based interactive image segmentation. Scribble is a traditional manner of interaction. Most of scribbles-based methods are built on graph structures. GraphCut [7] is one of the most representative methods. It uses the max-flow/min-cut theorem to minimize energy function with hard constraints (i.e. labeled scribbles) and soft constraints. Grady et al. [8] adopted the random walk algorithm to assign a label to each unlabeled pixel based on the predefined seed pixels in discrete space. In [9], Bai et al. proposed a weighted geodesic distance based framework, which is fast for image and video segmentation and matting. Nguyen et al. [10] proposed a convex active contour model to segment objects, and their results were with smooth and accurate boundary contour. Spina et al. [11] presented a live markers methodology to reduce the user intervention for effective segmentation of target objects. Following the seed propagation strategy, Jian et al. [27] employed the adaptive constraint propagation to adaptively propagate the scribbles information into the whole image. Recently, Wang et al. [28] changed their view on interactive image segmentation and formulated it as a probabilistic estimation problem, proposing a pairwise likelihood learning based framework. These methods are friendly to clearly defined scribbles, but they cannot solve the ambiguity of fixations and their inference speed is usually slow. 2) Clicks-based interactive image segmentation. Click is a classical manner of interaction. It has been deeply studied in the deep learning era. The positive and negative clicks are transformed into two separate Euclidean distance maps for network input. Xu et al. [12] directly sent RGB image and two distance maps into a fully convolutional network. Liew et al. [13] proposed a two- branch fusion network with global prediction and local regional refinement. In addition to the RGB image and distance maps, Li et al. [14] included clicks in their network input and proposed an end-to-end segmentation-selection network. In [16], Jang et al. introduced the backpropagating refinement scheme to correct mislabeled locations in the initial segmentation map. Different from the direct concatenation of RGB image and interaction maps of the above methods, Hu et al. [17] separately input RGB image and interaction maps into two networks, and designed a fusion network for feature interactions. CNNs have greatly improved the performance of clicks-based interactive image segmentation, but when these methods are applied to fixations-based object segmentation, some background regions will be mistakenly segmented. To address the problem of erroneous localization, we explore the boundary information in our BPM to filter redundant background regions and guard the gazed object. 3) Bounding boxes-based interactive image segmentation. In a bounding box, the target object and background coexist, which is different from scribble and click. Rother et al. [18] extended the graph-cut approach, and segmented object with a rectangle, namely GrabCut. To overcome the looseness of the bounding box, Lempitsky et al. [19] incorporated the tightness prior into the global energy minimization function as hard constraints to further completed target object. Shi et al. [21] proposed a coarse-to-fine method with region- level and pixel-level segmentation. Similar to [12], Xu et al. [20] transformed the bounding box to a distance map and concatenated it with the RGB image to input into an encoder-decoder network. Although bounding box and fixation are similar (i.e. target object and background coexist in both interactions), the bounding box-based methods are difficult to transfer to fixations-based object segmentation. ### II-B Fixations-Based Object Segmentation Fixation plays an integral role in the human visual system and it is convenient for interaction. In an early study, Sadeghi et al. [29] constructed an eyegaze-based interactive segmentation system which adopts random walker to segment objects. Meanwhile, Mishra et al. [22] gave the definition of fixations-based object segmentation, that is, segmenting regions containing fixation points. They transformed the image to polar coordinate system, and found the optimal contour to fit the target object. Based on the interpretation of visual receptive field, Kootstra et al. [30] used symmetry to select fixations closer to the center of the object to obtain more complete segmentation. Differently, Li et al. [23] focused on selecting the most salient objects, and they ranked object proposals based on fixations. Similar to [23], Shi et al. [24] analyzed the fixation distribution and proposed three metrics to evaluate the score of each candidate region. In [31], Tian et al. first determined the uninterested regions, and then used superpixel-based random walk model to segment the gazed objects. Khosravan et al. [32] integrated fixations into the medical image segmentation and proposed a Gaze2Segment system. Li et al. [25] constructed a dataset where all fixations fall in objects (i.e. constrained fixations), and proposed a CNN-based model to simulate the human visual system to segment objects based on fixations. These studies have promoted the development of fixations-based object segmentation. However, all the fixations in [22, 30, 31, 25] fall in objects, which are hardly guaranteed in practice. These methods [22, 30, 31, 25] will get stuck in the ambiguity of unconstrained fixations, especially of personal fixations. For [23, 24], they are based on region proposal and cannot obtain accurate results. In summary, the above methods cannot solve the problem of ambiguous fixations, as shown in Fig. 1. In this paper, we take advantage of CNNs, and propose a bottom-up and top-down network to locate objects and preserve objects’ boundaries. Moreover, we construct a dataset to promote this special direction of interactive image segmentation, i.e. personal fixations- based object segmentation. ### II-C Boundary-Aware Segmentation The boundary/edge-aware segmentation idea is widely-used in salient object detection [33, 34, 35, 36] and semantic segmentation [37]. In [33], Wang et al. modeled the boundary information as an edge-preserving constraint, and included it as an additional supervision in loss function. In [34], Wang et al. proposed a two-branch network, including boundary and mask sub-networks, for jointly predicting masks of salient objects and detecting object boundaries. In [35], Wu et al. explored the logical interrelations between binary segmentation and edge maps in a multi-task network, and proposed a cross refinement unit in which the segmentation features and edge features are fused in a cross-task manner. In [36], Zhao et al. focused on the complementarity between salient edge information and salient object information. They integrated the local edge information of shallow layers and global location information of deep layers to obtain the salient edge features, and then the edge features were fed to the one-to-one guidance module to fuse the complementary region and edge information. In [37], Ding et al. first introduced the boundary information as an additional semantic class to enable the network to be aware of the boundary layout, and then proposed a boundary-aware feature propagation network to control the feature propagation based on the learned boundary information. In our method, we use the boundary information in two aspects: the multi-task structure (i.e. segmentation and boundary predictions) and the Boundary Preservation Module. Different from [34, 35], we integrate the learned boundary map into the prediction network in BPMs to preserve the completeness of the gazed objects, rather than fuse the segmentation features and boundary features. Compared with [36], our segmentation prediction is accompanied by the boundary prediction in a uniform prediction network, and the boundary supervision is employed at multiple scales. Different from [37], which uses the boundary map to control the region of feature propagation, our method uses the boundary map to filter the background of erroneous localization in features. In short, our use of boundary information is diverse and in-depth, which is suitable for the personal fixations-based object segmentation task. ## III Personal Fixations-Based Object Segmentation Problem Statement. Given an image ${\bf I}$ and a fixation map ${\bf FM}$ of a person, personal fixations-based object segmentation aims to segment the gazed objects of this person according to his/her personal ${\bf FM}$, producing a binary segmentation map. In general, different individuals generate different fixation maps when observing the same image, which means that individuals may be interested in different objects. In other words, segmentation results of different individuals on the same image vary with the observer. So, the special characteristic of this task is that an image has multiple binary segmentation maps due to multiple fixation maps. Although the ambiguity of fixations makes this task difficulty, the personal fixation map is the only information that can determine the gazed objects. Applications. This task has several meaningful applications. First, such a convenient manner of interaction is conducive to the development of special eye-control devices for patients with hand disability, ALS and polio, facilitating their lives and improving their quality of life. Second, fixation is advantageous to diagnose certain mental illnesses, such as autism spectrum disorder (ASD) [38, 39] and schizophrenia spectrum disorders (SSD) [40, 41]. This task understands personal fixations at the object level, which is helpful to improve the accuracy of disease diagnosis. For example, patients with ASD prefer to pay attention to background rather than foreground, so the proportion of foreground in their segmentation results will be less than that of healthy people. TABLE I: Categories of fixation map (FM) in the PFOS dataset. Constrained FM means that all fixations fall in the objects/foreground. Unconstrained FM represents that some fixations fall in the background. PFOS dataset | Constrained FM | Unconstrained FM ---|---|--- 10,500 | 3,683 (35.1%) | 6,817 (64.9%) ## IV Dataset Construction and Transformation Currently, there are many prevalently used datasets for fixation prediction, such as MIT1003 [42], OSIE [26] and SALICON [43], and for interactive image segmentation, such as GrabCut [18], Berkeley [44] and PASCAL VOC [45]. However, there is no dataset for the personal fixations-based object segmentation task. Considering that it is time-consuming for dataset annotations, we propose a convenient way to collect suitable data from existing datasets for this task. Obviously, the PFOS dataset must contain fixation data and pixel-level annotations for objects. Among the existing datasets, some datasets, such as DUTS-OMRON [46], PASCAL-S [23] and OSIE [26], are potential candidates. The pixel-level annotations of DUTS-OMRON and PASCAL-S are for salient object detection [47, 48, 49], that is, these annotations only focus on the most visually attractive objects but ignore other objects, which could be fixated by different individuals, in a scene. Therefore, they are not perfect for constructing a PFOS dataset. Fortunately, the pixel-level annotations of OSIE have semantic attributes. This means that we can select objects, which the user is interested in, based on personal fixations. In other words, we can create the pixel-level binary ground truths (GTs) for personal fixations-based object segmentation. So, we transform the fixation prediction dataset OSIE to our PFOS dataset. For each image in the OSIE dataset, it has corresponding fixation maps and semantic GTs of different subjects. The detailed steps for dataset transformation are as follows: 1) Semantic labels collection. We get the position of each fixation point from the fixation map, and we collect the semantic label of each position in the corresponding semantic GT. 2) Semantic labels distillation. As mentioned in Sec. I and shown in Fig. 1, some fixation points fall in the background or the same object. For semantic labels collected from Step 1, we discard the semantic label “0” which indicates background. Then, if there are several same semantic labels, we keep only one. 3) Binary GT creation. Based on the distilled semantic labels from Step 2, we can determine the gazed objects and create the binary GT. We reserve the regions with the distilled semantic labels in the semantic GT, and set them as foreground. We set the regions with the other unrelated semantic labels as background. In this convenient way, we efficiently create the binary GTs and successfully construct the PFOS dataset. The PFOS dataset retains all 700 images and 10,500 free-view personal fixation maps from the OSIE dataset. In the PFOS dataset, the image resolution is $800\times 600$. Each image has 15 personal fixation maps from 15 subjects and the transformed binary GTs. In the constructed PFOS dataset, there are two categories of fixation maps. The first category is that all fixations fall in the objects/foreground, i.e. the constrained fixation map in [25]. The second category is that some fixations fall in the background, namely the unconstrained fixation map. We present the details of them in Tab. I. In our PFOS dataset, the unconstrained fixation maps account for 64.9% and the constrained fixation maps hold 35.1%. The large proportion of unconstrained fixation maps increase the ambiguity of our PFOS dataset and make this dataset challenging. \begin{overpic}[width=432.75322pt]{Figs/Fig2_example.pdf} \put(4.5,0.0){ Image with fixations } \put(45.5,0.0){ FDM } \put(79.5,0.0){ GT } \put(-1.0,57.0){ \begin{sideways}{Subject A}\end{sideways} } \put(-1.0,33.2){ \begin{sideways}{Subject B}\end{sideways} } \put(-1.0,8.5){ \begin{sideways}{Subject C}\end{sideways} } \end{overpic} Figure 2: Examples of the PFOS dataset. Green dots in each image indicate fixations, FDM is fixation density map, and GT represents ground truth. ## V Methodology In this section, we first conduct data preprocessing which transforms the fixation points into fixation density maps in Sec. V-A. Then, we present the overview and motivation of the proposed Object Localization and Boundary Preservation (OLBP) network in Sec. V-B. Next, we give the detailed formulas of the Object Localization Module (OLM) and the Boundary Preservation Module (BPM) in Sec. V-C and Sec. V-D, respectively. Finally, we clarify the implementation details of OLBP network in Sec. V-E. Figure 3: The overall architecture of the proposed OLBP network. OLBP network is organized in the mixed bottom-up and top-down manner. We employ the modified VGG-16 to extract five blocks of features from an input image. Then in each OLM, FDM is analyzed by several dilated and normal convolutional layers to determine the location of objects in the corresponding block features. Based on the object localization in each feature block, the top-down prediction is established. During the prediction process, the boundary information is introduced into BPMs to guard the completeness of objects and to filter background of erroneous localization. We also construct a multi-task prediction structure, which contains object segmentation branch and boundary prediction branch, to exploit the complementarity between regions and boundaries. ### V-A Data Preprocessing The fixation points in each fixation map are sparse. With only a few pixels per fixation map, there is too little valuable information to supply. The similar problem arises in the clicks-based interactive image segmentation. Xu et al. [12] transformed the clicks into Euclidean distance maps. Inspired by this, we employ the Gaussian blur to transform the sparse fixation map (i.e. FM) into the fixation density map (i.e. FDM): $\mathbf{FDM}={\mathrm{nor}_{\mathrm{min-max}}}(\mathbf{FM}\circledast G_{\sigma}(x,y;\sigma)),$ (1) where ${\mathrm{nor}_{\mathrm{min-max}}}(\cdot)$ is the min-max normalization, $\circledast$ denotes convolution operator, and $G_{\sigma}(\cdot)$ is a Gaussian filter with parameter $\sigma$ which is the standard deviation. $\sigma$ is set corresponding to 1∘ visual angle in the OSIE dataset [26]. It is 24 pixels of an $800\times 600$ image by default. The effect of Gaussian blur is similar to the receptive field of eye, that is, the center of fixation is with a high resolution and the surrounding of fixation is with a low resolution. Thus, after performing Gaussian blur and linear transformation on FM, the dense FDM contains more prior information of objects. In this paper, we adopt the dense FDM rather than the raw FM. We present an image with the personal fixations of three subjects of the PFOS dataset in Fig. 2. The fixation maps of Subject A and Subject B are constrained fixation maps, while the fixation map of Subject C is an unconstrained fixation map. ### V-B Network Overview and Motivation The proposed OLBP network has three critical components: the feature extractor, the object locator and the prediction network with boundary preservation. The overall architecture of OLBP network is illustrated in Fig. 3. Feature Extractor. In the OLBP network, we adopt the modified VGG-16 [50], from which the last three fully connected layers have been removed, as the feature extractor. We denote its input image as ${\bf I}\\!\in\\!\mathbb{R}^{H\\!\times\\!W\\!\times\\!C}$, and initialize its parameters by the image classification model [50]. The feature extractor has five convolutional blocks, as shown in Fig. 3. We operate on the feature map of the last convolutional layer in each block, i.e. conv1-2, conv2-2, conv3-3, conv4-3 and conv5-3, which are denoted as $\\{{\bf F}^{(i)}_{r}$: ${\bf F}^{(i)}_{r}\\!\in\\!\mathbb{R}^{\textit{h}_{i}\\!\times\\!\textit{w}_{i}\\!\times\\!\textit{c}_{i}},i=1,2,...,5\\}$. Notably, the feature resolution at the i-th block, i.e. $[\textit{h}_{i},\textit{w}_{i}]$, is $[\frac{\textit{H}}{2^{i-1}},\frac{\textit{W}}{2^{i-1}}]$ and $\textit{c}_{i\in\\{1,2,3,4,5\\}}=\\{64,128,256,512,512\\}$. In reality, the input resolution $[\textit{H},\textit{W},\textit{C}]$ of ${\bf I}$ is set to $288\times 288\times 3$. Object Localization Module. Although FDM is a probability map, it is a critical interaction that reflects the intention of the user. It is important to effectively explore the object location information of FDM. However, when we construct a CNN-based model for the personal fixations-based object segmentation task, it is natural to directly concatenate FDM and the input image for the network input. Since there are three channels for image and only one channel for FDM, the direct concatenation operation may drown out the critical interaction information of FDM. Based on the above analysis, we propose the Object Localization Module to process FDM. The parallel convolution structure is effective to explore meaningful information in CNN features [51], especially with the dilated convolution [52]. Thus, in OLM, we employ several parallel dilated convolutions with different dilation rates to profoundly analyze the personal FDM to obtain object location information, which are a group of response maps. These response maps belong to ${[0,1]}^{\textit{h}_{i}\\!\times\\!\textit{w}_{i}\\!\times\\!\textit{c}_{i}}$, which shows they have the same number of channels as the features of image at the i-th block. They are applied to re-weight features of image to highlight the gazed objects at channel-wise and spatial-wise. To enhance the location presentation of the response maps, we apply deep supervision [53] in OLM. As presented in Fig. 3, the OLM is performed in a bottom-up manner, and it is assembled after each block of feature extractor for strong object localization. The detailed description of OLM is presented in Sec. V-C. We show the ablation study of OLM in Sec. VI-C, including a variant of direct concatenation of image and FDM. Boundary Preservation Module and Prediction Network. Since some fixations fall in the background, there may be some noise on the re-weighted feature of OLM. The ambiguity over the fixations causes great disturbance to the segmentation result. Fortunately, there is a priori knowledge that the background usually does not have a regular boundary. Thus, we introduce the boundary information into the prediction network, and propose the Boundary Preservation Module to filter the background of erroneous localization and preserve the completeness of the gazed objects. BPM is a momentous component to purify the segmentation result. We also attach the pixel-level segmentation supervision and boundary supervision to BPM. As shown in Fig. 3, BPMs are equipped between convolutional blocks in the prediction network from top to down. To make full use of the boundary information, we also construct a multi-task structure in the prediction network. We elaborate the formulation and ablation study of BPM in Sec. V-D and Sec. VI-C, respectively. TABLE II: Detailed parameters of each OLM. We present the kernel size and channel number of each dilated/normal convolutions. Besides, we also present the dilation rates and the size of output feature. For instance, $(3\times 3,32)$ denotes that the kernel size is $3\times 3$ and the channel number is 32. Aspects | | Dilation --- conv | Dilation --- rate 2$\times$Conv | Output size OLM-1 | $(3\times 3,32)$ | $1/3/5/7$ | $(7\times 7,128)$ | $[288\times 288\times 128]$ OLM-2 | $(3\times 3,64)$ | $1/3/5/7$ | $(5\times 5,256)$ | $[144\times 144\times 256]$ OLM-3 | $(3\times 3,128)$ | $1/3/5/7$ | $(5\times 5,512)$ | $[72\times 72\times 512]$ OLM-4 | $(3\times 3,256)$ | $1/2/3/4$ | $(3\times 3,1024)$ | $[36\times 36\times 1024]$ OLM-5 | $(3\times 3,256)$ | $1/2/3/4$ | $(3\times 3,1024)$ | $[18\times 18\times 1024]$ ### V-C Object Localization Module As the OLM-5 shown in Fig. 3, there are three main parts in the Object Localization Module: location analysis unit, feature re-weighting (i.e. Re- wei) and segmentation supervision (i.e. Seg sup). Its objective is to extract object location information of personal FDM and to highlight objects in feature of image ${\bf F}^{(i)}_{r}$. OLM is the most indispensable part of the whole OLBP network. Concretely, in OLM-i, the ${\bf FDM}\\!\in\\!\mathbb{R}^{H\\!\times\\!W\\!\times\\!1}$ is first downsampled to fit the resolution of ${\bf F}^{(i)}_{r}$ and to generate $\mathbf{F}^{(i)}_{fdm}\\!\in\\!\mathbb{R}^{\textit{h}_{i}\\!\times\\!\textit{w}_{i}\\!\times\\!1}$ which is formulated as: $\displaystyle\mathbf{F}^{(i)}_{fdm}=\mathrm{MaxPool}(\mathbf{FDM};{W}^{(i)}_{ks}),$ (2) where $\mathrm{MaxPool}(\cdot)$ is the max pooling with parameters ${W}^{(i)}_{ks}$, which are $2^{i-1}\times 2^{i-1}$ kernel with $2^{i-1}$ stride. Then, we design the location analysis unit, which contains four parallel dilated convolutions [52] with different dilation rates, to analyze $\mathbf{F}^{(i)}_{fdm}$, and obtain the multi-interpretation feature $\mathbf{F}^{(i)}_{mi}$. The process in this unit can be formulated as: $\displaystyle\mathbf{F}^{(i)}_{mi}=\mathrm{concat}\big{(}$ $\displaystyle C_{d}(\mathbf{F}^{(i)}_{fdm};{W}^{(i_{1})}_{d}),C_{d}(\mathbf{F}^{(i)}_{fdm};{W}^{(i_{2})}_{d}),$ (3) $\displaystyle C_{d}(\mathbf{F}^{(i)}_{fdm};{W}^{(i_{3})}_{d}),C_{d}(\mathbf{F}^{(i)}_{fdm};{W}^{(i_{4})}_{d})\big{)},$ where $\mathrm{concat}(\cdot)$ is the cross-channel concatenation, and $C_{d}(\cdot;{W}^{(i_{n})}_{d})$ is the dilated convolution with parameters ${W}^{(i_{n})}_{d}$ for $n\in\\{1,2,3,4\\}$. Notably, ${W}^{(i_{n})}_{d}$ are comprised of kernel size, channel number and dilation rate. Considering the resolution difference of each $\mathbf{F}^{(i)}_{r}$, the dilation rates of each unit are different and the details are presented in Tab. II. In this unit, the dilated convolutions large the receptive field without increasing the computation. They are performed in a parallel manner, which makes $\mathbf{F}^{(i)}_{mi}$ effectively capture the local and global location information of the gazed objects. The multi-scale features in $\mathbf{F}^{(i)}_{mi}$ are complementary to each other. They are blended to produce the location response maps $\mathbf{r}^{(i)}_{loc}\\!\in\\!{[0,1]}^{\textit{h}_{i}\\!\times\\!\textit{w}_{i}\\!\times\\!\textit{c}_{i}}$ via: $\displaystyle\mathbf{F}^{(i)}_{int}={2C}(\mathbf{F}^{(i)}_{mi};{W}^{(i)}_{2c}),$ (4) $\displaystyle\mathbf{r}^{(i)}_{loc}=\psi(C(\mathbf{F}^{(i)}_{int};{W}^{(i)}_{c})),$ (5) where $\mathbf{F}^{(i)}_{int}$ is the interim feature, $2C(\ast;{W}^{(i)}_{2c})$ are two convolutional layers with the same parameters ${W}^{(i)}_{2c}$, $\psi(\cdot)$ is the sigmoid function, and $C(\ast;{W}^{(i)}_{c})$ is the convolutional layer with parameters ${W}^{(i)}_{c}$ which are $3\times 3$ kernel with ${c}_{i}$ channels. ${W}^{(i)}_{2c}$ contain kernel size and channel number, which are different in different OLMs. Their details are shown in the column with “2$\times$Conv” of Tab. II. \begin{overpic}[width=433.62pt]{Figs/Fig4_enhanced_Feature.png} \put(5.9,-2.4){ Image } \put(25.9,-2.4){ FDM } \put(46.6,-2.4){ GT } \put(65.2,-3.1){ $\mathbf{r}^{(2)}_{loc}$ } \put(84.4,-3.1){ $\mathbf{F}^{(2)}_{loc}$} \end{overpic} Figure 4: Feature visualization in OLM-2. $\mathbf{r}^{(2)}_{loc}$ is the location response map, and $\mathbf{F}^{(2)}_{loc}$ is the location-enhanced feature. After completing the FDM interpretation in location analysis unit, we successfully obtain $\mathbf{r}^{(i)}_{loc}$, which are the protagonists of the feature re-weighting (i.e. Re-wei) part. We employ $\mathbf{r}^{(i)}_{loc}$ to re-weight $\mathbf{F}^{(i)}_{r}$ at channel-wise and spatial-wise, and receive the location-enhanced feature $\mathbf{F}^{(i)}_{loc}\\!\in\\!\mathbb{R}^{\textit{h}_{i}\\!\times\\!\textit{w}_{i}\\!\times\\!\textit{c}_{i}}$, which is computed as: $\displaystyle\mathbf{F}^{(i)}_{loc}=\mathbf{F}^{(i)}_{r}\otimes\mathbf{r}^{(i)}_{loc},$ (6) where $\otimes$ is element-wise multiplication. Besides, in Re-wei, to balance the information of image and location, we concatenate $\mathbf{F}^{(i)}_{r}$ to $\mathbf{F}^{(i)}_{loc}$ and obtain the output feature $\mathbf{F}^{(i)}_{olm}$ of OLM. The size of $\mathbf{F}^{(i)}_{olm}$ is shown in Tab. II. Notably, at the training phase, we apply the pixel-level segmentation supervision (i.e. Seg sup) to each OLM. In Fig. 4, we visualize feature in OLM-2 to verify the effectiveness of the location enhancement. Concretely, in OLM-2, the conv2-2 is re-weighted by the location response map. As shown in Fig. 4, the location response map $\mathbf{r}^{(2)}_{loc}$ contains rich location information of the gazed objects. After using Eq. 6 to perform the location enhancement operation on conv2-2, we observe that the gazed objects are highlighted in $\mathbf{F}^{(2)}_{loc}$ (with darker color). In summary, the location- enhanced feature $\mathbf{F}^{(i)}_{loc}$ of OLM has strong location expression ability and contributes to the subsequent segmentation prediction network. ### V-D Boundary Preservation Module The Boundary Preservation Module is built to restrain the falsely highlighted part of the re-weighted feature of OLM and to preserve the completeness of the gazed objects for the segmentation prediction. As the BPM-5 shown in Fig. 3, the structure of BPM is succinct, but it is a key bridge to connect convolutional blocks of the prediction network. Let $\\{{\bf F}^{(i)}_{p}$: ${\bf F}^{(i)}_{p}\\!\in\\!\mathbb{R}^{\textit{h}_{i-1}\\!\times\\!\textit{w}_{i-1}\\!\times\\!\textit{c}_{i-1}},i=2,3,4,5\\}$ denote the output feature of each deconvolutional layer in the prediction network. In BMP, $\mathbf{F}^{(i)}_{p}$ is processed by a convolutional layer to generate the boundary mask $\mathbf{B}^{(i)}$, which is defined as: $\displaystyle\mathbf{B}^{(i)}=C(\mathbf{F}^{(i)}_{p},{W}^{(i)}_{c}).$ (7) To increase the accuracy of $\mathbf{B}^{(i)}\\!_{i\in\\{2,3,4,5\\}}$, we introduce the pixel-level boundary supervision (i.e. “Bound sup” on BPM-5 in Fig. 3) in BPM. Since that there are no pixel-level boundary annotations in the PFOS dataset, we employ the morphological operation on binary segmentation GT $\mathbf{G}_{s}$ to produce the boundary GT $\mathbf{G}_{b}$, as follow: $\displaystyle\mathbf{G}_{b}=\mathrm{Dilate}(\mathbf{G}_{s};\theta)-\mathbf{G}_{s},$ (8) where $\mathrm{Dilate}(\ast;\theta)$ is the morphological dilation operation with dilation coefficient $\theta$ which is 2 pixels. Then, $\mathbf{B}^{(i)}$ is concatenated to $\mathbf{F}^{(i)}_{p}$ to generate the output feature $\mathbf{F}^{(i)}_{bpm}$ of BPM. We also put the pixel- level segmentation supervision behind $\mathbf{F}^{(i)}_{bpm}$, such as “Seg sup” on BPM-5 in Fig. 3. The segmentation supervision and the boundary supervision cooperate well with each other, improving the feature representation of the gazed objects. In this way, we novelly introduce boundary information into the BPM, and $\mathbf{F}^{(i)}_{bpm}$ carries the feature de-noising and boundary preservation capabilities into the prediction network. ### V-E Implementation Details TABLE III: Quantitative results including Jaccard index, S-measure, weighted F-measure, E-measure and F-measure on the PFOS dataset (in percentage %). Semantic Segmentation means semantic segmentation-based method. Clicks means clicks-based interactive image segmentation method. Fixations means fixations-based object segmentation method. FDM-Guided Semantic Segmentation means embedding FDM into semantic segmentation method. FDM-Guided Salient Object Detection means embedding FDM into salient object detection method. The best three results are shown in red, blue, and green. $\uparrow$ denotes larger is better. The subscript of each method represents the publication year. † means CNNs-based method. Aspects | Methods | PFOS Dataset ---|---|--- $\mathcal{J}\uparrow$ | $\mathcal{S}_{\lambda}\uparrow$ | $w\mathcal{F}_{\beta}\uparrow$ | $\mathcal{E}_{\xi}\uparrow$ | $\mathcal{F}_{\beta}\uparrow$ Semantic Segmentation | PSPNet17† [54] | 51.0 | 58.9 | 55.5 | 64.2 | 60.2 SegNet17† [55] | 58.7 | 70.4 | 66.6 | 78.4 | 72.5 DeepLab18† [51] | 52.8 | 65.7 | 60.5 | 72.9 | 66.9 EncNet18† [56] | 55.5 | 62.2 | 60.5 | 69.0 | 65.3 DeepLabV3+18† [57] | 45.6 | 61.4 | 53.1 | 67.8 | 59.3 HRNetV219† [58] | 46.1 | 50.7 | 49.0 | 53.8 | 53.2 Clicks | ISLD18† [14] | 61.2 | 73.4 | 71.2 | 82.5 | 77.9 FCTSFN19† [17] | 62.4 | 72.9 | 69.9 | 82.8 | 75.1 BRS19† [16] | 62.1 | 73.0 | 69.1 | 82.3 | 74.6 Fixations | AVS12 [22] | 40.9 | 56.0 | 48.7 | 65.1 | 56.6 SOS14 [23] | 42.6 | 57.5 | 51.4 | 67.8 | 60.0 GBOS17 [24] | 38.0 | 56.7 | 48.0 | 63.9 | 58.1 CFPS19† [25] | 70.5 | 78.9 | 76.7 | 87.4 | 81.3 FDM-Guided Semantic Segmentation | DeepLabV3+18† [57] | 71.0 | 79.5 | 78.3 | 87.6 | 83.2 HRNetV219† [58] | 58.8 | 71.3 | 68.6 | 80.4 | 75.7 FDM-Guided Salient Object Detection | CPD19† [59] | 69.2 | 78.4 | 76.4 | 86.2 | 81.7 GCPA20† [60] | 72.3 | 80.3 | 78.9 | 88.1 | 83.6 Personal Fixations | OLBP (Ours) | 73.7 | 81.1 | 80.0 | 88.7 | 84.3 Prediction Network. The prediction network is constructed in the top-down manner to gradually restore resolution. It consists of five convolutional blocks, four BPMs and four deconvolutional layers. A dropout layer [61] is placed before each deconvolutional layer to prevent the prediction network from overfitting. In addition, we attach the boundary prediction branch to the prediction network to assist the object segmentation branch. We initialize parameters of the prediction network by xavier method [62]. Overall Loss. As shown in Fig. 3, there are totally 15 losses in the OLBP network, including 10 segmentation losses and 5 boundary losses. The overall loss $\mathbb{L}$ can be divided into three parts: losses of multi-task prediction, losses on OLMs and losses on BPMs. $\mathbb{L}$ is calculated as: $\displaystyle\mathbb{L}=$ $\displaystyle[\mathcal{L}_{s}(\mathbf{S}^{(1)},\mathbf{G}_{s})+\mathcal{L}_{s}(\mathbf{B}^{(1)},\mathbf{G}_{b})]+\sum\limits_{i=1}^{5}\mathcal{L}_{s}(\mathbf{S}^{(i)}_{olm},\mathbf{G}_{s})$ (9) $\displaystyle+\sum\limits_{i=2}^{5}[\mathcal{L}_{s}(\mathbf{S}^{(i)}_{bpm},\mathbf{G}_{s})+\mathcal{L}_{s}(\mathbf{B}^{(i)},\mathbf{G}_{b})],$ where $\mathcal{L}_{s}(\cdot,\cdot)$ is the softmax loss, $\mathbf{S}^{(1)}$ is the predicted segmentation map, and $\mathbf{B}^{(1)}$ is the predicted boundary map. $\mathbf{S}^{(i)}_{olm}$ and $\mathbf{S}^{(i)}_{bpm}$ present the side output segmentation results in OLM and BPM, respectively. $\mathbf{B}^{(i)}\\!_{i\in\\{2,3,4,5\\}}$ is the boundary mask in BPM. Notably, for each softmax loss, we resize the resolutions of $\mathbf{G}_{s}$ and $\mathbf{G}_{b}$ to fit the resolutions of corresponding $\mathbf{S}^{(i)}_{olm}$, $\mathbf{S}^{(i)}_{bpm}$ and $\mathbf{B}^{(i)}$. Network Training. The PFOS dataset is separated into training set and testing set. The training set contains 600 images with 9,000 personal fixation maps, including 3,075 constrained fixation maps and 5,925 unconstrained fixation maps. The testing set consists of 100 images with 1,500 personal fixations, including 608 constrained fixation maps and 892 unconstrained fixation maps. The OLBP network is implemented on Caffe [63] and experimented using a NVIDIA Titan X GPU. The data of training set and testing set are resized to $288\times 288$ for training and inference. We adopt the standard stochastic gradient descent (SGD) method [64] to optimize our OLBP network for 30,000 iterations. The learning rate is set to $8\times 10^{-8}$, and it will be divided by 10 after 14,000 iterations. The dropout ratio, batch size, iteration size, momentum and weight decay are set to 0.5, 1, 8, 0.9 and 0.0001, respectively. ## VI Experiments In this section, we present comprehensive experiments on the proposed PFOS dataset. We introduce evaluation metrics in Sec. VI-A. In Sec. VI-B, we compare the proposed OLBP network with state-of-the-art methods. Then, we conduct ablation studies in Sec. VI-C and show some personal segmentation results in Sec. VI-D. Finally, we present some discussions on the connections between fixation-based object segmentation and salient object detection in Sec. VI-E. ### VI-A Evaluation Metrics We use five evaluation metrics, i.e. Jaccard index ($\mathcal{J}$), S-measure ($\mathcal{S}_{\lambda}$) [65], F-measure ($\mathcal{F}_{\beta}$), weighted F-measure ($w\mathcal{F}_{\beta}$) [66], and E-measure ($\mathcal{E}_{\xi}$) [67], to evaluate the performance of different methods. Jaccard Index $\mathcal{J}$. Jaccard index is also called intersection-over- union (IoU), which can compare similarities and differences between two binary maps. It is defined as: $\mathcal{J}=\frac{|\mathbf{S}\cap\mathbf{G}_{s}|}{|\mathbf{S}\cup\mathbf{G}_{s}|},$ (10) where $\mathbf{S}$ is the predicted segmentation map, and $\mathbf{G}_{s}$ is the binary segmentation GT. S-measure $\mathcal{S}_{\lambda}$. S-measure focuses on the structural similarity between the predicted segmentation map and the binary segmentation GT. It evaluates the structural similarity of region-aware ($S_{r}$) and object-aware ($S_{o}$) simultaneously. S-measure is defined as: $\displaystyle\mathcal{S}_{\lambda}=\lambda\ast S_{o}+(1-\lambda)\ast S_{r},$ (11) where $\lambda$ is set to 0.5 by default. F-measure $\mathcal{F}_{\beta}$. F-measure is a weighted harmonic mean of precision and recall, which considers precision and recall comprehensively. It is defined as: $\displaystyle\mathcal{F}_{\beta}=\frac{(1+\beta^{2})\times Precision\times Recall}{\beta^{2}\times Precision+Recall},$ (12) where $\beta^{2}$ is set to 0.3 following previous studies [47, 48]. Weighted F-measure $w\mathcal{F}_{\beta}$. Weighted F-measure has the ability to evaluate the non-binary and binary map. It focuses on evaluating the weights errors of predicted pixels according to their location and their neighborhood, which is formulated as: $\displaystyle w\mathcal{F}_{\beta}=\frac{(1+\beta^{2})\times Precision^{w}\times Recall^{w}}{\beta^{2}\times Precision^{w}+Recall^{w}},$ (13) where $\beta^{2}$ is set to 1 following previous studies [68, 69]. \begin{overpic}[width=888.9223pt]{Figs/Fig5_Visual_example.png} \put(1.9,-1.3){ Image} \put(10.4,-1.3){ GT } \put(17.3,-1.3){ {Ours} } \put(24.75,-1.3){ CFPS } \put(32.1,-1.3){ GBOS } \put(40.3,-1.3){ SOS } \put(48.0,-1.3){ AVS } \put(55.75,-1.3){ BRS } \put(62.0,-1.3){ FCTSFN } \put(70.65,-1.3){ ISLD } \put(77.8,-1.3){ EncNet } \put(85.2,-1.3){ Deeplab } \put(93.1,-1.3){ SegNet } \end{overpic} Figure 5: Visualization comparison to some representative methods on the PFOS dataset. Zoom-in for the best view. E-measure $\mathcal{E}_{\xi}$. E-measure is based on cognitive vision studies. It evaluates the local errors (i.e. pixel-level) and the global errors (i.e. image-level) together. We introduce it to provide a more comprehensive evaluation. It could be computed as: $\displaystyle\mathcal{E}_{\xi}=\frac{1}{W\times H}\sum\limits_{x=1}^{W}\sum\limits_{y=1}^{H}f\Big{(}\frac{2\varphi_{\mathbf{G}_{s}}\circ\varphi_{\mathbf{s}}}{\varphi_{\mathbf{G}_{s}}\circ\varphi_{\mathbf{G}_{s}}+\varphi_{\mathbf{s}}\circ\varphi_{\mathbf{s}}}\Big{)},$ (14) where $\varphi_{\mathbf{G}_{s}}$ and $\varphi_{\mathbf{s}}$ are distance bias matrices for binary segmentation GT and predicted segmentation map, respectively, $\circ$ is the Hadamard product, and $f(\cdot)$ is the quadratic form. ### VI-B Comparison with the State-of-the-arts Comparison Methods. We compare our OLBP network against three types of state- of-the-art methods, including semantic segmentation-based methods, clicks- based interactive image segmentation methods and fixations-based object segmentation methods. For a reasonable comparison of the first type of method, we follow [12, 25], which convert the segmentation problem into the selection problem. Concretely, we first apply semantic segmentation methods, i.e. PSPNet [54], SegNet [55], DeepLab [51], EncNet [56], DeepLabV3+ [57], and HRNetV2 [58], to image, and then use the fixations to select the gazed objects. The second type of method includes ISLD [14], FCTSFN [17], and BRS [16]. The last type of method includes AVS [22], SOS [23], GBOS [24] and CFPS [25]. For all the above compared methods, we use the implementations with recommend parameter settings for a fair comparison. In addition, we modify several semantic segmentation methods (i.e. DeepLabV3+ [57] and HRNetV2 [58]) and recent salient object detection methods (i.e. CPD [59] and GCPA [60]) by embedding FDM in them to guide object segmentation. Two types of comparison methods are thus generated, namely FDM-guided semantic segmentation and FDM-guided salient object detection, respectively. Specifically, for DeepLabV3+, we embed FDM into features (i.e. low-level features and features generated from the ASPP) to bridge the encoder and decoder; for HRNetV2, we embed FDM between the second stage and the third stage; for CPD, we embed FDM into two partial decoders; and, for GCPA, we embed FDM into four self refinement modules. We retrain these modified methods with the same training dataset as our method, and their parameters are adjusted for better convergence. Notably, we use the well-known OTSU method [70] to binarize the generated probability map of our method and other CNNs- based methods. TABLE IV: Robustness evaluation of our method and several representative methods, such as the modified GCPA [60], CFPS [25] and the modified CPD [59], on the test part of the PFOS dataset in terms of Jaccard Index. The best result of each row is shown in bold. Notably, “+15% noise” means an additional 15% increase in the number of unconstrained fixations of the total number of fixations in a fixation map. We add the noise (i.e. unconstrained fixations) at three levels, i.e. 15%, 30%, and 45%. | OLBP | GCPA20 | CFPS19 | CPD19 ---|---|---|---|--- Dataset | (Ours) | [60] | [25] | [59] PFOS | 73.7 | 72.3 | 70.5 | 69.2 +15% noise | 72.2 | 70.9 | 69.6 | 68.7 +30% noise | 71.3 | 70.1 | 69.2 | 68.4 +45% noise | 70.3 | 69.7 | 68.8 | 68.1 Quantitative Performance Evaluation. We evaluate our OLBP network and other 17 state-of-the-art methods on the PFOS dataset using above five evaluation metrics. The quantitative results are presented in Table III. Our OLBP network favorably outperforms all the compared methods in terms of different metrics. Concretely, compared with the best method CFPS [25] in fixations-based object segmentation methods, the performance of our method is improved by $3.2\%$, $2.2\%$ and $3.0\%$ in $\mathcal{J}$, $\mathcal{S}_{\lambda}$ and $w\mathcal{F}_{\beta}$, respectively. The performance of our method is $5.9\%$ better than FCTSFN [17] in $\mathcal{E}_{\xi}$, and is $6.4\%$ better than ISLD [14] in $\mathcal{F}_{\beta}$. Note that the performance of our method is far better than that of three traditional methods AVS [22], SOS [24] and GBOS [24]. We attribute the performance superiority of the proposed OLBP network to the scheme of object localization and boundary preservation. In addition, semantic segmentation-based methods get an average of $51.6\%$ in $\mathcal{J}$. This may be due to the fact that semantic segmentation methods cannot accurately segment all objects, resulting in the failure of the object selection process. Clicks-based interactive image segmentation methods achieve an average of $61.9\%$ in $\mathcal{J}$, while our OLBP network obtains $73.7\%$ in $\mathcal{J}$. This demonstrates that our method is more robust than clicks-based interactive image segmentation methods in adapting the ambiguity of fixations. Fixations-based object segmentation methods contain three traditional methods and one CNN-based method, obtaining an average of $48.0\%$ in $\mathcal{J}$. Specifically, we present the results of the FDM-guided semantic segmentation methods, including the modified DeepLabV3+ and HRNetV2, in Table III. The modified DeepLabV3+ achieves a promising performance, but does not exceed our OLBP network (e.g. 71.0% vs 73.7% in $\mathcal{J}$). Although the FDM guidance brings some advantages to HRNetV2, but the modified HRNetV2 still does not perform well. For the FDM-guided salient object detection, both modified CPD and GCPA perform well, though our OLBP still outperforms them (e.g. 4.5% and 1.4% better than the modified CPD and GCPA in $\mathcal{J}$, respectively). In summary, there is a large room for performance improvement on the proposed PFOS dataset, suggesting that the PFOS dataset is challenging to all compared methods including OLBP. TABLE V: Ablation analyses for the proposed OLBP network on the PFOS dataset (in percentage %). As can be observed, each component in OLBP network plays an important role and contributes to the performance. The best result in each column is bold. Baseline: encoder-decoder network, OLM: object localization module, and BPM: boundary preservation module. | Baseline | OLM | BPM | $\mathcal{J}\uparrow$ | $\mathcal{S}_{\lambda}\uparrow$ | $w\mathcal{F}_{\beta}\uparrow$ ---|---|---|---|---|---|--- 1 | ✓∗ | | | 67.2 | 75.9 | 72.2 2 | ✓∗ | | ✓ | 68.0 | 76.4 | 72.5 3 | ✓ | | | 70.7 | 78.3 | 75.0 4 | ✓ | ✓ | | 73.0 | 80.7 | 79.5 5 | ✓ | | ✓ | 71.4 | 78.7 | 75.6 6 | ✓ | ✓ | ✓ | 73.7 | 81.1 | 80.0 ✓∗ means the image and FDM are concatenated. ✓means the image and FDM are fed to network separately. \begin{overpic}[width=429.28616pt]{Figs/Fig6_Ablation.pdf} \put(4.0,-1.0){ Image } \put(21.9,-1.0){ GT } \put(38.5,-1.0){ Ba${}^{*}$ } \put(51.1,-1.0){ Ba+OLM } \put(66.65,-1.0){ Ba${}^{*}$+BPM } \put(87.0,-1.0){ Ours } \end{overpic} Figure 6: Visual comparisons of different variants. “Ba∗” is the baseline network, whose input is the concatenated image and FDM. Qualitative Performance Evaluation. In Fig. 5, we show some representative visualization results of our OLBP network and other methods. Obviously, the visual segmentation maps of three traditional methods GBOS [24], SOS [24] and AVS [22] are rough. However, the CNN-based method CFPS [25], which belongs to the same type as GBOS, SOS and AVS, basically captures the gazed objects and brings in less background regions. The gazed objects in the segmentation results of clicks-based interactive image segmentation methods BRS [16], FCTSFN [17], and ISLD [14] are partially segmented and the details are relatively coarse. As for the EncNet [56], DeepLab [51] and SegNet [55], the object segmentation maps of them depend on the semantic segmentation results, which are great uncertainty. This results in their object segmentation maps that are sometimes accurate and sometimes bad. In contrast, our OLBP network is equipped with the scheme of object localization and boundary preservation, which precisely analyzes the location information of fixations and completes the gazed objects. The segmentation maps of “Ours” in Fig. 5 are very localized in the gazed objects with pretty fine details, even under the interference of some ambiguous fixations. Robustness Evaluation. We provide a robustness evaluation of our method and several representative methods, including the modified GCPA [60], CFPS [25] and the modified CPD [59], on the test dataset of the PFOS dataset. Concretely, we add the noise, i.e. unconstrained fixations, to the fixation map by random sampling on the background regions at three levels, i.e. different percentages (15%, 30%, 45%) increase in the number of unconstrained fixations of the total number of fixations. The performance of above methods after adding noise are presented in Table IV. Our method consistently outperforms the compared methods under three challenging situations, showing excellent robustness. ### VI-C Ablation Studies We comprehensively evaluate the contribution of each vital component to performance in our OLBP network. Specifically, we assess 1) the overall contributions of OLM and BPM; 2) the effectiveness of the three parts in OLM; and 3) the usefulness of BPM and the top-down manner in prediction network. The variants are retrained with the same hyper-parameters and training set as aforementioned settings in Sec. V-E, and the experiments are conducted on the PFOS dataset. TABLE VI: Ablation results of the OLM on the PFOS dataset (in percentage %). The best result in each column is bold. The corresponding structures of the listed variants are presented in Fig. 7. OLM variants | $\mathcal{J}\uparrow$ | $\mathcal{S}_{\lambda}\uparrow$ | $w\mathcal{F}_{\beta}\uparrow$ ---|---|---|--- w/o dilated convs | 72.7 -1.0 | 80.7 -0.4 | 79.0 -1.0 w/o multiply | 72.6 -1.1 | 80.5 -0.6 | 79.2 -0.8 w/o concat | 72.8 -0.9 | 80.8 -0.3 | 79.6 -0.4 w/o Seg sup | 72.9 -0.8 | 80.9 -0.2 | 79.4 -0.6 Ours | 73.7 | 81.1 | 80.0 \begin{overpic}[width=429.28616pt]{Figs/Fig7_OLM.pdf} \end{overpic} Figure 7: Structures of four OLM variants. w/o dilated convs: the four dilated convolutions are replaced by one convolutional layer; w/o multiply: without using response maps to re-weight image feature in Re-wei; w/o concat: without concatenating re-weighted feature and image feature in Re-wei; w/o Seg sup: without segmentation supervision. TABLE VII: The performance of side output segmentation maps of with/without BPM on PFOS dataset (in percentage %). The number in the lower right corner of the performance of w/o BPM is the difference between it and the performance of w/ BPM. The best result in each column is bold. Side outputs | w/ BPM (Ours) | w/o BPM ---|---|--- $\mathcal{J}\uparrow$ | $\mathcal{S}_{\lambda}\uparrow$ | $w\mathcal{F}_{\beta}\uparrow$ | $\mathcal{J}\uparrow$ | $\mathcal{S}_{\lambda}\uparrow$ | $w\mathcal{F}_{\beta}\uparrow$ $\mathbf{S}^{(5)}_{bpm}$ | 62.0 | 71.6 | 67.8 | 60.2 -1.8 | 70.4 -1.2 | 66.0 -1.8 $\mathbf{S}^{(4)}_{bpm}$ | 69.2 | 77.5 | 75.5 | 68.2 -1.0 | 76.9 -0.6 | 74.8 -0.7 $\mathbf{S}^{(3)}_{bpm}$ | 72.6 | 80.2 | 78.9 | 71.9 -0.7 | 79.8 -0.4 | 78.3 -0.6 $\mathbf{S}^{(2)}_{bpm}$ | 73.7 | 81.1 | 79.9 | 72.9 -0.8 | 80.6 -0.5 | 79.4 -0.5 $\mathbf{S}^{(1)}$ | 73.7 | 81.1 | 80.0 | 73.0 -0.7 | 80.7 -0.4 | 79.5 -0.5 1\. Does the proposed OLM and BPM contribute to OLBP network? To evaluate the contribution of the proposed OLM and BPM to OLBP network, we derive three variants: baseline network (denoted by “Ba”/“Ba∗”), baseline network with only OLMs (“Ba+OLM”), and baseline network with only BPMs (“Ba/Ba∗+BPM”). In particular, we provide two types of baseline network: the first one is an encoder-decoder network, whose input is the concatenated image and FDM (denoted by “Ba∗”); the second one is an encoder-decoder network with the down-sampled FDMs being concatenated to each skip-layer (denoted by “Ba”), i.e. the image and FDM are fed to network separately. We report the quantitative results in Tab. V. We observe that the first baseline network “Ba∗” (the 1st line in Tab. V) only obtains $67.2\%$ in $\mathcal{J}$, and the second baseline network “Ba” (the 3rd line in Tab. V) obtains $70.7\%$ in $\mathcal{J}$. This confirms that direct concatenation of the image and FDM results in the location information of FDM being submerged by image information; by contrast, concatenating FDM with image features at each scale benefits object location. OLM significantly improves the performance of the baseline network (e.g. $\mathcal{J}\\!:67.2\%/70.7\%\\!\rightarrow\\!73.0\%$ and $w\mathcal{F}_{\beta}\\!:72.2\%/75.0\%\\!\rightarrow\\!79.5\%$). This shows that the contribution of OLM is remarkable, and OLM does capture the location information. Comparing with OLM, the contribution of BPM to baseline networks is slightly inferior (e.g. $\mathcal{J}\\!:67.2\%\\!\rightarrow\\!68.0\%$; $70.7\%\\!\rightarrow\\!71.4\%$), but BPM also shows its effectiveness to improve performance of “Ba+OLM” (e.g. $w\mathcal{F}_{\beta}\\!:79.5\%\\!\rightarrow\\!80.0\%$). This demonstrates that BPM can further complete the objects and filter background of erroneous localization. With the cooperation between OLM and BPM, the performance of the whole OLBP network is improved by $6.5\%/3.0\%$ in $\mathcal{J}$, $5.2\%/2.8\%$ in $\mathcal{S}_{\lambda}$ and $7.8\%/5.0\%$ in $w\mathcal{F}_{\beta}$ compared with the baseline network “Ba∗”/“Ba”. This demonstrates that the scheme of bottom-up object localization and top-down boundary preservation is successfully embedded into the baseline network. Additionally, the segmentation maps of variants based on the first baseline network “Ba∗” are shown in Fig. 6. We observe that “Ba∗” almost segments all the objects in images. With the assistance of OLM, “Ba∗+OLM” determines the location of the gazed objects, and the gazed objects on the segmentation maps of “Ba+OLM” are much clearer. Finally, with the help of BPM, the segmentation maps of ours (i.e. OLBP network) are satisfactory. 2\. How effective are the three parts in OLM? As described in Sec. V-C, OLM consists of location analysis unit, feature re-weighting (i.e. Re-wei) and segmentation supervision (i.e. Seg sup). To validate the effectiveness of the three parts in OLM, we modify the structure of OLM and provide four variants: a) the four dilated convolutions are replaced by one convolutional layer in the location analysis unit (w/o dilated convs); b) without using response maps to re-weight image feature in Re-wei (w/o multiply); c) without concatenating re-weighted feature and image feature in Re-wei (w/o concat); and d) without segmentation supervision (w/o Seg sup). The ablation results are reported in Tab. VI, and the detailed structures of the above four OLM variants are presented in Fig. 7. \begin{overpic}[width=429.28616pt]{Figs/Fig8_PersonlVS.pdf} \put(29.5,66.3){ Visual individuation } \put(17.8,68.5){ 0.341 } \put(56.3,68.5){ 0.400 } \put(17.8,33.3){ 0.222 } \put(56.3,33.3){ 0.123 } \put(30.5,-2.3){ Visual consistency } \put(17.8,-0.1){ 0.126 } \put(56.3,-0.1){ 0.219 } \par\put(-1.0,91.8){ \begin{sideways}{Image}\end{sideways} } \put(-1.0,83.7){ \begin{sideways}{GT}\end{sideways} } \put(-1.0,73.05){ \begin{sideways}{Ours}\end{sideways} } \par\put(-1.0,56.45){ \begin{sideways}{Image}\end{sideways} } \put(-1.0,48.35){ \begin{sideways}{GT}\end{sideways} } \put(-1.0,37.7){ \begin{sideways}{Ours}\end{sideways} } \par\put(-1.0,23.0){ \begin{sideways}{Image}\end{sideways} } \put(-1.0,14.9){ \begin{sideways}{GT}\end{sideways} } \put(-1.0,4.25){ \begin{sideways}{Ours}\end{sideways} } \end{overpic} Figure 8: Visual examples of personal segmentation results. There are two basic properties of personal visual systems: visual individuation and visual consistency. The value of each image is the mean JS score. We discover that the performances of the four variants are worse than ours. Concretely, the performance degradation of w/o dilated convs (e.g. $\mathcal{J}\\!:73.7\%\\!\rightarrow\\!72.7\%$) validates that the parallel dilated convolutions analyze FDM thoroughly and one convolutional layer cannot mine sufficient location information from FDM. The performance drop of w/o multiply (e.g. $\mathcal{S}_{\lambda}\\!:81.1\%\\!\rightarrow\\!80.5\%$) confirms that the location response maps are more suitable to highlight objects on CNN feature of image than using them directly. The reason behind this is that location response maps are a group of probability maps, without rich object, texture and color information. Besides, w/o concat brings $0.9\%$ performance penalty in $\mathcal{J}$, which shows that the information balance between image and location is important. w/o Seg sup carries $0.6\%$ performance drop in $w\mathcal{F}_{\beta}$. This demonstrates that the segmentation supervision can enhance representation of the gazed objects. 3\. Is it useful to adopt BPM and the top-down manner in prediction network? To investigate the usefulness of the top-down manner in prediction network, we report the performance of side output segmentation maps of BPM in Tab. VII. Besides, we also report the side output performance of w/o BPM in Tab. VII to evaluate the importance of BPM. We observe that the quantitative results of side outputs ($\mathbf{S}^{(5)}_{bpm}$, $\mathbf{S}^{(4)}_{bpm}$, $\mathbf{S}^{(3)}_{bpm}$, $\mathbf{S}^{(2)}_{bpm}$ and $\mathbf{S}^{(1)}$) are incremental in terms of both w/ BPM (e.g. $w\mathcal{F}_{\beta}\\!:67.8\%\\!\rightarrow\\!75.5\%\\!\rightarrow\\!78.9\%\\!\rightarrow\\!79.9\%\\!\rightarrow\\!80.0\%$) and w/o BPM (e.g. $\mathcal{S}_{\lambda}\\!:70.4\%\\!\rightarrow\\!76.9\%\\!\rightarrow\\!79.8\%\\!\rightarrow\\!80.6\%\\!\rightarrow\\!80.7\%$). This confirms that the top-down manner is useful for the prediction network. The differences between the performance of w/o BPM and w/ BPM are also reported in Tab. VII. We discover that all the differences are negative, which shows that BPM works well for each side output of the top-down prediction network. ### VI-D Personal Segmentation Results Due that the personal fixations are closely related to age and gender, different users are interested in different objects when observing the same scene. We define the visual difference of different personal visual systems as visual individuation. Some examples of visual individuation are presented in the first part of Fig. 8. We can observe that there are multiple different types of objects and complex backgrounds in these images. The personal fixations of different users are located on different objects, which correspond to the distinctive GTs. \begin{overpic}[width=429.28616pt]{Figs/Fig9.pdf} \put(4.9,-2.5){ Image} \put(21.3,-2.5){ {Ours} } \put(34.8,-2.5){GT of SOD } \put(52.1,-2.5){ CPD${}_{\mathrm{sod}}$ } \put(67.3,-2.5){ GCPA${}_{\mathrm{sod}}$ } \end{overpic} Figure 9: Visual comparisons between our method, which is proposed for fixation-based object segmentation, and recent state-of-the-art salient object detection methods, including CPD [59] and GCPA [60], on the DUTS-OMRON [46] and PASCAL-S [23] datasets. “GT of SOD” means that the GT is for SOD task. “CPDsod” means the original CPD method for SOD. “GCPAsod” means the original GCPA method for SOD. In addition, we discover that personal visual systems are also consistent in some scenes, which is denoted as visual consistency. We show some examples of visual consistency in the second and third parts of Fig. 8. The images in the second part contain simple backgrounds and sparse objects, and the images in the third part contain more competitive situation, i.e. complex background and partially selected objects. In both parts, we observe that the locations of different personal fixations are similar, resulting in the identical GTs of different users. Notably, in either case, our method show the ability to segment the gazed objects consistent with the corresponding GT. We also provide the quantitative analysis of visual individuation and visual consistency with Jensen-Shannon (JS) divergence. JS divergence evaluates the similarity of two probability distributions $\mathbf{S}^{1}$ and $\mathbf{S}^{2}$, and it is based on Kullback-Leibler (KL) divergence. Its value belongs to [0, 1]. The closer its value is to zero, the smaller the difference between $\mathbf{S}^{1}$ and $\mathbf{S}^{2}$ is and the more similar they are. It can be expressed as follows: $\displaystyle\mathrm{JS}(\mathbf{S}^{1},\mathbf{S}^{2})=\frac{1}{2}\mathrm{KL}(\mathbf{S}^{1},\frac{\mathbf{S}^{1}+\mathbf{S}^{2}}{2})+\frac{1}{2}\mathrm{KL}(\mathbf{S}^{2},\frac{\mathbf{S}^{1}+\mathbf{S}^{2}}{2}),$ (15) $\displaystyle\mathrm{KL}(\mathbf{P},\mathbf{Q})=\sum^{N}_{i=1}\mathbf{P}_{\textit{i}}\mathrm{log}\left(\epsilon+\frac{\mathbf{P}_{\textit{i}}}{\epsilon+\mathbf{Q}_{\textit{i}}}\right),$ (16) where KL($\cdot$) is Kullback-Leibler divergence, which is often used as an evaluation metric in fixation prediction [71, 72, 73, 74], i indicates the ith pixel in the probability distribution, $N$ is the total number of pixels, and $\epsilon$ is a regularization constant. We introduce JS to measure the similarity of fixation points maps of each image in Fig. 8. First, we transform the fixation points map (green dots in each image) to FDM using Eq. 1; then we compute the JS score of each two FDMs; finally we report the mean JS score for each image in Fig. 8. It is obvious that the mean JS scores (i.e. 0.222, 0.123, 0.126, and 0.219) of images which belong to visual consistency are relatively smaller than those (i.e. 0.341 and 0.400) of images which belong to visual individuation. And the mean JS scores of images which belong to visual consistency are close to zero, which indicates that the distributions of FDMs are very similar, i.e. people may look at the same object(s). ### VI-E Discussions Salient Object Detection (SOD) is widely explored in color images [59, 60, 75, 76, 77], RGB-D images [78, 79] and videos [80, 81, 82], and it is closely related to our fixation-based object segmentation task. In this section, we discuss the connections between fixation-based object segmentation and SOD. SOD aims to highlight the most visually attractive object(s) in a scene, while fixation-based object segmentation aims to segment the gazed objects according to the fixation map, as defined in Sec. III. To illustrate the differences and connections between these two tasks, we conduct experiments on two SOD datasets, i.e. DUTS-OMRON [46] and PASCAL-S [23], and show visual comparisons with two state-of-the-art SOD methods, i.e. CPD [59] and GCPA [60], in Fig. 9, which summarizes three situations. First, in the $1^{\mathrm{st}}$ and $2^{\mathrm{nd}}$ rows, we present the differences of these two tasks: our method not only segments the salient objects, such as the bird and the big tent, but also segments the gazed wood stake and cloth that are not found in the GT of SOD and the results of CPD and GCPA. Second, in the $3^{\mathrm{rd}}$ and $4^{\mathrm{th}}$ rows, we find that the results of CPD and GCPA are similar to ours, but different from the GT of SOD. This shows that to some extent, the results of SOD methods CPD and GCPA are consistent with the fixation maps, even if the fixation maps are not exploited in these methods. Third, in the $5^{\mathrm{th}}$ and $6^{\mathrm{th}}$ rows, we can clearly observe that our results are consistent with the fixation points in images, while the other three maps are different. This shows that different SOD methods may cause confusion in some complicated scenes, resulting in inaccurate saliency maps. Furthermore, we find that the salient objects always appear in the results of our method, while there is ambiguity among different SOD methods, which may highlight different salient objects. So, to improve the accuracy of different SOD methods, we believe that the fixation-based object segmentation can be a pre-processing operation for SOD to determine the salient object proposals. ## VII Conclusion In this paper, we propose a three-step approach to transform the available fixation prediction dataset OSIE to the PFOS dataset for personal fixations- based object segmentation. The PFOS dataset is meaningful to promote the development of fixations-based object segmentation. Moreover, we present a novel OLBP network with the scheme of bottom-up object localization and top- down boundary preservation to segment the gazed objects. Our OLBP network is equipped with two essential components: the object localization module and the boundary preservation module. OLM is object locator, which is in charge of location analysis of fixations and object enhancement. BPM emphasizes erroneous localization distillation and object completeness preservation. Besides, we provide comprehensive experiments of our OLBP network and other three types of methods on the PFOS dataset, which demonstrate the excellence of our OLBP network and validate the challenges of the PFOS dataset. In our future work, we plan to apply the proposed OLBP network to some eye-control devices, facilitating the lives of patients with hand disability, ALS and polio. In addition, we plan to recruit subjects to collect fixation points and corresponding ground truths on the PASCAL VOC [45] and MS COCO [83] datasets for further exploring personal fixation-based object segmentation. ## References * [1] O. Le Meur, A. Coutrot, Z. Liu, P. Rämä, A. Le Roch, and A. Helo, “Visual attention saccadic models learn to emulate gaze patterns from childhood to adulthood,” _IEEE Trans. Image Process._ , vol. 26, no. 10, pp. 4777–4789, Oct. 2017. * [2] A. Mahdi, M. Su, M. Schlesinger, and J. Qin, “A comparison study of saliency models for fixation prediction on infants and adults,” _IEEE Trans. Cogn. Devel. Syst._ , vol. 10, no. 3, pp. 485–498, Sep. 2018. * [3] J. Hewig, R. H. Trippe, H. Hecht, T. Straube, and W. H. R. Miltner, “Gender differences for specific body regions when looking at men and women,” _J. Nonverbal Behav._ , vol. 32, no. 2, pp. 67–78, Jun. 2008. * [4] N. Alwall, D. Johansson, and S. Hansen, “The gender difference in gaze-cueing: Associations with empathizing and systemizing,” _Pers. Individ. Differ._ , vol. 49, no. 7, pp. 729–732, Nov. 2010. * [5] A. Li and Z. Chen, “Personalized visual saliency: Individuality affects image perception,” _IEEE Access_ , vol. 6, pp. 16 099–16 109, Jan. 2018. * [6] Y. Xu, S. Gao, J. Wu, N. Li, and J. Yu, “Personalized saliency and its prediction,” _IEEE Trans. Pattern Anal. Mach. Intell._ , vol. 41, no. 12, pp. 2975–2989, Dec. 2019. * [7] Y. Boykov and M.-P. Jolly, “Interactive graph cuts for optimal boundary region segmentation of objects in N-D images,” in _Proc. IEEE ICCV_ , Jul. 2001, pp. 105–112. * [8] L. Grady, “Random walks for image segmentation,” _IEEE Trans. Pattern Anal. Mach. Intell._ , vol. 28, no. 11, pp. 1768–1783, Nov. 2006. * [9] X. Bai and G. Sapiro, “Geodesic matting: A framework for fast interactive image and video segmentation and matting,” _Int. J. Comput. Vis._ , vol. 82, no. 2, pp. 113–132, Apr. 2009. * [10] T. N. A. Nguyen, J. Cai, J. Zhang, and J. Zheng, “Robust interactive image segmentation using convex active contours,” _IEEE Trans. Image Process._ , vol. 21, no. 8, pp. 3734–3743, Aug. 2012. * [11] T. V. Spina, P. A. V. de Miranda, and A. Xavier Falcão, “Hybrid approaches for interactive image segmentation using the live markers paradigm,” _IEEE Trans. Image Process._ , vol. 23, no. 12, pp. 5756–5769, Dec. 2014. * [12] N. Xu, B. Price, S. Cohen, J. Yang, and T. Huang, “Deep interactive object selection,” in _Proc. IEEE CVPR_ , Jun. 2016, pp. 373–381. * [13] J. Liew, Y. Wei, W. Xiong, S.-H. Ong, and J. Feng, “Regional interactive image segmentation networks,” in _Proc. IEEE ICCV_ , Oct. 2017, pp. 2746–2754. * [14] Z. Li, Q. Chen, and V. Koltun, “Interactive image segmentation with latent diversity,” in _Proc. IEEE CVPR_ , Jun. 2018, pp. 577–585. * [15] S. Mahadevan, P. Voigtlaender, and B. Leibe, “Iteratively trained interactive segmentation,” in _Proc. BMVC_ , Sep. 2018. * [16] W.-D. Jang and C.-S. Kim, “Interactive image segmentation via backpropagating refinement scheme,” in _Proc. IEEE CVPR_ , Jun. 2019, pp. 5292–5301. * [17] Y. Hu, A. Soltoggio, R. Lock, and S. Carter, “A fully convolutional two-stream fusion network for interactive image segmentation,” _Neural Netw._ , vol. 109, pp. 31–42, Jan. 2019. * [18] C. Rother, V. Kolmogorov, and A. Blake, “‘GrabCut’: Interactive foreground extraction using iterated graph cuts,” _ACM Trans. Graph._ , vol. 23, no. 3, pp. 309–314, Aug. 2004. * [19] V. Lempitsky, P. Kohli, C. Rother, and T. Sharp, “Image segmentation with a bounding box prior,” in _Proc. IEEE ICCV_ , Sep. 2009, pp. 277–284. * [20] N. Xu, B. Price, S. Cohen, J. Yang, and T. Huang, “Deep grabcut for object selection,” in _Proc. BMVC_ , Sep. 2017, pp. 1–12. * [21] R. Shi, K. N. Ngan, S. Li, and H. Li, “Interactive object segmentation in two phases,” _Signal Process. Image Commun._ , vol. 65, pp. 107–114, Jul. 2018\. * [22] A. K. Mishra, Y. Aloimonos, L. F. Cheong, and A. Kassim, “Active visual segmentation,” _IEEE Trans. Pattern Anal. Mach. Intell._ , vol. 34, no. 4, pp. 639–653, Apr. 2012. * [23] Y. Li, X. Hou, C. Koch, J. M. Rehg, and A. L. Yuille, “The secrets of salient object segmentation,” in _Proc. IEEE CVPR_ , Jun. 2014, pp. 280–287. * [24] R. Shi, N. K. Ngan, and H. Li, “Gaze-based object segmentation,” _IEEE Signal Process. Lett._ , vol. 24, no. 10, pp. 1493–1497, Oct. 2017. * [25] G. Li, Z. Liu, R. Shi, and W. Wei, “Constrained fixation point based segmentation via deep neural network,” _Neurocomputing_ , vol. 368, pp. 180–187, Nov. 2019. * [26] J. Xu, M. Jiang, S. Wang, M. S. Kankanhalli, and Q. Zhao, “Predicting human gaze beyond pixels,” _J. Vis._ , vol. 14, no. 1, pp. 1–20, Jan. 2014. * [27] M. Jian and C. Jung, “Interactive image segmentation using adaptive constraint propagation,” _IEEE Trans. Image Process._ , vol. 25, no. 3, pp. 1301–1311, Mar. 2016. * [28] T. Wang, J. Yang, Z. Ji, and Q. Sun, “Probabilistic diffusion for interactive image segmentation,” _IEEE Trans. Image Process._ , vol. 28, no. 1, pp. 330–342, Jan. 2019. * [29] M. Sadeghi, G. Tien, G. Hamarneh, and M. S. Atkins, “Hands-free interactive image segmentation using eyegaze,” in _SPIE Medical Imaging 2009: Computer-Aided Diagnosis_ , vol. 7260, Mar. 2009, pp. 441–450. * [30] G. Kootstra, N. Bergström, and D. Kragic, “Using symmetry to select fixation points for segmentation,” in _Proc. IEEE ICPR_ , Aug. 2010, pp. 3894–3897. * [31] X. Tian and C. Jung, “Point-cut: Fixation point-based image segmentation using random walk model,” in _Proc. IEEE ICIP_ , Sep. 2015, pp. 2125–2129. * [32] N. Khosravan, H. Celik, B. Turkbey, R. Cheng, E. McCreedy, M. McAuliffe, S. Bednarova, E. Jones, X. Chen, P. Choyke, B. Wood, and U. Bagci, “Gaze2Segment: A pilot study for integrating eye-tracking technology into medical image segmentation,” in _MICCAIW_ , Jul. 2017, pp. 94–104. * [33] N. Wang and X. Gong, “Adaptive fusion for RGB-D salient object detection,” _IEEE Access_ , vol. 7, pp. 55 277–55 284, May 2019. * [34] Y. Wang, X. Zhao, X. Hu, Y. Li, and K. Huang, “Focal boundary guided salient object detection,” _IEEE Trans. Image Process._ , vol. 28, no. 6, pp. 2813–2824, Jan. 2019. * [35] Z. Wu, L. Su, and Q. Huang, “Stacked cross refinement network for edge-aware salient object detection,” in _Proc. IEEE ICCV_ , Oct. 2019, pp. 7263–7272. * [36] J.-X. Zhao, J.-J. Liu, D.-P. Fan, Y. Cao, J. Yang, and M.-M. Cheng, “EGNet: Edge guidance network for salient object detection,” in _Proc. IEEE ICCV_ , Oct. 2019, pp. 8778–8787. * [37] H. Ding, X. Jiang, A. Q. Liu, N. M. Thalmann, and G. Wang, “Boundary-aware feature propagation for scene segmentation,” in _Proc. IEEE ICCV_ , Oct. 2019, pp. 6818–6828. * [38] W. Wei, Z. Liu, L. Huang, A. Nebout, and O. Le Meur, “Saliency prediction via multi-level features and deep supervision for children with autism spectrum disorder,” in _Proc. IEEE ICME Grand Challenges_ , Jul. 2019, pp. 621–624. * [39] X. Yang, M.-L. Shyu, H.-Q. Yu, S.-M. Sun, N.-S. Yin, and W. Chen, “Integrating image and textual information in human–robot interactions for children with autism spectrum disorder,” _IEEE Trans. Multimedia_ , vol. 21, no. 3, pp. 746–759, Mar. 2019. * [40] J. E. Silberg, I. Agtzidis, M. Startsev, T. Fasshauer, K. Silling, A. Sprenger, M. Dorr, and R. Lencer, “Free visual exploration of natural movies in schizophrenia,” _Eur. Arch. Psychiatry Clin. Neurosci._ , vol. 269, no. 4, pp. 407–418, Jun. 2019. * [41] J. Polec, R. Vargic, F. Csóka, E. Smolejová, A. Heretik, M. Bieliková, M. Svrček, and R. Móro, “Detection of schizophrenia spectrum disorders using saliency maps,” in _Proc. IEEE AICT_ , Sep. 2017, pp. 1–5. * [42] T. Judd, K. Ehinger, F. Durand, and A. Torralba, “Learning to predict where humans look,” in _Proc. IEEE ICCV_ , Sep. 2009, pp. 2106–2113. * [43] M. Jiang, S. Huang, J. Duan, and Q. Zhao, “SALICON: Saliency in context,” in _Proc. IEEE CVPR_ , Jun. 2015, pp. 1072–1080. * [44] K. McGuinness and N. E. O’Connor, “A comparative evaluation of interactive segmentation algorithms,” _Pattern Recognit._ , vol. 43, no. 2, pp. 434–444, Feb. 2010. * [45] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman, “The PASCAL visual object classes (voc) challenge,” _Int. J. Comput. Vis._ , vol. 88, no. 2, pp. 303–338, Jun. 2010. * [46] C. Yang, L. Zhang, H. Lu, X. Ruan, and M.-H. Yang, “Saliency detection via graph-based manifold ranking,” in _Proc. IEEE CVPR_ , Jun. 2013, pp. 3166–3173. * [47] A. Borji, M.-M. Cheng, H. Jiang, and J. Li, “Salient object detection: A benchmark,” _IEEE Trans. Image Process._ , vol. 24, no. 12, pp. 5706–5722, Dec. 2015. * [48] W. Wang, Q. Lai, H. Fu, J. Shen, and H. Ling, “Salient object detection in the deep learning era: An in-depth survey,” _arXiv preprint arXiv:1904.09146_ , 2019. * [49] Z. Liu, W. Zou, and O. Le Meur, “Saliency tree: A novel saliency detection framework,” _IEEE Trans. Image Process._ , vol. 23, no. 5, pp. 1937–1952, May 2014. * [50] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” _arXiv preprint arXiv:1409.1556_ , 2014. * [51] L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille, “DeepLab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs,” _IEEE Trans. Pattern Anal. Mach. Intell._ , vol. 40, no. 4, pp. 834–848, Apr. 2018. * [52] F. Yu and V. Koltun, “Multi-scale context aggregation by dilated convolutions,” in _Proc. ICLR_ , May 2016. * [53] S. Xie and Z. Tu, “Holistically-nested edge detection,” in _Proc. IEEE ICCV_ , Dec. 2015, pp. 1395–1403. * [54] H. Zhao, J. Shi, X. Qi, X. Wang, and J. Jia, “Pyramid scene parsing network,” in _Proc. IEEE CVPR_ , Jul. 2017, pp. 6230–6239. * [55] V. Badrinarayanan, A. Kendall, and R. Cipolla, “SegNet: A deep convolutional encoder-decoder architecture for image segmentation,” _IEEE Trans. Pattern Anal. Mach. Intell._ , vol. 39, no. 12, pp. 2481–2495, Dec. 2017. * [56] H. Zhang, K. Dana, J. Shi, Z. Zhang, X. Wang, A. Tyagi, and A. Agrawal, “Context encoding for semantic segmentation,” in _Proc. IEEE CVPR_ , Jun. 2018, pp. 7151–7160. * [57] L.-C. Chen, Y. Zhu, G. Papandreou, F. Schroff, and H. Adam, “Encoder-decoder with atrous separable convolution for semantic image segmentation,” in _Proc. ECCV_ , Sep. 2018, pp. 833–851. * [58] K. Sun, Y. Zhao, B. Jiang, T. Cheng, B. Xiao, D. Liu, Y. Mu, X. Wang, W. Liu, and J. Wang, “High-resolution representations for labeling pixels and regions,” in _Proc. IEEE CVPR_ , Jul. 2019. * [59] Z. Wu, L. Su, and Q. Huang, “Cascaded partial decoder for fast and accurate salient object detection,” in _Proc. IEEE CVPR_ , Jun. 2019, pp. 3902–3911. * [60] Z. Chen, Q. Xu, R. Cong, and Q. Huang, “Global context-aware progressive aggregation network for salient object detection,” in _Proc. AAAI_ , Feb. 2020. * [61] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Dropout: A simple way to prevent neural networks from overfitting,” _J. Mach. Learn. Res._ , vol. 15, no. 1, pp. 1929–1958, Jun. 2014. * [62] X. Glorot and Y. Bengio, “Understanding the difficulty of training deep feedforward neural networks,” in _Proc. AISTATS_ , May 2010, pp. 249–256. * [63] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell, “Caffe: Convolutional architecture for fast feature embedding,” in _Proc. ACM MM_ , Nov. 2014, pp. 675–678. * [64] L. Bottou, “Large-scale machine learning with stochastic gradient descent,” in _Proc. COMPSTAT_ , Aug. 2010, pp. 177–186. * [65] D.-P. Fan, M.-M. Cheng, Y. Liu, T. Li, and A. Borji, “Structure-measure: A new way to evaluate foreground maps,” in _Proc. IEEE ICCV_ , Oct. 2017, pp. 4548–4557. * [66] R. Margolin, L. Zelnik-Manor, and A. Tal, “How to evaluate foreground maps,” in _Proc. IEEE CVPR_ , Jun. 2014, pp. 248–255. * [67] D.-P. Fan, C. Gong, Y. Cao, B. Ren, M.-M. Cheng, and A. Borji, “Enhanced-alignment measure for binary foreground map evaluation,” in _Proc. IJCAI_ , Jul. 2018, pp. 698–704. * [68] G. Li, Z. Liu, and H. Ling, “ICNet: Information conversion network for RGB-D based salient object detection,” _IEEE Trans. Image Process._ , vol. 29, pp. 4873–4884, Mar. 2020. * [69] G. Li, Z. Liu, L. Ye, Y. Wang, and H. Ling, “Cross-modal weighting network for RGB-D salient object detection,” in _Proc. ECCV_ , Aug. 2020, pp. 665–681. * [70] N. Otsu, “A threshold selection method from gray-level histograms,” _IEEE Trans. Syst. Man Cybern._ , vol. 9, no. 1, pp. 62–66, Jan. 1979. * [71] W. Wang and J. Shen, “Deep visual attention prediction,” _IEEE Trans. Image Process._ , vol. 27, no. 5, pp. 2368–2378, May 2018. * [72] M. Cornia, L. Baraldi, G. Serra, and R. Cucchiara, “Predicting human eye fixations via an LSTM-based saliency attentive model,” _IEEE Trans. Image Process._ , vol. 27, no. 10, pp. 5142–5154, Oct. 2018. * [73] N. Liu, J. Han, T. Liu, and X. Li, “Learning to predict eye fixations via multiresolution convolutional neural networks,” _IEEE Trans. Neural Netw. Learn. Syst._ , vol. 29, no. 2, pp. 392–404, Feb. 2018. * [74] Z. Che, A. Borji, G. Zhai, X. Min, G. Guo, and P. L. Callet, “How is gaze influenced by image transformations? Dataset and model,” _IEEE Trans. Image Process._ , vol. 29, pp. 2287–2300, 2020. * [75] K. Fu, C. Gong, I. Y.-H. Gu, and J. Yang, “Normalized cut-based saliency detection by adaptive multi-level region merging,” _IEEE Trans. Image Process._ , vol. 24, no. 12, pp. 5671–5683, Oct. 2015. * [76] K. Fu, Q. Zhao, and I. Y.-H. Gu, “Refinet: A deep segmentation assisted refinement network for salient object detection,” _IEEE Trans. Multimedia_ , vol. 21, no. 2, pp. 457–469, Feb. 2019. * [77] K. Fu, Q. Zhao, I. Y.-H. Gu, and JieYang, “Deepside: A general deep framework for salient object detection,” _Neurocomputing_ , vol. 365, pp. 69–82, Sep. 2019. * [78] K. Fu, D.-P. Fan, G.-P. Ji, Q. Zhao, J. Shen, and C. Zhu, “Siamese network for RGB-D salient object detection and beyond,” _arXiv preprint arXiv:2008.12134_ , 2020. * [79] K. Fu, D.-P. Fan, G.-P. Ji, and Q. Zhao, “JL-DCF: Joint learning and densely-cooperative fusion framework for RGB-D salient object detection,” in _Proc. IEEE CVPR_ , Jun. 2020, pp. 3049–3059. * [80] X. Zhou, Z. Liu, C. Gong, and W. Liu, “Improving video saliency detection via localized estimation and spatiotemporal refinement,” _IEEE Trans. Multimedia_ , vol. 20, no. 11, pp. 2993–3007, Nov. 2018. * [81] D.-P. Fan, W. Wang, M.-M. Cheng, and J. Shen, “Shifting more attention to video salient object detection,” in _Proc. IEEE CVPR_ , Jun. 2019, pp. 8546–8556. * [82] W. Wang, J. Shen, J. Xie, M.-M. Cheng, H. Ling, and A. Borji, “Revisiting video saliency prediction in the deep learning era,” _IEEE Trans. Pattern Anal. Mach. Intell._ , vol. 43, no. 1, pp. 220–237, Jan. 2021. * [83] T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. L. Zitnick, “Microsoft COCO: Common objects in context,,” in _Proc. ECCV_ , Sep. 2014, pp. 740–755.
# Unsupervised Technical Domain Terms Extraction using Term Extractor Suman Dowlagar LTRC IIIT-Hyderabad suman.dowlagar@ research.iiit.ac.in Radhika Mamidi LTRC IIIT-Hyderabad radhika.mamidi@ iiit.ac.in ###### Abstract Terminology extraction, also known as term extraction, is a subtask of information extraction. The goal of terminology extraction is to extract relevant words or phrases from a given corpus automatically. This paper focuses on the unsupervised automated domain term extraction method that considers chunking, preprocessing, and ranking domain-specific terms using relevance and cohesion functions for ICON 2020 shared task 2: TermTraction. ## 1 Introduction The aim of Automatic Term Extraction (ATE) is to extract terms such as words, phrases, or multi-word expressions from the given corpus. ATE is widely used in many NLP tasks, such as machine translation, summarization, clustering the documents, and information retrieval. Unsupervised algorithms for domain term extraction are not labeled and trained on the corpus and do not have any pre-defined rules or dictionaries. They often use statistical information from the text. Most of these algorithms use stop word lists and can be applied to any text datasets. The standard unsupervised automated term extraction pipeline consists of * • Simple Rules: using chunking or POS tagging to extract Noun phrases for multi- word extraction. * • Naive counting: that counts how many terms each word occurs in the corpus. * • Preprocessing: Removing punctuation and common words such as stop words from the text. * • Candidate generation and scoring: using statistical measures and ranking algorithms to generate the possible set of domain terms * • Final set: Arrange the ranked terms in descending order based on the scores and take the top N keywords as the output. Currently, there are many methods for automatic term recognition. Evans and Lefferts (1995) used TF-IDF measure for term extraction.Navigli and Velardi (2002) used domain consensus which is designed to recognize the terms uniformly distributed over the whole corpus. The most popular method C-value Frantzi et al. (2000) is also a statistical measure that extracts a term based on the term’s frequency, length of the term, and the set of the candidates that enclose the term such that the term is in their substring. Bordea et al. (2013) proposed the method called Basic, which is a modification of the C-value for recognizing terms of average specificity. The successor of C-value statistic called the NC value Frantzi et al. (2000) considered scored the term based on the condition if it exists in a group of common words or if it contains nouns, verbs, or adjectives that immediately precede or follow the term. The methods proposed by Ahmad et al. (1999); Kozakov et al. (2004); Sclano and Velardi (2007) are based on extracting the terms of a text by considering the frequency of occurrence of terms in the general domain. A detailed survey of the existing automated term extraction algorithms and their evaluation are presented in papers by Astrakhantsev et al. (2015); Šajatović et al. (2019) In this paper, we used the term extractor algorithm Sclano and Velardi (2007) present in the pyate111https://pypi.org/project/pyate/ library for domain term extraction. The term extractor algorithm is developed initially for ontology extraction from large corpora. It uses domain pertinence/relevance, domain consensus, and lexical cohesion for extracting terms. A detailed description of the modules is given in the next section. The paper is organized as follows. Section 2 gives a detailed description of the term extraction algorithm used. Section 3 gives information about the datasets used and results. Section 4 concludes the paper. ## 2 Our Approach In this section, we describe in detail the methods used in the term extractor algorithm. Initially, TermExtractor performs chunking and proper name recognition and then extracts structures based on linguistic rules and patterns, including stop words, detection of misspellings, and acronyms. The extraction algorithm uses Domain Pertinence, Domain Cohesion, and Lexical Cohesion to decide if a term is considered a domain term. Domain Pertinence, or Domain Relevance (DR), requires a contrastive corpus and compares a candidate’s occurrence in the documents belonging to the target domain to its occurrence in other domains, but the measure only depends on the contrastive domain where the candidate has the highest frequency. The Domain Pertinence is based on a simple formula, $DR_{D_{i}}(t)=\frac{tf_{i}}{max_{j}(tf_{j})}$ (1) Where $tf_{i}$ is the frequency of the candidate term in the input domain- specific document collection and $max_{j}(tf_{j})$ is the general corpus domain, where the candidate has the highest frequency, and $D_{i}$ is the domain in consideration. Domain Consensus (DC) assumes that several documents represent a domain. It measures the extent to which the candidate is evenly distributed on these documents by considering normalized term frequencies $(\phi)$, $DC_{D_{i}}(t)=\sum_{d_{k}\epsilon D_{i}}\phi_{k}log\phi_{k}$ (2) Here, we assume $k$ distinct documents for the domain $D_{i}$. Lexical cohesion involves the choice of vocabulary. It is concerned with the relationship that exists between lexical items in a text, such as words and phrases. It compares the in-term distribution of words that make up a term with their out-of-term distribution. $LC_{D_{i}}(t)=\frac{n*tf_{i}*logtf_{i}}{\sum_{j}tf_{w_{j}}i}$ (3) Where $n$ is the number of documents in which the term $t$ occurs. The final weight of a term is computed as a weighted average of the three filters above, $score(t,D_{i})=\alpha*DR+\beta*DC+\gamma*LC$ (4) where $\alpha$, $\beta$, $\gamma$ are the weights, and they are equal to $1/3$ ## 3 Experiments This section describes the dataset used for domain terms extraction, implementation of the above approach, followed by results, and error analysis. ### 3.1 Dataset We used the dataset provided by the organizers of TermTraction ICON-2020. The task is to extract domain terms from the given English documents from the four technical domains like Computer Science, Physics, Life Science, Law. The data statistics of the documents in the respective domains are shown in the table 1. Domain | #Train docs | #Test docs ---|---|--- Bio-Chemistry | 229 | 10 Communication | 127 | 10 Computer-Science | 201 | 8 Law | 70 | 16 Table 1: Data statistics Biochemistry | Communication | Computer Science | Law ---|---|---|--- Data | run 1 | run 2 | Data | run 1 | run 2 | Data | run 1 | run 2 | Data | run 1 | run 2 M12S1 | 0.247 | 0.222 | M2-1 | 0.109 | 0.086 | KL2 | 0.220 | 0.225 | A01 | 0.079 | 0.077 M15S2 | 0.208 | 0.195 | M2-2 | 0.102 | 0.104 | KL4 | 0.241 | 0.246 | A02 | 0.099 | 0.066 M16S2 | 0.224 | 0.207 | M2-3 | 0.094 | 0.074 | KL8 | 0.138 | 0.146 | A03 | 0.144 | 0.126 M23S3 | 0.266 | 0.233 | M3-1 | 0.240 | 0.236 | W12 | 0.143 | 0.122 | FA1 | 0.104 | 0.116 M26S2 | 0.096 | 0.081 | M3-2 | 0.159 | 0.148 | W1332 | 0.216 | 0.195 | FA2 | 0.077 | 0.067 T18 | 0.463 | 0.427 | M3-3 | 0.140 | 0.132 | W13 | 0.108 | 0.089 | FC1 | 0.082 | 0.073 T25 | 0.310 | 0.282 | RM16 | 0.101 | 0.088 | W1436 | 0.181 | 0.165 | FC2 | 0.032 | 0.021 T39 | 0.265 | 0.247 | RM17 | 0.067 | 0.065 | W921 | 0.221 | 0.188 | FC3 | 0.016 | 0.014 T4 | 0.271 | 0.234 | RM18 | 0.098 | 0.115 | | | | FR1 | 0.149 | 0.113 T9 | 0.323 | 0.315 | SW1AW | 0.120 | 0.113 | | | | FR2 | 0.144 | 0.112 | | | | | | | | | FR3 | 0.073 | 0.062 | | | | | | | | | G3 | 0.103 | 0.098 | | | | | | | | | G4 | 0.056 | 0.052 | | | | | | | | | R1 | 0.022 | 0.055 | | | | | | | | | R2 | 0.033 | 0.026 | | | | | | | | | R3 | 0.044 | 0.048 Table 2: Term Extraction macro-F1 score. Template Sentence | Domain terms identified ---|--- We are not going to that , remove it completely, but nevertheless this is an indication that , NO plus is going to be a poorer donor , compared to carbon monoxide . So , this drastic reduction in the stretching frequency can only happen if you have , a large population of the anti - bonding orbitals of NO plus . And it has got a structure , which is very similar , a structure which is very similar to the structure of nickel tetra carbonyl . You will see that , while carbon monoxide is ionized with 15 electron volts , if you supply 15 electron volts , carbon monoxide can be oxidized or ionized . | | large population --- similar ionized carbonyl frequency poorer donor anti - bonding orbitals indication carbon monoxide electron volts nickel plus drastic reduction structure Table 3: Error analysis on the template sentence ### 3.2 Implementation We used Pyate (python automated term extraction library) that contains the term extractor method and is trained on the general corpus. With the help of the term extraction method, we extracted the relevant terms from the given corpus. We have submitted two runs, one run (run 1) is the term extractor function itself, and the other run (run 2) is term extractor combined with NP chunks of phrase length ¿ 2 obtained from NLTK ConsecutiveNPChunkTagger222ConsecutiveNPChunkTagger . ### 3.3 Results and Error Analysis We evaluated the performance of the method using average precision. The results are tabulated in Table 2. For the template sentence given in Table 3, our algorithm failed to recognize the domain terms NO plus and nickel tetra carbonyl. It considered NO as the stop word (no or negation) and discarded it while preprocessing. The algorithm also misunderstood words like “similar” as domain terms and failed to identify nickel tetra carbonyl as a domain term. It indicates that further study is necessary, which considers the candidate terms’ capitalization and uses better methods that support the more reliable form of compound words or multi-word expressions. ## 4 Conclusion For domain term extraction from technical domains like Bio-Chemistry, Law, Computer-Science, and communication, We used the term extractor method from pyate library for obtaining technical terms. The term extractor method uses keywords from the general corpora, and it considers Domain Pertinence, Domain Cohesion, and Lexical Cohesion methods for extracting domain terms in the given corpus. As mentioned above, it did not give preference to capitalized terms and did not consider some compound words. So we have to work towards better methods that consider capitalization, better formation of compound words for the more reliable performance of the automated domain term extractor. ## References * Ahmad et al. (1999) Khurshid Ahmad, Lee Gillam, Lena Tostevin, et al. 1999. University of surrey participation in trec8: Weirdness indexing for logical document extrapolation and retrieval (wilder). In _TREC_ , pages 1–8. * Astrakhantsev et al. (2015) Nikita A Astrakhantsev, Denis G Fedorenko, and D Yu Turdakov. 2015. Methods for automatic term recognition in domain-specific text collections: A survey. _Programming and Computer Software_ , 41(6):336–349. * Bordea et al. (2013) Georgeta Bordea, Paul Buitelaar, and Tamara Polajnar. 2013. Domain-independent term extraction through domain modelling. In _The 10th international conference on terminology and artificial intelligence (TIA 2013), Paris, France_. 10th International Conference on Terminology and Artificial Intelligence. * Evans and Lefferts (1995) David A Evans and Robert G Lefferts. 1995. Clarit-trec experiments. _Information processing & management_, 31(3):385–395. * Frantzi et al. (2000) Katerina Frantzi, Sophia Ananiadou, and Hideki Mima. 2000. Automatic recognition of multi-word terms:. the c-value/nc-value method. _International journal on digital libraries_ , 3(2):115–130. * Kozakov et al. (2004) Lev Kozakov, Youngja Park, T Fin, Youssef Drissi, Yurdaer Doganata, and Thomas Cofino. 2004. Glossary extraction and utilization in the information search and delivery system for ibm technical support. _IBM Systems Journal_ , 43(3):546–563. * Navigli and Velardi (2002) Roberto Navigli and Paola Velardi. 2002. Semantic interpretation of terminological strings. In _Proc. 6th Int’l Conf. Terminology and Knowledge Eng_ , pages 95–100. * Šajatović et al. (2019) Antonio Šajatović, Maja Buljan, Jan Šnajder, and Bojana Dalbelo Bašić. 2019. Evaluating automatic term extraction methods on individual documents. In _Proceedings of the Joint Workshop on Multiword Expressions and WordNet (MWE-WN 2019)_ , pages 149–154. * Sclano and Velardi (2007) Francesco Sclano and Paola Velardi. 2007. Termextractor: a web application to learn the shared terminology of emergent web communities. In _Enterprise Interoperability II_ , pages 287–290. Springer.
fifi # Experimentally Realizing Efficient Quantum Control with Reinforcement Learning Ming-Zhong Ai These two authors contributed equally to this work. CAS Key Laboratory of Quantum Information, University of Science and Technology of China, Hefei 230026, China CAS Center For Excellence in Quantum Information and Quantum Physics, University of Science and Technology of China, Hefei 230026, China Yongcheng Ding These two authors contributed equally to this work. International Center of Quantum Artificial Intelligence for Science and Technology (QuArtist) and Department of Physics, Shanghai University, 200444 Shanghai, China Department of Physical Chemistry, University of the Basque Country UPV/EHU, Apartado 644, 48080 Bilbao, Spain Yue Ban Department of Physical Chemistry, University of the Basque Country UPV/EHU, Apartado 644, 48080 Bilbao, Spain School of Materials Science and Engineering, Shanghai University, 200444 Shanghai, China José D. Martín-Guerrero IDAL, Electronic Engineering Department, University of Valencia, Avgda. Universitat s/n, 46100 Burjassot, Valencia, Spain Jorge Casanova Department of Physical Chemistry, University of the Basque Country UPV/EHU, Apartado 644, 48080 Bilbao, Spain IKERBASQUE, Basque Foundation for Science, Plaza Euskadi 5, 48009 Bilbao, Spain Jin-Ming Cui<EMAIL_ADDRESS>CAS Key Laboratory of Quantum Information, University of Science and Technology of China, Hefei 230026, China CAS Center For Excellence in Quantum Information and Quantum Physics, University of Science and Technology of China, Hefei 230026, China Yun-Feng Huang<EMAIL_ADDRESS>CAS Key Laboratory of Quantum Information, University of Science and Technology of China, Hefei 230026, China CAS Center For Excellence in Quantum Information and Quantum Physics, University of Science and Technology of China, Hefei 230026, China Xi Chen<EMAIL_ADDRESS>International Center of Quantum Artificial Intelligence for Science and Technology (QuArtist) and Department of Physics, Shanghai University, 200444 Shanghai, China Department of Physical Chemistry, University of the Basque Country UPV/EHU, Apartado 644, 48080 Bilbao, Spain Chuan-Feng Li<EMAIL_ADDRESS>CAS Key Laboratory of Quantum Information, University of Science and Technology of China, Hefei 230026, China CAS Center For Excellence in Quantum Information and Quantum Physics, University of Science and Technology of China, Hefei 230026, China Guang-Can Guo CAS Key Laboratory of Quantum Information, University of Science and Technology of China, Hefei 230026, China CAS Center For Excellence in Quantum Information and Quantum Physics, University of Science and Technology of China, Hefei 230026, China ###### Abstract Robust and high-precision quantum control is crucial but challenging for scalable quantum computation and quantum information processing. Traditional adiabatic control suffers severe limitations on gate performance imposed by environmentally induced noise because of a quantum system’s limited coherence time. In this work, we experimentally demonstrate an alternative approach to quantum control based on deep reinforcement learning (DRL) on a trapped ${}^{171}\mathrm{Yb}^{+}$ ion. In particular, we find that DRL leads to fast and robust digital quantum operations with running time bounded by shortcuts to adiabaticity (STA). Besides, we demonstrate that DRL’s robustness against both Rabi and detuning errors can be achieved simultaneously without any input from STA. Our experiments reveal a general framework of digital quantum control, leading to a promising enhancement in quantum information processing. ††preprint: APS/123-QED ## I INTRODUCTION Two-level systems physically realize qubits, which are the basic units of digital quantum computing. In this paradigm, externally controllable parameters should be designed to manipulate the qubits, implementing fast and robust gate operations. Thus, one can construct a universal fault-tolerant quantum computer with physical platforms based on trapped ions and superconducting circuits Nielsen and Chuang (2010). In this way, quantum error correction can also be realized physically to reduce the effects of quantum noises and systematic errors. From this perspective, quantum control is bridged to quantum information processing and quantum computing. This connection leads to enormous researches devoted to producing precise quantum control of qubits with driving fields, including adiabatic passages Král _et al._ (2007), optimized resonant $\pi$ pulses Remizov _et al._ (2015), composite pulses Brown _et al._ (2004); Torosov _et al._ (2011); Rong _et al._ (2015), pulse-shape engineering Steffen and Koch (2007); Barnes and Sarma (2012); Daems _et al._ (2013), and other optimizations Glaser _et al._ (2015); Caneva _et al._ (2009); Guérin _et al._ (2011); Hegerfeldt (2013); Garon _et al._ (2013); Van Damme _et al._ (2017); Arenz _et al._ (2017). A most straightforward approach to transit less dynamics obeys the adiabatic theorem by tuning the time-dependent parameter sufficiently slow. However, prolonged operation time destructs the quantum information by induced decoherence, affecting information processing efficiency. The concept of shortcuts to adiabaticity (STA) Guéry-Odelin _et al._ (2019); Torrontegui _et al._ (2013) is proposed, which combines the advantages of both adiabatic passages and resonant pulses. It breaks the adiabatic regime by various techniques, including inverse engineering Chen _et al._ (2010), counter-diabatic driving Deffner _et al._ (2014); An _et al._ (2016), fast- forward scaling Masuda and Nakamura (2010); Masuda (2012), which has been well developed over the past decade. Specifically, inverse engineering emanates from the Lewis-Riesenfeld theory, allowing superadiabatic state evolution on dynamical modes with boundary conditions. In addition, inverse engineering leaves enough freedom to further allow other tasks such as, e.g., suppressing systematic errors by collaborating with optimal control theory Daems _et al._ (2013); Ruschhaupt _et al._ (2012); Lu _et al._ (2013), dynamical decoupling techniques Munuera-Javaloy _et al._ (2020), and machine learning methods Zahedinejad _et al._ (2016); Liu _et al._ (2019); Ding _et al._ (2020). However, invariant-based STA requires continuously tunable parameters, limiting the genre of quantum control as analog-only. We consider a more complicated task: designing digital pulses instead of an analog controller with the same output and similar features. In this manner, we would deliver a framework that can be naturally integrated in current quantum computing paradigms based on the application of several digital quantum gates. We look for the optimal digital pulses design, which is similar to invariant- based STA for realizing robust quantum control. The optimal design is indeed a combinational optimization problem, being equivalent to dynamic programming, which is no longer analytically solvable. As artificial intelligence approach, Reinforcement Learning is a well-known tool for system control Sutton and Barto (2018), and deep learning has been developed for conquering complicated tasks in many areas Mnih _et al._ (2015, 2013); Silver _et al._ (2016, 2017), later applied in studying physics Carleo and Troyer (2017); Nagy and Savona (2019); Hartmann and Carleo (2019); Vicentini _et al._ (2019); Yoshioka and Hamazaki (2019); Iten _et al._ (2020). The framework of deep learning can be combined with reinforcement learning, searching control pulses for quantum state preparation Henson _et al._ (2018); Zhang _et al._ (2019), gate operation An and Zhou (2019), and quantum Szilard engine Sørdal and Bergli (2019). Since recent researches have employed Deep Reinforcement Learning (DRL) for quantum control Bukov _et al._ (2018); Porotti _et al._ (2019); Niu _et al._ (2019); Zhang _et al._ (2018); Wu _et al._ (2019); Wang _et al._ (2020), we are inspired to investigate the connection between DRL and STA. An optimistic expectation is that one can extend STA’s concept, introducing DRL as a new technique if it learns the features of STA protocols. In this paper, we present an experimental demonstration of a robust and high- precision quantum control task based on the deep reinforcement learning method on a trapped ${}^{171}\mathrm{Yb}^{+}$ ion. To be more specific, we train an Agent in a computer through DRL to achieve a single qubit X gate with time prior information bounded by STA. The multi-pulses control sequences produced by the DRL model is more robust than the standard $\pi$ pulse method (interacting with a constant amplitude for a period of time) with constant Rabi frequency in the presence of system noise. Besides, the robustness against both Rabi and detuning errors at the same time by DRL sequences is also verified. To demonstrate the application in the real laboratory noise environment, we examine the DRL models in the Zeeman energy level of the ion, which is sensitive to magnetic field noise. The results show that these DRL models can combat real system noises. Figure 1: (color online) Experimental sequences and model wave-forms. (a) Optimized detuning with time under STA method. The time is normalized to $[0,1]$. (b) Optimized detuning with time under DRL method. The time is normalized to $[0,1]$. (c) Energy level of ${}^{171}\textrm{Yb}^{+}$ ion. (d) Evolution of state in Bloch sphere under the driving of DRL model. Red solid line represents the trajectory optimized for $\Omega$ errors while blue solid line represents the trajectory optimized for $\Delta$ errors. Hollow circle and hollow triangle represent the state at the end of each driving step. (e) Experimental sequences in DRL model. After laser cooling and pumping, the ion is initialized to $|0\rangle$ state. Then a 20-steps microwave which contains DRL driving information is transmitted to the ion. finally a detecting laser is used to detect the probability in $|1\rangle$ state of the ion. ## II THEORETICAL MODELS Consider the coherent manipulation of a single qubit, whose Hamiltonian reads $H=\frac{\hbar}{2}\left[\Omega\sigma_{x}+\Delta(t)\sigma_{z}\right],$ (1) where the Rabi frequency $\Omega$ is fixed, while the detuning $\Delta(t)$ is time-varying. To achieve a robust qubit flipping from $|0\rangle$ to $|1\rangle$, a standard $\pi$ pulse, which corresponds to the Hamiltonian $\frac{\hbar}{2}\Omega\sigma_{x}$, is convenient and adequate. However, this operation is sensitive to systematic noise and decoherence. The invariant-based STA suggests that, one can achieve nonadiabatic quantum control of high fidelity and robustness by designed protocols, which satisfy the auxiliary equations derived from Lewis-Riesenfeld (LR) invariant. The LR invariant of a two-level system is constructed by $I(t)=\frac{\hbar}{2}\Omega_{0}\sum_{\pm}|\psi_{\pm}(t)\rangle\langle\psi_{\pm}(t)|$, where the eigenstates are $|\psi_{+}(t)\rangle=\left(\cos\frac{\theta}{2}e^{-i\frac{\beta}{2}},\sin\frac{\theta}{2}e^{i\frac{\beta}{2}}\right)^{\text{T}}$ and $|\psi_{-}(t)\rangle=\left(\sin\frac{\theta}{2}e^{-i\frac{\beta}{2}},-\cos\frac{\theta}{2}e^{i\frac{\beta}{2}}\right)^{\text{T}}$. The dynamics of the Hamiltonian is governed by time-dependent Schrödinger’s equation, whose solution is in superposition of these eigenstates as $|\Psi(t)\rangle=\sum_{\pm}c_{\pm}\exp(i\gamma_{\pm})|\psi_{\pm}(t)\rangle$, with LR phase calculated as $\gamma_{\pm}=\pm\frac{1}{2}\int_{0}^{t}\left(\frac{\dot{\theta}\cot\beta}{\sin\theta}\right)dt^{\prime}.$ (2) According to the condition for invariant $dI(t)/dt=\partial I(t)/\partial t+(1/i\hbar)[I(t),H(t)]=0$, we have the auxiliary equations $\displaystyle\dot{\theta}$ $\displaystyle=$ $\displaystyle-\Omega\sin\beta,$ (3) $\displaystyle\dot{\beta}$ $\displaystyle=$ $\displaystyle-\Omega\cot\theta\cos\beta+\Delta(t),$ (4) describing the state evolution along the dynamical modes with angular parameters $\theta$ and $\beta$, which characterize the trajectory on the Bloch sphere. As proposed in Ref. Ding _et al._ (2020), the framework can be applied to design robust quantum control, e.g., qubit flipping, against systematic errors with an adequate ansatz of free parameter $a$, such that $\theta(t)=\frac{\Omega T}{a}\left[as-\frac{\pi^{2}}{2}(1-s)^{2}+\frac{\pi^{3}}{3}(1-s)^{3}+\cos(\pi s)+A\right],$ (5) where $T=-\pi a/[(2-a-\pi^{2}/6)\Omega]$, $s=t/T$, and $A=\pi^{2}/6-1$ determined by boundary conditions $\theta(0)=0,~{}\dot{\theta}(0)=\Omega,~{}\ddot{\theta}(0)=0$ and $\theta(T)=\pi,~{}\dot{\theta}(T)=\Omega,~{}\ddot{\theta}(T)=0$. Specifically, one can nullify the probability of the first-order transition $P=\frac{\hbar^{2}}{4}\left|\int_{0}^{T}\langle\Psi_{-}(t)|\left(\delta_{\Omega}\Omega\sigma_{x}+\delta_{\Delta}\sigma_{z}\right)|\Psi_{+}(t)\rangle\right|^{2},$ (6) which yields the condition for error cancellation $\left|\int_{0}^{T}dte^{i2\gamma_{+}(t)}\left(\delta_{\Delta}\sin\theta-i2\delta_{\Omega}\dot{\theta}\sin^{2}\theta\right)\right|=0,$ (7) where systematic errors are characterized by $\Delta(t)\rightarrow\Delta(t)+\delta_{\Delta}$ and $\Omega\rightarrow\Omega(1+\delta_{\Omega})$, resulting in the configuration $a=0.604$ and $0.728$ for eliminating $\Delta$ and $\Omega$-error, respectively. Indeed, smooth detuning pulse $\Delta(t)$ as analog control of single-component is inversely engineered by substituting the ansatz into the following expression $\Delta(t)=-\frac{\ddot{\theta}}{\Omega\sqrt{1-\left(\frac{\dot{\theta}}{\Omega}\right)^{2}}}+\Omega\cot\sqrt{1-\left(\frac{\dot{\theta}}{\Omega}\right)^{2}}.$ (8) which is derived from combining auxiliary equations. The wave-forms of $\Delta(t)$ optimized for different systematic errors in STA are shown in fig. 1(a) and the maximum detuning $\Delta_{\max}$ for $\Delta$ and $\Omega$ errors are $1.5\Omega$ and $1.7\Omega$, respectively. Concerning our physical realization in trapped ions, the Rabi frequency $\Omega=(2\pi)3.3$ kHz is fixed, where we calculate the corresponding operation time for robust qubit flipping against $\Delta$ and $\Omega$-errors as $T_{\Delta}=364$ $\mu$s and $T_{\Omega}=293$ $\mu$s. Since an analog quantum control can be derived from the STA framework, it is more challenging to consider the digital quantum control of Landau-Zener problem. The problem is reformulated to the following expression: how should we manipulate a quantum system for a certain target with a step controller of $N$ intervals within a fixed time? The combinational optimization problem is equivalent to dynamic programming, i.e., a multi-step decision problem whose complexity grows exponentially with step number, allowing an approximation solution by artificial neural networks (ANN) or other universal function approximators; the use of deep ANN architectures with many layers leads to the concept of deep learning, and this, in turn, to DRL. In the framework of DRL, one assumes that there exists an unknown global optimal policy $\pi$ for a task, which gives an action $\textbf{a}(t_{i})$ once observing an arbitrary state $\textbf{s}(t_{i})$ at time $t_{i}$. The state-action relation $\pi(\textbf{s}|\textbf{a})$ is approximated by an Agent ANN, containing propagation of information between layers and nonlinear activation of neurons, whose parameters are tuned by optimizing algorithms for maximizing the accumulated reward. Details about the implementation of deep reinforcement learning can be found in supplementary materials. In our numerical experiments, the tunable range of detuning $[-\Delta_{\max},\Delta_{\max}]$ is renormalized into $\tilde{\Delta}\in[0,1]$ with $\Delta_{\max}$ being the maximal reachable value of $\Delta(t)$ in STA, which is the output of ANN as the encoded action at time step $t_{i}$: $\tilde{\Delta}(t_{i})=[\Delta(t_{i})+\Delta_{\max}]/2\Delta_{\max}$. Information of the two-level system, specifically, the expectation of spin on Z direction $\langle\sigma_{z}\rangle$, the renormalized detuning $\tilde{\Delta}(t_{i-1})$ that drives the system to the current state, and the system time $i/N$, are fed to the input layer of the ANN. The quantum dynamics are simulated by Liouville-von Neumann equation, which can be generalized to the Lindblad master equation for taking quantum noises into consideration. While network configuration, hyperparameters, and training details are explained in the literature Ding _et al._ (2020), we introduce the reward functions that we artificially design, which are similar to invariant-based STA that chooses an ansatz for obtaining quantum control. For converging the Agent to robust control of LZ-type, we firstly pre-train the Agent with $r(t_{i})=-|\tilde{\Delta}(t_{i})-\frac{i-1}{N-1}|$, punishing the deviations from linear growth of detuning, later rewarding a constant if $\langle\sigma_{z}\rangle>0.997$ at the final time step for fine-tuning under random systematic errors. For evaluating the DRL-inspired robust quantum control, we perform two numerical experiments as follows: (i) We set the operation time as $T_{\Delta}=364$ $\mu$s and $T_{\Omega}=293$ $\mu$s, being split uniformly by 20 pulses as the only hint from STA. The digital wave-forms output from our DRL model optimized for different systematic errors are shown in fig. 1(b). We emphasize that the STA framework clarifies the upper bound of robustness in Landau-Zener problems, which could be employed for benchmarking the capability of the Agent, as an artificial intelligence approach to digital quantum control with the alike feature. (ii) The operation time is arbitrarily set to be $T=300$ $\mu$s for checking if the Agent can explore desired protocols against hybrid systematic errors without any field knowledge of STA. We clarify that DRL is more general for this task since invariant-based STA no longer eliminates the hybrid errors perfectly but on certain proportion of $\delta_{\Delta}$ and $\delta_{\Omega}$ instead. All wave-forms used in real experiments are from these two numerical experiment models. ## III EXPERIMENTAL REALIZATION Our experiments are performed on a ${}^{171}{\rm Yb}^{+}$ ion trapped in a harmonic Paul trap, with the simplified structure being described in detail in supplementary materials. As shown in fig. 1(c), the two level system (TLS) is encoded in the ${}^{2}{\rm S}_{1/2}$ ground state of the ion, with $\left|0\right\rangle=\left|{}^{2}{\rm S}_{1/2},F=0,m_{F}=0\right\rangle$ and $\left|1\right\rangle=\left|{}^{2}{\rm S}_{1/2},F=1,m_{F}=0\right\rangle$. The difference of energy level $\left|0\right\rangle$ and $\left|1\right\rangle$ is about $\omega_{01}=12.6428$ GHz. The microwaves used to drive the TLS are generated through mixing method. More specifically, a microwave around 12.4428 GHz generated from signal generator (Agilent E8257D) is mixed with a 200 MHz microwave signal which is generated from a arbitrary waveform generator (AWG) and is used to modulate the microwave. After a high pass filter (HPF), this signal will be amplified to about 10 W and then transmitted to the ion with a microwave horn (Cui _et al._ , 2016). Our trap device is shielded with a 1.5 mm thick single layer Mu-metal (Farolfi _et al._ , 2019), making the final coherence time about 200 ms for $\left|0\right\rangle\leftrightarrow\left|1\right\rangle$ transition, which is characterized by Ramsey experiments. In each cycle, the experiment takes the following process: after 1 ms Doppler cooling, the state of the ion is initialized to $\left|0\right\rangle$ state through 20 $\mu$s optical pumping with $99.5\%$ fidelity. The wave-form output from DRL model is transformed into driving microwave through modulating the detuning, which is shown in fig. 1(e). Then the driving microwave is transmitted to the ion to drive the TLS. finally, a NA (numerical aperture) = 0.4 objective is used for state dependent fluorescence detection to determine the probability in state $\left|1\right\rangle$. In all of our experiments we set the Rabi frequency to $\Omega=(2\pi)$ 3.3 kHz, that is to say, the corresponding $2\pi$ time is about 300 $\mu$s. Figure 2: (color online) Noise robustness comparison of $\pi$ pulse, STA and DRL methods in single-qubit X gate task. (a) and (b) The performance of three control methods under different $\Omega$ and $\Delta$ errors respectively. The DRL method is as robust as STA method in most cases, except in big $\Omega$ and $\Delta$ errors. But they are all more robust than $\pi$ pulse in both kinds of errors. (c) and (d) The performance of feedback DRL. The feedback DRL agrees well with theoretical DRL in both $\Omega$ and $\Delta$ errors, which indicates our DRL model is robust to the disturbance of control pulses. The error bars indicate the standard deviation, and each data point is averaged over 2000 realizations. To verify the robustness of the DRL control method against systematic errors, we compare the performance of STA, DRL, and standard $\pi$ pulse method in the single-qubit X gate task under different $\Delta$ and $\Omega$ errors. The DRL models are pre-trained according to the time preliminary information provided by STA methods optimized in $\Delta$ and $\Omega$ errors, respectively. The state evolution under STA driving in Bloch sphere is shown in fig. 1(d). As shown in fig. 2 (a) and 2 (b), the DRL method performs as well as STA in most cases, in addition to the case that $\Omega$ error or $\Delta$ error is too large. Meanwhile, they are all more robust than the $\pi$ pulse method under system errors. To further explore our DRL model’s robustness, we also perform a feedback DRL experiment. In this experiment, 19 cycles are carried out. In cycle $n$, where $n$ belongs to [1,19], we measure the experimental result after $n$ control pulses, and feedback this result to the DRL model to obtain the next control pulses. After 19 cycles, we get the final 20 control pulses, and these pulses are only a little different from theoretical DRL pulses. As shown in fig. 2 (c) and 2 (d), the experimental feedback DRL results agree well with theoretical DRL, which means that our DRL model is robust to the disturbance of control pulses. Figure 3: (color online) The performance of $\pi$ pulse and DRL model under hybrid errors. The $\pi$ pulse method performs a little better than DRL in the case of almost no errors, because the pulses of DRL are more complex that it is easy to accumulate operation errors. However, with the increase of hybrid errors, the performance of DRL model is much better than $\pi$ pulse control. Then we examine the DRL model under $\Delta$ and $\Omega$ hybrid errors. It is worthwhile to mention that we set operation time and tunable range of detuning without any knowledge from STA when pre-training the DRL model locally. The performance of $\pi$ pulse and DRL method under hybrid errors is shown in fig. 3, in which the probabilities are taken logarithm to better distinguish the difference between these two methods The DRL method is more likely to accumulate errors than $\pi$ pulse due to the multi pulses driving operation on the one hand, on the other hand we just stop our training once $\langle\sigma_{z}\rangle>0.997$, which can be further improved theoretically. As we can expect, the $\pi$ pulse method performs a little better than DRL in the case of almost no errors. Nevertheless, with the increase of hybrid errors, the DRL method’s performance is much better than $\pi$ pulse in most cases, which is essential in precise quantum manipulation. Besides, we also examine the DRL model in the Zeeman energy level of the ion with $\left|1\right\rangle_{\textrm{z}}=\left|{}^{2}{\rm S}_{1/2},F=1,m_{F}=1\right\rangle$. The Zeeman energy level is first-order sensitive to the disturbance of the magnetic field, which could induce the realistic laboratory noise into TLS, and the corresponding coherence time is about 0.35 ms for $\left|0\right\rangle\leftrightarrow\left|1\right\rangle_{\textrm{z}}$ transition. The experimental results demonstrate that DRL’s performance is a little worse than theoretical expectation both in $\Omega$ and $\Delta$ errors due to extra decoherence, which is shown in fig. 4 (a) and 4 (b). We also compare the performance of $\pi$ pulse and DRL method in the single-qubit X gate under only the magnetic field noise with different Rabi time and different number of $\pi$ flips. As shown in fig. 4 (c) and 4 (d), the final probability decreases rapidly with the Rabi time and number of $\pi$ flips in $\pi$ pulse method owing to inevitable decoherence. However, the DRL method is more robust with the increase of Rabi time and number of $\pi$ flips, which is important in noisy quantum information processing. Figure 4: (color online) Noise-resilient feature of DRL method in Zeeman energy level. (a) and (b) The performance of DRL in Zeeman energy level. The DRL performs a little worse than theoretical values both in $\Omega$ and $\Delta$ errors due to disturbance of laboratory noise. (c) The comparison of $\pi$ pulse and DRL under different Rabi time. With the increase of Rabi time, the performance of $\pi$ pulse decreases rapidly while the DRL is more robust against decoherence. (d) The comparison of $\pi$ pulse and DRL under different number of $\pi$ flips. With the increase of $\pi$ flips, the performance of $\pi$ pulse decreases rapidly while the DRL is more robust against decoherence. The error bars indicate the standard deviation, and each data point is averaged over 2000 realizations ## IV CONCLUSION In summary, we experimentally demonstrate a robust quantum control task based on deep reinforcement learning. The DRL model’s multi-pulse control sequences are more robust than $\pi$ pulse in the presence of systematic errors. We also verify that robustness against both Rabi and detuning errors simultaneously can be achieved by DRL without any input from STA. In addition, we confirm that these DRL models can be significant in the real laboratory environment, which will lead to a promising enhancement in quantum information processing. ###### Acknowledgements. This work was supported by the National Key Research and Development Program of China (Nos. 2017YFA0304100, 2016YFA0302700), the National Natural Science Foundation of China (Nos. 11874343, 61327901, 11774335, 11474270, 11734015, 11874343), Key Research Program of Frontier Sciences, CAS (No. QYZDY-SSW- SLH003), the Fundamental Research Funds for the Central Universities (Nos. WK2470000026, WK2470000018), An-hui Initiative in Quantum Information Technologies (AHY020100, AHY070000), the National Program for Support of Topnotch Young Professionals (Grant No. BB2470000005). The theoretical part of the work is also partially supported from NSFC (12075145), STCSM (2019SHZDZX01-ZX04, 18010500400 and 18ZR1415500), Program for Eastern Scholar, HiQ funding for developing STA (YBN2019115204), QMiCS (820505) and OpenSuperQ (820363) of the EU Flagship on Quantum Technologies, Spanish Government PGC2018-095113-B-I00 (MCIU/AEI/FEDER, UE), Basque Government IT986-16, EU FET Open Grant Quromorphic (828826) as well as EPIQUS (899368). X. C. acknowledges Ramón y Cajal program (RYC-2017-22482). J. C. acknowledges the Ramón y Cajal program (RYC2018-025197-I) and the EUR2020-112117 project of the Spanish MICINN, as well as support from the UPV/EHU through the grant EHUrOPE. ## References * Nielsen and Chuang (2010) M. A. Nielsen and I. Chuang, _Quantum computation and quantum information_ (Cambridge University Press, 2010). * Král _et al._ (2007) P. Král, I. Thanopulos, and M. Shapiro, Reviews of modern physics 79, 53 (2007). * Remizov _et al._ (2015) S. V. Remizov, D. S. Shapiro, and A. N. Rubtsov, Physical Review A 92, 053814 (2015). * Brown _et al._ (2004) K. R. Brown, A. W. Harrow, and I. L. Chuang, Physical Review A 70, 052318 (2004). * Torosov _et al._ (2011) B. T. Torosov, S. Guérin, and N. V. Vitanov, Physical Review Letters 106, 233001 (2011). * Rong _et al._ (2015) X. Rong, J. Geng, F. Shi, Y. Liu, K. Xu, W. Ma, F. Kong, Z. Jiang, Y. Wu, and J. Du, Nature communications 6, 1 (2015). * Steffen and Koch (2007) M. Steffen and R. H. Koch, Physical Review A 75, 062326 (2007). * Barnes and Sarma (2012) E. Barnes and S. D. Sarma, Physical review letters 109, 060401 (2012). * Daems _et al._ (2013) D. Daems, A. Ruschhaupt, D. Sugny, and S. Guerin, Physical Review Letters 111, 050404 (2013). * Glaser _et al._ (2015) S. J. Glaser, U. Boscain, T. Calarco, C. P. Koch, W. Köckenberger, R. Kosloff, I. Kuprov, B. Luy, S. Schirmer, T. Schulte-Herbrüggen, _et al._ , The European Physical Journal D 69, 1 (2015). * Caneva _et al._ (2009) T. Caneva, M. Murphy, T. Calarco, R. Fazio, S. Montangero, V. Giovannetti, and G. E. Santoro, Physical review letters 103, 240501 (2009). * Guérin _et al._ (2011) S. Guérin, V. Hakobyan, and H. Jauslin, Physical Review A 84, 013423 (2011). * Hegerfeldt (2013) G. C. Hegerfeldt, Physical review letters 111, 260501 (2013). * Garon _et al._ (2013) A. Garon, S. Glaser, and D. Sugny, Physical Review A 88, 043422 (2013). * Van Damme _et al._ (2017) L. Van Damme, Q. Ansel, S. Glaser, and D. Sugny, Physical Review A 95, 063403 (2017). * Arenz _et al._ (2017) C. Arenz, B. Russell, D. Burgarth, and H. Rabitz, New Journal of Physics 19, 103015 (2017). * Guéry-Odelin _et al._ (2019) D. Guéry-Odelin, A. Ruschhaupt, A. Kiely, E. Torrontegui, S. Martínez-Garaot, and J. G. Muga, Reviews of Modern Physics 91, 045001 (2019). * Torrontegui _et al._ (2013) E. Torrontegui, S. Ibánez, S. Martínez-Garaot, M. Modugno, A. del Campo, D. Guéry-Odelin, A. Ruschhaupt, X. Chen, and J. G. Muga, in _Advances in atomic, molecular, and optical physics_ , Vol. 62 (Elsevier, 2013) pp. 117–169. * Chen _et al._ (2010) X. Chen, A. Ruschhaupt, S. Schmidt, A. del Campo, D. Guéry-Odelin, and J. G. Muga, Physical review letters 104, 063002 (2010). * Deffner _et al._ (2014) S. Deffner, C. Jarzynski, and A. del Campo, Physical Review X 4, 021013 (2014). * An _et al._ (2016) S. An, D. Lv, A. Del Campo, and K. Kim, Nature communications 7, 12999 (2016). * Masuda and Nakamura (2010) S. Masuda and K. Nakamura, Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences 466, 1135 (2010). * Masuda (2012) S. Masuda, Physical Review A 86, 063624 (2012). * Ruschhaupt _et al._ (2012) A. Ruschhaupt, X. Chen, D. Alonso, and J. Muga, New Journal of Physics 14, 093040 (2012). * Lu _et al._ (2013) X.-J. Lu, X. Chen, A. Ruschhaupt, D. Alonso, S. Guerin, and J. G. Muga, Physical Review A 88, 033406 (2013). * Munuera-Javaloy _et al._ (2020) C. Munuera-Javaloy, Y. Ban, X. Chen, and J. Casanova, arXiv preprint arXiv:2007.15394 (2020). * Zahedinejad _et al._ (2016) E. Zahedinejad, J. Ghosh, and B. C. Sanders, Physical Review Applied 6, 054005 (2016). * Liu _et al._ (2019) B.-J. Liu, X.-K. Song, Z.-Y. Xue, X. Wang, and M.-H. Yung, Physical Review Letters 123, 100501 (2019). * Ding _et al._ (2020) Y. Ding, Y. Ban, J. D. Martín-Guerrero, E. Solano, J. Casanova, and X. Chen, arXiv preprint arXiv:2009.04297 (2020). * Sutton and Barto (2018) R. S. Sutton and A. G. Barto, _Reinforcement Learning: An Introduction_ , 2nd ed. (The MIT Press, 2018). * Mnih _et al._ (2015) V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G. Bellemare, A. Graves, M. Riedmiller, A. K. Fidjeland, G. Ostrovski, _et al._ , nature 518, 529 (2015). * Mnih _et al._ (2013) V. Mnih, K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, D. Wierstra, and M. Riedmiller, arXiv preprint arXiv:1312.5602 (2013). * Silver _et al._ (2016) D. Silver, A. Huang, C. J. Maddison, A. Guez, L. Sifre, G. Van Den Driessche, J. Schrittwieser, I. Antonoglou, V. Panneershelvam, M. Lanctot, _et al._ , nature 529, 484 (2016). * Silver _et al._ (2017) D. Silver, T. Hubert, J. Schrittwieser, I. Antonoglou, M. Lai, A. Guez, M. Lanctot, L. Sifre, D. Kumaran, T. Graepel, _et al._ , arXiv preprint arXiv:1712.01815 (2017). * Carleo and Troyer (2017) G. Carleo and M. Troyer, Science 355, 602 (2017). * Nagy and Savona (2019) A. Nagy and V. Savona, Physical review letters 122, 250501 (2019). * Hartmann and Carleo (2019) M. J. Hartmann and G. Carleo, Physical review letters 122, 250502 (2019). * Vicentini _et al._ (2019) F. Vicentini, A. Biella, N. Regnault, and C. Ciuti, Physical review letters 122, 250503 (2019). * Yoshioka and Hamazaki (2019) N. Yoshioka and R. Hamazaki, Physical Review B 99, 214306 (2019). * Iten _et al._ (2020) R. Iten, T. Metger, H. Wilming, L. Del Rio, and R. Renner, Physical Review Letters 124, 010508 (2020). * Henson _et al._ (2018) B. M. Henson, D. K. Shin, K. F. Thomas, J. A. Ross, M. R. Hush, S. S. Hodgman, and A. G. Truscott, Proceedings of the National Academy of Sciences 115, 13216 (2018). * Zhang _et al._ (2019) X.-M. Zhang, Z. Wei, R. Asad, X.-C. Yang, and X. Wang, npj Quantum Information 5, 1 (2019). * An and Zhou (2019) Z. An and D. Zhou, EPL (Europhysics Letters) 126, 60002 (2019). * Sørdal and Bergli (2019) V. B. Sørdal and J. Bergli, Physical Review A 100, 042314 (2019). * Bukov _et al._ (2018) M. Bukov, A. G. Day, D. Sels, P. Weinberg, A. Polkovnikov, and P. Mehta, Physical Review X 8, 031086 (2018). * Porotti _et al._ (2019) R. Porotti, D. Tamascelli, M. Restelli, and E. Prati, Communications Physics 2, 1 (2019). * Niu _et al._ (2019) M. Y. Niu, S. Boixo, V. N. Smelyanskiy, and H. Neven, npj Quantum Information 5, 1 (2019). * Zhang _et al._ (2018) X.-M. Zhang, Z.-W. Cui, X. Wang, and M.-H. Yung, Physical Review A 97, 052333 (2018). * Wu _et al._ (2019) R.-B. Wu, H. Ding, D. Dong, and X. Wang, Physical Review A 99, 042327 (2019). * Wang _et al._ (2020) Z. T. Wang, Y. Ashida, and M. Ueda, Physical Review Letters 125, 100401 (2020). * Cui _et al._ (2016) J.-M. Cui, Y.-F. Huang, Z. Wang, D.-Y. Cao, J. Wang, W.-M. Lv, L. Luo, A. Del Campo, Y.-J. Han, C.-F. Li, _et al._ , Scientific reports 6, 33381 (2016). * Farolfi _et al._ (2019) A. Farolfi, D. Trypogeorgos, G. Colzi, E. Fava, G. Lamporesi, and G. Ferrari, Review of Scientific Instruments 90, 115114 (2019). Supplemental Material: Experimentally Realizing Efficient Quantum Control with Reinforcement Learning ## I EXPERIMENTAL PLATFORM AND WAVEFORM OF THE DRIVING MICROWAVE The type of platform used in our experiments is needle trap. As shown in Fig. S1, the needle trap consists of 6 needles. Two opposite needles are connected to radio frequency (RF) potential to trap the ion and the others are connected to direct current (DC) potential to fine tuning the position of ion. The size of the needle trap depends mainly on the distance between the two needles tips near the trap center, which is set to 180 $\mu$m in our experiment. The trap is installed in an ultrahigh vacuum below $10^{-11}$ torr, and a helical resonator provides the RF signal with frequency 24 MHz and amplitude of 180 V to the trap. Ion fluorescence is collected by an objective lens with 0.4 numerical aperture, and detected by a photo-multiplier tube (PMT). The total fluorescence detection efficiency is about 2$\%$. We generate required waveform of the microwave field through setting the waveform of AWG for modulation. The carrier microwave $B_{c}(t)=A_{1}\mathrm{sin}(\omega_{c}t)$, where $A_{1}$ is amplitude and $f_{c}=\omega_{c}/2\pi=12.4$ GHz is the frequency. The waveform generated by AWG for modulation is $I(t)=A_{2}\mathrm{sin}(\phi(t))$. After mixing, the microwave field will be $B(t)=\frac{A_{1}A_{2}}{2}(\mathrm{sin}(\omega_{c}t+\phi(t))+\mathrm{sin}(\omega_{c}t-\phi(t)))$, where the phase function $\phi(t)$ can be expressed in a piece-wise function for the microwave composed of 20 steps in our DRL experiments. With the qubit resonance frequency $f_{0}=\omega_{0}/2\pi=12.6$ GHz, we filter out the low frequency components of the microwave through a high pass filter. In our experiments, we only adjust $\Delta(t)$ with discrete steps by changing phase $\phi(t)$ as follows: $\phi(t)=\begin{cases}(\omega_{0}-\omega_{c})t+\Delta_{1}t,&(0,t_{1})\\\ (\omega_{0}-\omega_{c})t+\Delta_{2}t+\phi_{1},&(0,t_{2}-t_{1})\\\ (\omega_{0}-\omega_{c})t+\Delta_{3}t+\phi_{2},&(0,t_{3}-t_{2})\\\ \cdots\\\ (\omega_{0}-\omega_{c})t+\Delta_{20}t+\phi_{19},&(0,t_{20}-t_{19})\end{cases}$ (1) where $\Delta_{n}(n\in[1,20])$ is the step-wise detuning and $\phi_{1}=(\omega_{0}-\omega_{c})t_{1}+\Delta_{1}t_{1}$, $\phi_{2}=(\omega_{0}-\omega_{c})(t_{2}-t_{1})+\Delta_{2}(t_{2}-t_{1})+\phi_{1}$ and so on. Figure S1: (color online) Experimental setup. A single ${}^{171}\rm{Yb}^{+}$ ion is trapped in center of the needle trap. Two 369 nm and 935 nm lasers are used to cooling the ion and 369 nm laser is also used to detect the state of ion. The microwave used to drive the ion is generated through mixing method. The whole experimental sequences are controlled by a TTL sequences board based on Field Programmable Gate Array (FPGA). ## II QUBIT STATE PREPARATION AND MEASUREMENT In ion trap experiments, the state preparation and measurement cannot be perfect and there will always be some limitations. We characterize these errors as follows. The ion is prepared in $|0\rangle$ state through optical pumping and ideally, no photon should be detected as the ion is in dark state. However, due to the dark counts of the photon detector as well as photons scattered from the environment, we will collect some photons sometimes. Then we apply a $\pi$ pulse to flip the $|0\rangle$ state to $|1\rangle$ state and detect the fluorescence. Because the collection efficiency problem, no photon will be collected sometimes. The histograms of dark and bright state is shown in Fig. S2. The threshold is selected as 2 in our experiments. When the photon number is $>2$, the qubit is identified as bright state and the probability of being mistaken as dark state is $\epsilon_{D}$. By contrary, the probability of being mistaken as bright state when photon number is $\leq 2$ is $\epsilon_{B}$. The total error can be taken as $\epsilon=(\epsilon_{B}+\epsilon_{D})/2$. Figure S2: (color online) Histogram for photon counts in state preparation and detection experiments. The distribution of photon counts is shown when the qubit state is prepared in $|0\rangle$ (dark state) and $|1\rangle$ (bright state). ## III IMPLEMENTATION OF DEEP REINFORCEMENT LEARNING By combining reinforcement learning and deep learning, deep reinforcement learning (DRL) aims to solve decision-making problems, allowing a computational agent to make decisions from input data by trial. A mathematical model called Markov decision process describes the problem, where an agent at every time step $t$ observes a state $s_{t}$, takes an action $a_{t}$, receives a reward $r_{t}$ and transits to the state at the next time step $s_{t+1}$ according to the dynamics of the environment $P(s_{t+1}|s_{t},a_{t})$. The agent’s goal is to learn a policy $\pi(a|s)$ that maximizes the total reward $\sum_{i}\gamma^{i}r_{i}$, with $\gamma$ being the discount rate and $r_{i}$ being the scalar rewards. DRL employs an artificial neural network (ANN) as a general function approximator for the policy $\pi(a|s)$, leading to specialized algorithms for obtaining an optimal approximation. The simplified model framework can be found in Fig. S3. Figure S3: (color online) DRL framework for quantum control with qubit of one time step in training. The agent (DNN), consists of three hidden layers, observes a state from environment. After propagation between layers and nonlinear activations of the DNN nodes, output layer gives an action $a_{t}$. The environment rewards $r_{t}$ enable the agent to learn how to achieve the goal. Here we use Proximal Policy Optimization (PPO), which performs comparably state-of-the-art, remaining simplicity for implementation and tuning. It is worthwhile to mention that it is also the default RL algorithm at OpenAI. As an on-policy algorithm, PPO attempts to evaluate and improve the behavior policy that is used to make decisions. Its objective function is $L_{\text{clip}}(\theta)=\hat{E}_{t}\left\\{\min\left[r_{t}(\theta)\hat{A}_{t},\text{clip}\left(r_{t}(\theta),1+\epsilon,1-\epsilon\right)\hat{A}_{t}\right]\right\\},$ (2) , where $\theta$ is the policy parameter (the set that contains all weights and biases), $\hat{E}_{t}$ is the expectation over time steps, $r_{t}$ is the ratio of the probability under the new and old policies, $\hat{A}_{t}$ is the estimated advantage, and $\epsilon$ is the hyperparameter for bounding the clipping range. There is also a variant of PPO based on an adaptive Kullback- Leiber penalty, which controls the change of $\pi(a|s)$ at each iteration. A detailed explanation of PPO, as well as its pseudocodes, are already clearly presented in the original paper Schulman _et al._ (2017). Although there are arguments about the origin of performance enhancement from Trust Region Policy Optimization Schulman _et al._ (2015) (whether it is from the clipping or code-level tricks Engstrom _et al._ (2019)), we reckon these topics, including if PPO-like algorithms can be further optimized, go beyond the scope of this work. Thus, we implement a minimal PPO for our quantum control task. We use an open-source Python library, TensorForce (version 0.5.2) Schaarschmidt _et al._ (2017), for a quick implementation. The library is based on TensorFlow, a well-known framework for deep learning with GPU acceleration. The two-level system’s quantum dynamics in our training environment are numerically simulated by QuTiP (version 4.4.1) Johansson _et al._ (2012). We set a batch size of 16, and the learning rate is 1e-4 for both pre-training and fine-tuning. The ANN contains three hidden layers, where each of the layers consists of 32 fully-connected neurons activated by ReLU. Other hyperparameters and settings are the default configuration of the PPO Agent provided by TensorForce. Another evaluation environment can interact with the trapped ion system for verifying quantum control with feedback. Codes are compatible with both CPU and GPU versions of TensorFlow 1.13.1., which are available from the corresponding authors upon reasonable request. ## References * Schulman _et al._ (2017) J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, arXiv preprint arXiv:1707.06347 (2017). * Schulman _et al._ (2015) J. Schulman, S. Levine, P. Abbeel, M. Jordan, and P. Moritz, in _International conference on machine learning_ (2015) pp. 1889–1897. * Engstrom _et al._ (2019) L. Engstrom, A. Ilyas, S. Santurkar, D. Tsipras, F. Janoos, L. Rudolph, and A. Madry, in _International Conference on Learning Representations_ (2019). * Schaarschmidt _et al._ (2017) M. Schaarschmidt, A. Kuhnle, and K. Fricke, https:// github.com/tensorforce/tensorforce (2017). * Johansson _et al._ (2012) J. R. Johansson, P. D. Nation, and F. Nori, Computer Physics Communications 183, 1760 (2012).
# Online Packing to Minimize Area or Perimeter Mikkel Abrahamsen Lorenzo Beretta∗ Basic Algorithms Research Copenhagen (BARC), University of Copenhagen. BARC is supported by the VILLUM Foundation grant 16582. Lorenzo Beretta received funding from the European Union’s Horizon 2020 research and innovation program under the Marie Skłodowska-Curie grant agreement No. 801199. (January 21, 2021) ###### Abstract We consider online packing problems where we get a stream of axis-parallel rectangles. The rectangles have to be placed in the plane without overlapping, and each rectangle must be placed without knowing the subsequent rectangles. The goal is to minimize the perimeter or the area of the axis-parallel bounding box of the rectangles. We either allow rotations by $90^{\circ}$ or translations only. For the perimeter version we give algorithms with an absolute competitive ratio slightly less than $4$ when only translations are allowed and when rotations are also allowed. We then turn our attention to minimizing the area and show that the asymptotic competitive ratio of any algorithm is at least $\Omega(\sqrt{n})$, where $n$ is the number of rectangles in the stream, and this holds with and without rotations. We then present algorithms that match this bound in both cases and the competitive ratio is thus optimal to within a constant factor. We also show that the competitive ratio cannot be bounded as a function of Opt. We then consider two special cases. The first is when all the given rectangles have aspect ratios bounded by some constant. The particular variant where all the rectangles are squares and we want to minimize the area of the bounding square has been studied before and an algorithm with a competitive ratio of $8$ has been given [Fekete and Hoffmann, Algorithmica, 2017]. We improve the analysis of the algorithm and show that the ratio is at most $6$, which is tight. The second special case is when all edges have length at least $1$. Here, the $\Omega(\sqrt{n})$ lower bound still holds, and we turn our attention to lower bounds depending on Opt. We show that any algorithm for the translational case has an asymptotic competitive ratio of at least $\Omega(\sqrt{\textsc{Opt}})$. If rotations are allowed, we show a lower bound of $\Omega(\sqrt[4]{\textsc{Opt}})$. For both versions, we give algorithms that match the respective lower bounds: With translations only, this is just the algorithm from the general case with competitive ratio $O(\sqrt{n})=O(\sqrt{\textsc{Opt}})$. If rotations are allowed, we give an algorithm with competitive ratio $O(\min\\{\sqrt{n},\sqrt[4]{\textsc{Opt}}\\})$, thus matching both lower bounds simultaneously. ## 1 Introduction Problems related to packing appear in a plethora of big industries. For instance, two-dimensional versions of packing arise when a given set of pieces have to be cut out from a large piece of material so as to minimize waste. This is relevant to clothing production where cutting patterns are cut out from a roll of fabric, and similarly in leather, glass, wood, and sheet metal cutting. In some applications, it is important that the pieces are placed in an _online_ fashion. This means that the pieces arrive one by one and we need to decide the placement of one piece before we know the ones that will come in the future. This is in contrast to _offline_ problems, where all the pieces are known in advance. Problems related to packing were some of the first for which online algorithms were described and analyzed. Indeed, the first use of the terms “online” and “offline” in the context of approximation algorithms was in the early 1970s and used for algorithms for bin-packing problems [14]. In this paper, we study online packing problems where the pieces can be placed anywhere in the plane as long as they do not overlap. The goal is to minimize the region occupied by the pieces. The pieces are axis-parallel rectangles, and they may or may not be rotated by $90^{\circ}$. We want to minimize the size of the axis-parallel bounding box of the pieces, and the size of the box is either the perimeter or the area. This results in four problems: PerimeterRotation, PerimeterTranslation, AreaRotation, and AreaTranslation. #### Competitive analysis The _competitive ratio_ of an online algorithm is the equivalent of the _approximation ratio_ of an (offline) approximation algorithm. The usual definitions [7, 9, 11] of competitive ratio (or _worst case ratio_ , as it may also be called [11]) can only be used to describe that the cost of the solution produced by an online algorithm is at most some constant factor higher than the cost Opt of the optimal (offline) solution. In the study of approximation algorithms, it is often the case that the approximation ratio is described not just as a constant, but as a more general function of the input. In the same way, we generalize the definition of competitive ratios to support such statements about online algorithms. Consider an algorithm $A$ for one of the packing problems studied in this paper. Let $\mathcal{L}$ be the set of non-empty streams of rectangular pieces. For a stream $L\in\mathcal{L}$, we define $A(L)$ to be the cost of the packing produced by $A$ and let $\textsc{Opt}(L)$ be the cost of the optimal (offline) packing. We say that $A$ has an _absolute competitive ratio_ of $f(L)$, for some function $f:\mathcal{L}\longrightarrow\mathbb{R}^{+}$ which may just be a constant, if $\sup_{L\in\mathcal{L}}\frac{A(L)}{\textsc{Opt}(L)f(L)}\leq 1.$ We say that $A$ has an _asymptotic competitive ratio_ of $f(L)$ if $\limsup_{c\longrightarrow\infty}\left(\sup\left\\{\frac{A(L)}{\textsc{Opt}(L)f(L)}\mid L\in\mathcal{L}\text{ and }\textsc{Opt}(L)=c\right\\}\right)\leq 1.$ In this paper, the functions $f(L)$ that we consider will be (i) constants, (ii) functions of the number of pieces $n=|L|$, (iii) functions of $\textsc{Opt}(L)$. By definition, if $A$ has an absolute competitive ratio of $f(L)$, then $A$ also has an asymptotic competitive ratio of $f(L)$, but $A$ may also have a smaller asymptotic competitive ratio $g(L)<f(L)$. However, the following easy lemma shows that for the problems studied in this paper, any constant asymptotic competitive ratio can be matched to within an arbitrarily small difference by an absolute competitive ratio. ###### Lemma 1. For the problems studied in this paper, if an algorithm $A$ has an asymptotic competitive ratio of some constant $c>1$, then for every $\varepsilon>0$, there is an algorithm $A^{\prime}$ with absolute competitive ratio $c+\varepsilon$. It follows that any constant lower bound on the absolute competitive ratio is also a lower bound on the asymptotic competitive ratio. ###### Proof. Let $n>0$ be so large that when $\textsc{Opt}(L)\geq n$, we have $\frac{A(L)}{c\textsc{Opt}(L)}\leq 1+\varepsilon/c$. When the first piece $p$ of a stream $L$ is given, $A^{\prime}$ chooses a scale factor $\lambda>0$ big enough that when $p$ is scaled up by $\lambda$, the resulting piece $p^{\prime}:=\lambda p$ alone has cost $n$ (i.e., the area or the perimeter of $p^{\prime}$ is $n$). The algorithm $A^{\prime}$ now imitates the strategy of $A$ on the stream $\lambda L$ we get by scaling up all pieces of $L$ by $\lambda$. We then get that $\frac{A^{\prime}(L)}{(c+\varepsilon)\textsc{Opt}(L)}=\frac{A(\lambda L)}{(c+\varepsilon)\textsc{Opt}(\lambda L)}\leq\frac{(1+\varepsilon/c)c}{c+\varepsilon}=1.\qed$ For this reason, we do not distinguish between absolute and asymptotic competitive ratios when the ratio is a constant. Note that the argument does not work when the competitive ratio is a non-constant function of Opt. #### Results and structure of the paper We develop online algorithms for the perimeter versions PerimeterRotation and PerimeterTranslation, both with a competitive ratio slightly less than $4$. These algorithms are described in Section 2. The idea is to partition the positive quadrant into _bricks_ , which are axis-parallel rectangles with aspect ratio $\sqrt{2}$. In each brick, we build a stack of pieces which would be too large to place in a brick of smaller size. Online packing algorithms using higher-dimensional bricks were described by Januszewski and Lassak [15] and our algorithms are inspired by an algorithm of Fekete and Hoffmann [13] that we will get back to. Interestingly, we show in Section 2.2 that a more direct adaptation of the algorithm of Fekete and Hoffmann has a competitive ratio of at least $4$, and is thus inferior to the algorithm we describe. We also give a lower bound of $4/3$ for the version with translations and $5/4$ for the version with rotations. In Section 3, we study the area versions AreaRotation and AreaTranslation. We show in Section 3.1 that for any algorithm $A$ processing a stream of $n$ pieces cannot achieve a better competitive ratio than $\Omega(\sqrt{n})$, and this holds for all online algorithms and with and without rotations allowed. It also holds in the special case where all the edges of pieces have length at least $1$. We furthermore show that when the pieces can be arbitrary, there can be given no bound on the competitive ratio as a function of Opt for AreaRotation nor AreaTranslation. In Section 3.2 we describe the algorithms DynBoxTrans and DynBoxRot, which achieve a $O(\sqrt{n})$ competitive ratio for AreaTranslation and AreaRotation, respectively, for an arbitrary stream of $n$ pieces. This is thus optimal up to a constant factor when measuring the competitive ratio as a function of $n$. Both algorithms use a row of boxes of exponentially increasing width and dynamically adjusted height. In these boxes, we pack pieces using a next-fit shelf algorithm, which is a classic online strip packing algorithm first described by Baker and Schwartz [6]. We then turn our attention to two special cases. The first special case is when the aspect ratio is bounded by a constant $\alpha\geq 1$. A case of particular interest is when all pieces are squares, i.e., $\alpha=1$. It is natural to have the same requirement to the container as to the pieces, so let us assume that the goal is to minimize the area of the axis-parallel bounding square of the pieces, and call the problem SquareInSquareArea. This problem was studied by Fekete and Hoffmann [13], and they gave an algorithm for the problem and proved that it was $8$-competitive. We prove that the same algorithm is in fact $6$-competitive and that this is tight. It easily follows that if the aspect ratio is bounded by an arbitrary constant $\alpha\geq 1$ or if the goal is to minimize the area of the axis- parallel bounding rectangle, we also get a $O(1)$-competitive algorithm. The second special case is when all edges are _long_ , that is, when they have length at least $1$ (any other constant will work too). In Section 3.4, we show that under this assumption, there is a lower bound of $\Omega(\sqrt{\textsc{Opt}})$ for the asymptotic competitive ratio of AreaTranslation, whereas for AreaRotation, we get the lower bound $\Omega(\sqrt[4]{\textsc{Opt}})$. In Section 3.5, we provide algorithms for the area versions when the edges are long. For both problems AreaRotation and AreaTranslation, we give algorithms that match the lower bounds of Section 3.4 to within a constant factor. With translations only, this is just the algorithm from the general case with competitive ratio $O(\sqrt{n})=O(\sqrt{\textsc{Opt}})$. The algorithm with ratio $O(\sqrt[4]{\textsc{Opt}})$ for the rotational case follows the same scheme as the algorithms for arbitrary rectangles of Section 3.2, but differ in the way we dynamically increase boxes’ heights. We finally describe an algorithm for the rotational case with competitive ratio $O(\min\\{\sqrt{n},\sqrt[4]{\textsc{Opt}}\\})$, thus matching the lower bounds $\Omega(\sqrt{n})$ and $\Omega(\sqrt[4]{\textsc{Opt}})$ simultaneously. Actually, the two lower bounds for AreaRotation can be summarized by $\Omega(\max\\{\sqrt{n},\sqrt[4]{\textsc{Opt}}\\})$, while we manage to achieve a competitive ratio of $O(\min\\{\sqrt{n},\sqrt[4]{\textsc{Opt}}\\})$. However, this gives no contradiction, it simply proves that the _edge cases_ that have a competitive ratio of at least $\Omega(\sqrt[4]{\textsc{Opt}})$ must satisfy $\textsc{Opt}=O(n^{2})$, and those for which the competitive ratio is at least $\Omega(\sqrt{n})$ satisfy $n=O(\sqrt{\textsc{Opt}})$. We summarize the results in Table 1. Measure | Version | Trans./Rot. | Lower bound | Upper bound ---|---|---|---|--- Perimeter | General | Translation | $4/3$, Sec. 2.3 | $4-\varepsilon$, Sec. 2.1 Rotation | $5/4$, Sec. 2.3 | $4-\varepsilon$, Sec. 2.1 Area | General | Translation | $\Omega(\sqrt{n})$ & $\forall f:\Omega(f(\textsc{Opt}))$, Sec. 3.1 | $O(\sqrt{n})$, Sec. 3.2 Rotation | $\Omega(\sqrt{n})$ & $\forall f:\Omega(f(\textsc{Opt}))$, Sec. 3.1 | $O(\sqrt{n})$, Sec. 3.2 Sq.-in-sq. | N/A | 16/9, Sec. 3.3 | $6$, Sec. 3.3 Long edges | Translation | $\Omega(\sqrt{\textsc{Opt}})$, Sec. 3.4 | $O(\sqrt{n})=O(\sqrt{\textsc{Opt}})$, Sec. 3.5 Rotation | $\Omega(\max\\{\sqrt{n},\sqrt[4]{\textsc{Opt}}\\})$, Sec. 3.1 and 3.4 | $O(\min\\{\sqrt{n},\sqrt[4]{\textsc{Opt}}\\})$, Sec. 3.5 Table 1: Results of this paper. #### Related work The literature on online packing problems is rich. See the surveys of Christensen, Khan, Pokutta, and Tetali [9], van Stee [25, 26], and Csirik and Woeginger [11] for an overview. It seems that the vast majority of previous work on online versions of two-dimensional packing problems is concerned with either bin packing (packing the pieces into a minimum number of unit squares) or strip packing (packing the pieces into a strip of unit width so as to minimize the total height of the pieces). From a mathematical point of view, we find the problems studied in this paper perhaps even more fundamental than these important problems in the sense that we give no restrictions on where to place the pieces, whereas the pieces are restricted by the boundaries of the bins and the strip in bin and strip packing. Another related problem is to find the critical density of online packing squares into a square. In other words, what is the maximum $\Sigma\leq 1$ such that there is an online algorithm that packs any stream of squares of total area at most $\Sigma$ into the unit square? This was studied, among others, by Fekete and Hoffmann [13] and Brubach [8]. Lassak [16] and Januszewski and Lassak [15] studied higher-dimensional versions of this problem. Milenkovich [20] studied generalized offline versions of the minimum area problem: Translate $k$ given $m$-gons into a convex container of minimum area with edges in $n$ fixed directions. When the $m$-gons can be non-convex, the running time is $O((m^{2}+n)^{2k-2}(n+\log m))$, and when they are convex, the running times are $O((m+n)^{2k}(n+\log m))$ or $O(m^{k-1}(n^{2k+1}+\log m))$. Milenkovich and Daniels [22] described different algorithms for the same problems. Milenkovich [21] also studied the same problem when arbitrary rotations are allowed and the container is either a strip with a fixed width, a homothet of a given convex polygon, or an arbitrary rectangle (as in our work). He gave $(1+\varepsilon)$-approximation algorithms (no explicit running times are given, but they are apparently also exponential). Some algorithms have been described for computing the packing of two or three convex polygons that minimizes the perimeter or area of the convex hull or the bounding box [1, 5, 17, 23]. Alt [2] demonstrated how a $\rho$-approximation algorithm for strip packing (axis-parallel rectangles with translations) can be turned into a $(1+\varepsilon)\rho$-approximation algorithm for the offline version of AreaTranslation, for any constant $\varepsilon>0$. The same technique works for AreaRotation. The idea is to apply the strip packing algorithm to strips of increasing widths and in the end choose the packing that resulted in the smallest area. Therefore, the same technique cannot be applied in the online setting, where we need to choose a placement for each piece and stick with it. Alt also mentioned that finding a minimum area bounding box of a set of convex polygons with arbitrary rotations allowed can be reduced to the problem where the pieces are axis-parallel rectangles with only translations allowed. This reduction increases the approximation ratio by a factor by $2$. The reduction does not work when the pieces can be only translated, but Alt, de Berg, and Knauer [4] gave a $17.45$-approximation algorithm for this problem using different techniques. Lubachevsky and Graham [18] used computational experiments to find the rectangles of minimum area into which a given number $n\leq 5000$ of congruent circles can be packed; see also the follow-up work by Specht [24]. In another paper, Lubachevsky and Graham [19] studied the problem of minimizing the perimeter instead of the area. Another fundamental packing problem is to find the smallest square containing a given number of _unit_ squares, with arbitrary rotations allowed. A long line of mathematical research has been devoted to this problem, initiated by Erdős and Graham [12] in 1975, and it is still an active research area [10]. ## 2 The perimeter versions In Section 2.1, we present two online algorithms to minimize the perimeter of the bounding box: the algorithm BrickTranslation solves the problem PerimeterTranslation, where we can only translate pieces; the algorithm BrickRotation solves the problem PerimeterRotation, where also rotations are allowed. Both algorithms achieve a competitive ratio of $4$. In Section 2.3, we show a lower bound of $4/3$ for the version with translations and $5/4$ for the version with rotations. ### 2.1 Algorithms to minimize perimeter #### Algorithm for translations We pack the pieces into non-overlapping _bricks_ ; a technique first described by Januszewski and Lassak [15] which was also used by Fekete and Hoffmann [13] for the problem SquareInSquareArea. Let a _$k$ -brick_ be a rectangle of size $\sqrt{2}^{-k}\times\sqrt{2}^{-k-1}$ if $k$ is even and $\sqrt{2}^{-k-1}\times\sqrt{2}^{-k}$ if $k$ is odd. A _brick_ is a $k$-brick for some integer $k$. We tile the positive quadrant using one $k$-brick $B_{k}$ for each integer $k$ as in Figure 1 (left): if $k$ is even, $B_{k}$ is the $k$-brick with lower left corner $(0,\sqrt{2}^{-k-1})$ and otherwise, $B_{k}$ is the $k$-brick with lower left corner $(\sqrt{2}^{-k-1},0)$. The bricks $B_{k}$ are called the _fundamental_ bricks. We define $B_{>k}:=\bigcup_{i>k}B_{i}$ and $B_{\geq k}:=B_{>k-1}$, so that $B_{>k}$ is the $k$-brick immediately below (if $k$ is even) or to the left (if $k$ is odd) of $B_{k}$. An important property of a $k$-brick $B$ is that it can be split into two $(k+1)$-bricks: $B\dagger 1$ and $B\dagger 2$; see Figure 1 (middle). We introduce a uniform naming and define $B\dagger 1$ to be the left half of $B$ if $k$ is even and the lower half of $B$ if $k$ is odd. We define a _derived_ brick recursively as follows: a derived brick is either (i) a fundamental brick $B_{k}$ or (ii) $B\dagger 1$ or $B\dagger 2$, where $B$ is a derived brick. We introduce an ordering $\prec$ of the derived $k$-bricks as follows. Consider two derived $k$-bricks $D_{1}$ and $D_{2}$ such that $D_{1}\subset B_{i}$ and $D_{2}\subset B_{j}$. If $i>j$, then $D_{1}\prec D_{2}$. Else, if $i=j$ then the bricks $D_{1}$ and $D_{2}$ are both obtained by splitting the fundamental brick $B_{i}$, and the number of splits is $\ell:=i-k$. Hence the bricks have the forms $D_{1}=B_{i}\dagger b_{11}\dagger b_{12}\dagger\ldots\dagger b_{1\ell}$ and $D_{2}=B_{i}\dagger b_{21}\,b_{22}\,\ldots\,b_{2\ell}$, where $b_{ij}\in\\{1,2\\}$ for $i\in\\{1,2\\}$ and $j\in\\{1,\ldots,\ell\\}$. We then define $D_{1}\prec D_{2}$ if $(b_{11},b_{12},\ldots,b_{1\ell})$ precedes $(b_{21},b_{22},\ldots,b_{2\ell})$ in the lexicographic ordering. We say that a $k$-brick is _suitable_ for a piece $p$ of size $w\times h$ if the width and height of the brick are at least $w$ and $h$, respectively, and if that is not the case for a $(k+1)$-brick. We will always pack a given piece $p$ in a derived $k$-brick that is suitable for $p$. Figure 1: Left: Fundamental bricks. Middle: Splitting a brick. Right: Rectangular pieces packed in a brick. We now explain how we pack pieces into one specific brick; see Figure 1 (right). The first piece $p$ that is packed in a brick $B$ is placed with the lower left corner of $p$ at the lower left corner of $B$. Suppose now that some other pieces $p_{1},\ldots,p_{i}$ have been packed in $B$. If $k$ is even, then $p_{1},\ldots,p_{i}$ form a stack with the left edges contained in the left edge of $B$, and we place $p$ on top of $p_{i}$ (again, with the left edge of $p$ contained in the left edge of $B$). Otherwise, $p_{1},\ldots,p_{i}$ form a stack with the bottom edges contained in the bottom edge of $B$, and we place $p$ to the right of $p_{i}$ (again, with the bottom edge of $p$ contained in the bottom edge of $B$). We say that a brick _has room_ for a piece $p$ if the packing scheme above places $p$ within $B$, and it is apparent that an empty suitable brick for $p$ has room for $p$. Figure 2: Left: Some pieces have been packed by the algorithm. The bricks in $\mathcal{D}$ are drawn with fat edges. Right: A new piece arrives. There is already a brick of the suitable size in $\mathcal{D}$, but there is not enough room, so a new brick of the same size is added to $\mathcal{D}$ where the piece is placed. The algorithm BrickTranslation maintains the collection $\mathcal{D}$ of non- overlapping derived bricks, such that one or more pieces have been placed in each brick in $\mathcal{D}$; see Figure 2. Before the first piece arrives, we set $\mathcal{D}:=\emptyset$. Suppose that some stream of pieces have been packed, and that a new piece $p$ appears. Choose $k$ such that a $k$-brick is suitable for $p$. If there exists a derived $k$-brick $D\in\mathcal{D}$ such that $D$ has room for $p$, then we pack $p$ in $D$. Else, let $D$ be the minimum derived $k$-brick (with respect to the ordering $\prec$ described before) such that $D$ is interior-disjoint from each brick in $\mathcal{D}$; we then add $D$ to $\mathcal{D}$ and pack $p$ in $D$. ###### Theorem 2. The algorithm BrickTranslation has a competitive ratio strictly less than 4 for PerimeterTranslation. ###### Proof. We can assume, without loss of generality, that after we have packed the last rectangle, we have $\bigcup\mathcal{D}\subseteq B_{\geq 0}$ and $\bigcup\mathcal{D}\not\subseteq B_{\geq 1}$. As shown in Figure 3, we define a derived $k$-brick $B\subseteq B_{\geq 0}$ to be * • _sparse_ if $B\in\mathcal{D}$ and the total height (if $k$ is even, else width) of pieces stacked in $B$ is less than half of the height (if $k$ is even, else width) of $B$, * • _dense_ if $B\in\mathcal{D}$ and $B$ it is not sparse, * • _free_ if $B$ is interior-disjoint from each brick in $\mathcal{D}$, and * • _empty_ if $B$ is a maximal (w.r.t. inclusion) free brick. Figure 3: Brick $B_{3}$ is _sparse_ , brick $B_{2}\dagger 2$ is _empty_ , brick $B_{1}\dagger 1$ is _dense_. Brick $B_{1}\dagger 2\dagger 2$ is free, but not empty, since it is contained in $B_{1}\dagger 2$. ###### Remark 3. Sparse, dense and empty bricks together cover $B_{\geq 0}$, in fact every brick in $\mathcal{D}$ is either sparse or dense, and any brick in $B_{\geq 0}$ that is interior-disjoint from bricks in $\mathcal{D}$ is contained in some empty brick. ###### Remark 4. Every $k$-brick $D\in\mathcal{D}$ contains pieces for which it is suitable. Therefore, if $k$ is odd $D$ contains a piece of height at least $\sqrt{2}^{-k}/2$, and if $k$ is even $D$ contains a piece of width at least $\sqrt{2}^{-k}/2$. ###### Remark 5. Every $k$-brick $D\in\mathcal{D}$ that is dense contains pieces with total area at least $1/4$ of the area of $D$. To see this, suppose that $k$ is even, so that $D$ is $\sqrt{2}^{-k}\times\sqrt{2}^{-k-1}$, then thanks to density the total height of pieces in $D$ is at least half of its height, moreover thanks to Remark 4 all the pieces contained in $D$ have width at least $\sqrt{2}^{-k}/2$. If $k$ is odd we prove it analogously. ###### Remark 6. Consider two $k$-bricks $M$ and $N$. If $M\prec N$ and $M$ is free, then $N$ is free. To prove this it is sufficient to consider the step in which the first piece $p$ is placed within $N$ and $N\dagger b_{1}\dagger\ldots\dagger b_{\ell}$ is added to $\mathcal{D}$. Then, $N\dagger b_{1}\dagger\ldots\dagger b_{\ell}$ should be the $\prec$-minimum free suitable $k$-brick, but $M\dagger b_{1}\dagger\ldots\dagger b_{\ell}\prec N\dagger b_{1}\dagger\ldots\dagger b_{\ell}$ gives a contradiction. It follows that whenever we have a set $S$ of $k$-bricks that contains a free $k$-brick, then also $\max_{\prec}S$ is free. This turns out to be useful multiple times along the proof, choosing $S$ to be the set of $k$-bricks not contained in a strictly larger empty brick. ###### Remark 7. There exists no empty $0$-brick, otherwise $D\subseteq B_{\geq 1}$. Moreover, for every $k\geq 1$ we can have at most one sparse $k$-brick and one empty $k$-brick. In fact, a new empty (resp. sparse) $k$-brick is created only when no empty (resp. sparse) $k$-brick exists. In the following we prove an upper bound on the competitive ratio $\textsc{Alg}/\textsc{Opt}$, where Alg is the perimeter of the bounding box achieved by our online algorithm and Opt is the optimal perimeter computed offline. Hence, we need some techniques to provide an upper bound on Alg and a lower bound on Opt. For Alg, we will simply show a bounding box, in fact the perimeter of any bounding box containing all the pieces provides an upper bound to the minimum perimeter bounding box. For Opt, let $A$ be the total area of pieces and $L$ be the maximum length of an edge of a piece. If $L^{2}>A$ then the minimum perimeter bounding box cannot have a smaller perimeter than a box of size $L\times A/L$. Otherwise, if $L^{2}\leq A$ we have a weaker lower bound given by the box $\sqrt{A}\times\sqrt{A}$. Throughout the analysis we consider semiperimeters instead of perimeters to improve readability. We denote with $A(empty),A(sparse),A(dense)$ the total area of empty, sparse and dense bricks respectively. Thanks to Remark 3, we have that $A(empty)+A(sparse)+A(dense)=A(B_{\leq 0})=\sqrt{2}$. We denote with $A_{pcs}$ the total area of pieces in the stream. Thanks to Remark 5, we have $A_{pcs}\geq A(dense)/4$. From now on the proof branches in many cases and subcases. We will perform a depth-first visit of the case tree, and for each leaf of this tree we will prove that the competitive ratio is strictly less than $4$. Let $k$ be the smallest integer such that there exists a $k$-brick in $\mathcal{D}$, and let $M\in\mathcal{D}$ be the $\prec$-maximal $k$-brick. From our assumptions, it follows that $k\geq 0$. #### Case Tree Figure 4: Some of the cases listed in the proof of Theorem 2 are shown. The grey area must fit within the bounding box considered in the case analysis. Case (1) [_$M$ is a $0$-brick_] Thanks to Remark 4 every piece in $M$ has width at least $1/2$. Let $h$ be the total height of pieces stacked in $M$, then a bounding box of size size $1\times(\sqrt{2}/2+h)$ is obtained cutting the topmost part of $M$; see Figure 4. We can easily bound Opt with $1/2\times h$, and we get $\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{1+\frac{\sqrt{2}}{2}+h}{\frac{1}{2}+h}\leq 2+\sqrt{2}<4.$ Case (2) [_$M$ is a $k$-brick for $k\geq 2$_] Here we have two cases. Case (2.1) [_There exist a $1$-brick $N_{1}$ and a $2$-brick $N_{2}$ that are empty_] Thanks to Remark 6 we can choose $N_{1}$ = $B_{0}\dagger 2$ and $N_{2}=B_{0}\dagger 1\dagger 2$. In fact, for $B_{0}\dagger 2$ it is sufficient to choose $S$ as the set of all $1$-bricks, while for $B_{0}\dagger 1\dagger 2$ we can choose $S$ to be the set of all $2$-bricks that are not contained in a larger free brick. Thus, we can cut the topmost half of $B_{0}$ and get $\textsc{Alg}\leq 1+3/4\cdot\sqrt{2}$; see Figure 4. We have $\displaystyle A(empty)\leq\sum_{i\geq 1}A(B_{i})\leq\frac{\sqrt{2}}{2}$ $\displaystyle A(sparse)\leq\sum_{i\geq 2}A(B_{i})=\frac{\sqrt{2}}{4}\quad\text{(thanks to case (2) clause there is no sparse $1$-brick)}$ $\displaystyle A_{pcs}\geq\frac{A(dense)}{4}\geq\frac{A(B_{\geq 0})-A(sparse)-A(empty)}{4}\geq\frac{\sqrt{2}}{16}$ Now we are ready to bound Opt: $\displaystyle\textsc{Opt}\geq 2\cdot\sqrt{A_{pcs}}=\sqrt{\frac{\sqrt{2}}{4}}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{1+\frac{3}{4}\sqrt{2}}{\sqrt{\frac{\sqrt{2}}{4}}}\approx 3.47<4.$ Case (2.2) [_For $j=1$ or $j=2$ there does not exist an empty $j$-brick_] In this case we just use $\textsc{Alg}\leq 1+\sqrt{2}$. Then we have $\displaystyle A(empty)\leq\sum_{i\geq 1\land i\neq j}A(B_{i})\leq\frac{3}{8}\sqrt{2}\quad\text{(worst case is when $j=2$)}$ $\displaystyle A(sparse)\leq\sum_{i\geq 2}A(B_{i})=\frac{\sqrt{2}}{4}$ therefore performing the same computations of case (2.1), $A_{pcs}\geq 3/32\cdot\sqrt{2}$, and finally $\displaystyle\textsc{Opt}\geq 2\cdot\sqrt{\frac{3}{32}\sqrt{2}}=\sqrt{\frac{3}{8}\sqrt{2}}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{1+\sqrt{2}}{\sqrt{\frac{3}{8}\sqrt{2}}}\approx 3.32<4.$ Case (3) [_$M$ is a $1$-brick_] For the rest of the proof $L$ will be the length of the longest edge among all pieces. Since $M$ is a $1$-brick, we have $\sqrt{2}/4<L\leq\sqrt{2}/2$. Here we have two cases. Case (3.1) [_There does not exists an empty $1$-brick_] Here we have two cases. Case (3.1.1) [_For $j=2$ and $j=3$ there exists an empty $j$-brick_] Here we have three cases. Case (3.1.1.1) [_$M$ is the fundamental brick $B_{1}$_] Thanks to Remark 6 we can assume $B_{0}\dagger 2\dagger 2$ and $B_{0}\dagger 2\dagger 1\dagger 2$ to be empty. Here we have two cases. Case (3.1.1.1.1) [_$M$ is dense_] Since $M=B_{1}$ is the $\prec$-maximal $k$-brick in $\mathcal{D}$, then there does not exist a sparse $1$-brick. $\displaystyle A(empty)\leq\sum_{i\geq 2}A(B_{i})\leq\frac{\sqrt{2}}{4}$ $\displaystyle A(sparse)\leq\sum_{i\geq 2}A(B_{i})=\frac{\sqrt{2}}{4}$ therefore $A_{pcs}\geq\frac{\sqrt{2}}{8}$, and finally $\displaystyle\textsc{Opt}\geq 2\cdot\sqrt{\frac{\sqrt{2}}{8}}=\sqrt{\frac{\sqrt{2}}{2}}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{1+\sqrt{2}}{\sqrt{\frac{\sqrt{2}}{2}}}\approx 2.87<4.$ Case (3.1.1.1.2) [_$M$ is sparse_] Then, we can cut the rightmost part of $B_{\geq 0}$ and get a $3/4\times\sqrt{2}$ bounding box; see Figure 4. We have $\displaystyle A(empty)\leq\sum_{i\geq 2}A(B_{i})\leq\frac{\sqrt{2}}{4}$ $\displaystyle A(sparse)\leq\sum_{i\geq 1}A(B_{i})\leq\frac{\sqrt{2}}{2}$ hence $A_{pcs}\geq\sqrt{2}/16$. Since $L^{2}>1/8>\sqrt{2}/16$ we finally have $\displaystyle\textsc{Opt}\geq L+\frac{A_{pcs}}{L}\geq\frac{\sqrt{2}}{4}+\frac{1}{4}\quad\text{(minimizing over $L\in[\sqrt{2}/4,\sqrt{2}/2]$)}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{3/4+\sqrt{2}}{\frac{\sqrt{2}}{4}+\frac{1}{4}}\approx 3.59<4.$ Case (3.1.1.2) [_$M=B_{0}\dagger 1$_] Thanks to Remark 6 we can assume $B_{0}\dagger 2\dagger 2$ to be empty. Then, we can cut the topmost part of $B_{\geq 0}$ and get a $1\times\sqrt{2}/2+L$ bounding box; see Figure 4. We have $\displaystyle A(empty)\leq\sum_{i\geq 2}A(B_{i})\leq\frac{\sqrt{2}}{4}$ $\displaystyle A(sparse)\leq\sum_{i\geq 1}A(B_{i})\leq\frac{\sqrt{2}}{2}$ hence $A_{pcs}\geq\sqrt{2}/16$. Since $L^{2}>1/8>\sqrt{2}/16$ we finally have $\displaystyle\textsc{Opt}\geq L+\frac{A_{pcs}}{L}\geq L+\frac{\sqrt{2}}{16L}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{1+\sqrt{2}/2+L}{L+\frac{\sqrt{2}}{16L}}\leq 2+\sqrt{2}<4.\quad\text{(maximizing over $L\in[\sqrt{2}/4,\sqrt{2}/2]$)}$ Case (3.1.1.3) [_$M=B_{0}\dagger 2$_] This case is analogous to the previous one, in fact thanks to Remark 6 we can assume $B_{0}\dagger 1\dagger 2$ to be empty and cut the topmost part of $B_{\geq 0}$. Case (3.1.2) [_For $j=2$ or $j=3$ there does not exist an empty $j$-brick_] $\displaystyle A(empty)\leq\sum_{i\geq 2\land i\neq j}A(B_{i})\leq\frac{3}{16}\sqrt{2}\quad\text{(worst case is when $j=3$)}$ $\displaystyle A(sparse)\leq\sum_{i\geq 1}A(B_{i})\leq\frac{\sqrt{2}}{2}$ hence $A_{pcs}\geq\frac{5}{64}\cdot\sqrt{2}$. Since $L^{2}>1/8>\frac{5}{64}\cdot\sqrt{2}$ we finally have $\displaystyle\textsc{Opt}\geq L+\frac{A_{pcs}}{L}\geq\frac{\sqrt{2}}{4}+\frac{5}{16}\quad\text{(minimizing over $L\in[\sqrt{2}/4,\sqrt{2}/2]$)}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{1+\sqrt{2}}{\frac{\sqrt{2}}{4}+\frac{5}{16}}\approx 3.62<4.$ Case (3.2) [_There exists an empty $1$-brick_] Thanks to Remark 6 we can assume $B_{0}\dagger 2$ to be empty. Here we have two cases. Case (3.2.1) [_$M$ is the fundamental brick $B_{1}$_] Let $w$ be the total width of pieces stacked in $M$. Since $B_{0}\dagger 2$ is empty, we can cut the rightmost part of $B_{\geq 0}$ and get a $(1/2+w)\times\sqrt{2}$ bounding box; see Figure 4. Since increasing $w$ only improves our estimates, we consider the corner case $w=0$. Now we have two cases. Case (3.2.1.1) [_There does not exist an empty $2$-brick_] Here we have two cases. Case (3.2.1.1.1) [_For $j=3$ and $j=4$ there exists an empty $j$-brick_] Thanks to Remark 6 we can assume $B_{0}\dagger 1\dagger 2\dagger 2$ and $B_{0}\dagger 1\dagger 2\dagger 1\dagger 2$ to be empty. Thus, we can cut the topmost part of $B_{\geq 0}$ and get a $1/2\times(7/8\cdot\sqrt{2})$ bounding box; see Figure 4. We have $\displaystyle A(empty)\leq\sum_{i\geq 1\land i\neq 2}A(B_{i})\leq\frac{3}{8}\sqrt{2}$ $\displaystyle A(sparse)\leq\sum_{i\geq 1}A(B_{i})\leq\frac{\sqrt{2}}{2}$ hence $A_{pcs}\geq\sqrt{2}/32$. Since $L^{2}>1/8>\sqrt{2}/32$ we finally have $\displaystyle\textsc{Opt}\geq L+\frac{A_{pcs}}{L}\geq\frac{\sqrt{2}}{4}+\frac{1}{8}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{1/2+(7/8\cdot\sqrt{2})}{\frac{\sqrt{2}}{4}+\frac{1}{8}}\approx 3.63<4.$ Case (3.2.1.1.2) [_For $j=3$ or $j=4$ there does not exist an empty $j$-brick_] $\displaystyle A(empty)\leq\sum_{i\geq 1\land i\neq 2,j}A(B_{i})\leq\frac{11}{32}\sqrt{2}\quad\text{(worst case is when $j=4$)}$ $\displaystyle A(sparse)\leq\sum_{i\geq 1}A(B_{i})\leq\frac{\sqrt{2}}{2}$ hence $A_{pcs}\geq 5/128\cdot\sqrt{2}$. Since $L^{2}>1/8>5/128\cdot\sqrt{2}$ we finally have $\displaystyle\textsc{Opt}\geq L+\frac{A_{pcs}}{L}\geq\frac{\sqrt{2}}{4}+\frac{5}{32}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{1/2+\sqrt{2}}{\frac{\sqrt{2}}{4}+\frac{5}{32}}\approx 3.75<4.$ Case (3.2.1.2) [_There exists an empty $2$-brick_] Thanks to Remark 6 we can assume $B_{0}\dagger 1\dagger 2$ to be empty. Thus, we can cut the topmost part of $B_{\geq 0}$ and get a $1/2\times(3/4\cdot\sqrt{2})$ bounding box; see Figure 4. Here we have two cases. Case (3.2.1.2.1) [_There does not exist an empty $3$-brick_] $\displaystyle A(empty)\leq\sum_{i\geq 1\land i\neq 3}A(B_{i})\leq\frac{7}{16}\sqrt{2}$ $\displaystyle A(sparse)\leq\sum_{i\geq 1}A(B_{i})\leq\frac{\sqrt{2}}{2}$ hence $A_{pcs}\geq\sqrt{2}/64$. Since $L^{2}>1/8>\sqrt{2}/64$ we finally have $\displaystyle\textsc{Opt}\geq L+\frac{A_{pcs}}{L}\geq\frac{\sqrt{2}}{4}+\frac{1}{16}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{\frac{1}{2}+\frac{3}{4}\sqrt{2}}{\frac{\sqrt{2}}{4}+\frac{1}{16}}\approx 3.75<4.$ Case (3.2.1.2.2) [_There exists an empty $3$-brick_] Thanks to Remark 6 we can assume $B_{0}\dagger 1\dagger 1\dagger 2$ to be empty. Here we have two cases. Case (3.2.1.2.2.1) [_There does not exist an empty $4$-brick_] Here we have two cases. Case (3.2.1.2.2.1.1) [_For $j=5$ or $j=6$ there does not exist an empty $j$-brick_] $\displaystyle A(empty)\leq\sum_{i\geq 1\land i\neq 4,j}A(B_{i})\leq\frac{59}{128}\sqrt{2}\quad\text{(worst case is when $j=6$)}$ $\displaystyle A(sparse)\leq\sum_{i\geq 1}A(B_{i})\leq\frac{\sqrt{2}}{2}$ hence $A_{pcs}\geq 5/512\cdot\sqrt{2}$. Since $L^{2}>1/8>5/512\cdot\sqrt{2}$ we finally have $\displaystyle\textsc{Opt}\geq L+\frac{A_{pcs}}{L}\geq\frac{\sqrt{2}}{4}+\frac{5}{128}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{\frac{1}{2}+\frac{3}{4}\sqrt{2}}{\frac{\sqrt{2}}{4}+\frac{5}{128}}\approx 3.98<4.$ Case (3.2.1.2.2.1.2) [_For $j=5$ and $j=6$ there exists an empty $j$-brick_] Thanks to Remark 6 we can assume $B_{0}\dagger 1\dagger 1\dagger 1\dagger 2\dagger 2$ and $B_{0}\dagger 1\dagger 1\dagger 1\dagger 2\dagger 1\dagger 2$ to be empty. Then, we can cut the topmost part of $B_{\geq 0}$ and get a $1/2\times(11/16\cdot\sqrt{2})$ bounding box; see Figure 4. We have $\displaystyle A(empty)\leq\sum_{i\geq 1\land i\neq 4}A(B_{i})\leq\frac{15}{32}\sqrt{2}$ $\displaystyle A(sparse)\leq\sum_{i\geq 1}A(B_{i})\leq\frac{\sqrt{2}}{2}$ hence $A_{pcs}\geq\sqrt{2}/128$. Since $L^{2}>1/8>\sqrt{2}/128$ we finally have $\displaystyle\textsc{Opt}\geq L+\frac{A_{pcs}}{L}\geq\frac{\sqrt{2}}{4}+\frac{1}{32}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{\frac{1}{2}+\frac{11}{16}\sqrt{2}}{\frac{\sqrt{2}}{4}+\frac{1}{32}}\approx 3.83<4.$ Case (3.2.1.2.2.2) [_There exists an empty $4$-brick_] Thanks to Remark 6 we can assume $B_{0}\dagger 1\dagger 1\dagger 1\dagger 2$ to be empty. Then, we can cut the topmost part of $B_{\geq 0}$ and get a $1/2\times(5/8\cdot\sqrt{2})$ bounding box; see Figure 4. Now it remains to bound Opt, and we just assume $\textsc{Opt}\geq L\geq\sqrt{2}/4$, finally $\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{\frac{1}{2}+\frac{5}{8}\sqrt{2}}{\frac{\sqrt{2}}{4}}\approx 3.92<4.$ Case (3.2.2) [_$M=B_{0}\dagger 1$_] For the rest of the proof let $L$ be the length of the longest of pieces’ edges then, according to Remark 4, $\sqrt{2}/4\leq L\leq\sqrt{2}/2$. We can cut the topmost part of $B_{\geq 0}$ and get a $1\times(\sqrt{2}/2+L)$ bounding box; see Figure 4. Here we have two cases. Case (3.2.2.1) [_There exists a $2$-brick in $\mathcal{D}$_] Thanks to Remark 4, we have a piece of width at least $1/4$, and combining this with the fact that we have a piece of height $L$, it is apparent that $\textsc{Opt}\geq 1/4+L$; see Figure 4. Thus, $\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{1+\frac{\sqrt{2}}{2}+L}{\frac{1}{4}+L}\leq 2+\sqrt{2}<4\quad\text{(maximizing over $L\in[\sqrt{2}/4,\sqrt{2}/2]$).}$ Case (3.2.2.2) [_There does not exist a $2$-brick in $\mathcal{D}$_] Here we have two cases. Case (3.2.2.2.1) [_There does not exist an empty $2$-brick_] $\displaystyle A(empty)\leq\sum_{i\geq 1\land i\neq 2}A(B_{i})\leq\frac{3}{8}\sqrt{2}$ $\displaystyle A(sparse)\leq\sum_{i\geq 1}A(B_{i})\leq\frac{3}{8}\sqrt{2}$ hence $A_{pcs}\geq\sqrt{2}/16$. Since $L^{2}>1/8>\sqrt{2}/16$ we finally have $\displaystyle\textsc{Opt}\geq L+\frac{A_{pcs}}{L}\geq L+\frac{\sqrt{2}}{16L}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{1+\frac{\sqrt{2}}{2}+L}{L+\frac{\sqrt{2}}{16L}}\leq 2+\sqrt{2}<4\quad\text{(maximizing over $L\in[\sqrt{2}/4,\sqrt{2}/2]$).}$ Case (3.2.2.2.2) [_There exists an empty $2$-brick_] Thanks to Remark 6 we can assume $B_{1}\dagger 2$ to be empty. Here we have two cases. Case (3.2.2.2.2.1) [_There exists an empty $3$-brick_] Thanks to Remark 6 we can assume $B_{1}\dagger 1\dagger 2$ to be empty. Then, we can cut the rightmost part of $B_{\geq 0}$ and get a $3/4\times(\sqrt{2}/2+L)$ bounding box; see Figure 4. Now it remains to bound Opt. We have $\displaystyle A(empty)\leq\sum_{i\geq 1}A(B_{i})\leq\frac{\sqrt{2}}{2}$ $\displaystyle A(sparse)\leq\sum_{i\geq 1\land i\neq 2}A(B_{i})\leq\frac{3}{8}\sqrt{2}$ hence $A_{pcs}\geq\sqrt{2}/32$. Since $L^{2}>1/8>\sqrt{2}/32$ we finally have $\displaystyle\textsc{Opt}\geq L+\frac{A_{pcs}}{L}\geq L+\frac{\sqrt{2}}{32L}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{\frac{3}{4}+\frac{\sqrt{2}}{2}+L}{L+\frac{\sqrt{2}}{32L}}\leq 3.79<4\quad\text{(maximizing over $L\in[\sqrt{2}/4,\sqrt{2}/2]$).}$ Case (3.2.2.2.2.2) [_There does not exist an empty $3$-brick_] $\displaystyle A(empty)\leq\sum_{i\geq 1\land i\neq 3}A(B_{i})\leq\frac{7}{16}\sqrt{2}$ $\displaystyle A(sparse)\leq\sum_{i\geq 2\land i\neq 2}A(B_{i})\leq\frac{3}{8}\sqrt{2}$ hence $A_{pcs}\geq 3/64\cdot\sqrt{2}$. Since $L^{2}>1/8>3/64\cdot\sqrt{2}$ we finally have $\displaystyle\textsc{Opt}\geq L+\frac{A_{pcs}}{L}\geq L+\frac{3\sqrt{2}}{64L}$ $\displaystyle\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{1+\frac{\sqrt{2}}{2}+L}{L+\frac{3\sqrt{2}}{64L}}\leq 3.82<4\quad\text{(maximizing over $L\in[\sqrt{2}/4,\sqrt{2}/2]$).}$ ∎ #### Algorithm using rotations The algorithm BrickRotation is almost identical to BrickTranslation, but with the difference that we rotate each piece so that its height is at least its width. ###### Theorem 8. The algorithm BrickRotation has a competitive ratio of strictly less than 4 for PerimeterRotation. ###### Proof. The analysis of BrickTranslation carried out in the proof of Theorem 2 still holds, in fact all the estimates on Opt derived from consideration about area are still valid, and the only delicate spot is case (3.2.2.1). In that case we assume to have a piece $p$ having an edge of length $L\in[\sqrt{2}/4,\sqrt{2}/2]$, and that there exists a $2$-brick in $\mathcal{D}$. Thanks to Remark 4 there exists a piece $q$ of size $w_{q}\times h_{q}$ with $w_{q}\geq 1/4$, moreover we rotate every piece so that $1/4\leq w_{q}\leq h_{q}$. Finally, a box that contains both $p$ and $q$ must have size at least $L\times w_{q}$ or $L\times h_{q}$, hence $\textsc{Opt}\geq\min\left\\{L+w_{q},L+h_{q}\right\\}\geq L+\frac{1}{4}.$ This gives exactly the same bound showed in case (3.2.2.1) and completes the proof. ∎ ### 2.2 A similar but inferior algorithm Here we consider the algorithm we get by making a slight change to BrickTranslation. Suppose that the very first piece $p$ arrives and that a $k$-brick is suitable for $p$. Instead of placing $p$ in $B_{k}$ (as BrickTranslation would do), we consider the brick $B_{>k}$ to be a fundamental brick (although in the original algorithm, it was an infinite union of fundamental bricks) and we place $p$ in $B_{>k}$. Thus, we are never going to use the fundamental bricks $B_{i}$ individually, for $i>k$. From here on, the algorithm does as BrickTranslation: Whenever a new piece arrives, we place it in the first derived brick of the suitable size that has room. This behavior is similar to the algorithm for the problem SquareInSquareArea that was described by Fekete and Hoffmann [13]. That problem is studied in more detail in Section 3.3, and for that problem, the algorithm seems to be no worse than ours. Interestingly, the following theorem together with Theorem 2 implies that the modified algorithm is worse for the problem PerimeterTranslation. Figure 5: Left: A configuration produced by the modified version of BrickTranslation. Right: The configuration produced by the original algorithm BrickTranslation. ###### Theorem 9. The modified version of BrickTranslation has a competitive ratio of at least $4$ for the problem PerimeterTranslation. ###### Proof. For any $\varepsilon^{\prime}>0$, we can make an instance realizing a competitive ratio of more than $4-\varepsilon^{\prime}$ as follows. Figure 5 shows the packing produced by the modified and the original algorithm. We first give the algorithm the rectangle $(1/2\sqrt{2}+\varepsilon)\times\varepsilon$ for an infinitesimal $\varepsilon>0$. The rectangle is placed in $B_{>1}$ by the modified algorithm. For a large odd integer $k$, we then feed the algorithm with small rectangles of size $(\sqrt{2}^{-k-1}+\varepsilon)\times(\sqrt{2}^{-k}+\varepsilon)$ until $B_{1}\dagger 1$ has been completely split into $(k-2)$-bricks, each of which contains one small rectangle. We now give the algorithm a piece of size $\varepsilon\times(1/4+\varepsilon)$, which is placed in $B_{1}\dagger 2$. We again give the algorithm many small rectangles until $B_{0}\dagger 1\dagger 1\dagger 1$ has been split into $(k-2)$-bricks. Now follows a rectangle of size $(1/4\sqrt{2}+\varepsilon)\times\varepsilon$, which is placed in $B_{0}\dagger 1\dagger 1\dagger 2$. Finally, we fill $B_{0}\dagger 1\dagger 2\dagger 1$ with small rectangles. Note that as $k\longrightarrow\infty$, the bounding box of the produced packing converges to $B_{\geq 0}$, so it has a perimeter as $B_{-1}$. On the other hand, observe that as $\varepsilon\longrightarrow 0$, we have $\Sigma\longrightarrow A(B_{1})/4=A(B_{3})$, since the small rectangles fill out bricks with a total area of $A(B_{1})$ and with density $1/4$. In the limit, all the pieces can actually be packed into $B_{3}$, so Opt is at most the perimeter of $B_{3}$. But the perimeter of $B_{-1}$ is $4$ times that of $B_{3}$, which finishes the proof. ∎ ### 2.3 Lower bounds ###### Lemma 10. Consider any algorithm $A$ for the problem PerimeterTranslation. Then the competitive ratio of $A$ is at least $4/3$. ###### Proof. We first feed $A$ with two unit squares. Let the bounding box of the two squares have size $a\times b$ and suppose without loss of generality that $a\leq b$. Then $a\geq 1$ and $b\geq 2$. We now give $A$ a rectangle of size $2\times\varepsilon$ for a small value $\varepsilon>0$. The produced packing has a bounding box of perimeter more than $8$, whereas the optimal has perimeter $6+2\varepsilon$. Therefore, the competitive ratio is $\frac{8}{6+2\varepsilon}=\frac{4}{3+\varepsilon}$. By letting $\varepsilon\longrightarrow 0$, we get that the ratio is at least $4/3$. ∎ ###### Lemma 11. Consider any algorithm $A$ for the problem PerimeterRotation. Then the competitive ratio of $A$ is at least $5/4$. ###### Proof. We first feed $A$ with three unit squares. Let the bounding box of the three squares have size $a\times b$ and suppose without loss of generality that $a\leq b$. Suppose first that $b<3$. Then we must have $a\geq 2$ for the box to contain the squares. We then give the algorithm the rectangle $\varepsilon\times 3$ for a small value $\varepsilon>0$. The produced packing has a bounding box of size at least $(2+\varepsilon)\times 3$ and perimeter more than $10$, while the optimal solution has size $(1+\varepsilon)\times 3$ and perimeter $8+2\varepsilon$. On the other hand, if $b\geq 3$, we give the algorithm one more unit square. The produced packing has a bounding box of size at least $2\times 3$ or at least $1\times 4$, and thus perimeter at least $10$, while the optimal packing has size $2\times 2$ and perimeter $8$. We get that the competitive ratio is at least $\frac{10}{8+2\varepsilon}=\frac{5}{4+\varepsilon}$, and by letting $\varepsilon\longrightarrow 0$, we get that the ratio is at least $5/4$. ∎ ## 3 Area versions ### 3.1 General lower bounds In this section we show that, if we allow pieces to be arbitrary rectangles, we cannot bound the competitive ratio for neither AreaTranslation nor AreaRotation as a function of the area Opt of the optimal packing. However we will be able to bound the competitive ratio as a function of the total number $n$ of pieces in the stream. ###### Lemma 12. Consider any algorithm $A$ solving AreaTranslation or AreaRotation and let any $m\in\mathbb{N}$ and $p\in\mathbb{R}$ be given. There exists a stream of $n=m^{2}+1$ rectangles such that (i) the rectangles can be packed into a bounding box of area $2p^{2}$, and (ii) algorithm $A$ produces a packing with a bounding box of area at least $mp^{2}$. ###### Proof. We first feed $A$ with $m^{2}$ rectangles of size $p\times\frac{p}{m^{2}}$. These rectangles have total area $p^{2}$. Let $a\times b$ be the size of the bounding box of the produced packing. Suppose first that $a\geq\frac{p}{m}$ and $b\geq\frac{p}{m}$ hold. We then feed $A$ with a long rectangle of size $pm^{2}\times\frac{p}{m^{2}}$. The produced packing has a bounding box of area at least $\frac{p}{m}\cdot pm^{2}=mp^{2}$. The optimal packing is to pack the $m^{2}$ small rectangles along the long rectangle, which would produce a packing with bounding box of size $pm^{2}\times\frac{2p}{m^{2}}=2p^{2}$. Otherwise, we must have $b>pm$ or $a>pm$, since $ab\geq p^{2}$. We then feed $A$ with a square of size $p\times p$. The produced packing has a bounding box of area at least $p\cdot pm=mp^{2}$. The optimal packing is obtained stacking the $m^{2}$ thin rectangles on top of the big square, which produces a packing with bounding box of size $p\times 2p=2p^{2}$. ∎ ###### Corollary 13. Let $A$ be an algorithm for AreaTranslation or AreaRotation. Then $A$ does not have an asymptotic, and hence also absolute, competitive ratio which is a function of Opt. ###### Proof. Let $f$ be any function of Opt. For any value $\textsc{Opt}=c$, we choose $p:=\sqrt{c/2}$. We now choose $m>2f(c)$ and obtain that the competitive ratio is at least $\frac{mp^{2}}{2p^{2}}=m/2>f(c)=f(\textsc{Opt})$. ∎ ###### Corollary 14. Let $A$ be an algorithm for AreaTranslation or AreaRotation. If $A$ has an asymptotic competitive ratio of $f(n)$, where $n=|L|$ is the number of pieces in the stream, then $f(n)=\Omega(\sqrt{n})$. This holds even when all edges of the pieces are required to have length at least $1$. ###### Proof. We choose $p:=m^{2}$. Then all edges have length at least $1$, and the competitive ratio is at least $\frac{mp^{2}}{2p^{2}}=m/2=\Omega(\sqrt{n})$. Here, Opt can be arbitrarily big by choosing $m$ big enough, so it is a lower bound on the asymptotic competitive ratio. ∎ ### 3.2 Algorithms for arbitrary pieces In this section we provide algorithms that solve AreaTranslation and AreaRotation with a competitive ratio of $O(\sqrt{n})$, where $n$ is the total number of pieces. Thus we match the bounds provided in the previous section. We first describe the algorithm DynBoxTrans that solves AreaTranslation. We assume to receive a stream of pieces $p_{1},\dots,p_{n}$ of unknown length $n$, such that piece $p_{i}$ has size $w_{i}\times h_{i}$. For each $k\in\mathbb{Z}$, we define a rectangular box $B_{k}$ with a size varying dynamically. After pieces $p_{1},\dots,p_{j}$ have been processed $B_{k}$ has size $2^{k}\times T_{j}$, where $T_{j}:=H_{j}\sqrt{j}+7H_{j}$ and $H_{j}:=\max_{i=1,\dots,j}h_{i}$. We place the boxes with their bottom edges on the $x$-axis and in order such that the right edge of $B_{k-1}$ is contained in the left edge of $B_{k}$; see Figure 6. Furthermore, we place the lower left corner of box $B_{0}$ at the point $(1,0)$. It then holds that all the boxes are to the right of the point $(0,0)$. Figure 6: The algorithm DynBoxTrans packs pieces into the boxes $B_{k}$ that form a row. Every box has height $T_{j}$ that is dynamically updated. We say that the box $B_{k}$ is _wide enough_ for a piece $p_{i}=w_{i}\times h_{i}$ if $w_{i}\leq 2^{k}$. If a box $B_{k}$ is wide enough for $p_{i}$, we can pack $p_{i}$ in $B_{k}$ using the online strip packing algorithm $\textsc{NFS}_{k}$ that packs rectangles into a strip of width $2^{k}$. The algorithm $\textsc{NFS}_{k}$ is the _next-fit shelf algorithm_ first described by Baker and Schwartz [6]. The algorithm packs pieces in _shelves_ (rows), and each shelf is given a fixed height of $2^{j}$ for some $j\in\mathbb{Z}$ when it is created; see Figure 7. The width of each shelf is $2^{k}$, since this is the width of the box $B_{k}$. Figure 7: A packing produced by the next-fit shelf algorithm using four shelves. A piece of height $h$, where $2^{j-1}<h\leq 2^{j}$, is packed in a shelf of height $2^{j}$. We divide the shelves into two types. If the total width of pieces in a shelf is more than $2^{k-1}$ we call that shelf _dense_ , otherwise we say it is _sparse_. The algorithm $\textsc{NFS}_{k}$ places each piece as far left as possible into the currently sparse shelf of the proper height. If there is no sparse shelf of this height or the sparse shelf has not room for the piece, a new shelf of the appropriate height is created on top of the top shelf, and the piece is placed there at the left end of this new shelf. This ensures that at any point in time there exists at most one sparse shelf for each height $2^{j}$. If we allow the height of the box $B_{k}$ to grow large enough with respect to shelves’ heights, the space wasted by sparse shelves becomes negligible and we obtain a constant density strip packing, as stated in the following lemma. ###### Lemma 15. Let $\widetilde{H}$ be the total height of shelves in $B_{k}$, and $H_{max}$ be the maximum height among pieces in $B_{k}$. If $\widetilde{H}\geq 6H_{max}$, then the pieces in $B_{k}$ are packed with density at least $1/12$. ###### Proof. Let $2^{m-1}<H_{max}\leq 2^{m}$, so that $\widetilde{H}\geq 3\cdot 2^{m}$. For each $i\leq m$ we have at most one sparse shelf of height $2^{i}$ and each shelf of $B_{k}$ has height at most $2^{m}$, hence the total height of sparse shelves is at most $\sum_{i\leq m}2^{i}=2^{m+1}$, so the total height of dense shelves is at least $\widetilde{H}-2^{m+1}\geq\widetilde{H}/3$. Thus, the total area of the dense shelves is at least $2^{k}\cdot\widetilde{H}/3$. Consider a dense shelf of height $2^{i}$. Into that shelf, we have packed pieces of height at least $2^{i-1}$, and the total width of these pieces is at least $2^{k-1}$. Hence, the density of pieces in the shelf is at least $1/4$. Therefore, the total area of pieces in $B_{k}$ is at least $2^{k}\cdot\widetilde{H}/12$. On the other hand, the area of the bounding box is $2^{k}\cdot\widetilde{H}$, that yields the desired density. ∎ Now we are ready to describe how the algorithm works. When the first piece $p_{1}$ arrives, let $2^{k-1}<w_{1}\leq 2^{k}$, then we pack it in the box $B_{k}$ according to $\textsc{NFS}_{k}$ and define $B_{k}$ to be the _active box_. Suppose now that $B_{i}$ is the active box when the piece $p_{j}$ arrives, first we update the value of the threshold $T_{j-1}$ to $T_{j}$, then we have two cases. If $w_{j}>2^{i}$ we choose $\ell$ such that $2^{\ell-1}<w_{j}\leq 2^{\ell}$, pack $p_{j}$ in $B_{\ell}$ and define $B_{\ell}$ to be the active box. Else, $B_{i}$ is wide enough for $p_{j}$ and we try to pack $p_{j}$ into $B_{i}$. Since $B_{i}$ has size $2^{i}\times T_{j}$ it may happen that $\textsc{NFS}_{i}$ exceeds the threshold $T_{j}$ while packing $p_{j}$, generating an overflow. In this case, instead of packing $p_{j}$ in $B_{i}$, we pack $p_{j}$ into $B_{i+1}$ and define that to be the active box. ###### Theorem 16. The algorithm DynBoxTrans has an absolute competitive ratio of $O(\sqrt{n})$ for the problem AreaTranslation on a stream of $n$ pieces. ###### Proof. First, define $\Sigma_{j}$ as the total area of the first $j$ pieces, $W:=\max_{i=1,\dots,n}w_{i}$ and recall that $H_{j}=\max_{i=1,\dots,j}h_{i}$ and $T_{j}=H_{j}\sqrt{n}+7H_{j}$. Let $B_{k}$ be the last active box, so that we can enclose all the pieces in a bounding box of size $2^{k+1}\times T_{n}$, and bound the area returned by the algorithm as $\textsc{Alg}=O(2^{k}H_{n}\sqrt{n})$. On the other hand we are able to bound the optimal offline packing as $\textsc{Opt}=\Omega(\Sigma_{n}+WH_{n})$. If the active box never changed, then we have $2^{k}<2W$ that implies $\textsc{Alg}=O(WH_{n}\sqrt{n})=\textsc{Opt}\cdot O(\sqrt{n})$. Otherwise, let $B_{\ell}$ be the last active box before $B_{k}$, and $p_{j}$ be the first piece put in $B_{k}$. Here we have two cases. Case (1) [_$w_{j} >2^{\ell}$_] In this case we have $2^{k}<2W$ that implies $\textsc{Alg}=O(WH_{n}\sqrt{n})=\textsc{Opt}\cdot O(\sqrt{n})$. Case (2) [_$w_{j}\leq 2^{\ell}$_] In this case we have $k=\ell+1$. Denote with $\widetilde{H_{i}}$ the total height of shelves in $B_{i}$. Then we have $\widetilde{H_{\ell}}\geq T_{j}-H_{j}=H_{j}\sqrt{n}+6H_{j}$, otherwise we could pack $p_{j}$ in $B_{\ell}$. Thus, we can apply Lemma 15 and conclude that the box $B_{\ell}$ of size $2^{\ell}\times T_{j}$ is filled with constant density. Here we have two cases. Case (2.1) [_$\widetilde{H_{k}}\leq T_{j}$_] In this case we have $\textsc{Alg}=O(2^{k}T_{j})$ and, thanks to the constant density packing of $B_{\ell}$ we have $\Sigma_{j}=\Theta(2^{\ell}\widetilde{H_{\ell}})=\Theta(2^{k}T_{j})$. Since $\textsc{Opt}\geq\Sigma_{j}$, we get $\textsc{Alg}=O(\textsc{Opt})$. Case (2.2) [_$\widetilde{H_{k}} >T_{j}$_] In this case we have $\textsc{Alg}=O(2^{k}\widetilde{H_{k}})$. Moreover, $\widetilde{H_{k}}=O(H_{n}+\Sigma_{n}/2^{k})$, in fact if $2^{s-1}<H_{n}\leq 2^{s}$, then the total height of sparse shelves is $\sum_{i\leq s}2^{i}=2^{s+1}=O(H_{n})$. Furthermore, dense shelves are filled with constant density, therefore their total height is at most $O(\Sigma_{n}/2^{k})$. Finally, we need to show that $2^{k}=O(W\sqrt{n})$. Thanks to the constant density packing of $B_{\ell}$, we have $2^{k}H_{j}\sqrt{j}=O(2^{\ell}T_{j})=O(\Sigma_{j})$. We can upper bound the size of every piece $p_{i}$ for $i\leq j$ with $W\times H_{j}$ and obtain $\Sigma_{j}\leq n\cdot WH_{j}$. Plugging it in the previous estimate and dividing both sides by $H_{j}\sqrt{n}$ we get $2^{k}=O(W\sqrt{n})$. Now we have $\textsc{Alg}=O(2^{k}\widetilde{H_{k}})=O(2^{k}H_{n}+\Sigma_{n})=O(WH_{n}\sqrt{n}+\Sigma_{n})=\textsc{Opt}\cdot O(\sqrt{n})$. ∎ The algorithm DynBoxRot is obtained from DynBoxTrans with a slight modification: before processing any piece $p_{i}$ we rotate it so that $w_{i}\leq h_{i}$. In this way, it still holds that $\textsc{Opt}=\Omega(\Sigma_{n}+WH_{n})$ and the proof of Theorem 16 works also for the following. ###### Theorem 17. The algorithm DynBoxRot has an absolute competitive ratio of $O(\sqrt{n})$ for the problem AreaRotation on a stream of $n$ pieces. ### 3.3 Bounded aspect ratio In this section, we will consider the special case where the aspect ratio of all pieces is $\alpha=1$, i.e., all the pieces are squares. Furthermore, we will measure the size of the packing as the area of the minimum axis-parallel bounding _square_ , and we call the resulting problem SquareInSquareArea. Since we get a constant competitive ratio in this case, it follows that for other values of $\alpha$ and when allowing the bounding box to be a general rectangle, one can likewise achieve a constant competitive ratio. We first give a lower bound. ###### Lemma 18. Consider any algorithm $A$ for the problem SquareInSquareArea. Then the competitive ratio of $A$ is at least $16/9$. ###### Proof. We first give $A$ four $1\times 1$ squares. Let the bounding square have size $\ell\times\ell$. If $\ell\geq 3$, the bounding square of the four $1\times 1$ squares has size at least $3\times 3$, while the optimal packing has size $2\times 2$, which gives ratio at least $9/4$. Otherwise, if $\ell<3$, we give a $2\times 2$ square and we will prove that the bounding square has size at least $4\times 4$ while the optimal packing fits in a $3\times 3$ square, so the ratio is at least $16/9$. Let us assume by contradiction that there exists a $(4-\varepsilon)\times(4-\varepsilon)$ bounding square containing both a $2\times 2$ square and four $1\times 1$ squares, with the additional hypothesis that the $1\times 1$ squares fit in a $(3-\delta)\times(3-\delta)$ bounding box. We refer to notation in Figure 8 (left) and notice that we have $a<1$ or $b<1$, and analogously $c<1$ or $d<1$. Without loss of generality, we can assume $a,d<1$. Hence, starting from the configuration in Figure 8 (left) we can drag the $2\times 2$ square to the bottom left corner and obtain the configuration in Figure 8 (right), that still fulfill the hypotheses we assumed by contradiction. Figure 8: Left: A $2\times 2$ square inside a bounding square having edges shorter than $4$. Right: The $2\times 2$ square has been dragged in the bottom left corner of the bounding square. Four $1\times 1$ squares $Q_{1},\dots,Q_{4}$ are placed within the bounding square. From now on we employ the notation of Figure 8 (right). Let $(x_{i},y_{i})$ be the coordinates of the bottom left corner of square $Q_{i}$. Stating that $Q_{i}$ and $Q_{j}$ are disjoint is equivalent to $\max\\{|x_{i}-x_{j}|,|y_{i}-y_{j}|\\}\geq 1$. Consider now the two rectangular regions $ABDE$ and $GCDF$: note that each of them can contain at most two squares. Indeed, given $Q_{i}$ and $Q_{j}$ completely contained in $ABDE$, it holds $|y_{i}-y_{j}|\leq 1-\varepsilon$ thus $|x_{i}-x_{j}|\geq 1$. If three squares $Q_{1},Q_{2},Q_{3}$ are completely contain in $ABDE$ then we have, without loss of generality, $x_{1}\leq x_{2}-1\leq x_{3}-2$ and the minimal bounding square of $Q_{1},Q_{2},Q_{3}$ has size at least $3\times 3$, that gives a contradiction. The same holds for $GCDF$. Finally, every $Q_{i}$ is either fully contained in $ABDE$ or $GCDF$ hence, without loss of generality, we can assume that $Q_{1},Q_{2}$ are contained in $ABDE$ and $Q_{3},Q_{4}$ are contained in $GCDF$. This implies that $x_{1}\leq x_{2}-1$ and $y_{4}\leq y_{3}-1$, again without loss of generality. Observe that $x_{2}\leq x_{3}+1-\varepsilon$ and $y_{3}\leq y_{2}+1-\varepsilon$. $Q_{2}$ and $Q_{3}$ are disjoint, using the previous characterization we have two cases. First, $|x_{2}-x_{3}|\geq 1$ and thanks to the observation above it cannot be $x_{2}>x_{3}$, therefore we have $x_{1}\leq x_{2}-1\leq x_{3}-2$. Else, $|y_{2}-y_{3}|\geq 1$ and thanks to the observation above we have $y_{4}\leq y_{3}-1\leq y_{2}-2$. In both cases that gives a contradiction since we cannot pack all $Q_{i}$s in a $(3-\delta)\times(3-\delta)$ bounding square. ∎ We are now going to analyze the competitive ratio of the algorithm BrickTranslation (in fact, the algorithm BrickRotation has the exact same behavior when the pieces are squares). Note that a brick can never contain more than one piece. The algorithm is almost the same as the one described by Fekete and Hoffmann [13]. The slight difference is addressed in Section 2.2 and it is shown there that the behavior as described by Fekete and Hoffmann makes a worse algorithm for the problem PerimeterTranslation. However, even though the two algorithms will not always produce identical packings for the problem SquareInSquareArea, the analysis of the following theorem seems to hold for both versions, so for the problem SquareInSquareArea, the algorithms are equally good. ###### Theorem 19. The algorithm BrickTranslation has a competitive ratio of $6$ for SquareInSquareArea. The analysis is tight. ###### Proof. Suppose a stream of squares have been packed by BrickTranslation, and let Alg be the area of the bounding square of the resulting packing. Let $B_{k}$ be the largest elementary brick in which a square has been placed. Suppose without loss of generality that $k=0$, so that $B_{k}$ has size $1\times 1/\sqrt{2}$ and $B_{\geq k}$, which contains all the packed squares, has size $1\times\sqrt{2}$. Figure 9: Left: A $2k$-packing. The grey bricks are non-empty and may have been split into smaller bricks. Right: The $2k$-packing produced by BrickTranslation when providing the algorithm with enough copies of the square $S_{k}$ (the small grey squares), showing that the competitive ratio can be arbitrarily close to $6$. We now recursively define a type of packing that we call a $2k$-packing, for a non-negative integer $k$; see Figure 9 (left). As $k$ increases, so do the requirements to a $2k$-packing, in the sense that a $(2k+2)$-packing is also a $2k$-packing, but the other way is in general not the case. Define $F_{0}:=B_{\geq 1}$ and $U_{0}:=B_{0}$. A packing is a $0$-packing if pieces have been placed in $U_{0}$ (the brick $U_{0}$ may or may not have been split in smaller bricks). Hence, the considered packing is a $0$-packing by the assumption that a piece has been placed in $B_{0}$. Suppose that we have defined a $2k$-packing for some integer $k$. A $(2k+2)$-packing is a $2k$-packing with the additional requirements that * • the brick $U_{2k}$ has been split into $L:=U_{2k}\dagger 1$ and $E_{2k+1}:=U_{2k}\dagger 2$, * • the right brick $E_{2k+1}$ is empty, * • the left brick $L$ has been split into $F_{2k+2}:=L\dagger 1$ and $U_{2k+2}:=L\dagger 2$, and * • $U_{2k+2}$ is non-empty, and thus also $F_{2k+2}$ is non-empty. The symbols $U_{j},E_{j},F_{j}$ have been chosen such that the brick is a $j$-brick, i.e., the index tells the size of the brick. Consider a $2k$-packing. It follows from the definition that along the top edge of $B_{\geq 0}$ from the right corner $(1,\sqrt{2})$ to the left corner $(0,\sqrt{2})$, we meet a sequence $E_{1},E_{3},\ldots,E_{2k-1}$ of empty bricks of decreasing size, and finally meet a non-empty brick $U_{2k}$ which may have been split into smaller bricks. ###### Claim 20. If the packing is a $2k$-packing and not a $(2k+2)$-packing, then $\textsc{Alg}/\textsc{Opt}<6$. Since we pack a finite number of squares, the produced packing is a $2k$-packing but not a $(2k+2)$-packing for some sufficiently large $k$, so Claim 20 implies Theorem 19. Let us now prove Claim 20. We first compute the area of the brick $U_{2k}$ and the total areas of the bricks $F_{0},F_{2},\ldots,F_{2k}$, as these areas will be used often: $\displaystyle u_{k}$ $\displaystyle:=|U_{2k}|=2^{-2k}/\sqrt{2}$ (1) $\displaystyle f_{k}$ $\displaystyle:=\sum_{i=0}^{k}|F_{2i}|=\frac{2|B_{\geq 0}|-u_{k}}{3}=\frac{4-4^{-k}}{3\sqrt{2}}.$ (2) * 1) Suppose first that $U_{2k}$ has not been split into smaller bricks. Then, since $U_{2k}$ is non-empty by assumption, we know that $U_{2k}$ contains a square $S$ of size $s\times s$ where $s\in(s_{l},s_{h}]=\left(\sqrt{2}^{-2k-2},\sqrt{2}^{-2k-1}\right]$. Since the bricks $E_{1},E_{3},\ldots,E_{2k-1}$ are all empty, we get that the upper edge of the bounding square coincides with the upper edge of $S$, and we thus have $\textsc{Alg}\leq\textsc{Alg}(s):=(\sqrt{2}-(\sqrt{2}^{-2k-1}-s))^{2}.$ The largest empty brick in the bricks $F_{2i}$ can have size $|U_{2k}|/2$, so the total size of empty bricks in $F_{0},F_{2},\ldots,F_{2k}$ is $|U_{2k}|$. Moreover, the density of squares into bricks is at least $1/2\sqrt{2}$ and by (2), we get that $\textsc{Opt}\geq\textsc{Opt}(s):=\frac{f_{k}-u_{k}}{2\sqrt{2}}+s^{2}=\frac{1-4^{-k}}{3}+s^{2}.$ In the case that $k=0$, we get $\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{\textsc{Alg}(s)}{\textsc{Opt}(s)}=\frac{2s\sqrt{2}+2s^{2}+1}{2s^{2}}.$ A simple analysis shows that the fraction is largest when $s=s_{l}$, so we get the bound $\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{2s_{l}\sqrt{2}+2s_{l}^{2}+1}{2s_{l}^{2}}=3+2\sqrt{2}<5.83$ Suppose now that $k>0$. We divide into two cases of whether $s$ is in the lower or the upper half of the range $(s_{l},s_{h}]$. For the lower half, that is, $s\in(s_{l},\frac{s_{l}+s_{h}}{2}]$, we get $\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{\textsc{Alg}(\frac{s_{l}+s_{h}}{2})}{\textsc{Opt}(s_{l})}=\frac{96\cdot 4^{k}+(24\sqrt{2}-48)\cdot 2^{k}-6\sqrt{2}+9}{16\cdot 4^{k}-4}.$ It is straightforward to check that $(24\sqrt{2}-48)\cdot 2^{k}-6\sqrt{2}+9<6\cdot(-4)$ for all $k\geq 1$, so it follows that the ratio is less than $6$. For the upper half, that is, $s\in[\frac{s_{l}+s_{h}}{2},s_{h}]$, we get $\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{\textsc{Alg}(s_{h})}{\textsc{Opt}(\frac{s_{l}+s_{h}}{2})}=\frac{96\cdot 4^{k}}{16\cdot 4^{k}+6\sqrt{2}-7}.$ As $6\sqrt{2}-7>0$, the ratio is less than $6$. * 2) We now assume that $U_{2k}$ has been split into a $L$ and $E_{2k+1}$, which are the left and right halfs of $U_{2k}$, respectively. * 2.1) We first suppose that $E_{2k+1}$ is not empty. This implies that there is no empty $(2k+1)$-brick in $F_{0},F_{2},\ldots,F_{2k},U_{2k}$. Hence, each empty brick in the bricks $F_{0},F_{2},\ldots,F_{2k},U_{2k}$ is a $(2k+2)$-brick or smaller, so these empty bricks have total size at most $u_{k}/2$. We then get $\textsc{Opt}\geq\frac{f_{k}+u_{k}-u_{k}/2}{2\sqrt{2}}=\frac{1}{3}+\frac{4^{-k}}{24}>\frac{1}{3}.$ Since $\textsc{Alg}\leq 2$, it follows that $\frac{\textsc{Alg}}{\textsc{Opt}}<6$. * 2.2) We now suppose that $E_{2k+1}$ is empty. * 2.2.1) Suppose now that $L$ has not been split into smaller bricks. Then $L$ contains a square $S$ of size $s\times s$ for $s\in(s_{l},s_{h}]=\left(\sqrt{2}^{-2k-3},\sqrt{2}^{-2k-2}\right]$. As in case 1, we get $\textsc{Alg}\leq\textsc{Alg}(s):=(\sqrt{2}-(\sqrt{2}^{-2k-1}-s))^{2}.$ Note that there is no empty $(2k+1)$-brick in the bricks $F_{0},F_{2},\ldots,F_{2k}$, so these bricks contain a total area of at most $u_{k}/2$ empty bricks. We then get $\textsc{Opt}\geq\textsc{Opt}(s):=\frac{f_{k}-u_{k}/2}{2\sqrt{2}}+s^{2}.$ We then get the bound $\frac{\textsc{Alg}}{\textsc{Opt}}\leq\frac{\textsc{Alg}(s_{h})}{\textsc{Opt}(s_{l})}={\frac{24\cdot{4}^{k}+(12\sqrt{2}-24)\cdot{2}^{k}-6\,\sqrt{2}+9}{4\cdot{4}^{k}-1}}.$ Here, it is straightforward to verify that $(12\sqrt{2}-24)\cdot{2}^{k}-6\,\sqrt{2}+9<6\cdot(-1)$ for all $k\geq 0$, and hence the ratio is less than $6$. * 2.2.2) We now assume that $L$ has been split into $F_{2k+2}$ and $U_{2k+2}$, which are the bottom and top parts, respectively. * 2.2.2.1) Suppose that $U_{2k+2}$ is empty. Since also $E_{1},E_{3},\ldots,E_{2k+1}$ are empty, we get that $\textsc{Alg}\leq(\sqrt{2}-\sqrt{2}^{-2k-1}/2)^{2}$. Note that each empty bricks in the bricks $F_{0},F_{2},\ldots,F_{2k+2}$ can have size at most $u_{k}/8$, so the total size of the empty bricks is at most $u_{k}/4=u_{k+1}$, and we get $\textsc{Opt}\geq\frac{f_{k+1}-u_{k+1}}{2\sqrt{2}}.$ We therefore get $\frac{\textsc{Alg}}{\textsc{Opt}}\leq{\frac{48\cdot{4}^{k}-24\cdot{2}^{k}+3}{8\cdot{2}^{2\,k}-2}}.$ Here, it is straightforward to check that $-24\cdot{2}^{k}+3<6\cdot(-2)$ for all $k\geq 0$, so the ratio is less than $6$. * 2.2.2.2) We are finally left with the case that $U_{2k+2}$ is not empty. But then all the requirements are satisfied for the packing to be a $(2k+2)$-packing. We now observe that the analysis is tight. To this end, we show that for any given $k$ and a small $\varepsilon>0$, we can force the algorithm to produce a $2k$-packing, such that as $k\longrightarrow\infty$ and $\varepsilon\longrightarrow 0$, the ratio $\frac{\textsc{Alg}}{\Sigma}$ tends to $6$, where $\Sigma$ is the total area of the packed squares. Let $\varepsilon_{k}:=\varepsilon\sqrt{2}^{-k}$, $\ell_{k}:=\sqrt{2}^{-k}/2+\varepsilon_{k}$, and let $S_{k}$ be a square of size $\ell_{k}\times\ell_{k}$. We now feed the algorithm with copies of $S_{k}$. This will eventually result in a $2k$-packing, where each non-empty brick is a $2k$-brick; see Figure 9 (right). Let $n_{k}$ be the number needed to produce the $2k$-packing. The density in each non-empty brick is $\rho_{\varepsilon}:=\frac{|S_{k}|}{|B_{2k}|}$. As $\varepsilon\longrightarrow 0$, we get that $\rho_{\varepsilon}\longrightarrow\frac{1}{2\sqrt{2}}$. As $k\longrightarrow\infty$, the area of non-empty bricks converges to $\frac{2|B_{\leq 0}|}{3}=\frac{2\sqrt{2}}{3}$. Hence, we have $\Sigma\longrightarrow\frac{1}{2\sqrt{2}}\cdot\frac{2\sqrt{2}}{3}=\frac{1}{3}$. We then get $\frac{\textsc{Alg}}{\Sigma}\longrightarrow\frac{2}{1/3}=6$. Furthermore, the optimal packing of the squares is to place them so that their bounding box is a square of size $\lceil\sqrt{n}_{k}\rceil\ell_{k}\times\lceil\sqrt{n}_{k}\rceil\ell_{k}$. As $k\longrightarrow\infty$, we then have $\frac{\Sigma}{\textsc{Opt}}\longrightarrow 1$. Hence, we have $\frac{\textsc{Alg}}{\textsc{Opt}}\longrightarrow 6$. ∎ ### 3.4 More lower bounds when edges are long We already saw in Corollary 14 that as a function of $n$, the competitive ratio of an algorithm for AreaTranslation or AreaRotation must be at least $\Omega(\sqrt{n})$, even when all edges have length $1$. In this section, we give lower bounds in terms of Opt for the same case. Note that the assumption that the edges are long is needed for these bounds to be matched by actual algorithms, since Corollary 13 states that without the assumption, the competitive ratio cannot be bounded as a function of Opt. ###### Theorem 21. Consider any algorithm $A$ for the problem AreaTranslation with the restriction that all edges of the given rectangles have length at least $1$. If $A$ has an asymptotic competitive ratio $f(\textsc{Opt})$ as a function of Opt, then $f(\textsc{Opt})=\Omega(\sqrt{\textsc{Opt}})$. ###### Remark 22. Note that when the edges are long, $\Omega(\sqrt{\textsc{Opt}})=\Omega(\sqrt{n})$, so this bound is stronger than the $\Omega(\sqrt{n})$ bound of Corollary 14. ###### Proof of Theorem 21.. For any $n\in\mathbb{N}$, we do as follows. We first provide $A$ with $n^{2}$ unit squares. Let the bounding box of the produced packing of these squares have size $a\times b$. Assume without loss of generality that $a\leq b$, so that $b\geq n$. We now give $A$ the rectangle $n^{2}\times 1$. The optimal offline solution to this set of rectangles has a bounding box of size $n^{2}\times 2$. The packing produced by $A$ has a bounding box of size at least $n^{2}\times n=\Omega(\sqrt{\textsc{Opt}})\cdot\textsc{Opt}$. ∎ ###### Theorem 23. Consider any algorithm $A$ for the problem AreaRotation with the restriction that all edges of the given rectangles have length at least $1$. If $A$ has a competitive ratio $f(\textsc{Opt})$ as a function of Opt, then $f(\textsc{Opt})=\Omega(\sqrt[4]{\textsc{Opt}})$. ###### Proof. For any $n\in\mathbb{N}$, we do as follows. We first provide $A$ with $n^{2}$ unit squares. Let the bounding box of the produced packing of these squares have size $a\times b$. Assume without loss of generality that $a\leq b$. If $a\geq n^{1/2}$, we give $A$ the rectangle $1\times n^{2}$. Otherwise, we have $b>n^{3/2}$, and then we give $A$ the square $n\times n$. In either case, there is an optimal offline solution of area $2n^{2}$, but the bounding box of the packing produced by $A$ has area at least $n^{5/2}=\Omega(\sqrt[4]{\textsc{Opt}})\cdot\textsc{Opt}$. ∎ ### 3.5 Algorithms when edges are long In this section, we describe algorithms that match lower bounds of Section 3.4. We analyze these algorithms under the assumption that we feed them with rectangles with edges of length at least $1$ (of course, any other positive constant will also work), but we require no bound on the aspect ratio. Under this assumption, we observe that DynBoxTrans has absolute competitive ratio $O(\sqrt{\textsc{Opt}})$ for AreaTranslation. We then describe the algorithm DynBoxRot${}_{\sqrt[4]{\textsc{Opt}}}$, which we prove to have absolute competitive ratio $O(\sqrt[4]{\textsc{Opt}})$ for AreaRotation. By Theorems 21 and 23, both algorithms are optimal to within a constant factor. In previous sections we proved lower bounds of $\Omega(\sqrt{n})$ and $\Omega(\sqrt[4]{\textsc{Opt}})$ for AreaRotation. They can be summarized stating that AreaRotation has a competitive ratio of $\Omega(\max\\{\sqrt{n},\sqrt[4]{\textsc{Opt}}\\})$. The last theorem of this section, describes the algorithm $\textsc{DynBoxRot}_{\sqrt{n}\,\wedge\sqrt[4]{\textsc{Opt}}}$ that simultaneously matches both lower bounds achieving a competitive ratio of $O(\min\\{\sqrt{n},\sqrt[4]{\textsc{Opt}}\\}$. At a first sight it may seem that this algorithm contradicts the lower bound of $\Omega(\max\\{\sqrt{n},\sqrt[4]{\textsc{Opt}}\\})$; however this simply proves that the _edge cases_ that have a competitive ratio of at least $\Omega(\sqrt[4]{\textsc{Opt}})$ must satisfy $\textsc{Opt}=O(n^{2})$. Likewise, those for which the competitive ratio is at least $\Omega(\sqrt{n})$ satisfy $n=O(\sqrt{\textsc{Opt}})$. #### Translations only Under the long edge assumption, we have $n\leq\textsc{Opt}$. Therefore, DynBoxTrans achieves a competitive ratio of $O(\sqrt{n})=O(\sqrt{\textsc{Opt}})$ for AreaTranslation and matches the bound stated in Theorem 21. #### Rotations allowed Now we tackle the AreaRotation problem and describe the algorithm DynBoxRot${}_{\sqrt[4]{\textsc{Opt}}}$. We define the threshold function $T_{j}=\Sigma_{j}^{3/4}+7H_{j}$, where $H_{j}=\max_{i=1,\dots,j}h_{i}$ and $\Sigma_{j}$ is the total area of pieces $p_{1},\dots,p_{j}$. $\textsc{DynBoxRot}_{\sqrt[4]{\textsc{Opt}}}$ is obtained by running DynBoxRot, as described in Section 3.2, employing this new threshold $T_{j}$. ###### Theorem 24. The algorithm DynBoxRot${}_{\sqrt[4]{\textsc{Opt}}}$ has an absolute competitive ratio of $O(\sqrt[4]{\textsc{Opt}})$ for the problem AreaRotation, where Opt is the area of the optimal offline packing. ###### Proof. This proof is similar to the one of Theorem 16. Define $W:=\max_{i=1,\dots,n}w_{i}$. Recall that in DynBoxRot we preprocess every piece $p$ rotating it so the $w_{p}\leq h_{p}$, hence $W\leq\sqrt{\Sigma_{n}}$. Let $B_{k}$ be the last active box, so that we can enclose all the pieces in a bounding box of size $2^{k+1}\times T_{n}$, and bound the area returned by the algorithm as $\textsc{Alg}=O(2^{k}H_{n}+2^{k}\Sigma_{n}^{3/4})$. On the other hand we are able to bound the optimal offline packing as $\textsc{Opt}=\Omega(\Sigma_{n}+WH_{n})$. If the active box never changed, then we have $2^{k}<2W$ that implies $\textsc{Alg}=O(WH_{n}+\Sigma_{n}^{5/4})=\textsc{Opt}\cdot O(\sqrt[4]{\textsc{Opt}})$. Otherwise, let $B_{\ell}$ be the last active box before $B_{k}$, and $p_{j}$ be the first piece put in $B_{k}$. Here we have two cases. Case (1) [_$w_{j} >2^{\ell}$_] In this case we have $2^{k}<2W$ that implies $\textsc{Alg}=O(WH_{n}+\Sigma_{n}^{5/4})=\textsc{Opt}\cdot O(\sqrt[4]{\textsc{Opt}})$. Case (2) [_$w_{j}\leq 2^{\ell}$_] In this case we have $k=\ell+1$. Denote with $\widetilde{H_{i}}$ the total height of shelves in $B_{i}$. Then we have $\widetilde{H_{\ell}}\geq T_{j}-H_{j}=\Sigma_{j}^{3/4}+6H_{j}$, otherwise we could pack $p_{j}$ in $B_{\ell}$. Thus, we can apply Lemma 15 and conclude that the box $B_{\ell}$ of size $2^{\ell}\times T_{j}$ is filled with constant density. Here we have two cases. Case (2.1) [_$\widetilde{H_{k}}\leq T_{j}$_] In this case we have $\textsc{Alg}=O(2^{k}T_{j})$ and, thanks to the constant density packing of $B_{\ell}$ we have $\Sigma_{j}=\Theta(2^{\ell}\widetilde{H_{\ell}})=\Theta(2^{k}T_{j})$. Since $\textsc{Opt}\geq\Sigma_{j}$, we get $\textsc{Alg}=O(\textsc{Opt})$. Case (2.2) [_$\widetilde{H_{k}} >T_{j}$_] In this case we have $\textsc{Alg}=O(2^{k}\widetilde{H_{k}})$. Moreover, $\widetilde{H_{k}}=O(H_{n}+\Sigma_{n}/2^{k})$, in fact if $2^{s-1}<H_{n}\leq 2^{s}$, then the total height of sparse shelves is $\sum_{i\leq s}2^{i}=2^{s+1}=O(H_{n})$. Furthermore, dense shelves are filled with constant density, therefore their total height is at most $O(\Sigma_{n}/2^{k})$. Finally, we need to show that $2^{k}=O(\sqrt[4]{\Sigma_{n}})$. Thanks to the constant density packing of $B_{\ell}$, we have $2^{k}\Sigma_{j}^{3/4}=O(2^{\ell}T_{j})=O(\Sigma_{j})$. Dividing both sides by $\Sigma_{j}^{3/4}$ we get $2^{k}=O(\Sigma_{j}^{1/4})$. In the end notice that, thanks to the long edge hypotheses $H_{n}\leq\Sigma_{n}$ and we have $\textsc{Alg}=O(2^{k}\widetilde{H_{k}})=O(2^{k}H_{n}+\Sigma_{n})=O(\Sigma_{n}^{5/4})=\textsc{Opt}\cdot O(\sqrt[4]{\textsc{Opt}})$. ∎ So far we managed to match the competitive ratio lower bounds of $\Omega(\sqrt{n})$ and $\Omega(\sqrt[4]{\textsc{Opt}})$ employing two different algorithms: DynBoxRot and $\textsc{DynBoxRot}_{\sqrt[4]{\textsc{Opt}}}$. A natural question is whether is it possible to match the performance of these algorithms simultaneously, having an algorithm that achieves a competitive ratio of $O(\min\\{\sqrt{n},\sqrt[4]{\textsc{Opt}}\\})$. We give an affirmative answer by describing the algorithm $\textsc{DynBoxRot}_{\sqrt{n}\,\wedge\sqrt[4]{\textsc{Opt}}}$. Again, we employ the same scheme of DynBoxRot with a different threshold function. This time the definition of $T_{j}$ is slightly more involved. First define $\widetilde{T}_{j}=\begin{cases}\Sigma_{j}^{3/4}+7H_{j},&\text{ if }\Sigma_{j}<j^{2}\\\ H_{j}\sqrt{n}+7H_{j},&\text{ otherwise.}\end{cases}$ Later we will write $\widetilde{T}_{j}$ as $\widetilde{T}_{j}=\mathbbm{1}_{\\{\Sigma_{j}<j^{2}\\}}\cdot\Sigma_{j}^{3/4}+\mathbbm{1}_{\\{\Sigma_{j}\geq{j}^{2}\\}}\cdot H_{j}\sqrt{n}+7H_{j}$. We now define $T_{j}=\begin{cases}0,&\text{ if }j=0\\\ \max\left\\{T_{j-1},\widetilde{T_{j}}\right\\},&\text{ if }j\geq 1.\end{cases}$ This two-step definition is necessary for a correct implementation of the algorithm because we must guarantee that $T_{j}$ does not decrease. ###### Theorem 25. When used on the problem AreaRotation, the algorithm $\textsc{DynBoxRot}_{\sqrt{n}\,\wedge\sqrt[4]{\textsc{Opt}}}$ has an absolute competitive ratio of $O(\min\\{\sqrt{n},\sqrt[4]{\textsc{Opt}}\\})$, where Opt is the area of the optimal offline packing and $n$ is the total number of pieces in the stream. ###### Proof. Again, we define $W:=\max_{i=1,\dots,n}w_{i}$. Recall that in DynBoxRot we preprocess every piece $p$ rotating it so the $w_{p}\leq h_{p}$, hence $W\leq\sqrt{\Sigma_{n}}$. Let $B_{k}$ be the last active box, so that we can enclose all the pieces in a bounding box of size $2^{k+1}\times T_{n}$. There exists a $n^{\prime}\leq n$ such that $T_{n}=\widetilde{T}_{n^{\prime}}$. We can bound the area returned by the algorithm as $\textsc{Alg}=O\left(2^{k}\widetilde{T}_{n^{\prime}}\right)=O\left(2^{k}H_{n^{\prime}}+\mathbbm{1}_{\\{\Sigma_{n^{\prime}}<{n^{\prime}}^{2}\\}}\cdot 2^{k}\Sigma_{n^{\prime}}^{3/4}+\mathbbm{1}_{\\{\Sigma_{n^{\prime}}\geq{n^{\prime}}^{2}\\}}\cdot 2^{k}H_{n^{\prime}}\sqrt{{n^{\prime}}}\right).$ We bound the optimal offline packing as $\textsc{Opt}=\Omega(\Sigma_{n}+WH_{n})$. If the active box never changed, then we have $2^{k}<2W$ that implies $\displaystyle ALG$ $\displaystyle=O\left(WH_{n}+\mathbbm{1}_{\\{\Sigma_{n^{\prime}}<{n^{\prime}}^{2}\\}}\cdot W\Sigma_{n^{\prime}}^{3/4}+\mathbbm{1}_{\\{\Sigma_{n^{\prime}}\geq{n^{\prime}}^{2}\\}}\cdot WH_{n^{\prime}}\sqrt{{n^{\prime}}}\right)$ $\displaystyle=O\left(WH_{n}+\mathbbm{1}_{\\{\Sigma_{n^{\prime}}<{n^{\prime}}^{2}\\}}\cdot\Sigma_{n}\sqrt[4]{\Sigma_{n^{\prime}}}+\mathbbm{1}_{\\{\Sigma_{n^{\prime}}\geq{n^{\prime}}^{2}\\}}\cdot WH_{n}\sqrt{{n^{\prime}}}\right)$ $\displaystyle\leq\textsc{Opt}\cdot O\left(\min\left\\{\sqrt[4]{\Sigma_{n^{\prime}}},\sqrt{n^{\prime}}\right\\}\right)=\textsc{Opt}\cdot O\left(\min\left\\{\sqrt[4]{\textsc{Opt}},\sqrt{n}\right\\}\right).$ Otherwise, let $B_{\ell}$ be the last active box before $B_{k}$, and $p_{j}$ be the first piece put in $B_{k}$. Here we have two cases. Case (1) [_$w_{j} >2^{\ell}$_] In this case we have, again, $2^{k}<2W$ and we use the same argument employed above. Case (2) [_$w_{j}\leq 2^{\ell}$_] In this case we have $k=\ell+1$. Denote with $\widetilde{H_{i}}$ the total height of shelves in $B_{i}$. Then we have $\widetilde{H_{\ell}}\geq T_{j}-H_{j}\geq\widetilde{T}_{j}-H_{j}\geq 6H_{j}$, otherwise we could pack $p_{j}$ in $B_{\ell}$. Thus, we can apply Lemma 15 and conclude that the box $B_{\ell}$ of size $2^{\ell}\times T_{j}$ is filled with constant density. Here we have two cases. Case (2.1) [_$\widetilde{H_{k}}\leq T_{j}$_] In this case we have $\textsc{Alg}=O(2^{k}T_{j})$ and, thanks to the constant density packing of $B_{\ell}$ we have $\Sigma_{j}=\Theta(2^{\ell}\widetilde{H_{\ell}})=\Theta(2^{k}T_{j})$. Since $\textsc{Opt}\geq\Sigma_{j}$, we get $\textsc{Alg}=O(\textsc{Opt})$. Case (2.2) [_$\widetilde{H_{k}} >T_{j}$_] In this case we still have $\textsc{Alg}=O(2^{k}\widetilde{H_{k}})$. Moreover, $\widetilde{H_{k}}=O(H_{n}+\Sigma_{n}/2^{k})$, in fact if $2^{s-1}<H_{n}\leq 2^{s}$, then the total height of sparse shelves is $\sum_{i\leq s}2^{i}=2^{s+1}=O(H_{n})$. Furthermore, dense shelves are filled with constant density, therefore their total height is at most $O(\Sigma_{n}/2^{k})$. Finally, we need to show that $2^{k}=O(\min\\{\sqrt[4]{\Sigma_{n}},\sqrt{n}\\})$. Let $T_{j}=\widetilde{T}_{j^{\prime}}$, we have two cases. Case (2.2.1) [_$\Sigma_{j^{\prime}} <{j^{\prime}}^{2}$_] We have $\widetilde{T}_{j^{\prime}}\geq\Sigma_{j^{\prime}}^{3/4}$. And thanks to the constant density packing of $B_{\ell}$, we have also $2^{k}\Sigma_{j^{\prime}}^{3/4}=O(2^{\ell}T_{j})=O(\Sigma_{j^{\prime}})$. Dividing both sides by $\Sigma_{j^{\prime}}^{3/4}$ we get $2^{k}=O(\sqrt[4]{\Sigma_{j^{\prime}}})$. Case (2.2.2) [_$\Sigma_{j^{\prime}}\geq{j^{\prime}}^{2}$_] In this case we have $\widetilde{T}_{j^{\prime}}\geq H_{j^{\prime}}\sqrt{j^{\prime}}$. Using the constant density argument we get $2^{k}H_{j^{\prime}}\sqrt{j^{\prime}}=O(2^{k}\widetilde{T}_{j^{\prime}})=O(\Sigma_{j^{\prime}})\leq O(j^{\prime}\cdot WH_{j^{\prime}})$. Dividing both sides by $H_{j^{\prime}}\sqrt{j^{\prime}}$ we obtain $2^{k}=O(W\sqrt{j^{\prime}})$. Therefore, we have $2^{k}=\begin{cases}O(\sqrt[4]{\Sigma_{j^{\prime}}})&\text{ if }\Sigma_{j^{\prime}}<{j^{\prime}}^{2}\\\ W\sqrt{j^{\prime}}&\text{otherwise.}\end{cases}$ In the end notice that, thanks to the long edge hypotheses $H_{n}\leq\Sigma_{n}$, thus Alg $\displaystyle=O\left(2^{k}\widetilde{H_{k}}\right)=O\left(2^{k}H_{n}+\Sigma_{n}\right)$ $\displaystyle=O\left(\mathbbm{1}_{\\{\Sigma_{j^{\prime}}<{j^{\prime}}^{2}\\}}\cdot H_{n}\sqrt[4]{\Sigma_{j^{\prime}}}+\mathbbm{1}_{\\{\Sigma_{j^{\prime}}\geq{j^{\prime}}^{2}\\}}\cdot WH_{n}\sqrt{j^{\prime}}+\Sigma_{n}\right)$ $\displaystyle\leq\textsc{Opt}\cdot O\left(\min\left\\{\sqrt[4]{\Sigma_{j^{\prime}}},\sqrt{j^{\prime}}\right\\}\right)=\textsc{Opt}\cdot O\left(\min\left\\{\sqrt[4]{\textsc{Opt}},\sqrt{n}\right\\}\right).$ ∎ ## 4 Further questions It is natural to consider problems where the given pieces are more general, such as convex polygons. Here, we may allow the pieces to be rotated by arbitrary angles. In that case, it follows from the technique described by Alt [2] that one can obtain a constant competitive ratio for computing a packing with a minimum perimeter bounding box: For each new piece, we rotate the piece so that a diameter of the piece is horizontal. We then use the algorithm BrickRotation to pack the bounding boxes of the pieces. Since the area of each piece is at least half of the area of its bounding box, the density of the produced packing is at least half of the density of the packing of the bounding boxes. This results in an increase of the competitive ratio by a factor of at most $\sqrt{2}$. For the problem of minimizing the perimeter of the bounding box (or convex hull) with convex polygons as pieces and only translations allowed, we do not know if it is possible to get a competitive ratio of $O(1)$, and this seems to be a very interesting question for future research. In order to design such an algorithm, it would be sufficient to show that for some constants $\delta>0$ and $\Sigma>0$, there is an online algorithm that packs any stream of convex polygons of diameter at most $\delta$ and total area at most $\Sigma$ into the unit square, which is in itself an interesting problem. The three-dimensional version of this question has a negative answer, even for offline algorithms: Alt, Cheong, Park, and Scharf [3] showed that for any $n\in\mathbb{N}$, there exists a finite number of 2D unit disks embedded in 3D that cannot all be packed by translation in a cube with edges of length $n$. ## References * [1] Hee-Kap Ahn and Otfried Cheong. Aligning two convex figures to minimize area or perimeter. Algorithmica, 62(1-2):464–479, 2012. * [2] Helmut Alt. Computational aspects of packing problems. Bulletin of the EATCS, 118, 2016. * [3] Helmut Alt, Otfried Cheong, Ji-won Park, and Nadja Scharf. Packing 2D disks into a 3D container. In International Workshop on Algorithms and Computation (WALCOM 2019), pages 369–380, 2019. * [4] Helmut Alt, Mark de Berg, and Christian Knauer. Approximating minimum-area rectangular and convex containers for packing convex polygons. In 23rd Annual European Symposium on Algorithms (ESA 2015), pages 25–34, 2015. * [5] Helmut Alt and Ferran Hurtado. Packing convex polygons into rectangular boxes. In 18th Japanese Conference on Discrete and Computational Geometry (JCDCGG 2000), pages 67–80, 2000. * [6] Brenda S. Baker and Jerald S. Schwarz. Shelf algorithms for two-dimensional packing problems. SIAM Journal on Computing, 12(3):508–525, 1983. * [7] Allan Borodin and Ran El-Yaniv. Online computation and competitive analysis. Cambridge University Press, 2005. * [8] Brian Brubach. Improved bound for online square-into-square packing. In 12th International Workshop on Approximation and Online Algorithms (WAOA 2014), pages 47–58, 2014. * [9] Henrik I. Christensen, Arindam Khan, Sebastian Pokutta, and Prasad Tetali. Approximation and online algorithms for multidimensional bin packing: A survey. Computer Science Review, 24:63–79, 2017. * [10] Fan Chung and Ron Graham. Efficient packings of unit squares in a large square. Discrete & Computational Geometry, 2019. * [11] János Csirik and Gerhard J. Woeginger. On-line packing and covering problems. In Amos Fiat and Gerhard J. Woeginger, editors, Online Algorithms: The State of the Art, pages 147–177. Springer, 1998. * [12] Paul Erdős and Ron Graham. On packing squares with equal squares. Journal of Combinatorial Theory, Series A, 19(1):119–123, 1975\. * [13] Sándor P. Fekete and Hella-Franziska Hoffmann. Online square-into-square packing. Algorithmica, 77(3):867–901, 2017. * [14] Amos Fiat and Gerhard J. Woeginger. Competitive analysis of algorithms. In Amos Fiat and Gerhard J. Woeginger, editors, Online Algorithms: The State of the Art, pages 1–12. Springer, 1998. * [15] Janusz Januszewski and Marek Lassak. On-line packing sequences of cubes in the unit cube. Geometriae Dedicata, 67(3):285–293, 1997. * [16] Marek Lassak. On-line potato-sack algorithm efficient for packing into small boxes. Periodica Mathematica Hungarica, 34(1-2):105–110, 1997. * [17] Hyun-Chan Lee and Tony C. Woo. Determining in linear time the minimum area convex hull of two polygons. IIE Transactions, 20(4):338–345, 1988. * [18] Boris D. Lubachevsky and Ronald L. Graham. Dense packings of congruent circles in rectangles with a variable aspect ratio. In Boris Aronov, Saugata Basu, János Pach, and Micha Sharir, editors, Discrete and Computational Geometry: The Goodman-Pollack Festschrift, pages 633–650. 2003. * [19] Boris D. Lubachevsky and Ronald L. Graham. Minimum perimeter rectangles that enclose congruent non-overlapping circles. Discrete Mathematics, 309(8):1947–1962, 2009. * [20] Victor J. Milenkovic. Translational polygon containment and minimal enclosure using linear programming based restriction. In Proceedings of the twenty-eighth annual ACM symposium on Theory of Computing (STOC 1996), pages 109–118, 1996. * [21] Victor J. Milenkovic. Rotational polygon containment and minimum enclosure using only robust 2D constructions. Computational Geometry, 13(1):3–19, 1999. * [22] Victor J. Milenkovic and Karen Daniels. Translational polygon containment and minimal enclosure using mathematical programming. International Transactions in Operational Research, 6(5):525–554, 1999. * [23] Dongwoo Park, Sang Won Bae, Helmut Alt, and Hee-Kap Ahn. Bundling three convex polygons to minimize area or perimeter. Computational Geometry, 51:1–14, 2016. * [24] E. Specht. High density packings of equal circles in rectangles with variable aspect ratio. Computers & Operations Research, 40(1):58 –69, 2013. * [25] Rob van Stee. SIGACT news online algorithms column 20: the power of harmony. SIGACT News, 43(2):127–136, 2012. * [26] Rob van Stee. SIGACT news online algorithms column 26: Bin packing in multiple dimensions. SIGACT News, 46(2):105–112, 2015.
Let $(X,g)$ be a compact manifold with boundary $M^n$ and $\sigma$ a defining function of $M$. To these data, we associate natural conformally covariant polynomial one-parameter families of differential operators $C^\infty(X) \to C^\infty(M)$. They arise through a residue construction which generalizes an earlier construction in the framework of Poincaré-Einstein metrics and are referred to as residue families. Residue families may be viewed as curved analogs of conformal symmetry breaking differential operators. The main ingredient of the definition of residue families are eigenfunctions of the Laplacian of the singular metric $\sigma^{-2}g$. We prove that if $\sigma$ is an approximate solution of a singular Yamabe problem, i.e., if $\sigma^{-2}g$ has constant scalar curvature $-n(n+1)$, up to a sufficiently small remainder, these families can be written as compositions of certain degenerate Laplacians (Laplace-Robin operators). This result implies that the notions of extrinsic conformal Laplacians and extrinsic $Q$-curvature introduced in recent works by Gover and Waldron can naturally be rephrased in terms of residue families. This new spectral theoretical perspective allows easy new proofs of several results of Gover and Waldron. Moreover, it enables us to relate the extrinsic conformal Laplacians and the critical extrinsic $Q$-curvature to the scattering operator of the asymptotically hyperbolic metric $\sigma^{-2}g$ extending the work of Graham and Zworski. The relation to the scattering operators implies that the extrinsic conformal Laplacians are self-adjoint. We describe the asymptotic expansion of the volume of a singular Yamabe metric in terms of Laplace-Robin operators (reproving results of Gover and Waldron). We also derive new local holographic formulas for all extrinsic $Q$-curvatures (critical and sub-critical ones) in terms of renormalized volume coefficients, the scalar curvature of the background metric, and the asymptotic expansions of eigenfunctions of the Laplacian of the singular metric $\sigma^{-2}g$. These results naturally extend earlier results in the Poincaré-Einstein case. Furthermore, we prove a new formula for the singular Yamabe obstruction $\B_n$. The simple structure of these formulas shows the benefit of the systematic use of so-called adapted coordinates. We use the latter formula for $\B_n$ to derive explicit expressions for the obstructions in low-order cases (confirming earlier results). Finally, we relate the obstruction $\B_n$ to the supercritical $Q$-curvature $\QC_{n+1}$. [2020]Primary 35J30 53B20 53B25 53C18; Secondary 35J70 35Q76 53C25 58J50 August 27, 2024 § INTRODUCTION Conformal differential geometry studies natural geometric quantities associated to Riemannian (and pseudo-Riemannian) manifolds (such as curvature invariants and natural differential operators) which transform nicely under conformal changes of the metric. In recent years, it has developed in profound and surprising ways, which are connected to the spectral theory of Laplace-type operators, scattering theory, holographic principles (AdS-CFT correspondence), Cartan geometry, non-linear partial differential equations, and representation theory. Particular powerful tools are tractor calculus and the Poincaré-Einstein and ambient metrics in the sense of Fefferman-Graham <cit.>. A central role in those parts of geometric analysis which are related to conformal differential geometry play the conformally invariant powers of the Laplacian, which are known as GJMS-operators <cit.> and the related Branson's $Q$-curvatures <cit.>. The GJMS-operators $P_{2N}(h)$ act on the space $C^\infty(M)$ on any Riemannian manifold $(M,h)$. They are of the form $\Delta_h^N + \mbox{lower-order terms}$, where the lower-order terms are given in terms of the metric and covariant derivatives of its curvature. We recall that for general $h$, the operators $P_{2N}(h)$ exist for all orders $2N$ if $n$ is odd but only for $2N \le n$ if $n$ is even. The GJMS-operators $P_{2N}$ generalize the conformal Laplacian (Yamabe P_2 = \Delta - \left(\frac{n}{2}-1\right) \J, \quad 2 (n-1) \J = \scal and the Paneitz operator P_4 = \Delta^2 - \delta ((n-2) \J - 4 \Rho) d + \left(\frac{n}{2}-2\right) \left(\frac{n}{2}\J^2 - 2 |\Rho|^2 - \Delta (\J)\right), where $\Rho$ is the Schouten tensor of $h$ (for the notation, we refer to Section <ref>). The scalar curvature quantities Q_2 = \J \quad \mbox{and} \quad Q_4 = \frac{n}{2} \J^2 - 2 |\Rho|^2 - \Delta(\J) are the lowest-order cases of Branson's $Q$-curvatures. In particular, the quantity $Q_4$ plays a central role in geometric analysis <cit.>. In <cit.>, one of the authors analyzed the structure of GJMS-operators and Branson's $Q$-curvatures of a Riemannian manifold $(M^n,h)$ through a theory of conformally covariant polynomial one-parameter families of differential operators $C^\infty(X) \to C^\infty(M)$, where $X^{n+1}$ is a manifold of one more dimension. Such an extrinsic perspective on objects living on $M$ is sometimes referred to as a holographic point of view following <cit.>. These families of differential operators can be regarded as curved versions of certain intertwining operators in representation theory. Following the framework introduced in a series of works by T. Kobayashi and his coauthors, these intertwining operators are now known as conformal symmetry breaking operators $C^\infty(S^{n+1}) \to C^\infty(S^n)$. The notion is motivated by the fact that they are equivariant only with respect to the subgroup of the conformal group of $S^{n+1}$ consisting of diffeomorphisms that leave the equatorial subsphere $S^n \hookrightarrow S^{n+1}$ invariant. Such intertwining operators acting on functions have analogs acting on sections of homogeneous vector bundles on spheres as well as in contexts where other groups replace the conformal group of the sphere. We refer the interested reader to <cit.>. In the curved case, the conformal covariance property takes the role of the intertwining property, and the definition of the operators uses a Poincaré-Einstein metric in the sense of Fefferman and Graham <cit.>. The curved versions of the symmetry breaking operators then are defined in terms of the asymptotic expansions of eigenfunctions of the Laplacian of a Poincaré-Einstein metric on $X$ and the resulting so-called renormalized volume coefficients. The latter quantities are curvature invariants of a metric on $M$, the study of which originally was motivated by the AdS/CFT-correspondence <cit.>. Since the construction of the curved versions of the symmetry breaking differential operators can be expressed as a residue construction, these were termed residue families in <cit.>. A key feature of the theory of residue families is that each GJMS-operator $P_{2N}(h)$ comes along with a residue family (see (<ref>)), and one can effectively utilize the family parameter to study the structure of these special values <cit.>. Another basic perspective on GJMS-operators and $Q$-curvatures of a metric $h$ on $M$ was developed in <cit.> by showing how the scattering operator $\Sc(\lambda)$ of the Laplacian of a Poincaré-Einstein metric on $X$ in normal form relative to $h$ naturally encodes both quantities. The scattering operator $\Sc(\lambda)$ is a one-parameter meromorphic family of conformally covariant pseudo-differential operators, which may be regarded as a curved version of the Knapp-Stein intertwining operator for spherical principal series representations. Thus, both $\Sc(\lambda)$ and residue families describe GJMS-operators and $Q$-curvatures. But whereas the former are meromorphic families of pseudo-differential operators, the latter are polynomial families of differential In <cit.>, Gover and Waldron developed an alternative perspective on the above holographic descriptions of GJMS-operators and $Q$-curvatures. In this approach, the conformal geometry of the metric $h$ on $M$ and the associated Poincaré-Einstein metric $g_+$ on $X$ play a central role. In particular, this leads to a deeper understanding of the role of the Einstein property of $g_+$ in constructions of GJMS-operators. The conformal tractor calculus on $X$ naturally associates a conformally covariant polynomial one-parameter family of second-order differential operators on functions on $X$ to a metric $g$ on $X$ and a defining function $\sigma$ of the hypersurface $\iota^*: M \hookrightarrow X$. Later, Gover and Waldron termed this operator the Laplace-Robin operator <cit.>.[In <cit.>, the Laplace-Robin operators are regarded as operators acting on densities.] We shall follow this terminology here. The Laplace-Robin operators degenerate on $M$ in the sense that the coefficients of its leading parts vanish on $M$. The data $(g,\sigma)$ induce a metric $\iota^*(g)$ on $M$ and a singular metric $\sigma^{-2} g$ on the complement of $M$ in $X$. It is the latter metric that generalizes the Poincaré-Einstein metric. Now, if $\sigma^{-2} g$ is Einstein, then certain compositions of Laplace-Robin operators on $X$ reduce to GJMS-operators on $M$. Even if the Einstein condition is violated, analogous compositions of (renormalized) Laplace-Robin operators still reduce to conformally covariant operators on $M$. We emphasize that the notion of conformal covariance in the latter statement concerns constructions that depend on a metric $g$ on $X$ and a defining function $\sigma$ of the boundary $M$ of $X$, i.e., constructions which are invariant under conformal changes of both $g$ and $\sigma$: $\hat{g} = e^{2\varphi} g$ and $\hat{\sigma} = e^\varphi \sigma$. Note that $\sigma^{-2} g$ is invariant under such substitutions. Now a version of the singular Yamabe problem asks to find (for a given metric $g$ on $X$) a defining function $\sigma$ so that the scalar curvature of $\sigma^{-2}g$ is a negative constant. This conformally invariant condition for pairs $(g,\sigma)$ actually determines the defining function $\sigma$ by the background metric $g$ (at least to some extent) and the embedding $\iota$. If $\sigma$ solves the singular Yamabe problem, the above constructions of conformally covariant operators on $M$ only depend on the metric $g$ and the embedding $\iota$. In other words, these operators live on $M$, their definition depends on the embedding of $M \hookrightarrow X$, and they are conformally covariant with respect to conformal changes of the background metric $g$ on $X$. Gover, Waldron, and coauthors used this idea in a series of works to develop a theory of conformal invariants of hypersurfaces. To a large extent, this theory rests on the conformal tractor calculus <cit.>. In recent years, Laplace-Robin operators independently also appeared in other contexts, albeit not under this name. As described above, their original and most general definition comes from conformal tractor calculus. But already in <cit.>, it was observed that a special case was crucial in <cit.> in the setting of Poincaré-Einstein metrics. Some other special cases were found to be interesting in representation theory. Clerc <cit.> proved that the symmetry breaking differential operators $C^\infty(\R^{n+1}) \to C^\infty(\R^n)$ introduced in <cit.> also arise as compositions of some second-order intertwining operator for spherical principal series representations on functions on $\R^{n+1}$. In <cit.>, an attempt to generalize the result of Clerc led to a description of the general residue families of <cit.> as compositions of some second-order operators, which turned out to be Laplace-Robin operators. In other words, the latter result could be viewed as a curved version of the quoted result of Clerc. In <cit.>, the notions of spectral shift operators and Bernstein-Sato operators are used in the context of symmetry breaking operators on functions, differential forms, and spinors. These results suggest extensions of the notion of Laplace-Robin operators on forms. Such extensions on forms using tractor calculus have been developed in the monograph <cit.>. The present work sheds new light on the above results of Gover and Waldron. We connect the various strands of development by introducing an extension of the notion of residue families. Following the principles of <cit.>, we prove a series of new results, and - as by-products - we confirm and give alternative proofs of some already known results. The treatment replaces tractor calculus and distributional calculus by more classical arguments. Among other things, we stress the role of the scattering operator extending <cit.>. We hope that our methods enhance the understanding of the We continue with a more detailed description of the main new results. We consider a compact Riemannian manifold $(X,g)$ with boundary $M$. Let $\iota: M \hookrightarrow X$ be the canonical embedding. Let $h = \iota^*(g)$ be the induced metric on $M$ and $\sigma \in C^\infty(X)$ a boundary defining function, i.e., $\sigma \ge 0$ on $X$, $\sigma^{-1}(0)=M$ and $d\sigma|_M \ne 0$. The data $(g,\sigma)$ define a Laplace-Robin operator <cit.> \begin{equation*} L(g,\sigma;\lambda) \st (n+2\lambda-1) (\nabla_{\grad_g (\sigma)} + \lambda \rho) - \sigma (\Delta_g + \lambda \J), \; \lambda \in \C \end{equation*} on $C^\infty(X)$. Here $2n \J = 2n \J^g = \scal^g$ and $(n+1) \rho(g,\sigma) = -\Delta_g (\sigma) - \sigma \J$. The significance of this operator rests on its conformal covariance property \begin{equation}\label{CTL-basic} L(\hat{g}, \hat{\sigma}; \lambda) \circ e^{\lambda\varphi} = e^{(\lambda-1)\varphi} \circ L(g,\sigma;\lambda), \; \lambda \in \C \end{equation} for all conformal changes $(\hat{g},\hat{\sigma})=(e^{2\varphi}g,e^{\varphi}\sigma)$ with $\varphi \in C^\infty(X)$. This property implies that all \begin{equation}\label{LR-basic} \st L(g,\sigma;\lambda\!-\!N\!+\!1) \circ \cdots \circ L(g,\sigma;\lambda), \; N \in \N \end{equation} are conformally covariant: $L_N(\lambda)$ shifts the conformal weight $\lambda$ into $\lambda-N$. The notion of residue families has its origin in a far-reaching generalization of Gelfand's distribution $r_+^\lambda$ <cit.>. As the first generalization of $r_+^\lambda$ we consider the distribution \left\langle M(\lambda),\psi \right\rangle = \int_X \sigma^\lambda \psi dvol_g, \; \Re(\lambda) > -1 with $\psi \in C_c^\infty(X)$. Later $\sigma$ will be a solution of the singular Yamabe problem, i.e., $\sigma^{-2}g$ has constant scalar curvature $-n(n+1)$.[Then $\sigma$ is not necessarily $C^\infty$ up to the boundary.] Note that $\left\langle M(0),1 \right\rangle = \int_X dvol_g$. The one-parameter family of distributions $M(\lambda)$ admits a meromorphic continuation to $\C$ with simple poles in $-\N$. In order to describe the residues of $M(\lambda)$, we introduce coordinates on $X$ near its boundary as follows. Let $\mathfrak{X} \st \NV/|\NV|^2$ with $\NV = \grad (\sigma)$ and define a local diffeomorphism \eta: I \times M \to X, \; (s,x) \mapsto \Phi^s_{\mathfrak{X}}(x), \quad I = (0,\varepsilon) using the flow $\Phi^s_\mathfrak{X}$ of $\mathfrak{X}$. Such coordinates will be called adapted coordinates. Then $\eta^*(\sigma) = s$ and $\sigma^\lambda$ pulls back to $s^\lambda$. Moreover, it holds dvol_{\eta^*(g)} = v(s) ds dvol_h with some $v \in C^\infty(I \times M)$. Hence studying the residues of $M(\lambda)$ reduces to studying the residues of \left \langle M(\lambda),\psi \right\rangle = \int_I \int_M s^\lambda v(s) \psi(s,x) ds dvol_h. Now Gelfand's formula \Res_{\lambda=-N-1} \left(\int_0^\infty r^\lambda \psi dr \right) = \frac{\psi^{(N)}(0)}{N!} shows that \begin{equation}\label{M-res-g} \Res_{\lambda=-N-1} \left \langle M(\lambda),\psi \right \rangle = \frac{1}{N!} \int_M (v\psi)^{(N)}(0) dvol_h. \end{equation} In particular, if $\psi = 1$ near the boundary, then \begin{equation}\label{M-res} \Res_{\lambda=-N-1} \left \langle M(\lambda),\psi \right \rangle = \int_M v_N dvol_h, \end{equation} where the expansion $v(s) = \sum_{j \ge 0} s^j v_j$ defines the coefficients $v_j(g,\sigma) \in C^\infty(M)$. The above definitions immediately imply that the residue of $\left \langle M(\lambda),1\right\rangle$ at $\lambda=-n-1$ is a conformal invariant in the sense that I(\hat{g},\hat{\sigma}) = I(g,\sigma), \quad I(g,\sigma) \st \int_M v_n dvol_g. where $\hat{g} = e^{2\varphi}g$ and $\hat{\sigma} = e^{\varphi} \sigma$. Moreover, the formula \left\langle M(\lambda),1 \right\rangle = \int_X \sigma^{\lambda+n+1} dvol_{\sigma^{-2}g}, \; \Re(\lambda) > -1 suggests to regard the finite part of $\left\langle M(\lambda),1\right\rangle$ at $\lambda = -n-1$ as a renormalized volume of the singular metric $\sigma^{-2}g$. In fact, this should be called the Riesz renormalization of the volume of $\sigma^{-2} g$ <cit.>. In Theorem <ref> and Theorem <ref>, we shall return to the issue of renormalized volumes. We shall see that the residues in (<ref>) for $N \le n$ determine the singular terms in the Hadamard renormalization of the volume of $\sigma^{-2}g$. Similar renormalization techniques have also been used in the context of Möbius invariant energies of knots and their generalizations <cit.>. Now we continue with a definition of residue families. Here we assume that $|d\sigma|^2_g = 1$ on $M$. This assumption implies that the metric $\sigma^{-2}g$ is asymptotically hyperbolic. The data $(g,\sigma)$ give rise to a holomorphic one-parameter family \begin{equation}\label{MU} \lambda \mapsto \langle M_u(\lambda),\psi \rangle \st \int_X \sigma^\lambda u \psi dvol_g, \; \Re(\lambda) \gg 0 \end{equation} of distributions $M_u(\lambda)$ on $X$. Here $\psi$ is a test function on $X$ with support up to the boundary, and the additional datum $u$ is an eigenfunction of the Laplacian of the singular metric $\sigma^{-2} g$: -\Delta_{\sigma^{-2} g} u = \mu (n-\mu) u, \; \Re(\mu) = n/2, \, \mu \ne n/2 with boundary value $f \in C^\infty(M)$. In particular, it holds $M(\lambda) = M_1(\lambda)$ since $u=1$ is an eigenfunction of $\Delta_{\sigma^{-2}g}$ with eigenvalue $0$ and boundary value $1$. Since $\sigma^{-2}g$ is asymptotically hyperbolic, there is an eigenfunction $u$ for any $f \in C^\infty(M)$ so that in its asymptotic expansion $f$ defines one of the leading terms. Now $M_u(\lambda)$ admits a meromorphic continuation to $\C$ with simple poles in $\{ -\mu-N-1, \, |\, N \in \N \}$. For $N \in \N_0$, we set \begin{equation}\label{res-fam-basic} \D_N^{res}(g,\sigma;\lambda) \st N!(2\lambda\!+\!n\!-\!2N\!+\!1)_N \delta_N(g,\sigma;\lambda\!+\!n\!-\!N), \end{equation} where $(a)_N = a (a+1) \cdots (a+N-1)$ is the Pochhammer symbol and \Res_{\lambda=-\mu-1-N} (\langle M_u(\lambda), \psi \rangle) = \int_M f \delta_N(g,\sigma;\mu)(\psi) dvol_h for a meromorphic family $\delta_N(g,\sigma;\mu)$ of differential operators $C^\infty(X) \to C^\infty(M)$ of order $\le N$. $M_u(\lambda)$ should be regarded as a further generalization of Gelfand's distribution $r_+^\lambda$ <cit.>. The factor in (<ref>) guarantees that the residue family $\D_N^{res}(\lambda)$ is polynomial in $\lambda$. Its degree equals $2N$. It easily follows from these definitions that $\D_N^{res}(\lambda)$ satisfies a conformal transformation law which is analogous to that of $L_N(\lambda)$. In view of $M(\lambda) = M_1(\lambda)$, the residue formula (<ref>) can be restated as \int_M \D_N^{res}(g,\sigma;0)(\psi) dvol_h \sim \int_M (v \psi)^{(N)}(0) dvol_h (for $N < n$). The above definition generalizes the notion of residue families introduced in <cit.>. For the details, we refer to Section <ref>. We also emphasize that while $L_N(\lambda)$ maps functions on $X$ to functions on $X$, residue families map functions on $X$ to functions on the boundary $M$. The families $L_N(g,\sigma;\lambda)$ and $D_N^{res}(g,\sigma;\lambda)$ are the main objects of the present paper. These objects and most related discussions will only depend on approximations of the data $(g,\sigma)$ in a sufficiently small neighborhood of the boundary $M$. As a preparation for the following discussion, we briefly recall the role of residue families for Poincaré-Einstein metrics in <cit.>. These polynomial families of differential operators are defined through a residue construction as in (<ref>), where $g_+ = r^{-2}(dr^2 + h_r)$ is a Poincaré-Einstein metric on $X = (0,\varepsilon) \times M$ in normal form relative to a given metric $h$ on $M$. Let $\iota: M \hookrightarrow X$ be the embedding $m \mapsto (0,m)$. The theory in <cit.> deals with approximations of Poincaré-Einstein metrics which are completely determined by the metric $h$. In particular, for even $n$, these define even-order residue families $D_{2N}^{res}(h;\lambda)$ of order $2N \le n$. Since \begin{equation}\label{GJMS-res} D_{2N}^{res}\left(h;-\frac{n}{2}+N\right) = P_{2N}(h) \iota^*, \end{equation} the residue family $D_{2N}^{res}(h;\lambda)$ can be regarded as a perturbation of the GJMS-operator $P_{2N}(h)$. Moreover, residue families satisfy systems of recursive relations that involve lower-order residue families and GJMS-operators. These imply recursive relations among GJMS-operators and recursive relations for $Q$-curvatures. In other words, residue families may be viewed as a device to study the GJMS-operators and $Q$-curvatures on the boundary $M$. The above more general construction deals with a general metric $g$ on $X$. It induces a metric $h = \iota^*(g)$ on $M$, but of course is not determined by $h$. In that case, the resulting residue families depend on the metric $g$ and a boundary defining function $\sigma$. To get residue families, which are determined only by the metric $g$ and the embedding $\iota$, we put an extra condition on $\sigma$. At the same time, we restrict considerations to residue families for $N \le n$. Similarly, as in the Poincaré-Einstein case, this restriction means that it suffices to deal with finite approximations of true eigenfunctions. This way, we obtain specific conformally covariant polynomial one-parameter families of differential operators $C^\infty(X) \to C^\infty(M)$. However, we emphasize that the problem of describing all such conformally covariant families is more complicated <cit.>. Now, following <cit.>, the extra condition which we pose is that $\sigma$ solves a singular Yamabe problem. The version of that problem of interest here asks to find a defining function $\sigma$ of $M$ so that $\scal(\sigma^{-2}g) = -n(n+1)$. By <cit.> such $\sigma$ exist and are unique. However, in general, $\sigma$ is not smooth up to the boundary. More precisely, the existence of smooth solutions is obstructed by a conformally covariant scalar curvature invariant $\B_n$ called the singular Yamabe obstruction. Although this lack of smoothness will only play a minor role for our main purposes, we also have some new results for the invariant $\B_n$. By the conformal transformation law of scalar curvature, the condition that $\sigma$ solves the singular Yamabe problem can be restated as \SC(g,\sigma) \st |d\sigma|_g^2 + 2 \sigma \rho \stackrel{!}{=} 1. This role of the functional $\SC(g,\sigma)$ also implies its conformal invariance: \SC(\hat{g},\hat{\sigma}) = \SC(g,\sigma). The main results of the present work only require us to assume that $\sigma$ satisfies the weaker condition \begin{equation}\label{condition-Y} \SCY: \hspace{0.5cm} \SC(g,\sigma) = 1 + \sigma^{n+1} R_{n+1} \qquad \qquad \end{equation} with a smooth remainder term $R_{n+1}$. The restriction of the remainder term to $\sigma = 0$ then defines the singular Yamabe obstruction $\B_n$. For more details and background on the singular Yamabe problems, we refer to Section <ref>. Now we are ready to state the first main result. Let $\iota: M \hookrightarrow X$ be the embedding of $M$ into $X$. Let $\N \ni N \le n$ and assume that the condition $\SCY$ is satisfied. Then \begin{equation}\label{equivalence} \iota^* L_N(g,\sigma;\lambda) = \D_N^{res}(g,\sigma;\lambda), \; \lambda \in \C. \end{equation} Some comments on this result are in order. The identity (<ref>) is to be interpreted as an identity of operators acting on smooth functions with support in a sufficiently small neighborhood of $M$. Although the assumptions guarantee that the operators on both sides of (<ref>) only depend on the metric $g$ and the embedding $\iota$, we keep the notation of the general case indicating the dependence of the operators on $g$ and $\sigma$. If $g = r^2 g_+$ is a Poincaré-Einstein metric in normal form relative to the metric $h$ on $M$, then the equality (<ref>) was proven in <cit.>.[<cit.> uses a different normalization of residue families.] In particular, if $g$ is the Euclidean metric on $\R^{n+1}$, then Theorem <ref> reduces to the main result of <cit.> (for more details, we refer to <cit.>). Theorem <ref> is a consequence of the following identity. For any boundary defining function $\sigma \in C^\infty(X)$, it \begin{equation*}\label{M-CF} L(g,\sigma;\lambda) + \sigma^{\lambda-1} \circ \left(\Delta_{\sigma^{-2}g} - \lambda(n\!+\!\lambda) \id \right) \circ \sigma^{-\lambda} = \lambda (n\!+\!\lambda) \sigma^{-1}(\SC(g,\sigma)-1) \id,\; \lambda \in \C. \end{equation*} Again, Theorem <ref> is to be interpreted as an identity near the boundary of $X$. We apply Theorem <ref> to prove the existence of a meromorphic continuation of the family $M_u(\lambda)$ using a Bernstein-Sato argument. We recall that for any polynomial $p$, the classical Bernstein-Sato argument <cit.> proves the existence of a meromorphic distribution-valued function $p^\lambda$ by using the functional equation D(\lambda) (p^{\lambda+1}) = b(\lambda) p^\lambda for some polynomial $b(\lambda)$ and a family $D(\lambda)$ of differential operators. Here $\sigma^\lambda u$ takes the role of $p^\lambda$ and $L(\lambda)$ takes the role of the family $D(\lambda)$. The Bernstein-Sato argument yields formulas for the residues and proves Theorem <ref>. The conjugation formula in Theorem <ref> easily implies the commutator relations \begin{equation}\label{sl2} L_N(g,\sigma;\lambda) \circ \sigma - \sigma \circ L_N(g,\sigma;\lambda\!-\!1) = N(2\lambda\!+\!n\!-\!N) L_{N-1}(g,\sigma;\lambda\!-\!1) \end{equation} if $\SC(g,\sigma)=1$. The relation (<ref>) was discovered in <cit.>, where the special case $N=1$ is regarded as one commutator relation in a basic $sl(2)$-structure. In turn, it follows that for $2\lambda = -n+N$ the operator $L_N(g,\sigma;\lambda)$ reduces to a tangential operator $\PO_N(g,\sigma)$ on $M$. By Theorem <ref>, we obtain \begin{equation}\label{PO-Dres} \PO_N(g,\sigma) \iota^* = \D_N^{res}\left(g,\sigma;\frac{-n+N}{2}\right). \end{equation} In other words, each operator $\PO_N(g,\sigma)$ on $M$ comes with a polynomial one-parameter family. The definition of $\PO_N$ in terms of $L_N$ is due to <cit.>. The spectral theoretic description (<ref>) of $\PO_N$ will be shown to have a series of significant consequences. The conformal covariance of the families $L_N(\lambda)$ implies that the operators $\PO_N$ on $C^\infty(M)$ are conformally covariant in the sense that e^{\frac{n+N}{2} \iota^*(\varphi)} \circ \PO_N(\hat{g},\hat{\sigma}) = \PO_N(g,\sigma) \circ e^{\frac{n-N}{2} \iota^* (\varphi)}, \; \varphi \in C^\infty(X). For $N \le n$, these conformally covariant operators are completely determined by the metric $g$ and the embedding $M \hookrightarrow X$. Following <cit.>, we call them extrinsic conformal Laplacians. The notion is motivated by the fact that for even $N$ the leading term of $\PO_N(g,\sigma)$ is given by a constant multiple of $\Delta_h^{N/2}$, where $h =\iota^*(g)$ (see Theorem <ref>). If the background metric $g$ is the conformal compactification $r^2 g_+$ of a Poincaré-Einstein metric in normal form relative to $h$, then these even-order operators actually are constant multiples of GJMS-operators of $h$. But, in general, the operators $\PO_N(g,\sigma)$ depend on the metric $g$ in a neighborhood of $M$ and the embedding $M \hookrightarrow X$. The following result makes the leading parts of all extrinsic conformal Laplacians explicit. Its proof rests on the spectral theoretic interpretation (<ref>) of $\PO_N$. It holds \begin{equation}\label{LT-P-even-M} \PO_{2N} = (2N\!-\!1)!!^2 \Delta^N + LOT \end{equation} for $2N \le n$ and \begin{equation}\label{LT-P-odd-M} \PO_N = (2N\!-\!2) (N\!-\!1)! \sum_{r=0}^{\frac{N-3}{2}} m_N(r) \Delta^r \delta(\lo d) \Delta^{\frac{N-3}{2}-r} + LOT \end{equation} for odd $N$ with $n \ge N \ge 3$ with rational coefficients given by (<ref>). Here $\Delta$ is the Laplacian of $h$, and $\lo$ is the trace-free part of the second fundamental form $L$. LOT refers to terms of order $2N-2$ and $N-3$, respectively. In the Poincaré-Einstein case, <cit.> identifies the operator $\PO_{2N}$ with a constant multiple of the GJMS operator $P_{2N}$. An alternative proof of that result is given in <cit.>. In the general case, formula (<ref>) is also stated in <cit.> although the proof only refers to results in the Poincaré-Einstein case in <cit.>. Note also that in the Poincaré-Einstein case, the operators $\PO_N$ for odd $N$ vanish. Note that the order of $\PO_{2N}$ equals $2N$ and the order of $\PO_N$ for odd $N$ equals $N-1$ in the generic case. Now we use the operators $\PO_N$ to define analogs of Branson's $Q$-curvatures. Theorem <ref> implies that for $N < n$ the identity \PO_N(g,\sigma) (1) = \left(\frac{n-N}{2}\right) \QC_N(g,\sigma) defines a function $\QC_N(g,\sigma)$. These functions will be called the subcritical extrinsic $Q$-curvatures. The critical extrinsic $Q$-curvature $\QC_n(g,\sigma)$ can be defined either through a limiting argument from the subcritical ones or more elegantly by \begin{equation}\label{Q-crit-M} \QC_n^{res}(g,\sigma) = - \dot{\D}_n^{res}(g,\sigma;0)(1). \end{equation} We recall that the condition $\SCY$ guarantees that the quantities in(<ref>) are determined by the metric $g$ and the embedding $M \hookrightarrow X$. In the Poincaré-Einstein case, these definitions reduce to constant multiples of Branson's $Q$-curvatures. If $\SC(g,\sigma)=1$, it follows from Theorem <ref> that the definition (<ref>) is a special case of the notion of $Q$-curvature defined in <cit.>.[The numbering in the published version differs from that in the arXiv version.] By differentiating the conformal transformation law of $\D_n^{res}(g,\sigma;\lambda)$ (see Theorem <ref>) at $\lambda=0$, it follows that \begin{equation}\label{fundamental-M} e^{n\iota^*(\varphi)} \QC_n(\hat{g},\hat{\sigma}) = \QC_n(g,\sigma) + \PO_n(g,\sigma)(\iota^*(\varphi)) \end{equation} for $\varphi \in C^\infty(X)$. As noticed above, the metric $\sigma^{-2}g$ is asymptotically hyperbolic if $|d\sigma|^2_g=1$ on $M$. Associated to such metrics, there is a scattering operator $\Sc(\lambda)$ acting on $C^\infty(M)$. It describes the asymptotic expansion of eigenfunctions of the Laplacian of $\sigma^{-2}g$. In <cit.>, Graham and Zworski showed how the GJMS-operators $P_{2N}$ and the critical Branson $Q$-curvature $Q_n$ of a metric $h$ on $M$ are encoded in the scattering operator of a Poincaré-Einstein metric with the conformal class of $h$ as conformal infinity. The following result extends these interpretations to the present framework. Suppose that condition $\SCY$ is satisfied. Let $N \in \N$ with $2 \le N \le n$ and suppose that $(n/2)^2-(N/2)^2$ is not in the discrete spectrum of $-\Delta_{\sigma^{-2}g}$. Then \PO_N = 2 (-1)^N (N-1)! N! \Res_{\frac{n-N}{2}}(\Sc(\lambda)). The function $\Sc(\lambda)(1)$ is regular at $\lambda=0$ and its value at $\lambda=0$ is denoted by $\Sc(0)(1)$. Then \QC_n = 2 (-1)^n (n-1)! n! \Sc(0)(1). Note that the scattering operator $\Sc(\lambda)$ is a pseudo-differential operator with principal symbol 2^{2\lambda-n} \frac{\Gamma(\lambda-\frac{n}{2})}{\Gamma(-\lambda+\frac{n}{2})}|\xi|_h^{n-2\lambda}. This shows that the residues of $\Sc(\lambda)$ at $\lambda=\frac{n}{2}-N$ are caused by the $\Gamma$-factors and that the residue of its pole at $\lambda = \frac{n}{2}-N$ is an operator with principal part given by a constant multiple of $\Delta^N$. Moreover, Theorem <ref> yields an identification of this residue. In particular, these residues are elliptic operators. On the other hand, the above formula yields no information on the structure of the residues of $\Sc(\lambda)$ at $\frac{n-N}{2}$ for odd $N$. This makes the second part of Theorem <ref> interesting. Similarly as in <cit.>, the self-adjointness of $\Sc(\lambda)$ for $\lambda \in \R$ combined with Theorem <ref> implies Let $\SCY$ be satisfied. Then the operators $\PO_N(g)$ are formally self-adjoint with respect to the scalar product on $C^\infty(M)$ defined by $h$. For closed $M$, combining (<ref>) with the self-adjointness of $\PO_n$ and $\PO_n(1)=0$ shows that the integral \int_M \QC_n(g) dvol_h is a global conformal invariant. Next, we describe a local formula for the critical extrinsic $Q$-curvature. It extends the holographic formula for Branson's critical $Q$-curvature proved in <cit.>. The formulation of that result requires two more ingredients: renormalized volume coefficients $v_k$ and solution operators $\T_k(0)$. These data are defined in local coordinates. We choose a diffeomorphism $\eta$ between the product $[0,\varepsilon) \times M$ with coordinates $(s,x)$ and a neighborhood of $M$ in $X$. $\eta$ is defined by a renormalized gradient flow of $\sigma$. It satisfies $\eta^*(\sigma) = s$. These coordinates will be called adapted coordinates. Then \eta^* (dvol_{\sigma^{-2}g}) = s^{-n-1} v(s,x) ds dvol_h, and the expansion $v(s,x) = \sum_{k\ge 0} v_k(x) s^k$ defines the coefficients $v_k$. Let $(v\J)_k$ be the coefficients in the analogous expansion of $v \eta^*(\J)$, where $\J = \J^g = \scal^g/2n$. Secondly, the differential operators $\T_k(0)$ describe the asymptotic expansions of harmonic functions of the Laplacian of $\sigma^{-2} g$. The details of these definitions are given in Sections [Holographic formula for $\QC_n$] Let $n$ be even. If $\SCY$ is satisfied, then it holds \begin{equation*}\label{Q-holo-form-M} \QC_n(g,\sigma) = (n\!-\!1)!^2 \sum_{k=0}^{n-1} \frac{1}{n\!-\!1\!-\!2k} \T_k^*(g,\sigma;0) \left( (n\!-\!1)(n\!-\!k) v_{n-k} + 2k (v \J)_{n-k-2} \right), \end{equation*} where $\J = \J^g$. For $k=n-1$, the second term in the sum is defined as $0$. The proof of Theorem <ref> generalizes a proof of the holographic formula for $Q_n$ for Poincaré-Einstein metrics given in <cit.>. The arguments rest on Theorem <ref> and an alternative description of residue families in terms of the coefficients $v_k$ and the operators $\T_k(\lambda)$. We conjecture that the result extends to odd $n$ (Conjecture <ref>). There are analogous formulas for all subcritical extrinsic $Q$-curvatures which extend <cit.>. Theorem <ref> implies the representation \QC_n = n! (n\!-\!1)! v_n + \sum_{k=1}^{n-1} \T_k^*(0)(\cdots). Integration of this identity for closed $M$ implies the equality \begin{equation}\label{GI} \int_M \QC_n dvol_h = n! (n\!-\!1)! \int_M v_n dvol_h \end{equation} of global conformal invariants. This extends a result of <cit.> for Poincaré-Einstein metrics and reproves a special case of <cit.>. We shall also give a second proof of the identity (<ref>) which utilizes residue families and extends to odd $n$. We emphasize again that the integrands on both sides depend on the embedding of $M$ in $X$, i.e., these are global conformal invariants of the embedding. The integral on the right-hand side of (<ref>) appears as a coefficient of $\log \varepsilon$ in the asymptotic expansion of the volume \begin{equation}\label{reno-volume-M} vol_{\sigma^{-2}g}(\{\sigma > \varepsilon \}) = \int_{\sigma > \varepsilon} dvol_{\sigma^{-2}g} \end{equation} of the singular metric $\sigma^{-2}g$ for $\varepsilon \to 0$. This implies its conformal invariance. Through that interpretation of the right-hand side of (<ref>), this identity is a special case of <cit.>. The singular terms in this expansion of the volume are given by the integrals of the renormalized volume coefficients $v_k$ for $k <n$. All these integrals admit the uniform expressions \begin{equation}\label{HF-v} \int_{M} v_{k} dvol_h = (-1)^k \frac{(n\!-\!1\!-\!k)!}{ (n-1)! k!} \int_{M} \iota^* L_k(-n\!+\!k)(1) dvol_h \end{equation} in terms of Laplace-Robin operators. This reproves a special case of <cit.>. The above results for the expansion of the volumes (<ref>) have analogs for general defining functions $\sigma$. In that case, the operators $L_k$ are replaced by composition $\tilde{L}_k = (L \circ \SC^{-1})_k$ and the coefficient of $\log \varepsilon$ is a constant multiple of the integral of a curvature quantity $\tilde{\QC}_n(g,\sigma)$ (Theorem <ref>, (<ref>)). The total integral of $\tilde{\QC}_n$ is a conformal invariant of the pair $(g,\sigma)$ (Lemma <ref>). The integrals in (<ref>) are generalizations of the Willmore energy of a closed surface in $\R^3$. In fact, for a surface $M^2 \hookrightarrow X^3$ with Gauss curvature $K$, it holds 2 v_2 = \QC_2 = -K + \frac{1}{2} |\lo|^2 (by (<ref>), (<ref>) and $\J^h = K$). But if $X$ is the flat $\R^3$, then $H^2 - K = |\lo|^2$. Hence for closed $M \hookrightarrow \R^3$ the integral $\int_M v_2 dvol_h$ is a linear combination of the Euler characteristic $\chi(M)$ and the Willmore energy $\W$ Willmore energy \W = \int_M H^2 dvol_h. The Willmore energy also contributes to the rigid string action introduced in <cit.>, Finally, the usage of adapted coordinates leads to a new formula for the singular Yamabe obstruction $\B_n$ (see Section <ref> for its definition). In order to state that formula, we note that the pull-back $\eta^*(g)$ of $g$ takes the form $a^{-1} ds^2 + h_s$ with some coefficient $a \in C^\infty ((0,\varepsilon) \times M)$ and a one-parameter family $h_s$ of metrics on $M$. For \mathring{v} \st dvol_{h_s}/dvol_h, it holds \frac{\mathring{v}'}{\mathring{v}} = \frac{1}{2} \tr (h_s^{-1} h_s'). Let $\J$ and $\rho$ denote the pull-backs of $\J^g$ and $\rho(g,\sigma)$ by $\eta$, respectively. The following result restates Theorem <ref>. [The obstruction $\B_n$] If $\sigma$ satisfies $\SCY$, then \begin{equation}\label{ob-magic-2} (n\!+\!1)! \B_n = -2\partial_s^n \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 + 4 \sum_{j=1}^n j \binom{n}{j} \partial_s^{j-1}(\rho)|_0 \partial_s^{n-j} \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 - 2n \partial_s^{n-1}(\J)|_0 \end{equation} for $n \ge 1$. For $n=1$, the identity (<ref>) yields $\B_1 = 0$. Moreover, the Taylor coefficients of $\rho$ obey a recursive relation in terms of the Taylor coefficients of $\mathring{v}'/\mathring{v}$ and $\J$ (Proposition <ref>). This result has several consequences of independent interest. For a flat background metric, the scalar curvature term on the right-hand side of (<ref>) vanishes, and the recursive structure of all terms implies that for even $n$ \begin{equation*} (n\!+\!1)! \B_n = - 2 (n\!-\!1)!!/(n\!-\!2)!! \Delta^\frac{n}{2} (H) + LOT \end{equation*} (Theorem <ref>), where LOT refers to terms of lower differential order. Up to the numerical coefficients, that description of the leading term of $\B_n$ was first formulated in <cit.> <cit.> (for general backgrounds). We also use Theorem <ref> to reproduce the known explicit formulas for $\B_2$ (for general backgrounds) and to calculate $\B_3$ for (conformally) flat backgrounds confirming earlier results. it is unclear what LOT means in GW we also can prove that the LOT contains at most n-1 derivatives of the curvature tensor of the background Moreover, Theorem <ref> and Theorem <ref> play a key role in the proof of the residue formula \begin{equation*} \Res_{n=N-1} (\QC_N) = (-1)^{n-1} n! (n\!+\!1)! \frac{n}{2} \B_n \end{equation*} for even $n$ (Theorem <ref>), where we regard $\QC_N$ as a function of $n$ which is singular at $n=N-1$. This formula relates $\B_n$ for even $n$ to the super-critical $Q$-curvature $\QC_{n+1}$. It proves a conjecture in <cit.>. We conjecture that this result extends to all dimensions. <cit.> established a coordinate-free formula for the Yamabe obstructions $\B_n$ in terms of tractor calculus constructions. This formula involves an analog of the critical extrinsic conformal Laplacian acting on the so-called normal tractor of the hypersurface $M$ as a key ingredient.[This result coincides with Theorem 7.7 in <arXiv:150602723v4>.] We finish this section with an outline of the content of the paper. We combine this review with additional comments on the relations of the new results to the literature. Section <ref> briefly recalls basic results on the version of the singular Yamabe problem of interest here. In Section <ref>, we establish Theorem <ref> and describe first consequences. This result is the basic link that connects the Laplace-Robin operators of Gover and Waldron with the spectral theory of the Laplacian of the singular metric $\sigma^{-2}g$. In Section <ref>, we describe some of the representation theoretical aspects of the Laplace-Robin operators. Section <ref> introduces the notion of adapted coordinates. This notion is basic for the whole work. Adapted coordinates give rise to the definition of renormalized volume coefficients $v_k$, which enables us to prove local formulas for $Q$-curvatures (Theorem <ref>) and to prove the formula for the obstruction $\B_n$ in Theorem <ref>. In this connection, we show that in adapted coordinates, the function $\rho$ obeys an extremely beneficial ordinary differential equation in the variable $s$, which implies recursive relations for the Taylor coefficients of $\rho$. We emphasize that Graham <cit.> utilizes a different notion of renormalized volume coefficients which are defined in terms of the normal exponential map. This notion seems to be less appropriate from the present point of view. Section <ref> introduces the notion of residue families. Here we extend earlier definitions in the context of Poincaré-Einstein metrics <cit.>. The constructions are built on basic facts in the spectral theory of asymptotically hyperbolic metrics, which are detailed in <cit.>. In particular, we use a version of a Poisson transform and introduce the scattering operator $\Sc(\lambda)$. We derive a formula for residue families in terms of the coefficients $v_k$ and the operators $\T_k(\lambda)$. Section <ref> contains the proof of Theorem <ref>. The identity (<ref>) is a direct consequence. Here we also prove that these identities can be regarded as consequences of a beautiful formula for the action of $L_N(-1)$ on the distribution $\sigma^*(\delta) = \delta_M$ of $M$ (Theorem <ref>). This distributional formula reproves a basic technical result in <cit.>. In Section <ref>, we introduce the notion of extrinsic conformal Laplace operators $\PO_N$. Here we derive the spectral theoretical interpretation (<ref>) of $\PO_N$, use it to determine the leading terms of $\PO_N$ (Theorem <ref>) and recognize these operators as residues of the scattering operator (first part of Theorem <ref>). These results extend results of <cit.>. Section <ref> defines extrinsic $Q$-curvatures and establishes the second part of Theorem <ref> which extends another result of <cit.>. Moreover, we supply two proofs of the equality (<ref>). The integrated coefficients $v_k$ are shown to describe the singular coefficients in the asymptotic expansion of (<ref>). These expansions and their generalizations (Theorem <ref>) reprove results of <cit.>. In Section <ref>, we establish extensions of the holographic formulas for $Q$-curvatures in <cit.>. In particular, we prove Theorem <ref>. Section <ref> contains comments on further perspectives. In the main body of the text, we usually suppress detailed calculations and the discussion of examples and special cases. However, the reader can find this material in Section <ref>. It starts with an overview of its own. The results presented here may be used to gain a deeper understanding of the material. In particular, this section contains full details on low-order Yamabe obstructions, low-order cases of conformal Laplacians, extrinsic $Q$-curvatures, and renormalized volume coefficients. All proofs are independent of the literature. Hopefully, the attached list of symbols facilitates reading. Finally, we like to emphasize that, although the present work is deeply inspired by the pioneering works of Gover and Waldron, the current treatment is fully independent and self-contained. We also stress again that our perspective is (via the residue families and their applications) one of a spectral-theoretic nature exploiting the structure of eigenfunctions of the Laplacian. After the present work had been posted, the paper <cit.> again discussed extrinsic conformal Laplacians from the perspective of scattering theory. It defines extrinsic conformal Laplacians in terms of the scattering operator of the singular Yamabe metric $\sigma^{-2}g$ by mimicking the known relations between GJMS-operators and the scattering operator of Poincaré-Einstein metrics <cit.>. However, the relations between these definitions and the notions introduced by Gover and Waldron are established only here. Acknowledgments. The work on this project started during a visit of the first author at the University of Århus in autumn 2019. A large part of the paper's final version was written during a stay of the first author at IHES in early 2020. He is grateful to both organizations for financial support and very stimulating atmospheres. Finally, we thank the anonymous referee who provided valuable detailed comments on an earlier version of the manuscript. § GENERAL NOTATION $\N$ natural numbers $\N_0$ non-negative integers $(a)_N$ Pochhammer symbol $\mathfrak{X}(X)$ space of vector fields on $X$ $\N$ is the set of natural numbers, and $\N_0$ is the set of non-negative integers. For a complex number $a\in\C$ and an integer $N\in\N$, the Pochhammer symbol $(a)_N$ is defined by $(a)_N \st a(a+1) \cdots (a+N-1)$. We set $(a)_0 \st 1$. $R$ curvature tensor $C^{-\infty}(X)$ space of distributions on $X$ All manifolds are smooth. For a manifold $X$, $C^\infty(X)$ and $C_c^\infty(X)$ denote the respective spaces of smooth functions and smooth functions with compact support on $X$. If $X$ is a manifold with boundary, then $C^\infty(X)$ is the space of functions that are smooth up to the boundary. $C^{-\infty}(X)$ is the space of distributions on $X$. Let $\mathfrak{X}(X)$ be the space smooth vector fields on $X$. Metrics on $X$ usually are denoted by $g$. $dvol_g$ is the Riemannian volume element defined by $g$. The Levi-Civita connection of $g$ is denoted by $\nabla_X^g$ or simply $\nabla_X$ for $X \in \mathfrak{X}(X)$ if $g$ is understood. In these terms, the curvature tensor $R$ of the Riemannian manifold $(X,g)$ is defined by $R(X,Y)Z =\nabla_X \nabla_Y (Z) - \nabla_Y \nabla_X (Z) - \nabla_{[X.Y]}(Z)$ for vector fields $X,Y,Z \in \mathfrak{X}(X)$. We also set $\nabla_X (u) = \langle du,X \rangle$ for $X \in \mathfrak{X}(X)$ and $u \in C^\infty(X)$. $dvol_g$ volume element $\Omega^p$ space of $p$-forms $\nabla^g$ Levi-Civita connection of $g$ $\grad_g(u)$ gradient field $\delta^g$ divergence operator $\Delta_g$ Laplacian of $g$ For a metric $g$ on $X$ and $\sigma \in C^\infty(X)$, let $\grad_g(\sigma)$ be the gradient of $\sigma$ with respect to $g$, i.e., it holds $g(\grad_g(\sigma),V) = \langle d\sigma,V \rangle$ for all vector fields $V \in \mathfrak{X}(X)$. $g$ defines pointwise scalar products and norms $|\cdot|_g$ on $\mathfrak{X}(X)$ and on forms $\Omega^*(X)$. Then $|\grad_g(\sigma)|_g^2 = |d\sigma|_g^2$. $\delta^g$ is the divergence operator on differential forms or symmetric bilinear forms. On forms, it coincides with the negative adjoint $-d^*$ of the enaböe differential $d$ with respect to the Hodge scalar product defined by $g$. Let $\Delta_g = \delta^g d$ be the non-positive Laplacian on $C^\infty(X)$. On the Euclidean space $\R^n$, it equals $\sum_i \partial_i^2$. $\Ric^g$ Ricci tensor of $g$ $\scal^g$ scalar curvature of $g$ $\scal(g)$ scalar curvature of $g$ $\Rho^g$ Schouten tensor of $g$ A metric $g$ on a manifold $X$ with boundary $M$ induces a metric $h$ on $M$. Curvature quantities of $g$ and $h$ have the respective metric as an index if required by clarity. In particular, the scalar curvature of the metric $g$ on $X$ is denoted by $\scal^g$ or $\scal(g)$. $\Ric^g$ denote the Ricci tensor of $g$. On a manifold $(M,g)$ of dimension $n$, we set $\J^g =\frac{1}{2(n-1)} \scal^g$ if $n \ge 2$ and define the Schouten tensor $\Rho^g$ of $g$ by $(n-2)\Rho^g = \Ric^g - \J^g g$ if $n \ge 3$. $L$ second fundamental form Let $M$ be a hypersurface in $(X,g)$ with the induced metric $h$. The second fundamental form $L$ of $M$ is defined by $L(X,Y) = -g(\nabla_X(Y),N)$ for vector fields $X,Y \in \mathfrak{X}(M)$ and a unit normal vector field $N$. In particular, if $X$ is a manifold with boundary $M$ with defining function $\sigma$ so that $|d\sigma|_g=1$ on $M$, then we set $L(X,Y) = -g(\nabla_X(Y),\NV)$ with $\NV \st \grad_g(\sigma)$. With these conventions, $L=g$ for the round sphere $S^n \subset \R^{n+1}$ if $\sigma$ is the distance function of $S^n$. We set $n H = \tr_h(L)$ if $M$ has dimension $n$. $H$ is the mean curvature of $M$. Let $\lo = L - H h$ be the trace-free part of $L$. We sometimes identify $L$ with the shape operator $S$ defined by $h(X,S(Y)) = L(X,Y)$. For a bilinear form $b$ on $T(M)$, we denote its trace-free part with respect to a metric known by context by $\mathring{b}$ or $b_\circ$. $\NV$ gradient of $\sigma$ $H$ mean curvature $\lo$ trace-free part of $L$ $\iota$ embedding $b_\circ$ trace-free part of $b$ $\iota$ denotes various canonical embeddings such as $\iota: M \hookrightarrow X$. The symbol $\iota^*$ will be used for the induced pull-back of functions, forms, and metrics. For any diffeomorphism $f$, the symbol $f_* = (f^{-1})^*$ denotes the push-forward by $f$. Differentiation with respect to the variable $\lambda$ will often be denoted by $^\cdot$. In contrast, differentiation with respect to the variables $r$ and $s$ will usually be denoted by $'$. The symbol $\circ$ denotes compositions of operators. The symbol $\sim$ indicates a proportionality. § THE SINGULAR YAMABE PROBLEM Let $(X,g)$ be a compact manifold with boundary $M$ of dimension $n$. The problem to ask for a defining function $\sigma$ of $M$ so that \begin{equation}\label{syp} \scal (\sigma^{-2}g) = -n(n+1) \end{equation} is known as (a version of) the singular Yamabe problem <cit.>. The metric $\sigma^{-2} g$ is called a singular Yamabe metric if (<ref>) is true.[Later we shall often deal with a weaker condition.] The conformal transformation law of scalar curvature shows that \scal(\sigma^{-2}g) = -n(n+1) |d\sigma|_g^2 + 2n \sigma \Delta_g(\sigma) + \sigma^2 \scal(g). Following <cit.>, we write this equation in the form \scal(\sigma^{-2}g) = -n(n+1) \SC(g,\sigma), \SC(g,\sigma) = |d\sigma|_g^2 + 2 \rho \sigma (see Definition (<ref>)).[In <cit.>, the quantity $\SC(g,\sigma)$ is termed the $\SC$-curvature of $(g,\sigma)$.] In these terms, $\sigma$ is a solution of (<ref>) iff $\SC(g,\sigma)=1$. Such $\sigma$ exist <cit.> and are unique <cit.>. However, in general, $\sigma$ is not smooth up to the boundary. The smoothness is obstructed by a locally determined conformally invariant scalar function on $M$, called the singular Yamabe obstruction. Moreover, the solution is smooth up to the boundary iff the obstruction vanishes In order to describe the structure of $\sigma$ more precisely, we follow <cit.> and <cit.>. We use geodesic normal coordinates (see Section <ref>). Let $r = d_M$ be the distance function of $M$ for the background metric $g$. Then there are uniquely determined coefficients $\sigma_{(k)} \in C^\infty(M)$ for $2 \le k \le n+1$ so that the smooth defining function $\sigma_F$ \begin{equation}\label{sigma-finite} \sigma_F \st r + \sigma_{(2)} r^2 + \dots + \sigma_{(n+1)} r^{n+1} \end{equation} \begin{equation}\label{Yamabe-finite} \SC(g,\sigma_F) = 1 + R r^{n+1} \end{equation} with a smooth remainder term $R$. We briefly describe how these coefficients are recursively determined. In geodesic normal coordinates, the metric $g$ takes the form $dr^2 + h_r$ with a one-parameter family $h_r$ of metrics on $M$. The condition (<ref>) is equivalent to |d\sigma_F|_g^2 - \frac{2}{n+1} \sigma_F \Delta_g(\sigma_F) - \frac{1}{n(n+1)} \sigma_F^2 \scal^g = 1 + R r^{n+1}. We write the left-hand side of this equation in the form \begin{align}\label{Y-F} & \partial_r(\sigma_F)^2 + h_r^{ij} \partial_i (\sigma_F) \partial_j (\sigma_F) \notag \\ & - \frac{2}{n+1} \sigma_F \left (\partial_r^2 (\sigma_F) + \frac{1}{2} \tr (h_r^{-1} h_r') \partial_r (\sigma_F) + \Delta_{h_r} (\sigma_F) \right) - \frac{1}{n(n+1)} \sigma_F^2 \scal^g \end{align} and expand this sum into a Taylor series in the variable $r$. Then the vanishing of the coefficient of $r^k$ for $k \le n$ is equivalent to an identity of the form (k-1-n) \sigma_{(k+1)} = LOT, where $LOT$ involves only lower-order Taylor coefficients of $\sigma$. The latter relation also shows that there is a possible obstruction to the existence of an improved solution $\sigma_F'$ which contains a term $\sigma_{(n+2)} r^{n+2}$ and satisfies $\SC(g,\sigma_F') = 1 + R r^{n+2}$. However, by setting $\LO_n$ \begin{equation}\label{sol-log-d} \sigma = \sigma_F + \LO_n r^{n+2} \log r \end{equation} with an appropriate coefficient $\LO_n \in C^\infty(M)$ one may get a solution of \begin{equation}\label{sol-log} \SC(g,\sigma) = 1 + O(r^{n+2} \log r). \end{equation} The coefficient $\LO_n$ is determined by the condition that the coefficient of $r^{n+1}$ in the expansion of (<ref>) vanishes. But that coefficient equals \left( r^{-n-1} (\SC(g,\sigma_F) - 1) \right)|_{r=0} - 2 \frac{n+2}{n+1} \LO_n. The first term exists by the construction of $\sigma_F$ and the second term is generated by the action of the terms $\partial_r(\sigma)^2$ and $\sigma \partial_r^2(\sigma)$ in (<ref>) on the log-term in $\sigma$.[In fact, $\partial_r (r^{n+2} \log r) = r^{n+1} + \cdots$ and $\partial_r^2 (r^{n+2} \log r) = (2n+3) r^n + \cdots$.] Hence for \LO_n \st \frac{1}{2} \frac{n+1}{n+2} \left( r^{-n-1} (\SC(g,\sigma_F) - 1) \right)|_{r=0} the condition (<ref>) is satisfied. Following <cit.>, we define the singular Yamabe obstruction by $\B_n$ singular Yamabe obstruction \begin{equation}\label{B-def} \B_n \st \left( r^{-n-1} (\SC(g,\sigma_F) - 1) \right)|_{r=0} . \end{equation} In these terms, we see that with \begin{equation}\label{obstruction-two} \LO_n = \frac{1}{2} \frac{n+1}{n+2} \B_n \end{equation} the improved $\sigma$ defined in (<ref>) satisfies (<ref>). By <cit.>, the unique solution $\sigma$ of the singular Yamabe problem has an expansion of the form \sigma = r + \sigma_{(2)} r^2 + \dots + \sigma_{(n+1)} r^{n+1} + \LO_n r^{n+2} \log r + \dots. Graham <cit.> calls $\LO_n$ the singular Yamabe obstruction. Since $\sigma_F$ is determined by $g$ (and the embedding $M \hookrightarrow X$), we regard $\B_n$ as a functional of $g$ (and the embedding). It is a key result that $\B_n$ is a conformal invariant of $g$. More precisely, we write $\hat{\B}_n$ for the obstruction defined by $\hat{g}=e^{2\varphi} g$ with $\varphi \in C^\infty(X)$. Then $e^{(n+1) \iota^*(\varphi)} \hat{\B}_n = \B_n$. Let $r$ be the distance function of $M$ for $g$. Let S_N \st 1 + \sum_{j=2} r^j \sigma_{(j)}. Then for $N \le n+1$ the condition \begin{equation}\label{scalar-exp} \SC (g,S_N) = 1 + 0 r + \cdots + 0 r^{N-1} + O(r^N) = 1 + O(r^N) \end{equation} determines the coefficient $\sigma_{(N)}$ in terms of lower-order coefficients $\sigma_{(2)}, \dots, \sigma_{(N-1)}$. In that case, we also write $S_N = S_N(g)$. Moreover, we recall that the coefficient of $r^k$ ($k \le N-1$) in the expansion of $\SC(g,S_N)$ depends only on $\sigma_{(2)}, \dots, \sigma_{(k+1)}$. Now we have the obvious relation \SC(\hat{g},\hat{\sigma}) = \SC(g,\sigma) for any $\sigma \in C^\infty(X)$ and $\hat{\sigma} \st e^{\varphi} \sigma$. We write the expansion (<ref>) (for $N \le n+1$) as an expansion in terms of the distance function $\hat{r}$ of $M$ for $\hat{g}$. Hence \begin{equation}\label{scalar-exp-hat} \SC(\hat{g},e^\varphi S_N(g)) = 1 + O(\hat{r}^N). \end{equation} This expansion determines the coefficients $\hat{\sigma}_{(2)}, \dots, \hat{\sigma}_{(N)}$ in the expansion e^\varphi S_N(g) = \hat{r} + \hat{r}^2 \hat{\sigma}_{(2)} + \dots + \hat{r}^{N} \hat{\sigma}_{(N)} + \cdots = S_N(\hat{g}) + \mbox{higher-order terms}. In general, this expansion involves higher-order terms, i.e., $e^\varphi S_N(g) \ne S_N(\hat{g})$. In particular, the remainder term in (<ref>) depends on $\hat{\sigma}_{N+1}$. However, for $N=n+1$, the coefficient of $\hat{r}^{n+1}$ in (<ref>) does not depend on $\hat{\sigma}_{(n+2)}$ since it appears with a prefactor $0$. Therefore, \begin{align*} (\hat{r}^{-n-1} \SC(\hat{g},S_{n+1}(\hat{g})))|_{\hat{r}=0} & = (\hat{r}^{-n-1} \SC(\hat{g},e^\varphi S_{n+1}(g)))|_{\hat{r}=0} \\ &= e^{-(n+1) \iota^*(\varphi)} (r^{-n-1} \SC(g,S_{n+1}(g)))|_{r=0}. \end{align*} This relation implies the assertion. In <cit.>, Gover and Waldron describe an elegant algorithm that recursively determines the solution $\sigma$ of the singular Yamabe problem as a power series of some boundary defining function $\sigma_0$ (like the distance function $\sigma_0 = d_M$). Note that these power series are not Taylor series: their coefficients still live on the ambient space $X$. This algorithm rests on the interpretation of the quantity $\SC(g,\sigma)$ as the squared length of the scale tractor associated to $\sigma$. For our purposes, it will be enough to consider smooth approximate solutions of the singular Yamabe problem. This motivates the following definition. The defining function $\sigma \in C^\infty(X)$ of $M$ is said to satisfy the condition $\SCY$ iff \sigma = r + \sigma_{(2)} r^2 + \dots + \sigma_{(n+1)} r^{n+1} + O(r^{n+2}) \SC(g,\sigma) = 1 + R_{n+1} r^{n+1} for a smooth remainder term $R_{n+1}$. Equivalently, it holds \SC(g,\sigma) = 1 + R_{n+1} \sigma^{n+1} for another smooth remainder term $R_{n+1}$. The restriction of either remainder terms to $M$ is the singular Yamabe obstruction: $\B_n = \iota^* R_{n+1}$. We recall that the obstruction $\B_n$ satisfies \begin{equation}\label{B-CI} e^{(n+1) \iota^*(\varphi)} \B_n(\hat{g}) = \B_n(g). \end{equation} Graham <cit.> determined the first two non-trivial coefficients $\sigma_{(2)}$ and $\sigma_{(3)}$ in the expansion of $\sigma$ (Lemma <ref>). An explicit formula for the next coefficient $\sigma_{(4)}$ is given in <cit.>. We reproduce these results in a slightly different form in Section <ref> and also display a formula for $\sigma_{(5)}$. Graham <cit.> shows that the obstruction $\B_1$ vanishes. The obstruction for surfaces in a three-manifold is given by the formula $\B_2$ $\B_3$ \begin{align}\label{B2} \B_2 & = -\frac{1}{3} (\delta^h \delta^h (\lo) + H |\lo|^2 + (\lo,\iota^*(\Rho^g))) \notag \\ & = -\frac{1}{3} (\Delta_h (H) + \delta^h (\Ric^g (\NV,\cdot)) + H |\lo|^2 + (\lo,\iota^*(\Rho^g))) \end{align} (<cit.>). The equivalence of both expressions follows from the Codazzi-Mainardi equation. For the details of that argument and a derivation of these formulas from Theorem <ref>, we refer to Section <ref>. The first formula reproduces a result in <cit.>. Explicit formulas for $\B_3$ were first derived in <cit.> and [3] from a general universal tractor formula found in <cit.>. In particular, it was proved that for a conformally flat background metric $g$ the obstruction $\B_3$ is given by the closed formula \begin{equation}\label{B3-closed} 6 \B_3 = 3 (\delta^h \delta^h + (\Rho^h,\cdot))((\lo^2)_\circ) + |\lo|^4, \end{equation} where $b_\circ$ denotes the trace-free part of the symmetric bilinear form $b$. For general background metrics $g$, the formula for $\B_3$ contains additional terms defined by the Weyl tensor of $g$. For full details, we refer to [3]. In the more recent work <cit.>, these formulas for $\B_3$ were derived without utilizing tractor calculus. In Section <ref>, we shall deduce (<ref>) from the general formula in Theorem <ref>. Formula (<ref>) manifestly implies the conformal invariance $e^{4 \varphi} \hat{\B}_3 = \B_3$ since the operator $b \mapsto \delta \delta (b) + (\Rho,b)$ is conformally covariant on trace-free symmetric bilinear forms on $M$. Up to constant multiples, $\B_2$ and $\B_3$ are the respective variations of the functionals \int_M |\lo|_h^2 dvol_h \quad \mbox{and} \quad \int_M (\lo,\JF)_h dvol_h (for the definition of the Fialkov tensor $\JF$ we refer to Section <ref>). These are special cases of the variation formulas of the functional $\A \st \int_M v_n dvol_h$ (with respect to a one-parameter family of hypersurfaces $M \hookrightarrow X$) which were proved in <cit.> and <cit.>. They state that the variation of $\A$ is proportional to the obstruction $\B_n$. The equivalence of both results follows from (<ref>). This result may be regarded as an analog of the result that the metric variation of the total critical $Q$-curvature of an even-dimensional closed manifold is given by the corresponding Fefferman-Graham obstruction tensor <cit.>. Poincaré-Einstein metrics are an important special class of singular Yamabe metrics. In fact, if $g_+ = r^{-2} (dr^2 + h_r)$ is a Poincaré-Einstein metric in normal form relative to $h=h_0$ <cit.>, then $\scal(g_+) = -n(n+1)$, the background metric is $dr^2+h_r$ and the corresponding defining function is $\sigma=r$. In this case, the singular Yamabe obstruction vanishes. We shall refer to this case as the Poincaré-Einstein case. § THE CONJUGATION FORMULA In this section, we introduce Laplace-Robin operators (or degenerate Laplacians) following <cit.>. We relate them to the spectral theory of the Laplacian of singular metrics $\sigma^{-2}g$ and use this relation to prove basic properties of the Laplace-Robin operators. Let $X$ be a manifold of dimension $n+1$. $L(g,\sigma)$ Laplace-Robin operator For any pair $(g,\sigma)$ consisting of a metric $g$ on $X$ and $\sigma \in C^\infty(X)$, the one-parameter family \begin{equation}\label{LR-OP} L(g,\sigma;\lambda) \st (n\!+\!2\lambda\!-\!1) (\nabla_{\grad_g(\sigma)} + \lambda \rho) - \sigma (\Delta_g + \lambda \J): C^\infty(X) \to C^\infty(X) \end{equation} of differential operators is called the Laplace-Robin operator of the pair $(g,\sigma)$. Here $\lambda \in \C$, 2 n \J \st \scal(g) \quad \mbox{and} \quad (n+1) \rho(g,\sigma) \st - \Delta_g (\sigma) - \sigma \J. Moreover, we set $\SC(g,\sigma)$ \begin{equation}\label{SC} \SC(g,\sigma) \st |d\sigma|_g^2 + 2 \sigma \rho. \end{equation} Similarly, we define $L(g,\sigma;\lambda)$ for a manifold $X$ with boundary $M$. In this case, $g$ and $\sigma$ are assumed to be smooth up to the boundary. Then $L(\lambda)$ acts on the space $C^\infty(X)$ of smooth functions up to the boundary and on the space $C^\infty(X^\circ)$ of smooth functions on the open interior $X^\circ = X \setminus M$ of $X$. From now on, we assume that $(X,g)$ is a compact manifold with boundary $M$ and $\sigma$ is a defining function of $M$. We recall that $\sigma$ is a defining function of $M$ if $\sigma^{-1}(0)=M$, $\sigma > 0$ on $X^\circ$ and $d\sigma|_M \ne 0$. Let $\iota: M \hookrightarrow X$ be the embedding and set $h \st \iota^*(g)$. Then the operator $\iota^* L(g,\sigma;\lambda)$ degenerates to the first-order operator \begin{equation}\label{bv-op} C^\infty(X) \ni u \mapsto (n\!+\!2\lambda\!-\!1) \iota^* \left[\nabla_{\grad_g(\sigma)}(u) - \frac{\lambda}{n\!+\!1}\Delta_g(\sigma) u \right] \in C^\infty(M). \end{equation} Since certain linear combinations of Dirichlet and Neumann boundary values are also known as Robin boundary values, this naturally motivates the above notion of a Laplace-Robin operator. If $\sigma^{-2} g$ has constant scalar curvature $-n(n+1)$, the boundary operator (<ref>) reduces to the conformally covariant boundary operator u \mapsto (n\!+\!2\lambda\!-\!1) \iota^* (\nabla_{\grad_g(\sigma)} - \lambda H) u, where $H$ is the mean curvature of $M$ (<cit.>, <cit.>). If $g_+ = r^{-2} g$ is Poincaré-Einstein in the sense that $\Ric(g_+) = - n g_+$, then $L(g,r;\lambda-n+1)$ equals the shift operator $S(g_+;\lambda)$ of <cit.>. Assume that $\sigma \in C^\infty(X)$ is a defining function of the boundary $M$ of $X$. Then it holds \begin{equation}\label{CF} L(g,\sigma;\lambda) + \sigma^{\lambda-1} \circ \left(\Delta_{\sigma^{-2}g} - \lambda(n\!+\!\lambda) \id \right) \circ \sigma^{-\lambda} = \lambda (n\!+\!\lambda) \sigma^{-1}(\SC(g,\sigma)-1) \id \end{equation} as an identity of operators acting on $C^\infty(X^\circ)$. Let $\Con(g,\sigma;\lambda)$ denote the operator \sigma^{\lambda-1} \circ \left(\Delta_{\sigma^{-2}g} - \lambda(n\!+\!\lambda) \id \right) \circ \sigma^{-\lambda}. The relation \Delta_{\sigma^{-2}g} = \sigma^2 \Delta_g - (n\!-\!1) \sigma \nabla_{\grad_g(\sigma)} shows that \begin{align*} \Con(g,\sigma;\lambda) (u) & = \sigma^{\lambda+1} \Delta_g (\sigma^{-\lambda} u) - (n\!-\!1) \sigma^\lambda \nabla_{\grad_g(\sigma)}(\sigma^{-\lambda} u) - \lambda(n\!+\!\lambda) \sigma^{-1} u \\ & = \sigma \Delta_g (u) - (n\!+\!2\lambda\!-\!1) \nabla_{\grad_g(\sigma)}(u) + \Con(g,\sigma;\lambda)(1) u \end{align*} for $u \in C^\infty(X^\circ)$. Thus, it only remains to calculate the constant term $\CT(\Con)(\lambda) \st \Con(\lambda)(1) \in C^\infty(X^\circ)$ of $\Con(\lambda)$. Note that \begin{align*} \CT(\Con)(\lambda) & = \sigma^{\lambda-1}(\Delta_{\sigma^{-2}g} - \lambda(n\!+\!\lambda) \id )(\sigma^{-\lambda}) \\ & = \sigma^{\lambda+1} \Delta_{g} (\sigma^{-\lambda}) + \lambda (n\!-\!1) \sigma^{-1} |\grad_g(\sigma)|^2 - \lambda(n\!+\!\lambda) \sigma^{-1} \end{align*} shows that $\CT(\Con)(\lambda)$ is a quadratic polynomial in $\lambda$. It is obvious that $\CT(\Con)(0)=0$. Next, we determine the leading coefficient of that polynomial. We choose orthonormal bases $\left\{\partial_i \right\}$ on the tangent spaces of the level hypersurfaces $\sigma^{-1}(c)$ of $\sigma$. The sets $\sigma^{-1}(c)$ are smooth manifolds if $c$ is sufficiently small. $\NV$ is perpendicular to these hypersurfaces. Let $\left\{ \alpha, dx^i \right\}$ be the dual basis of $\left\{ \NV, \partial_i \right\}$. Then \begin{equation}\label{Laplace-basis} \Delta_g (u) = \frac{1}{|\NV|^2} \nabla_{\NV}^2 (u) + \Delta_{g_\sigma}(u) + \frac{1}{|\NV|} H_\sigma \nabla_\NV (u) - \frac{1}{|\NV|^2} \langle du, \nabla_{\NV}(\NV) \rangle, \end{equation} where $-H_\sigma = \langle \alpha,\nabla_{\partial_i} (\partial_i) \rangle$ and $\Delta_{g_\sigma}$ denotes the tangential Laplacians for the induced metrics on the leaves $\sigma^{-1}(c)$; for more details, see Section <ref>. Since $\Delta_{g_\sigma} (\sigma^{-\lambda}) = 0$, it follows that the coefficient of $\lambda^2$ in $\sigma^{\lambda+1} \Delta_{g}(\sigma^{-\lambda})$ is given by $\sigma^{-1} |\NV|^{-2} \nabla_\NV(\sigma)^2 = \sigma^{-1} |\NV|^2$. Hence the leading coefficient of the quadratic polynomial $\CT(\Con)(\lambda)$ equals \sigma^{-1 }|\NV|^2 - \sigma^{-1}. Finally, we calculate \CT(\Con)(-1) = \Delta_g(\sigma) - (n-1) \sigma^{-1} |\NV|^2 + (n-1) \sigma^{-1}. These arguments prove that \begin{equation}\label{C-CT} \CT(\Con)(\lambda) = \lambda \left((n\!+\!\lambda) \sigma^{-1} (|\NV|^2-1) - \Delta_g(\sigma)\right). \end{equation} \begin{align*} \CT(L)(\lambda) + \CT(\Con)(\lambda) & = \lambda (n\!+\!2\lambda\!-\!1) \rho - \lambda \sigma \J + \lambda \left((n\!+\!\lambda) \sigma^{-1} (|\NV|^2-1) - \Delta_g(\sigma)\right) \\ & = \lambda (n\!+\!\lambda) \sigma^{-1}(2 \rho \sigma + |\NV|^2-1) \\ & = \lambda (n\!+\!\lambda) \sigma^{-1} (\SC(g,\sigma)-1) \end{align*} by the definition of $\rho$. This completes the proof. The above proof shows the identity (<ref>) only for functions with support near the boundary $M$. This will be enough for all later applications. For simplicity, we shall interpret (<ref>) and similar identities in this way without further mentioning. Let $g_+ = r^{-2} (dr^2 + h_r)$ be a Poincaré-Einstein metric in normal form relative to the metric $h$ on $M$. It lives on the space $(0,\varepsilon) \times M$ and satisfies $\Ric(g_+) = -n g_+$. Hence $\scal(g_+)=-n(n+1)$. Then $\SC(dr^2+h_r,r;\lambda) = \SC(g_+,1;\lambda)=1$, and the conjugation formula reads - L(dr^2+h_r,r;\lambda) = r^{\lambda-1} \circ (\Delta_{g_+} - \lambda(n+\lambda) \id ) \circ r^{-\lambda}. We refer to <cit.> for the discussion of the relation between this conjugation formula and a formula in <cit.>. The conjugation formula is equivalent to the identity \begin{equation}\label{CF-s} L(g,\sigma;\lambda) + \sigma^{\lambda-1} \circ \Delta_{\sigma^{-2}g} \circ \sigma^{-\lambda} = \lambda (n\!+\!\lambda) \sigma^{-1} \SC(g,\sigma) \id \end{equation} of operators acting on $C^\infty(X^\circ)$. The conformal covariance of the Laplace-Robin operator is an immediate consequence of these identities. More precisely, we have The Laplace-Robin operator satisfies \begin{equation}\label{CT-ID} L(\hat{g}, \hat{\sigma}; \lambda) \circ e^{\lambda\varphi} = e^{(\lambda-1)\varphi} \circ L(g,\sigma;\lambda), \; \lambda \in \C \end{equation} for all conformal changes $(\hat{g},\hat{\sigma}) = (e^{2\varphi}g,e^{\varphi}\sigma)$, $\varphi \in It suffices to note that $\hat{\sigma}^{-2} \hat{g} = \sigma^{-2} g$ and $\SC(\hat{g},\hat{\sigma}) = \SC(g,\sigma)$. Strictly speaking, the above arguments prove the conformal covariance of the operator $L(g,\sigma;\lambda)$ for boundary defining $\sigma$ when acting on $C^\infty(X^\circ)$. In <cit.>, the conformal covariance of the operator $L(g,\sigma;\lambda)$ for any pair $(g,\sigma)$ acting on $C^\infty(X)$ follows from its interpretation in terms of tractor calculus. For a direct proof, see <cit.>. The following consequence of the conjugation formula will be of central significance in the rest of the paper. We continue to assume that $\sigma$ is a boundary defining function and statements are valid near $M$. It holds \begin{equation}\label{L-Delta} - L(g,\sigma;\lambda) = \sigma^{\lambda-1} \circ (\Delta_{\sigma^{-2}g} - \lambda(n+\lambda) \id ) \circ \sigma^{-\lambda} \end{equation} for $\lambda \in \C$ iff $\SC(g,\sigma) = 1$. More generally, it holds -L(g,\sigma;\lambda) = \sigma^{\lambda-1} \circ (\Delta_{\sigma^{-2}g} - \lambda(n+\lambda) \id ) \circ \sigma^{-\lambda} + O(\sigma^n) if $\sigma$ satisfies $\SCY$. These identities are identities of operators acting on $C^\infty(X^\circ)$. For $N \in \N$, we define \begin{equation}\label{LN} L_N(g,\sigma;\lambda) \st L(g,\sigma;\lambda\!-\!N\!+\!1) \circ \cdots \circ L(g,\sigma;\lambda). \end{equation} In these terms, iterated application of (<ref>) implies Assume that $\SC(g,\sigma)=1$. Then it holds \begin{equation*} \sigma^{-\frac{n}{2}-N} \circ \prod_{j=0}^{2N-1} \left(\Delta_{\sigma^{-2}g} + \left(\frac{n}{2}\!+\!N\!-\!j\right) \left(\frac{n}{2}\!-\!N\!+\!j\right) \id \right) \circ \sigma^{\frac{n}{2}-N} = L_{2N} \left(g,\sigma;-\frac{n}{2}\!+\!N\right) \end{equation*} for $2N \le n$. In particular, we have \begin{equation} \sigma^{-n} \circ \prod_{j=0}^{n-1} \left(\Delta_{\sigma^{-2}g} + (n\!-\!j)j \id \right) = L_n (g,\sigma;0) \end{equation} The conjugation formula also sheds new light on the fact that the formal adjoint of a Laplace-Robin operator is another Laplace-Robin operator. The Laplace-Robin operator satisfies \begin{equation}\label{L-ad} L(g,\sigma;\lambda)^* = L(g,\sigma;-\lambda\!-\!n), \; \lambda \in \C, \end{equation} where $^*$ denotes the adjoint operator with respect to the Riemannian volume of $g$. More precisely, it holds \begin{equation}\label{adjoint} \int_X L(g,\sigma;\lambda)(\varphi) \psi dvol_g = \int_X \varphi L(g,\sigma;-\lambda\!-\!n)(\psi) dvol_g \end{equation} for $\varphi, \psi \in C_c^\infty(X^\circ)$. Let $\varphi,\psi \in C_c^\infty(X^\circ)$. The identity (<ref>) yields \begin{align*} & \int_X L(g,\sigma;\lambda)(\varphi) \psi dvol_g \\ & = - \int_X \sigma^{\lambda-1} \Delta_{\sigma^{-2}g} (\sigma^{-\lambda} \varphi) \psi dvol_g + \lambda (n\!+\!\lambda) \int_X \sigma^{-1} \SC(g,\sigma) \varphi \psi dvol_g. \end{align*} Note that $\lambda(n+\lambda)$ is invariant under the substitution $\lambda \mapsto -\lambda-n$. We rewrite the first integral in terms of volumes with respect to the metric $\sigma^{-2}g$ and apply the self-adjointness of $\Delta_{\sigma^{-2}g}$ with respect to the volume of the metric $\sigma^{-2}g$. Using $dvol_{\sigma^{-2}g} = \sigma^{-n-1} dvol_g$, we find - \int_X \sigma^{-\lambda} \varphi \Delta_{\sigma^{-2}g} (\sigma^{\lambda+n} \psi) dvol_{\sigma^{-2} g} = - \int_X \varphi \sigma^{-\lambda-n-1} \Delta_{\sigma^{-2}g} (\sigma^{\lambda+n} \psi) dvol_g. Now another application of (<ref>) implies the assertion (<ref>). For later applications, we need an extension of Corollary <ref> to another class of test functions. The proof of the following result will be given in Section <ref>. The identity (<ref>) continues to be true for $\psi \in C^\infty(X)$ and $\varphi \in C^2(X)$ so that $\iota^*(\varphi)=0$. The proof of this result actually shows that \begin{align}\label{adjoint-g} &\int_X L(g,\sigma;\lambda)(\varphi) \psi dvol_g - \int_X \varphi L(g,\sigma;-\lambda\!-\!n)(\psi) dvol_g \notag \\ & = (n\!+\!2\lambda) \int_M \iota^*(\varphi \psi |\NV|) dvol_h \end{align} for $\varphi , \psi \in C^2(X)$. The proof of this identity rests on a calculation of the left-hand side (see also <cit.>). Finally, we derive some basic commutator relations. For any $N \in \N$, it holds L(g,\sigma;\lambda\!+\!N) \circ \sigma^N - \sigma^N \circ L(g,\sigma;\lambda) = N(n\!+\!2\lambda\!+\!N) \sigma^{N-1} \SC(g,\sigma) \id. In particular, it holds \begin{equation}\label{sl2-comm} L(g,\sigma;\lambda\!+\!1) \circ \sigma - \sigma \circ L(g,\sigma;\lambda) = (n\!+\!2\lambda\!+\!1) \SC(g,\sigma) \id. \end{equation} \begin{align}\label{LN-gen-b} & L_N(g,\sigma;\lambda) \circ \sigma - \sigma \circ L_N(g,\sigma;\lambda\!-\!1) = N(n\!+\!2\lambda\!-\!N) L_{N-1}(g,\sigma;\lambda-1) \notag \\ & + \sum_{j=1}^N (n\!+\!2\lambda\!-\!2j\!+\!1) \underbrace{L(g,\sigma;\lambda\!-\!N\!+\!1) \circ \cdots \circ (\SC(g,\sigma)-1) \circ \cdots \circ L(g,\sigma;\lambda\!-\!1)}_{N factors}. \end{align} Moreover, if $\SC(g,\sigma)$ is nowhere zero, then for any $N \in \N$ it holds \begin{equation}\label{LN-gen-tilde} \tilde{L}_N(g,\sigma;\lambda\!+\!1) \circ \sigma - \sigma \circ \tilde{L}_N(g,\sigma;\lambda) = N(n\!+\!2\lambda\!-\!N\!+\!2) \tilde{L}_{N-1}(g,\sigma;\lambda), \end{equation} \begin{equation}\label{LR-general} \index{$\tilde{L}(g,\sigma)$} \tilde{L}(g,\sigma;\lambda) \st L(g,\sigma;\lambda) \circ \SC(g,\sigma)^{-1} \end{equation} \begin{equation}\label{LN-tilde} \index{$\tilde{L}_N(g,\sigma)$} \tilde{L}_N(g,\sigma;\lambda) \st \tilde{L}(g,\sigma;\lambda\!-\!N\!+\!1) \circ \cdots \circ \tilde{L}(g,\sigma;\lambda). \end{equation} The identity (<ref>) implies \begin{align*} & L(g,\sigma;\lambda) \circ \sigma^N - \sigma^N \circ L(g,\sigma;\lambda-N) \\ & = \lambda (n\!+\!\lambda) \sigma^{N-1} \SC(g,\sigma) \id - (\lambda\!-\!N) (n\!+\!\lambda\!-\!N) \sigma^{N-1} \SC(g,\sigma) \id \\ & = N(n\!+\!2\lambda\!-\!N) \sigma^{N-1} \SC(g,\sigma) \id. \end{align*} This proves the first commutator relation. The remaining claims are consequences. This completes the proof. The above commutator relations substantially simplify if $\SC=1$. Although we proved the identities in Corollary <ref> as identities of operators acting on $C^\infty(X^\circ)$, they are also valid for operators acting on $C^\infty(X)$ (<cit.>). § SYMMETRY BREAKING OPERATORS In the present section, we discuss some representation theoretical aspects of the results in Section <ref>. The simplest special case of the Laplace-Robin operator $L$ appears for the hyperplane $M=\R^n$ in $X=\R^{n+1}$ with the flat Euclidean metric $g_0$. Let $M$ be given by the zero locus of the defining function $\sigma_0 = x_{n+1}$. We shall also write $\sigma_0=r$ and $g_+ = r^{-2} g_0$. Then $\J = \rho = 0$ and we obtain \begin{equation}\label{L-flat} L(g_0,\sigma_0;\lambda) = (n+2\lambda-1) \partial_{n+1} - x_{n+1} \Delta_{\R^{n+1}}: C^\infty(\R^{n+1}) \to C^\infty(\R^{n+1}). \end{equation} An easy calculation shows the conjugation formula L(g_0,r;\lambda) = r^{\lambda-1} \circ (-\Delta_{g_+} + \lambda(n+\lambda)) \circ r^{-\lambda}. It implies that the operator $L(g_0,r;\lambda)$ is an intertwining operator for spherical principal series representations. Indeed, let $\gamma \in SO(1,n+1)$ be an isometry of the hyperbolic metric $g_+=r^{-2} g_0$ acting on the upper-half space $r > 0$. Then we calculate \begin{align*} & L(g_0,r;\lambda) \left( \left(\frac{\gamma_*(r)}{r}\right)^{-\lambda} \gamma_*(u)\right) \\ & = r^{\lambda-1} (-\Delta_{g_+} + \lambda(n+\lambda)) \left(r^{-\lambda} \left(\frac{\gamma_*(r)}{r}\right)^{-\lambda} \gamma_*(u)\right) \\ & = r^{\lambda-1} (-\Delta_{g_+} + \lambda(n+\lambda)) (\gamma_*(r^{-\lambda} u)) \\ & = r^{\lambda-1}\gamma_* (-\Delta_{g_+} + \lambda(n+\lambda))(r^{-\lambda} u)) \\ & = \left(\frac{\gamma_*(r)}{r}\right)^{-\lambda+1} \gamma_* (r^{\lambda-1} (-\Delta_{g_+}+\lambda(n+\lambda))(r^{-\lambda} u) \\ & = \left(\frac{\gamma_*(r)}{r}\right)^{-\lambda+1} \gamma_* L(g_0,r;\lambda)(u). \end{align*} In other words, it holds \begin{equation}\label{rep-flat} L(g_0,r;\lambda) \circ \pi^0_{-\lambda}(\gamma) = \pi^0_{-\lambda+1}(\gamma) \circ L(g_0,r;\lambda) \end{equation} with $\pi_\lambda^0$ non-compact model of spherical principal series representation \pi^0_\lambda(\gamma) \st \left( \frac{\gamma_*(r)}{r} \right)^\lambda \gamma_*. Note that \frac{\gamma_*(r)}{r} = e^{\Phi_\gamma}, where $\Phi_\gamma$ is the conformal factor of the conformal transformation induced by $\gamma$ with respect to the Euclidean metric, i.e., $\gamma_*(g_0) = e^{2\Phi_\gamma} g_0$. The representation $\pi_\lambda^0(\gamma)$ is actually well-defined for all $\gamma \in SO(1,n+2)$ acting on $\R^{n+1}$ (viewed as the boundary of hyperbolic space of dimension $n+2$). However, the intertwining property (<ref>) holds true only for the subgroup of $SO(1,n+1)$ leaving the boundary $r=0$ of the upper half-space invariant. The fact that $L(g_0,r;\lambda)$ is an intertwining operator for a subgroup of the conformal group of the Euclidean metric on $\R^{n+1}$ connects it with the theory of symmetry breaking operators. In fact, it follows from the above that the compositions D_N(\lambda) \st \iota^* L(\lambda-N+1) \circ \cdots \circ L(\lambda): C^\infty(\R^{n+1}) \to C^\infty(\R^n), \; N \in \N D_N(\lambda) \circ \pi_{-\lambda}^0(\gamma) = \pi_{-\lambda+N}^{0 \prime}(\gamma) \circ D_N(\lambda), \; \gamma \in SO(1,n+1) and Clerc <cit.> proved that $D_N(\lambda)$ coincides with the symmetry breaking operator introduced in <cit.>.[$\pi_\lambda^{0 \prime}$ denotes the analogous representation on functions on the subspace $\R^n$.] Similarly, let $M = S^n$ be an equatorial subsphere of $X=S^{n+1}$ with the round metric $g$. Let $M$ be defined as the zero locus of the height function $\sigma=\He$ being defined as the restriction of $x_{n+2}$ to $S^{n+1}$. Then $\J = \frac{n+1}{2}$ and $\He$ height (n+1) \rho = - \Delta_{S^{n+1}} \He + \J \He = -(n+1) \He + \frac{n+1}{2} \He = - \frac{n+1}{2} \He using the fact that $\He$ is an eigenfunction of the Laplacian on the sphere $S^{n+1}$. Thus, $\rho = \frac{1}{2} \He$ and we obtain \begin{equation}\label{L-sphere} L(g,\He;\lambda) = (n+2\lambda-1) \nabla_{\grad (\He)} - \He \Delta_{S^{n+1}} + \lambda(\lambda+1) \He: C^\infty(S^{n+1}) \to C^\infty(S^{n+1}). \end{equation} A calculation shows that L(g,\He;\lambda) = \He^{\lambda-1} \circ (-\Delta_{\He^{-2}g} + \lambda(n+\lambda)) \circ \He^{-\lambda}. Again, the operator $L(g,\He;\lambda)$ is an intertwining operator for spherical principal series representations. Indeed, it holds L(g,\He;\lambda) \circ \pi_{-\lambda}(\gamma) = \pi_{-\lambda+1}(\gamma) \circ L(g,\He;\lambda) for all $\gamma \in SO(1,n+1)$ acting on the upper hemisphere $\He > 0$ of $S^{n+1}$. Here $\pi_\lambda$ spherical principal series representation \pi_\lambda(\gamma) \st \left( \frac{\gamma_*(\He)}{\He} \right)^\lambda \gamma_*. The latter representations are well-defined for $\gamma \in SO(1,n+2)$ acting on $S^{n+1}$. Note that \frac{\gamma_*(\He)}{\He} = e^{\Phi_\gamma}, where $\Phi_\gamma$ is the conformal factor of the conformal transformation induced by $\gamma$ with respect to the round metric $g$, i.e., $\gamma_*(g) = e^{2\Phi_\gamma} g$. We also note that the operator (<ref>) is equivalent to the intertwining operator displayed in <cit.>. We omit the details of that calculation. Finally, we observe that the above two models of the Laplace-Robin operator are conformally equivalent. In fact, let $\kappa: S^{n+1} \to \R^{n+1}$ be the stereographic projection. Then \kappa^* (\He) = \Phi x_{n+1} \quad \mbox{and} \quad \kappa^*(g) = \Phi^2 \sum_{i=1}^{n+1} dx_i^2 = \Phi^2 g_0 with $\Phi = 2/(1+|x|^2)$ <cit.>. Hence \begin{align*} \kappa^* L(g,\He;\lambda) \kappa_* & = L(\kappa^*(g),\kappa^*(\He);\lambda) \\ & = L(\Phi^2 g_0,\Phi x_{n+1};\lambda) \\ & = \Phi^{\lambda-1} L(g_0,x_{n+1};\lambda) \Phi^{-\lambda} \end{align*} using a very special case of the conformal invariance of the Laplace-Robin operator (Corollary <ref>). By combining this conjugation formula with the results in the later sections, it follows that the equivariant families $D_{2N}^c(\lambda): C^\infty(S^{n+1}) \to C^\infty (S^n)$ constructed in <cit.> can be regarded as residue families $\D_{2N}^{res}(g,\He;\lambda)$ (as defined in Section <ref>). § ADAPTED COORDINATES, RENORMALIZED VOLUME COEFFICIENTS AND A FORMULA FOR $\B_N$ Let $X$ be compact with closed boundary $M$ and let $\sigma$ be a defining function of $M$, i.e., $\sigma^{-1}(0)=M$, $\sigma > 0$ on $X \setminus M$ and $d\sigma|_M \ne 0$. Let $\iota: M \hookrightarrow X$ be the embedding and $h = \iota^*(g)$. We start with the definition of two different types of local coordinates of $X$ near the boundary: geodesic normal coordinates and adapted coordinates. $w_j$ renormalized volume coefficients (normal coordinates) Geodesic normal coordinates are defined by the normal geodesic flow of the hypersurface $M$, i.e., we consider a diffeomorphism of $I \times M$ (with a small interval $I = [0,\varepsilon)$) onto a neighborhood of $M$ in $X$, which is defined by $\Phi^r$ geodesic flow \kappa: I \times M \ni (r,x) \mapsto \Phi^r(x) \in X, where $\Phi^r$ is the geodesic flow with initial speed given by a unit normal field on $M$. Then $\kappa^*(g)$ has the form $dr^2 + h_r$ for a one-parameter family $h_r$ on $M$. Let $u(r)$ $u_j$ volume coefficients \begin{equation}\label{v-geo} u(r) \st dvol_{h_r}/dvol_h, \quad u(r) = \sum_{j \ge 0} r^j u_j, \quad u_j \in C^\infty(M). \end{equation} Now, if $\SC(g,\sigma) = 1$, then the volume form of the singular metric $\sigma^{-2} g$ has the form \begin{align}\label{def-w} dvol_{\kappa^*(\sigma^{-2} g)} & = \sigma(r)^{-n-1} u(r) dr dvol_h \notag \\ & = r^{-n-1} w(r) dr dvol_h. \end{align} for $\sigma(r) = \kappa^*(\sigma)$ and some $w \in C^\infty(I \times M)$. Moreover, we have expansions \begin{equation}\label{RVC-A} w(r) = 1 + \sum_{j \ge 1} r^j w_j \end{equation} with $w_j \in C^\infty(M)$ and dvol_{\kappa^*(\sigma^{-2} g)} = \sum_{j \ge 0} r^{-n-1+j} w_j dr dvol_h. Following <cit.>, the coefficients $w_j \in C^\infty(M)$ for $j \le n$ are called singular Yamabe renormalized volume coefficients. Note that the definition of the coefficients $w_j \in C^\infty(M)$ involves the Taylor expansion of $\sigma(r)$ in $r$. Special interest deserves the critical coefficient $w_n$ since for closed $M$, the total integral \int_M w_n dvol_h is conformally invariant <cit.>. $\eta$ adapted coordinates Similarly, adapted coordinates are associated to the data $(g,\sigma)$ through a diffeomorphism \eta: I \times M \ni (s,x) \mapsto \Phi_\mathfrak{X}^s(x) \in X onto a open neighborhood of $M$ in $X$ (with some small interval $I = [0,\varepsilon)$), where $\Phi_\mathfrak{X}^s$ denotes the flow of the vector field $\mathfrak{X}$ \begin{equation}\label{X-field} \mathfrak{X} \st \NV /|\NV|^2, \quad \NV = \grad_g(\sigma) \end{equation} with $\Phi^0_\mathfrak{X} = \id$. We shall also use the notation $\mathfrak{X}_\sigma$ in cases where the dependence on $\sigma$ is important. Note that $|\mathfrak{X}| = 1/|\NV|$. Then (d/ds)(\sigma \circ \eta) = \langle d\sigma, \mathfrak{X} \rangle = \langle d \sigma, \NV \rangle / |\NV|^2 \stackrel{!}{=} 1 and the differential of $\eta$ maps the vector field $\partial_s$ to the vector field $\mathfrak{X}$ (see Section <ref>). This implies the important relation \begin{equation}\label{key-pb} \eta^*(\sigma) = s \end{equation} and the intertwining property \begin{equation}\label{intertwine} \eta^* \circ \mathfrak{X} = \partial_s \circ \eta^*, \end{equation} where $\partial_s$ and $\mathfrak{X}$ are viewed as first-order differential operators. Therefore, \eta^* \circ \mathfrak{X}^k = \partial_s^k \circ \eta^*, and by composition with $\iota^*$, we obtain \begin{equation}\label{translate} \iota^* \partial_s^k \circ \eta^* = \iota^* \mathfrak{X}^k = \iota^* \left(|\NV|^{-2} \nabla_\NV\right)^k. \end{equation} Now if $\SC(g,\sigma)=1$, i.e., if $|\NV|^2=1-2\sigma \rho$, then it follows from (<ref>) that the Taylor coefficients in the variable $s$ of any function $\eta^*(u)$ with $u \in C^\infty(X)$ can be written as linear combinations of iterated gradients $\iota^* \nabla_\NV^k (u)$ with coefficients that are polynomials in the quantities $\iota^* \nabla_\NV^k(\rho) \in C^\infty(M)$. In particular, it \begin{equation}\label{trans-1-2} \iota^* \partial_s \circ \eta^* = \iota^* \nabla_\NV \quad \mbox{and} \quad \iota^* \partial^2_s \circ \eta^* = \iota^* \nabla_\NV^2 - 2 H \nabla_\NV. \end{equation} For more details, we refer to Section <ref>. If $\sigma$ satisfies only the weaker condition $\SCY$, then $|\NV|^2 = 1 - 2\sigma \rho + O(\sigma^{n+1})$ and the same conclusions are true for sufficiently small $k$. Note that the metric $\eta^*(g)$ has the form \begin{equation}\label{normal-adapted} \eta^*(|\NV|^{-2}) ds^2 + h_s \end{equation} with a one-parameter family $h_s$ of metrics on $M$ so that $h_0 = h = \iota^*(g)$. We shall refer to (<ref>) as the normal form of $g$ in adapted coordinates. We expand $h_s = \sum_{j \ge 0} h_{(j)} s^j$. It follows from (<ref>) that \iota^* \partial_s^k (\eta^*(|\NV|^{-2})) = \iota^* (\nabla_\NV(|\NV|^{-2}))^k. Thus, if $\sigma$ satisfies $\SCY$, the Taylor coefficients of the coefficient $\eta^*(|\NV|^{-2})$ in the variable $s$ are polynomials in the quantities In later calculations in adapted coordinates, we shall often use the same notation for quantities like $\rho$ and $\J$ and their pull-backs by $\eta$ without further mentioning. $v_j$ renormalized volume coefficients (adapted coordinates) Now (<ref>) implies dvol_{\eta^*(g)} = \eta^*(|\NV|)^{-1} ds dvol_{h_s} = v(s) ds dvol_h for some $v(s) \in C^\infty(I \times M)$. Since the condition $\SCY$ implies $|\NV|=1$ on $M$, we get $v(0,x)=1$, and we have an expansion \begin{equation}\label{RVC-B} v(s) = 1 + \sum_{j \ge 1} s^j v_j \quad \mbox{with $v_j \in C^\infty(M)$}. \end{equation} The coefficients $v_j$ for $j \le n$ also will be called singular Yamabe renormalized volume coefficients. They describe the volume of the singular metric $\sigma^{-2} g$ through the expansion dvol_{\eta^*(\sigma^{-2} g)} = s^{-n-1} v(s) ds dvol_h = \sum_{j \ge 0} s^{-n-1+j} v_j ds dvol_h. Again, special interest deserves the critical coefficient $v_n$ since \int_M v_n dvol_h is conformally invariant for closed $M$. This follows from the equality \begin{equation}\label{w=v} \int_M v_n dvol_h = \int_M w_n dvol_h \end{equation} which can be proved by the following argument of <cit.>. The identity $|d\sigma|^2_{\tilde{g}}=1$ for $\tilde{g} = |d\sigma|_g^2 g$ shows that $\sigma$ can be viewed as the distance function $d_M^{\tilde{g}}$ of $M$ in the metric $\tilde{g}$. Hence it holds vol_{\sigma^{-2}g} (\{ \sigma > \varepsilon \}) = vol_{\sigma^{-2} g} (\{ d_M^{\tilde{g}} > \varepsilon \}) = vol_{\tilde{\sigma}^{-2} \tilde{g}} (\{ d_M^{\tilde{g}} > \varepsilon \}) with $\tilde{\sigma}= |d\sigma|_g \sigma$. By comparing the coefficients of $\log \varepsilon$ in the expansions of both sides, we find \int_M v_n(g) dvol_{\iota^*(g)} = \int_M w_n(\tilde{g}) dvol_{\iota^*(\tilde{g})}. Now the conformal invariance of the latter integral implies the equality (<ref>). Following <cit.>, the integral $\A$ anomaly \begin{equation}\label{def-A} \A \st \int_{M^n} v_n dvol_h \end{equation} for a closed $M$ is called the singular Yamabe energy of $M$. The quantity $\A$ appears in the asymptotic expansion of the volume of the singular metric $\sigma^{-2}g$ (Theorem <ref>). Now we continue with the We use adapted coordinates. It suffices to prove that the operator $\mathbf{L}(\lambda) \st L(\eta^*(g),s;\lambda)$ satisfies \int_X \mathbf{L}(\lambda)(\varphi) \psi dvol_{\eta^*(g)} = \int_X \varphi \mathbf{L}(-\lambda-n)(\psi) dvol_{\eta^*(g)} if $X=[0,\varepsilon) \times M$, $\psi \in C_c^2([0,\varepsilon) \times M)$ and $\varphi \in C^2([0,\varepsilon) \times M)$ so that $\varphi(0,x)=0$. Now, by definition \mathbf{L}(\lambda) = (n\!+\!2\lambda\!-\!1) (\eta^*(|\NV|^2) \partial_s + \lambda \eta^*(\rho)) - s (\Delta_{\eta^*(g)} + \lambda \eta^*(\J)). In the following, we simplify the notation by writing $g$, $\rho$, and $\J$ instead of the pull-backs of these quantities by $\eta$. Then $a$ $h_s$ \begin{equation}\label{LR-adapted} \mathbf{L}(\lambda) = (n\!+\!2\lambda\!-\!1) (a \partial_s + \lambda \rho) - s (\Delta_g + \lambda \J), \end{equation} where $a \st |\NV|^2$ . In these terms, the background metric reads $g = a^{-1} ds^2 + h_s$ and we obtain $dvol_g = a^{-1/2} ds dvol_{h_s}$. Hence \begin{equation*} v = a^{-1/2} (\det (h_s)/\det(h))^{1/2} \end{equation*} \begin{equation}\label{vol-g} \frac{v'}{v} = -\frac{1}{2} \frac{a'}{a} + \frac{1}{2} \tr (h_s^{-1} h'_s), \end{equation} where $'$ denotes the derivative in the variable $s$. An easy calculation shows that \begin{equation}\label{Laplace-adapted} \Delta_g = a \partial_s^2 + \frac{a}{2} \tr (h_s^{-1} h'_s) \partial_s + \frac{1}{2} a' \partial_s - \frac{1}{2} (d\log a,d \cdot)_{h_s} + \Delta_{h_s}. \end{equation} Now we observe that \begin{align*} & \int_X a \varphi' \psi dvol_g = \int_X a \varphi' \psi v ds dvol_h \\ & = - \int_X \varphi \left[a \psi' + a \frac{v'}{v} \psi + a' \psi \right] v ds dvol_h = - \int_X \varphi \left[a \psi' + a \frac{v'}{v} \psi + a' \psi \right] dvol_g \end{align*} \int_X s \Delta_g(\varphi) \psi dvol_g = \int_X \varphi \Delta_g(s\psi) dvol_g using Green's formula and the assumptions. The expression (<ref>) shows that \Delta_g(s\psi) = s \Delta_g(\psi) + 2 a \psi' + \frac{a}{2} \tr (h_s^{-1} h'_s) \psi + \frac{1}{2} a' \psi. \begin{align*} \int_X \mathbf{L}(\lambda)(\varphi) \psi dvol_g & = -(n\!+\!2\lambda\!-\!1) \int_X \varphi \left( a \psi' + a \frac{v'}{v} \psi + a' \psi \right) dvol_g \\ & - \int_X \varphi \left(s \Delta_g(\psi) + 2 a \psi' + \frac{a}{2} \tr (h_s^{-1} h'_s) \psi + \frac{1}{2} a' \psi \right) dvol_g \\ & + \int_X \varphi \psi (\lambda(n\!+\!2\lambda\!-\!1)\rho - \lambda s \J) dvol_g. \end{align*} On the other hand, we have \begin{align*} & \int_X \varphi \mathbf{L}(-\lambda\!-\!n)(\psi) dvol_g \\ & = \int_X \varphi \left(-(n\!+\!2\lambda\!+\!1) a \psi' - s \Delta_g (\psi) + (n\!+\!2\lambda\!+\!1)(\lambda\!+\!n) \rho \psi + (\lambda\!+\!n) s \J \psi \right) dvol_g. \end{align*} It follows that the assertion is equivalent to the identity \begin{align*} & -(n\!+\!2\lambda\!-\!1) a \frac{v'}{v} - (n\!+\!2\lambda\!-\!1) a' - \frac{a}{2} \tr (h_s^{-1} h'_s) - \frac{1}{2} a' + \lambda(n\!+\!2\lambda\!-\!1) \rho - \lambda s \J \\ & = (n\!+\!2\lambda\!+\!1)(\lambda\!+\!n) \rho + (\lambda \!+\!n) s\J. \end{align*} By (<ref>), this identity is equivalent to -a \frac{v'}{v} - a' = (n+1) \rho + s \J. The identities (<ref>) and (<ref>) also show that \begin{equation}\label{Laplace-s} \Delta_g(s) = \frac{a}{2} \tr (h_s^{-1} h'_s) + \frac{1}{2} a' \stackrel{!}{=} a \frac{v'}{v} + a'. \end{equation} Thus, we have reduced the assertion to the identity - \Delta_g(s) = (n+1) \rho + s \J. But this is just the definition of $\rho$. The proof is complete. The above arguments also prove the relation (<ref>). In fact, partial integration and Green's formula yield the additional terms (n\!+\!2\lambda\!-\!1) \int_M \iota^*(a \varphi \psi v) dvol_h + \int_M \iota^* (\varphi \psi |\NV|) dvol_h = (n+2\lambda) \int_M \iota^*(\varphi \psi |\NV|) dvol_h since the unit normal field on $M$ is $|\NV| \partial_s$ and $v_0 = |\NV|^{-1}$ Note that equation (<ref>) can be written in the form \begin{equation}\label{bL2} v(s) \Delta_{g} (s) = \partial_s (v(s) a); \end{equation} we recall that $a = \eta^*(|\NV|^2)$. As a corollary of this formula, we obtain a useful formula for $v(s)$ in terms of $\rho$ and $\J$. If $\sigma$ satisfies $\SCY$, then it holds \begin{equation}\label{bL} \frac{v'}{v} = \frac{-(n-1)\rho + 2 s \rho' - s \J}{1-2s\rho} + O(s^n). \end{equation} Here $\rho$ and $\J$ are identified with their pull-backs by $\eta$. If $\rho=0$, then it holds $v'/v = - s \J + O(s^n)$. The latter case contains the Poincaré-Einstein case. We write (<ref>) in the form v(s) (-(n+1) \rho - s \J) = \partial_s (v(s) a). Now the assumption implies $a = \SC - 2s \rho = 1- 2s \rho + O(s^{n+1})$. Hence \begin{align*} v(s) (-(n+1) \rho - s \J) & = \partial_s (v(s) (1-2s \rho+ O(s^{n+1}))) \\ & = v'(s) (1-2s\rho) - 2 v(s) \rho - 2 v(s) s \rho' + v O(s^n). \end{align*} Now simplification proves the claim. In the Poincaré-Einstein case, one easily shows that $\rho=0$ and $v'/v = - s \J$ (see Example <ref>). Lemma <ref> can be used to derive formulas for the coefficients $v_k$ with $k \le n$ in terms of the Taylor coefficients of $\rho$ and $\J$. In particular, we obtain For $\N \ni k \le n$, we have v_k = -(n\!-\!2k\!+\!1) \iota^* \frac{1}{k!} \partial_s^{k-1}(\rho) + LOT, where LOT refers to terms with lower-order derivatives of $\rho$ and $\J$. We consider the coefficient of $s^{k-1}$ in the expansion of $v'/v$. On the one hand, it equals $k v_k$. On the other hand, (<ref>) yields the expression \iota^* \partial_s^{k-1} (\rho) \left( - \frac{n-1}{(k-1)!} + \frac{2}{(k-2)!} \right) for this coefficient. This implies the assertion. In particular, the critical coefficient $v_n$ involves the quantity $\iota^* \partial_s^{n-1}(\rho)$. Conversely, a version of Lemma <ref> implies a recursive formula for the Taylor coefficients of $\rho$. For the discussion of that formula, we introduce the notation $\mathring{v}(s)$ \begin{equation}\label{ring-v} \mathring{v}(s) \st dvol_{h_s} / dvol_h. \end{equation} \frac{\mathring{v}'}{\mathring{v}} = \frac{1}{2} \tr (h_s^{-1} h_s'). We also recall that $a = 1- 2s \rho + O(s^{n+1})$ if $\sigma$ satisfies $\SCY$. The recursive formula for the Taylor coefficients of $\rho$ will be a consequence of a first-order differential equation. If $\sigma$ satisfies $\SCY$, then $\rho$ solves the differential equation \begin{equation}\label{magic-rec} - s \rho' + n \rho + a \frac{\mathring{v}'}{\mathring{v}} + s \J = O(s^n) \end{equation} with the initial condition $\rho(0) = - H$. The identity (<ref>) implies \begin{equation}\label{v-deco} \frac{v'}{v} = - \frac{1}{2} \frac{a'}{a} + \frac{\mathring{v}'}{\mathring{v}}. \end{equation} We use this decomposition on the left-hand side of (<ref>) and multiply the resulting identity with $a$. Then (n-1) \rho - 2 s \rho' + s \J - \frac{1}{2} a' + a \frac{\mathring{v}'}{\mathring{v}} = O(s^n). By $a' = - 2\rho - 2 s\rho' + O(s^n)$, this identity simplifies to (<ref>). By restriction of (<ref>) to $s=0$, we obtain $n \rho(0) + \tr (L) = 0$ using $h_{(1)} = 2L$ (see (<ref>)). Hence $\rho(0) = - H$. Another proof of $\rho(0) = -H$ will be given in Lemma <ref>. By repeated differentiation of the identity (<ref>) in the variable $s$, it follows that the Taylor coefficients of $\rho$ can be determined recursively using the Taylor coefficients of $h_s$ and $\J$. More precisely, we obtain Assume that $\sigma$ satisfies $\SCY$. Then \begin{equation}\label{rec-rho-full} (n-k) \partial_s^k(\rho)|_0 = -\partial_s^k \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 + 2 \sum_{j=1}^k j \binom{k}{j} \partial_s^{j-1}(\rho)|_0 \partial_s^{k-j} \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 - k \partial_s^{k-1}(\J)|_0 \end{equation} for $1 \le k \le n-1$. For a discussion of more details of such types of formulas in low-order cases, we refer to Section <ref>. In particular, we use (<ref>) to derive explicit formulas for the first two derivatives of $\rho$ in $s$ at $s=0$. Let $g_+ = s^{-2} (ds^2 + h_s)$ be a Poincaré-Einstein metric. Assume that $g=ds^2+h_s$ is smooth up to the boundary. In particular, the obstruction tensor vanishes. In that case, adapted coordinates coincide with geodesic normal coordinates. Now it holds $\rho=0$. We show that the vanishing of the Taylor coefficients of $\rho$ (in the variable $s$) up to order $n-1$ recursively follows from (<ref>). In fact, assume that we know that $\partial_s^j (\rho)|_0=0$ for $j=0,\dots,k-1$. Then the right-hand side of (<ref>) simplifies to -\partial_s^k \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 - k \partial_s^{k-1}(\J)|_0. k \partial_s^{k-1}(\J)|_0 = \partial_s^k(s \J)|_0 = - \frac{1}{2} \partial_s^k (\tr (h_s^{-1}h_s')|_0 = - \partial_s^k\left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 \begin{equation}\label{J-trace} \J = -\frac{1}{2s} \tr (h_s^{-1}h_s') \end{equation} (which follows by combining the Einstein condition with the conformal transformation law for scalar curvature - for the details see <cit.>). Hence (<ref>) implies $\partial_s^k(\rho)|_0=0$. Alternatively, we could note that the relation (<ref>) transforms the differential equation (<ref>) into -s \rho' + n\rho - 2s \rho \frac{\mathring{v}'}{\mathring{v}} = O(s^n) with the initial condition $\rho(0) = 0$. Then $\rho=0$ is the unique solution of this initial value problem. For $k=n$, the coefficient on the left-hand side of (<ref>) vanishes. This suggests the following formula for the singular Yamabe obstruction. Assume that $\sigma$ satisfies $\SCY$. Then \begin{equation}\label{obstruction-magic} (n\!+\!1)! \B_n = -2\partial_s^n \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 + 4 \sum_{j=1}^n j \binom{n}{j} \partial_s^{j-1}(\rho)|_0 \partial_s^{n-j} \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 - 2n \partial_s^{n-1}(\J)|_0. \end{equation} We start with general data $(g,\sigma)$. As before, we identify $|\NV|^2$ and $\SC$ with their respective pull-backs by $\eta$. Then $\SC = |\NV|^2 + 2 s \rho = a + 2s\rho$. In these terms, the identity (<ref>) reads v (-(n+1) \rho - s \J) = v' a + v a'. \begin{align*} a \frac{v'}{v} & = - (n+1) \rho - s \J - \partial_s(a) \\ & = -(n+1) \rho - s \J + 2 \partial_s (s\rho) - \partial_s (\SC-1) \\ & = -(n-1) \rho - s \J + 2 s \rho' - \partial_s(\SC-1). \end{align*} We decompose the left-hand side using (<ref>) and reorder. This gives a \frac{\mathring{v}'}{\mathring{v}} - \frac{1}{2} a' + (n-1) \rho + s \J - 2 s \rho' + \partial_s(\SC-1) = 0. Now, using $a = (1-2s\rho) + (\SC-1)$, we obtain the relation \begin{equation}\label{Basic-R} - s \rho' + n \rho + a \frac{\mathring{v}'}{\mathring{v}} + s \J = -\frac{1}{2} \partial_s (\SC-1) \end{equation} which improves (<ref>). Now, assuming that $\sigma$ satisfies $\SCY$, differentiate (<ref>) $n$ times in $s$. By $\partial_s^j(a)|_0 = -2j \partial_s^{j-1}(\rho)|_0$ for $1 \le j \le n$ and \partial_s^{n+1}(\SC -1)|_0 = (n+1)! \B_n, this proves the assertion. The basic relation (<ref>) will be confirmed in a number of special cases with $\SC = 1$ in Examples <ref>–<ref>. Proposition <ref> and Theorem <ref> should be compared with <cit.>. The latter result establishes formulas for the restrictions of normal derivatives $\nabla_\NV^k(\rho)$ of $\rho$ to $M$ and for the obstruction $\B_n$ in terms of lower-order normal derivatives of $\rho$ and additional terms. The above results clarify the structure of all such additional terms. Here it is crucial to work in adapted coordinates. Note that the formula (<ref>) shows that the obstruction $\B_n$ involves the Taylor coefficients $h_{(k)}$ of $h_s$ (in the normal form (<ref>) of $g$ in adapted coordinates) for $k \le n+1$. In Section <ref>, we shall derive the classical formula for $\B_2$ (see (<ref>) and <cit.>) from (<ref>). Similarly, in Section <ref> we evaluate the formula (<ref>) for the obstruction $\B_3$ in case of a (conformally) flat background. Finally, we apply the above results to determine the leading term of the obstruction $\B_n$ for an embedding $M^n \hookrightarrow \R^{n+1}$ if $n$ is even. First, we note that Theorem <ref> and Proposition <ref> show that $\B_n$ is a functional of the second fundamental form $L$. Here we also use the formulas which recursively determine the coefficients of $h_s$ and the formula for the first two coefficients. The claim for the Taylor coefficients of $h_s$ is a bit more complicated since it involves the proof that the variations of the Christoffel symbols also lead to functionals of $L$. For a flat background metric and even $n$, it holds \begin{equation}\label{Bn-deco} (n+1)! \B_n = c_n \Delta^\frac{n}{2} (H) + nl \end{equation} c_n = - 2 \frac{(n-1)!!}{(n-2)!!} and a non-linear functional $nl$ of $L$. For odd $n$, the obstruction $\B_n$ is non-linear in $L$. The non-linear part in (<ref>) can also be described as a term of lower differential order. In fact, we can write $\B_n$ as a sum of terms that are homogeneous in $L$. In each such term, the sum of the number of derivatives and the homogeneous degree in $L$ is $n+1$. But the non-linear terms in (<ref>) consist of homogeneous terms of degree at least $2$. One should compare that version of the structural result for $\B_n$ with <cit.>. Theorem <ref> extends the following observations. By the second formula in (<ref>), $\B_2$ is the sum of a constant multiple of $\Delta(H)$ being linear in $L$ and a term that is cubic in $L$ and does not contain derivatives. Similarly, the first three terms in (<ref>) are homogeneous of degree $2$ in $L$ and each such term involves $2$ derivatives. Let $n$ be even. We extract from the formula \begin{equation*}\label{ob-magic} (n\!+\!1)! \B_n = -2 \partial_s^n \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 + 4 \sum_{j=1}^n j \binom{n}{j} \partial_s^{j-1}(\rho)|_0 \partial_s^{n-j} \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 \end{equation*} the contributions which are linear in $L$. In the following, the symbol $nl$ indicates non-linear terms. First, we ignore in this sum all products with at least two factors. Hence (n+1)! \B_n = - 2 \partial_s^n \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 + nl. Moreover, the expansion \frac{\mathring{v}'}{\mathring{v}} = \frac{1}{2} \tr (h_s^{-1} h_s') = \frac{1}{2} \sum_{k \ge 1} ( k \tr (h_{(k)}) + nl ) s^{k-1} \partial_s^n \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 = \frac{1}{2} (n+1)! \tr (h_{(n+1)}) + nl. In order to evaluate $h_{(n+1)}$, we $(n-1)$-times differentiate in $s$ the determining relation (<ref>) for $h_s$. Then \frac{1}{2}\partial_s^{n+1}(h_s)|_0 = - \Hess (\partial_s^{n-1}(s\rho))|_0) + nl using $g^{00} = a = 1 - 2s \rho$. Hence \frac{1}{2} (n+1)! h_{(n+1)} = - (n-1) \Hess (\partial_s^{n-2}(\rho)|_0 )+ nl. These results imply \begin{align}\label{Bn-leading} (n+1)! \B_n & = - (n+1)! \tr (h_{(n+1)}) + nl \notag \\ & = 2 (n-1) \Delta (\partial_s^{n-2}(\rho)|_0) + nl. \end{align} Now Proposition <ref> shows that \begin{align*} (n-k) \partial_s^k(\rho)|_0 & = - \partial_s^k \left(\frac{\mathring{v}'}{\mathring{v}}\right)|_0 + nl \\ & = - \frac{1}{2} (k+1)! \tr (h_{(k+1)}) + nl. \end{align*} But $(k-1)$-times differentiating in $s$ the determining relation for $h_s$, shows that \begin{align*} \frac{1}{2} \partial_s^{k+1} (h_s)|_0 & = - \Hess(\partial_s^{k-1}(s \rho)|_0) + nl \\ & = - (k-1) \Hess (\partial_s^{k-2}(\rho)|_0) +nl. \end{align*} \begin{equation}\label{rho-leading} (n-k) \partial_s^k(\rho)|_0 = (k-1) \Delta (\partial_s^{k-2}(\rho)|_0) + nl. \end{equation} Combining (<ref>) and (<ref>) gives \begin{align*} (n+1)! \B_n & = 2 (n-1) \Delta (\partial_s^{n-2}(\rho)|_0) + nl \\ & = (n-1)(n-3) \Delta^2 (\partial_s^{n-4}(\rho)|_0) + nl \\ & = \cdots = 2 (n-1)!!/(n-2)!! \Delta^\frac{n}{2}(\rho|_0) + nl. \end{align*} This implies the assertion using $\rho|_0 = -H$ (Lemma <ref>). For odd $n$, the same arguments show that $\B_n$ is a constant multiple of $\Delta^{\frac{n-1}{2}}(\partial_s(\rho)|_0) + nl$. Since $\partial_s(\rho)|_0$ is a constant multiple of $|\lo|^2$ (Lemma <ref>), this completes the proof. For $n=2$ and $n=4$, we find 3! \B_2 = - 2 \Delta(H) + nl \quad \mbox{and} \quad 5! \B_4 = - 3 \Delta^2 (H) + nl, respectively. The first decomposition fits with (<ref>). For a flat background and odd $n$, the proof of Theorem <ref> shows that one contribution to $\B_n$ is a constant multiple of $\Delta^{\frac{n-1}{2}}(|\lo|^2)$. But $\B_n$ has further contributions of the same differential order, which are quadratic in $L$. For instance, (<ref>) shows that in addition to $\Delta(|\lo|^2)$, $\B_3$ contains the contributions $|dH|^2$ and $(\lo,\Hess(H))$ of differential order $2$. We finish this section with a representation theoretical argument proving the vanishing of the obstruction $\B_n$ for the equatorial subsphere $S^n \hookrightarrow S^{n+1}$. In the following, we use the notation of Section <ref>. First, assume that $M^n \hookrightarrow \R^{n+1}$ (with the flat metric $g_0$). Let $\B_n^M(g_0)$ be the singular Yamabe obstruction of $M$. Let $\gamma \in SO(1,n+2)$ be a conformal diffeomorphism of $g_0$, i.e., $\gamma_*(g_0) = e^{2 \Phi_\gamma} g_0$. Then e^{(n+1) \iota^* \Phi_\gamma} \B_n^M (\gamma_*(g_0)) = \B_n^M(g_0) by Lemma <ref>. This relation is equivalent to \begin{equation}\label{B-diffeo} e^{(n+1) \iota^* \Phi_\gamma} \gamma_* (\B_n^{\gamma(M)}(g_0)) = \B_n^M(g_0), \; \gamma \in SO(1,n+2). \end{equation} In particular, all $\gamma \in SO(1,n+1) \hookrightarrow SO(1,n+2)$ leave invariant the hypersurface $M^n = \R^n \hookrightarrow \R^{n+1}$, and it holds \iota^* \left(\frac{\gamma_*(r)}{r}\right)^{n+1} \gamma_* (\B_n^M (g_0) ) = \B_n^M(g_0), \pi_{n+1}^0(\gamma) (\B_n^M(g_0)) = \B_n^M(g_0). But since the identical representation is not a subrepresentation of $\pi_{n+1}^0$, it follows that $\B_n^M(g_0)=0$. A similar argument proves the vanishing of $\B_n$ for the equatorial subsphere $S^n \hookrightarrow S^{n+1}$. By the analog of (<ref>) for hypersurfaces of $S^{n+1}$, the obstruction of any $\gamma(S^n)$ vanishes, too. § RESIDUE FAMILIES In the present section, we associate residue families to any pair $(g,\sigma)$ which satisfies the condition $\SCY$ (see Section <ref>). Residue families are defined in terms of the residues of one-parameter families \lambda \mapsto \langle M_u(\lambda),\psi \rangle = \int_X \sigma^\lambda u \psi dvol_g, \; \Re(\lambda) \gg 0 of distributions on $X$, where $u$ are eigenfunctions of the Laplacian $\Delta_{\sigma^{-2}g}$ of the singular metric $\sigma^{-2}g$. These residues express the obstruction extending $u$ as a distribution up to the boundary of $X$. The restriction of $\SC(g,\sigma)$ to the boundary $M$ equals $|\grad_g(\sigma)|_g^2 = |d\sigma|^2_g$. Therefore, the condition $\SCY$ implies that $|d\sigma|_g^2=1$ on $M$. That property is equivalent to the property that the sectional curvatures of $\sigma^{-2}g$ tend to $-1$ at the boundary $M$, i.e., the metric $\sigma^{-2}g$ is asymptotically hyperbolic. Next, we recall some basic results in the spectral theory of the Laplacian of asymptotically hyperbolic metrics. For more details, we refer to <cit.>. The spectrum of $-\Delta_{\sigma^{-2}g}$ is the union of a finite pure point spectrum $\sigma_{pp} \subset (0,(n/2)^2)$ and an absolutely continuous spectrum $\sigma_{ac} = [(n/2)^2,\infty)$ of infinite multiplicity. The generalized eigenfunctions with smooth functions on $M$ as boundary values are described by a Poisson operator. This operator is a far-reaching generalization of the well-known Poisson transform of Helgason <cit.> which relates generalized eigenfunctions of the commutative algebra of invariant differential operators on a symmetric space of the non-compact type to hyperfunctions on a naturally associated boundary. It is defined by an integral transform. In the present situation, the family $\Po(\lambda)$ of Poisson operators is meromorphic for $\Re(\lambda)<n/2$, $\lambda \ne n/2$ with poles in $\lambda$ iff $\lambda(n-\lambda) \in \sigma_{pp}$ such that $\Po(\lambda)$ Poisson operator (\Delta_{\sigma^{-2}g} + \lambda(n-\lambda)) \Po(\lambda) (f) = 0 for any $f \in C^\infty(M)$. In contrast to Helgason's definition of a Poisson operator by an integral transform, it is defined in terms of the resolvent of the Laplacian. In both theories, the argument $f$ is seen in the leading terms of the asymptotic expansion of the eigenfunction $\Po(\lambda)(f)$. To describe the asymptotic expansion of eigenfunctions in the range of the Poisson operator, we choose coordinates on $X$ near the boundary. Indeed, there is a unique defining function $\theta$ and a diffeomorphism $\tau$ mapping $[0,\varepsilon] \times M$ with coordinates $(t,x)$ to a neighborhood of $M$ in $X$ so that $\tau$ $\theta$ \tau^* (\sigma^{-2} g) = t^{-2} (dt^2 + h_t) \quad \mbox{and} \quad \; \iota^*(\theta^2 \sigma^{-2}g) = h_0 \stackrel{!}{=} h, \; \tau^*(\theta) = t. This is the normal form of an asymptotically hyperbolic metric with prescribed conformal infinity as used in <cit.>. The function $\theta$ is a solution of the eikonal equation $|d\theta|_{\theta^2 \sigma^{-2} g}=1$ near $M$ and the gradient flow of $\theta$ with respect to the metric $\theta^2 \sigma^{-2} g$ defines the diffeomorphism $\tau$. Note that $\iota^*(\theta^2 \sigma^{-2}) = 1$. In these terms, the eigenfunctions $\Po(\lambda)(f)$ have the following properties. (i) $ \tau^* \Po(\lambda) (f) = t^{\lambda} f + t^{n-\lambda} g + O(t^{n/2+1})$ for some $g \in C^\infty(M)$ if $\Re(\lambda)=n/2$, $\lambda \ne n/2$.[Of course, the function $g$ should not be confused with the metric $g$.] (ii) $\tau^* \Po(\lambda) (f) = t^\lambda F + t^{n-\lambda} G$ with smooth $F$ and $G$ on $[0,\varepsilon) \times M$ so that $\iota^*(F)=f$ if $\Re(\lambda) \le n/2$ with $\lambda \not\in \{n/2-N/2 \,|\, N \in \N_0\}$ and $\lambda(n-\lambda) \notin \sigma_{pp}$. The function $f$ is called the boundary value of $u = \Po(\lambda)(f)$. The function $F$ in (ii) depends on $\lambda$ and has poles in $\lambda \in \{n/2-N/2 \,|\, N \in \N\}$. But these poles cancel against poles of the second term $G= G(\lambda)$ in the decomposition of $\Po(\lambda)$. If $\lambda$ is as in (ii), we define $\SC(\lambda)$ scattering operator \begin{equation}\label{scatt-def} \Sc(\lambda)(f) = \iota^*(G). \end{equation} The operator $\Sc(\lambda)$ is called the scattering operator of the asymptotically hyperbolic metric $\sigma^{-2}g$.[The substitution $\lambda \mapsto n-\lambda$ maps the operator $\Sc(\lambda)$ to the scattering operator in <cit.>.] It is a family of pseudo-differential operators with principal symbol being a constant multiple of $|\xi|^{n-2\lambda}$. It is meromorphic in $\Re(\lambda) < n/2$ with poles in the set $\{\frac{n-N}{2} \,|\, N \in \N\}$ and if $\lambda(n-\lambda) \in \sigma_{pp}$. The poles in $\lambda = \frac{n-N}{2}$ are sometimes referred to as its trivial poles. Its nontrivial poles in $\Re(\lambda) > n/2$ will not be of interest here. Now let $u = \Po(\mu)(f) $ be a solution of -\Delta_{\sigma^{-2}g} u = \mu (n-\mu)u, \quad \Re(\mu) = n/2, \; \mu \ne n/2 with boundary value $f \in C^\infty(M)$. Instead of the asymptotic expansions of $u$ as above, we will consider asymptotic expansion in terms of powers of $\sigma$. The following formal arguments describe the expansion of $u$ using the Laplace-Robin operators $L(g,\sigma;\mu)$.[The arguments are only formal since the products $\sigma^j f_j$ are not functions on $X$ without a specification of coordinates.] In the Poincaré-Einstein case, the following algorithm is contained in the proof of <cit.>. We start with $f_0 = f$ and define $f_N \in C^\infty(M)$ recursively by N(n-2\mu-N) f_N = \iota^*\left(\sigma^{-N+1} L(-\mu) \left(\sum_{j=0}^{N-1} \sigma^j f_j \right)\right). But the definition of $L(\lambda)$ implies L(\lambda) (\sigma^j f_j) = j(n+2\lambda-j) \sigma^{j-1} f_j + O(\sigma^j), \; \lambda \in \C. L(-\mu) \left(\sum_{j=0}^N \sigma^j f_j\right) = O(\sigma^N) and the conjugation formula yields -(\Delta_{\sigma^{-2}g} + \mu(n-\mu)) \left(\sigma^\mu \sum_{j=0}^{N} \sigma^j f_j \right) = \sigma^{\mu+1} L(-\mu) \left(\sum_{j=0}^{N} \sigma^j f_j \right) = \sigma^{\mu} The coefficients $f_j$ are given by differential operators $f \mapsto \T_j(\mu)(f)$. The construction shows that the operator $\T_N(\mu)$ has simple poles in the set \left \{ \frac{n-j}{2} \; |\, j \le N \right \}. These are the poles that appeared above in (ii). Note that, using $\iota^*(\rho)=-H$, it easily follows that $\T_1(\mu)=-\mu H$ (see also Lemma <ref>). We also observe that $L(\mu)(1) \in C^\infty(X)$ is a multiple of $\mu$. By an easy induction, this implies that \begin{equation}\label{T-1} \T_N(\mu)(1) = \frac{\mu}{N! (n\!-\!2\mu\!-\!1) \cdots (n\!-\!2\mu\!-\!N)} \QC_{N}(\mu) \end{equation} for some polynomial $\QC_{N}(\mu) \in C^\infty(M)$. In particular, the function $\T_n(\mu)(1)$ is regular at $\mu=0$. Now, in order to justify the above arguments, we use adapted coordinates. Let $u$ be as above. Then $\tau^*(u)$ has the form $t^\mu F + t^{n-\mu} G$ with $\iota^*(F)=f$ and $\iota^*(G) = \SC(\mu)(f)$. Let $\zeta \st \tau^{-1} \circ \eta$. Then $\zeta$ \zeta^*(t) = \eta^* \tau_* (t) = \eta^* \left(\sigma \frac{\tau_*(t)}{\sigma}\right) = s \eta^* \left(\frac{\tau_*(t)}{\sigma}\right) by $\eta^*(\sigma)=s$. It follows that the pull-back by $\zeta$ of $\tau^*(u)$ equals \eta^*(u) = s^\mu \Omega^\mu \zeta^*(F) + s^{n-\mu} \Omega^{n-\mu} \zeta^*(G) with $\Omega \st \tau_*(t)/\sigma$. But $\iota^* \zeta^* (F) = \iota^*(F)$, $\iota^* \zeta^*(G) = \iota^*(G)$ \begin{equation}\label{Omega-rest} \iota^* (\Omega) = 1 \end{equation} imply that the leading terms in the expansion of $\eta^*(u)$ are $\iota^*(F) = f$ and $\iota^*(G) = \Sc(\mu)(f)$. In order to prove the restriction property (<ref>), we recall that \tau_* (t^{-2} (dt^2+h_t)) = \sigma^{-2} g. \Omega^2 g = dt^2 + h_t. In particular, $\iota^* (\Omega^2) h = h_0$. But in the construction of $(\tau,\theta)$ we required that $h_0=h$. This proves (<ref>). Thus, the asymptotic expansion of the eigenfunction $\eta^*(u)$ of $\Delta_{\eta^*(\sigma^{-2}g)}$ takes the form $\T_j(\lambda)$ solution operators \begin{equation}\label{asymp} \sum_{j \ge 0} s^{\mu+j} \T_j(\mu)(f)(x) + \sum_{j \ge 0} s^{n-\mu+j} \T_j(n\!-\!\mu) \Sc(\mu)(f)(x), \quad s \to 0, \end{equation} where $\T_j(\mu)$ are families of differential operators on $M$; we shall refer to these operators as solution operators. One easily find that the order of $\T_j(\mu)$ is $\le 2 [\frac{j}{2}]$. The above formal arguments show that the families $\T_N(\mu)$ are rational in $\mu$ with simple poles in the set \left\{ \frac{n-j}{2} \; |\, j \le N \right\}. Of course, the coefficients $\T_j(\mu)$ can be determined recursively in terms of the Laplace-Robin operator in adapted coordinates. In the following, it will often suffice to work with a finite version of the expansion The solution operators $\T_j(\lambda): C^\infty(M) \to C^\infty(M)$ describe formal asymptotic expansions of eigenfunctions (with smooth boundary value) of the Laplacian of asymptotically hyperbolic metrics $\sigma^{-2}g$. Another type of asymptotic expansions of eigenfunctions appears in <cit.>. In an even more general setting, these are expansions, say in powers of a defining function, the coefficients of which are functions on the space $X$ but not on its boundary $M$. Comparing both types of expansions would require additional expansions of the coefficients. Later, we shall use the fact that the scattering operator $\Sc(\lambda)$ for $\lambda \in \R$ is formally selfadjoint with respect to the scalar product on $C^\infty(M)$ defined by $h$. For the convenience of the reader, we include a proof that directly derives this property from the expansion (<ref>) (without invoking the definition of $\Sc(\lambda)$ in terms of expansions in power series of $t$) (compare with <cit.>). $\Sc(\lambda)^* = \Sc(\lambda)$ for $\lambda \in \R$, $\lambda < n/2$ such that $\lambda \notin \left\{ \frac{n-N}{2} \,|\, N \in \N \right\}$ and $\lambda(n-\lambda) \notin \sigma_{pp}$. Note that the assumptions guarantee that the ladders $\lambda+\N$ and $n-\lambda +\N$ are disjoint. We recall Green's formula \begin{equation}\label{Green} \int_X (du,dv)_g dvol_g + \int_X u \Delta_g(v) dvol_g = \int_{\partial X} u \star_g dv = \int_{\partial X} u i_{N(v) N} (dvol_g) \end{equation} on a compact Riemannian manifold $(X,g)$ with boundary $\partial X$. Here $N$ denotes a unit normal field. Now let $\R \ni \lambda < n/2$ be as above. Let $u_1$ and $u_2$ be real solutions of -\Delta_{\eta^*(\sigma^{-2}g)} u = \lambda(n-\lambda) u on $(0,\varepsilon) \times M$ of the form $u = s^{\lambda} F + s^{n-\lambda} G$ with smooth $F,G$. These are defined by the Poisson transforms of smooth boundary functions $f_1$ and $f_2$. Let k \st \eta^*(\sigma^{-2}g) = s^{-2}( a^{-1} ds^2 + h_s). Then $dvol_k = s^{-n-1} a^{-1/2} ds dvol_{h_s}$ and the restriction of $\nu = s \sqrt{a} \partial_s$ to the boundary $s=\varepsilon$ defines a unit normal field. By (<ref>), we find \begin{align}\label{Green-a} & \int_{s > \varepsilon} ((du_1,du_2)_k - \lambda(n-\lambda) u_1 u_2) dvol_k \notag \\ & = \varepsilon^{-n} \int_{s=\varepsilon} u_1 \partial_{\nu}(u_2) dvol_{h_\varepsilon} = \varepsilon^{-n+1} \int_{s=\varepsilon} u_1 \partial_s(u_2) \sqrt{a} dvol_{h_\varepsilon}. \end{align} The finite part of the expansion of the last integral in $\varepsilon$ is the coefficient of $\varepsilon^{n-1}$ in the expansion of \int_{s=\varepsilon} u_1 \partial_s(u_2) \sqrt{a} dvol_{h_\varepsilon}. By plugging in the expansions of $u_1$, $u_2$, $\sqrt{a}$ and $dvol_{h_\varepsilon}$, we obtain the expression \int_M \iota^* (\lambda F_2 G_1 + (n-\lambda) F_1 G_2) dvol_h for this coefficient. Since the left-hand side of (<ref>) is symmetric in $u_1$ and $u_2$, the latter result equals \int_M \iota^* (\lambda F_1 G_2+ (n-\lambda) F_2 G_1) dvol_h. It follows that \int_M \iota^* (F_1 G_2) dvol_h = \int_M \iota^* (F_2 G_1) dvol_h. In terms of the expansion (<ref>), this means that \int_M f_1 \SC(\lambda) (f_2) dvol_h = \int_M f_2 \SC(\lambda) (f_1) dvol_h, i.e., $\Sc(\lambda)^* = \Sc(\lambda)$. Now we are ready to define residue families. $\D_N^{res}(g,\sigma;\lambda)$ residue family Let $\N \ni N \le n$ and assume that the condition $\SCY$ is satisfied. We consider an eigenfunction $u$ of $\Delta_{\sigma^{-2}} g$ with boundary value $f \in C^\infty(M)$ satisfying \begin{equation}\label{eigen} \Delta_{\sigma^{-2}g} u + \mu (n-\mu) u = 0 \quad \mbox{for $\Re(\mu) = n/2$, $\mu \ne n/2$}. \end{equation} Such eigenfunctions have asymptotic expansions (as above) near the boundary. We consider the integral \begin{equation}\label{Mu-L} \left\langle M_u(\lambda),\psi \right\rangle = \int_X \sigma^\lambda u \psi dvol_g \end{equation} with $\psi \in C_c^\infty(X)$ and $\lambda \in \C$. The function $\lambda \mapsto \left\langle M_u(\lambda),\psi \right\rangle$ is holomorphic if $\Re(\lambda) \gg 0$ and we regard $M_u(\lambda)$ as a holomorphic family of distributions on $X$. If $\supp(\psi) \cap M = \emptyset$, then $M_u(\lambda)$ admits a holomorphic continuation to $\C$. $M_u(\lambda)$ generalizes the meromorphic family of distributions $M(\lambda) = M_1(\lambda)$ discussed in Section <ref>. Likewise as $M(\lambda)$, the family $M_u(\lambda)$ admits a meromorphic continuation with simple poles in $\{-\mu - N - 1 \}$. The proof of this fact is similar as for $u=1$ and follows by expanding the integrand near the boundary. In addition to these poles, $M_u(\lambda)$ has simple poles in the set $\{\mu + n - N - 1\}$. However, this second ladder of poles will be ignored in the following. The details are given in the proof of Theorem <ref>. This proof also shows that the residues have the form \begin{equation}\label{delta-def} \Res_{\lambda=-\mu-1-N} \left( \int_X \sigma^\lambda u \psi dvol_g \right) = \int_M f \delta_N(g,\sigma;\mu)(\eta^*(\psi)) dvol_h \end{equation} with some meromorphic families $\delta_N(g,\sigma;\mu)$ of differential operators $C^\infty([0,\varepsilon)\times M) \to C^\infty(M)$ of order $\le N$. The residues of $M_u(\lambda)$ are distributions on $X$ with support on the boundary $M$ of $X$. They may be regarded as obstructions to extending $M_u(\lambda)$ as a distribution to $X$. Let $\N \ni N \le n$ and assume that the condition $\SCY$ is satisfied. Then the one-parameter family \begin{equation}\label{D-res-def} \D_N^{res}(g,\sigma;\lambda) = N!(2\lambda\!+\!n\!-\!2N\!+\!1)_N \delta_N(g,\sigma;\lambda\!+\!n\!-\!N), \; \lambda \in \C \end{equation} is called the residue family of order $N$. The family $\D_n^{res}(\lambda)$ will be called the critical residue family. residue families are not conformally covariant: we need η^* for that! we distinguish two versions: one without and one with η^* Some comments are in order. The families $\delta_N(\mu)$ are defined for $\Re(\mu) = \frac{n}{2}$ (with $\mu \ne \frac{n}{2}$). Hence $\D_N^{res}(\lambda)$ is defined for $\Re (\lambda) = -\frac{n}{2}+N$ (with $\lambda \ne -\frac{n}{2}+N$). But the normalizing coefficient in (<ref>) is chosen so that $\D_N^{res}(g,\sigma;\lambda)$ actually extends to a polynomial family of order $N$ and degree $2N$ in $\lambda \in \C$. This fact will follow from Theorem <ref>. The proof of the latter result actually contains a second proof of the existence of the meromorphic continuation of $\langle M_u(\lambda),\psi \rangle$. The method is a version of a Bernstein-Sato-type argument. It provides explicit knowledge of the position of poles of $M_u(\lambda)$ and formulas for the residues. However, the following result gives a more direct description of the operator $\delta_N(g,\sigma;\lambda)$ in terms of solution operators. The equivalence of both descriptions of residues will have interesting consequences. Let $\N \ni N \le n$. Then \begin{align}\label{D-res-sol} \D_N^{res}(g,\sigma;\lambda) & = N! (2\lambda\!+\!n\!-\!2N\!+\!1)_N \\ & \times \sum_{j=0}^{N} \frac{1}{j!} \left[ \T_{N-j}^*(g,\sigma;\lambda\!+\!n\!-\!N) v_0 + \cdots + \T_0^*(g,\sigma;\lambda\!+\!n\!-\!N) v_{N-j} \right] \iota^* \partial_s^j. \notag \end{align} The relation (<ref>) implies \begin{equation}\label{M-adapted} \langle M_u(\lambda),\psi \rangle = \int_X \sigma^\lambda u \psi dvol_g = \int_{(0,\varepsilon) \times M} s^\lambda \eta^*(u) \eta^*(\psi) dvol_{\eta^*(g)}, \; \Re(\lambda) \gg 0. \end{equation} Here $\eta^*(u) $ is an eigenfunction of the Laplacian of the metric $s^{-2} \eta^*(g)$. In order to simplify the notation, we write the latter integral as \int_{[0,\varepsilon) \times M} s^\lambda u \psi dvol_{\eta^*(g)} = \int_0^\infty \int_M s^\lambda u \psi v ds dvol_h with an appropriate eigenfunction $u$ and test functions $\psi$ on the space $[0,\varepsilon) \times M$. Now we expand $v$ according to (<ref>) and $u$ according to (<ref>). The classical formula <cit.> \begin{equation}\label{Gelfand} \Res_{\lambda=-N-1} \left( \int_0^\infty s^\lambda \psi(s) ds \right) = \frac{\psi^{(N)}(0)}{N!}, \; N \in \N_0 \end{equation} for test functions $\psi \in C_c^\infty(\overline{\R_+})$ shows that \begin{align*} & \Res_{\lambda=-\mu-1-N} \left( \int_X \sigma^\lambda u \psi dvol_g \right) \\ & = \Res_{\lambda=-\mu-1-N} \left( \int_0^\infty \sum_{a=0}^N \sum_{j+k=a} \int_M s^{\lambda+\mu+a} \T_k(\mu)(f) v_j \psi dvol_h \right) \\ & = \sum_{a=0}^N \frac{1}{(N-a)!} \sum_{j+k=a} \int_M \T_k(\mu)(f) v_j \iota^* \partial_s^{N-a}(\psi) dvol_h. \end{align*} Since $f$ is arbitrary, taking adjoints proves the assertion. Definition <ref> generalizes the notion of residue families $D_N^{res}(h;\lambda)$ introduced in <cit.>. In that case, $g = r^2 g_+$ is the conformal compactification of a Poincaré-Einstein metric $g_+$. However, the definitions in <cit.> use a different normalizing coefficient. That choice is motivated by the fact that in these references, the expansion of $g$ involves only even powers of $r$. More precisely, if $g_+$ is in normal form relative to $h$, i.e., $\iota^*(r^2g_+)=h$, then it holds \begin{align*} \D^{res}_{2N}(g,r;\lambda) & = (-2N)_N \left(\lambda\!+\!\frac{n}{2}\!-\!2N\!+\!\frac{1}{2}\right)_N D^{res}_{2N}(h;\lambda), \\ \D^{res}_{2N+1}(g,r;\lambda) & = -2(-2N\!-\!1)_{N+1}\left(\lambda\!+\!\frac{n}{2}\!-\!2N\!-\!\frac{1}{2}\right)_{N+1} D^{res}_{2N+1}(h;\lambda). \end{align*} If $\lambda$ is a zero of the prefactors in these factorization identities, then the family $\D_N^{res}(g,r;\lambda)$ vanishes and therefore hides the non-trivial operator $D_N^{res}(h;\lambda)$. In particular, if $2N+1=n$, then \D_n^{res} (g,r;0) = 0. However, for general $g$ and $\sigma$ satisfying $\SCY$, the critical value $\D_n^{res}(g,\sigma;0)$ for odd $n$ need not vanish. For the case $n=3$, we refer to Section <ref>. The degrees of $D_{2N}^{res}(h;\lambda)$ and $D_{2N+1}^{res}(h;\lambda)$ both equal $N$. Hence the above relations show that the respective degrees of $\D_{2N}^{res}(g,r;\lambda)$ and $\D_{2N+1}^{res}(g,r,\lambda)$ are $2N$ and $2N+1$. Since in the generic case $\D_N^{res}(\lambda)$ has degree $2N$, it follows that in the Poincaré-Einstein case the degrees fall on half. This drop in degree reflects the vanishing of curvature invariants in the Poincaré-Einstein case. For instance, (<ref>) and Lemma <ref> show that $\D_1^{res}(g,r;\lambda)$ and $\D_2^{res}(g,r;\lambda)$ have respective degrees $1$ and $2$. In these cases, the vanishing of $H$, $\lo$, and $\Rho_{00}$ are responsible for the drop in degree. The following result implies that residue families of order $N \le n$ are completely determined by the metric $g$ and the embedding $M \hookrightarrow X$. Let $\N \ni N \le n$. Then $\D_N^{res}(g,\sigma;\lambda)$ is determined only by the coefficients $\sigma_{(j)}$ for $j \le N+1$ in the expansion of $\sigma$ in geodesic normal coordinates. The claim follows by evaluating the residue definition of residue families in terms of geodesic normal coordinates. We use the diffeomorphism $\kappa$ to write \begin{equation}\label{M-geodesic} \langle M_u(\lambda),\psi \rangle = \int_X \sigma^\lambda u \psi dvol_g = \int_{[0,\varepsilon)} \int_M \left( \frac{\kappa^*(\sigma)}{r} \right)^\lambda r^\lambda \kappa^*(u) \kappa^*(\psi) dr dvol_{h_r} \end{equation} for $\Re(\lambda) \gg 0$ and test functions $\psi$ with sufficiently small support. The eigenfunction $\kappa^*(u)$ of $\Delta_{\kappa^*(\sigma)^{-2} g}$ has an asymptotic expansion in $r$ of the form \sum_{j \ge 0} r^{\mu+j} \T_j(\mu)(f) + \cdots, where the dots indicate an asymptotic expansion with exponents $n-\mu+j$. In that expansion, the operators $\T_j(\mu)$ are determined by recursive relations. An induction argument using the formula \Delta_{\kappa^*(\sigma)^{-2} g} = (\kappa^*(\sigma))^2 \Delta_{dr^2 + h_r} - (n-1) \kappa^*(\sigma) \nabla_{\grad(\kappa^*(\sigma))} shows that $\T_N(\mu)$ is determined only by the coefficients of $r^j$ for $j \le N+1$ in the expansion of $\kappa^*(\sigma)$, i.e., by $\sigma_{(j)}$ for $j \le N+1$. Now the residue of the left-hand side of (<ref>) at $\lambda=-\mu-1-N$ is determined by the coefficient of $r^{\lambda+\mu+N}$ in the expansion of the integrand. That coefficient involves the operators $\T_j(\mu)$ with $j \le N$ and the coefficients of $r^{j}$ for $j \le N$ in the expansion of $\kappa^*(\sigma)/r$. The latter are determined by $\sigma_{(j)}$ for $j \le N+1$. All other ingredients of the integrand do not depend on $\sigma$. The proof is complete. Since the coefficients $\sigma_{(j)}$ for $j \le n+1$ are determined by the metric $g$ and the embedding $\iota$, the residue families $\D_N^{res}(g,\sigma;\lambda)$ for $N \le n$ are completely determined by the metric $g$ and the embedding $\iota$, and it is justified to use the simplified notation to $\D_N^{res}(g;\lambda)$. The definition of residue families can be extended to a wider setting. Let $\sigma \in C^\infty(X)$ be a boundary defining function so that $|d\sigma|_g^2=1$ on the boundary $M$. Then the singular metric $\sigma^{-2}g$ is asymptotically hyperbolic. This implies the existence of a Poisson operator and the existence of an eigenfunction $u$ of $\Delta_{\sigma^{-2}g}$ with eigenvalue $-\mu(n-\mu)$ and arbitrary given boundary value $f\in C^\infty(M)$. The asymptotic expansion of $u$ in terms of adapted coordinates can be stated as an asymptotic expansion in powers of $\sigma$. For $N \in \N$, we define an operator \delta_N(g,\sigma;\mu): C^\infty(X) \to C^\infty(M) \Res_{\lambda=-\mu-1-N} \left(\int_X \sigma^\lambda u \psi dvol_g \right) = \int_M f \delta_N(g,\sigma;\mu) (\psi) dvol_{\iota^*(g)} and let \D_N^{res}(g,\sigma;\lambda) \st N! (2\lambda\!+\!n\!-\!2N\!+\!1)_N \delta_N(g,\sigma;\lambda\!+\!n\!-\!N). These general residue families are conformally covariant in the following sense. The residue family $\D_N^{res}(g,\sigma;\lambda)$ is conformally covariant in the sense that \D_N^{res}(\hat{g},\hat{\sigma};\lambda) \circ e^{\lambda \varphi} = e^{(\lambda-N) \iota^* (\varphi)} \circ \D_N^{res}(g,\sigma;\lambda) for all conformal changes $(\hat{g},\hat{\sigma}) = (e^{2\varphi}g,e^{\varphi}\sigma)$, $\varphi \in C^\infty(X)$. Let $u \in \ker(\Delta_{\sigma^{-2}g}+\mu(n-\mu))$ be an eigenfunction with leading term $f \in C^\infty(M)$ in its expansion into powers of $\sigma$. We calculate the \Res_{\lambda=-\mu-1-N} \left(\int_X \sigma^\lambda u \psi dvol_g \right) in two ways. On the one hand, it equals \int_M f \delta_N(g,\sigma;\mu) (\psi) dvol_{\iota^*(g)}. On the other hand, for $\Re(\lambda) \gg 0$, we have \int_X \sigma^\lambda u \psi dvol_g = \int_X \hat{\sigma}^\lambda u (e^{(-\lambda-n-1)\varphi} \psi) dvol_{\hat{g}} and the leading term in the $\hat{\sigma}$-expansion of $u$ equals $e^{-\mu \iota^*(\varphi)}f$. Hence the residue equals \int_M e^{-\mu \iota^*(\varphi)} f \delta_N(\hat{g},\hat{\sigma};\mu)(e^{(\mu-n+N)\varphi} \psi) But $dvol_{\iota^*(\hat{g})} = e^{n \iota^* \varphi} dvol_{\iota^*(g)}$. Since $f \in C^\infty(M)$ is arbitrary, we find \delta_N(g,\sigma;\mu) = e^{(-\mu+n)\iota^*(\varphi)} \circ \delta_N(\hat{g},\hat{\sigma};\mu) \circ e^{(\mu+N-n)\varphi}. This result implies the assertion. As a special case, it follows that the compositions of residue families of order $N \le n$ in the sense of Definition <ref> with $\eta^*$ (and $\sigma$ being a solution of the Yamabe problem) are conformally covariant. Whereas the general residue families in Theorem <ref> depend on $g$ and $\sigma$, Proposition <ref> shows that the special cases in Definition <ref> only depend on $g$ (and the embedding $\iota$). Although Definition <ref> breaks the conformal covariance (by omitting $\eta^*$), for those values of $\lambda$ for which residue families are tangential, the resulting operators on $M$ are still conformally covariant. This observation will play a central role in Section <ref>. In the following sections, it will always be clear from the context which notion of residue families is being used. § RESIDUE FAMILIES AS COMPOSITIONS OF $L$-OPERATORS In the present section, we show that the composition of residue families as defined in Definition <ref> with $\eta^*$ (defining adapted coordinates) can be identified with compositions of Laplace-Robin operators and the restriction operator $\iota^*$. We recall the notation $L_N(g,\sigma;\lambda) \st L(g,\sigma;\lambda\!-\!N\!+\!1) \circ \cdots \circ L(g,\sigma;\lambda)$ and set $L_0 = \id$ (see (<ref>)). Let $\N \ni N \le n$ and assume that $\sigma$ satisfies the condition $\SCY$. Then \D_N^{res}(g,\sigma;\lambda) \circ \eta^* = \iota^* L_N(g,\sigma;\lambda). It suffices to prove that \begin{equation}\label{red-main} \delta_N(g,\sigma;\lambda) \circ \eta^* = \frac{1}{N! (2\lambda\!-\!n\!+\!1)_N} \iota^* L(g,\sigma;\lambda\!-\!n\!+\!1) \circ \cdots \circ L(g,\sigma;\lambda\!-\!n\!+\!N). \end{equation} Let $u$ be an eigenfunction with boundary value $f \in C^\infty(M)$ satisfying (<ref>) with $\Re(\mu) = n/2$, $\mu \ne n/2$. In the following, it will be convenient to use the notation \begin{equation*} A((\lambda)_N) \st A(\lambda) \circ A(\lambda+1) \circ \cdots \circ A(\lambda+N-1) \end{equation*} for any $\lambda$-dependent family $A(\lambda)$ of operators. Then $L_N(\lambda) = L((\lambda-N+1)_N)$. On the one hand, (<ref>) states that \begin{equation}\label{eq:h1} \Res_{\lambda=-\mu-1-N} \left(\int_{X} \sigma^{\lambda} u \psi dvol_g \right) = \int_M f \delta_N(g,\sigma;\mu)(\eta^*(\psi)) dvol_h. \end{equation} Now we first assume that $\sigma$ satisfies the stronger assumption $\SC(g,\sigma)=1$. We apply Corollary <ref> to calculate \begin{align} - L(g,\sigma;\lambda+1) (\sigma^{\lambda+1} u) & = \sigma^\lambda \left(\Delta_{\sigma^{-2}g} - (\lambda\!+\!1)(n\!+\!\lambda\!+\!1) \id \right) u \\ & = \sigma^\lambda (-\mu(n\!-\!\mu) - (\lambda\!+\!1)(n\!+\!\lambda\!+\!1)) u \\ & = -(\lambda\!+\!\mu\!+\!1)(\lambda\!-\!\mu\!+\!n\!+\!1) \sigma^\lambda u. \end{align} We regard this relation as a Bernstein-Sato-type functional equation. Hence for $\Re(\lambda) \notin -\frac{n}{2}-\N$ we obtain \begin{align*} \sigma^\lambda u & = \frac{L(g,\sigma;\lambda\!+\!1)(\sigma^{\lambda+1} u)} {(\lambda\!+\!\mu\!+\!1)(\lambda\!-\!\mu\!+\!n\!+\!1)} \\ & = \frac{L(g,\sigma;\lambda\!+\!1)L(g,\sigma;\lambda\!+\!2)(\sigma^{\lambda+2} u)} {(\lambda\!+\!\mu\!+\!1)(\lambda\!+\!\mu\!+\!2)(\lambda\!-\!\mu\!+\!n\!+\!1)(\lambda\!-\!\mu\!+\!n\!+\!2)} \\ & = \dots = \frac{L(g,\sigma;(\lambda\!+\!1)_N) (\sigma^{\lambda+N} u)} {(\lambda\!+\!\mu\!+\!1)_N \end{align*} It follows that the integral \lambda \mapsto \int_X \sigma^\lambda u \psi dvol_g, \; \Re(\lambda) \gg 0 admits a meromorphic continuation to $\C$ with simple poles in the set -\mu-1-\N_0 \cup \mu-n-1-\N_0. More precisely, we get \begin{equation}\label{int-N} \int_X \sigma^\lambda u \psi dvol_g = \frac{1}{(\lambda\!+\!\mu\!+\!1)_N (\lambda\!-\!\mu\!+\!n\!+\!1)_N} \int_X L(g,\sigma;(\lambda\!+\!1)_N)(\sigma^{\lambda+N} u) \psi dvol_g \end{equation} for $\Re(\lambda) > - \frac{n}{2}-1$ and $N \ge 1$. In the following, it will be convenient to choose $\lambda$ so that $\Re(\lambda) > - \frac{n}{2}+1$. Now we note that a function in $C^\infty(X^\circ)$ with an asymptotic expansion of the form $\sum_{j\ge 0} \sigma^{\nu + j} a_j$ with $\Re(\nu) > 2$ and $a_j \in C^\infty(M)$ satisfies the assumptions in Proposition <ref>. Thus, by a repeated application of Proposition <ref>, the right-hand side of (<ref>) equals \frac{ 1}{(\lambda\!+\!\mu\!+\!1)_N (\lambda\!-\!\mu\!+\!n\!+\!1)_N} \int_X \sigma^{\lambda+N} u L(g,\sigma;(-\lambda\!-\!n\!-\!N)_N)(\psi) dvol_g. By the assumptions, the zeros of the product \mu \mapsto (\lambda\!+\!\mu\!+\!1)_N (\lambda\!-\!\mu\!+\!n\!+\!1)_N are simple for $\Re(\lambda) > - \frac{n}{2}+1$. Thus, using the residue formula \begin{equation}\label{Delta0} \Res_{\lambda=-\mu-1} \left(\int_{X} \sigma^\lambda u \psi dvol_g \right) = \int_M f \iota^*(\psi) dvol_h, \end{equation} we find \begin{align}\label{residue-2} & \Res_{\lambda=-\mu-1-N} \left(\int_{X} \sigma^\lambda u \psi dvol_g \right) \notag \\ & = \frac{(-1)^N}{(-N)_N (2\mu\!-\!n\!+\!1)_N} \int_M f \iota^* L(g,\sigma;(\mu\!-\!n\!+\!1)_N)(\psi) dvol_h \notag \\ & = \frac{1}{N!(2\mu\!-\!n\!+\!1)_N} \int_M f \iota^* L_N(g,\sigma;\mu\!-\!n\!+\!N)(\psi) dvol_h \end{align} for $N \ge 1$. Comparing this result with (<ref>), completes the proof of (<ref>) for $\Re(\mu) = n/2$, $\mu \ne n/2$. The assertion then follows by meromorphic continuation. If $\sigma$ satisfies only the assumption $\SCY$, analogous arguments show that the right-hand side of (<ref>) contains an additional integral \int_X u \psi R_{n+1} \sigma^{\lambda+n+1} dvol_g. Since $R_{n+1}$ is smooth up to the boundary and $N \le n$, this integral is regular at $\lambda=-\mu-1-N$, i.e., does not contribute to the residue. The proof is Let $N \in \N$ with $N \le n$ and assume that $\sigma$ satisfies $\SCY$. Then \D_N^{res}(g,\sigma;\lambda) = \D_{N-1}^{res}(g,\sigma;\lambda-1) \circ L(g,\sigma;\lambda). Theorem <ref> identifies the composition of residue families with $\eta^*$ with compositions of $L$-operators if $\sigma$ satisfies $\SCY$. By Theorem <ref>, residue families are linear combinations of compositions of tangential operators and iterated normal derivatives $\iota^* \partial^k_s$. We may use formula (<ref>) to write their composition with $\eta^*$ in terms of iterated gradients $\nabla_\NV^k$. This yields a formula for the composition of residue families with $\eta^*$ in terms of iterated gradients and tangential For closed $M$, Theorem <ref> implies formulas for integrated renormalized volume coefficients in terms of compositions of Laplace-Robin operators. First, we observe that, for $N \le n-1$, Theorem <ref> implies \begin{align*} \int_M \D_N^{res}(g,\sigma;-n\!+\!N)(1) dvol_h & = N! (-n\!+\!1)_N \sum_{j=0}^{N} \int_M \T_{N-j}^*(g,\sigma;0) (v_{j}) dvol_h \\ & = (-1)^N \frac{(n\!-\!1)! N!}{(n\!-\!1\!-\!N)!} \sum_{j=0}^{N} \int_M v_{j} \T_{N-j}(g,\sigma;0)(1) dvol_h \\ & = (-1)^N \frac{(n\!-\!1)! N!}{(n\!-\!1\!-\!N)!} \int_M v_{N} dvol_h. \end{align*} In the last equality, we used the fact that all coefficients except the leading one in the expansion of the harmonic function $u=1$ vanish. Combining this identity with Theorem <ref> we obtain Let $\N \ni N \le n-1$ and assume that $\sigma$ satisfies the condition $\SCY$. Then \begin{equation}\label{HF-volume} \int_{M} v_N dvol_h = (-1)^N \frac{(n\!-\!1\!-\!N)!}{ (n-1)!N!} \int_{M} \iota^* L_{N}(g,\sigma;-n\!+\!N)(1) dvol_h. \end{equation} We shall see later in Theorem <ref> that this reproves the special case $\tau=1$ of <cit.>. The critical case $N = n$ will be discussed in Section <ref>. These identities for integrated renormalized volume coefficients admit a natural interpretation as special cases of an interesting identity for distributions. In order to describe that point of view, we smoothly extend $g$ and $\sigma$ to a sufficiently small neighborhood $\tilde{X}$ of $M$ so that $|\NV| \ne 0$ on $\tilde{X}$ (this is always possible <cit.>). It will be convenient to assume that $\tilde{X} = \eta (I \times M)$ with a sufficiently small interval $I=(-\varepsilon,\varepsilon)$ around $0$. For any $u \in C^\infty(\R)$, the pull-back $\sigma^*(u) \in C^\infty(X)$ defines a current by \left\langle \sigma^*(u),\psi dvol_g \right\rangle = \int_{\tilde{X}} \sigma^*(u) \psi dvol_g, \; \psi \in C_c^\infty(\tilde{X}). By approximating the delta distribution $\delta$ at $0$ by test functions, we obtain a current $\sigma^*(\delta)$. We recall that the pull-back $\sigma^*(u)$ of a distribution $u$ on the real line by $\sigma$ exists since the differential of $\sigma$ is surjective. The pull-back operation itself then is continuous on distributions (and currents) <cit.>. Since $|\NV|=1$ on $M$, it holds \begin{equation}\label{Ho-simple} \left\langle \sigma^*(\delta),\psi dvol_g \right\rangle = \int_M \iota^*(\psi) dvol_h \end{equation} by an extension of <cit.>. We also use the notation $\delta_M$ for the latter distribution and call it the delta distribution of $M$. $\delta_M$ delta distribution of $M$ We define the action of a differential operator $D$ on currents $u$ on $\tilde{X}$ by \langle D (u), \psi dvol_g \rangle = \langle u, D^* (\psi) dvol_g \rangle, \; \psi \in C_c^\infty(\tilde{X}). Here the formal adjoint $D^*$ of $D$ is determined by the relation \begin{equation}\label{pair} \int_{\tilde{X}} D(\varphi) \psi dvol_g = \int_{\tilde{X}} \varphi D^*(\psi) dvol_g, \; \varphi \in C^\infty(\tilde{X}), \psi \in C_c^\infty(\tilde{X}). \end{equation} Assume that $N \le n-1$ and assume that $\sigma$ satisfies $\SCY$. Then \begin{equation}\label{Shift-delta} L(g,\sigma;-N) \circ \cdots \circ L(g,\sigma;-1) (\sigma^*(\delta)) = a_N\mathfrak{X}^N (\sigma^*(\delta)), \end{equation} where $a_N = (n\!-\!1)!/(n\!-\!1\!-\!N)!$ and $\mathfrak{X} = \NV / |\NV|^2$ is defined in Section <ref>. Corollary <ref> implies that $L_N(\lambda)$ acts on $\sigma^*(\delta)$ by \begin{equation}\label{dual-1} \langle L_N(\lambda)(\sigma^*(\delta)),\psi dvol_g \rangle = \langle \sigma^*(\delta), L_N(-n\!-\!\lambda\!+\!N\!-\!1) (\psi) dvol_g \rangle. \end{equation} \begin{equation*} \langle L_N(\lambda)(\sigma^*(\delta)), \psi dvol_g \rangle = \int_M \iota^* L_N(-n\!-\!\lambda\!+\!N\!-\!1) (\psi) dvol_h. \end{equation*} Thus, Theorem <ref> yields \langle L_{N}(\lambda)(\sigma^*(\delta)), \psi dvol_g \rangle = \int_M \D_{N}^{res}(-\lambda\!-\!n+\!N\!-\!1)(\eta^*(\psi)) dvol_h. Note that this formula implies that $\langle L_{N}(\lambda)(\sigma^*(\delta)), \psi dvol_g \rangle$ only depends on the first $N$ terms in the expansion of $\sigma$. Now, by Theorem <ref>, the latter integral equals \begin{align*} & N!(-2\lambda\!-\!n\!-\!1)_N \\ & \times \sum_{j=0}^{N} \frac{1}{(N\!-\!j)!} \int_M [\T_{j}^*(-\lambda\!-\!1) \circ v_0 + \cdots + \T_0^*(-\lambda\!-\!1) \circ v_j ] (\iota^* \partial_s^{N-j}(\eta^*(\psi))) dvol_h. \end{align*} Hence using partial integration, we obtain \begin{align*} & \langle L_{N}(\lambda)(\sigma^*(\delta)), \psi dvol_g \rangle = N!(-2\lambda\!-\!n\!-\!1)_N \\ & \times \sum_{j=0}^{N} \frac{1}{(N\!-\!j)!} \int_M [\T_{j}(-\lambda\!-\!1)(1) + \cdots + \T_0(-\lambda\!-\!1)(1) v_{j}] \iota^* \partial_s^{N-j}(\eta^*(\psi)) dvol_h. \end{align*} Now let $\lambda = -1$. Since $\T_{j}(0)(1) = 0$ for $j \ge 1$ and $\T_0 = \id$, it follows that \begin{align}\label{int-L} \langle L_N(-1)(\sigma^*(\delta)), \psi dvol_g \rangle & = N!(-n\!+\!1)_N \sum_{j=0}^{N} \frac{1}{(N\!-\!j)!} \int_M v_{j} \iota^* \partial_s^{N-j}(\eta^*(\psi)) dvol_h \notag \\ & = (-n\!+\!1)_N \int_M \iota^* \partial_s^{N} (v \eta^*(\psi)) dvol_h. \end{align} On the other hand, partial integration shows \begin{align*} \langle \mathfrak{X}^N(u),\psi dvol_g \rangle & = \int_{\tilde{X}} \mathfrak{X}^N(u) \psi dvol_g \\ & = \int_{I \times M} \eta^* \mathfrak{X}^N(u) \eta^*(\psi) v ds dvol_h \\ & = \int_{I \times M} \partial_s^N \eta^*(u) \eta^*(\psi) v ds dvol_h & \mbox{(by \eqref{intertwine})} \\ & = (-1)^N \int_{I \times M} \eta^*(u) v^{-1} \partial_s ^N(\eta^*(\psi) v) v ds dvol_h \\ & = (-1)^N \int_{\tilde{X}} u \eta_*(v^{-1} \partial_s^N (v \eta^*(\psi)) dvol_g \end{align*} for $u \in C^\infty(\tilde{X})$ and $\psi \in C^\infty_c (\tilde{X})$. Hence $(\mathfrak{X}^N)^* (\psi) = (-1)^N \eta_*(v^{-1} \partial_s^N (v \eta^*(\psi))$ and \begin{equation}\label{fund-dist} \langle \mathfrak{X}^N(\sigma^*(\delta)),\psi dvol_g \rangle = (-1)^N \int_M \iota^* \partial_s^N (v \eta^*(\psi)) dvol_h. \end{equation} The proof is complete. Note that, for $\psi = 1$ near $M$, the arguments in the above proof show that \begin{align*} \int_M \iota^* L_N(-n\!+\!N) (1) dvol_h & = \frac{(n\!-\!1)!}{(n\!-\!1\!-\!N)!} \langle \mathfrak{X}^N (\sigma^*(\delta)), dvol_g \rangle \\ & = (-1)^N \frac{(n\!-\!1)!}{(n\!-\!1\!-\!N)!}\int_M \iota^* \partial_s^N (v) dvol_h \end{align*} for $N \le n-1$. This proves that the identity (<ref>) is a special case of Theorem <ref>. Theorem <ref> results from our attempt to understand the distributional formula in <cit.>. In <cit.>, this distributional identity is the key to prove formulas like (<ref>). Here we follow a reverse logic. The following shift-property of residue families either follows by combining Theorem <ref> with Corollary <ref> or directly from the residue definition of residue families. Let $1 \le N \le n$ and assume that $\sigma$ satisfies the condition $\SCY$. Then \D_N^{res}(g,\sigma;\lambda) \circ s = N (2\lambda\!+\!n\!-\!N) \D_{N-1}^{res}(g,\sigma;\lambda\!-\!1). We use the residue definition of residue families (Definition <ref>). In particular, $u$ is an eigenfunction of the Laplacian of $\sigma^{-2}g$ for the eigenvalue $-\mu(n-\mu)$. Now the calculation \begin{align*} \int_M f \delta_N(\mu) (\eta^*(\sigma \psi)) dvol_h & = \Res_{\lambda=-\mu-1-N} \left( \int_X \sigma^\lambda u (\sigma \psi) dvol_g \right) \\ & = \Res_{\lambda=-\mu-1-N} \left( \int_X \sigma^{\lambda+1} u \psi dvol_g \right) \\ & = \Res_{\lambda=-\mu-N} \left( \int_X \sigma^\lambda u \psi dvol_g \right) \\ & = \int_M f \delta_{N-1}(\mu) (\eta^* (\psi)) dvol_h \end{align*} shows that $\delta_N(\mu) \circ \eta^* \circ \sigma = \delta_{N-1}(\mu) \circ \eta^*$, i.e., $\delta_N(\mu) \circ s = \delta_{N-1}(\mu)$. The claim is a direct consequence. Finally, we note that the relation $\delta_N(\mu) \circ s = \delta_{N-1}(\mu)$ also is an immediate consequence of Theorem <ref>. § EXTRINSIC CONFORMAL LAPLACIANS If $\SC(g,\sigma) \ne 0$ and $N \in \N$, the commutator relation (<ref>) in Corollary <ref> shows that the composition \tilde{L}_N(g,\sigma) \st \tilde{L}_N \left(g,\sigma;\frac{-n\!+\!N}{2}\right) = \tilde{L}\left(g,\sigma;\frac{-n\!-\!N}{2}\!+\!1\right) \circ \cdots \circ \tilde{L}\left(g,\sigma;\frac{-n\!+\!N}{2}\right) is a tangential operator, i.e., it holds $\tilde{L}_N \circ \sigma = \sigma \circ \tilde{L}'_N$ for some operator $\tilde{L}'_N$. Thus, the operator $\tilde{L}_N(g,\sigma)$ on $C^\infty(X)$ induces an operator $\tilde{\PO}_N(g,\sigma)$ on $C^\infty(M)$ according to \begin{equation}\label{conformal-power} \iota^* \tilde{L}_N(g,\sigma) = \tilde{\PO}_N(g,\sigma) \iota^*. \end{equation} This observation is a special case of <cit.>. If additionally $\sigma$ satisfies the condition $\SCY$ and $N \le n$, then the resulting operator on $C^\infty(M)$ will be denoted by $\PO_N(g,\sigma)$. In the latter case, one may also directly apply (<ref>). Now, following Gover and Waldron, we define Let $\N \ni N \le n$. Assume that $\sigma$ satisfies the condition $\SCY$. Then the operators $\PO_N(g,\sigma)$ are called extrinsic conformal Laplacians. The operator $\PO_n(g,\sigma)$ is called the critical extrinsic conformal Laplacian. The notion is justified by the fact that for even $N$ these operators generalize the GJMS-operators $P_{2N}$ which are of the form $\Delta^N + LOT$. More precisely, let $g_+$ be a Poincaré-Einstein metric in normal form relative to $h$. Then \begin{equation}\label{P-GJMS} \PO_{2N}(r^2g_+,r) = (2N\!-\!1)!!^2 P_{2N}(h). \end{equation} In the general case, the operator $\PO_{2N}(g,\sigma)$ is of the form (2N\!-\!1)!!^2 \Delta^N_M + LOT. Here the lower order terms depend on the embedding $M \hookrightarrow X$ (Proposition <ref>). For odd $N$, the leading part of $\PO_N(g)$ is not given by a power of the Laplacian but involves $\lo$ (Proposition <ref>). Note that $\D_1^{res}(\frac{-n+1}{2})=0$ shows that $\PO_1 = 0$. Note also that the L(g,\sigma;0)(1) = 0 \begin{equation}\label{crit-van} \PO_n(g,\sigma)(1) = 0. \end{equation} Let $\N \ni N \le n$. Assume that $\sigma$ satisfies $\SCY$. \begin{equation}\label{D-res-power} \D_N^{res}\left(g,\sigma;\frac{-n\!+\!N}{2}\right) = \PO_N(g,\sigma) \iota^*. \end{equation} More generally, the factorization identities \D_N^{res}\left(g,\sigma;\frac{-n\!-\!k}{2}\!+\!N\right) = \PO_k(g,\sigma) \circ \D_{N-k}^{res}\left(g,\sigma;\frac{-n\!-\!k}{2}\!+\!N\right) for $1\le k \le N$ hold true. Theorem <ref> implies \D_N^{res}\left(\frac{-n\!-\!k}{2}\!+\!N\right) = \iota^* L\left(\frac{-n\!-\!k}{2}\!+\!1\right) \circ \cdots \circ L\left(\frac{-n\!-\!k}{2}\!+\!N\right). We decompose this product as \left( \iota^* L\left(\frac{-n\!-\!k}{2}\!+\!1\right) \circ \cdots \circ L\left(\frac{-n\!+\!k}{2}\right) \right) \circ \left(L\left(\frac{-n\!+\!k}{2}+1\right) \circ \cdots \circ L\left(\frac{-n\!-\!k}{2}\!+\!N\right) \right). By the definition of $\PO_k$ and Theorem <ref>, this composition equals \PO_k \circ \D_{N-k}^{res}\left(\frac{-n\!-\!k}{2}\!+\!N\right). The proof is complete. Note that $\PO_1=0$ implies the vanishing property $\D_N^{res}\left(\frac{-n-1}{2}\!+\!N\right)=0$. The trivial zero of $D_N^{res}(\lambda)$ at $\lambda=-\frac{n+1}{2}\!+\!N$ is actually one of the zeros in the prefactor in the definition (<ref>). It appears here as a trivial zero since the solution operator $\T_1(\lambda)$ is regular (Lemma <ref>). Theorem <ref> extends factorization formulas in the setting of Poincaré-Einstein metrics. In that case, the left factors are GJMS-operators on the boundary. For details, we refer to <cit.>. Finally, we notice an interesting direct formula for the critical extrinsic conformal Laplacian $\PO_n(g,\sigma)$ in terms of the Laplacian of the singular Yamabe metric $\sigma^{-2} g$. By composition with $\iota^*$, Corollary <ref> shows that \PO_n(g,\sigma) \iota^*(f) = \iota^* \left( \sigma^{-n} \prod_{j=0}^{n-1} \left(\Delta_{\sigma^{-2}g} + (n\!-\!j)j \right) \right) (f) for any $f \in C^\infty(X)$. We omit the analogous formulas in the subcritical cases. Next, we provide a spectral theoretical description of the extrinsic conformal Laplacians. Let $N \in \N$ with $2 \le N \le n$. Assume that $\sigma$ satisfies the condition $\SCY$. Then \PO_N = (-1)^{N-1} 2 (N\!-\!1)! N! \Res_{\frac{n-N}{2}}(\T_N^*(\lambda)). On the right-hand side of formula (<ref>) in Theorem <ref>, all solution operators except $\T_N(\lambda)$ are regular at $\lambda=\frac{n-N}{2}$. The result follows by combining that observation with the identity (<ref>) in Theorem <ref>. We use the spectral theoretical interpretation of the operators $\PO_N$ to separate their leading parts. Let $\N \ni N \le n$. Assume that $\sigma$ satisfied $\SCY$. Then \LT( \Res_{\lambda=\frac{n}{2}-N}(\T_{2N}(\lambda)) = - \frac{1}{2^{2N} (N-1)! N!} \Delta^N_h. \LT(\PO_{2N}) = (2N-1)!!^2 \Delta^N_h. The remaining terms are of order $\le N-2$. As a preparation for the proof, we observe that formula (<ref>) and the identity \Delta_{\sigma^{-2}g} = \sigma^2 \Delta_g - (n\!-\!1) \sigma \nabla_{\grad(\sigma)} imply that the Laplacian of the metric $\eta^*(\sigma^{-2}g) = s^{-2} \eta^*(g)$ takes the form \begin{align}\label{L-op} & s^2 a \partial_s^2 + \frac{1}{2} s^2 a \tr (h_s^{-1} h'_s) \partial_s + \frac{1}{2} s^2 a' \partial_s - (n\!-\!1) s a \partial_s - \frac{1}{2} s^2 (d \log a, d \cdot)_{h_s} + s^2 \Delta_{h_s}, \end{align} where $a = \eta^*(|\NV|^2)$. In the special case $|\NV|=1$, the formula (<ref>) simplifies to s^2 \partial_s^2 + \frac{1}{2} s^2 \tr (h_s^{-1} h'_s) \partial_s - (n\!-\!1) s \partial_s + s^2 \Delta_{h_s}. In particular, this reproduces the formula for the Laplacian of a Poincaré-Einstein metric <cit.>. By combining the formula (<ref>) with (<ref>), we can give another proof of the conjugation formula (<ref>). Here the identities (<ref>) and (<ref>) are useful. We omit the details. The solution operators $\T_N(\lambda)$ are determined by the ansatz \sum_{N\ge 0} s^{\lambda+N} \T_N(\lambda)(f), \; \T_0 = \id for an approximate solution $u$ of the equation $-\Delta_{s^{-2} \eta^*(g)} u = \lambda(n\!-\!\lambda) u$ with boundary value $f$. By formula (<ref>), this means that the sum \begin{align}\label{L-A-Sol} & a \sum_{N \ge 0} (\lambda\!+\!N)(\lambda\!+\!N\!-\!1) \T_N(\lambda) s^{\lambda+N} \notag \\ + & \frac{a}{2} \sum_{N \ge 0} (\lambda\!+\!N) \tr (h_s^{-1} h'_s) \T_N(\lambda) s^{\lambda+N+1} \notag \\ + & \frac{a'}{2} \sum_{N \ge 0} (\lambda\!+\!N) \T_N(\lambda) s^{\lambda+N+1} \notag \\ - & (n\!-\!1) a \sum_{N \ge 0} (\lambda\!+\!N) \T_N(\lambda) s^{\lambda+N} \notag \\ - & \frac{1}{2} \sum_{N \ge 0} (d \log a,d \T_N(\lambda))_{h_s} s^{\lambda+N+2} \notag \\ + & \sum_{N \ge 0} \Delta_{h_s} \T_N(\lambda) s^{\lambda+N+2} \end{align} coincides with the sum \begin{equation}\label{L-A-Sol-2} -\lambda(n\!-\!\lambda) \sum_{N\ge 0} \T_N(\lambda) s^{\lambda+N}. \end{equation} In order to compare coefficients of powers of $s$ in (<ref>) and (<ref>), we also insert the expansions of $a$ and of $h_s$. Note that the equality of the coefficients of $s^\lambda$ in (<ref>) and (<ref>) is trivially satisfied using $\iota^*a = 1$. The equality of coefficients of $s^{\lambda+N}$ in (<ref>) and (<ref>) yields a recursive formula for $\T_N(\lambda)$. For the coefficient $\T_1(\lambda)$, we find $\T_1(\lambda) = -H \id$ (Lemma <ref>). An easy induction shows that the order of $\T_{2N}(\lambda)$ and $\T_{2N+1}(\lambda)$ is $2N$. Now we give the proof of Proposition <ref>. Let $N$ be even. Then (<ref>) implies that N(2\lambda-n+N) \LT(\T_N(\lambda)) + \Delta_h \LT(\T_{N-2}(\lambda)) = 0. It follows that the leading term of $\T_{2N}(\lambda)$ is given by \begin{equation}\label{LT-T} \prod_{j=1}^N \frac{1}{2j (n-2\lambda-2j)} \Delta_h^N = \frac{1}{N! 2^{2N}} \frac{\Gamma(\frac{n}{2}-\lambda-N)}{\Gamma(\frac{n}{2}-\lambda)} \Delta_h^N. \end{equation} Note that this observation fits with <cit.>. Hence \LT( \Res_{\lambda=\frac{n}{2}-N}(\T_{2N}(\lambda)) = - \frac{1}{2^{2N} (N-1)! N!} \Delta_h^N. The second claim follows from that result by combining it with Theorem <ref>. For odd $N$, the operator $\PO_N$ is of order $N-1$ for general metrics. In the following result, we separate from the residue $\Res_{\lambda=\frac{n-N}{2}}(\T_N(\lambda))$ and from $\PO_N$ respective self-adjoint leading terms $\LT(\cdot)$ so that the remaining terms are of order $N-3$ for general metrics. Let $3 \le N \in \N$ be odd. Assume that $\sigma$ satisfies $\SCY$. Then \LT( \Res_{\lambda=\frac{n-N}{2}}(\T_N(\lambda)) = \frac{1}{N (N\!-\!2)!} \sum_{r=0}^{\frac{N-3}{2}} m_N(r) \Delta^r \delta(\lo d) \Delta^{\frac{N-3}{2}-r} \begin{equation}\label{m-coeff} m_N(r) \st \binom{N-1}{r} \prod_{j=1}^{\frac{n-3}{2}} (N-j) \prod_{j=1}^r \frac{1}{(N-2j)} \prod_{j=1}^{\frac{n-3}{2}-r} \frac{1}{(N-2j)}. \end{equation} \begin{equation}\label{LT-P-odd} \LT(\PO_N) = (2N\!-\!2) (N\!-\!1)! \sum_{r=0}^{\frac{N-3}{2}} m_N(r) \Delta^r \delta(\lo d) \Delta^{\frac{N-3}{2}-r}. \end{equation} Let $N$ be odd. Note that $a = 1 + 2 H s + \cdots$ and $\tr (h_s^{-1} h'_s ) = 2 \tr(L) + \cdots$ by Lemma <ref> and (<ref>). Comparing the coefficients of $s^{\lambda+N}$ in (<ref>) and (<ref>) yields the relation \begin{align*} & N (2\lambda\!-\!n\!+\!N) \T_N(\lambda) \\ & + 2 (\lambda\!+\!N\!-\!1) (\lambda\!+\!N\!-\!2) H \T_{N-1}(\lambda) + (\lambda\!+\!N\!-\!1) \tr (L) \T_{N-1}(\lambda) \\ & + (\lambda\!+\!N\!-\!1) H \T_{N-1}(\lambda) - 2(n\!-\!1)(\lambda\!+\!N\!-\!1) \T_{N-1}(\lambda) \\ & - (dH,d\T_{N-3}(\lambda))_h \\ & + \Delta_h \T_{N-2}(\lambda) + \Delta_h' \T_{N-3}(\lambda) = 0, \end{align*} up to operators of order $\le N-3$. Here $\Delta_{h_s} = \Delta_h + s \Delta_h' + \cdots$. Simplification shows that \begin{align}\label{RR-odd} & N(2\lambda\!-\!n\!+\!N) \T_N(\lambda) + (\lambda\!+\!N\!-\!1)(2\lambda\!+\!2N\!-\!n\!-\!1) H \T_{N-1}(\lambda) - (dH,d\T_{N-3}(\lambda))_h \notag \\ & + \Delta \T_{N-2}(\lambda) + \Delta' \T_{N-3}(\lambda) = 0, \end{align} up to operators of order $\le N-3$.[From now on, $\Delta$ is the Laplacian of $h$.] The leading terms of the solution operators $\T_{N-1}(\lambda)$ and $\T_{N-3}(\lambda)$ are multiplies of powers of $\Delta$ (see (<ref>)). Moreover, we recall the variation formula \begin{equation}\label{Laplace-var-g} (d/dt)|_0(\Delta_{g+sh}) = - (\Hess_g (\cdot), h)_g - (\delta^g (h),d \cdot)_g + \frac{1}{2} (d \tr_g(h),d\cdot)_g \end{equation} (<cit.>). Hence $\Delta'$ \begin{align}\label{Delta-var} \Delta'_h \st (d/dt)|_0(\Delta_{h+2sL}) & = - (\Hess_h(\cdot),2L)_h - 2(\delta^h (L),d \cdot)_h + n (dH,d\cdot)_h \notag \\ & = - 2\delta^h ( L d) + n (dH,d \cdot)_h \notag \\ & = - 2 H \Delta_h + (n-2) (dH,d\cdot)_h - 2 \delta^h (\lo d). \end{align} In other words, the first variation $\Delta_h'$ of the Laplacian with respect to the variation $h_s$ of $h$ is given by \begin{equation}\label{var-1} -2H \Delta + (n-2) (dH,d\cdot)_h - 2 \delta (\lo d). \end{equation} It follows that $N(2\lambda\!-\!n\!+\!N)\T_N(\lambda)$ is a linear combination of terms of the form \begin{equation}\label{H-terms} H \Delta^{\frac{N-1}{2}} \quad \mbox{and} \quad (dH,d\Delta^{\frac{N-3}{2}}), \end{equation} \begin{equation}\label{L-terms} \Delta^r \delta ( \lo d) \Delta^{\frac{N-3}{2}-r}, \; r=0,\dots,\tfrac{N-3}{2} \end{equation} and terms of order $\le N-3$. In order to determine the coefficients of the terms (<ref>) in $\T_N(\lambda)$, we let $\mathring{\T}_N(\lambda)$ be the sum of these contributions to $\T_N(\lambda)$. Then (<ref>) implies \begin{equation}\label{RR-odd-L} a_N(\lambda) \mathring{\T}_N(\lambda) + \Delta \mathring{\T}_{N-2}(\lambda) - 2 (-1)^{\frac{N-3}{2}} \prod_{j=1}^\frac{N-3}{2} \frac{1}{a_{2j}(\lambda)} \delta (\lo d) \Delta^{\frac{N-3}{2}} = 0, \end{equation} where $a_N(\lambda) \st N(2\lambda-n+N)$. That recursive relation is solved by (-1)^\frac{N-3}{2} a_N(\lambda) \mathring{\T}_N(\lambda) = 2 \sum_{r=0}^{\frac{N-3}{2}} \prod_{j=1}^{\frac{N-3}{2}-r} \frac{1}{a_{2j}(\lambda)} \prod_{j=1}^{r} \frac{1}{a_{N-2j}(\lambda)} \Delta^r \delta (\lo d) \Delta^{\frac{N-3}{2}-r}. \begin{align*} & N \Res_{\lambda=\frac{n-N}{2}}(\mathring{\T}_N(\lambda)) \\ & = 2^{-\frac{N-3}{2}} \sum_{r=0}^{\frac{N-3}{2}} \frac{1}{(\frac{N-3}{2}\!-\!r)! r!} \prod_{j=1}^{\frac{N-3}{2}-r} \frac{1}{(N-2j)} \prod_{j=1}^r \frac{1}{(N-2j)} \Delta^r \delta (\lo d) \Delta^{\frac{N-3}{2}-r}. \end{align*} Now the inverse of the coefficient of the term for $r=0$ equals 2^{\frac{N-3}{2}} \left(\frac{N-3}{2}\right)! \prod_{j=1}^{\frac{N-3}{2}} (N-2j) = (N-2)!. Therefore, we obtain \Res_{\lambda=\frac{n-N}{2}} \left(\T_N(\lambda)\right) = \frac{1}{N (N-2)!} \sum_{r=0}^{\frac{N-3}{2}} m_N(r) \Delta^r \delta (\lo d) \Delta^{\frac{N-3}{2}-r}, up to contributions by the terms in (<ref>) (containing $H$) and lower-order terms. However, the terms in (<ref>) do not contribute. This is a consequence of the conformal covariance of $\PO_N$. The proof is complete. The formula (<ref>) also makes clear that, if $\lo$ vanishes, the operator $\PO_N$ is of order $< N-2$. For a discussion of the special cases of $\PO_3$ and $\PO_5$, we refer to Lemma <ref> and Lemma <ref>. In particular, the proof of Lemma <ref> confirms the vanishing of the terms $H \Delta$ and Next, we relate the operators $\PO_N$ to the scattering operator $\Sc(\lambda)$ generalizing a result of <cit.>. Let $N \in \N$ with $2 \le N \le n$. Assume that $\sigma$ satisfies $\SCY$ and that $(n/2)^2 - (N/2)^2 \notin \sigma_{pp}$. Then \begin{equation}\label{P-S} \PO_N = (-1)^N 2 (N-1)! N! \Res_{\frac{n-N}{2}}(\Sc(\lambda)). \end{equation} The assumptions guarantee that the scattering operator is well-defined and that $\lambda(n-\lambda) \notin \sigma_{pp}$ for $\lambda=\frac{n-N}{2}$. If $\Re(\lambda) < \frac{n}{2}$ so that $\lambda(n-\lambda) \notin \sigma_{pp}$ and $\lambda \notin \frac{n}{2} - \N$, the Poisson transform $\Po(\lambda)(f)$ yields an eigenfunction $u$ of the Laplacian of the metric $\eta^*(\sigma^{-2}g) = s^{-2} \eta^*(g)$ with boundary value $f \in C^\infty(M)$ and with an asymptotic expansion of the form \begin{equation} \sum_{j\ge 0} s^{\lambda+j} \T_j(\lambda)(f) + \sum_{j \ge 0} s^{n-\lambda+j} \T_j(n-\lambda) \Sc(\lambda)(f). \end{equation} Although the families $\T_N(\lambda)$ and $\Sc(\lambda)$ have simple poles at $\lambda=\frac{n-N}{2}$, the Poisson transform $\Po(\lambda)(f)$ is holomorphic at $\lambda=\frac{n-N}{2}$ <cit.>. That means that \frac{\Res_{\frac{n-N}{2}}(\T_N(\lambda)) s^{\lambda+N}}{\lambda - \frac{n-N}{2}} + \frac{\Res_{{\frac{n-N}{2}}}(\Sc(\lambda)) s^{n-\lambda}}{\lambda-\frac{n-N}{2}} is regular at $\lambda = \frac{n-N}{2}$. Hence \Res_{\frac{n-N}{2}}(\T_N(\lambda)) + \Res_{{\frac{n-N}{2}}}(\Sc(\lambda)) = 0 and the asymptotic expansion of $u$ involves a $\log$-term 2 \Res_{\frac{n-N}{2}}(\T_N(\lambda))(f) s^{\frac{n+N}{2}} \log (s). Now the claim follows from Theorem <ref>. In <cit.>, the scattering operator is defined as $\Sc(n-\lambda)$ and the GJMS-operators $P_{2N}$ have leading part $(-\Delta)^N$. Now \begin{align*} P_{2N}(h) & = \frac{1}{(2N\!-\!1)!!^2} \PO_{2N}(r^2 g_+,r) & \mbox{(by \eqref{P-GJMS})} \\ & = \frac{2(2N\!-\!1)! (2N)!} {(2N\!-\!1)!!^2} \Res_{\frac{n}{2}-N}(\Sc(\lambda)) & \mbox{(by Theorem \ref{LS})} \\ & = 2^{2N} N!(N\!-\!1)! \Res_{\frac{n}{2}-N}(\Sc(\lambda)). \end{align*} This shows that Theorem <ref> extends <cit.>. Let $\N \ni N \le n$. Assume that $\sigma$ satisfied $\SCY$. Then the operators $\PO_N(g,\sigma)$ are formally self-adjoint as operators on $C^\infty(M)$ with respect to the scalar product defined by $h$. Lemma <ref> shows that $\Sc(\lambda)$ is self-adjoint on $\R \setminus \left\{\frac{n-N}{2}\,|\, N \in \Z \right\}$. Since $\Sc(\lambda)$ is meromorphic, its residue at $\lambda = \frac{n-N}{2}$ is self-adjoint, too. Then Theorem <ref> proves the assertion. § EXTRINSIC $Q$-CURVATURES AND RENORMALIZED VOLUME COEFFICIENTS theoremsection equationsection The zeroth-order terms of the GJMS-operators $P_{2N}$ led to the notion of Branson's $Q$-curvature <cit.>. In the present section, we use residue families to extend the notion of Branson's $Q$-curvatures to the framework of the singular Yamabe problem. This extends the discussion of $Q$-curvatures in <cit.>. The resulting curvature quantities will be called extrinsic $Q$-curvatures.[In <cit.>, the critical extrinsic $Q$-curvature is called the extrinsically coupled $Q$-curvature.] We relate the integrated critical renormalized volume coefficient $v_n$ to the integrated critical extrinsic $Q$-curvature. Moreover, we discuss the Hadamard renormalization of the volume of singular Yamabe metrics. Here the total critical extrinsic $Q$-curvature plays an important role. Although the treatment is inspired by <cit.>, our arguments differ and may continue to illuminate these topics. By Theorem <ref>, it holds $\PO_n(g,\sigma) \iota^* = \D_n^{res}(g,\sigma;0)$ if $\sigma$ satisfies $\SCY$. Now, following the philosophy of <cit.>, it is natural to consider the pair $(\D_n^{res}(g,\sigma;0),\dot{\D}_n^{res}(g,\sigma;0)(1))$. $\QC_n(g,\sigma)$ critical extrinsic $Q$-curvature Assume that $\sigma$ satisfies $\SCY$. Then the function \begin{equation}\label{Q-critical} \QC_n(g,\sigma) \st - \dot{\D}_n^{res}(g,\sigma;0)(1) \in C^\infty(M) \end{equation} is called the critical extrinsic $Q$-curvature of $g$. Since $L(g,\sigma;0)(1) = 0$, the identification of residue families with products of $L$-operators (Theorem <ref>) implies that \begin{equation}\label{QL} \QC_n(g,\sigma) = - \iota^* L(g,\sigma;-n\!+\!1) \circ \cdots \circ L(g,\sigma;-1) \circ \dot{L}(g,\sigma;0)(1). \end{equation} Moreover, the definition of $L$ yields \begin{equation}\label{b} \dot{L}(g,\sigma;0)(1) = (n-1) \rho - \sigma \J. \end{equation} Therefore, we obtain \begin{equation}\label{QL-2} \QC_n(g,\sigma) = - \iota^* L(g,\sigma;-n\!+\!1) \circ \cdots \circ L(g,\sigma;-1) ((n-1) \rho - \sigma \J). \end{equation} Next, we define subcritical versions of $\QC_n$. Let $N < n$. The definition of $L$ implies that \begin{equation}\label{aa} L\left(g,\sigma;\frac{-n+N}{2}\right)(1) = \left(\frac{n-N}{2}\right) (-(N\!-\!1) \rho + \sigma \J). \end{equation} Hence the function $\PO_N(g,\sigma)(1)$ is of the form \left(\frac{n-N}{2}\right) \QC_N(g,\sigma) with a scalar curvature quantity $\QC_N(g,\sigma) \in C^\infty(M)$. It follows that \begin{equation}\label{a} \QC_N(g,\sigma) = -\iota^* L\left(g,\sigma;\frac{-n\!-\!N}{2}\!+\!1\right) \circ \cdots \circ L\left(g,\sigma;\frac{-n\!+\!N}{2}\!-\!1\right) ((N\!-\!1)\rho - \sigma \J) \end{equation} We shall call these quantities subcritical extrinsic $Q$-curvatures. In terms of residue families, these definitions are equivalent to the following definition. $\QC_N(g,\sigma)$ extrinsic $Q$-curvature Let $\N \ni N < n$. Assume that $\sigma$ satisfies $\SCY$. The functions $\QC_N(g,\sigma) \in C^\infty(M)$ which are determined by the equation \begin{equation}\label{Q-sub-residue} \D_N^{res}\left(g,\sigma;\frac{-n+N}{2}\right) (1) = \left(\frac{n-N}{2}\right) \QC_N(g,\sigma) \end{equation} are called subcritical extrinsic $Q$-curvatures. Since the residue families $\D_N^{res}(g,\sigma;\lambda)$ for $N \le n$ are completely determined by $g$ (and $\iota$) (Proposition <ref>), the quantities $\QC_N(g,\sigma)$ for $N \le n$ are also completely determined by $g$ (and $\iota$). Therefore, we shall also use the notation $\QC_N(g)$. The identities (<ref>) and (<ref>) show that the critical extrinsic $Q$-curvature is a limiting case $\lim_{n \to N}$ of the subcritical extrinsic $Q$-curvatures (continuation in dimension). The subcritical $Q$-curvature $\QC_N$ is directly linked to the solution operator $\T_N(\lambda)$ through the relation \begin{equation}\label{QT} \QC_N = (-1)^N \QC_N\left(\frac{n-N}{2}\right), \end{equation} where the polynomial $\QC_N(\lambda)$ is defined in (<ref>). In fact, (<ref>) implies \Res_{\frac{n-N}{2}}(\T_N)(1) = -\frac{\frac{n-N}{2}}{2 (N-1)! N!} \QC_N\left(\frac{n-N}{2}\right). On the other hand, Theorem <ref> shows that \Res_{\frac{n-N}{2}}(\T_N)(1) = - (-1)^{N} \frac{1}{2 (N-1)!N!} \PO_N (1) = - (-1)^N \frac{\frac{n-N}{2}}{2 (N-1)!N!} \QC_N. The identity (<ref>) follows by combining both relations. Continuation in $n$ also gives the relation \begin{equation}\label{QT-critical} \QC_n = (-1)^n \QC_n(0) \end{equation} in the critical case. The extrinsic $Q$-curvatures $\QC_N$ are related to Branson's $Q$-curvatures $Q_{2N}$ as follows. We use the convention that $Q_{2N}$ for even $n$ and $2N <n$ is defined by P_{2N} (g) (1) = \left({\frac{n}2}-N\right) (-1)^N Q_{2N}(g), where $P_{2N} = \Delta^N + LOT$ is the GJMS-operator of order $2N$ with $\Delta$ denoting the non-positive Laplacian. Now let $g_+$ be a Poincaré-Einstein metric in normal form relative to $h$. Then (<ref>) implies \QC_{2N}(r^2 g_+,r) = (-1)^N(2N-1)!!^2 Q_{2N}(h). Formula (<ref>) for the critical extrinsic $Q$-curvature $\QC_n$ is closely related to the definition of the critical extrinsic $Q$-curvature used in <cit.>. In fact, under the assumption $\SC(g,\sigma)=1$, <cit.> defines $\QC_n(g,\sigma)$ by (-1)^{n} \iota^* L(g,\sigma;-n+1) \circ \cdots \circ L(g,\sigma;-1) \circ L_{\log}(g,\sigma;1) (\log (1)). Here $L_{\log}(g,\sigma;\lambda)$ is a version of $L$ which acts on log-densities by[For the definition of the notion of log-densities, we refer to <cit.>.] L_{\log}(g,\sigma;\lambda)(u) \st (n-1) (\nabla_{\grad(\sigma)}(u) + \lambda \rho) - \sigma (\Delta_g (u) + \lambda \J). In particular, we have $L_{\log}(g,\sigma;\lambda)(\log(1)) = \lambda ((n-1) \rho -\sigma \J)$ and hence L_{\log}(g,\sigma;1)(\log(1)) = (n-1) \rho - \sigma \J = \dot{L}(g,\sigma;0)(1) using (<ref>). Differentiation of the conformal transformation law for the critical residue family $\D_n^{res}(g,\sigma;\lambda)$ at $\lambda=0$ yields the following result. Assume that $\sigma$ satisfies $\SCY$. Then \begin{equation}\label{CTL-Q2} e^{n \iota^*(\varphi)} \QC_n(\hat{g},\hat{\sigma}) = \QC_n(g,\sigma) + \PO_n(g,\sigma)(\iota^*(\varphi)) \end{equation} for all conformal changes $(\hat{g},\hat{\sigma}) = (e^{2\varphi}g,e^\varphi \sigma)$, $\varphi \in C^\infty(X)$. Theorem <ref> implies e^{-(\lambda-n)\iota^*(\varphi)} \circ \D_n^{res}(\hat{g},\hat{\sigma};\lambda) = \D_n^{res}(g,\sigma;\lambda) \circ e^{-\lambda \varphi}. By differentiating this identity with respect to $\lambda$ at $\lambda=0$, we obtain -e^{n \iota^*(\varphi)} \iota^*(\varphi) \circ \D_n^{res}(\hat{g},\hat{\sigma};0) + e^{n \iota^*(\varphi)} \circ \dot{\D}_n^{res}(\hat{g},\hat{\sigma};0) = \dot{\D}_n^{res}(g,\sigma;0) - \D_n^{res}(g,\sigma;0) \circ \varphi. -\iota^*(\varphi) \D_n^{res}(g,\sigma;0)(1) + e^{n \iota^*(\varphi)} \dot{\D}_n^{res}(\hat{g},\hat{\sigma};0)(1) = \dot{\D}_n^{res}(g,\sigma;0)(1) - \D_n^{res}(g,\sigma;0) (\varphi). Now $\D_n^{res}(g,\sigma;0) =\PO_n(g,\sigma) \iota^*$ and (<ref>) imply the For closed $M^n$, integration of(<ref>) implies \int_{M} \QC_n(\hat{g},\hat{\sigma}) dvol_{\hat{h}} = \int_{M} \QC_n(g,\sigma) dvol_h + \int_{M} \PO_n(g,\sigma)(\iota^*(\varphi)) dvol_h. Later we shall prove that $\PO_n$ is self-adjoint. Hence the second integral on the right-hand side equals $\int_M \PO_n (1) \iota^*(\varphi) dvol_h$. By (<ref>), this integral vanishes. This shows Let $M^n$ be closed. Assume that $\sigma$ satisfies $\SCY$. Then the integral \int_M \QC_n(g) dvol_h is conformally invariant as a functional of $g$ and the embedding $M \hookrightarrow X$. An alternative argument proving this invariance will be given in Theorem <ref>. A generalization of Corollary <ref> to general $\sigma$ will be discussed in Lemma <ref>. Next, we recall that Theorem <ref> relates the operators $\PO_N$ to the scattering operator $\Sc(\lambda)$. In particular, it holds \PO_n = (-1)^n 2 (n-1)! n! \Res_0(\Sc(\lambda)). Since $\PO_n(1)=0$, it follows that the function $\Sc(\lambda)(1)$ is regular at $\lambda=0$. Its value at $\lambda=0$ will be denoted by $\Sc(0)(1)$. Assume that $\sigma$ satisfies $\SCY$. Then \QC_n = 2 (-1)^n (n-1)! n! \Sc(0)(1). We consider the coefficient of $s^n$ in the asymptotic expansion (<ref>) of the eigenfunction $\Po(\lambda)(1)$ for $\lambda=0$. Since $\Po(0)(1)=1$, that coefficient vanishes. On the other hand, it is given by \T_n(0)(1) + \Sc(0)(1); we recall that the function $\T_n(\lambda)(1)$ is regular at $\lambda=0$. Hence $\T_n(0)(1) = - \Sc(0)(1)$. But \begin{equation}\label{T-Q} \T_n(0)(1) = -\frac{1}{2 (n-1)! n!} \QC_n(0) = - (-1)^n \frac{1}{2 (n-1)!n!} \QC_n \end{equation} using (<ref>) and (<ref>). This implies the assertion. Theorem <ref> extends <cit.> for the scattering operator of Poincaré-Einstein metrics. In fact, \begin{align*} Q_n & = (-1)^\frac{n}{2} \frac{1}{(n-1)!!^2} \QC_n & \mbox{(by Remark \ref{Q-GJMS})} \\ & = (-1)^\frac{n}{2} \frac{2 (n-1)! n!}{(n-1)!!^2} \Sc(0)(1) & \mbox{(by Theorem \ref{Q-S})} \\ & = (-1)^\frac{n}{2} 2^n \left(\frac{n}{2}\right)! \left(\frac{n}{2}-1\right)! \Sc(0)(1). \end{align*} In that case, the crucial relation between the values of $\Sc(\lambda)(1)$ and $\T_n(\lambda)(1)$ at $\lambda=0$ is provided by <cit.>. The following result (<cit.> for $\tau=1$ and $\SC=1$) is an analog of the identity (<ref>) in the critical case. Let $M$ be closed and assume that $\sigma$ satisfies $\SCY$. Let $n\ge 2$ be even. Then it holds the equality \begin{equation}\label{vQ} \LA = \int_M v_n(g) dvol_h = \frac{(-1)^{n} }{(n-1)! n!} \int_M \QC_n(g) dvol_h \end{equation} of conformal invariants. The quantity $\LA$ is sometimes referred to as an anomaly. This is motivated by the fact that, in the Poincaré-Einstein case, the function $v_n$ is the infinitesimal conformal anomaly of the renormalized volume <cit.>. We shall give two proofs of that result. The arguments in the first proof will also play a role in Section <ref>. The second proof resembles the proof of the analogous result in the subcritical cases. We work in adapted coordinates. In particular, the notation will not distinguish between objects on $X$ and their pull-backs by $\eta$. First, we note that \begin{equation}\label{help-2} \iota^* \partial_s^{n-1} (v \dot{L}(0)(1)) = \iota^* \partial_s^n (v). \end{equation} Since $\dot{L}(0)(1) = (n-1) \rho - s \J$ (see (<ref>)), this local identity is a special case of the local identity (<ref>). The current assumption implies $|\NV|^2 = 1 - 2 s \rho + O(s^{n+1})$ and the same arguments as in the proof of Theorem <ref> yield the assertion. Now we integrate (<ref>). It holds \int_M \iota^* \partial_s^n (v) dvol_h = n! \int_M v_n dvol_h and the integral of the left-hand side of (<ref>) equals \begin{align*} & (-1)^{n-1} \langle \mathfrak{X}^{n-1} (\sigma^*(\delta)), \dot{L}(0)(1) dvol_g \rangle & \mbox{(by \eqref{fund-dist})} \\ & = c_n \langle L(-n\!+\!1) \circ \cdots \circ L(-1) (\sigma^*(\delta)), \dot{L}(0)(1) dvol_g \rangle & \mbox{(by Theorem \ref{dist-volume})} \\ & = c_n \langle \sigma^*(\delta), L(-n\!+\!1) \circ \cdots \circ L(-1) \circ \dot{L}(0)(1) dvol_g \rangle & \mbox{(by \eqref{dual-1})} \\ & = c_n \int_M \iota^* L(-n\!+\!1) \circ \cdots \circ L(-1) \circ \dot{L}(0)(1) dvol_h & \mbox{(by \eqref{Ho-simple})} \\ & = -c_n \int_M \QC_n dvol_h & \mbox{(by \eqref{QL})} \end{align*} with $c_n = (-1)^{n-1}/(n\!-\!1)!$. This completes the proof. In the Poincaré-Einstein case, it holds $v' = v \dot{L}(0)(1)$. This immediately proves (<ref>). The proof of Theorem <ref> rests on the local identity (<ref>). This identity will also play a role in Section <ref>. We continue with a Theorem <ref> yields the identity \begin{equation}\label{critical-c} \int_M \D_n^{res}(\lambda)(1) dvol_h = n! (2\lambda\!-\!n\!+\!1)_n \sum_{j=0}^n \int_M v_j \T_{n-j}(\lambda)(1) dvol_h. \end{equation} We split the sum on the right-hand side as \sum_{j=1}^{n-1} \int_M v_j \T_{n-j}(\lambda)(1) dvol_h + \int_M v_n dvol_h + \int_M \T_n(\lambda)(1) dvol_h. Now we differentiate (<ref>) at $\lambda=0$. Since, $\T_j(0)(1)=0$ for $j=1,\dots,n-1$, we find \int_M \dot{\D}_n^{res}(0)(1) dvol_h = 2 (-1)^{n-1} (n-1)! n! \left( \int_M v_n dvol_h + \int_M \T_n(0)(1) dvol_h \right) using that $\T_n(\lambda)(1)$ is regular at $\lambda=0$. Now (<ref>) and (<ref>) imply -2 \int_M \QC_n dvol_h = 2 (-1)^{n-1} (n-1)!n! \int_M v_n dvol_h. This implies (<ref>). We finish this section with a discussion of renormalized volumes of singular metrics $\sigma^{-2} g$. First, we combine the above results to prove Let $M$ be closed and assume that $\sigma$ satisfied $\SCY$. Then the volume vol_{\sigma^{-2}g}(\{\sigma > \varepsilon \}) = \int_{\sigma > \varepsilon} dvol_{\sigma^{-2}g} admits the expansion \sum_{k=0}^{n-1} \frac{c_k}{n-k} \varepsilon^{-n+k} - \LA \log \varepsilon + V + o(1), \; \varepsilon \to 0, c_k = \int_M v_k(g) dvol_h \quad \mbox{and} \quad \LA = \int_M v_n(g) dvol_h. $V$ is called the renormalized volume. The coefficients in the expansion are natural functionals of the metric background $g$, which can be written in the form c_k = (-1)^k \frac{(n\!-\!1\!-\!k)!}{(n\!-\!1)! k!} \int_M \iota^* L_k(-n\!+\!k) (1) dvol_h for $k=0,\dots,n-1$ and \begin{equation}\label{anomaly-SY} \LA = (-1)^{n-1} \frac{1}{(n\!-\!1)! n!} \int_M \iota^* \dot{L}_n(0) (1) dvol_h. \end{equation} Using $\eta^*(dvol_{\sigma^{-2} g}) = dvol_{s^{-2} \eta^*(g)} = s^{-n-1} v(s) ds dvol_h$, we obtain the asymptotic expansion \int_{\sigma > \varepsilon} dvol_{\sigma^{-2}g} = \sum_{k=0}^{n-1} \frac{1}{n-k} \varepsilon^{-n+k} \int_M v_k dvol_h - \log \varepsilon \int_M v_n dvol_h + V + o(1). Now the expressions for the coefficients in terms of Laplace-Robin operators follow from Corollary <ref> and Theorem <ref>. The above definition of the renormalized volume is also known as the Hadamard renormalization <cit.>. In the proof of Theorem <ref>, we deduced the formulas for the coefficients $c_k$ from the relation between residue families and iterated Laplace-Robin operators (Theorem <ref>). We recall that this relation requires assuming that $\sigma$ solves the singular Yamabe problem. However, the only consequences of Theorem <ref> which are relevant in this context already follow from a study of the residues of the meromorphic continuation of the integral \begin{equation}\label{basic-int} \lambda \mapsto \int_X \sigma^\lambda dvol_g, \; \Re(\lambda) > -1. \end{equation} Therefore it is of interest to include a discussion of an extension of Theorem <ref> to general $\sigma$ which only rests on the study of that integral. First, we note that the coefficients $c_k$ in Theorem <ref> are related to the residues of (<ref>): $c_k = \int_M v_k dvol_h$. Moreover, the Hadamard renormalization $V$ of the volume of $\sigma^{-2}g$ is related to the Riesz renormalization of the volume of $\sigma^{-2} g$ which is defined as the constant term in the Laurent series of (<ref>) at $\lambda=-n-1$ <cit.>. We first observe that Remark <ref> implies \begin{equation*} \SC(g,\sigma)^{-1} \circ L(g,\sigma;\lambda) + \sigma^{\lambda-1} \circ \SC(g,\sigma)^{-1} \circ \Delta_{\sigma^{-2}g} \circ \sigma^{-\lambda} = \lambda (n\!+\!\lambda) \sigma^{-1} \id \end{equation*} in a sufficiently small neighborhood of the boundary, where $\SC(g,\sigma) \ne 0$. Hence \begin{equation}\label{BS-shift} \SC(g,\sigma)^{-1} L(g,\sigma;\lambda) (\sigma^{\lambda}) = \lambda (n\!+\!\lambda) \sigma^{\lambda-1}. \end{equation} This is a Bernstein-Sato-type functional equation. Let $\chi \in C_c^\infty(X)$ be a cut-off function of the boundary $M$ so that $\SC \ne 0$ on the support of $\chi$. In the following, we shall simplify the notation by suppressing the arguments $(g,\sigma)$ of $L$ and $\SC$. The second integral on the right-hand side of the \int_X \sigma^\lambda dvol_g = \int_X \chi \sigma^\lambda dvol_g + \int_X (1-\chi) \sigma^\lambda dvol_g is holomorphic on $\C$. Now (<ref>) implies \int_X \chi \sigma^\lambda dvol_g = \frac{1}{(\lambda\!+\!1) (n\!+\!\lambda\!+\!1)} \int_X \chi \SC^{-1} L(\lambda\!+\!1)(\sigma^{\lambda+1}) dvol_g for $\Re(\lambda) > -1$. Partial integration using Proposition <ref> shows that for $\Re(\lambda) \gg 0$ the latter integral equals \int_X \sigma^{\lambda+1} L(-\lambda\!-\!n\!-\!1)(\chi \SC^{-1}) dvol_g. Next, we note that $v_0 = \iota^* |\NV|^{-1}$ implies the residue formula \begin{equation}\label{res-formula} \Res_{\lambda=-1} \left( \int_X \sigma^\lambda \psi dvol_g \right) = \int_M \iota^* \frac{1}{|\NV|} \psi dvol_h. \end{equation} Now combining the above result with the residue formula (<ref>) yields \Res_{\lambda=-2} \left( \int_X \sigma^\lambda dvol_g \right) = - \frac{1}{n\!-\!1} \int_M \iota^* \frac{1}{|\NV|} L(-n\!+\!1)(\SC^{-1}) dvol_h. More generally, arguments as in the proof of Theorem <ref> resting on a repeated application of the functional equation (<ref>) and Proposition <ref> show that \begin{equation}\label{vol-pi} \int_X \chi \sigma^\lambda dvol_g = \frac{1}{(\lambda\!+\!1)_k (\lambda\!+\!n\!+\!1)_k} \int_X \sigma^{\lambda+k} \tilde{L}_k (-\lambda\!-\!n\!-\!1)(\chi) dvol_g, \end{equation} for $\Re(\lambda) \gg 0$, and it follows that \begin{equation}\label{res-vol-g} \Res_{\lambda=-k-1} \left( \int_X \sigma^\lambda dvol_g \right) = \frac{1}{(-n\!+\!1)_k k!} \int_M \iota^* \frac{1}{|\NV|} \tilde{L}_k (-n\!+\!k) (1) dvol_h \end{equation} for $k < n$ using the relation $\iota^* \tilde{L}_k(\lambda)(\chi) = \iota^* \tilde{L}_k(\lambda)(1)$. The assumption $k < n$ guarantees that the prefactor on the right-hand side of (<ref>) is regular at $\lambda = -k-1$. Here we use the notation \tilde{L}(\lambda) \st L(\lambda) \circ \SC^{-1} \quad \mbox{and} \quad \tilde{L}_k(\lambda) \st \tilde{L}(\lambda\!-\!k\!+\!1) \circ \cdots \circ \tilde{L}(\lambda) (see (<ref>) and (<ref>)). If $\SC =1$, then $\iota^* \NV=1$ and \begin{align*} \Res_{\lambda=-2} \left( \int_X \sigma^\lambda dvol_g \right) & = - \frac{1}{n\!-\!1} \int_M \iota^* L(-n\!+\!1)(1) dvol_h \\ & = -(n-1) \int_M \iota^* \rho dvol_h = (n-1) \int_M H dvol_h \end{align*} using $\iota^* \rho = - H$ (Lemma <ref>). This fits with the fact that $v_1 = (n-1)H$ (Example <ref>). Similarly, if $\sigma = d_M$ is the distance function of $M$, then $\SC = 1+2\sigma \rho$ implies $\iota^* L(-n+1)(\SC^{-1}) = (n-1)^2 \iota^* \rho + 2(n-1) \iota^* \rho = (n-1)(n+1) \iota^* \rho$. Hence \begin{align*} \Res_{\lambda=-2} \left( \int_X \sigma^\lambda dvol_g \right) = \int_M \iota^* \Delta_g(\sigma) dvol_h = n \int_M H dvol_h. \end{align*} These results prove the first part of the following theorem. Let $M$ be closed and $\sigma$ be a defining function of $M$. Then the volume vol_{\sigma^{-2}g}(\{\sigma > \varepsilon \}) = \int_{\sigma > \varepsilon} dvol_{\sigma^{-2}g} admits the expansion \sum_{k=0}^{n-1} \frac{c_k}{n-k} \varepsilon^{-n+k} - \LA \log \varepsilon + V + o(1), \; \varepsilon \to 0, c_k = \int_M v_k dvol_h = (-1)^k \frac{(n\!-\!1\!-\!k)!}{(n\!-\!1)! k!} \int_M \iota^* \frac{1}{|\NV|} \tilde{L}_k(-n\!+\!k)(1) dvol_h for $k=0,\dots,n-1$. Moreover, it holds \begin{equation}\label{anomaly-g} \LA = (-1)^{n-1} \frac{1}{(n\!-\!1)! n!} \int_M \iota^* \frac{1}{|\NV|} \tilde{L}_{n-1}(-1) \left( \dot{L}(0)(1) \SC^{-1} + (d \SC^{-1}, d\sigma)_g \right) dvol_h. \end{equation} Theorem <ref> is due to <cit.>. For a discussion of the relations between the current arguments and the proofs in these references, we refer to Remark <ref>. We work in adapted coordinates. The form of the expansion follows as in the proof of Theorem <ref>. Since $c_k = \int_M v_k dvol_h$ is the residue of $\int_X\sigma^\lambda dvol_g$ at $\lambda=-k-1$, the asserted formula for $c_k$ follows by a calculation of these residues using the functional equation (<ref>). It only remains to prove the formula for the anomaly $\LA$. Note that the proof for $c_k$ with $k \le n-1$ does not extend to the present case since for $k=n$ the right-hand side of (<ref>) has a double pole at $\lambda=-n-1$. To bypass this difficulty, we first prove the local relation[This relation can be interpreted as a local version of the global relation in <cit.>.] \begin{equation}\label{Myst-a} \iota^* \partial_s^{n-1} \left(v \left[((n-1) \rho - s\J) \SC^{-1} + |\NV|^2 \partial_s(\SC^{-1})\right] \right) = \iota^* \partial_s^n(v) \end{equation} near $M$. Let I \st \iota^* \partial_s^{n-1} (v \left[(n-1) \rho - s\J \right] \SC^{-1}). The definition $\rho = -\frac{1}{n+1}(\Delta (s) + s \J)$ implies that \begin{equation}\label{help-1} I = - \frac{n-1}{n+1} \left(\iota^* \partial_s^{n-1} (v \Delta (s) \SC^{-1}) + 2n \iota^* \partial_s^{n-2} (v \J \SC^{-1}) \right). \end{equation} Now we consider the first term on the right-hand side of (<ref>). By (6.17), it holds v \Delta(s) = \partial_s (v |\NV|^2). Hence we obtain \iota^* \partial_s^{n-1} (v \Delta (s) \SC^{-1}) = \iota^* \partial_s^n (v |\NV|^2 \SC^{-1}) - \iota^* \partial_s^{n-1} (v |\NV|^2 \partial_s(\SC^{-1})). But $\SC = |\NV|^2 + 2 s \rho$ implies $|\NV|^2 = \SC -2s \rho$. Hence \iota^* \partial_s^{n-1} (v \Delta (s) \SC^{-1}) = \iota^* \partial_s^n (v) - 2 n \iota^* \partial_s^{n-1} (v \rho \SC^{-1}) - \iota^* \partial_s^{n-1} (v |\NV|^2 \partial_s(\SC^{-1})). Now another application of the definition of $\rho$ shows that the middle term on the right-hand side of the last equation \frac{2n}{n+1} \iota^* \partial_s^{n-1} ((v \Delta(s) + v s \J)\SC^{-1}). Simplification of the resulting equation yields \begin{align}\label{red-id} & -\frac{n-1}{n+1} \iota^* \partial_s^{n-1} (v \Delta (s) \SC^{-1}) \notag \\ & = \iota^* \partial_s^n (v) + \frac{2n(n-1)}{n+1} \iota^* \partial_s^{n-2} (v \J) - \iota^* \partial_s^{n-1} (v |\NV|^2 \partial_s(\SC^{-1})). \end{align} By combining this result with (<ref>), we have proved I = \iota^* \partial_s^n (v) - \iota^* \partial_s^{n-1} (v |\NV|^2 \partial_s(\SC^{-1})). This implies (<ref>). Now the relation (<ref>) shows that \begin{align*} \int_M v_n dvol_h = \frac{1}{n!} \int_M \iota^* \partial_s^n(v) dvol_h = \frac{1}{n!} \int_M \iota^* \partial_s^{n-1} \left( v \E \right) dvol_h, \end{align*} \begin{equation}\label{E} \E \st ((n-1) \rho - s\J) \SC^{-1} + |\NV|^2 \partial_s(\SC^{-1}). \end{equation} \int_M v_n dvol_h = \frac{(n-1)!}{n!} \Res_{\lambda=-n} \left(\int_X \sigma^\lambda \E dvol_g \right) = \frac{1}{n} \Res_{\lambda=-n} \left(\int_X \sigma^\lambda \E dvol_g \right). Now we apply the functional equation (<ref>). We assume that $\Re(\lambda) \gg 0$ and let $\chi \in C_c^\infty(X)$ be a cut-off function as in the discussion following (<ref>). First, the relation \sigma^\lambda = \SC^{-1} \frac{1}{(\lambda\!+\!1)(n\!+\!\lambda\!+\!1)} L(\lambda\!+\!1)(\sigma^{\lambda+1}) \begin{align*} \int_X \chi \sigma^\lambda \E dvol_g & = \frac{1}{(\lambda\!+\!1)(n\!+\!\lambda\!+\!1)} \int_X \chi \SC^{-1} L(\lambda\!+1\!)(\sigma^{\lambda+1}) \E dvol_g \\ & = \frac{1}{(\lambda\!+\!1)(n\!+\!\lambda\!+\!1)} \int_X \sigma^{\lambda+1} \tilde{L}(-n\!-\!\lambda\!-\!1) (\chi \E) dvol_g \end{align*} by partial integration using Proposition <ref>. We continue to apply this argument and find \int_X \chi \sigma^\lambda \E dvol_g = \frac{1}{(\lambda\!+\!1)_{n-1} (n\!+\!\lambda\!+\!1)_{n-1}} \int_X \sigma^{\lambda+n-1} \tilde{L}_{n-1}(-n\!-\!\lambda\!-\!1)(\chi \E) dvol_g. Hence the residue formula (<ref>) yields \begin{align*} \Res_{\lambda=-n} \left(\int_X \sigma^\lambda \E dvol_g \right) &= \frac{1}{(-n\!+\!1)\dots(-1) (n\!-\!1)!} \int_M \iota^* \tilde{L}_{n-1}(-1)(\chi \E) dvol_h \\ & = (-1)^{n-1} \frac{1}{(n\!-\!1)! (n\!-\!1)!} \int_M \iota^* \frac{1}{|\NV|} \tilde{L}_{n-1}(-1)(\E) dvol_h. \end{align*} Thus, we find \int_M v_n dvol_h = (-1)^{n-1} \frac{1}{n! (n-1)!} \int_M \iota^* \frac{1}{|\NV|} \tilde{L}_{n-1}(-1)(\E) dvol_h. This proves the formula for the anomaly using $\dot{L}(0)(1) = (n-1) \rho - s \J$ (see (<ref>)) and the fact that $\grad(\sigma)$ in adapted coordinates equals $|\NV|^2 \partial_s$ (see (<ref>)). The formulas for the coefficients $c_k$ in Theorem <ref> and Theorem <ref> are special cases of <cit.> (for $\tau=1$). Similarly, the formulas for $\LA$ are special cases of <cit.>. But note that in <cit.> there are no local coefficients $v_k$ (defined in adapted coordinates). The present proofs differ from those in <cit.>. Whereas the above proofs rest on the conjugation formula and a Bernstein-Sato-type argument, the latter rest on a certain distributional calculus (see also Remark <ref>). Since the expansion in Theorem <ref> can be written in the form \sum_{k=0}^{n-1} \frac{1}{n-k} \left\langle \sigma^*(\delta^{(k)}),dvol_g \right\rangle \varepsilon^{-n+k} - (-1)^{n-1}\left\langle \sigma^*(\delta^{(n)}),dvol_g \right\rangle \log \varepsilon + V + o(1) using the formula \begin{equation}\label{BR} \left\langle \sigma^*(\delta^{(k)}),dvol_g \right\rangle = (-1)^k k! \int_M v_k dvol_h \end{equation} for the currents $\sigma^*(\delta^{(k)})$, this proves the equivalence of Theorem <ref> and <cit.> combined with <cit.> (for $\tau =1$). We continue with a proof of the relation (<ref>). First, we extend $g$ and $\sigma$ smoothly to a sufficiently small neighborhood $\tilde{X}$ of $M$ as in the discussion after Corollary <ref>. Then $\sigma^*(\delta^{(k)})$ is defined as a continuous functional on smooth volume forms on $\tilde{X}$ with compact support.[Since $\sigma^*(\delta)$ has compact support, we can pair it with any smooth volume form.] Next, we note the commutation rule \begin{equation}\label{B-1} \sigma^*(u') = \mathfrak{X}_\sigma (\sigma^*(u)), \; u \in C^\infty(\R) \end{equation} \mathfrak{X}(\tilde{X}) \ni \mathfrak{X}_\sigma \st \NV/|\NV|^2, \; \NV = \grad_g(\sigma). \begin{equation}\label{B-2} \sigma^*(u^{(k)}) = \mathfrak{X}^k_\sigma (\sigma^*(u)) \end{equation} for $k \in \N$. Now any $\varphi \in C^\infty(\tilde{X})$ defines a current on $\tilde{X}$ by \langle \varphi, \psi dvol_g \rangle = \int_{\tilde{X}} \varphi \psi dvol_g, \; \psi \in C_0^\infty(\tilde{X}). An extension of (<ref>) to $u=\delta$ yields an analogous relation for currents; it suffices to approximate $\delta$ by test functions in the weak topology. The relation (<ref>) for $u=\delta$ implies \begin{align*} \left \langle \sigma^*(\delta^{(k)}), \psi dvol_g \right \rangle & = \left\langle \mathfrak{X}_\sigma^k (\sigma^*(\delta)), \psi dvol_g \right \rangle \\ & = \left \langle \sigma^*(\delta), (\mathfrak{X}_\sigma^*)^k (\psi) dvol_g \right \rangle, \end{align*} where the adjoint operator $\mathfrak{X}^*_\sigma$ is defined by \int_{\tilde{X}} \mathfrak{X}_\sigma(\varphi) \psi dvol_g = \int_{\tilde{X}} \varphi \mathfrak{X}_\sigma^*(\psi) dvol_g, \; \psi \in C_c^\infty(\tilde{X}). Now we calculate (using adapted coordinates and partial integration) \begin{align*} \int_{\tilde{X}} \mathfrak{X}^k_\sigma (\varphi) \psi dvol_g & = \int_{I \times M} \eta^*(\mathfrak{X}_\sigma^k(\varphi)) \eta^*(\psi) v ds dvol_h \\ & = \int_{I \times M} \partial_s^k (\eta^*(\varphi)) \eta^*(\psi) v ds dvol_h \qquad \mbox{(by \eqref{intertwine})} \\ & = (-1)^k \int_{I \times M} \eta^*(\varphi) v^{-1} \partial_s^k(\eta^*(\psi) v) v ds dvol_h \qquad \mbox{(by partial integration)} \\ & = (-1)^k \int_X \varphi \eta_* (v^{-1} \partial_s^k (v \eta^*(\psi))) dvol_g. \end{align*} This shows that $\mathfrak{X}_\sigma^* (\psi) = \eta_* (v^{-1} \partial_s^k (v \eta^*(\psi)))$. Next, we observe that \left \langle \sigma^*(\delta), \psi dvol_g \right\rangle = \int_M \iota^* \frac{\psi}{|\NV|} dvol_h. This formula is an extension of <cit.> (for a flat background).[With appropriate interpretation, this identity coincides with <cit.>. The current reference to Hörmander replaces the reference for this result which has been used in <cit.>. This reference actually only utilizes formal arguments for domains in flat space $\R^n$.] $\iota^*(v) = v_0 = \frac{1}{|\NV|}$, these results imply \begin{align*} \left\langle \sigma^*(\delta^{(k)}),\psi dvol_g \right\rangle & = (-1)^k \int_M \iota^* \partial_s^k (v \eta^*(\psi)) dvol_h. \end{align*} In particular, for $\psi = 1$ we find (<ref>). By (<ref>), the anomaly $\LA$ is proportional to \int_M \iota^* \frac{1}{\sqrt{\SC}} \tilde{L}_{n-1}(-1) (\E) dvol_h with $\E = \dot{L}(0)(1) \SC^{-1} + (d \SC^{-1},d\sigma)_g$ (see (<ref>)). In the singular Yamabe case, this integral reduces to \int_M \iota^* L_{n-1}(-1) \dot{L}(0)(1) dvol_h = \int_M \iota^* \dot{L}_n(0)(1) dvol_h = - \int_M \QC_n dvol_h. This motivates Gover and Waldron <cit.> to regard the function \begin{equation}\label{tilde-Qn} \tilde{\QC}_n(g,\sigma) \st \iota^* \frac{1}{\sqrt{\SC}} \tilde{L}_{n-1}(g,\sigma;-1) (\E) \end{equation} as a generalized critical extrinsic $Q$-curvature. We show that this quantity shares basic properties with $\QC_n$. In this context, we consider conformal changes $(\hat{g},\hat{\sigma}) = (e^{2\varphi} g, e^\varphi \sigma)$ and we let $L = L(g,\sigma)$ and $\hat{L} = L(\hat{g},\hat{\sigma})$. Similarly $\hat{\cdot}$ also denotes other functionals of $(g,\sigma)$ for such conformal changes. Now differentiating the conformal transformation law e^{-(\lambda-1) \varphi} \circ \hat{L}(\lambda) = L(\lambda) \circ e^{-\lambda \varphi} at $\lambda = 0$, gives e^\varphi \dot{\hat{L}}(0)(1) = \dot{L}(0)(1) - L(0)(\varphi) using $L(0)(1)=0$. Hence e^\varphi \hat{\E} = \E - \SC^{-1} L(0)(\varphi) + \sigma (d\SC^{-1},d\varphi), and the conformal covariance of $\tilde{L}(\lambda)$ implies \begin{align*} e^{n \iota^*(\varphi)} \hat{\tilde{\QC}}_n & = \iota^* \frac{1}{\sqrt{\SC}} e^{n \varphi} \hat{\tilde{L}}_{n-1}(-1) (\hat{\E}) = \iota^* \frac{1}{\sqrt{\SC}}\tilde{L}_{n-1}(-1) (e^\varphi \hat{\E}) \\ & = \iota^* \frac{1}{\sqrt{\SC}} \tilde{L}_{n-1}(-1) (\E - \SC^{-1} L(0)(\varphi) + \sigma (d\SC^{-1},d\varphi)) \\ & = \tilde{\QC}_n - \iota^* \frac{1}{\sqrt{\SC}} \tilde{L}_{n-1}(-1) (\SC^{-1} L(0)(\varphi) - \sigma (d\SC^{-1},d\varphi)). \end{align*} This proves the conformal transformation law \begin{equation}\label{CTL-tilde} e^{n \iota^*(\varphi)} \hat{\tilde{\QC}}_n = \tilde{\QC}_n - \iota^* \tilde{\PO}_n(\varphi) \end{equation} \tilde{\PO}_n \st \frac{1}{\sqrt{\SC}} \tilde{L}_{n-1} (-1) ( \SC^{-1} L(0)(\cdot) - \sigma (d\SC^{-1},d\cdot)). Note that the latter operator again is conformally covariant: $e^{n\varphi} \tilde{\PO}_n(\hat{g},\hat{\sigma}) = \tilde{\PO}_n(g,\sigma)$. This immediately follows from the conformal covariance of $L$ and the invariance of $\SC$. The following result generalizes Corollary <ref> and clarifies the content of <cit.>. For closed $M^n$, the integral \int_M \tilde{\QC}_n (g,\sigma) dvol_h is invariant under conformal changes of the pair $(g,\sigma)$. We give two proofs. The first proof uses an argument of <cit.>. Since the integral of $\tilde{\QC}_n$ is proportional to the anomaly $\LA$ in Theorem <ref>, it suffices to prove that $\LA(\hat{g},\hat{\sigma}) = \LA(g,\sigma)$. We recall that $\hat{\sigma}^{-2} \hat{g} = \sigma^{-2} g$ and prove that the expansion of the difference \int_{\hat{\sigma} > \varepsilon} dvol_{\sigma^{-2} g} - \int_{\sigma > \varepsilon} dvol_{\sigma^{-2} g} does not contain a $\log \varepsilon$ term. Note that $\{\hat{\sigma} > \varepsilon \} = \{\sigma > \varepsilon e^{-\varphi} \}$. Now, using adapted coordinates, we find that the above difference equals \begin{align*} & \int_{\varepsilon}^{\varepsilon e^{-\varphi}} s^{-n-1} \left( \int_M v(s) dvol_h \right) ds \\ & = \sum_{k=0}^{n-1} \varepsilon^{-n+k} \int_M \frac{v_k}{n-k} \left(1-e^{(n-k)\varphi}\right) dvol_h - \int_M \varphi v_n dvol_h + o(1). \end{align*} This expansion does not contain a $\log \varepsilon$ term. The second proof rests on the conformal transformation law (<ref>). It shows that it suffices to prove that $\int_M \iota^* \tilde{\PO}_n(\varphi) dvol_h = 0$ for all $\varphi \in C^\infty(X)$. Now the generalization[With appropriate interpretations as in Remark <ref>, the relation (<ref>) follows from <cit.>.] \begin{equation}\label{extension} \int_M \iota^* \frac{1}{|\NV|} \tilde{L}_{n-1}(-1)(\psi) dvol_h \sim \int_M \iota^* \partial_s^{n-1}(v \psi) dvol_h \end{equation} \begin{align*} \langle \sigma^*(\delta), L_{n-1}(-1) (\psi) dvol_g \rangle \sim \int_M \iota^* \partial_s^{n-1} (v \psi) dvol_h \end{align*} (see (<ref>)) shows that the assertion is equivalent to \begin{align}\label{van2} \int_M \iota^* \partial_s^{n-1} \left( v \left[\SC^{-1} L(0)(\varphi) - s (d\SC^{-1},d\varphi)_g \right] \right) dvol_h= 0. \end{align} We recall that $L(0)(\varphi) = (n-1) a \partial_s (\varphi) - s \Delta_g(\varphi)$ with $a = |\NV|^2$ and calculate \begin{align*} & \iota^* \partial_s^{n-1} \left( v f L(0)(\varphi) \right) \\ & = (n-1) \iota^* \partial_s^{n-1} (v f a \varphi') - (n-1) \iota^* \partial_s^{n-2} (v f \Delta_g (\varphi)) \\ & = (n-1) \iota^* \partial_s^{n-2} ( v' f a \varphi' + v f' a \varphi' + v f a' \varphi' + v f a \varphi'') \\ & - (n-1) \iota^* \partial_s^{n-2}\left(v f \left (a \varphi'' + a \frac{1}{2} \tr (h_s^{-1} h_s') \varphi' + \frac{1}{2} a' \varphi' - \frac{1}{2} (d \log a,d\varphi)_{h_s} + \Delta_{h_s}(\varphi) \right)\right) \end{align*} using (<ref>). Simplification of that result gives (n-1) \iota^* \partial_s^{n-2} \left( v a f' \varphi' + v f \frac{1}{2} (d \log a, d \varphi)_{h_s} - v f \Delta_{h_s} (\varphi)\right). Here we used that $v = a^{-\frac{1}{2}} dvol_{h_s}/dvol_h = a^{-\frac{1}{2}} \mathring{v}$ implies \frac{\mathring{v}'}{\mathring{v}} = \frac{1}{2} \tr (h_s^{-1} h_s') = \frac{v'}{v} + \frac{1}{2} \frac{a'}{a} (see (<ref>)). Thus, we obtain \begin{align*} & \int_M \iota^* \partial_s^{n-1} \left( v f L(0)(\varphi) \right) dvol_h \\ & = (n-1) \iota^* \partial_s^{n-2} \left( \int_M v \left(a f' \varphi' + f \frac{1}{2} (d \log a, d \varphi)_{h_s} - f \Delta_{h_s}(\varphi)\right) dvol_h \right). \end{align*} Now partial integration on $M$ yields \begin{align*} \int_M v f \Delta_{h_s}(\varphi) dvol_h & = \int_M a^{-\frac{1}{2}} f \Delta_{h_s}(\varphi) dvol_{h_s} = - \int_M (d (a^{-\frac{1}{2}} f, d\varphi)_{h_s} dvol_{h_s} \\ & = - \int_M v a^{\frac{1}{2}} (d (a^{-\frac{1}{2}} f), d\varphi)_{h_s} dvol_h \\ & = - \int_M v (d f,d\varphi)_{h_s} dvol_h + \frac{1}{2} \int_M v f (d \log a,d\varphi)_{h_s} dvol_h \end{align*} using $a^{\frac{1}{2}} (d (a^{-\frac{1}{2}} f), d\varphi)_{h_s} = - \frac{1}{2} f (d \log a, d\varphi)_{h_s} + (d f,d\varphi)_{h_s}$. Differentiating this relation by $s$ implies \begin{align*} \int_M \iota^* \partial_s^{n-1} \left( v f L(0)(\varphi) \right) dvol_h & = (n-1) \iota^* \partial_s^{n-2} \left( \int_M v a f' \varphi' + v (df,d\varphi)_{h_s} dvol_h \right) \\ & = (n-1) \iota^* \partial_s^{n-2} \left( \int_M v (df,d\varphi)_g dvol_h \right). \end{align*} For $f = \SC^{-1}$, this identity implies the vanishing of (<ref>). Note that similar arguments show that L(-n) ( f \sigma^*(\delta^{(n-1)})) = L(0)(f) \sigma^*(\delta^{(n-1)}) for $f \in C^\infty(X)$ (see <cit.>). We omit the details. A calculation shows that in adapted coordinates \begin{equation}\label{TQ} \tilde{\QC}_2(g,\sigma) = \iota^* (|\NV|^{-5} ( \rho^2 - \rho \SC' - 2 (\SC')^2) + |\NV|^{-3} (\J - \rho' + \SC'')). \end{equation} In particular, if $\iota^* \SC'$ and $\iota^* \SC''$ vanish (singular Yamabe case), then $\QC_2 = \iota^* (\rho^2 - \rho' + \J)$. By $\iota^* (\rho) = -H$ and $\iota^* (\rho') = \Rho_{00} + |\lo|^2$ (Lemma <ref>), we get $\QC_2 = - \Rho_{00} - |\lo|^2 + \iota^*(\J)$. Now the Gauss equation $\iota^* \J = \J^h + \Rho_{00} + \frac{1}{2} |\lo|^2 - H^2$ implies \QC_2 = \J^h - \frac{1}{2} |\lo|^2. This is the known formula for the critical extrinsic $Q$-curvature in dimension $n=2$ (Example <ref>). If $\sigma= d_M$ is the distance function, then $|\NV|=1$ and geodesic normal coordinates are adapted coordinates. Now $\SC = 1 + 2 s \rho$ implies $\iota^*(\SC') = 2 \iota^*(\rho)$ and $\iota^*(\SC'') = 4 \iota^*(\rho')$. Thus, we get \tilde{\QC}_2 = \iota^* (\J + 3 \rho' - 9 \rho^2). But $\iota^*(\rho) = - \frac{2}{3} H$ and $3 \iota^*(\rho') = - \iota^* \partial_s (\Delta_g (s) + s \J) = - \iota^* (2 H' + \J) = \iota^* (\Ric_{00} - \J) + |L|^2 = \Rho_{00} + |L|^2$ (using $2 H' = -|L|^2 - \Ric_{00}$ - see (<ref>)) show that $\tilde{\QC}_2 = \Ric_{00} + |\lo|^2 - 2 H^2$. This confirms the formula $-2 \LA = \int_M u_2 dvol_h$ (with $u_2$ as in (<ref>)). Finally, we use similar arguments as above to establish analogous formulas for the integrated renormalized volume coefficients $w_k$, which are defined in terms of geodesic normal coordinates. Let $d_M$ be the distance function of $M$. Let $M$ be closed. Assume that $\sigma$ satisfies $\SCY$. Then \int_M w_k dvol_h = (-1)^k \frac{(n\!-\!1\!-\!k)!}{(n\!-\!1)! k!} \int_M \iota^* L_k(g,\sigma;-n\!+\!k) \left( \left(\frac{d_M}{\sigma}\right)^{n-k}\right) dvol_h for $0 \le k \le n-1$ and \int_M w_n dvol_h = \frac{(-1)^{n-1} }{(n\!-\!1)! n!} \int_M \iota^* \dot{L}_n(g,\sigma;0)(1) dvol_h. We compare two different calculations of the residues of the family \lambda \mapsto \int_X \sigma^\lambda \psi dvol_g, \; \Re(\lambda) > -1 for appropriate test functions $\psi$. On the one hand, arguments as above using the functional equation (<ref>) prove the relation \begin{equation}\label{res-cal-1} \Res_{\lambda=-k-1} \left( \int_X \sigma^\lambda \psi dvol_g \right) = \frac{1}{k!(-n\!+\!1)_k} \int_M \iota^* L_k(g,\sigma;-n\!+\!k)(\psi) dvol_h. \end{equation} On the other hand, we use geodesic normal coordinates and asymptotic expansions of the resulting integrand. By $\kappa^*(d_M)=r$ and \kappa^* (dvol_g) = \left( \frac{\kappa^*(\sigma)}{r}\right)^{n+1} w(r) dr dvol_h (see (<ref>)), we obtain \begin{align*} \int_X \sigma^\lambda \psi dvol_g & = \int_X d_M^\lambda \left(\frac{\sigma}{d_M}\right)^\lambda \psi dvol_g \\ & = \int_{[0,\varepsilon)} \int_M r^\lambda \left( \frac{\kappa^*(\sigma)}{r}\right)^{\lambda+n+1} \kappa^*(\psi) w(r) dr dvol_h \end{align*} for test functions $\psi$ with appropriate support. The classical formula (<ref>) implies that \Res_{\lambda=-k-1} \left( \int_X \sigma^\lambda \psi dvol_g \right) = \frac{1}{k!} \int_M \left( \left(\frac{\kappa^*(\sigma)}{r}\right)^{-k+n} \kappa^*(\psi) w(r)\right)^{(k)}(0) dvol_h. For the test function $\psi_k = (d_M/\sigma)^{n-k} \chi$ (with an appropriate cut-off function $\chi$), the latter result yields \begin{equation}\label{res-cal-2} \Res_{\lambda=-k-1} \left( \int_X \sigma^\lambda \psi_k dvol_g \right) = \int_M w_k dvol_h. \end{equation} Now comparing (<ref>) and (<ref>) proves the first assertion. The assertion in the critical case follows from $\int_M w_n dvol_h = \int_M v_n dvol_h$ (see (<ref>)) and Theorem <ref>. The first part of Theorem <ref> is a special case of <cit.>. § HOLOGRAPHIC FORMULÆ FOR EXTRINSIC $Q$-CURVATURES We work in adapted coordinates. In particular, $\J=\J^g$ is identified with $\eta^*(\J)$. The following result is a local version of Theorem <ref>. Let $n$ be even and assume that $\sigma$ satisfies $\SCY$. Then \begin{equation}\label{Q-holo-form} \QC_n (g) = (-1)^n (n\!-\!1)!^2 \sum_{k=0}^{n-1} \frac{1}{n\!-\!1\!-\!2k} \T_k^*(g;0) \left( (n\!-\!1)(n\!-\!k) v_{n-k} + 2k (v \J)_{n-k-2} \right). \end{equation} For $k=n-1$, the second term on the right-hand side is defined as $0$. The assumption that $n$ is even guarantees that the fractions on the right-hand side are well-defined. For odd $n$, we refer to Conjecture <ref>. Assume that $g_+ = r^{-2}(dr^2 + h_r)$ is an even Poincaré-Einstein metric in normal form relative to $h$ and let $g = r^2 g_+ = dr^2 + h_r$. Then \J^g= - \frac{1}{2r} \tr (h_r^{-1} h'_r) = - \frac{1}{r} \frac{v'}{v}(r) by <cit.> or Lemma <ref>, and the second term on the right-hand side of (<ref>) yields $-2k (n-k) v_{n-k}$. Therefore, the sum simplifies to \sum_{k=0}^{n-1} (n-k) \T_k^*(0) (v_{n-k}) and the assertion reduces to the main result of <cit.> by noting that it only contains contributions for even $k$. Note that (<ref>) can be written in the form \QC_n = (-1)^{n} n! (n\!-\!1)! v_n + \sum_{k=1}^{n-1} \T_k^*(0)(\cdots). Since $\T_k(0)(1) = 0$ for $k \ge 1$, integration of this identity (for closed $M$) reproduces Theorem <ref>, and the holographic formula provides a formula for the lower-order terms. There is a generalization of Theorem <ref> to subcritical $Q$-curvatures. Let $n$ be even and $\N \ni N < n$. Assume that $\sigma$ satisfies $\SCY$. Then \begin{align}\label{Q-holo-form-gen} \QC_N(g) & = (-1)^{N} (N\!-\!1)!^2 \sum_{k=0}^{N-1} \frac{1}{(2N\!-\!n\!-\!1\!-\!2k)} \notag \\ & \times \T_k^*\left(g;\frac{n-N}{2}\right) \left( (N\!-\!1)(N\!-\!k) v_{N-k} + (n\!-\!N\!+\!2k) (v \J)_{N-k-2} \right). \end{align} For $k=N-1$, the second term on the right-hand side is defined as $0$. Again, the assumption that $n$ is even guarantees that the fractions on the right-hand side are well-defined. For odd $n$, we refer to Conjecture <ref>. In the even Poincaré-Einstein case, the formula is non-trivial only for even $N$. In that case, the solution operators in (<ref>) act on (N\!-\!1)(N\!-\!k) v_{N-k} - (n\!-\!N\!+\!2k) (N\!-\!k) v_{N-k} = (2N\!-\!n\!-\!1\!-\!2k)(N\!-\!k) v_{N-k} and the sum simplifies to \sum_{k=0}^{N-1} \T_k^*\left(g;\frac{n-N}{2}\right) (N-k) v_{N-k}. Here only terms with even $k$ contribute. Thus the formula reduces to the main result of <cit.>. The following conjecture implies that for odd $n$ the respective terms with the singular fractions in the sums (<ref>) and (<ref>) do not contribute. For odd $n\ge 3$, it holds \begin{equation}\label{van-id} \tfrac{n+1}{2} v_{\frac{n+1}{2}} + (v \J)_{\frac{n-3}{2}} = 0. \end{equation} Moreover, the holographic formulas (<ref>) and (<ref>) are valid also for odd $n$. For $n=3$ and $n=5$, the respective relations (<ref>) read \begin{align*} 2v_2 + \J_0 & = 0, \\ 3 v_3 + \J'_0 + v_1 \J_0 & = 0. \end{align*} For details, we refer to the discussion in Examples <ref>–<ref>. These identities are consequences of the relation (<ref>). More generally, the relation (<ref>) implies that any coefficient $v_N$ can be written as a linear combination of products of derivatives of $\rho$ and $\J$ at $s=0$. The calculations of the resulting formulas for odd $n \le 11$ confirm the relation (<ref>) in these special cases. For the discussion of the holographic formula for $\QC_3$ in general dimensions, we refer to Example <ref>. Now we present the proof of Theorem <ref>. We evaluate the quantity $\dot{\D}_n^{res}(0)(1)$ using the factorization formula \D_n^{res}(\lambda) = \D_{n-1}^{res}(\lambda-1) L(\lambda) (Corollary <ref>). In view of $L(0)(1)=0$, it follows that \begin{equation}\label{start-crit} \dot{\D}_n^{res}(0)(1) = \D_{n-1}^{res}(-1) \dot{L}(0)(1). \end{equation} Now we apply the representation formula \D_{n-1}^{res}(-1) = (-1)^{n-1} (n\!-\!1)!^2 \sum_{j=0}^{n-1} \frac{1}{(n\!-\!1\!-\!j)!} \left[ \T_j^*(0) v_0 + \cdots + \T_0^*(0) v_j\right] \iota^* \partial_s^{n-1-j} (Theorem <ref>). We find \dot{\D}_n^{res}(0)(1) = (-1)^{n-1} (n-1)!^2 \sum_{j=0}^{n-1} \frac{1}{(n\!-\!1\!-\!j)!} \left[ \T_j^*(0) v_0 + \cdots + \T_0^*(0) v_j\right] \iota^* \partial_s^{n-1-j} (\dot{L}(0)(1)). In the latter sum, the operator $\T_k^*(0)$ acts on the sum \left(v_0 \frac{1}{(n\!-\!1\!-\!k)!} \iota^* \partial_s^{n-1-k} + \cdots + v_{n-1-k} \iota^* \right)(\dot{L}(0)(1)). But this quantity equals the $(n\!-\!1\!-\!k)$'th Taylor coefficient $(v \dot{L}(0)(1))_{n-1-k}$ of $v \dot{L}(0)(1)$. Now the \begin{equation}\label{reduction} \iota^* \partial_s^k (v \dot{L}(0)(1)) = -\frac{n\!-\!1}{(n\!-\!1\!-\!2k)} \iota^* \partial_s^{k+1}(v) - \frac{2k(n\!-\!1\!-\!k)}{(n\!-\!1\!-\!2k)} \iota^* \partial_s^{k-1}(v\J) \end{equation} for $0 \le k \le n-1$ (which extends (<ref>))[For $k=0$, the second term on the right-hand side is defined as $0$.] yields \begin{align*} (v \dot{L}(0)(1))_{n-1-k} & = \frac{1}{(n\!-\!1\!-\!k)!} \iota^* \partial_s^{n-1-k} (v \dot{L}(0)(1)) \\ & = \frac{n\!-\!1}{(n\!-\!1\!-\!2k)(n\!-\!1\!-\!k)!} \iota^* \partial_s^{n-k} (v) + \frac{2k (n\!-\!1\!-\!k)}{(n\!-\!1\!-\!2k)(n\!-\!1\!-\!k)!} \iota^* \partial_s^{n-2-k}(v\J) \\ & = \frac{(n\!-\!1)(n\!-\!k)}{(n\!-\!1\!-\!2k)} v_{n-k} + \frac{2k}{(n\!-\!1\!-\!2k)} (v\J)_{n-2-k} \end{align*} for $0 \le k \le n-1$. This implies the claim. It only remains to prove the identity (<ref>). Note that $\dot{L}(0)(1) = (n-1)\rho-s\J$. Using the definition of $\rho$, we obtain \begin{equation}\label{d-delta} \iota^* \partial_s^k (v\dot{L}(0)(1)) = - \frac{n-1}{n+1} \left( \iota^* \partial_s^k(v \Delta (s)) + \frac{2kn}{n-1} \iota^* \partial_s^{k-1}(v\J)\right). \end{equation} Now the relation (<ref>) implies \begin{align*} \iota^* \partial_s^k(v \Delta (s)) & = \iota^* \partial_s^{k+1}(v |\NV|^2) \\ & = \iota^* \partial_s^{k+1}(v) - 2 (k+1) \iota^* \partial_s^k(v \rho) \\ & = \iota^* \partial_s^{k+1}(v) + \frac{2(k+1)}{n+1} \iota^* \partial_s^k(v \Delta(s) + s v\J) \end{align*} for $0 \le k \le n-1$ using $|\NV|^2 = 1-2s\rho + O(s^{n+1})$. We rewrite this identity in the form \frac{n-2k-1}{n+1} \iota^* \partial_s^k(v \Delta (s)) = \iota^* \partial_s^{k+1}(v) + \frac{2k(k+1)}{n+1} \iota^* \partial_s^{k-1}(v \J). By substituting this result into (<ref>), we obtain (<ref>). Note that the above arguments extend those in the proof of Theorem <ref>. Finally, we sketch a proof of Theorem <ref>. For $N < n$, we have \D_N^{res} \left(\frac{-n+N}{2}\right) (1) = \left(\frac{n-N}{2}\right) \QC_N by (<ref>). Hence \begin{align*} \QC_N & = \D_N^{res}\left(\frac{-n+N}{2}\right) (1) \left(\frac{n-N}{2}\right)^{-1} \\ & = \D_{N-1}^{res}\left(\frac{-n+N}{2}-1\right) L\left(\frac{-n+N}{2}\right) (1) \left(\frac{n-N}{2}\right)^{-1} & \mbox{(by Corollary \ref{factor})} \\ & = \D_{N-1}^{res}\left(\frac{-n+N}{2}-1\right) (-(N-1) \rho + s \J) & \mbox{(by \eqref{aa})}. \end{align*} This formula generalizes (<ref>). Now we proceed as in the proof of Theorem <ref>. In particular, Theorem <ref> and the above identity imply \begin{align*} \QC_N & = (-1)^{N} (N\!-\!1)!^2 \\ & \times \sum_{j=0}^{N-1} \frac{1}{j!} \left[ \T_{N-1-j}^*\left(\frac{n\!-\!N}{2}\right) v_0 + \cdots + \T_0^*\left(\frac{n\!-\!N}{2}\right) v_j \right] \iota^* \partial_s^j ((N\!-\!1)\rho-s\J). \end{align*} But in the latter double sum the operator $\T_k^*\left(\frac{n-N}{2}\right)$ acts on the $(N-1-k)$'th Taylor coefficient of $((N-1)\rho-s\J)v$. A calculation using the definition of $\rho$ yields the extension \begin{equation}\label{reduction-2} \iota^* \partial_s^k (v ((N\!-\!1)\rho - s\J)) = - \frac{N\!-\!1}{(n\!-\!1\!-\!2k)} \iota^* \partial_s^{k+1}(v) - \frac{k(n\!-\!2k\!+\!N\!-\!2)}{(n\!-\!1\!-\!2k)} \iota^* \partial_s^{k-1}(v\J) \end{equation} of (<ref>) for $k=0,\dots,N-1$.[For $k=0$, the second term on the right-hand side is defined as $0$.] Hence \begin{equation*} (v ((N\!-\!1)\rho-s\J))_{N-1-k} = - \frac{(N\!-\!1)(N\!-\!k)}{(n\!-\!2N\!+\!2k\!+\!1)} v_{N-k} - \frac{n\!-\!N\!+\!2k}{(n\!-\!2N\!+\!2k\!+\!1)} (v\J)_{N-2-k} \end{equation*} for $k=0,\dots,N-1$. This implies the assertion. $\square$ We finish this section with an application to the singular Yamabe obstruction $\B_{n}$ (see Section <ref>). We recall that our discussion of extrinsic conformal Laplacians $\PO_N$ and extrinsic $Q$-curvatures $\QC_N$ is restricted to the range $N \le n$. Already the first super-critical $\QC$-curvature $\QC_{n+1}$ is not well-defined. Calculations in low-order cases point to the origin of its non-existence: $\QC_N$ has a pole in $n=N-1$. The following result interprets its residue. Let $n$ be even and $N \ge 3$. Then it holds \begin{equation}\label{QB-F} \Res_{n=N-1} (\QC_N) = (-1)^{n-1} n! (n\!+\!1)! \frac{n}{2} \B_n. \end{equation} Assume that $N \le n$. Theorem <ref> and Theorem <ref> imply that \QC_N = (-1)^N (N\!-\!1)! N! \frac{N\!-\!1}{2N\!-\!n\!-\!1} v_N + \cdots, where the hidden terms are regular at $n=N-1$. However, the term $v_N$ has a simple pole at $n=N-1$. More precisely, it follows from Corollary <ref> that v_N = -(n\!+\!1\!-\!2N) \frac{1}{N!} \partial^{N-1}_s (\rho)|_0 + \cdots, where the hidden terms are regular at $n=N-1$, and Proposition <ref> explains the origin of the pole of $\partial^{N-1}_s (\rho)|_0$. But a comparison of Proposition <ref> and Theorem <ref> shows that[See also <cit.>.] \begin{equation}\label{B-rho} \B_{n} = \frac{2}{(n+1)!} \Res_{n=N-1} (\partial_s^{N-1}(\rho))|_0. \end{equation} The result follows from these facts. For a detailed discussion of the relation between $\B_2$ and $\QC_3$, we refer to Section <ref>. This proof rests on the key relation between $Q$ and $v$ (holographic formula)! Some possible additions for later versions: 1. the above proof requires some improvements: we should emphasize that we do a continuation in dimension argument. 2. There is no difference between taking partials in $s$ or partials in $N$: lower-order derivatives of $\rho$ are not singular. 3. the arguments also give the conformal invariance of $\B_n$: this requires that the pole of $\PO_{n+1}$ is only in the constant term. Combining Theorem <ref> with $\Res_{n=N-1} (\PO_N) \sim \Res_{n=N-1} (\QC_N)$ allows us to derive the conformal invariance $e^{(n+1) \iota^*(\varphi)} \hat{\B}_{n} = \B_{n}$ of the obstruction from the conformal covariance of $\PO_N$. The above observations resemble the result that the residues \Res_{n=4}(P_6) = -16 (\delta (\B d) - (\B,\Rho)) \Res_{n=6}(P_8) = -48 (\delta(\OB_6 d) - (\OB_6,\Rho)) of super-critical GJMS-operators $P_6$ and $P_8$ are conformally covariant. Here $\B$ is the Bach tensor, and $\OB_6$ is the Fefferman-Graham obstruction tensor in dimension $n=6$. For more details, we refer to <cit.>. § COMMENTS ON FURTHER DEVELOPMENTS We recall that $\PO_n \iota^* = \D_n^{res}(0)$. For even $n$, this critical extrinsic conformal Laplacian is elliptic. For odd $n$, the leading part of this operator is determined by $\lo$. The lower-order terms are not known in general. For $n=3$, the explicit formula in Proposition <ref> shows that $\PO_3$ vanishes iff $\lo=0$, i.e., iff $M$ is totally umbilic. The vanishing of $\PO_n$ is a conformally invariant condition. Now if $\PO_n(g) = 0$, we define \begin{equation} \PO_n'(g) = \dot{\D}_n^{res}(g;0) \quad \mbox {and} \quad \QC_n'(g) = \frac{1}{2} \ddot{\D}_n^{res}(g;0). \end{equation} Then $\PO_n': C^\infty(X) \to C^\infty(M)$ is a conformally covariant operator e^{n \iota^*(\varphi)} \PO_n'(\hat{g}) = \PO_n'(g), \; \varphi \in C^\infty(X) such that the pair (\PO_n', \QC_n') satisfies the fundamental identity \begin{equation}\label{FI-prime} e^{n \iota^*(\varphi)} \QC_n'(\hat{g}) = \QC_n'(g) - \PO_n'(g)(\varphi) + \iota^*(\varphi) \PO_n'(g)(1) \end{equation} for all $\varphi \in C^\infty(X)$. This identity follows by twice differentiation of the conformal transformation law of the critical residue family $\D_n^{res}(\lambda)$ at $\lambda=0$. Note that $\PO_n'(g)(1) = - \QC_n(g)$. It is interesting to analyze this construction further. First of all, it is a question of independent interest to characterize the vanishing of $\PO_n(g)$ for odd $n$. Is it true that $\QC_n(g)(1) = 0$? The operator $\PO_n$ may be regarded as a boundary operator for a conformally covariant boundary value problem for the critical GJMS-operator $P_{n+1}$ on $X$. For $n=3$, such boundary value problems were recently analyzed in <cit.>. A conformally covariant boundary operator of third order together with a $Q$-curvature like scalar curvature quantity was discovered in <cit.> in connection with the study of Polyakov formulas on four-dimensional manifolds with boundary. Later it was studied from various perspectives. For details, we refer to <cit.> and the references in these works. It would be interesting to develop an extension of the present theory in higher codimension situations again using solutions of singular Yamabe problems. We briefly describe some aspects of a special case of such a theory. Let $S^m \hookrightarrow S^n$ be an equatorial subsphere of $S^n$, $m \le n-1$. It is well-known that the complement of $S^m$ in $S^n$ with the round metric $g$ is conformally equivalent to the product $\mathbb{H}^{m+1} \times S^{n-m-1}$ with the respective hyperbolic and round metric of constant scalar curvature $-(m+1)(m+2)$ and $(n-m-1)(n-m)$ on the factors. The conformal factor $\sigma \in C^\infty(S^n\setminus S^m)$ can be defined in terms of a Knapp-Stein intertwining operator on $S^n$ applied to the delta-distribution of $S^m$ <cit.>. More explicitly, assume that $S^m$ is defined by the equations $x_{m+2} = \cdots = x_{n+1}=0$. Then $\sigma$ is the restriction of $(\sum_{j=m+2}^{n+1} x_j^2)^{1/2}$ to $S^n$. A stereographic projection yields an isometry of $\sigma^{-2} g$ and (x_{m+1}^2 + \cdots + x_{n}^2)^{-1} \sum_{i=1}^n dx_i^2 = r^{-2} \sum_{i=1}^{m+1} dx_i^2 + g_{S^{n-m-1}}, where $r^2 = \sum_{i=m+1}^{n} x_i^2$. The above $\sigma$ is a generalization of the height function $\He$ in Section <ref> (the case $m=n-1$). Now, for any eigenfunction $\tau$ of the Laplacian on $S^{n-m-1}$ (spherical harmonics), there is an intertwining operator D_{N,\tau}(\lambda): C^\infty(S^n) \to C^\infty(S^{m}) for the subgroup $SO(1,m+1) \subset SO(1,n+1)$ leaving $S^m$ invariant (symmetry breaking operator). These families can be constructed in terms of the residues of the family \lambda \mapsto \int_{S^n} \sigma^\lambda u \psi dvol_{S^n}, where $u$ is an eigenfunction of $\Delta_{\sigma^{-2}g}$ on the complement of $S^m$ which is compatible with $\tau$. The conformal factor $\sigma$ is a solution of the singular Yamabe problem on the complement of $S^m$. In fact, the scalar curvature of $\sigma^{-2}g$ is $-(n+1)(2m-n+2)$. It is negative iff $m > (n-2)/2$, i.e., iff the dimension of the hyperbolic space exceeds the dimension of the sphere. This is a special case of <cit.>, where it is proved that if $M \subset S^n$ is a smooth submanifold of dimension $m$, a solution of the singular Yamabe problem with negative scalar curvature exists iff $m > (n-2)/2$. The analogous result for $S^n$ replaced by an arbitrary $X$ is in <cit.>. For a description of the asymptotic expansions of solutions of the singular Yamabe problem in the negative case, we refer to <cit.>. Related representation theoretical aspects of the case $S^m \hookrightarrow S^n$ are studied in <cit.>. Here the spectral decomposition of spherical principal series representations of $O(1,n+1)$ under restriction to $O(1,m+1) \times O(n-m)$ is made fully explicit. The conformal invariance of the integral \int_M \QC_n(g) dvol_h for closed $M$ (see (<ref>)) generalizes the conformal invariance of the total Branson \int_M Q_n(h) dvol_h. The classification of scalar Riemannian curvature quantities of a manifold $(M,h)$ which - like $Q_n$ - upon integration give rise to a conformal invariant has been the subject of the Deser-Schwimmer conjecture <cit.>. S. Alexakis has achieved this classification in a series of works (see <cit.> and its references). The present context suggests asking for an analogous classification of scalar Riemannian invariants of a manifold $(X,g)$ which, upon integration over closed submanifolds, yield conformal invariants of $g$. For first results in that direction, we refer to <cit.>. For related results around ${\bf Q}_4$, we refer to <cit.> and <cit.>. § CALCULATIONS AND FURTHER RESULTS theoremsubsection equationsubsection In the present section, we collect formulas and computational details used in the main body of the text. In addition, we illustrate various aspects of the general theory by low-order examples. This often gives additional insight into the situation's nature and complexity. The results may also serve as material for future research. Moreover, we add a few further results. For the reader's convenience, we start with an outline of the content of this section. In Section <ref>, we recall basic forms of the hypersurface Gauss equations, recall the transformation laws of the second fundamental form and of some derived constructions under conformal changes of the metric, derive a basic formula for the Laplacian, which plays a key role in the proof of the conjugation formula and clarify the relation between high-order iterated normal derivatives $\nabla_\NV^k$ and their analogs in adapted coordinates. In Section <ref>, we determine the first three terms in the expansion of a general metric $g$ in geodesic normal and adapted coordinates. These results are fundamental for the later calculations. We also illustrate the results in several model cases with constant curvature and vanishing trace-free part of $L$. These model cases may serve as valuable test examples of identities of the general theory. In Section <ref>, we first determine the first four terms in the asymptotic expansion of solutions of the singular Yamabe problem in geodesic normal coordinates. Then, we use these results to calculate the obstructions $\B_2$ and $\B_3$ in these terms. Finally, we prove a formula for the obstructions in terms of a formal residue of the super-critical coefficient $\sigma_{(n+2)}$. The explicit form of the first few coefficients $\sigma_{(k)}$ enables us to confirm this formula in low-order cases. In Section <ref>, we derive explicit formulas for the first three renormalized volume coefficients $v_k$ (in adapted coordinates). In Section <ref>, we derive explicit formulas for the first two normal derivatives of $\rho$. The discussion illustrates the efficiency of the recursive relation for the Taylor coefficients of $\rho$ expressed in Proposition <ref>. We further use these results in Section <ref> for a detailed discussion of $\QC_2$ and $\QC_3$. In Section <ref>, we show that the well-known formula for the obstruction $\B_2$ naturally follows from Theorem <ref>. Moreover, in Section <ref>, we evaluate the special case $n=3$ of Theorem <ref> for a flat background metric. We find that the formula coincides with the one derived in Section <ref> as well with a formula of Gover and Waldron. In addition, we verify that in the conformally flat case, the result fits with the formula for $\B_3$ established in [3]. In Section <ref>, we illustrate the role of the obstruction in the variational formula for the singular Yamabe energy <cit.>, <cit.> in low-order cases. In <cit.> and [3], the authors developed a new variational calculus. In contrast to these references, here we only use classical style arguments as in <cit.>, for instance. In Section <ref>, we provide direct proofs of the conformal covariance of $\PO_3$ and the fundamental conformal transformation law of $\QC_3$. Section <ref> is devoted to a derivation of explicit formulas for the first two solution operators $\T_1(\lambda)$ and $\T_2(\lambda)$ and the resulting first two residue families $\D_1^{res}(\lambda)$ and $\D_2^{res}(\lambda)$. In addition, we determine the leading term of $\PO_3$ through the leading term of the third solution operator $\T_3(\lambda)$. In the last section, we describe low-order renormalized volume coefficients $w_k$ in terms of Laplace-Robin operators. These results imply low-order cases of Theorem <ref>. Throughout this section, we apply some additional conventions. We use indices $i,j$ for tensorial objects on $M$ and $0$ for $\NV$ when viewed as a normal vector of $M$. In particular, $h_{ij}$ are the components of the metric $h$ on the boundary and $\Rho_{00} = \Rho_X (\NV,\NV)$ is the restriction to $M$ of the Schouten tensor of $g$ for the normal vector $\NV$. Similar conventions are used in adapted coordinates. Sometimes, it will be convenient to distinguish curvature quantities of $g$ and $h$ not by superscripts but by adding a bar to those of $g$ and leaving those of $h$ unbared. Then $\bar{R}$, $\overline{\Ric}$ and $\overline{\scal}$ are the curvature tensor, the Ricci tensor and the scalar curvature of $g$, respectively. For instance, $\Ric^g(\NV,\cdot)$ and $\overline{\Ric}_0$ are the same $1$-forms on $M$. As before, we often use the same notation for a function on $X$ and its pull-back by $\kappa$ or $\eta$ without mentioning it. A prime denotes derivatives in the variable $r$ and $s$. The restriction of a function $f(s)$ to $s=0$ is also denoted by $f_0$. This notation often replaces $\iota^* (f)$. For instance, we write $\rho'_0$ for the restriction $\iota^*(\partial_s (\rho))$ of $\partial_s(\rho)$ to $s=0$. If confusion is excluded, we sometimes even omit the symbols indicating restriction. §.§ Some basic identities §.§.§ Gauss equations Let $M^n$ be a hypersurface in $(X^{n+1},g)$ with the induced metric $h=\iota^*(g)$. Then it holds \begin{align} \Ric^g_{ij} - \Ric^h_{ij} & = R^g_{0ij0} + L^2_{ij} - n H L_{ij}, \label{GRicci} \\ \scal^g - \scal^h & = 2 \Ric^g_{00} + |L|^2 - n^2 H^2 \label{G1} \end{align} on $M$. In the bar-notation, (<ref>) reads \iota^* \overline{\scal} - \scal = 2 \overline{\Ric}_{00} +|L|^2 - n^2 H^2. The following result follows from the Gauss equation (<ref>). For $n \ge 2$, it holds \begin{equation}\label{G2} \iota^* \J^g - \J^h = \Rho^g_{00} + \frac{1}{2(n\!-\!1)} |\lo|^2 - \frac{n}{2} H^2. \end{equation} The Gauss equation (<ref>) is equivalent to 2n \iota^* \J^g - 2(n\!-\!1) \J^h = 2((n\!-\!1) \Rho^g_{00} + \J^g) + |L|^2 - n^2 H^2. 2(n\!-\!1) \iota^* \J^g - 2(n\!-\!1) \J^h = 2(n\!-\!1)\Rho^g_{00} + |\lo|^2 - n(n-1) H^2. This implies the assertion. $\JF$ Fialkow tensor Next, let \begin{equation}\label{Fialkow-tensor} \JF \st \iota^* (\Rho^g) - \Rho^h + H \lo + \frac{1}{2} H^2 h. \end{equation} $\JF$ is a conformally invariant symmetric bilinear form, i.e., it holds $\hat{\JF} = \JF$ (<cit.>.[In <cit.>, the tensor $\JF$ is called the Fialkow tensor following <cit.> (in turn being inspired by <cit.>). For more details on the relation to Fialkow's classical work, we refer to <cit.>.] Moreover, it satisfies the fundamental relation \begin{align}\label{FW-relation} (n\!-\!2) \JF_{ij} & = (n\!-\!2) \left(\iota^* \Rho^g_{ij} - \Rho^h_{ij} + H \lo_{ij} + \frac{1}{2} H^2 h_{ij} \right) \notag \\ & \stackrel{!}{=} W_{0ij0} + \lo^2_{ij} - \frac{|\lo|^2}{2(n\!-\!1)} h_{ij} \end{align} (<cit.>), where $W$ is the Weyl tensor of $g$. We recall that $W$ Weyl tensor \begin{equation}\label{RW-deco} R = W - \Rho \owedge g, \end{equation} where the Kulkarni-Nomizu product of the bilinear forms $b_1$ and $b_2$ is defined by \begin{align*} & (b_1 \owedge b_2) (X,Y,Z,W) \\ & \st b_1 (X,Z) b_2 (Y,W) - b_1 (Y,Z) b_2(X,W) + b_1(Y,W) b_2(X,Z) - b_1(X,W) b_2(Y.Z). \end{align*} $\owedge$ Kulkarni-Nomizu product §.§.§ Conformal change and the second fundamental form Let $N$ be a fixed unit normal field of $M$ defining $L$. Let $\hat{L}$ denote the second fundamental form of $M$ with respect to the metric $\hat{g} = e^{2\varphi} g$. Then it holds e^{-\varphi} \hat{L} = L + \nabla_N(\varphi) h. As a consequence, we find e^{\varphi} \hat{H} = H + \nabla_n(\varphi). Both relations combine into the conformal invariance property \begin{equation}\label{CTL-L} e^{-\varphi} \loh = \lo \end{equation} of the trace-free part $\lo = L - H h$ of $L$. It follows that $\loh^2 = \lo^2$, where $(\lo^2) _{ij} \st h^{ab} \lo_{ia} \lo_{jb}$. For $|L|^2 = \tr_h (L^2) = h^{ij} (L^2)_{ij} = h^{ij} h^{ab} L_{ia} L_{jb}$, we find $e^{2\varphi} |\loh|^2 = |\lo|^2$.[Of course, the norm on the left-hand side is taken with respect to $\hat{h}$.] \begin{equation}\label{tf-square} (\loh^2)_\circ = (\lo^2)_\circ. \end{equation} §.§.§ Some formulas for the Laplacian Here we discuss some useful formulas for the Laplacian. In particular, we prove the identity Let $\sigma$ be a defining function of $M$ and $\partial_0 = \NV = \grad_g(\sigma)$. We assume that $|\partial_0|_g =1$ on $M$. Then it holds \begin{equation}\label{LBM} \Delta_g (u) = \partial_0^2 (u) + \Delta_h (u) + n H \partial_0 (u) - \langle du, \nabla_{\partial_0}(\partial_0) \rangle \end{equation} on $M$. Let $\left \{\partial_i \right \}$ be an orthonormal basis of the tangent spaces of the level surfaces $\sigma^{-1}(c)$ (for small $c$) and let $\partial_0 = \grad_g(\sigma)$. By assumption, these form an orthonormal basis on $M$. Let $\left\{ dx^i \right\}$ together with $\left\{dx^0\right\}$ be the dual basis. We calculate $\iota^* \Delta_g (f)$ using $\Delta = \tr (\nabla^2)$. First of all, we have $\nabla (u) = \partial_i (u) dx^i + \partial_0 (u) dx^0$. Hence[As usual, we sum over repeated indices.] \begin{align*} \nabla^2(u) & = \partial_{ij}(u) dx^i \otimes dx^j + \partial_i(u) \nabla_{\partial_j}(dx^i) \otimes dx^j \\ & + \partial_{i0}(u) dx^i \otimes dx^0 + \partial_i(u) \nabla_{\partial_0} (dx^i) \otimes dx^0 \\ & + \partial_{0j}(u) dx^0 \otimes dx^j + \partial_0(u) \nabla_{\partial_j}(dx^0) \otimes dx^j \\ & + \partial_{00}(u) dx^0 \otimes dx^0 + \partial_0(u) \nabla_{\partial_0}(dx^0) \otimes dx^0. \end{align*} Now taking traces gives \begin{align*} \tr (\nabla^2)(u) & = \sum_{i=0}^n \partial_{ii}(u) + \partial_i(u) \langle \nabla_{\partial_j}(dx^i), \partial_j \rangle + \partial_i(u) \langle \nabla_{\partial_0} (dx^i), \partial_0 \rangle + \partial_0(u) \langle \nabla_{\partial_j}(dx^0), \partial_j \rangle \\ & + \partial_0^2(u) + \partial_0(u) \langle \nabla_{\partial_0}(dx^0), \partial_0 \rangle \end{align*} on $M$. Thus, we obtain \begin{align*} \tr (\nabla^2)(u) & = \sum_{i=0}^n \partial_{ii}(u) + \partial_i(u) \langle \nabla_{\partial_j}(dx^i), \partial_j \rangle - \partial_i(u) \langle dx^i, \nabla_{\partial_0} (\partial_0) \rangle - \partial_0(u) \langle dx^0, \nabla_{\partial_j}(\partial_j) \rangle \\ & + \partial_0^2(u) - \partial_0(u) \langle dx^0, \nabla_{\partial_0}(\partial_0) \rangle. \end{align*} But since $\nabla^g_{\partial_i}(\partial_j) = \nabla^h_{\partial_i}(\partial_j)$ on $M$, the last display simplifies to \begin{align*} \Delta_g (u) & = \partial_0^2(u) + \Delta_h (u) - \partial_0(u) \langle dx^0, \nabla_{\partial_j}(\partial_j) \rangle \\ & - \partial_i(u) \langle dx^i, \nabla_{\partial_0} (\partial_0) \rangle - \partial_0(u) \langle dx^0, \nabla_{\partial_0}(\partial_0) \rangle \end{align*} on $M$. By $L (\partial_i,\partial_j) = - \langle \nabla_{\partial_i} (\partial_j), dx^0 \rangle $, we obtain \Delta_g (u) = \partial_0^2(u) + \Delta_h (u) + n H \partial_0(u) - \langle du,\nabla_{\partial_0}(\partial_0) \rangle on $M$. This completes the proof. Without the assumption that $|\partial_0|=1$ on $M$, an extension of the above arguments shows that \begin{equation*} \Delta_g (u) = \frac{1}{|\partial_0|^2} \partial_0^2 (u) + \Delta_h (u) + n H \frac{1}{|\partial_0|} \partial_0 (u) - \frac{1}{|\partial_0|^2} \langle du, \nabla_{\partial_0}(\partial_0) \rangle \end{equation*} on $M$. Similar arguments prove (<ref>). In the situation of Lemma <ref>, assume that $\sigma$ is the distance function of $M$. Then $\nabla_\NV (\NV) = 0$ and we recover the well-known formula \begin{equation*} \Delta_g = \nabla_{\NV}^2 + \Delta_h + n H \nabla_{\NV} \end{equation*} on $M$. The following result reproves <cit.>. If $\sigma$ satisfies $\SCY$, then \begin{equation}\label{LR} \Delta_g = \nabla_\NV^2 + \Delta_h + (n-1) H \nabla_\NV \end{equation} on $M$. By Lemma <ref>, it suffices to prove that $\nabla_\NV(\NV) = H \NV$ on $M$. For any $X \in \mathfrak{X}(X)$, we calculate \langle \nabla_\NV(\NV),X\rangle = \Hess_g(\sigma)(\NV,X) = \Hess_g(\sigma)(X,\NV) = \langle \nabla_X(\NV),\NV \rangle = 1/2 \langle d (|\NV|^2),X \rangle. Now the assertion follows from $\SCY$ using $\rho = -H$ on $M$ (Lemma <ref>). §.§.§ Iterated normal derivatives The identity \iota^* \partial_s^k \circ \eta^* = \iota^* (|\NV|^{-2} \nabla_\NV)^k (see (<ref>)) relates iterated normal derivatives $\iota^* \partial_s^k \circ \eta^*$ with respect to $s$ to iterated weighted gradients $\iota^* (|\NV|^{-2} \nabla_\NV)^k$ of $\sigma$. Moreover, if $\SC(g,\sigma)=1$, i.e., if $|\NV|^2 = 1- 2 \sigma \rho$, any iterated weighted gradient $\iota^* (|\NV|^{-2} \nabla_\NV)^k$ can be written as a composition with $\iota^*$ of a linear combination of iterated gradients $\nabla^j_\NV$ for $j \le k$ and polynomials in the curvature quantities $\iota^* \nabla_\NV^j(\rho) \in C^\infty(M)$ for $j \le k-1$. This follows by an easy induction using $\nabla_\NV (\sigma) = |\NV|^2$. In particular, we obtain the following low-order formulas. If $\SC(g,\sigma)=1$, then it holds \begin{equation*} \iota^* \partial_s \eta^* = \iota^* \nabla_\NV \quad \mbox{and} \quad \iota^* \partial_s^2 \eta^* = \iota^* (\nabla_\NV^2 + 2 \rho \nabla_\NV). \end{equation*} We calculate \iota^* \partial_s^2 \eta^* = \iota^* (|\NV|^{-2} \nabla_\NV)^2 = \iota^* \nabla_\NV (1 + 2 \sigma \rho) \nabla_\NV = \iota^* (\nabla_\NV^2 + 2 \nabla_\NV(\sigma) \rho \nabla_\NV) using $|\NV|^2 = 1-2\sigma \rho$. Now $\iota^* \nabla_\NV (\sigma) = \iota^* (|\NV|^2) = 1$ implies the second identity. These formulas can easily be inverted to express iterated normal gradients in terms of iterated normal derivatives with respect to $s$. In particular, we obtain the following identities. If $\SC(g,\sigma)=1$, then it holds \begin{equation*} \iota^* \nabla_\NV = \iota^* \partial_s \eta^* \quad \mbox{and} \quad \iota^* \nabla_\NV^2 = (\iota^* \partial_s^2 - 2 \rho_0 \iota^* \partial_s ) \eta^*. \end{equation*} The above discussion obviously generalizes to the case that $\sigma$ only satisfies the condition $\SCY$ with a non-trivial remainder. §.§ Expansions of the metric. Model cases We start by discussing the normal forms of a given metric $g$ in geodesic normal coordinates and in adapted coordinates (as defined in Section <ref>). The formulas for these normal forms contain respective families $h_r$ and $h_s$ of metrics on $M$. In order to simplify notation, we shall use the same notation for the coefficients of their Taylor series in $r$ and $s$. It always will be clear from the context which coefficients are meant. We derive formulas for the first few Taylor coefficients of $h_r$ and $h_s$. A series of geometrically intuitive examples follow the discussion. $h_{(k)}$ coefficients of $h_r$ or $h_s$ We expand the family $h_r$ in the normal form $dr^2 + h_r$ of $g$ in geodesic normal coordinates as h_r = h + h_{(1)} r + h_{(2)} r^2 + \cdots. The coefficients $h_{(k)}$ can be expressed in terms of the curvature of the metric $g$, its covariant derivatives, and the second fundamental form $L$. The expansion starts \begin{equation}\label{h-geodesic} (h_r)_{ij} = h_{ij} + 2L_{ij} r + ((L^2)_{ij} - R_{0ij0}) r^2 + (h_{(3)})_{ij} r^3 + \cdots \end{equation} \begin{equation}\label{h-cubic} 3 (h_{(3)})_{ij} = - \nabla_{\partial_r} (R)_{0ij0} - 2 L_i^k R_{0jk0} - 2 L_j^k R_{0ik0}. \end{equation} Here $(L^2)_{ij} = L_{ik} L^k_j = L_{ik} L_{js} h^{ks}$. Next, assume that $\sigma$ satisfies $\SCY$. We expand the family $h_s$ in the normal form (<ref>) as h_s = h + h_{(1)} s + h_{(2)} s^2 + \cdots. The coefficients $h_{(k)}$ can be expressed in terms of the curvature of the metric $g$, its covariant derivatives, and the second fundamental form $L$. The expansion starts with \begin{equation}\label{h-adapted} (h_s)_{ij} = h_{ij} + 2 L_{ij} s + ((L \lo)_{ij} - R_{0ij0}) s^2 + (h_{(3)})_{ij} s^3 + \cdots \end{equation} \begin{align}\label{h-adapted-cubic} 3 (h_{(3)})_{ij} & = - \nabla_{\partial_s} (R)_{0ij0} - 2 \lo_i^k R_{0jk0} - 2 \lo_j^k R_{0ik0} \notag \\ & + \Hess_{ij}(H) - H R_{0ij0} - 3 H (L \lo)_{ij} + 2 L_{ij} \rho_0'. \end{align} Here $(L \lo)_{ij} = L_{ik} \lo_{js} h^{ks}$ and $\rho_0' = \Rho_{00} + |\lo|^2/(n-1)$. The notation $\rho_0'$ will be justified in Lemma <ref>. We recall the standard formulas R^s_{ijk} = \Gamma_{jk}^l \Gamma_{il}^s - \Gamma_{ik}^l \Gamma_{jl}^s + \partial_i (\Gamma_{jk}^s) - \partial_j (\Gamma_{ik}^s), \quad R_{ijkl} = R_{ijk}^s g_{sl} and $\Gamma_{ij}^k$ Christoffel symbol \Gamma_{ij}^m = \frac{1}{2} g^{km} (\partial_i (g_{jk}) + \partial_j (g_{ik}) - \partial_k (g_{ij})) for the components of the curvature tensor of a metric $g$.[As usual, we sum over repeated indices and denote derivatives by a lower index.] For the metric $g = dr^2 + h_r$, we find $\Gamma_{ij}^0 = -\frac{1}{2} \partial_r(h_r)_{ij}$. Hence $L_{ij} = - \Gamma_{ij}^0 = \frac{1}{2} (h_{(1)})_{ij}$ on $M$. This proves $h_{(1)} = 2L$. In order to verify the formula for $h_{(2)}$ in (<ref>), we calculate the components $R_{0ij0}$ of the curvature tensor of the metric $dr^2 + h_r$. We decompose $R_{0jk}^0$ as R_{0jk}^0 = \left( \Gamma_{jk}^l \Gamma_{0l}^0 - \Gamma_{0k}^l \Gamma_{jl}^0 + \partial_r (\Gamma_{jk}^0) - \partial_j (\Gamma_{0k}^0) \right) + \Gamma_{jk}^0 \Gamma_{00}^0 - \Gamma_{0k}^0 \Gamma_{j0}^0, where the summations run only over tangential indices. Now, for the metric $g = dr^2 + h_r$, we find the Christoffel symbols \Gamma_{ij}^0 = - \frac{1}{2} g_{ij}', \quad \Gamma_{0j}^0 = 0, \quad \Gamma_{00}^0 = 0, \quad \Gamma_{0k}^l = \frac{1}{2} g^{rl} g_{kr}', where $^\prime$ denotes $\partial_r$. It follows that \begin{equation}\label{curv-geo} R_{0jk}^0 = \frac{1}{4} g^{rl} g_{kr}' g_{jl}' - \frac{1}{2} g_{jk}''. \end{equation} We evaluate this formula at $r=0$. Using $g_{jk}' = 2 L_{jk}$, we find R_{0jk0} = R_{0jk}^0 = h^{rl} L_{kr} L_{jl} - (h_{(2)})_{jk} = (L^2)_{jk} - (h_{(2)})_{jk}. This implies the formula for $h_{(2)}$ in (<ref>). Next, we prove the formula for the cubic term. First, we note that \nabla_{\partial_r}(R)_{0jk0} = \partial_r (R_{0jk0}) - R(\partial_r,\nabla_{\partial_r}(\partial_j),\partial_k,\partial_r) - R(\partial_r,\partial_j,\nabla_{\partial_r}(\partial_k),\partial_r) using $\nabla_{\partial_r}(\partial_r) = 0$. Now (<ref>) implies \begin{align*} \partial_r (R_{0jk0}) = \partial_r (R_{0jk}^0 g_{00}) = \partial_r (R_{0jk}^0) = \frac{1}{4} (g^{rl})' g_{kr}' g_{jl}' + \frac{1}{4} g^{rl} g_{kr}'' g_{jl}' + \frac{1}{4} g^{rl} g_{kr}' g_{jl}''- \frac{1}{2} g_{jk}'''. \end{align*} Evaluation of this formula for $r=0$ yields \begin{align*} \partial_r (R_{0jk0}) & = - 2 L^{rl} L_{kr} L_{jl} + h^{rl} (h_{(2)})_{kr} L_{jl} + h^{rl} L_{kr} (h_{(2)})_{jl} - 3 (h_{(3)})_{jk} \\ & = - L_j^r R_{0kr0} - L_k^l R_{0jl0} - 3 (h_{(3)})_{jk}. \end{align*} The two remaining terms in the formula for $\partial_r (R_{0jk0})$ for $r=0$ are - R_{0jl0} L_k^l - R_{0lk0} L_j^l. Thus, we find \nabla_{\partial_r}(R)_{0jk0} = -2 L_j^r R_{0kr0} - 2 L_k^l R_{0jl0} - 3 (h_{(3)})_{jk}. This proves (<ref>). Similarly, the formula for $h_s$ in (<ref>) and (<ref>) follows from a calculation of the Christoffel symbols and the components $R_{0ij0} = R_{0ij}^0 g_{00}$ of the curvature tensor of the metric \eta^*(g) = \eta^*(|\NV|^2)^{-1} ds^2 + h_s = a^{-1} ds^2 + h_s with $a = \eta^*(|\NV|^2)$. In order to simplify the notation, we shall write $g$ instead of $\eta^*(g)$ and $\rho$ instead of $\eta^*(\rho)$. For the above metric, we find the Christoffel symbols[As usual the prime means derivative in $s$.] \Gamma_{ij}^0 = - \frac{1}{2} g^{00} g_{ij}', \quad \Gamma_{0j}^0 = \frac{1}{2} g^{00} \partial_j (g_{00}), \quad \Gamma_{00}^0 = \frac{1}{2} g^{00} g_{00}', \quad \Gamma_{0k}^l = \frac{1}{2} g^{rl} g_{kr}'. Here $g^{00} = a = 1- 2s\rho$ (by assumption) and $g_{00} = a^{-1} = 1 + 2s\rho + \cdots$. In particular, $g^{00}=1$, $g_{00}' = -2H$ and $(g^{00})' = 2H$ on $M$. Here we used that $\rho = -H$ on $M$ (Lemma <ref>). Hence $L_{ij} = - \Gamma_{ij}^0 = \frac{1}{2} (h_{(1)})_{ij}$. The components $R_{0jk0}$ of the curvature tensor are given by R_{0jk0} = R_{0jk}^0 g_{00} \begin{align}\label{R-adapted} R_{0jk}^0 & = \frac{1}{2} \Gamma_{jk}^l g^{00} \partial_l (g_{00}) + \frac{1}{4} g^{00} g^{rl} g_{kr}' g_{jl}' - \frac{1}{2} ((g^{00})' g_{jk}' + g^{00} g_{jk}'') \notag\\ & - \frac{1}{2} ( \partial_j (g^{00}) \partial_k(g_{00}) + g^{00} \partial^2_{kj}(g_{00})) \notag \\ & - \frac{1}{4} (g^{00})^2 g_{jk}' g_{00}' - \frac{1}{4} (g^{00})^2 \partial_j(g_{00}) \partial_k(g_{00}). \end{align} Evaluation of this formula for $s=0$ gives R_{0jk0} = h^{rl} L_{kr} L_{jl} - (h_{(2)})_{jk} - 2H L_{jk} + H L_{jk} = (L^2)_{jk} - H L_{jk} - (h_{(2)})_{jk}. But $L^2 - H L = L (L - H h) = L \lo$. Hence $R_{0jk0} = (L \lo)_{jk} - (h_{(2)})_{jk}$. This implies the formula for $h_{(2)}$ in (<ref>). Finally, we prove the formula for the cubic term. Here we utilize the expansions g_{00} = (1-2 \rho s)^{-1} = 1 + 2\rho s + 4 \rho^2 s^2 + \cdots = 1 + 2\rho_0 s + (2\rho_0'+4\rho_0^2) s^2 + \cdots g^{00} = 1 -2\rho s = 1- 2\rho_0 s - 2\rho_0' s^2 + \cdots. We proceed as above and calculate $\nabla_{\partial_s} (R)_{0jk0}$ for $s=0$. In the present case, it holds \begin{align*} \nabla_{\partial_s}(R)_{0jk0} & = \partial_s (R_{0jk0}) - R(\partial_s,\nabla_{\partial_s}(\partial_j),\partial_k,\partial_s) - R(\partial_s,\partial_j,\nabla_{\partial_s}(\partial_k),\partial_s) \\ & - R(\nabla_{\partial_s}(\partial_s),\partial_j,\partial_k,\partial_s) - R(\partial_s,\partial_j,\partial_k,\nabla_{\partial_s}(\partial_s)); \end{align*} note that $\nabla_{\partial_s}(\partial_s)$ does not vanish in general. First, we calculate the term $\partial_s (R_{0jk0})$ for $s=0$ using (<ref>). We \begin{align*} \partial_s (R_{0jk0}) & = \partial_s (R_{0jk}^0) - 2 H R_{0jk}^0 \\ & = -\Gamma_{jk}^l \partial_l(H) + \frac{1}{2} H h^{rl} g_{kr}' g_{jl}' + \frac{1}{4} (g^{rl})' g_{kr}' g_{jl}' + \frac{1}{4} h^{rl} g_{kr}'' g_{jl}' + \frac{1}{4} h^{rl} g_{kr}' g_{jl}'' \\ & + 4 \rho_0' L_{jk} - 2 H g_{jk}'' - \frac{1}{2} g_{jk}''' + \partial^2_{kj}(H) \\ & + \frac{1}{2} H g_{jk}'' - \rho_0' g_{jk}' - 2 H R_{0jk0} \end{align*} using the expansions of $g_{00}$ and $g^{00}$. We further simplify that sum by using the known formula for the first derivative of $h_s$ and $\Hess_{jk} = \partial^2_{jk} - \Gamma_{jk}^l \partial_l$. Then \begin{align*} \partial_s (R_{0jk0}) & = \Hess_{jk}(H) + 2 H L_k^l L_{jl} - 2 L^{rl} L_{kr} L_{jl} + L_j^r (h_{(2)})_{kr} + L_k^l (h_{(2)})_{jl} \\ & + 4 L_{jk} \rho_0' - 4 H (h_{(2)})_{jk} - 3 (h_{(3)})_{jk} + H (h_{(2)})_{jk} - 2 L_{jk} \rho_0' \\ & = \Hess_{jk}(H) + 2 H (L^2)_{jk} - 2 (L^3)_{jk} + L_j^r (h_{(2)})_{kr} + L_k^l (h_{(2)})_{jl} \\ & -3 H (h_{(2)})_{jk} + 2 L_{jk} \rho_0' - 3 (h_{(3)})_{jk} - 2 H R_{0jk0} \end{align*} on $M$. The four remaining contributions to $\nabla_{\partial_s}(R)_{0jk0}$ for $s=0$ equal \begin{align*} & - \Gamma_{0j}^l R_{0lk0} - \Gamma_{0k}^l R_{0jl0} - \Gamma_{00}^0 R_{0jk0} - \Gamma_{00}^0 R_{0jk0} \\ & = - \frac{1}{2} g^{rl} g_{jr}' R_{0lk0} - \frac{1}{2} g^{rl} g_{kr}' R_{0jl0} - g^{00} g_{00}' R_{0jk0} \\ & = - L_j^l R_{0lk0} - L_k^l R_{0jl0} + 2 H R_{0jk0} \\ & = -\lo_j^l R_{0lk0} - \lo_k^l R_{0jl0}. \end{align*} \begin{align*} \nabla_{\partial_s}(R)_{0jk0} & = \Hess_{jk}(H) + 2 H (L^2)_{jk} - 2 (L^3)_{jk} \\ & + L_j^r ((L \lo)_{kr} - R_{0kr0}) + L_k^l ((L \lo)_{jl} - R_{0jl0}) - 3 H ((L \lo)_{jk} - R_{0jk0}) \\ & - \lo_j^l R_{0lk0} - \lo_k^l R_{0jl0} + 2 L_{jk} \rho_0' - 3 (h_{(3)})_{jk} - 2 H R_{0jk0} \\ & = \Hess_{jk}(H) - L_j^r R_{0kr0} - L_k^l R_{0jl0} - 3 H (L \lo)_{jk} + 3 H R_{0jk0} \\ & -\lo_j^l R_{0lk0} - \lo_k^l R_{0jl0} + 2 L_{jk} \rho_0' - 3 (h_{(3)}) _{jk} - 2 H R_{0jk0} \\ & = \Hess_{jk}(H) - 2 \lo_j^r R_{0kr0} - 2 \lo_k^r R_{0jr0} - 3 H (L \lo)_{jk} - H R_{0jk0} \\ & + 2 L_{jk} \rho_0' - 3 (h_{(3)}) _{jk} \end{align*} by the formula for $h_{(2)}$. This proves (<ref>). Formula (<ref>) was given in <cit.> and the formula (<ref>) for the cubic coefficient was displayed in <cit.>.[In our conventions, $R$ and $L$ have opposite signs as in <cit.>.] The proof of Proposition <ref> shows that $h_{(2)}$ depends only on $\rho_0$. Similarly, $h_{(3)}$ only depends on $\rho_0$ and $\rho_0'$. More generally, it is easy to see that $h_{(k)}$ only depends on $\partial_s^j(\rho)|_0$ for $j \le k-2$. Formula (<ref>) for the cubic term $h_{(3)}$ of the expansion of $g$ in adapted coordinates yields \begin{equation}\label{h12-trace-n} \tr (h_{(1)}) = 2 n H, \quad \tr(h_{(2)}) = |\lo|^2 - \Ric_{00} \end{equation} \begin{equation}\label{h3-trace-n} 3 \tr (h_{(3)}) = - \nabla_{\partial_s}(\Ric)_{00} - 4 \lo^{ik} R_{0ik0} + \Delta H - H \Ric_{00} - 3 H|\lo|^2 + 2 n H \rho_0'. \end{equation} If $g_+ = r^{-2} g = r^{-2} (dr^2 + h_r)$ is a Poincaré-Einstein metric in normal form relative to $h = h_0$, then $h_{(1)}=0$ and $h_{(2)} = -\Rho^h$. Comparing this result with (<ref>), implies that $R^g_{0ij0} = \Rho^h_{ij}$. Hence $\Ric^g_{00} = \J^h$. We illustrate the above results in some simple model cases. Let $S^n \subset \R^{n+1}$ be defined by $|x|=1$. Let $g_0$ be the Euclidean metric on $\R^{n+1}$. Then the function $\sigma = (1-|x|^2)/2$ yields the Poincaré metric \sigma^{-2} g_0 = \frac{4}{(1-|x|^2)^2} \sum_{i=1}^{n+1} dx_i^2 on the unit ball $|x|<1$. Now $\NV = \grad(\sigma) = - \sum_i x_i \partial_i$. Hence the normalized gradient field $\mathfrak{X} = \NV/|\NV|^2$ is given by $\mathfrak{X} = -\frac{1}{|x|^2} \sum_i x_i \partial_i$. It follows that the map \eta: (-1/2,1/2) \times S^n \ni (s,x) \mapsto \sqrt{1-2s} x \in \R^{n+1} defines the adapted coordinates. Note that $\eta^*(\sigma)=s$. Hence we obtain \eta^*(g_0) = \frac{1}{1-2s} ds^2 + (1-2s) h, where the round metric $h$ on $S^n$ is induced by $g$. In particular, $h_s = h - 2sh$, i.e., $h_{(1)} = -2h$ and $h_{(2)}=0$ The coefficient $h_{(1)}$ is to be interpreted as $2L$ being defined by the unit normal field $-\partial_r$. The vanishing of the quadratic and the cubic term is confirmed by the general formulas. Note also that $\rho = 1$. It follows that $v(s)=(1-2s)^{\frac{n-1}{2}}$ and $\mathring{v}(s) = (1-2s)^{\frac{n}{2}}$. These results confirm the relation (<ref>) using $a=1-2s$. Moreover, the identity (<ref>) in Conjecture <ref> is trivially satisfied. Let $S^n \subset S^{n+1}$ be an equatorial subsphere. Let $g$ be the round metric on $S^{n+1}$. Then the height-function $\sigma = \He \in C^\infty(S^{n+1})$ defines the metric $\He^{-2} g$ on both connected components of the complement of the zero locus of $\He$. It is isometric to the Poincaré-metric on the unit ball (see Section <ref>). The map \eta: (-1,1) \times S^n \ni (s,x) \mapsto (\sqrt{1-s^2}x,s) \in S^{n+1} defines the adapted coordinates. Note that $\eta^*(\sigma)=s$. Hence we obtain \eta^*(g) = \frac{1}{1-s^2} ds^2 + (1-s^2) h, where the round metric $h$ on $S^n$ is induced by $g$. In particular, $h_{(1)} = 0$ and $h_{(2)}= -h$. The coefficient $h_{(1)}$ vanishes since $L=0$ and the coefficient $h_{(2)}$ is to be interpreted as $-R_{0ij0}$. The vanishing of the cubic term is confirmed by the general formula. Note that $\He$ is an eigenfunction of the Laplacian on $S^{n+1}$ with eigenvalue $-(n+1)$ and $\J = \frac{n+1}{2}$. Hence $\rho=\frac{1}{2} \He$ and $\eta^*(\rho) = \frac{1}{2} s$. Finally, we have $v(s)=(1-s^2)^{\frac{n-1}{2}}$ and $\mathring{v}(s) = (1-s^2)^{\frac{n}{2}}$. These results confirm the relation (<ref>) using $a = 1-s^2$. Moreover, the identity (<ref>) in Conjecture <ref> reduces to the trivial relation \binom{\frac{n-1}{2}}{\frac{n+1}{4}} (-1)^{\frac{n+1}{4}} + \binom{\frac{n-1}{2}}{\frac{n-3}{4}} (-1)^{\frac{n-3}{4}} = 0. In the above two examples, either the curvature of the background metric or the second fundamental form vanishes. We finish this section with the discussion of a model case with non-trivial curvature and non-trivial second fundamental form. Let $\mathbb{H}^{n+1}$ be the upper half-space with the hyperbolic metric $g_+ = r^{-2} (dr^2 + dx^2)$. For $c>0$, we let $X = \{r \ge c\}$ with boundary $M = \{ r=c \}$ and background metric $g=g_+$. The metric $g$ restricts to $h = c^{-2} dx^2$ on $M$. The defining function $\sigma = 1 - c/r$ is smooth up to the boundary $M$. It solves the singular Yamabe problem since $\sigma^{-2} g_+ = (r-c)^{-2} (dr^2 + dx^2)$ has scalar curvature $-n(n+1)$. The map \eta: (0,1) \times \R^n \ni (s,x) \mapsto \left(\frac{c}{1-s},x\right) \in X defines the adapted coordinates. In fact, $\eta^*(\sigma) = s$. We obtain the normal form \eta^*(g) = \frac{1}{(1-s)^2} ds^2 + \frac{(1-s)^2}{c^2} dx^2 of $g$ in adapted coordinates. In particular, $h_{(1)} = -2h$, $h_{(2)} = h$ and $h_{(3)} = 0$. These results fit with the general formulas in Proposition <ref> since $L=-h$ and $R_{0ij0} = -h_{ij}$. The vanishing of $h_{(3)}$ follows using $\lo = 0$, $H = -1$ and $\Rho_{00} = - 1/2$. A calculation using the formula $\Delta_{g_+} = r^2 \partial_r^2 - (n-1) r \partial_r + \Delta_{\R^n}$ for the Laplacian of $g_+$ yields \rho = \frac{1}{2} + \frac{c}{2r}. Hence $\eta^*(\rho) = 1- s/2$. Note that $\NV = \grad_g(\sigma) = -c \partial_r$, $|\NV|^2 = r^{-2} c^2$ and $\eta^*(|\NV|^2) = (1-s)^2 \stackrel{!}{=} 1- 2s\eta^*(\rho)$. Finally, we have $v(s) = (1-s)^{n-1}$ and $\mathring{v}(s) = (1-s)^n$. By $\J = - \frac{n+1}{2}$, the identity (<ref>) in Conjecture <ref> reduces to the trivial relation \binom{n-1}{\frac{n+1}{2}} (-1)^{\frac{n+1}{2}} - \binom{n-1}{\frac{n-3}{2}} (-1)^{\frac{n-3}{2}} = 0. Finally, these results confirm the relation (<ref>) using $a=(1-s)^2$. §.§ Approximate solutions of the singular Yamabe problem. The residue formula In the present section, we determine the first few terms in the expansion $\sigma_F = r + \sigma_{(2)} r^2 + \dots$ of a solution of the equation \begin{equation}\label{start-sol} \SC(g,\sigma_F) = 1+O(r^{n+1}). \end{equation} We also describe the obstruction $\B_n$ in terms of a formal residue of the supercritical term $\sigma_{n+2}$. By (<ref>), equation (<ref>) takes the form \begin{align*} & \partial_r(\sigma_F)^2 + h_r^{ij} \partial_i (\sigma_F) \partial_j (\sigma_F) \notag \\ & - \frac{2}{n+1} \sigma_F \left (\partial_r^2 (\sigma_F) + \frac{1}{2} \tr (h_r^{-1} h_r') \partial_r (\sigma_F) + \Delta_{h_r} (\sigma_F) + \bar{\J} \sigma_F \right) = 1+ O(r^{n+1}). \end{align*} Here and in the following, we use the bar notation for curvature quantities of $g$. We shall formulate the results in terms of the volume coefficients $u_k$ of the metric $g$ in geodesic normal coordinates (see (<ref>)) using \begin{equation}\label{volume-geodesic} \frac{u'(r)}{u(r)} = \frac{1}{2} \tr (h_r^{-1} h_r'). \end{equation} The following results describe the first three Taylor coefficients of a solution $\sigma_F$ of the singular Yamabe problem in general dimensions. $\sigma_{(2)}$, $\sigma_{(3)}$, $\sigma_{(4)}$ It holds \begin{align*} \sigma_{(2)} & = \frac{1}{2n} u_1, \\ \sigma_{(3)} & = \frac{2}{3(n-1)} u_2 - \frac{1}{3n} u_1^2 + \frac{1}{3(n-1)} \bar{\J} \end{align*} \begin{align}\label{sigma4-g} \sigma_{(4)} & = \frac{3}{4(n-2)} u_3 - \frac{9n^2-20n+7}{12n(n-1)(n-2)} u_1 u_2 + \frac{6n^2-11n+1}{24n^2(n-2)} u_1^3 \notag \\ & + \frac{2n-1}{6n(n-1)(n-2)} u_1 \bar{\J} + \frac{1}{4(n-2)} \bar{\J}' + \frac{1}{4(n-2)} \Delta (\sigma_{(2)}). \end{align} Alternatively, the above formulas can be derived from the description of the solution $\sigma$ in <cit.>. In fact, these formulas describe expansions of $\sigma$ into power series of any defining function with coefficients that live on the background space $X$. The calculation then requires expanding these coefficients into power series of the distance function. We omit the details. The volume coefficients $u_j$ may be expressed in terms of the Taylor coefficients of $h_r$. Such relations follow from (<ref>) by Taylor expansion in $r$ and resolving the resulting relations for $u_j$. We find \begin{align*} u_1 & = \frac{1}{2} \tr (h_{(1)}), \\ u_2 & = \frac{1}{8} (\tr (h_{(1)})^2 + 4 \tr (h_{(2)}) - 2 \tr (h_{(1)}^2)) \end{align*} \begin{align*} u_3 & = \frac{1}{48} (\tr (h_{(1)})^3 + 12 \tr (h_{(1)}) \tr (h_{(2)}) + 24 \tr (h_{(3)}) - 6 \tr(h_{(1)}) \tr (h_{(1)}^2) \\ & - 24 \tr(h_{(1)} h_{(2)}) + 8 \tr (h_{(1)}^3)). \end{align*} These formulas are valid in general dimensions. The expressions for $h_{(j)}$ (for $j \le 3$) in Proposition <ref> imply \begin{align*} u_1 & = n H, \\ 2 u_2 & = -\overline{\Ric}_{00} - |L|^2 + n^2 H^2 = \overline{\Ric}_{00} + \scal - \overline{\scal} \end{align*} (by the Gauss equation) or equivalently \begin{equation}\label{u2} 2 u_2 = -\overline{\Ric}_{00} - |\lo|^2 + n(n-1) H^2. \end{equation} Moreover, we find \begin{equation}\label{u3} 6 u_3 = - \bar{\nabla}_0 (\overline{\Ric})_{00} + 2 (L,\bar{\G}) - 3 n H \overline{\Ric}_{00} + 2 \tr(L^3) - 3n H |L|^2 + n^3 H^3, \end{equation} where $\bar{\G}_{ij} \st \bar{R}_{0ij0}$, or equivalently $\bar{\G}$ \begin{align*} 6 u_3 & = -\bar{\nabla}_0 (\overline{\Ric})_{00} + 2 (\lo,\bar{\G}) - (3n-2) H \overline{\Ric}_{00} \\ & + 2 \tr (\lo^3) - 3(n-2) H |\lo|^2 + n(n-1)(n-2) H^3. \end{align*} For these formulas for $u_1,u_2,u_3$, see also <cit.>. Note that \begin{equation}\label{det-2} - |L|^2 + n^2 H^2 = \sum_{ij} \begin{vmatrix} L_{ii} & L_{ij} \\ L_{ji} & L_{jj} \end{vmatrix} = 2 \sigma_2(L) \end{equation} and $\sigma_k(L)$ elementary symmetric polynomial \begin{equation}\label{det-3} 2 \tr(L^3) - 3n H |L|^2 + n^3 H^3 = \sum_{ijk} \begin{vmatrix} L_{ii} & L_{ij} & L_{ik} \\ L_{ji} & L_{jj} & L_{jk} \\ L_{ki} & L_{kj} & L_{kk} \end{vmatrix} = 6 \sigma_3(L) \end{equation} in an orthonormal basis. Here $\sigma_k(L)$ denotes the $k$-th elementary symmetric polynomial in the eigenvalues of the shape operator. The identities (<ref>) and (<ref>) are special cases of Newton's identities relating elementary symmetric polynomials to power series. In these terms, we have \begin{align*} 2 u_2 & = -\overline{\Ric}_{00} + 2 \sigma_2(L) \\ 6 u_3 & = -\bar{\nabla}_0 (\overline{\Ric})_{00} + 2 (\lo,\bar{\G}) - (3n-2) H \overline{\Ric}_{00} + 6 \sigma_3(L). \end{align*} Note also that versions of the Gauss identity express the quantities $\sigma_2(L)$ and $\sigma_3(L)$ in terms of the curvatures of $g$ and of the induced metric on $M$ <cit.>. Finally, we note that (<ref>) vanishes in $n=2$ - this identity is equivalent to $\tr(\lo^3) = 0$ in $n=2$. The above results imply It holds $2 \sigma_{(2)} = H$ and \begin{equation}\label{sigma3} 6 \sigma_{(3)} = -2 \bar{\Rho}_{00} - \frac{2}{n-1} |\lo|^2 = 2( \J - \iota^* \bar{\J}) - \frac{1}{n-1} |\lo|^2 - n H^2 \end{equation} for $n \ge 2$. The second equality follows by combining the first equality with (<ref>). The results of Lemma <ref> are contained in <cit.>. See also <cit.>.[These references use a different convention for $L$ and $H$.] The above formulas for $v_j$ are equivalent to the corresponding formulas in <cit.>. We also refer to <cit.> for the corresponding results in higher codimensions. However, the methods of proof in these references are different. For $n=1$, the above results easily imply \begin{equation}\label{B1=0} \B_1 = (r^{-2} (\SC(S_2)-1))|_0 = -2u_2 - \bar{\J}_0 = \overline{\Ric}_{00} - \bar{\J}_0 = 0 \end{equation} using $\lo=0$ (see also Remark <ref>). This result is well-known <cit.>. It also is of interest to explicate the above formulas for flat backgrounds. In fact, it follows from the identity \begin{equation}\label{volume-flat} u(r) = \det (\id + r L) \end{equation} (see <cit.>) for a flat background that the formulas for $\sigma_{(k)}$ (for $k \le 4$) can be expressed in terms of $L$. Newton's identities imply \begin{align*} u_1 & = n H, \\ u_2 & = \frac{1}{2} (n (n-1) H^2 - |\lo|^2), \\ u_3 & = \frac{1}{6} ( H^3 n(n-1)(n-2) - 3 (n-2) H |\lo|^2 + 2 \tr(\lo^3)) \end{align*} and a direct calculation yields the following result. The expansion of a solution of the Yamabe problem for $M^n \hookrightarrow \R^{n+1}$ has the form \sigma_F = r + \frac{r^2}{2} H - \frac{r^3}{3(n-1)} |\lo|^2 + r^4 \sigma_{(4)} + \cdots with the coefficient \begin{align*} \sigma_{(4)} & = \frac{1}{24(n-2)} \left(6 \tr(\lo^3) + \frac{7n-11}{n-1} H |\lo|^2 + 3 \Delta(H)\right). \end{align*} Note that the coefficients in formula (<ref>) for $\sigma_{(4)}$ have a simple pole in $n=2$. The following result calculates the formal residue at $n=2$. It holds \res_{n=2}(\sigma_{(4)}) = \frac{3}{4} u_3 + \frac{1}{32} u_1^3 - \frac{1}{8} u_1 u_2 + \frac{1}{4} u_1 \bar{\J} + \frac{1}{4} \bar{\J}' + \frac{1}{4} \Delta (\sigma_{(2)}). For a flat background, we obtain \res_{n=2}(\sigma_{(4)}) = \frac{1}{8} (H |\lo|^2 + \Delta (H)). Note that for a flat background it holds $u_3 = 0$ in $n=2$. We also recall that $\tr(\lo^3) = 0$ for $n=2$. Alternatively, one can use formula <cit.> for $\sigma_{(4)}$ to confirm this residue formula for general backgrounds. Corollary <ref> and the following result are special cases of the residue formula in Lemma <ref>. It holds \res_{n=2}(\sigma_{(4)}) = - \frac{3}{8} \B_2. We recall that the obstruction $\B_2$ is defined by \B_2 = (r^{-3} (\SC(S_3)-1))|_0, where $S_3= r + \sigma_{(2)} r^2 + \sigma_{(3)} r^3$. A calculation yields \begin{equation*}\label{B2-new-g} \B_2 = - 2 u_3 - \frac{1}{12} u_1^3 + \frac{1}{3} u_1 u_2 - \frac{2}{3} u_1\bar{\J} - \frac{2}{3} \bar{\J}' - \frac{2}{3} \Delta (\sigma_{(2)}). \end{equation*} We omit the details. We recall that the term $u_3$ vanishes in $n=2$ in the flat case but not in the curved case. In Section <ref>, we shall derive another formula for $\B_2$ from Theorem <ref>. Although the equivalence of both formulas is non-trivial, we leave the check of consistency to the reader. Similarly, we may either directly evaluate the definition $\B_3 = (r^{-4} (\SC(S_4)-1))|_0$ or use the residue formula (<ref>) to derive a formula for $\B_3$ from the residue of $\sigma_{(5)}$ at $n=3$. The results read as follows. It holds \begin{align*}\label{B3-start} \B_3 & = - 2u_4 + \frac{1}{2} u_1 u_3 + \frac{1}{3} u_2^2 - \frac{7}{18} u_1^2 u_2 + \frac{2}{27} u_1^4 - \frac{1}{3} \bar{\J} u_2 - \frac{5}{12} \bar{\J}' u_1 - \frac{1}{4} \bar{\J}'' \notag \\ & - \frac{1}{2} \Delta (\sigma_{(3)}) - \frac{1}{3} u_1 \Delta(\sigma_{(2)}) - \frac{1}{2} \Delta' (\sigma_{(2)}) + |d\sigma_{(2)}|^2, \end{align*} \begin{equation*} 6 \sigma_{(2)} = u_1 \quad \mbox{and} \quad 9 \sigma_{(3)} = 3 u_2 - u_1^2 + \frac{3}{2} \bar{\J} \end{equation*} and $\Delta' = (d/dt)|_0 (\Delta_{h+2sL})$ (see (<ref>)). For a flat background, it holds $\bar{\J}=0$, the term $u_4$ vanishes, and we obtain \begin{equation}\label{B3-flat-algo} 12 \B_3 = \Delta (|\lo|^2) + 6 H \tr (\lo^3) + |\lo|^4 + 3 |dH|^2 - 6 H \Delta(H) - 3 \Delta' (H). \end{equation} The above formula for $\B_3$ also follows from the following result through the residue formula \begin{equation}\label{res-5} \res_{n=3}(\sigma_{(5)}) = - \frac{2}{5} \B_3. \end{equation} The evaluation of that formula rests on the following result for the coefficient $\sigma_{(5)}$. In general dimensions, it holds \begin{align*} \sigma_{(5)} & = - \frac{n+1}{10(n-3)} |d\sigma_{(2)}|^2 \\ & + \frac{1}{5(n-3)} \Delta'(\sigma_{(2)}) + \frac{1}{5(n-3)} \Delta(\sigma_{(3)}) + \frac{3n-1}{20(n-3)(n-2)n} \Delta(\sigma_{(2)}) u_1 \\ & + \frac{1}{10(n-3)} \bar{\J}'' + \frac{n-1}{4(n-3)(n-2)n} \bar{\J}' u_1 + \frac{2(3n-5)}{15(n-3)(n-1)^2} \bar{\J} u_2 \\ & - \frac{4n-3}{20 (n-2)(n-1)n} \bar{\J} u_1^2 + \frac{1}{30(n-1)^2} \bar{\J}^2 \\ & + \frac{48n^4-247n^3+387n^2-179n+3}{60(n-3)(n-2)(n-1)n^2} u_1^2 u_2 - \frac{2(3n^2-11n+10)}{15(n-3)(n-1)^2} u_2^2 \\ & - \frac{24n^4-110n^3+133n^2-24n-3}{120 (n-3)(n-2)n^3} u_1^4 - \frac{16n^2-53n+27}{20(n-3)(n-2)n} u_1 u_3 + \frac{4}{5(n-3)} u_4. \end{align*} We omit the details of the proof. An alternative formula for $\sigma_{(5)}$ is given in <cit.>. Here the same comments as after Lemma <ref> apply. Lemma <ref> implies \res_{n=2}(\sigma_{(5)}) = - \frac{1}{2} u_1 \res_{n=2} (\sigma_{(4)}) = \frac{3}{4} H \B_2. This relation extends the identities \begin{align*} \res_{n=1}(\sigma_{(4)}) & = - \frac{1}{2} u_1 \res_{n=1}(\sigma_{(3)}), \\ \res_{n=1}(\sigma_{(3)}) & = - \frac{1}{2} u_1 \res_{n=1}(\sigma_{(2)}) = 0. \end{align*} More generally, we conjecture the factorization identities \res_{n=k-3}(\sigma_{(k)}) = - \frac{1}{2} u_1 \res_{n=k-3} (\sigma_{(k-1)}) for $k \ge 4$. Combining (<ref>) with the variation formula $\Delta'(u) = -2 (\Hess(u),L) - 3 (du,dH)$ (see (<ref>)) shows that 12 \B_3 = \Delta (|\lo|^2) + 12 |dH|^2 + 6 (\Hess(H),\lo) + |\lo|^4 + 6 H \tr(\lo^3). In Section <ref>, we shall alternatively derive that result from Theorem <ref>. A direct proof that, for conformally flat backgrounds, Lemma <ref> is equivalent to Lemma <ref> will be given in a separate work. Finally, we establish the residue formula for the singular Yamabe obstruction in full generality. It holds \begin{equation}\label{res-f} \res_{n=k-2}(\sigma_{(k)}) =-\frac{k-1}{2k} \B_{k-2}, \; k \ge 4. \end{equation} Let $S_k$ be the Taylor polynomial $r + r^2 \sigma_{(2)} + \cdots + r^k \sigma_{(k)}$ of degree $k$. Then a calculation shows that \begin{equation}\label{S-deco} \SC(S_k) = \SC(S_{k-1}) + r^{k-1} \frac{2k (n-k+2)}{n+1} \sigma_{(k)} + O(r^k). \end{equation} Assume that $S_{k-1}$ is the $(k-1)$-th approximate solution of the singular Yamabe problem, i.e., $\SC(S_{k-1}) = 1+ O(r^{k-1})$. Then $S_k$ is the $k$-th approximate solution if in the expansion \SC(S_k) = 1 + r^{k-1} \left (\frac{2k (n-k+2)}{n+1} \sigma_{(k)} + \cdots \right) + O(r^k) with an unknown coefficient $\sigma_{(k)}$ the coefficient of $r^{k-1}$ vanishes, i.e., if $\SC(S_k) = 1 + O(r^k)$. This can be solved for $\sigma_{(k)}$ if $k =2,3,\dots,n+1$.[This algorithm yields Lemma <ref>.] In the case $k=n+2$, we only have \SC(S_{n+1}) = 1 + O(r^{n+1}), and the restriction of the latter remainder is the obstruction $\B_n$. Now (<ref>) implies \frac{k-1}{2k} (r^{-k+1} (\SC(S_{k-1})-1)|_0 + \res_{n=k-2} (\sigma_{(k)}) = \frac{k-1}{2k} (r^{-k+1}(\SC(S_k)-1)|_0 if $S_{k-1}$ is the $(k-1)$-th approximate solution of the singular Yamabe problem. Then the left-hand side is well-defined and the right-hand side vanishes. This implies the assertion. §.§ Renormalized volume coefficients (adapted coordinates) Here we consider the first three coefficients in the expansion of $v(s)$ as defined in (<ref>) in adapted coordinates. As usual we identify $\rho$ with $\eta^*(\rho)$. We also recall that derivatives with respect to $s$ are denoted by a prime. We first derive formulas for $v_1$ and $v_2$ from Proposition <ref>. Note that a^{-1/2} = (1-2s\rho)^{-1/2} = 1 + s \rho_0 + s^2 (\frac{3}{2} \rho_0^2 + \rho'_0) + \cdots. Hence the factorization $v(s) = a^{-1/2} \mathring{v}(s)$ and the identities \begin{equation*} \mathring{v}_1 = \frac{1}{2} \tr (h_{(1)}) \quad \mbox{and} \quad \mathring{v}_2 = \frac{1}{8} \tr (h_{(1)})^2 + \frac{1}{2} \tr (h_{(2)}) - \frac{1}{4} \tr (h_{(1)}^2)) \end{equation*} show that the expansion of $v(s)$ starts with \begin{align*} & 1 + \frac{1}{2} (2 \rho_0 + \tr (h_{(1)})) s \\ & + \frac{1}{2} \left(\tr (h_{(2)}) - \frac{1}{2} |h_{(1)}|^2 + \frac{1}{4} \tr(h_{(1)})^2 + \rho_0 \tr (h_{(1)}) + 3 \rho_0^2 + 2 \rho'_0 \right) s^2 + \cdots. \end{align*} Now (<ref>) and $\rho_0 = -H$ (Lemma <ref>) imply that the linear coefficient equals $(n-1)H$. Hence $v_1=(n-1)H$. This result fits with the formula <cit.> for its integral. Moreover, using Lemma <ref>, we \frac{1}{2} \left( |\lo|^2 - \overline{\Ric}_{00} - 2 |L|^2 + n^2 H^2 - 2n H^2 + 3 H^2 + 2 \bar{\Rho}_{00} + 2 \frac{|\lo|^2}{n-1}\right) for the quadratic coefficient. By $\overline{\Ric}_{00} = (n-1) \bar{\Rho}_{00} + \bar{\J}_0$ and $|L|^2 =|\lo|^2 + n H^2$, the latter sum simplifies to \begin{equation}\label{v2} v_2 = \frac{1}{2} \left(-\frac{n-3}{n-1} |\lo|^2 - (n-3) \bar{\Rho}_{00} + (n-1)(n-3) H^2 - \bar{\J}_0 \right). \end{equation} In particular, (<ref>) shows that $2 v_2 = - \bar{\J}_0$ for $n=3$. This proves the relation (<ref>) in Conjecture <ref> for $n=3$. In the following examples, we demonstrate how the identity (<ref>) can be used to calculate the renormalized volume coefficients $v_k$ (for $k=1,2,3$). This alternative method does not require the calculation of composition of $L$-operators and does not use explicit formulas for the Taylor coefficients of $h_s$. The restriction of (<ref>) to $s=0$ implies $v_1=-(n-1)\rho_0$. Now the fact $\rho_0 = \iota^* \rho = -H$ (Lemma <ref>) implies $v_1 = (n-1)H$. We restrict the derivative of (<ref>) in $s$ to $s=0$. Then 2v_2 - v_1^2 = -(n\!-\!3) \rho'_0 - \bar{\J}_0 - 2(n\!-\!1) \rho^2_0. Now we combine this result with the value of $v_1$ (Example <ref>) and Lemma <ref> to conclude that \begin{align}\label{v2n} -2v_2 & = \bar{\J}_0 + (n\!-\!3) \left(\bar{\Rho}_{00} + \frac{1}{n\!-\!1} |\lo|^2 \right) - (n\!-\!3)(n\!-\!1) H^2 \notag \\ & = \bar{\J}_0 + (n\!-\!3) \rho_0' - (n\!-\!3)(n\!-\!1) H^2. \end{align} This result fits with (<ref>) and with the formula <cit.> for its integral. Using (<ref>), we finally obtain \begin{equation}\label{v2n-2} -2v_2 = \J + (n\!-\!2) \bar{\Rho}_{00} + \frac{2n\!-\!5}{2(n\!-\!1)} |\lo|^2 - (n\!-\!2)(n\!-\!3/2) H^2. \end{equation} In particular, for $n=2$, we have \begin{equation}\label{v2-2} -2v_2 = \J - \frac{1}{2} |\lo|^2. \end{equation} For closed $M$, the total integral of this quantity is a conformal invariant (by Gauss-Bonnet). Note that $2v_2=\QC_2$ (by Example <ref>) which confirms Theorem <ref> for $n=2$. The equality of the coefficients of $s^2$ in (<ref>) yields the identity 3v_3 - 3 v_1 v_2 + v_1^3 = - \frac{n\!-\!5}{2} \rho''_0 - \bar{\J'}_0 - 4(n\!-\!2) \rho'_0 \rho_0 - 2 \rho_0 \bar{\J}_0 - 4(n\!-\!1) \rho_0^3. Hence, using $\rho_0 = -H$, we obtain 3v_3 = 3 v_1 v_2 - v_1^3 - \frac{n\!-\!5}{2} \rho''_0 - \bar{\J}'_0 + 4(n\!-\!2) \rho'_0 H + 2 H \bar{\J}_0 + 4(n\!-\!1) H^3. Combining this formula with the formulas for $v_1$ and $v_2$ in Example <ref> and Example <ref> gives 6 v_3 = (n\!-\!5)(n\!-\!3)(n\!-\!1) H^3 - (n\!-\!5)(3n\!-\!5) H \rho'_0 - (3n\!-\!7) H \bar{\J}_0 - (n\!-\!5) \rho''_0 - 2 \bar{\J}'_0. In particular, this formula implies 6 v_3 = - 8 H \bar{\J}_0 - 2 \bar{\J}'_0. if $n=5$. This proves the relation (<ref>) in Conjecture <ref> for $n=5$. Finally, we note that $\iota^* \nabla_{\NV}^2(\rho)$ corresponds to $\rho_0'' - 2 \rho_0 \rho_0'$ (see Example <ref>). Hence we can rewrite the latter formula as \begin{align}\label{v3-inter} 6 v_3 & = (n\!-\!5)(n\!-\!3)(n\!-\!1) H^3 - (3n\!-\!7) H \iota^* \bar{\J} \notag \\ & - (n\!-\!5)(3n\!-\!7) H \iota^* \nabla_\NV (\rho)- (n\!-\!5) \iota^* \nabla_\NV^2(\rho) - 2 \iota^* \nabla_\NV(\bar{\J}). \end{align} In particular, for $n=3$ we get 6 v_3 = - 2 H \iota^* \bar{\J} + 4 H \iota^* \nabla_\NV (\rho) + 2 \iota^* \nabla_\NV^2(\rho) - 2 \iota^* \nabla_\NV( \bar{\J} ). This quantity actually equals a multiple of $\QC_3$, up to a divergence term; for a discussion of $\QC_3$ we refer to Example <ref>. By <cit.>, the result (<ref>) implies that[There seems to be a misprint in the contribution of the term $H \iota^* \J$.] -\iota^* L(-n+1)L(-n+2)L(-n+3)(1) = 6 (n-1)(n-2)(n-3) v_3, up to a divergence term. Therefore, the result confirms the formula for $c_3$ in Theorem <ref>. In order to express the sum in (<ref>) in terms of standard curvature terms, it remains to determine $\iota^* \nabla_\NV^k (\rho)$ for $k=1,2$. Explicit formulas for these terms will be derived in Section <ref> (Lemma <ref>, Lemma <ref>). The case $k=2$ was first treated in <cit.> (see Remark <ref>). It is only here where we need the full information of Proposition Of course, the renormalized volume coefficients $v_k$, which are defined in terms of adapted coordinates, are to be distinguished from the renormalized volume coefficients $w_k$, which are defined in terms of geodesic normal coordinates <cit.>. By (<ref>), the latter ones are defined by the w(r) = (1 + \sigma_{(2)} r + \sigma_{(3)} r^2 + \cdots)^{-(n+1)} u(r). In particular, we find $w_1 = -(n+1) \sigma_{(2)} + u_1 = \frac{n-1}{2} H$. The above relation implies w_2 = 6 \sigma_{(2)}^2 - 3 \sigma_{(3)} - 3 \sigma_{(2)} u_1 + u_2 for $n=2$, and its evaluation yields 2 w_2 = - \J + \frac{1}{2} |\lo|^2. Note that $w_1 = v_1$ (for $n=1$) and $w_2 = v_2$ (for $n=2$). In general, the coefficients $w_n$ and $v_n$ differ by a non-trivial total divergence. In particular, we find w_3 = -20 \sigma_{(2)}^3 + 20 \sigma_{(2)} \sigma_{(3)} - 4 \sigma_{(4)} + 10 \sigma_{(2)}^2 u_1 - 4 \sigma_{(3)} u_1 - 4 \sigma_{(2)} u_2 + u_3 for $n=3$, and an evaluation yields \begin{align*} 6 v_3 & = 6 w_3 + \Delta(H). \end{align*} Explicit formulas for $w_1, w_2$ (in general dimensions) were derived in <cit.>. In Section <ref>, these coefficients will be described in terms of $L$-operators. the last relation follows from the later formula for v_3 and a calculation of w_3 (which we only did in a manuscript) §.§ Low-order Taylor coefficients of $\rho$ Assuming that $\sigma$ satisfies the condition $\SCY$, we derive formulas for the first few Taylor coefficients of $\rho$ in the variable $s$. We first use Lemma <ref> to derive formulas for the restrictions of $\rho$ and $\nabla_\NV(\rho)$ to $M$. The following result reproves part of <cit.>. Let $n \ge 2$. Then $\iota^* \rho = - H$ and \begin{equation}\label{rho-1} \iota^* \nabla_\NV(\rho) = \Rho_{00} + \frac{|\lo|^2}{n-1}. \end{equation} We calculate in geodesic normal coordinates. We expand the defining relation \begin{equation}\label{rho-def} \Delta_g (\sigma) = -(n\!+\!1) \rho - \sigma \J \end{equation} of $\rho$ into a power series of $r$. The Laplacian takes the form $\partial_r^2 + \frac{1}{2} \tr (h_r^{-1} h'_r) \partial_r + \Delta_{h_r}$. Hence the restriction of (<ref>) to $r=0$ implies $2 \sigma_{(2)} + n H = -(n+1)\rho_0$, and Lemma <ref> yields the first assertion. Next, we restrict the derivative of (<ref>) in $r$ to $r=0$. Then 6 \sigma_{(3)} + \frac{1}{2} \partial_r (\tr(h_r^{-1} h'_r))|_0 + \tr(h_r^{-1} h'_r)|_0 \sigma_{(2)} = -(n\!+\!1) \partial_r(\rho)|_0 - \J_0. But (<ref>) implies the identities \tr(h_r^{-1} h'_r)|_0 = 2 \tr (L) \quad \mbox{and} \quad \partial_r (\tr(h_r^{-1} h'_r))|_0 = - 2|L|^2 - 2 \Ric_{00}. Hence the above relation transforms into 6 \sigma_{(3)} - |L|^2 - \Ric_{00} + 2 \tr(L) \sigma_{(2)} = -(n\!+\!1) \partial_r(\rho)|_0 - \J_0. We combine this result with (<ref>) for $\Ric_{00}$, (<ref>) for $\sigma_{(3)}$ and $|L|^2 = |\lo|^2 + n H^2$ to obtain \partial_r(\rho)|_0 = \iota^* (\J^g) - \J^h + \frac{|\lo|^2}{2(n\!-\!1)} + \frac{n}{2} H^2. Now (<ref>) implies the assertion. The equation (<ref>) justifies the notation $\rho_0'$ in Proposition <ref>. Next, we provide an alternative proof of Lemma <ref>. This illustrates the efficiency of the differential equation for $\rho$ (Lemma <ref>), the resulting recursive formula in Proposition <ref> and the formula for the obstruction (Theorem <ref>) in low-order cases. First of all, the restriction of (<ref>) to $s=0$ yields $n \rho(0) + \frac{1}{2} \tr (h_{(1)}) = 0$. Using $h_{(1)} = 2L$ by (<ref>), we get $\rho(0) = -H$. This reproves the first part of Lemma <ref>. Next, the formula (<ref>) for $k=1$ yields (n-1) \rho'_0 - 2 \rho_0 \left(\frac{\mathring{v}'}{\mathring{v}}\right)|_0 + \partial_s \left(\frac{\mathring{v}'}{\mathring{v}} \right)|_0 + \J_0 = 0 or, equivalently, \begin{equation}\label{rho-prime} (n-1) \rho_0' - \rho_0 \tr (h_{(1)}) + \tr (h_{(2)}) - \frac{1}{2} \tr (h_{(1)}^2) + \J_0 = 0. \end{equation} By (<ref>) and $\tr(L^2) = \tr (\mathring{L}^2) + n H^2$, this gives \begin{align*} 0 & = (n-1) \rho_0' + 2n H^2 + \tr (\mathring{L}^2) - \Ric_{00} - 2 \tr (L^2) + \J_0 \\ & = (n-1) \rho_0' - \Ric_{00} - |\mathring{L}|^2 + \J_0 \\ & = (n-1) \rho_0' - (n-1) \Rho_{00}- |\mathring{L}|^2 . \end{align*} This reproves the second part of Lemma <ref>. For $n=1$, Theorem <ref> states that \B_1 = - \partial_s \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 + 2 \rho_0 \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 - \J_0 or, equivalently, \B_1 = -\tr (h_{(2)}) + \frac{1}{2} \tr (h_{(1)}^2) + \rho_0 \tr (h_{(1)}) - \J_0. By $\lo = 0$, $\tr(h_{(1)}) = 2H$, $\tr(h_{(1)}^2) = 4 H^2$, $\tr(h_{(2)}) = -\Ric_{00}$ and $\rho_0 = -H$, this formula simplifies to \B_1 = \Ric_{00} - \J_0 = \Ric_{00} - K_0 = 0 using $\Ric = K g$, $K$ being the Gauss curvature of $g$. We continue discussing the second-order derivative of $\rho$ in general dimensions $n \ge 3$. The arguments are parallel to the discussion in Section <ref>. Proposition <ref> for $k=2$ gives \begin{equation*} -(n\!-\!2) \rho_0'' = \partial_s^2 \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 - 4 \rho_0 \partial_s \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 - 4 \rho_0' \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 +2 \J'_0. \end{equation*} This formula is equivalent to \begin{align}\label{rho-second} & -(n\!-\!2) \rho_0'' \notag \\ & = \tr ( 3 h_{(3)} - 3 h_{(1)} h_{(2)} + h_{(1)}^3) - 4 \rho_0 \tr (h_{(2)}) + 2 \rho_0 \tr (h_{(1)}^2) - 2 \rho_0' \tr (h_{(1)}) + 2 \J_0'. \end{align} In order to make that formula explicit, we use Proposition <ref> and in particular Corollary <ref>. We obtain \begin{align}\label{rho-second-sum} & -(n\!-\!2) \rho_0'' = -\nabla_{\partial_s}(\Ric)_{00} + 2 \partial_s(\J) - 4 \lo^{ij} R_{0ij0} + \Delta H - H \Ric_{00} - 3 H|\lo|^2 - 2n H \rho_0' \notag \\ & - 6 \tr (L^2 \lo) + 6 L^{ij} R_{0ij0} + 8 \tr (L^3) + 4 H |\lo|^2 - 4 H \Ric_{00} - 8 H |L|^2. \end{align} Now, by the obvious relation -\nabla_{\partial_s}(\Ric)_{00} + 2 \partial_s(\J) = - \nabla_{\partial_s} (G)_{00} - (n-2) \partial_s(\J) for the Einstein tensor $G = \Ric - \frac{1}{2} \scal g$ and the identity (<ref>), we get \begin{align*} -(n\!-\!2) \rho_0'' & = \delta^h (\Ric(\partial_s,\cdot)) - \lo^{ij} \Ric_{ij} + (n\!+\!1) H \Ric_{00} - H \scal - (n\!-\!2) \J_0' - 2n H \rho_0' \notag \\ & - 4 \lo^{ij} R_{0ij0} + \Delta H - H \Ric_{00} - 3 H|\lo|^2 \notag \\ & - 6 \tr (L^2 \lo) + 6 L^{ij} R_{0ij0} + 8 \tr (L^3) + 4 H |\lo|^2 - 4 H \Ric_{00} - 8 H |L|^2 . \end{align*} The decomposition (<ref>) shows that \begin{align*} L^{ij} R_{0ij0} & = L^{ij} (W_{0ij0} + \Rho_{ij} + \Rho_{00} g_{ij}) \\ & = L^{ij} W_{0ij0} + L^{ij} \Rho_{ij} + n H \Rho_{00} \\ & = L^{ij} W_{0ij0} + \lo^{ij} \Rho_{ij} + H (\J - \Rho_{00}) + n H \Rho_{00} \\ & = L^{ij} W_{0ij0} + \lo^{ij} \Rho_{ij} + H \J + (n-1) H \Rho_{00} \\ & = \lo^{ij} W_{0ij0} + \lo^{ij} \Rho_{ij} + H \Ric_{00}. \end{align*} Similarly, we find \lo^{ij} R_{0ij0} = \lo^{ij} W_{0ij0} + \lo^{ij} \Rho_{ij}. The latter two results and the formula (<ref>) for $\rho_0'$ imply \begin{align*} -(n\!-\!2) \rho_0'' & = \delta^h (\Ric(\partial_s,\cdot)) - \lo^{ij} \Ric_{ij} + (n\!+\!1) H \Ric_{00} - H \scal - (n\!-\!2) \J_0' \notag \\ & - 4 \lo^{ij} W_{0ij0} - 4 \lo^{ij} \Rho_{ij} + \Delta H - H \Ric_{00} - 3 H|\lo|^2 \notag \\ & - 6 \tr (L^2 \lo) + 6 \lo^{ij} W_{0ij0} + 6 \lo^{ij} \Rho_{ij} + 6 H \Ric_{00} \notag \\ & + 8 \tr (L^3) + 4 H |\lo|^2 - 4 H \Ric_{00} - 8 H |L|^2 - 2n H \left(\Rho_{00} + \frac{|\lo|^2}{n\!-\!1} \right). \end{align*} Simplification gives \begin{align*} -(n\!-\!2) \rho_0'' & = \delta^h (\Ric(\partial_s,\cdot)) + \Delta H + 2 \lo^{ij} W_{0ij0} + 2 \lo^{ij} \Rho_{ij} - \lo^{ij} \Ric_{ij} \notag \\ & + (n\!+\!2) H \Ric_{00} - 2n H \Rho_{00} - H \scal \notag \\ & + H|\lo|^2 - 8 H |L|^2 - 6 \tr (L^2 \lo) + 8 \tr (L^3) - \frac{2n}{n\!-\!1} H |\lo|^2 - (n\!-\!2) \J_0'. \end{align*} Further simplification using |L|^2 = |\lo|^2 + n H^2, \quad \tr (L^2 \lo) = 2 H |\lo|^2 \quad \mbox{and} \quad \tr (L^3) = \tr (\lo^3) +3 H |\lo|^2 + n H^3 \begin{align}\label{rho2-a} -(n\!-\!2) \rho_0'' & = \delta^h (\Ric(\partial_s,\cdot)) + \Delta H + 2 (\lo^{ij} W_{0ij0} + \tr (\lo^3)) - (n\!-\!3) \lo^{ij} \Rho^g_{ij} \notag \\ & + \frac{(n\!-\!2)(n\!+\!1)}{n\!-\!1} H \Ric_{00} - \frac{n\!-\!2}{n\!-\!1} H \scal + \frac{3n\!-\!5}{n\!-\!1} H |\lo|^2 - (n\!-\!2) \J_0' . \end{align} Next, we apply the basic identity (<ref>). It implies that \begin{equation}\label{F-L-trace} \lo^{ij} W_{0ij0} + \tr (\lo^3) \stackrel{!}{=} (n-2) \lo^{ij} \JF_{ij}. \end{equation} Now, combining the formula (<ref>) for $\Delta H$ and (<ref>) with (<ref>), yields \begin{align*} -(n\!-\!2) \rho_0'' & = \frac{n\!-\!2}{n\!-\!1} \delta^h (\Ric(\partial_s,\cdot)) + \frac{1}{n\!-\!1} \delta \delta (\lo) + 2 (n\!-\!2) \lo^{ij} \JF_{ij} - (n\!-\!3) \lo^{ij} \Rho^g_{ij} \\ & + \frac{(n\!-\!2)(n\!+\!1)}{n\!-\!1} H \Ric_{00} - \frac{n\!-\!2}{n\!-\!1} H \scal + \frac{3n\!-\!5}{n\!-\!1} H |\lo|^2 - (n\!-\!2) \J_0'. \end{align*} \begin{align*} \rho_0'' & = - \frac{1}{(n\!-\!1)(n\!-\!2)} \delta \delta (\lo) - \frac{1}{n\!-\!1} \delta^h (\Ric(\partial_s,\cdot)) - 2 \lo^{ij} \JF_{ij} + \frac{n\!-\!3}{n\!-\!2} \lo^{ij} \Rho^g_{ij} \\ & - \frac{n\!+\!1}{n\!-\!1} H \Ric_{00} + \frac{1}{n\!-\!1} H \scal - \frac{3n\!-\!5}{(n\!-\!1)(n\!-\!2)} H |\lo|^2 + \J_0' . \end{align*} Finally, we use the defining relation $\iota^* \Rho^g = \JF + \Rho^h - H \lo - 1/2 h H^2$ of $\JF$ to replace in this identity the Schouten tensor $\Rho^g$ by the Schouten tensor $\Rho^h$. Thus, we have proved the following result. For $n \ge 3$, it holds \begin{align}\label{rho-2-ex} \rho_0'' & = - \frac{1}{(n\!-\!1)(n\!-\!2)} \delta \delta (\lo) - \frac{1}{n\!-\!1} \delta^h (\Ric^g(\partial_s,\cdot)) - \frac{n\!-\!1}{n\!-\!2} (\lo,\JF) + \frac{n\!-\!3}{n\!-\!2} (\lo,\Rho^h) \notag \\ & - \frac{n\!+\!1}{n\!-\!1} H \Ric_{00} + \frac{1}{n\!-\!1} H \scal - \frac{n\!+\!1}{n\!-\!1} H |\lo|^2 + \J_0'. \end{align} In Section <ref>, we shall connect this result with the holographic formula for $\QC_3$. Lemma <ref> is equivalent to <cit.> and <cit.>, up to the sign of the term $(\lo,\Rho^h)$. In fact, the quoted result calculates the restriction of $\nabla_{\NV}^2(\rho)$ to $M$. By Example <ref>, it corresponds to $\rho''_0+ 2 H \rho_0'$. But (<ref>) implies \begin{align*} \rho''_0+ 2 H\rho_0' & = (\cdots) -\frac{n\!+\!1}{n\!-\!1} H \Ric_{00} + \frac{1}{n\!-\!1} H \scal - \frac{n\!+\!1}{n\!-\!1} H |\lo|^2 + \J_0' \\ & + 2 H \left(\frac{1}{n\!-\!1} \Ric_{00} - \frac{1}{n\!-\!1} \J_0 + \frac{|\lo|^2}{n\!-\!1}\right) + \J_0' \\ & = (\cdots) - H \Ric_{00} + 2 H \J_0 - H|\lo|^2 + \J_0' \\ & = (\cdots) - H ((n\!-\!1) \Rho_{00} + |\lo|^2) + H \J_0 + \J_0', \end{align*} where $(\cdots)$ indicates the four terms in the first line of (<ref>). This proves the claim. The present alternative proof rests on the recursive formula in Proposition <ref> and the explicit formula for $h_{(3)}$ in Proposition <ref>. The proof of Proposition <ref> shows that the calculation of the coefficient $h_{(2)}$ involves $h_{(1)}$ and $\rho_0$. Once $h_{(2)}$ has been determined, we calculated $\rho_0'$ using (<ref>). Similarly, the calculation of $h_{(3)}$ in the proof of Lemma <ref> involves the lower-order coefficients of $h_s$ and $\rho_0$, $\rho_0'$. Once $h_{(3)}$ has been determined, the recursive formula (<ref>) yields $\rho_0''$. A continuation of that iterative process yields the higher-order coefficients, at least in principle. We finish this section with some comments concerning the functions $\rho$ and the singular Yamabe obstructions in Examples <ref>–<ref>. In Example <ref>, it holds $\mathring{v}'/\mathring{v} = - n/(1-2s)$. Hence the differential equation (<ref>) reduces to - s \rho' + n \rho - n(1-2s \rho)/(1-2s) = 0. One readily checks that $\rho=1$ is the unique solution with initial value $\rho(0) = 1$. The vanishing of the obstruction is reproduced by the formula (<ref>). Indeed, we obtain (n+1)! \B_n = 2n \partial_s^n (1/(1-2s))|_0 - 4 n^2 \partial_s^{n-1} (1/(1-2s))|_0 = 2n 2^n n! - 4 n^2 2^{n-1} (n-1)! = 0. Similarly, in Example <ref>, it holds $\mathring{v}'/\mathring{v} = - ns/(1-s^2)$ and the differential equation (<ref>) reduces to - s\rho' + n\rho - n s (1-2s\rho)/(1-s^2) + \tfrac{n+1}{2} s = 0. One easily checks that $\rho = \frac{1}{2} s$ is the unique solution with the initial value $\rho(0)=0$. Again, the vanishing of the obstruction is reproduced by (n+1)! \B_n = 2 \partial_s^n(ns/(1-s^2))|_0 - 4 \binom{n}{2} \partial_s^{n-2}(ns/(1-s^2))|_0 = 2n n! - 4n \binom{n}{2} (n-2)! = 0 for odd $n$ and trivially for even $n$. §.§ The obstruction $\B_2$ for general backgrounds In the present section, we prove the equivalence of both formulas for the singular Yamabe obstruction $\B_2$ displayed in (<ref>) and derive the second of these formulas from Theorem <ref>. The Codazzi-Mainardi identity states that \begin{equation}\label{CME} \nabla^h_Y (L)(X,Z) - \nabla^h_X(L)(Y,Z) = R^g (X,Y,Z,N) \end{equation} for $X,Y,Z \in \mathfrak{X}(M)$ if $L(X,Y) = - h(\nabla^g_X(Y),N)$ for some unit normal vector field $N$ <cit.>. Now, we decompose the curvature tensor $R$ as \begin{align}\label{KN} & R(X,Y,Z,W) = W(X,Y,Z,W) \\ & + \Rho(Y,Z) g (X,W) - \Rho(X,Z) g(Y,W) - \Rho(Y,W)g(X,Z) + \Rho(X,W)g(Y,Z), \notag \end{align} where $W$ is the trace-free Weyl tensor (see (<ref>)), and take traces in (<ref>) in the arguments $X,Z$. Then n \langle dH, Y \rangle - \langle \delta^h (L), Y \rangle = -(n-1) \Rho(N,Y). \delta^h (\lo) - (n-1) dH = (n-1) \Rho(N,\cdot) [Lemma 6.25.2]J1). Now, taking a further divergence, yields \begin{equation}\label{ddL} \delta^h \delta^h (\lo) - (n-1) \Delta H = (n-1) \delta^h (\Rho(N,\cdot)). \end{equation} For $n=2$, this proves the equivalence of both formulas in (<ref>). We continue by showing that the second formula for $\B_2$ in (<ref>) is a special case of Theorem <ref>. First, we observe that the formula - 3 \B_2 = \partial_s^2 \left( \frac{\mathring{v}'}{\mathring{v}} \right)|_0 - 4 \rho_0 \partial_s \left (\frac{\mathring{v}'}{\mathring{v}} \right)|_0 - 4 \rho_0' \left(\frac{\mathring{v}'}{\mathring{v}}\right)|_0 + 2 \J_0' in Theorem <ref> is equivalent to -3 \B_2 = \tr ( 3 h_{(3)} - 3 h_{(1)} h_{(2)} + h_{(1)}^3) - 4 \rho_0 \tr (h_{(2)}) + 2 \rho_0 \tr (h_{(1)}^2) - 2 \rho_0' \tr (h_{(1)}) + 2 \J_0'. In order to make that sum explicit, we use the results in Proposition <ref>. By Corollary <ref> for $n=2$ and $\rho_0 = - H$, we obtain \begin{align}\label{B2-sum} -3 \B_2 & = -\nabla^g_{\partial_s}(\Ric)_{00} + 2 \partial_s(\J) - 4 \lo^{ik} R_{0ik0} + \Delta H - H \Ric_{00} - 3 H|\lo|^2 \\ & - 6 \tr (L^2 \lo) + 6 L^{ij} R_{0ij0} + 8 \tr (L^3) + 4 H |\lo|^2 - 4 H \Ric_{00} - 8 H |L|^2 - 4 H \rho_0'. \notag \end{align} Now we prove the identity \begin{equation}\label{Einstein-00} \nabla^g_{\partial_s}(G)_{00} = -\delta^h (\Ric(\partial_s,\cdot)) + \lo^{ij} \Ric_{ij} - (n+1) H \Ric_{00} + H \scal \end{equation} for the Einstein tensor $G \st \Ric - \frac{1}{2} \scal g$. Note that $G= \Ric - 2 \J g$ in dimension $n=3$. It is well-known that the second Bianchi identity implies the relation $2 \delta^g (\Ric) = d\scal$. Hence \begin{align*} \nabla^g_{\partial_s}(\Ric)(\partial_s,\partial_s) & = \delta^g(\Ric)(\partial_s) - g^{ij} \nabla^g_{\partial_i}(\Ric)(\partial_j,\partial_s) \\ & = \frac{1}{2} \langle d \scal,\partial_s \rangle \\ & - g^{ij} \partial_i (\Ric(\partial_j,\partial_s)) + g^{ij} \Ric(\nabla^g_{\partial_i}(\partial_j),\partial_s) + g^{ij} \Ric(\partial_j,\nabla^g_{\partial_i}(\partial_s)) \\ & = \frac{1}{2} \langle d \scal,\partial_s \rangle \\ & - h^{ij} \partial_i (\Ric(\partial_j,\partial_s)) + h^{ij} \Ric (\nabla^h_{\partial_i}(\partial_j) - L_{ij} \partial_s,\partial_s) + h^{ij} \Ric(\partial_j,\nabla^g_{\partial_i}(\partial_s)) \\ & = \frac{1}{2} \langle d \scal,\partial_s \rangle - \delta^h (\Ric(\partial_s,\cdot)) - n H \Ric_{00} + h^{ij} \Ric(\partial_j,\nabla^g_{\partial_i}(\partial_s)) \end{align*} on $M$. Therefore, using $\nabla_{\partial_i}^g(\partial_s) = L_i^k \partial_k$, we obtain \begin{align*} \nabla^g_{\partial_s}(G)_{00} & = -\delta^h (\Ric(\partial_s,\cdot)) - n H \Ric_{00} + h^{ij} L_i^k \Ric_{jk} \\ & = -\delta^h (\Ric(\partial_s,\cdot)) - n H\Ric_{00} + L^{ij} \Ric_{ij}. \end{align*} This implies (<ref>). Next, we observe that the decomposition (<ref>) yields \begin{align*} L^{ij} R_{0ij0} & = L^{ij} (W_{0ij0} + \Rho_{ij} + \Rho_{00} g_{ij}) \\ & = L^{ij} \Rho_{ij} + 2 H \Rho_{00} \\ & = \lo^{ij} \Rho_{ij} + H (\J - \Rho_{00}) + 2 H \Rho_{00} = \lo^{ij} \Rho_{ij} + H \J + H \Rho_{00} \\ & = \lo^{ij} \Rho_{ij} + H \Ric_{00} \end{align*} since the Weyl tensor $W$ vanishes in dimension $3$. Similarly, we find \begin{equation*} \lo^{ij} R_{0ij0} = \lo^{ij} \Rho_{ij}. \end{equation*} These results and the formula (<ref>) for $\rho_0'$ show that the sum (<ref>) simplifies to (recall that $n=2$) \begin{align*} & \delta^h (\Ric(\partial_s,\cdot)) - \lo^{ij} \Ric_{ij} + 3 H \Ric_{00} - H \scal - 4 \lo^{ij} \Rho_{ij} + \Delta H - H \Ric_{00} - 3 H |\lo|^2 \\ & - 6 \tr (L^2 \lo) + 6 \lo^{ij} \Rho_{ij} + 6 H \Ric_{00} + 8 \tr(L^3) + 4 H |\lo|^2 - 4 H \Ric_{00} - 8 H |L|^2 \\ & - 4H \Ric_{00} + H \scal - 4 H |\lo|^2 \\ & = \delta^h (\Ric(\partial_s,\cdot)) + \Delta H + \lo^{ij} \Ric_{ij}, \end{align*} up to terms that are at least quadratic in $L$. In order to deal with these terms, we note that |L|^2 = |\lo|^2 + 2 H^2, \quad \tr (L^2 \lo) = 2 H |\lo|^2 \quad \mbox{and} \quad \tr (L^3) = 3 H |\lo|^2 + 2 H^3 using $\tr(\lo^3)=0$. It follows that the remaining terms are $(- 3 - 12 + 24 + 4 - 8 - 4) H |\lo|^2 = H |\lo|^2$. Summarizing we obtain -3 \B_2 = \delta^h (\Ric(\partial_s,\cdot)) + \Delta H + \lo^{ij} \Ric_{ij}+ H |\lo|^2. This proves the second formula for $\B_2$ in (<ref>). The identity (<ref>) is equivalent to <cit.> or <cit.>. In addition to the derivation of the above formula for the singular Yamabe obstruction $\B_2$, in these references, this crucial identity is used to prove a formula for the second normal derivative of $\rho$ and a formula for $\QC_3$. The calculations lead to the same results as here. Our discussion of the second normal derivative of $\rho$ and the explicit formula for $\QC_3$ is contained in Sections <ref> and <ref>. It uses the identity (<ref>). §.§ The obstruction $\B_3$ for conformally flat backgrounds We first evaluate Theorem <ref> for the obstruction $\B_3$ of a three-manifold $M \hookrightarrow \R^4$. The resulting formula is equivalent to a formula in <cit.> (see Lemma <ref>). This will imply a formula for $\B_3$ in a conformally flat background. The Yamabe obstruction of a hypersurface $(M^3,h)$ in the four-dimensional flat space $\R^4$ is given by the formula \begin{equation}\label{B3-final} 12 \B_3 = \Delta (|\lo|^2) + 12 |dH|^2 + 6 (\lo,\Hess (H)) - 2 |\lo|^4 + 6 \tr (\lo^4) + 6 H \tr (\lo^3) \end{equation} or, equivalently, \begin{equation}\label{B3-final2} 12 \B_3 = \Delta (|\lo|^2) + 12 |dH|^2 + 6 (\lo,\Hess (H)) + |\lo|^4 + 6 H \tr (\lo^3). \end{equation} Here the Hessian, the Laplacian, scalar products, and traces are taken with respect to the metric $h$ on $M$. Newton's identity 24 \sigma_4 (L) = \tr(L)^4 - 6 \tr(L)^2 |L|^2 + 3 |L|^4 + 8 \tr(L) \tr(L^3) - 6 \tr(L^4) for the elementary symmetric polynomial $\sigma_4(L)$ of the eigenvalues of $L$ (or rather of the shape operator) implies \begin{align*} 24 \sigma_4(L) & = n(n-1)(n-2)(n-3) H^4 - 6 (n-2)(n-3) H |\lo|^2 + 8(n-3) H \tr(\lo^3) \\ & + 3 (|\lo|^4 - 2 \tr(\lo^4)). \end{align*} Since $\sigma_4(L) = 0$ in dimension $n=3$, we get $|\lo|^4 = 2 \tr(\lo^4)$ for $n=3$. Corollary <ref> implies the equivalence of (<ref>) and (<ref>). We derive formula (<ref>) as a consequence of the identity \begin{equation}\label{3-flat} 12\B_3 = - \partial_s^3 \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 + 6 \rho_0 \partial_s^2 \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 + 12 \rho_0' \partial_s \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 + 6 \rho_0'' \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 - 3 \partial_s^2(\bar{\J})|_0 \end{equation} in Theorem <ref>. Note that, for a flat background, the last term vanishes. By the relations \begin{align*} \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 & = \frac{1}{2} \tr (h_{(1)}), \\ \partial_s \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 & = \frac{1}{2} \tr (2 h_{(2)} - h_{(1)}^2), \\ \partial_s^2 \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 & = 2! \frac{1}{2} \tr (3 h_{(3)} - 3 h_{(1)} h_{(2)} + h_{(1)}^3), \\ \partial_s^3 \left( \frac{\mathring{v}'}{\mathring{v}}\right)|_0 & = 3! \frac{1}{2} \tr (4 h_{(4)} - 4 h_{(1)} h_{(3)} - 2 h_{(2)}^2 + 4 h_{(1)}^2 h_{(2)} - h_{(1)}^4), \end{align*} the identity (<ref>) is equivalent to \begin{align}\label{B3-flat} 12 \B_3 & = - \tr( 12 h_{(4)} - 12 h_{(1)} h_{(3)} - 6 h_{(2)}^2 + 12 h_{(1)}^2 h_{(2)} - 3 h_{(1)}^4) \notag \\ & + 6 \rho_0 \tr (3 h_{(3)} - 3 h_{(1)} h_{(2)} + h_{(1)}^3) \notag \\ & + 6 \rho_0' \tr (2 h_{(2)} - h_{(1)}^2) \notag \\ & + 3 \rho_0'' \tr (h_{(1)}). \end{align} In order to evaluate that sum, we use the formulas for $h_{(k)}$ ($k \le 3$) in Proposition <ref> and the formulas for the first two normal derivatives of $\rho$ in Section <ref>. In addition, it remains to determine the coefficient $\tr(h_{(4)})$. The following result even provides a closed formula for the Taylor coefficient $h_{(4)}$ of $h_s$ (for a flat background). Let $n=3$. Then \begin{align}\label{h4-ex} 12 h_{(4)} & = - 9 H \Hess (H) + L \Hess (H) + \Hess (H) L - 6 dH \otimes dH - \gamma \notag \\ & - \Hess (|\lo|^2) + 4 \lo^2 |\lo|^2 + 15 H^2 ({(\lo^2)}_\circ + H \lo) - H \lo |\lo|^2 + 3 L \rho_0'', \end{align} \begin{equation}\label{gamma-form} \gamma_{jk} \st 2 h^{lm} (\nabla_j(L)_{km} + \nabla_k(L)_{jm} - \nabla_m(L)_{jk}) \partial_l(H) \end{equation} \begin{equation}\label{rho-pp} \rho_0'' = - \Delta (H) - 2 \tr (\lo^3) - 2 H |\lo|^2. \end{equation} We use the same notation as in the proof of Proposition <ref>. We expand the curvature components $R_{0jk}^0$ of the metric \eta^*(g) = a^{-1} ds^2 + h + h_{(1)} s + h_{(2)} s^2 + h_{(3)} s^3 + h_{(4)} s^4 + \cdots into power series of $s$. We recall that $a = \eta^*(|\NV|^2)$. In order to simplify the notation, we write $g$ for the metric $\eta^*(g)$ and identify $\eta^*(\rho)$ with $\rho$. By assumption, it holds $a = 1- 2s \rho$. As usual, the $0$-components refer to $\partial_s$. Since $g$ is flat, the components $R_{0jk}^0$ vanish. Now, for the above metric, we find the Christoffel symbols \Gamma_{ij}^0 = - \frac{1}{2} g^{00} g_{ij}', \quad \Gamma_{0j}^0 = \frac{1}{2} g^{00} \partial_j (g_{00}), \quad \Gamma_{00}^0 = \frac{1}{2} g^{00} g_{00}', \quad \Gamma_{0k}^l = \frac{1}{2} g^{rl} g_{kr}', where $'$ denotes the derivative in $s$. We recall that $g^{00} = a = 1- 2s \rho$ (by assumption) and $g_{00} = a^{-1} = 1 + 2s\rho + \cdots$. Hence \begin{align}\label{R-adapted-2} 0 \stackrel{!}{=} R_{0jk}^0 & = \frac{1}{2} \Gamma_{jk}^l g^{00} (g_{00})_l + \frac{1}{4} g^{00} g^{rl} g_{kr}' g_{jl}' - \frac{1}{2} ((g^{00})' g_{jk}' + g^{00} g_{jk}'') \notag \\ & - \frac{1}{2} ((g^{00})_j (g_{00})_k + g^{00} (g_{00})_{kj}) - \frac{1}{4} (g^{00})^2 g_{jk}' g_{00}' - \frac{1}{4} (g^{00})^2 (g_{00})_j (g_{00})_k. \end{align} Now we display the Taylor expansions of all $6$ terms in (<ref>) up to order $s^2$. Using these results, it is easy to see that the coefficients of $s$ reproduce the result of the earlier calculation of $h_{(3)}$ in the proof of Proposition <ref> (Remark <ref>). First, we observe that the Christoffel symbols $\Gamma_{jk}^l$ for $g$ restrict to the Christoffel symbols $\Gamma_{jk}^l$ for $h$. Moreover, we recall the general variation formula \delta (\Gamma_{jk}^l) = \frac{1}{2} g^{lm} (\nabla_j (\delta(g))_{km} + \nabla_k (\delta(g))_{jm} - \nabla_{m}(\delta(g))_{jk}). Let $(\Gamma_{jk}^l)' = \delta (\Gamma_{jk}^l)$ denote the variation of the Christoffel symbols for the variation $g = h + 2 L s + \cdots$. By $g_{00} = 1 + 2 s \rho + \cdots$ and $\rho_0 = - H$, it follows that the first term in (<ref>) contributes by \begin{align}\label{term0} s^2 \frac{1}{2} (\Gamma_{jk}^l)' (-2 \partial_l(H)) & = - s^2 (\Gamma_{jk}^l)' \partial_l(H) \notag \\ & = - s^2 h^{lm} (\nabla_j(L)_{km} + \nabla_k(L)_{jm} - \nabla_{m}(L)_{jk}) \partial_l(H) \notag \\ & = - s^2 \frac{1}{2} \gamma_{jk}. \end{align} The remaining contributions of the first term in (<ref>) match with the contributions by the second part of the fourth term to \begin{align}\label{term1} -\frac{1}{2} s [\Hess (2\rho_0)] - \frac{1}{2} s^2 \left[\Hess (2\rho_0' + 4 \rho_0^2) - 2 \rho_0 \Hess (2\rho_0) \right] + \cdots \end{align} with $\Hess$ defined with respect to $h$. Next, the second term contributes by $1/4$ times[Here we use the fact that $[h_{(1)},h_{(2)}] = 0$ for a flat background.] \begin{align}\label{term2} h_{(1)}^2 & + s \big[4h_{(1)} h_{(2)} - h_{(1)}^3 - 2 h_{(1)}^2 \rho_0\big] \notag \\ & + s^2 \big[3 h_{(1)} h_{(3)} + 3 h_{(3)} h_{(1)} + 4 h_{(2)}^2 - 5 h_{(1)}^2 h_{(2)} + h_{(1)}^4 \notag \\ & + 2 h_{(1)}^3 \rho_0 - 4 h_{(1)} h_{(2)} \rho_0 - 4 h_{(2)} h_{(1)} \rho_0 - 2 h_{(1)}^2 \rho_0'\big] + \cdots. \end{align} Finally, we find * Term three contributes by $-1/2$ times \begin{align}\label{term3} 2 h_{(2)} - 2 h_{(1)} \rho_0 & + s \big[6h_{(3)} - 8h_{(2)} \rho_0 - 4 h_{(1)} \rho_0'\big] \notag\\ & + s^2 \big[12 h_{(4)} - 18 h_{(3)}\rho_0 - 12 h_{(2)} \rho_0' - 3 h_{(1)} \rho_0''\big] + \cdots \end{align} * The first part of term four contributes by $-1/2$ times \begin{equation}\label{term4} s^2 (-4) \partial_j(\rho_0) \partial_k(\rho_0) \end{equation} * Term five contributes by $-1/4$ times \begin{equation}\label{term5} 2 h_{(1)} \rho_0 + s \big[4 h_{(2)} \rho_0 + 4 h_{(1)} \rho_0' \big] + s^2 \left[ 6 h_{(3)} \rho_0 + 8 h_{(2)} \rho_0' + 3 h_{(1)} \rho_0'' \right] + \cdots \end{equation} * Term six contributes by $-1/4$ times \begin{equation}\label{term6} s^2 4 \partial_j(\rho_0) \partial_k(\rho_0) \end{equation} Summarizing the coefficients of $s^2$ in (<ref>)-(<ref>) we obtain \begin{align}\label{h4-ev} 0 = & - \frac{1}{2} (\gamma + \Hess (2\rho_0' + 4 \rho_0^2) - 4 \rho_0 \Hess (\rho_0)) \notag \\ & + \frac{1}{4} \big[3 h_{(1)} h_{(3)} + 3 h_{(3)} h_{(1)} + 4 h_{(2)}^2 - 5 h_{(1)}^2 h_{(2)} + h_{(1)}^4 + 2 h_{(1)}^3 \rho_0 - 8 h_{(1)} h_{(2)} \rho_0 - 2 h_{(1)}^2 \rho_0'\big] \notag\\ & - \frac{1}{2} \big[12 h_{(4)} - 18 h_{(3)}\rho_0 - 12 h_{(2)} \rho_0' - 3 h_{(1)} \rho_0'' \big] \notag\\ & - \frac{1}{4} \left[ 6 h_{(3)} \rho_0 + 8 h_{(2)} \rho_0' + 3 h_{(1)} \rho_0'' \right] \notag\\ & + 2 d \rho_0 \otimes d \rho_0 \notag\\ & - d \rho_0 \otimes d\rho_0. \end{align} Note that the third and the fourth line can be summarized to -\frac{1}{4} (24 h_{(4)} - 30 h_{(3)} \rho_0 - 16 h_{(2)} \rho_0' - 3 h_{(1)} \rho_0''). Therefore, we find the formula \begin{align}\label{h4-gen} 24 h_{(4)} & = - 4 \Hess (\rho_0') - 8 \Hess (\rho_0^2) + 8 \rho_0 \Hess(\rho_0) - 2 \gamma \notag \\ & + \big[3 h_{(1)} h_{(3)} + 3 h_{(3)} h_{(1)} + 4 h_{(2)}^2 - 5 h_{(1)}^2 h_{(2)} + h_{(1)}^4 + 2 h_{(1)}^3 \rho_0 - 8 h_{(1)} h_{(2)} \rho_0 - 2 h_{(1)}^2 \rho_0'\big] \notag \\ & +30 h_{(3)} \rho_0 + 16 h_{(2)} \rho_0' + 3 h_{(1)} \rho_0'' \notag \\ & + 4 d \rho_0 \otimes d \rho_0. \end{align} Now we apply the known formulas for $h_{(j)}$ ($j \le 3$) (Proposition <ref>) and $\rho_0,\rho_0',\rho_0''$ (Lemma <ref>, Lemma <ref>) to make that sum fully explicit. First, we prove the remarkable simplification \begin{align}\label{rem-sim} & 3 h_{(1)} h_{(3)} + 3 h_{(3)} h_{(1)} + 4 h_{(2)}^2 - 5 h_{(1)}^2 h_{(2)} + h_{(1)}^4 + 2 h_{(1)}^3 \rho_0 - 8 h_{(1)} h_{(2)} \rho_0 - 2 h_{(1)}^2 \rho_0' \notag \\ & = 2 L \Hess(H) + 2 \Hess(H) L. \end{align} In fact, we calculate \begin{align*} & 3 h_{(1)} h_{(3)} + 3 h_{(3)} h_{(1)} \\ & = 2 L \Hess (H) + 2 \Hess(H) L - 6 H L^2 \lo + 2 L^2 |\lo|^2 - 6 H L \lo L + 2 L^2 |\lo|^2 \end{align*} \begin{align*} 2 h_{(1)}^3 \rho_0 - 8h_{(1)} h_{(2)} \rho_0 - 2 h_{(1)}^2 \rho_0' = -16 H L^3 + 16 H L^2 \lo - 4 L^2 |\lo|^2. \end{align*} The sum of these two results gives 2 L \Hess (H) + 2 \Hess(H) L + 4 H (\lo^3 + 2 H \lo^2 + H^2 \lo) - 16 H (\lo^3 + 3 H \lo^2 + 3 H^2 \lo + H^3 \id). Moreover, we get \begin{align*} & 4 h_{(2)}^2 - 5 h_{(1)}^2 h_{(2)} + h_{(1)}^4 \\ & = 4 \lo^4 + 8 H \lo^3 + 4 H^2 \lo^2 - 20 \lo^4-60 H \lo^3 - 60 H^2 \lo^2 - 20 H^3 \lo \\ & + 16 \lo^4 + 64 H \lo^3 + 96 H^2 \lo^2 + 64 H^3 \lo + 16 H^4 \id \\ & = 12 H \lo^3 + 40 H^2 \lo^2 + 44 H^3 \lo + 16 H^4 \id. \end{align*} Summing these identities proves (<ref>). The above results imply \begin{align*} 12 h_{(4)} & = - \Hess (|\lo|^2) - 4 \Hess (H^2) + 4 H \Hess (H) + 2 d H \otimes dH - \gamma \\ & + L \Hess(H) + \Hess(H) L + \alpha \\ & = - \Hess (|\lo|^2) + L \Hess(H) + \Hess(H) L - 8 H \Hess (H) - 6 dH \otimes dH - \gamma + \alpha, \end{align*} \begin{align*} & \alpha \st 15 h_{(3)} \rho_0 + 8 h_{(2)} \rho_0' + 3/2 h_{(1)} \rho_0'' \\ & = - 5 H \Hess (H) + 15 H^2 L \lo - 5 H L |\lo|^2 + 4 L \lo |\lo|^2 + 3 L \rho_0'' \\ & = - 5 H \Hess (H) + 15 H^2 \lo^2 + 15 H^3 \lo - 5 H \lo |\lo|^2 - 5 H^2 |\lo|^2 \id + 4 \lo^2 |\lo|^2 + 4 H \lo |\lo|^2 \\ & + 3 L \rho_0'' \\ & = -5 H \Hess (H) + 4 \lo^2 |\lo|^2 + 5 H^2 (3 \lo^2 - |\lo|^2 \id) - H \lo |\lo|^2 + 15 H^3 \lo + 3 L \rho_0''. \end{align*} Summarizing the last two results implies the first assertion. The formula for $\rho_0''$ in $n=3$ is a direct consequence of the formula for $\rho_0''$ for general $n$ (Lemma <ref>) using $(\lo,\JF) = \tr (\lo^3)$ and $\delta \delta (\lo) = 2 \Delta (H)$ (by Codazzi-Mainardi). Lemma <ref> implies 12 \tr (h_{(4)}) = - \Delta (|\lo|^2) - 9 H \Delta (H) + 2 (L,\Hess(H)) - 12 |dH|^2 + 4 |\lo|^4 + 9 H \rho_0''. It only remains to prove that \begin{equation}\label{trace-gamma} \tr (\gamma) = 6 (dH,dH). \end{equation} \begin{align*} \tr (\gamma) & = 2 h^{jk} h^{lm} (\nabla_j(L)_{km} + \nabla_k(L)_{jm} - \nabla_m(L)_{jk}) \partial_l(H) \\ & = 2 h^{lm} (\delta(L)_m + \delta (L)_m - \nabla_m (\tr(L))) \partial_l(H) \\ & = 6 h^{lm} \partial_m(H) \partial_l(H) \\ & = 6 (dH,dH) \end{align*} by $\delta(L) = 3 dH$ (Codazzi-Mainardi). The proof is complete. We proceed with the evaluation of (<ref>). \begin{align*} & \tr(12 h_{(1)} h_{(3)} + 6 h_{(2)}^2 - 12 h_{(1)}^2 h_{(2)} + 3 h_{(1)}^4) \\ & = 8 (L,\Hess(H)) + 6 \tr (\lo^4) + 8 |\lo|^4 + 36 H \tr (\lo^3) + 126 H^2 |\lo|^2 + 144 H^4. \end{align*} By the known formulas for the coefficients $h_{(k)}$ for $k \le 3$, we find \begin{align*} & \tr(12 h_{(1)} h_{(3)} + 6 h_{(2)}^2 - 12 h_{(1)}^2 h_{(2)} + 3 h_{(1)}^4) \\ & = 8 \tr (L \Hess (H)) - 24 H \tr (L^2 \lo) + 8 \tr(L^2) |\lo|^2 + 6 \tr (L^2 \lo^2) - 48 \tr (L^3 \lo) + 48 \tr (L^4). \end{align*} The latter sum equals the sum of $8 \tr (L \Hess (H))$ and \begin{align*} & -24 H \tr (\lo^3 + 2H \lo^2) + 8 \tr(\lo^2 + 2H \lo + H^2 \id) |\lo|^2 + 6 \tr( (\lo^2 + 2H \lo + H^2) \lo^2) \\ & - 48 \tr (\lo^4 + 3 H \lo^3 + 3 H^2 \lo^2) + 48 \tr(\lo^4 + 4 H \lo^3 + 6 H^2 \lo^2 + H^4 \id). \end{align*} The result follows by simplification. The following result evaluates the lower-order terms in (<ref>). \begin{align*} & 6 \rho_0 \tr (3 h_{(3)} - 3 h_{(1)} h_{(2)} + h_{(1)}^3) + 6 \rho_0' \tr (2 h_{(2)} - h_{(1)}^2) + 3 \rho_0'' \tr (h_{(1)}) \notag \\ & = - 6 H \Delta (H) - 6 |\lo|^4 - 12 H \tr (\lo^3) - 108 H^2 |\lo|^2 - 144 H^4 + 18 H \rho_0''. \end{align*} By $3 \tr (h_{(3)}) = \Delta (H)$, the sum equals \begin{align*} & -6 H (\Delta (H) - 6 \tr (L^2 \lo) + 8 \tr (L^3)) + 6 |\lo|^2 \tr (L \lo - 2 L^2) + 18 H \rho_0'' \\ & = -6 H \Delta (H) + 36 H \tr (L^2 \lo) - 48 H \tr (L^3) - 6 |\lo|^4 - 36 H^2 |\lo|^2 + 18 H \rho_0'' \\ & = -6 H \Delta (H) + 36 H (\tr(\lo^3) + 2H |\lo|^2) - 48 ( H\tr(\lo^3) + 3H^2 |\lo|^2 + 3 H^4) \\ & - 6 |\lo|^4 -36 H^2 |\lo|^2 + 18 H \rho_0''. \end{align*} Simplification completes the proof. Now we summarize the above results. We obtain \begin{align*} 12 \B_3 & = \Delta (|\lo|^2) + 9 H \Delta (H) - 2 (L,\Hess(H)) + 12 |dH|^2 - 4 |\lo|^4 - 9 H \rho_0'' \\ & + 8 (L,\Hess(H)) + 6 \tr (\lo^4) + 8 |\lo|^4 + 36 H \tr (\lo^3) + 126 H^2 |\lo|^2 + 144 H^4 \\ & - 6 H \Delta (H) - 6 |\lo|^4 - 12 H \tr (\lo^3) - 108 H^2 |\lo|^2 - 144 H^4 + 18 H \rho_0'' \\ & = \Delta (|\lo|^2) + 12 |dH|^2 + 6 (\lo,\Hess(H)) - 2 |\lo|^4 + 6 \tr (\lo^4) \\ & + 9 H \Delta (H) + 18 H^2 |\lo|^2 + 24 H \tr (\lo^3) + 9 H \rho_0''. \end{align*} The relation $\rho_0'' = - \Delta (H) - 2 \tr (\lo^3) - 2 H |\lo|^2$ implies the assertion. This completes the proof of (<ref>). The linear terms in the expansions (<ref>)-(<ref>) of Christoffel symbols show that \begin{align*} 0 & = \Hess (H) + \frac{1}{4} (2 h_{(1)} h_{(2)} + 2 h_{(2)} h_{(1)} - h_{(1)}^3 - 2 \rho_0 h_{(1)}^2) \\ & - \frac{1}{2} ( 6h_{(3)} - 8 \rho_0 h_{(2)} - 4 \rho_0' h_{(1)}) \\ & - \frac{1}{4} (4 \rho_0 h_{(2)} + 4 \rho_0' h_{(1)}). \end{align*} \begin{align*} 0 & = \Hess(H) + h_{(1)} h_{(2)} - \frac{1}{4} h_{(1)}^3 + \frac{1}{2} H h_{(1)}^2 - 3 h_{(3)} - 3 H h_{(2)} + \rho_0' h_{(1)} \\ & = \Hess(H) - 3 h_{(3)} - 3 H L \lo + 2 \rho_0' L \end{align*} using the formulas for $h_{(1)}$ and $h_{(2)}$ in Proposition <ref>. This reproduces the formula (<ref>) for $h_{(3)}$ for a flat background. We round up this section with a discussion of the relation of the formula for $\B_3$ in Proposition <ref> to alternative formulas in the literature. In <cit.> and <cit.>, it is stated that for a conformally flat background $\B_3$ equals $\BB_3$, where $\BB_3$ \begin{equation}\label{GGHW-B3} 6 \BB_3 \st |\nabla \lo|^2 + 2 (\lo,\Delta (\lo)) + 3/2 (\delta(\lo),\delta(\lo)) - 2 \J |\lo|^2 + |\lo|^4 \end{equation} with $\J = \J^h$. In the flat case, using $\delta(\lo) = 2 dH$ (Codazzi-Mainardi) and the Gauss identity \J = \frac{3}{2} H^2 - \frac{1}{4} |\lo|^2, this formula reads \begin{equation}\label{GW-B3} 6 \BB_3 = |\nabla \lo|^2 + 2 (\lo, \Delta (\lo)) + 6 |dH|^2 - 3H^2 |\lo|^2 + 3/2 |\lo|^4. \end{equation} For $n=3$, it holds \begin{equation}\label{Id-1} \frac{1}{2} \Delta (|\lo|^2) = 3 (\lo,\Hess(H)) + |\nabla \lo|^2 + 3 H \tr(L^3) - |L|^4 \end{equation} \begin{equation*}\label{Id-2} \delta \delta (\lo^2) = 4 (\lo,\Hess(H)) + |\nabla \lo|^2 + 2 |dH|^2 + 3H \tr(L^3) - |L|^4. \end{equation*} Hence we have the difference formula \begin{equation}\label{basic-div} \frac{1}{2} \Delta (|\lo|^2) - \delta \delta (\lo^2) = - (\lo,\Hess(H)) - 2 |dH|^2. \end{equation} As a consequence, we obtain $\BB_3 = \B_3$. On the one hand, we apply the identity (<ref>) to calculate \begin{align*} 12 \B_3 & = 6 (\lo,\Hess(H)) + 2 |\nabla \lo|^2 + 6 H \tr(L^3) - 2 |L|^4 \\ & + 12 |dH|^2 + 6 (\lo,\Hess(H)) - 2 |\lo|^4 + 6 \tr (\lo^4) + 6 H \tr (\lo^3) \\ & = 12 (\lo,\Hess(H)) + 2 |\nabla L|^2 + 6 |dH|^2 \\ & + 6 H \tr(L^3) + 6 H \tr(\lo^3) - 2 |L|^4 - 2 |\lo|^4 + 6 \tr(\lo^4) \end{align*} using the relation $|\nabla L|^2 = |\nabla \lo|^2 + 3 |dH|^2$. Hence 6 \B_3 = 6 (\lo,\Hess(H)) + |\nabla L|^2 + 3 |dH|^2 + 3 H \tr(L^3) + 3 H \tr(\lo^3) - |L|^4 - |\lo|^4 + 3 \tr(\lo^4). On the other hand, we use the identity $\Delta (|\lo|^2) = 2 (\lo,\Delta(\lo)) + 2 |\nabla \lo|^2$ and (<ref>) to find \begin{align*} 6 \BB_3 & = \Delta (|\lo|^2) - |\nabla \lo|^2 + 6 |dH|^2- 3 H^2 |\lo|^2 + 3/2 |\lo|^4 \\ & = 6 (\lo,\Hess(H)) + |\nabla L|^2 + 3 |dH|^2 + 6 H \tr (L^3) - 3 H^2 |\lo|^2 -2 |L|^4 + 3/2 |\lo|^4. \end{align*} Hence the difference $6 (\BB_3 - \B_3)$ equals \begin{align*} & 3 H \tr(L^3) - 3 H \tr(\lo^3) - |L|^4 + |\lo|^4 + 3/2 |\lo|^4 - 3 H^2 |\lo|^2 - 3 \tr(\lo^4) \\ & = 3 H (3 H \tr (\lo^2) + 3 H^3) - 6H^2|\lo|^2 - 9H^4 - 3 H^2 |\lo|^2 + 3 (1/2 |\lo|^4 - \tr (\lo^4)) \\ & = 3 ((1/2 |\lo|^4 - \tr (\lo^4)) \\ & = 0 \end{align*} by Corollary <ref>. This proves the assertion. Finally, we show that the formula for $\B_3$ established in Proposition <ref> implies For a conformally flat background, it holds \begin{equation}\label{B3-GW-0} 6 \B_3 = 3 \delta \delta ((\lo^2)_\circ) + 3 (\Rho,(\lo^2)_\circ) + |\lo|^4. \end{equation} Note that both $\B_3$ and the right-hand side of (<ref>) are conformally invariant. In fact, the operator $T: b \mapsto \delta \delta (b) + (\Rho,b)$ acting on trace-free symmetric bilinear forms $b$ on $M^3$ is well-known to be conformally invariant in the sense that $e^{4\varphi} \hat{T} (b) = T (b)$. In (<ref>), the operator $T$ acts on the trace-free part $(\lo^2)_\circ$ of $\lo^2$. The conformal invariance $\hat{\lo}^2 = \lo^2$ implies the conformal invariance of the trace-free part of $\lo^2$ (Section <ref>). This shows the claimed conformal invariance. In particular, the right-hand side of (<ref>) has the same conformal transformation law as $\B_3$. In more explicit terms, formula (<ref>) reads \begin{align}\label{B3-GW} 6 \B_3 & = 3 \delta \delta (\lo^2) - \Delta (|\lo|^2) + 3 (\Rho,\lo^2) - |\lo|^2 \J + |\lo|^4, \end{align} and it suffices to verify (<ref>) in the flat case. For a flat background, the formulas (<ref>) and (<ref>) are equivalent. The assertion is equivalent to the vanishing of the sum \begin{align*} & 6 \delta \delta (\lo^2) - 2 \Delta (|\lo|^2) + 6 (\Rho,\lo^2) - 2 |\lo|^2 \J + 2 |\lo|^4 \\ & - \Delta (|\lo|^2) - 12 |dH|^2 - 6 (\lo,\Hess (H)) - |\lo|^4 - 6 H \tr (\lo^3). \end{align*} But the identity \begin{equation}\label{Fial} \JF = \iota^* \bar{\Rho} - \Rho + H \lo + \frac{1}{2} H^2 h \stackrel{!}{=} \lo^2 - \frac{1}{4} |\lo|^2 h \end{equation} for the Fialkov tensor $\JF$ (see (<ref>)) and the Gauss identity $\J = 3/2 H^2 - 1/4 |\lo|^2$ (see (<ref>)) imply \begin{equation}\label{JP} 6(\lo^2,\Rho) - 2 |\lo|^2 \J = 6 H \tr(\lo^3) - |\lo|^4. \end{equation} Hence the above sum simplifies to 6 \delta \delta (\lo^2) - 3 \Delta (|\lo|^2) - 12 |dH|^2 - 6 (\lo,\Hess (H)). By (<ref>), this sum vanishes. The proof is complete. Alternatively, one may derive formula (<ref>) for conformally flat backgrounds by direct evaluation of the definition of the obstruction $\B_3$. For details, we refer to <cit.>. §.§ Variational aspects Here we relate the obstructions $\B_2$ (for general backgrounds) and $\B_3$ (for conformally flat backgrounds) to singular Yamabe energy functionals. The discussion illustrates the general results of <cit.> and connects with the classical literature. We first consider the classical situation of a variation of the Willmore functional. For a closed surface $f: M^2 \hookrightarrow \R^3$, we consider a normal variation $f_t: M^2 \hookrightarrow \R^3$ of $f$: f_t (x) = f (x) + t u(x) N_0, where $u \in C^\infty(M)$ and $N_0$ is the unit normal of $M$. The variation field of $f_t$ is $u N_0$. We set $M_t = f_t (M)$ and let $\W_2$ Willmore functional \begin{equation}\label{W2-flat} \W_2(f) \st \int_{f(M)} |\lo|^2 dvol_h, \end{equation} where the metric $h$ is induced by the Euclidean metric on $\R^3$. We often identify $M$ with $f(M)$. Set \var (\W_2)[u] \st (d/dt)|_0 (\W_2(M_t)). In order to calculate that variation, we recall the well-known variation formulas \begin{align*} \var (h)[u] & = 2 u L, \\ \var (L)[u] & = -\Hess(u) + u L^2, \\ 2 \var (H)[u] & = -\Delta (u) - u |L|^2 \end{align*} \var(dvol_h)[u] = 2 u H dvol_h, where $L^2$, $|L|^2$ and $\Delta (u)$ are defined by the metric $h$ on $M$. First, we note that \var( |\lo|^2)[u] = 2 (\lo ,\var(\lo)[u]) - 4 u (L,\lo^2); the second term comes from raising $2$ indices: $|\lo|^2 = \tr (\lo^2) = h^{ia} h^{jb} \lo_{ij}\lo_{kb}$. \begin{align*} \var (|\lo|^2)[u] & = 2 (\lo,\var(L)[u] - H \var (h)[u]) - 4 u \tr(\lo^3 + H \lo^2) \\ & = - 2 ((\lo,\Hess(u) - u L^2) + 2 H u (\lo,L)) + 4 u H |\lo|^2 \end{align*} using $\tr(\lo^3)=0$. Now partial integration gives \begin{align*} \var(\W_2)[u] & = - \int_M u \left[ 2 \delta \delta (\lo) -2 (\lo,L^2) + 4 H (\lo,L) + 4 H |\lo|^2 - 2 H |\lo|^2 \right] dvol_h; \end{align*} the last term comes from the variation of the volume. Simplification yields \begin{equation}\label{W2} \var (\W_2)[u] = -\int_M u ( 2 \Delta (H) + 2 H |\lo|^2) dvol_h \end{equation} using $\delta \delta (\lo) = \Delta(H)$ and again $\tr (\lo^3)=0$. This proves the classical result that in a flat background, the Euler-Lagrange equation of the Willmore functional $\W_2$ is \Delta(H) + H |\lo|^2 = 0. By $|\lo|^2 = 2 (H^2- K)$, where $K$ is the Gauss curvature, the Euler-Lagrange equation of the Willmore functional $\W_2$ reads \Delta (H) + 2 H(H^2-K) = 0. This equation is known as the Willmore equation. It was already mentioned in <cit.> and Schadow (1922). We refer to <cit.> for more details. The variation formula (<ref>) implies the special case \begin{equation}\label{AB-2} \var(\A_2)[u] = \frac{3}{2} \int_{M^2} u \B_2 dvol_h \end{equation} of the variation formula \begin{equation}\label{var-form-1} \var(\A_n)[u] = (n+2)(n-1) \int_M u \LO_n dvol_h = \frac{(n+1)(n-1)}{2} \int_{M^n} u \B_n dvol_h \end{equation} for $\A_n = \int_M v_n dvol_h$. The first equality in (<ref>) was proved in <cit.>. The variational formula in terms of $\B_n$ was established in <cit.> by different arguments. For the second equality, we refer to (<ref>). In fact, (<ref>) shows that \A_2 = \int_{M^2} v_2 dvol = \int_{M^2} \left(-\frac{1}{2} \J + \frac{1}{4} |\lo|^2 \right) dvol_h. \var(\A_2)[u] = \frac{1}{4} \var(\W_2)[u] = -\frac{1}{2} \int_M u (\Delta (H) + H |\lo|^2) dvol_h by Gauss-Bonnet. On the other hand, we have \B_2 = -\frac{1}{3} (\Delta (H) + H|\lo|^2). This implies (<ref>). These results generalize as follows to closed surfaces $M^2 \hookrightarrow (X^3,g)$ in general backgrounds. For the following discussion, we also refer to <cit.>. We consider normal variations with a variation field of the form $u N_0$ with a unit normal field $N_0$. The following formula is well-known (see <cit.>, <cit.>, <cit.>). It can be proved by calculation in geodesic normal coordinates. For $u=1$, it plays a central role in <cit.>. It holds \begin{equation}\label{VL} \var(L_{ij})[u] = -\Hess_{ij}(u) + u (L^2)_{ij} - u \bar{R}_{0ij0}. \end{equation} \begin{align*} \var(|\lo|^2)[u] & = 2 (\lo,\var(L)[u] - H \var(h)[u]) - 4 u \tr (\lo^3 + H \lo^2) \\ & = 2 (\lo, - \Hess (u) + u L^2) - 2 u \lo^{ij} R_{0ij0} - 4 H u (\lo,L) - 4 H |\lo|^2. \end{align*} Now simplification and partial integration gives \var(\W_2)[u] = - \int_M u (2 \delta \delta (\lo) + 2 H |\lo|^2 + 2 \lo^{ij} R_{0ij0}) dvol_h. Since the Weyl tensor vanishes in dimension $3$, we have $\lo^{ij} R_{0ij0} = \lo^{ij} (\Rho_{ij} + \Rho_{00} h_{ij}) = \lo^{ij} \Rho_{ij}$. Thus the integrand is given by letting the operator \begin{equation}\label{op-2} b \mapsto \delta \delta (b) + (\iota^*(\Rho), b) + H (\lo,b) \end{equation} act on $b = \lo$. The above operator is well-known to be conformally covariant on trace-free symmetric bilinear forms (see <cit.>). This implies the conformal invariance of the integrand. The above calculation shows that the Euler-Lagrange equation of the Willmore functional $\W_2$ is \begin{equation}\label{EL} \delta \delta (\lo) + H |\lo|^2 + \lo^{ij} \bar{R}_{0ij0} = 0. \end{equation} By Codazzi-Mainardi, $\delta \delta (\lo)$ equals $\Delta (H) + \delta (\bar{\Rho}_{0})$. Thus we obtain \Delta (H) + \delta (\overline{\Ric}_{0}) + H |\lo|^2 + \lo^{ij} \bar{R}_{0ij0} = 0. We also observe that the left-hand side of (<ref>) coincides with \delta \delta (\lo) + H |\lo|^2 + (\lo,\iota^* (\bar{\Rho})) = \delta \delta (\lo) + H |\lo|^2 + (\lo,\iota^* (\overline{\Ric})) since $\lo^{ij} \bar{R}_{0ij0} = \lo^{ij} \overline{\Ric}_{ij}$. In fact, since the Weyl tensor vanishes in dimension $3$, it holds \bar{R}_{0ij0} = \bar{\Rho}_{ij} + \bar{\Rho}_{00} h_{ij} and we obtain \lo^{ij} \bar{R}_{0ij0} = \lo^{ij} \bar{\Rho}_{ij} = \lo^{ij} \overline{\Ric}_{ij}. It follows that the variation of the Willmore functional $\W_2$ yields the Yamabe obstruction $\B_2$: \var (\A_2)[u] = - \frac{1}{4} \var (\W_2) [u] = -\frac{3}{2} \int u \B_2 dvol_h confirming (<ref>) for general backgrounds. $\W_3$ Willmore functional We continue with an analogous discussion of the variation of \begin{equation}\label{W3} \W_3 \st \int_{M^3} \tr(\lo^3) dvol_h \end{equation} for variations of a closed three-manifold $M^3 \hookrightarrow X^4$ in a conformally flat background $(X,g)$. We first determine the variation of $\W_3$ for a general background metric $g$. Here we use the variation \begin{align}\label{var-form} \var (h)[u] & = 2 u L, \notag \\ \var (L)[u] & = - \Hess(u) + u L^2 - u \bar{R}_{0 \cdot \cdot 0}, \notag \\ 3 \var (H)[u] & = - \Delta (u) - u |L|^2 - u \overline{\Ric}_{00} \end{align} \var(dvol_h)[u] = 3 u H dvol_h. Note that the operator $\Delta (u) + u |L|^2 + u \overline{\Ric}_{00}$ is the Jacobi operator appearing in the second variation formula for the area of minimal surfaces <cit.>. First, we observe that \var(\tr (\lo^3))[u] = 3 (\lo^2,\var(\lo)[u]) - 6 u (L,\lo^3); the second term comes from raising $3$ indices: $\tr (\lo^3) = h^{ia} h^{jb} h^{kc} \lo_{ij}\lo_{kb} \lo_{ca}$. \begin{align*} & \var (\tr (\lo^3))[u] \\ & = 3 (\lo^2,\var(\lo)[u]) - 6 u (L,\lo^3) \\ & = 3 (\lo^2,\var (L)[u] - \var (H)[u] h - H \var(h)[u]) - 6 u (L,\lo^3) \\ & = 3 (\lo^2,-\Hess(u) + u L^2) - 3 u (\lo^2,\bar{R}_{0 \cdot\cdot 0}) + (\Delta(u) + u |L|^2 + u \overline{\Ric}_{00}) |\lo|^2 \\ & - 6 u H (\lo^2,L) - 6 u (L,\lo^3). \end{align*} Now partial integration and simplification yields \begin{align}\label{var-W3-g} & \var\big(\int_M \tr(\lo^3) dvol_h\big)[u] \notag \\ & = \int_M u \big[ - 3 \delta \delta (\lo^2) + \Delta (|\lo|^2) - 3 \tr (\lo^4) + |\lo|^4 - 3 H \tr(\lo^3) \big] dvol_h \notag \\ & + \int_M u ( - 3 (\lo^2,\bar{R}_{0 \cdot \cdot 0}) + |\lo^2| \overline{\Ric}_{00}) dvol_h. \end{align} Now, for a conformally flat background, we reformulate this variation formula in a conformally invariant way. The following result is also covered in <cit.> using a different method. Assume that $n=3$ and that the Weyl tensor of $(X,g)$ vanishes. Then \begin{equation*}\label{var-W3-CI} \var \left( \int_M \tr(\lo^3) dvol_h \right)[u] = - 3 \int_M u \left(\delta \delta ((\lo^2)_\circ) + (\Rho, (\lo^2)_\circ) + \frac{1}{3} |\lo|^4\right) dvol_h. \end{equation*} By (<ref>), the claim is equivalent to the identity \begin{align*} & \delta \delta ((\lo^2)_\circ) + (\Rho, (\lo^2)_\circ) + \frac{1}{3} |\lo|^4 \\ & = \delta \delta (\lo^2) - \frac{1}{3} \Delta (|\lo|^2) + \tr (\lo^4) - \frac{1}{3}|\lo|^4 + H \tr(\lo^3) + (\lo^2,\bar{R}_{0 \cdot \cdot 0}) - \frac{1}{3} |\lo^2| \overline{\Ric}_{00}. \end{align*} Note that \delta \delta ((\lo^2)_\circ) = \delta \delta (\lo^2) - \frac{1}{3} \Delta (|\lo|^2) \begin{align*} (\Rho, (\lo^2)_\circ) = (\Rho,\lo^2) - \frac{1}{3} (\Rho,|\lo|^2 h) = (\Ric,\lo^2) - \J |\lo|^2 - \frac{1}{3} \J |\lo|^2 = (\Ric,\lo^2) - \frac{4}{3} \J |\lo|^2. \end{align*} Moreover, it holds \begin{align*} (\lo^2,\bar{R}_{0 \cdot \cdot 0}) - \frac{1}{3} |\lo^2| \overline{\Ric}_{00} & = (\lo^2,\bar{\Rho}) + \bar{\Rho}_{00} |\lo|^2 - \frac{1}{3} |\lo^2| \overline{\Ric}_{00} \\ & = (\lo^2,\bar{\Rho}) + \frac{1}{3} \bar{\Rho}_{00} |\lo|^2 - \frac{1}{3} \bar{\J} |\lo|^2 \end{align*} (by the vanishing of the Weyl tensor). Hence the claim reduces to \begin{align*} & (\Ric,\lo^2) - \frac{4}{3} \J |\lo|^2 + \frac{1}{3} |\lo|^4 \\ & = \tr (\lo^4) - \frac{1}{3}|\lo|^4 + H \tr(\lo^3) + (\bar{\Rho},\lo^2) + \frac{1}{3} \bar{\Rho}_{00} |\lo|^2 - \frac{1}{3} \bar{\J} |\lo|^2 . \end{align*} Now the Gauss equations \begin{align*} \Ric_{ij} = \overline{\Ric}_{ij} - \bar{R}_{0ij0} + 3 H L_{ij} - (L^2)_{ij} \quad \mbox{and} \quad \J - \overline{\J} = - \bar{\Rho}_{00} - \frac{1}{4} |\lo|^2 + \frac{3}{2} H^2 \end{align*} (see (<ref>) and (<ref>)) imply \begin{align*} (\Ric,\lo^2) - \frac{4}{3} \J |\lo|^2 & = (\overline{\Ric},\lo^2) - (\bar{\Rho},\lo^2) - \bar{\Rho}_{00} |\lo|^2 + 3 H (L,\lo^2) - (L^2,\lo^2) \\ & - \frac{4}{3} \left(\bar{\J} - \bar{\Rho}_{00} - \frac{1}{4} |\lo|^2 + \frac{3}{2} H^2\right) |\lo|^2 \\ & = 2 (\bar{\Rho},\lo^2) + \bar{\J} |\lo|^2 - (\bar{\Rho},\lo^2) - \bar{\Rho}_{00} |\lo|^2 + \frac{4}{3} \bar{\Rho}_{00} |\lo|^2 - \frac{4}{3} \bar{\J} |\lo|^2 \\ & + 3 H(L,\lo^2) - (L^2,\lo^2) + \frac{1}{3} |\lo|^4 - 2 H^2 |\lo|^2 \\ & = (\bar{\Rho},\lo^2) + \frac{1}{3} \bar{\Rho}_{00} |\lo|^2 - \frac{1}{3} \bar{\J} |\lo|^2 \\ & + 3 H(L,\lo^2) - (L^2,\lo^2) + \frac{1}{3} |\lo|^4 - 2 H^2 |\lo|^2. \end{align*} Hence it suffices to prove that 3 H(L,\lo^2) - (L^2,\lo^2) + \frac{2}{3} |\lo|^4 - 2 H^2 |\lo|^2 = \tr (\lo^4) - \frac{1}{3} |\lo|^4 + H \tr (\lo^3). By simplification, this identity is equivalent to \frac{2}{3}|\lo|^4 - \tr (\lo^4) = \tr (\lo^4) - \frac{1}{3} |\lo|^4, i.e., $|\lo|^4 = 2 \tr (\lo^4)$ (see Corollary <ref>). The proof is complete. In terms of \begin{equation}\label{A3} \A_3 = \int_{M^3} v_3 dvol_h = -\frac{2}{3} \int_{M^3} \tr (\lo^3) dvol_h = -\frac{2}{3} \W_3 \end{equation} we obtain \var (\A_3)[u] = 4 \int_{M^3} u \B_3 dvol_h. Combine (<ref>) with Lemma <ref>. This is a special case of (<ref>). Note that the second equality in (<ref>) follows by combining (<ref>) with (<ref>) and (<ref>). Another proof of Corollary <ref> (even for general backgrounds) has been given in [3] using a method that rests on a certain distributional calculus. Finally, the above classical style arguments have been extended to the general case in <cit.>. §.§ Low-order extrinsic $Q$-curvatures Here we discuss the low-order extrinsic $Q$-curvatures $\QC_2$ and $\QC_3$ from the perspective of their holographic formulas. We consider $\QC_2$ in general dimensions. The holographic formula (<ref>) states that -\QC_2 = \frac{1}{n-3} (2v_2 + (n-2) (v\J)_0) + \frac{1}{n-1} \T_1^*\left(\frac{n}{2}-1\right) (v_1) for even $n$. Using $v_1 = (n-1)H$ and the formula (<ref>) for $v_2$ as well as $\T_1(\lambda)=-\lambda H$ (Lemma <ref>), we obtain - \QC_2 = \J_0 - \rho_0' + \frac{n}{2}H^2 = \J_0 - \Rho_{00} - \frac{|\lo|^2}{n-1} + \frac{n}{2} H^2. In particular, we see that the holographic formula makes sense for all $n \ge 2$. By the hypersurface Gauss identity (<ref>), the above formula simplifies to $\QC_2$ extrinsic $Q$-curvature of order $2$ \begin{equation}\label{Q2-gen} -\QC_2 = \J^h - \frac{|\lo|^2}{2(n-1)}. \end{equation} Note that this result fits with the formula $\PO_2$ second-order conformal Laplacian \begin{equation}\label{PO2} \PO_2 = \Delta_h - \left(\frac{n}{2}-1\right)\left (\J^h - \frac{|\lo|^2}{2(n-1)}\right) \end{equation} <cit.>. Independently, the latter formula for $\PO_2$ will be derived below from the solution operator $\T_2(\lambda)$ (Lemma <ref> and the discussion following it). We also recall that $\QC_2 = - Q_2$ in the Poincaré-Einstein case. Finally, we note that the formula for $\QC_2$ is singular for $n=1$ and $\Res_{n=1}(\QC_2) = \frac{1}{2} |\lo|^2 = 0$. The holographic formula (<ref>) for $\QC_3$ states that \begin{align*} \tfrac{1}{4} \QC_3 & = \tfrac{1}{n-5} (6 v_3 + (n\!-\!3) (v\J)_1) + \tfrac{1}{n-3} \T_1^*(\tfrac{n-3}{2}) (4 v_2 + (n\!-\!1) (v\J)_0) + \tfrac{1}{n-1} \T_2^*(\tfrac{n-3}{2})(2v_1) \end{align*} for even $n$. As a byproduct of the following discussion, we will see that the fractions in that formula do not prevent its validity in odd dimensions. First, we note that the above formula is equivalent to \begin{align*} \frac{1}{4} \QC_3 & = \frac{1}{n\!-\!5} (6 v_3 + (n\!-\!3) \J'_0 + (n\!-\!3) v_1 \J_0) \\ & + \frac{1}{n\!-\!3} \T_1^*\left(\frac{n-3}{2}\right) (4v_2 + (n\!-\!1) \J_0) + \frac{1}{n\!-\!1} \T_2^*\left(\frac{n-3}{2}\right) (2v_1). \end{align*} Now the formulas \begin{align*} v_1 & = (n\!-\!1) H \\ 2 v_2 & = (n\!-\!3)(n\!-\!1) H^2 - \iota^* (\J) - (n\!-\!3) \iota^* \nabla_\NV(\rho) & \mbox{(by \eqref{v2n})} \\ 6 v_3 & = (n\!-\!5)(n\!-\!3)(n\!-\!1) H^3 - (3n\!-\!7) H \iota^* (\J) - (n\!-\!5)(3n\!-\!7) H \iota^* \nabla_\NV(\rho) \\ & - (n\!-\!5) \iota^* \nabla_\NV^2(\rho) - 2 \iota^* \nabla_\NV(\J) & \mbox{(by \eqref{v3-inter})} \end{align*} show that the fractions are reduced. Indeed, it holds \begin{align*} 6 v_3 + (n\!-\!3) \J'_0 + (n\!-\!3) v_1 \J_0 & = 0 \qquad \mbox{if $n=5$}, \\ 4 v_2 + (n\!-\!1) \J_0 & = 0 \qquad \mbox{if $n=3$} \end{align*} (these are the relations mentioned in Remark <ref>). Then we obtain \begin{align}\label{Q3-ex-holo} \tfrac{1}{4} \QC_3 & = (n\!-\!3)(n\!-\!1) H^3 - (3n\!-\!7) H \iota^* \nabla_\NV(\rho) + \iota^* \nabla_\NV(\J) - \iota^* \nabla_\NV^2(\rho) + (n\!-\!2) H \iota^* (\J) \notag \\ & + \T_1^*\left(\frac{n\!-\!3}{2}\right) (\iota^* (\J) - 2 \iota^* \nabla_\NV(\rho) + 2(n\!-\!1)H^2) + \T_2^*\left(\frac{n\!-\!3}{2}\right) (2H). \end{align} A calculation using Lemma <ref>, Lemmas <ref>–<ref> and the hypersurface Gauss identity (<ref>) shows that \begin{equation}\label{Q3-rho} \frac{1}{4} \QC_3 = \Delta H + H \iota^* \J + \iota^* \nabla_\NV (\J) - (n\!-\!1) H \iota^* \nabla_\NV(\rho) - \iota^* \nabla_\NV^2(\rho). \end{equation} In particular, for $n=3$, we find \begin{align*} \frac{1}{4} \QC_3 = \Delta H + H\iota^* (\J) + \iota^* \nabla_\NV(\J) - 2 H \iota^* \nabla_\NV(\rho) -\iota^* \nabla_\NV^2(\rho). \end{align*} By comparison with Example <ref>, we see that 12 v_3 = - \QC_3 + 4\Delta H and consequently \begin{equation}\label{vQ3} 12 \int_{M^3} v_3 dvol_h = - \int_{M^3} \QC_3 dvol_h. \end{equation} This confirms Theorem <ref> for $n=3$. Formula (<ref>) also follows from the last display in the proof of <cit.> or <cit.> which evaluates the composition of three Laplace-Robin operators. We continue with the discussion of the holographic formula for $\QC_3$ in general dimensions $n \ge 3$. In fact, we prove that an evaluation of (<ref>) shows that the explicit formula (<ref>) for $\rho_0''$ is equivalent to a simple formula for $\QC_3$. Assume that $n \ge 3$. Then the formula (<ref>) for $\rho_0''$ is equivalent to $\QC_3$ extrinsic $Q$-curvature of order $3$ \begin{align}\label{Q3F} \frac{1}{4} \QC_3 & = \frac{1}{n\!-\!2} (\delta \delta (\lo) - (n\!-\!3) (\lo, \Rho^h) + (n\!-\!1) (\lo,\JF)) \\ & = \frac{1}{n\!-\!2} (\delta \delta (\lo) - 2(n\!-\!2) (\lo,\Rho^h) + (n\!-\!1) (\lo,\Rho^g) + (n\!-\!1) H |\lo|^2). \notag \end{align} Note that (<ref>) fits with the explicit formula for $\PO_3$ in Proposition <ref>. The proof of Proposition <ref> will show that the displayed formula for $\QC_3$ follows by combining the holographic formula for $\QC_3$ (in the form (<ref>)) with the formulas for the first two normal derivatives of $\rho$. Note also that in dimension $n=3$, the above formula reads $\QC_3$ critical extrinsic $Q$-curvature of order $3$ \begin{equation}\label{Q3-ex} \QC_3 = 4 \delta \delta (\lo) + 8 (\lo,\JF). \end{equation} It immediately follows from that expression that the integral $\int_{M^3} \QC_3 dvol_h$ is conformally invariant as a functional of $g$. It suffices to prove the equivalence of (<ref>) and the first identity. For this, we make explicit the equality of (<ref>) and (<ref>). In terms of adapted coordinates, it states the equality \begin{equation*} \Delta H + H \J_0 + \J_0' - (n\!-\!1) H \rho_0' - \rho_0'' - 2 H \rho_0' = \frac{1}{n\!-\!2} \delta \delta (\lo) - \frac{n\!-\!3}{n\!-\!2} \lo^{ij} \Rho^h_{ij} + \frac{n\!-\!1}{n\!-\!2} \lo^{ij} \JF_{ij}. \end{equation*} Here we used that $\iota^* \nabla_\NV^2(\rho)$ corresponds to $\rho_0'' - 2 \rho_0 \rho_0' = \rho_0'' + 2 H \rho_0'$ (see Example <ref>). By the identity (<ref>) for $\Delta H$, this relation is equivalent to \begin{equation*} \rho_0'' = \frac{1}{n\!-\!1} \delta \delta (\lo) -\frac{1}{n\!-\!2} \delta \delta (\lo) + \frac{n\!-\!3}{n\!-\!2} \lo^{ij} \Rho^h_{ij} - \frac{n\!-\!1}{n\!-\!2} \lo^{ij} \JF_{ij} + H \J_0 - (n+1) H \rho_0' + \J_0'. \end{equation*} H \J_0 - (n\!+\!1) H \rho_0' = - \frac{n\!+\!1}{n\!-\!1} H \Ric_{00} + \frac{1}{n\!-\!1} H \scal - \frac{n\!+\!1}{n\!-\!1} H |\lo|^2 shows that this expression for $\rho_0''$ coincides with the one given in (<ref>). The proof is complete. The above formula for $\QC_3$ is singular for $n=2$. But the second formula in (<ref>) shows that \begin{equation}\label{QB2} \Res_{n=2} (\QC_3) = 4 (\delta \delta (\lo) + (\lo, \Rho^g) + H |\lo|^2), \end{equation} up to the term $-8(n-2) (\lo,\Rho^h) = -8(\lo,\Ric^h)$ (which vanishes in dimension $n=2$ by $\Ric^h = K h$, $K$ being the Gauss curvature). The right-hand side of (<ref>) is proportional to the obstruction $\B_2$. From that perspective, its conformal invariance follows from the conformal covariance of $\PO_3$. The above argument to derive $\B_2$ from the constant term of $\PO_3$ is due to <cit.>. For the general relation between the singular Yamabe obstruction and the super-critical $Q$-curvature $Q_{n+1}$, we refer to Theorem <ref>. §.§ The pair $(\PO_3,\QC_3)$ The following explicit formula for $\PO_3$ was first proven in <cit.> by evaluation of the relevant composition of three Laplace-Robin operators.[See also the identical <cit.>.] $\PO_3$ extrinsic conformal Laplacian of order $3$ For $n \ge 3$, the operator \PO_3 = 8 \delta (\lo d) + \frac{n\!-\!3}{2} \frac{4}{n\!-\!2} (\delta \delta (\lo) - (n\!-\!3) (\lo,\Rho^h) + (n\!-\!1) (\lo,\JF)) is conformally covariant: e^{\frac{n+3}{2} \varphi} \circ \PO_3(\hat{g}) = \PO_3(g) \circ e^{\frac{n-3}{2}\varphi}, \; \varphi \in C^\infty(X). The formula for the leading term of $\PO_3$ will be derived in Lemma <ref>. We first calculate \begin{align*} & e^{\frac{n+3}{2} \varphi} \hat{\delta} (\hat{\lo} d)(e^{-\frac{n-3}{2} \varphi} u) = \delta (e^{\frac{n-3}{2}\varphi} \lo d)(e^{-\frac{n-3}{2}\varphi} u) + \frac{n\!-\!3}{2} e^{\frac{n-3}{2}\varphi} i_{\grad(\varphi)}(\lo d)(e^{-\frac{n-3}{2} \varphi} u) \\ & = \delta (\lo d)(u) - \frac{n\!-\!3}{2} \delta (\lo d\varphi \cdot u) + \frac{n\!-\!3}{2} i_{\grad(\varphi)}(\lo du) - \left(\frac{n\!-\!3}{2}\right)^2 (d\varphi,\lo d\varphi) u \\ & = \delta (\lo d)(u) - \frac{n\!-\!3}{2} \delta (\lo d\varphi) u - \left(\frac{n\!-\!3}{2}\right)^2 (d\varphi,\lo d\varphi) u \end{align*} using $\hat{\lo} = e^{-\varphi} \lo$ for the endomorphism $\lo$ corresponding to the trace-free second fundamental form and the transformation law $e^{(a+2)\varphi} \hat{\delta} e^{-a\varphi} = \delta + (n\!-\!2\!-\!a) i_{\grad(\varphi)}$ on $\Omega^1(M)$. Next, the term with the Schouten tensor yields (\lo,\Rho^h) - (\lo,\Hess(\varphi)) + (\lo,d\varphi \otimes \varphi). Finally, it holds \begin{align*} & e^{3\varphi} \hat{\delta} \hat{\delta} (\hat{\lo}) = \delta(e^\varphi \hat{\delta} (e^\varphi \lo)) + (n\!-\!3) i_{\grad(\varphi)} e^{\varphi} \hat{\delta} (e^\varphi \lo) \\ & = \delta \delta (\lo) + (n\!-\!1) \delta (\lo d\varphi) + (n\!-\!3) i_{\grad(\varphi)} \delta (\lo) + (n\!-\!3)(n\!-\!1) i_{\grad(\varphi)} i_{\grad(\varphi)} (\lo) \end{align*} using $\hat{\lo} = e^{\varphi} \lo$ and the transformation law e^{(a+2)\varphi} \hat{\delta} (e^{-a\varphi} b) = \delta (b) + (n\!-\!2\!-\!a) i_{\grad(\varphi)}(b) - \tr(b) d\varphi on symmetric bilinear forms $b$. Now simplification proves the claim. The operator $\frac{n-2}{4} \PO_3$ differs from the tractor calculus operator D_M^A L_{AB} D_X^B by the contribution of $(\lo,\JF)$ <cit.>. Here $L_{AB}$ is a tractor calculus version of the second fundamental form. This identification again implies its conformal covariance. Proposition <ref> yields an explicit formula for $\QC_3$ (see (<ref>)). As a direct cross-check of that formula, we note that in dimension $n=3$ \begin{align*} e^{3\varphi} \hat{\QC}_3 & = 4 e^{3\varphi} \hat{\delta} \hat{\delta} (\hat{\lo}) + 8 e^{3\varphi} (\hat{\lo}, \hat{\JF})_{\hat{h}} \\ & = 4 \delta e^\varphi \hat{\delta} (e^{\varphi} \lo) +8 (\lo,\JF)_h \\ & = 4 (\delta \delta (\lo) + 2 \delta (\lo d)(\varphi)) + 8 (\lo,\JF)_h \\ & = \QC_3 + \PO_3 (\varphi). \end{align*} This also confirms Theorem <ref> for $n=3$. Similar arguments provide an elementary proof of the conformal invariance of the obstruction $\B_2$. §.§ Low-order solution operators Under the assumption $\SCY$, we make the low-order solution operators $\T_1(\lambda)$ and $\T_2(\lambda)$ explicit in general dimensions. For that purpose, we use the ansatz $u = s^\lambda f + s^{\lambda+1} \T_1(\lambda) f + s^{\lambda+2} \T_2(\lambda) f + \dots$ for an approximate solution of the equation $-\Delta_{\sigma^{-2}g}(u)=\lambda(n-\lambda)u$ in adapted coordinates. $\T_1(\lambda) = - \lambda H$. By the assumption $\SCY$, the coefficients in (<ref>) expand as $a = 1-2s\rho + O(s^{n+1}) = 1 + 2s H + \cdots$ (using $\iota^* (\rho) = -H$) and $\tr (h_s^{-1} h'_s) = 2 \tr(L) + \cdots$. Now we let the operator (<ref>) act on $u$. The expansion of the result starts with s^2 \partial_s^2 (s^\lambda) f - (n-1) s\partial_s (s^\lambda) f \stackrel{!}{=} -\lambda(n-\lambda) s^\lambda f. The next term in the expansion reads \begin{align*} & s^2 \partial_s^2(s^{\lambda+1}) \T_1(\lambda)(f)+ 2 s^3 H \partial_s^2(s^\lambda) f + s^2 \tr(L) \partial_s(s^\lambda) f \\ & - (n-1) s \partial_s(s^{\lambda+1}) \T_1(\lambda)(f) - 2(n-1) s^2 H \partial_s(s^\lambda) f + H s^2 \partial_s(s^\lambda) f \\ & = (\lambda\!-\!n\!+\!1)(\lambda\!+\!1) s^{\lambda+1} \T_1(\lambda)(f) + (2\lambda\!-\!n\!+\!1) \lambda H s^{\lambda+1} f. \end{align*} This result equals $-\lambda(n-\lambda) s^{\lambda+1} \T_1(\lambda)(f)$ iff $\T_1(\lambda)(f)=-\lambda H$. It is worth emphasizing that $\T_1(\lambda)$ is regular in $\lambda$. In fact, this property does not hold for general asymptotically hyperbolic metrics <cit.>. As a consequence of Lemma <ref>, we find \begin{align}\label{D1-exp} \D_1^{res}(\lambda) & = (2\lambda\!+\!n\!-\!1)(\iota^* \partial_s + (v_1 + \T_1^*(\lambda\!+\!n\!-\!1)) \iota^*) \notag \\ & = (2\lambda\!+\!n\!-\!1)(\iota^* \partial_s -\lambda H \iota^*). \end{align} This formula obviously confirms Theorem <ref> for the operator $\iota^* L(\lambda)$. The formula for $\T_2(\lambda)$ is a bit more complicated. Assume that $n \ge 2$. Then \begin{align*} \T_2(\lambda) & = \frac{-\Delta_h + \lambda \J^h}{2(2\lambda\!-\!n\!+\!2)} + \frac{\lambda}{2} \Rho_{00} - \frac{\lambda}{2(2\lambda\!-\!n\!+\!2)} \left(\frac{n\!-\!1\!-\!2\lambda}{n\!-\!1} - \frac{1}{2(n\!-\!1)} \right) |\lo|^2 \\ & + \frac{\lambda}{2(2\lambda\!-\!n\!+\!2)} \left((\lambda\!+\!1)(2\lambda\!-\!n\!+\!3) - \frac{n}{2}\right) H^2. \end{align*} By $\SCY$, the coefficients in (<ref>) expand as a = 1-2s\rho + \cdots = 1 + 2s H - 2s^2 \left(\Rho_{00} + \frac{|\lo|^2}{n-1}\right) + \cdots, up to a remainder in $O(s^{n+1})$, and \tr (h_s^{-1} h'_s) = 2 \tr(L) - s (2\Ric_{00} + 2 |\lo|^2 + 4n H^2) + \cdots. Hence the equality of the coefficients of $s^{\lambda+2}$ in the expansion of the eigenequation $\Delta_{s^{-2}\eta^*( g)} u = -\lambda(n-\lambda) u$ yields the \begin{align*} & 2(2\lambda\!-\!n\!+\!2) \T_2(\lambda)(f) + (\lambda\!+\!1)(2\lambda\!-\!n\!+\!3) H \T_1(\lambda)(f) \\ & = \left[-2\lambda(\lambda\!-\!1) \rho_0' - \lambda (\Ric_{00} + |\lo|^2) + 2(n\!-\!1) \lambda \rho_0' -2\lambda \rho_0' + \Delta \right](f) = 0. \end{align*} This condition for $\T_2(\lambda)$ simplifies to \begin{align*} & 2(2\lambda\!-\!n\!+\!2) \T_2(\lambda)(f) - \lambda(\lambda\!+\!1)(2\lambda\!-\!n\!+\!3) H^2 f \\ & = \lambda \left[ -2(n\!-\!1\!-\!\lambda) \rho_0' + \Ric_{00} + |\lo|^2 \right] f - \Delta f \\ & = \lambda \left [-2(n\!-\!1\!-\!\lambda)\left(\Rho_{00} + \frac{|\lo|^2}{n-1}\right) + ((n\!-\!1) \Rho_{00} + \J_0) + |\lo|^2 \right] f - \Delta f \\ & = -\lambda \left((n\!-\!1\!-\!2\lambda) \Rho_{00} + \frac{n\!-\!1\!-\!2\lambda}{n\!-\!1} |\lo|^2 - \J_0 \right) f - \Delta f \end{align*} using Lemma <ref>. Finally, (<ref>) implies \begin{align*} & 2(2\lambda\!-\!n\!+\!2) \T_2(\lambda)(f) \\ & = - \lambda \left((n\!-\!2\!-\!2\lambda) \Rho_{00} - \J^h + \left(\frac{n\!-\!1\!-\!2\lambda}{n\!-\!1} - \frac{1}{2(n\!-\!1)} \right) |\lo|^2 + \frac{n}{2} H^2 \right) f \\ & + \lambda(\lambda\!+\!1)(2\lambda\!-\!n\!+\!3) H^2 f - \Delta f. \end{align*} This completes the proof. In particular, Lemma <ref> implies -4 \Res_{\frac{n}{2}-1}(\T_2(\lambda)) = \Delta_h - \left( \frac{n}{2}-1\right) \left(\J^h - \frac{ |\lo|^2}{2(n-1)} \right) = \PO_2(g). This is a special case of Theorem <ref>. It follows that -\QC_2(g) = \J^h - \frac{1}{2(n-1)} |\lo|^2. For $n=2$, Lemma <ref> implies \T_2(0)(1) = \frac{1}{4} \J^h - \frac{1}{8} |\lo|^2. For the function $\QC_2(\lambda)$ defined in (<ref>), it follows that $\QC_2(0)= - \J^h + \frac{1}{2} |\lo|^2 \stackrel{!}{=} \QC_2$. This confirms (<ref>) for $n=2$. The above results yield an explicit formula for $\D_2^{res}(\lambda)$. \begin{align}\label{D2-final} \D_2^{res}(\lambda) & = (2\lambda\!+\!n\!-\!3 ) \Big[(2\lambda\!+\!n\!-\!2) \iota^* \partial_s^2 - 2 (2\lambda\!+\!n\!-\!2)(\lambda\!-\!1) H \iota^* \partial_s \notag \\ & - \Big[ \Delta_h + \lambda \J^h - \lambda(2\lambda\!+\!n\!-\!2) \Rho_{00} - \lambda \left (\frac{2\lambda\!+\!n\!-\!2}{2(n\!-\!1)} + \frac{2\lambda\!+\!n\!-\!1}{2(n\!-\!1)} \right) |\lo|^2 \notag \\ & -(2\lambda\!+\!n\!-\!2)(\lambda\!-\!1/2) \lambda H^2\Big] \iota^* \Big]. \end{align} We omit the details of the calculation. This result fits with the formula for $\iota^* L(\lambda-1) L(\lambda)$ in <cit.>, \D_2^{res}(\lambda) \circ \eta^* = \iota^* L(\lambda-1) L(\lambda). In order to see this, it only remains to express the normal derivatives in the variable $s$ by iterated gradients. But $\iota^* \nabla_\NV$ corresponds to $\iota^*\partial_s$ and $\iota^* \nabla_\NV^2$ corresponds to $\iota^* (\partial_s^2 + 2 H \partial_s)$ (see Example <ref>). Note that the prefactor in (<ref>) implies that $\D_2^{res}(-\frac{n-3}{2}) = 0$. The conformally covariant term in brackets (for $\lambda=-(n-3)/2$ and up to the contribution by $|\lo|^2$) has been used in <cit.> as a boundary operator associated to the Paneitz operator on $X$. For general $\lambda$, it appears in <cit.> (up to the term containing $|\lo|^2$). Concerning the classification of such boundary operators, we refer to <cit.>. Lemma <ref> immediately shows that $\D_2^{res}(-\frac{n}{2}+1)= \PO_2(g)\iota^*$. In addition, we find the remarkable identity \begin{equation}\label{factor-big} \D_2^{res}\left(\frac{-n+1}{2}\right) \circ \eta^* = 2 \iota^* P_2(g), \end{equation} where $P_2(g)$ is the Yamabe operator of $g$. In fact, we calculate \begin{align*} & \D_2^{res}\left(\frac{-n+1}{2}\right) \\ & = 2 \left(\iota^* \partial_s^2 + (n\!+\!1) H \iota^* \partial_s + \Delta_h \iota^* - \frac{n\!-\!1}{2} \left[\J^h + \Rho_{00} - \frac{n}{2} H^2 + \frac{1}{2(n\!-\!1)} |\lo|^2 \right] \iota^* \right) \\ & = 2 \left(\iota^* \partial_s^2 + (n\!+\!1) H \iota^* \partial_s + \Delta_h \iota^* - \frac{n\!-\!1}{2} \iota^* \J^g \right) \end{align*} using (<ref>). Hence \begin{align*} \D_2^{res}\left(\frac{-n\!+\!1}{2}\right) \circ \eta^* & = 2 \left(\iota^* \nabla_\NV^2 + (n\!-\!1) H \iota^* \nabla_\NV + \Delta_h \iota^* - \frac{n\!-\!1}{2} \iota^* \J^g \right) \\ & = 2 \iota^*\left(\Delta_g - \frac{n\!-\!1}{2}\J^g \right) = 2 \iota^* P_2(g) \end{align*} using $\iota^* \nabla_\NV^2 = (\iota^* \partial_s^2 + 2 H \iota^* \partial_s) \circ \eta^*$ (Example <ref>) and the identity (<ref>). For Poincaré-Einstein metrics, the relation (<ref>) is one of the identities in a second set of so-called factorization identities <cit.>. It is an open problem whether the higher-order residue families in the general case continue to satisfy such identities. Finally, we note that Lemma <ref> implies that \begin{equation}\label{Q2-van} \QC^{res}_2(0) \st \D_2^{res}(0)(1) =0. \end{equation} This is a special case of the following conjecture. $\QC_N^{res}(0) \st \D_N^{res}(0)(1) \stackrel{!}{=} 0$ for $N \ge 1$. This vanishing result is well-known for residue families of even order in the Poincaré-Einstein case A conformally covariant second-order family of differential operators which interpolates between the GJMS operators $\iota^* P_2(g)$ and $P_2(h) \iota^*$ (with $h = \iota^* (g))$ was given in <cit.>. This result suggests restating the above formula for $\D_2^{res}(\lambda)$ in the perhaps more enlightening \begin{align}\label{L2-nice} \iota^* L(\lambda-1) L(\lambda) & = (2\lambda\!+\!n\!-\!3) \Big[ (2\lambda\!+\!n\!-\!2) \iota^* P_2(g) - (2\lambda\!+\!n\!-\!1) \PO_2(g) \iota^* \notag \\ & -(2\lambda\!+\!n\!-\!1)(2\lambda\!+\!n\!-\!2) H \iota^* \nabla_\NV \notag \\ & + 2 \left(\lambda\!+\!\frac{n\!-\!1}{2}\right) \left(\lambda\!+\!\frac{n\!-\!2}{2}\right) (\QC_2(g) + \iota^* Q_2(g) + \lambda H^2) \iota^* \Big], \end{align} P_2(g) = \Delta_g - \frac{n\!-\!1}{2} \J^g = \Delta_g - \frac{n\!-\!1}{2} Q_2(g) is the Yamabe operator of $g$ and \PO_2(g) = \Delta_h - \left(\frac{n}{2}-1\right)\left (\J^h - \frac{|\lo|^2}{2(n-1)}\right) = \Delta_h + \left(\frac{n}{2}-1\right) \QC_2(g) is the extrinsic Yamabe operator on $M$ (see (<ref>)).[Note that we use the conventions $\PO_2(g)(1) = \frac{n-2}{2} \QC_2(g)$ on $M$ and $P_2(g)(1) = -\frac{n-1}{2} Q_2(g)$ on $X$.] In order to prove that formula, it is enough to relate the cubic polynomial in square brackets to the corresponding cubic polynomial in (<ref>). It is easy to relate the terms with derivatives. Moreover, the coincidence of the zeroth order terms follows from the Gauss identity - we omit the calculation. (<ref>) immediately shows that $\D_2^{res}(g;-\frac{n}{2}+1) = \PO_2(g) \iota^*$ and \begin{equation*} \D_2^{res}\left(g;\frac{-n+1}{2}\right) \circ \eta^* = 2 \iota^* P_2(g). \end{equation*} The formula (<ref>) leads to a simple formula for the $Q$-curvature polynomial \QC_2^{res}(g;\lambda) \st \D_2^{res}(g;\lambda)(1). It holds \begin{align}\label{Q2-simple} & \QC_2^{res}(g;\lambda) = (2\lambda\!+\!n\!-\!3) \lambda \notag \\ & \times \left[ (2\lambda\!+\!n\!-\!2) \iota^* Q_2(g) + (2\lambda\!+\!n\!-\!1) \QC_2(g) + 2\left(\lambda\!+\!\frac{n\!-\!1}{2}\right) \left(\lambda\!+\!\frac{n\!-\!2}{2}\right) H^2 \right]. \end{align} This result clearly implies (<ref>). It suggests considering the quadratic polynomial in brackets as the actual interesting object. In the critical dimension $n=2$, we also find $\dot{\D}_2^{res}(0)(1) = \dot{\QC}_2^{res}(0) = -\QC_2$. One should compare Lemma <ref> with <cit.>. Note that in the Poincaré-Einstein case (i.e., if $g = r^2 g_+$), it holds $\iota^* Q_2(g) = \iota^* \J^g = \J^h = Q_2(h)$, $\QC_2(g) = - Q_2(h)$ (Remark <ref>) and $H=0$. Hence (<ref>) reduces to (2\lambda\!+\!n\!-\!3) \lambda \left[(2\lambda\!+\!n\!-\!2) Q_2(h) - (2\lambda\!+\!n\!-\!1) \QC_2(h) \right] = (2\lambda\!+\!n\!-\!3) \lambda Q_2(h) (see also Remark <ref>). It is an open problem whether for $N \ge 3$ the $Q$-curvature polynomial $\QC_N^{res}(\lambda) \st \D_N^{res}(\lambda)(1)$ similarly can be reduced to a lower degree polynomial and whether these polynomials admit a recursive description as in the Poincaré-Einstein case Next, we determine the leading part of $\PO_3$ from the leading part of the solution operator $\T_3(\lambda)$. $\PO_3$ is an operator of second-order. Let $\LT(\PO_3)$ denote its terms that contain one or two derivatives. The same notation will be used for other second-order operators. It holds \LT( \Res_{\lambda=\frac{n-3}{2}}(\T_3(\lambda)) = \frac{1}{3} \delta (\lo d). \LT(\PO_3) = 8 \delta (\lo d). The leading part of the solution operator $\T_3(\lambda)$ is determined by the equation \begin{align*} & 3(2\lambda\!-\!n\!+\!3) \LT(\T_3(\lambda)) + 2 (\lambda\!+\!1)(\lambda\!+\!2) H \LT(\T_2(\lambda)) + (\lambda+2) \tr(L) \LT(\T_2(\lambda)) \\ & + (\lambda+2) H \LT(\T_2(\lambda)) - 2 (n\!-\!1)(\lambda+2) H \LT(\T_2(\lambda)) - (dH,d\cdot)_h \\ & + \LT(\Delta \T_1(\lambda)) + \Delta_h' = 0, \end{align*} where $\Delta_{h_s} = \Delta_h + s \Delta_h' + \frac{s^2}{2!} \Delta_h'' + \cdots$. Simplification yields \begin{align*} & -3(2\lambda\!-\!n\!+\!3) \LT(\T_3(\lambda)) \\ & = (\lambda\!+\!2)(2\lambda\!-\!n\!+\!5) H \LT(\T_2(\lambda)) - (dH,d\cdot)_h + \LT(\Delta \T_1(\lambda)) + \Delta'. \end{align*} But the first variation $\Delta'$ of the Laplacian with respect to the variation $h_s$ of $h$ is given by (<ref>). Using $\LT( \T_2(\frac{n-3}{2}))= \frac{1}{2} \Delta$ (Lemma <ref>), it follows that $$-6 \LT( \Res_{\lambda=\frac{n-3}{2}}(\T_3(\lambda))$$ \begin{align*} & \frac{n\!+\!1}{2} H \Delta - (dH,d\cdot)_h - \frac{n\!-\!3}{2} \LT(\Delta \circ H) -2H \Delta + (n\!-\!2) (dH,d\cdot)_h - 2 \delta (\lo d) \\ & = -(dH,d\cdot)_h - (n\!-\!3) (dH,d\cdot)_h + (n\!-\!2) (dH,d\cdot)_h - 2 \delta (\lo d) \\ & = - 2 \delta (\lo d). \end{align*} This completes the proof. Similar arguments prove It holds \LT( \Res_{\lambda=\frac{n-5}{2}}(\T_5(\lambda)) = \frac{1}{30} (\Delta \delta (\lo d) + \delta(\lo ) \Delta). \LT(\PO_5) = 192 (\Delta \delta (\lo d) + \delta(\lo d) \Delta). These results are special cases of Proposition <ref>. §.§ Low order cases of Theorem <ref> The following result illustrates Theorem <ref> for $k=1,2$. Assume that $\sigma$ satisfies $\SCY$. Then \begin{equation*} \iota^* L(-n\!+\!1) \left( \left( \frac{r}{\sigma} \right)^{n-1} \right) = - \frac{(n\!-\!1)^2}{2} H \end{equation*} for $n \ge 1$ and \begin{align*} & \iota^* L(-n\!+\!1) L(-n\!+\!2) \left( \left( \frac{r}{\sigma} \right)^{n-2} \right) \\ & = \frac{(n\!-\!1)(n\!-\!2)}{6} \left( -\frac{n\!-\!5}{n\!-\!1} |\lo|^2 - 2 (n\!-\!2) \iota^* \J^g + 2 (n\!-\!5) \J^h + \frac{(n\!-\!2)(n\!-\!3)}{2} H^2 \right) \end{align*} for $n \ge 2$. Here $L(\lambda)$ is short for $L(g,\sigma;\lambda)$. By <cit.>, it holds $w_1 = \frac{n-1}{2} H$ and the right-hand side of the second identity equals $2(n\!-\!1)(n\!-\!2) w_2$. Hence w_1 = - \frac{1}{n-1} \iota^* L(-n\!+\!1)\left( \left( \frac{r}{\sigma} \right)^{n-1} \right) for $n \ge 2$ and = \frac{1}{2(n\!-\!1)(n\!-\!2)} \iota^* L(-n\!+\!1) L(-n\!+\!2) \left( \left(\frac{r}{\sigma} \right)^{n-2} \right) for $n \ge 3$. We expand in geodesic normal coordinates. First, we note that \left( \frac{r}{\sigma} \right)^{n-1} = 1 - (n\!-\!1) \sigma_{(2)} r + \cdots and recall that $\sigma_{(2)} = \frac{1}{2} H$. Now we calculate \begin{align*} \iota^* L(-n\!+\!1) (1) = -(n\!-\!1)^2 H \quad \mbox{and} \quad \iota^* L(-n\!+\!1) (H r) = - (n\!-\!1) H \end{align*} using $\rho_0 = -H$. This proves the first identity. Similarly, we find \left( \frac{r}{\sigma} \right)^{n-2} = 1 - (n\!-\!2) \sigma_{(2)} r + \left(-(n\!-\!2) \sigma_{(3)} + \binom{n\!-\!1}{2} \sigma_{(2)}^2\right) r^2 + \cdots. Now $\grad_g(\sigma) = \partial_r + H r \partial_r + \cdots$ and we obtain \begin{align*} & \iota^* L(-n\!+\!1)L(-n\!+\!2)(1) \\ & = -(n\!-\!1) \iota^* (\partial_r + (n\!-\!1) H) ((-n\!+\!3) (\partial_r + H r \partial_r - (n\!-\!2) \rho) - r (\Delta_g - (n\!-\!2) \J^g))(1) \\ & = -(n\!-\!1)(n\!-\!2) ((n\!-\!3) \partial_r (\rho)|_0 + \J^g|_0 - (n\!-\!1)(n\!-\!3) H^2). \end{align*} Note that this result coincides with $2(n\!-\!1)(n\!-\!2)v_2$, where $v_2$ is as in (<ref>). Moreover, we get \begin{align*} \iota^* L(-n\!+\!1)L(-n\!+\!2)(\sigma_{(3)} r^2) & = (n\!-\!1)(n\!-\!2) (-2(n-3) \sigma_{(3)} - 2 \sigma_{(3)}) \\ & = 2(n\!-\!1)(n\!-\!2) \sigma_{(3)} \end{align*} \begin{equation*} \iota^* L(-n\!+\!1)L(-n\!+\!2)(\sigma_{(2)}^2 r^2) = 2(n\!-\!1)(n\!-\!2) \sigma_{(2)}^2. \end{equation*} Finally, we obtain \iota^* L(-n\!+\!1)L(-n\!+\!2)(\sigma_{(2)} r) = \frac{1}{2}(n\!-\!1)(n\!-\!2)(2n\!-\!3) H^2 \Delta_g (\sigma_{(2)} r) = \frac{1}{2} \tr (h_r^{-1} h_r') \sigma_{(2)} + r \Delta_h (\sigma_{(2)}) = n H \sigma_{(2)} + O(r). We omit the details. Note that only in the latter calculation the contribution $H r\partial_r$ in $\grad_g(\sigma)$ plays a role. Now, using Lemma <ref>, Lemma <ref> and Lemma <ref>, these results imply the second assertion. E. Abbena, A. Gray and L. Vanhecke, Steiner's formula for the volume of a parallel hypersurface in a Riemannian manifold, Annali Sc. Norm. Sup. Pisa 8 (3) (1981) 473-493. P. Albin, Renormalizing curvature integrals on Poincaré-Einstein manifolds, Adv. in Math. 221 (2009 (1) 140–169. <arXiv:math/0504161> S. Alexakis, The Decomposition of Global Conformal Invariants, Annals of Mathematics Studies, 182, Princeton University Press, Princeton, NJ, 2012, x+449. L. Andersson, P. Chruściel and H. Friedrich, On the regularity of solutions to the Yamabe equation and the existence of smooth hyperboloidal initial data for Einstein's field equations, Commun. Math. Phys. 149 (1992) 587–612. C. Arias, A. R. Gover and A. Waldron, Conformal geometry of embedded manifolds with boundary from universal holographic formulæ, Advances in Math. 384 (2021). P. Aviles and R. C. McOwen, Complete conformal metrics with negative scalar curvature in compact Riemannian manifolds, Duke Math. J. 56 (1988) 395–398. H. Baum and A. Juhl, Conformal Differential Geometry: $Q$-Curvature and Conformal Holonomy. Oberwolfach Seminars 40, 2010. I. N. Bernshtein, Modules over a ring of differential operators. Study of the fundamental solutions of equations with constant coefficients, Functional Analysis and Its Applications 5 (1971) 89–101. A. Besse, Einstein Manifolds, Ergebnisse der Mathematik und ihrer Grenzgebiete 10, Springer-Verlag, (1987). S. Blitz, R. Gover and A. Waldron, Generalized Willmore energies, $Q$-curvatures, extrinsic Paneitz operators, and extrinsic Laplacian powers. <arXiv:2111.00179v2> T. P. Branson, Sharp inequalities, the functional determinant, and the complementary series, Trans. Amer. Math. Soc. 347 (1995) 3671–3742. J. L. Brylinski, The beta function of a knot, Int. J. Math. 10 (4) (1999) 415–423. A. Čap and J. Slovák, Parabolic Geometries. I. Background and General Theory, Mathematical Surveys and Monographs, 154, American Mathematical Society, Providence, RI, (2009), x+628. J. S. Case, Boundary operators associated with the Paneitz operator, Indiana Univ. Math. J. 67 (1) (2018) 293–327. <arXic:1509.08342> S.-Y. Alice Chang and J. Qing, The zeta functional determinants on manifolds with boundary. I. The formula, J. Funct. Anal. 147 (2) (1997) 327–362. S.-Y. Alice Chang, Conformal invariants and partial differential equations, Bull. Amer. Math. Soc. (N.S.), 42 (3) (2005) 365–393. S.-Y. Alice Chang, Conformal Geometry on Four Manifolds, Proc. Int. Cong. of Math. (2018) 1 119–146. <arXiv:1809.06339> S.-Y. Alice Chang, S. McKeown and P. Yang, Scattering on singular Yamabe spaces. <arXiv:2109.02014>. J.-L. Clerc, Another approach to Juhl's conformally covariant differential operators from $S^n$ to $S^{n-1}$, SIGMA 13 (2017) 11. <arXiv:1612.01856> T. Colding and W. Minicozzi II, A Course in Minimal Surfaces, Graduate Studies in Mathematics 121, AMS, 2011. S. Deser and A. Schwimmer, Geometric classification of conformal anomalies in arbitrary dimensions, Phys. Lett. B 309 (3-4) (1993) 279–284. [1] Z. Djadli, C. Guillarmou and M. Herzlich, Opérateurs géométriques, invariants conformes et variétés asymptotiquement hyperboliques, Panoramas et Synthèses 26, Société Mathématique de France, 2008. [2] C. Fefferman and C. R. Graham, Conformal invariants, The mathematical heritage of Élie Cartan (Lyon, 1984), Astérisque, (1985), Numero Hors Serie, 95–116. C. Fefferman and C. R. Graham, The Ambient Metric. Annals of Math. Studies 178, Princeton University Press, 2012. <arXiv:0710.0919> C. Fefferman and R. Graham, Juhl's formulae for GJMS-operators and $Q$-curvatures, J. Amer. Math. Soc. 26 (4) (2013) 1191–1207. <arXiv:1203.0360> M. Fischmann, A. Juhl and P. Somberg, Conformal Symmetry Breaking Differential Operators on Differential Forms, Memoirs Amer. Math. Soc. 268 (2021) no. 1304. <arXiv:1605.04517> M. Fischmann and B. Ørsted, A family of Riesz distributions for differential forms on Euclidian space, Int. Math. Research Notices (2020). <arXiv:1702.00930> M. Fischmann, B. Ørsted and P. Somberg, Bernstein-Sato identities and conformal symmetry breaking operators, J. Funct. Anal. 277 11 (2019). <arXiv:1711.01546> J. Frahm and C. Weiske, Symmetry breaking operators for real reductive groups of rank one, J. Funct. Anal. 279, 5, (2020). I. M. Gelfand and G. E. Shilov, Generalized Functions, Vol. 1: Properties and Operations. Academic Press Inc., 1964. [3] M. Glaros, R. Gover, M. Halbasch and A. Waldron, Variational calculus for hypersurface functionals: singular Yamabe problem Willmore energies, J. Geom. Phys. 138 (2019) 168–193. R. Gover, Almost Einstein and Poincaré-Einstein manifolds in Riemannian signature, Journal of Geometry and Physics 60 (2) (2010) 182–204. <arXiv:0803.3510v1> A. R. Gover, E. Latini and A. Waldron, Poincaré-Einstein holography for forms via conformal geometry in the bulk, Memoirs Amer. Math. Soc. 235 (2015) no. 1106. A. R. Gover and L. Peterson, Conformal boundary operators, $T$-curvatures, and conformal fractional Laplacians of odd order. <arXiv:1802.08366> A. R. Gover and A. Waldron, Boundary calculus for conformally compact manifolds, Indiana Univ. Math. J. 63 (1) (2014) 119–163. <arXiv:1104.2991v2> A. R. Gover and A. Waldron, Generalising the Willmore equation: submanifold conformal invariants from a boundary Yamabe problem. <arXiv:1407.6742v1> A. R. Gover and A. Waldron, Conformal hypersurface geometry via a boundary Loewner-Nirenberg-Yamabe problem, Comm. in Analysis and Geometry 29 (4) (2021). A. R. Gover and A. Waldron, Renormalized volume, Comm. in Math. Physics 354 (3) (2017) 1205–1244. <arXiv:1603.07367> A. R. Gover and A. Waldron, Renormalized volume with boundary, Comm. Contemp. Math. 21 (2) (2019). R. Gover and A. Waldron, A calculus for conformal hypersurfaces and new higher Willmore energy functionals, Adv. Geom. 20, 1, (2020), 29–60. <arXiv:1611.04055v1>[This is basically the second of two parts of <cit.>. The other part corresponds to <arXiv:1506.02723v4> and is published as <cit.>.] A. R. Gover and A. Waldron, Singular Yamabe and Obata problems, in Differential Geometry in the Large, London Math. Soc. Lecture Notes Series 463, Cambridge Univ. Press (2021), 193–214. <arXiv:1912.13114v1> C. R. Graham, R. Jenne, L. J. Mason and G. A. J. Sparling, Conformally invariant powers of the Laplacian. I. Existence, J. London Math. Soc. (2) 46 (3) (1992), C. R. Graham and M. Zworski, Scattering matrix in conformal geometry, Inventiones math. 152 (1) (2003), 89–118. <arXiv:math/0109089> C. R. Graham and K. Hirachi, The ambient obstruction tensor and $Q$-curvature, in AdS/CFT Correspondence: Einstein Metrics and their Conformal Boundaries, IRMA Lectures in Mathematics and Theoretical Physics 8 (2005), 59–71. <arXiv:math/0405068> C. R. Graham and A. Juhl, Holographic formula for $Q$-curvature, Advances in Math. 216 (2) (2007) 841–853. <arXiv:0704.1673v1> C. R. Graham and M. Gursky, Chern-Gauss-Bonnet formula for singular Yamabe metrics in dimension four. Indiana Univ. Math. J. 70 (3), 1131–1166. <arXiv:1902.01562> C. R. Graham, Volume and area renormalizations for conformally compact Einstein metrics, Rend. Circ. Mat. Palermo (2) Suppl. 63 (2000) 31–42. <arXiv:math/9909042v1> C. R. Graham, Volume renormalization for singular Yamabe metrics, Proc. Amer. Math. Soc. 145 (2017) 1781–1792. <arXiv:1606.00069> D. H. Grant, A conformally invariant third order Neumann-type operator for hypersurfaces, Master's thesis, The University of Auckland, 2003. A. Gray, Tubes, Progress in Math. 221, Birkhäuser, 2004. C. Guillarmou, Meromorphic properties of the resolvent on asymptotically hyperbolic manifolds, Duke Math. J. 129 (1) (2005), 1–37. <arXiv:math/0311424> M. Halbasch, Asymptotic Yamabe-problem obstruction densities. Thesis, University California, Davis (2015). S. Helgason, Groups and Geometric Analysis. Integral Geometry, Invariant Differential Operators, and Spherical Functions, Academic Press, 1984. L. Hörmander, The Analysis of Linear Partial Differential Operators I. Springer 1983. G. Huisken and A. Polden, Geometric Evolution Equations for Hypersurfaces, in S. Hildebrand, M. Struwe (eds) Calculus of Variations and Geometric Evolution Problems, Lecture Notes in Mathematics 1713 (1999) 45–84. A. Juhl, Families of Conformally Covariant Differential Operators, $Q$-curvature and Holography, Birkhäuser, Progress in Math. 275, Birkhäuser, 2009. A. Juhl, Holographic formula for $Q$-curvature. II, Advances in Math. 226 (2011) 3409–3425. A. Juhl, Explicit formulas for GJMS-operators and $Q$-curvatures, Geom. Funct. Anal. 23 (4) (2013) 278–1370. <arXiv:1108.0273> A. Juhl, On the recursive structure of Branson's $Q$-curvature, Math. Res. Lett. 21 (3) (2014), 495–507. A. Juhl, Extrinsic Paneitz operators and $Q$-curvatures for hypersurfaces. <arXiv:1210.04838> A. Juhl and B. Ørsted, Shift operators, residue families and degenerate Laplacians, Pacific J. Math. 308 (1) (2020), 103–160. <arXiv:1806:02556> A. Juhl and B. Ørsted, On singular Yamabe obstructions. J. Geom. Anal. 32, 146 (2022). H. Kleinert, The membrane properties of condensing strings, Phys. Lett. B 174 (1986) 335–338. T. Kobayashi, T. Kubo and M. Pevzner, Conformal Symmetry Breaking Operators for Differential Forms on Spheres, Lecture Notes in Math. 2170 (2016). <arXiv:1605.05722> T. Kobayashi, B. Ørsted, P. Somberg and V. Souček, Branching laws for Verma modules and applications in parabolic geometry. I. Advances in Math. 285, (2015), 1796–1852. T. Kobayashi and B. Speh, Symmetry Breaking for Representations of Rank One Orthogonal Groups 238, Memoirs of AMS, Number 1126 (2015). <arXiv:1310.3213> T. Kobayashi and B. Speh, Symmetry Breaking for Representations of Rank One Orthogonal Groups II, Lecture Notes in Mathematics, 2234, Springer (2018), xv+342. <arXiv:1801.00158> C. Loewner and L. Nirenberg, Partial differential equations invariant under conformal or projective transformations, in Contributions to analysis (a collection of papers dedicated to Lipman Bers) (1974) 245–272. R. Mazzeo, Regularity for the singular Yamabe problem, Indiana Univ. Math. J., 40 (4) (1991) 1277–1299. A. Mondino and Huy The Nguyen, Global conformal invariants of submanifolds, Ann. Inst. Fourier (Grenoble), 68 (6) (2018) 2663–2695. <arXiv:1501.07527v2> J. Möllers and Y. Oshima, Restriction of most degenerate representations of $O(1,N)$ with respect to symmetric pairs, J. Math. Sci. Univ. Tokyo 22 (1) (2015) 279–338. O'Hara, J. Energy of Knots and Conformal Geometry, Series on Knots and Everything, 33, World Scientific Publishing Co. (2003). O'Hara, J. Residues of manifolds. <arXiv:2012.01713> A. Polyakov, Fine structure of strings, Nuclear Physics B 268 (1986) 406–412. S. Pigola and G. Veronelli, The smooth Riemannian extension problem: completeness. H. Rosenberg, Hypersurfaces of constant curvature in space forms, Bull. Sci. Math. 117 (2) (1993) G. Thomsen, Über konforme Geometrie I: Grundlagen der konformen Flächentheorie, Abh. Math. Sem. Hamburg 3 (1924) 31–56. Y. Vyatkin, Manufacturing conformal invariants of hypersurfaces, PhD thesis, University of Auckland, 2013. T. J. Willmore, Riemannian Geometry, Oxford Science Publications, 1993. E. Witten, Anti de Sitter space and holography, Adv. Theor. Math. Phys. 2 (2) (1998) 253–291.
# A Closer Look at Temporal Sentence Grounding in Videos: Dataset and Metric Yitian Yuan<EMAIL_ADDRESS>Tsinghua University , Xiaohan Lan <EMAIL_ADDRESS>Tsinghua University , Xin Wang <EMAIL_ADDRESS>Tsinghua University & Pengcheng Laboratory , Long Chen<EMAIL_ADDRESS>Columbia University , Zhi Wang <EMAIL_ADDRESS>Tsinghua University and Wenwu Zhu∗ <EMAIL_ADDRESS>Tsinghua University & Pengcheng Laboratory (2021) ###### Abstract. Temporal Sentence Grounding in Videos (TSGV), i.e., grounding a natural language sentence which indicates complex human activities in a long and untrimmed video sequence, has received unprecedented attentions over the last few years. Although each newly proposed method plausibly can achieve better performance than previous ones, current TSGV models still tend to capture the moment annotation biases and fail to take full advantage of multi-modal inputs. Even more incredibly, several extremely simple baselines without training can also achieve state-of-the-art performance. In this paper, we take a closer look at the existing evaluation protocols for TSGV, and find that both the prevailing dataset splits and evaluation metrics are the devils to cause unreliable benchmarking. To this end, we propose to re-organize two widely-used TSGV benchmarks (ActivityNet Captions and Charades-STA). Specifically, we deliberately make the ground-truth moment distribution _different_ in the training and test splits, i.e., out-of-distribution (OOD) testing. Meanwhile, we introduce a new evaluation metric “dR@$n$,IoU@$m$” to calibrate the basic IoU scores by penalizing on the bias-influenced moment predictions and alleviate the inflating evaluations caused by the dataset annotation biases such as overlong ground-truth moments. Under our new evaluation protocol, we conduct extensive experiments and ablation studies on eight state-of-the-art TSGV methods. All the results demonstrate that the re- organized dataset splits and new metric can better monitor the progress in TSGV. Our reorganized datsets are available at https://github.com/yytzsy/grounding_changing_distribution. temporal sentence grounding in videos, dataset bias, evaluation metric, dataset re-splitting, out-of-distribution testing ††journalyear: 2021††copyright: acmcopyright††conference: Proceedings of the 2nd International Workshop on Human-centric Multimedia Analysis; October 20, 2021; Virtual Event, China††booktitle: Proceedings of the 2nd International Workshop on Human-centric Multimedia Analysis (HUMA ’21), October 20, 2021, Virtual Event, China††price: 15.00††doi: 10.1145/3475723.3484247††isbn: 978-1-4503-8671-5/21/10††ccs: Computing methodologies Artificial intelligence††ccs: Computing methodologies Natural language processing††ccs: Computing methodologies Computer vision ## 1\. Introduction Figure 1. (a): TSGV aims to localize a moment with the start timestamp (21.3s) and end timestamp (30.7s). (b): The performance comparisons of some SOTA TSGV models with Bias-based baseline (orange bar) on Charades-STA with evaluation metric R<EMAIL_ADDRESS>(c): The performance comparisons of some SOTA TSGV models with PredictAll baseline (orange bar) on ActivityNet Captions with evaluation metric R<EMAIL_ADDRESS> Detecting human activities of interest from untrimmed videos is a prominent and fundamental problem in video scene understanding. Early video action localization works (Shou et al., 2016; Wang et al., 2016) mainly focus on detecting activities belonging to some predefined categories (Shou et al., 2016; Wang et al., 2016), which extremely restrict their flexibility and can hardly cover various human activities in life. For this purpose, a more challenging but meaningful task which extends the limited categories to open natural language descriptions was proposed (Anne Hendricks et al., 2017; Gao et al., 2017; Hendricks et al., 2018), dubbed as Temporal Sentence Grounding in Videos (TSGV). As the example shown in Figure 1 (a), given a natural language query and an untrimmed video, TSGV needs to identify the start and end timestamps of one segment (i.e., moment) in the video, which semantically corresponds to the language query. Due to its profound significance, the TSGV task has received unprecedented attentions over the last few years — a surge of datasets (Anne Hendricks et al., 2017; Gao et al., 2017; Krishna et al., 2017; Regneri et al., 2013) and methods (Chen et al., 2018, 2020; Duan et al., 2018; Gao et al., 2019; Ge et al., 2019; Hahn et al., 2019; He et al., 2019; Jiang et al., 2019; Liu et al., 2018a, b; Lu et al., 2019; Wang et al., 2019; Wu et al., 2020; Xu et al., 2019; Yuan et al., 2019a, b; Zeng et al., 2020; Zhang et al., 2019a, 2020, b) have been developed. Although each newly proposed method can plausibly achieve better performance and make progress over previous ones, a recent study (Otani et al., 2020) shows that even today’s state-of-the-art (SOTA) TSGV models still fail to make full use of multi-modal inputs, i.e., they over-rely on the ground-truth moment annotation biases in current benchmarks, and lack of sufficient understanding of the multi-modal inputs. Specifically, taking one prevailing benchmark Charades-STA (Gao et al., 2017) as an example, suppose that there is a Bias-based baseline model which only makes predictions by sampling a moment from the frequency statistics of the ground-truth moment annotations in the training set. As illustrated in Figure 1 (b), this naive Bias-based model unexpected surpasses several SOTA deep models, i.e., Charades-STA has obvious moment location annotation biases. Therefore, _we argue that current TSGV datasets with heavily biased annotations cannot accurately monitor the progress in TSGV research._ Meanwhile, another characteristic of the ground-truth moment annotations in current TSGV benchmarks is that they usually have relatively long temporal durations. For example, 40% queries in the ActivityNet Captions dataset refer to a moment occupying over 30% temporal ranges of the whole input video. These overlong ground-truth moments incidentally make the current evaluation metrics unreliable. Specifically, the most prevalent evaluation metric for today’s TSGV is “R@$n$,IoU@$m$”, i.e., the percentage of testing samples which have at least one of the top-$n$ results with IoU larger than $m$. Due to the difficulty of the TSGV task, almost all published TSGV works tend to use a _small_ IoU threshold $m$ (i.e., 0.3) for evaluation, especially for challenging datasets (i.e., ActivityNet Captions (Krishna et al., 2017)). However, _we argue that the metric R@ $n$,IoU@$m$ with small $m$ is unreliable for the datasets with overlong ground-truth moment annotations_. For example, a small IoU threshold can be easily achieved by a long duration moment prediction. As an extreme case, a simple baseline which always directly takes the whole input video as the prediction (cf. the PredictAll baseline in Figure 1 (c)) can still achieve a SOTA performance under this metric. In this paper, to help disentangle the effects of ground-truth moment annotation biases, we propose to resplit the widely-used TSGV benchmarks (i.e., Charades-STA and ActivityNet Captions) by changing their ground-truth moment annotation distributions, obtaining two new evaluation benchmarks: Charades-CD (Charades-STA under Changing Distributions) and ActivityNet-CD. These new splits are created by re-organizing all splits (the training, validation and test sets) of original datasets, and the ground-truth moment distributions are deliberately designed _different_ in the training and test splits, i.e., out-of-distribution (OOD) testing. To better evaluate models’ generalization ability and compare the performance between the OOD samples and the independent and identically distributed (IID) samples, we also maintain a test split with IID samples, denoted as test-iid set (vs. test-ood set). Meanwhile, we propose a more reliable evaluation metric — dR@$n$, IoU@$m$ — for small threshold $m$. This metric calibrates the basic IoU scores with the temporal location discrepancy between the predicted and ground-truth moments, which is expected to reduce the influence of moment durations and restraint the inflating evaluations caused by overlong ground-truth moments in the datasets. To demonstrate the difficulty of our new splits and monitor the progress in TSGV, we evaluate the performance of eight representative SOTA models on our new evaluation protocol. Our key finding is that the performance of most tested models drops significantly when evaluated on the OOD samples (i.e., the test-ood set) compared to the IID samples (i.e., the test-iid set). This finding provides further evidences that existing methods only fit the moment annotation biases, and fail to bridge the semantic gaps between the video contents and natural language queries. Meanwhile, the proposed metric (dR@$n$,Iou@$m$) can effectively reduce the inflating performance caused by the annotation biases when the IoU threshold $m$ is small. In summary, we make three contributions in this paper: * • We propose new splits of two prevailing TSGV datasets, which are able to disentangle the effects of annotation biases. * • We propose a new metric: dR@$n$,IoU@$m$, which is more reliable than the existing metrics, especially when IoU threshold is small. * • We conduct extensive studies with several SOTA models. Consistent performance gaps between IID and OOD samples have proven that our new evaluation protocol can better monitor the progress in TSGV. ## 2\. Related Works ### 2.1. Temporal Sentence Grounding in Videos In this section, we coarsely group existing TSGV methods into four categories: _Two-Stage Methods._ Early TSGV methods typically solve this problem in a two- stage fashion: They first extract numerous video segment candidates by temporal sliding windows, and then either match the query sentence with these candidates (Anne Hendricks et al., 2017) or fuse query and video segment features to regress the final position, e.g., CTRL (Gao et al., 2017), ACL-K (Ge et al., 2019), SLTA (Jiang et al., 2019), ACRN (Liu et al., 2018a), ROLE (Liu et al., 2018b), VAL (Song and Han, 2018) and BPNet (Xiao et al., 2021). To speed up the sliding window processing, Xu et.al. (Xu et al., 2019) proposed QSPN, which injects text features early to generate segment candidates, and helps to eliminate the unlikely segment candidates and increases the grounding accuracy. _End-to-End Methods._ Besides the two-stage framework, some other TSGV works seek to solve the grounding problem in an end-to-end manner (Chen et al., 2018; Yuan et al., 2019a, b; Zeng et al., 2020; Zhang et al., 2019a, 2020, b). Chen et.al. (Chen et al., 2018) proposed TGN, which sequentially scores a set of temporal candidates ended at each frame and generates the final grounding result in one single pass. Similarly, ABLR model also processes video sequences via LSTMs (Yuan et al., 2019b), where the start and end timestamps of the predicted segments are regressed from the attention weights yielded by the multi-pass interaction between videos and queries. There are also some works leveraging temporal convolutional networks to solve the TSGV problem. Zhang et.al. (Zhang et al., 2019a) presented MAN, which assigns candidate segment representations aligned with language semantics over different temporal locations and scales in hierarchical temporal convolutional feature maps. Yuan et.al. (Yuan et al., 2019a) introduced the SCDM, where query semantic is used to control the feature normalization between different temporal convolutional layers, making the query-related video activities tightly compose together. Both MAN and SCDM only consider 1D temporal feature maps, while 2D-TAN (Zhang et al., 2020) models the temporal relations between video segments by a 2D map. In the 2D map, 2D-TAN encodes the adjacent temporal relation, and learns discriminative features for matching video segments with queries. _RL-based Methods._ Some recent models also regard the TSGV task as a sequence decision making problem, and resort to Reinforcement Learning (RL) algorithms. Specifically, Wang et.al. (Wang et al., 2019) introduced a semantic matching RL (SM-RL) model by extracting semantic concepts of videos and fusing them with global context features. Then, video contents are selectively observed and associated with the given sentence in a matching-based manner. Hahn et.al. (Hahn et al., 2019) presented TripNet, which uses RL to efficiently localize relevant activity clips in long videos, by learning how to intelligently hop around the video. Wu et.al. (Wu et al., 2020) formulated a tree-structured policy based progressive RL (TSP-PRL) model to sequentially regulate the predicated temporal boundaries by an iterative refinement process. _Weakly Supervised Methods._ Since the ground-truth annotations for the TSGV task are manually consuming, some works start to extend this problem to a weakly supervised scenario where the ground-truth segments are unavailable in the training stage (Duan et al., 2018; Gao et al., 2019; Mithun et al., 2019; Song et al., 2020; Tan et al., 2019). Mithun et.al. (Mithun et al., 2019) utilized a latent alignment between video frames and sentence descriptions with Text-Guided Attention (TGA), and TGA was used during the test stage to retrieve relevant moments. Duan et.al. (Duan et al., 2018) took the TSGV task as an intermediate step for dense video captioning, and then they established a cycle system and leveraged the captioning loss to train the whole model. Song et.al. (Song et al., 2020) presented a multi-level attentional reconstruction network, which leverages both intra- and inter-proposal interactions to learn a language-driven attention map, and can directly rank the candidate proposals at the inference stage. ### 2.2. Biases in TSGV Datasets A recent work (Otani et al., 2020) also discusses the dataset bias problem in current TSGV benchmarks. The main contribution of (Otani et al., 2020) is to find and analyse the moment annotation biases in previal benchmarks and perform human studies to demonstrate the disagreement among different annotators. In contrast, we propose to re-split these datasets to reduce these ground-truth moment annotation biases, and introduce a new metric to alleviate the inflating performance of SOTA models, i.e., we go one step further to build a more reasonable and reliable evaluation protocol. Meanwhile, we reproduce and analyse eight different SOTA methods from four different categories on both original and new evaluation protocols. ## 3\. Dataset and Metric Analysis ### 3.1. Dataset Analysis So far, there are four available TSGV datasets in our communities: DiDeMo (Anne Hendricks et al., 2017), TACoS (Regneri et al., 2013), Charades-STA (Gao et al., 2017) and ActivityNet Captions (Krishna et al., 2017). Since both DiDeMo and TACoS have some inherent and obvious disadvantages (e.g., For DiDeMo, the unit interval of annotations is five seconds; for TACoS, the visual scene is restricted in kitchen), dataset Charades-STA and ActivityNet Captions gradually become the mainstream benchmarks for TSGV evaluation (Chen et al., 2018; Hahn et al., 2019; Xu et al., 2019; Yuan et al., 2019a; Zeng et al., 2020; Zhang et al., 2020). The details about these two datasets are as follows: Charades-STA. It is built upon the Charades (Sigurdsson et al., 2016) dataset. The average length of videos in Charades is 30 seconds, and each video is annotated with multiple descriptions, action labels, action intervals, and classes of interacted objects. Gao et.al. (Gao et al., 2017) extended the Charades dataset to the TSGV task by assigning the temporal intervals to text descriptions and matching the common key words in the interval action labels and texts. In the official split (Gao et al., 2017), there are 5,338 videos and 12,408 query-moment pairs in the training set, and 1,334 videos and 3,720 query-moment pairs in the test set (cf. Table 1). ActivityNet Captions. It is originally developed for the dense video captioning task (Krishna et al., 2017). Since the official test set is withheld, previous TSGV works (Yuan et al., 2019a, b) merge the two available validation subsets “val1” and “val2” as the test set. In summary, there are 10,009 videos and 37,421 query-moment pairs in the training set, and 4,917 videos and 34,536 query-moment pairs in the test set. (cf. Table 1). Figure 2. The ground-truth moment annotation distributions of all query-moment pairs in Charades-STA and ActivityNet Captions. The deeper the color, the larger density in distributions. To examine the ground-truth moment annotation distributions of these two datasets, we normalize the start and end timestamps of all annotated moments in both the training and test sets, and use Gaussian kernel density estimation to fit the joint distribution of these normalized start and end timestamps. As shown in Figure 2, for both two datasets, the moment annotation distributions are almost identical in the training and test sets. For Charades-STA, most of the moments start at the beginning of the videos and end at around $20\%-40\%$ of the length of the videos. The moment annotation distributions present a strip with relatively uniform width, which indicates that the length of moment in Charades-STA roughly concentrates within a certain range. For ActivityNet Captions, the distributions are significantly different from those of Charades-STA, which concentrates in three local areas, i.e., the three corners. All these areas show that a considerable number of ground-truth moments start at the beginning of the video or end at the ending of the video, even exactly the same as the whole video (the left top area). This may be due to that the dataset ActivityNet Captions is originally annotated for dense video captioning, and the captions (queries) are always annotated based on the whole video. Therefore, we can observe that the ground-truth moment annotations in both benchmarks consists of strong biases. In other word, by fitting these moment annotation biases, a simple baseline can also achieve a state-of-the-art performance (cf. Figure 1). ### 3.2. Evaluation Metric Analysis To evaluate the temporal grounding accuracy, almost all existing TSGV works adopt the “R@$n$,IoU@$m$” as a standard evaluation metric. Specifically, for each query $q_{i}$, it first calculates the Intersection-over-Union (IoU) between the predicted moment and its ground-truth, and this metric is formally defined as: (1) $\text{R@$n$,IoU@$m$}=\frac{1}{N_{q}}\sum_{i}r(n,m,q_{i}),$ where $r(n,m,q_{i})=1$ if there is at least one of top-$n$ predicted moments of query $q_{i}$ having an IoU larger than threshold $m$, otherwise it equals to 0. $N_{q}$ is the total number of all queries. Most of previous TSGV methods (Chen et al., 2018; Liu et al., 2018b; Xu et al., 2019; Yuan et al., 2019b; Zhang et al., 2020) always report their scores on some small IoU thresholds like $m\in\\{0.1,0.3,0.5\\}$. However, as shown in Figure 3 (b), for dataset ActivityNet Captions, a substantial proportion of ground-truth moments have relatively long durations. Statistically, 40%, 20%, and 10% of sentence queries refer to a moment occupying over 30%, 50%, and 70% of the length of the whole video, respectively. Such annotation biases can obviously increase the chance of correct predictions under small IoU thresholds. Taking an extreme case as example, if the ground-truth moment is the whole video, any predictions with duration longer than 0.3 can achieve R<EMAIL_ADDRESS>Thus, metric R@$n$,IoU@$m$ with small $m$ is unreliable for current biased annotated datasets. Figure 3. The histogram of the normalized ground-truth moment durations in Charades-STA and ActivityNet Captions. Figure 4. (a) and (b) illustrate the ground-truth moment annotation distributions of each split in Charades-CD and ActivityNet-CD, respectively. (c) presents the moment annotation distributions of the query-indicated moments which contain action cook in the training and test-ood sets of Charades-CD. The deeper the color, the larger the density in the distribution. Figure 5. The frequency distributions of the top-30 actions in the query-moment pairs of different splits. The longer the bar, the more frequently the action appears. ## 4\. Proposed Evaluation Protocol ### 4.1. Dataset Re-splitting To accurately monitor the research progress in TSGV and reduce the influence of moment annotation biases, we propose to re-organize the two datasets (i.e., Charades-STA and ActivityNet Captions) by deliberately assigning different moment annotation distributions in each split. Particularly, each dataset is re-splitted into four sets: _training_ , _validation (val)_ , _test-iid_ , and _test-ood_. All samples in the training, val, and test-iid sets satisfy the independent and identical distribution, and the samples in test-ood set are out-of-distribution. The performance gap between the test-iid set and test-ood set can effectively reflect the generalization ability of the models. We name the two new re-organized datasets as Charades-CD and ActivityNet-CD. Dataset Aggregation and Splitting. For each dataset, we merge the training and test sets by aggregating all the query-moment pairs, i.e., Charades-STA has 12,408 + 3,720 = 16,128 pairs overall and ActivityNet Captions has 37,421 + 34,536 = 71,957 pairs in total (cf. Table 1). We first regard each query- moment pair as a data sample. Then, we use the Gaussian kernel density estimation as mentioned in Section 3.1 (cf. Figure 2) to fit the moment annotation distribution among these data samples. In the fitted distribution, each moment has a density value based on its temporal location in the video. We rank all the moments (as well as their paired queries) based on their density values in a descending order, and take the lower 20% data samples as the preliminary test-ood set, i.e., the temporal locations distribution of the preliminary test-ood set is furthest _different_ from the distribution of the whole dataset. The remaining 80% data sample are divided into the preliminary training set. Conflicting Video Elimination. Since each video is associated with multiple sentence queries, another concern is that we need to make sure that there is no video overlap between the training and test sets. Thus, after obtaining the preliminary test-ood set, we check whether the videos of these samples also appear in the preliminary training set. If it is the case, we move all samples (i.e., query-moment pairs) referring to the same video into the split with most of samples. In addition, to avoid the inflating performance of overlong predictions in ActivityNet-CD (cf. the PredictAll baseline in Figure 1), we leave all samples with ground-truth moment occupying over 50% of the length of the whole video into the training set. After eliminating all conflicting videos, we obtain the final test-ood set, which consists of around 20% query-moment pairs of the whole dataset. Then, we randomly divide the remaining samples (based on videos) into three splits: the training, val, and test-iid sets, which consist of around 70%, 5%, and 5% data samples, respectively. The statistics of the new proposed splits are reported in Table 1. Dataset | Split | # Videos | # Pairs ---|---|---|--- Charades-STA | training | 5,338 | 12,408 test | 1,334 | 3,720 ActivityNet Captions | training | 10,009 | 37,421 test | 4,917 | 34,536 Charades-CD | training | 4,564 | 11,071 val | 333 | 859 test-iid | 333 | 823 test-ood | 1,442 | 3,375 ActivityNet-CD | training | 10,984 | 51,415 val | 746 | 3,521 test-iid | 746 | 3,443 test-ood | 2,450 | 13,578 Table 1. The detailed statistics of the number of videos and query-moment pairs in different datasets and splits. ### 4.2. Charades-CD and ActivityNet-CD Moment Annotation Distributions. The ground-truth moment annotation distributions of Charades-CD and ActivityNet-CD are illustrated in Figure 4. From the Figure 4, we can observe that the moment annotation distributions of the test-ood set are significantly different from those of the other three sets (i.e., training, val, and test-iid sets). Compared with the moment annotation distributions of original test split (cf. Figure 2), the proposed test-ood split has several improvements: 1) For Charades-CD, the distributions of the start timestamps of the moments are more diverse (vs. concentrating on the beginning of the videos). 2) For ActivityNet-CD, more moments locate in relatively central areas of the videos, i.e., models will not perform well by over relying on the annotation biases. Action Distributions. We also investigate the action distributions of the original and re-organized datasets. Specifically, for each dataset, we extract the verbs from all sentence queries and count the frequency of each verb. Since the verb frequencies satisfy a long-tail distribution, we select the top-30 frequent verbs, which cover 92.7% of all action types in Charades-CD and 52.9% for ActivityNet-CD, respectively. The statistical results are illustrated in Figure 5. From this figure, we can observe that the new test- ood sets on both two datasets still have similar action distributions with the training set and original splits, which shows the OOD of moment annotations comes from each verb type. As shown in Figure 4 (c), the moment annotation distribution of the new training and test-ood set are totally different for the verb _cook_. ### 4.3. Proposed Evaluation Metric As discussed in Section 3.2, the most prevailing evaluation metric — R@$n$,IoU@$m$ — is unreliable under small threshold $m$. To alleviate this issue, as shown in Figure 6, we propose to calibrate the $r(n,m,q_{i})$ value by considering the “temporal distance” between the predicted and ground-truth moments. Specifically, we propose a new metric discounted-R@$n$,IoU@$m$ (dR@$n$,IoU@$m$): (2) $\text{dR@$n$,IoU@$m$}=\frac{1}{N_{q}}\sum_{i}r(n,m,q_{i})\cdot\alpha^{s}_{i}\cdot\alpha^{e}_{i},$ where $\alpha^{*}_{i}=1-\text{abs}(p_{i}^{*}-g_{i}^{*})$, and $\text{abs}(p_{i}^{*}-g_{i}^{*})$ is the absolute distance between the boundaries of predicted and ground-truth moments. Both $p_{i}^{*}$ and $g_{i}^{*}$ are normalized to the range (0, 1) by dividing the whole video length. When the predicted and ground-truth moments are very close to each other, the discount ratio $\alpha^{*}_{i}$ will be close to 1, i.e., the new metric can degrade to R@$n$,IoU@$m$ with exactly accurate predictions. Otherwise, even the IoU threshold condition is met, the score $r(n,m,q_{i})$ will still be discounted by $\alpha^{*}_{i}$, which helps to alleviate the inflating recall scores under small IoU thresholds. With the proposed dR@$n$,IoU@$m$ metric, those speculation methods which over-rely on moments annotation biases (e.g., long moments annotations in ActivityNet Captions) will not perform well. Figure 6. An illustration of the proposed dR@$n$,IoU@$m$ metric. Figure 7. Performances (%) of SOTA TSGV methods on the test set of original splits (Charades-STA and ActivityNet Captions) and test sets (test-iid and test-ood) of proposed splits (Charades-CD and ActivityNet-CD). We use metric R<EMAIL_ADDRESS>in all cases. ## 5\. Experiments ### 5.1. Benchmarking the SOTA TSGV Methods To demonstrate the difficulty of the new proposed splits (i.e., Charades-CD and ActivityNet-CD), we compare the performance of two simple baselines and eight representative state-of-the-art methods on both the original and proposed splits. Specifically, we can categorize these methods into the following groups: * • _Bias-based Method_ : It uses the Gaussian kernel density estimation to fit the moment annotation distribution, and randomly samples several locations based on the fitted distribution as the final moment predictions. * • _PredictAll Method_ : It directly predicts the whole video as the final moment predictions. * • _Two-Stage Methods_ : Cross-modal Temporal Regression Localizer (CTRL) (Gao et al., 2017), and Attentive Cross-modal Retrieval Network (ACRN) (Liu et al., 2018a). * • _End-to-End Methods_ : Attention-Based Location Regression (ABLR) (Yuan et al., 2019b), Semantic Conditioned Dynamic Modulation (SCDM) (Yuan et al., 2019a), 2D Temporal Adjacent Network (2D-TAN) (Zhang et al., 2020), and Dense Regression Network (DRN) (Zeng et al., 2020). * • _RL-based Method_ : Tree-Structured Policy based Progressive Reinforcement Learning (TSP-PRL) (Wu et al., 2020). * • _Weakly-supervised Method_ : Weakly-Supervised Sentence Localizer (WSSL) (Duan et al., 2018). For all these SOTA methods, we use the public available official implementations to get the temporal grounding results. The results of the proposed test-iid and test-ood sets on two datasets come from the same model finetuned on the val set. For more fair comparisons, we have unified the feature representations of the videos and sentence queries. To cater for most of TSGV methods, we use I3D feature (Carreira and Zisserman, 2017) for the videos in dataset Charades-STA (Charades-CD), and C3D feature (Tran et al., 2015) for the videos in dataset ActivityNet Captions (Activity-CD). Each word in the query sentences is encoded by a pretrained GloVe (Pennington et al., 2014) word embedding. Method | Split | Charades-CD | ActivityNet-CD ---|---|---|--- m=0.1 | m=0.3 | m=0.5 | m=0.7 | m=0.9 | m=0.1 | m=0.3 | m=0.5 | m=0.7 | m=0.9 Bias-based | test-iid | 31.42 | 26.25 | 16.87 | 9.34 | 2.70 | 36.15 | 29.31 | 19.81 | 12.27 | 7.68 test-ood | 14.75 | 9.30 | 5.04 | 2.21 | 0.55 | 21.89 | 9.21 | 0.26 | 0.11 | 0.03 PredictAll | test-iid | 31.04 | 10.93 | 0.00 | 0.00 | 0.00 | 36.43 | 29.62 | 20.05 | 12.45 | 7.83 test-ood | 37.43 | 27.13 | 0.06 | 0.00 | 0.00 | 21.87 | 9.01 | 0.00 | 0.00 | 0.00 CTRL (Gao et al., 2017) | test-iid | 50.61 | 42.65 | 29.80 | 11.86 | 1.41 | 27.34 | 19.42 | 11.27 | 4.29 | 0.25 test-ood | 52.80 | 44.97 | 30.73 | 11.97 | 1.12 | 26.23 | 15.68 | 7.89 | 2.53 | 0.20 ACRN (Liu et al., 2018a) | test-iid | 53.22 | 47.50 | 31.77 | 12.93 | 0.71 | 27.69 | 20.06 | 11.57 | 4.41 | 0.75 test-ood | 53.36 | 44.69 | 30.03 | 11.89 | 1.38 | 27.03 | 16.06 | 7.58 | 2.48 | 0.17 ABLR (Yuan et al., 2019b) | test-iid | 59.26 | 52.26 | 41.13 | 23.50 | 3.66 | 55.62 | 46.86 | 35.45 | 20.57 | 6.32 test-ood | 54.09 | 44.62 | 31.57 | 11.38 | 1.39 | 46.88 | 33.45 | 20.88 | 10.03 | 2.31 SCDM (Yuan et al., 2019a) | test-iid | 62.47 | 58.14 | 47.36 | 30.79 | 6.62 | 55.15 | 46.44 | 35.15 | 22.04 | 6.07 test-ood | 59.08 | 52.38 | 41.60 | 22.22 | 3.81 | 45.08 | 31.56 | 19.14 | 9.31 | 1.94 2D-TAN (Zhang et al., 2020) | test-iid | 59.80 | 53.71 | 43.46 | 24.99 | 6.95 | 57.11 | 49.18 | 39.63 | 27.36 | 9.00 test-ood | 50.87 | 43.45 | 30.77 | 11.75 | 1.92 | 44.37 | 30.86 | 18.38 | 9.11 | 2.05 DRN (Zeng et al., 2020) | test-iid | 57.03 | 51.35 | 41.91 | 26.74 | 6.46 | 56.96 | 48.92 | 39.27 | 25.71 | 6.81 test-ood | 49.17 | 40.45 | 30.43 | 15.91 | 3.13 | 47.50 | 36.86 | 25.15 | 14.33 | 3.76 TSP-PRL (Wu et al., 2020) | test-iid | 54.60 | 46.44 | 35.43 | 17.01 | 3.57 | 53.98 | 44.93 | 33.93 | 19.50 | 4.79 test-ood | 42.21 | 31.93 | 19.37 | 6.20 | 1.16 | 44.23 | 29.61 | 16.63 | 7.43 | 1.46 WSSL (Duan et al., 2018) | test-iid | 45.90 | 34.99 | 14.06 | 4.27 | 0.00 | 36.67 | 26.06 | 17.20 | 6.16 | 1.24 test-ood | 49.92 | 35.86 | 23.67 | 8.27 | 0.06 | 30.71 | 17.00 | 7.17 | 1.82 | 0.17 Table 2. Performances (%) of SOTA TSGV methods on the Charades-CD and ActivityNet-CD datasets with metric dR@$1$,IoU@$m$. ### 5.2. Performance Comparisons on the Original and Proposed Data Splits We report the performance of all mentioned TSGV methods with metric R<EMAIL_ADDRESS>in Figure 7. From Figure 7, we can observe that almost all methods have a significant performance gap between the test-iid and test-ood sets, i.e., these methods always over-rely on the moment annotation biases, and fail to generalize to the OOD testing. Meanwhile, the performance results on the original test set and the proposed test-iid set are relatively close, which shows that the moment distribution of the test-iid set is similar to the majority of the whole dataset. We provide more detailed experimental result analyses in the following: Baseline Methods. After changing the moment annotation distributions in different splits, the Bias-based method cannot take advantage of the annotation biases and its performance degrades from 13.6% on the test-iid set to 0.1% on the test-ood set of ActivityNet-CD. For the PredictAll method, since all the ground-truth moments in Charades-CD are less than 50% range of the whole videos, naively predicting the whole video as the grounding results will inevitably cause the R<EMAIL_ADDRESS>scores to 0.0 on this dataset. Since the ground-truth moments in ActivityNet-CD are much longer, the PredictAll method achieves high results at 11.9% and 13.8% on the original test set and new test-iid set, respectively. However, in the test-ood set where the longer segments are excluded, the PredictAll method also degrades its performance to 0.0. Two-Stage Methods. We find that the two-stage methods (i.e., CTRL and ACRN) are less sensitive to the domain gaps between the test-iid and test-ood sets. This is due to that they use a sliding-window strategy to retrieve video moment candidates, and compare these moment candidates with each query sentence individually. In this manner, all moment candidates are independent to the overall video contents, and the moment annotation distributions have less influence on the model performance. We can also observe that the performance of these two methods on the test-ood set of ActivityNet-CD presents a more obvious drop compared to the performance on test-iid set. In contrast, the performance on the test-iid and test-ood sets of Charades-CD are competing. The main reason is that the ground-truth moments in the test-ood set of Charades-CD always occupy a longer range over the whole videos (cf. Figure 4 (a), which makes the sliding windows have more chance to hit the ground-truth moments. In summary, although CTRL and ACRN are less sensitive to the moment annotation biases, their grounding performances are still far behind other types of SOTA methods, e.g., SCDM and DRN. End-to-End Methods. As for the end-to-end methods (i.e., ABLR, SCDM, 2D-TAN and DRN), we can observe that their performances all drop significantly on the test-ood set compared to the test-iid set on both two datasets. These methods all have considered the whole video contexts and temporal information. The initial intention for this design is that numerous queries often contain some words referring to temporal orders and locations such as “before”, “after”, “begin” and “end”, or they want to encode the important temporal relations between video moments. Unfortunately, although our test-ood split does not break any video temporal relations, their performance on the OOD testing still drop significantly. This demonstrates that current methods do not play their advantages and fail to utilize the video temporal relation or vision-language interaction for TSGV. RL-based Method. The RL-based method TSP-PRL also suffers from obvious performance drops on the test-ood set compared to the test-iid set. Actually, TSP-PRL adopts IoU between the predicted and ground-truth moment as the training reward in the RL framework. In this case, the temporal annotations directly affect the model learning, and the changes of moment annotation distributions will inevitably influence the model performance. Weakly-supervised Method. The results of the weakly-supervised method WSSL is _thought-provoking_ : it achieves better performance on test-ood set compared to test-iid set in Charades-CD, but results of these splits in ActivityNet-CD are exactly the reverse. After carefully checking the predicted moment results, we find that the normalized (start, end) moment predictions on both two datasets converge on several certain predictions (i.e., (0, 1), (0, 0.5), (0.5, 1)). These results indicate that the WSSL method does not learn to align the video and sentence semantics at all. Instead, it only speculatively guesses several possible locations. ### 5.3. Performance Evaluation with dR@$n$,IoU@$m$ We report the performance of all mentioned TSGV methods with our proposed metric dR@$1$,IoU@$m$ in Table 2. The trend of performance drop on the test- ood set compared to the test-iid set in Table 2 is similar to that in Figure 7. These results verify again that current TSGV methods suffer from severe temporal annotation biases in the datasets, and fail to generalize to the OOD testing. Meanwhile, by comparing Table 2 and Figure 7, we can observe that the dR@$1$,IoU@$m$ values are smaller than the R@$1$,IoU@$m$ values. For example, the SCDM model achieves score 32.5% in R<EMAIL_ADDRESS>while score 30.8% in dR<EMAIL_ADDRESS>on the test-iid set of Charades-CD. Such phenomenon is adhere to our definition of dR@$1$,IoU@$m$. For more clearer illustration, we further compare the dR@$1$ and R@$1$ scores under different IoUs of some SOTA methods in Figure 8. When the IoU threshold is small, dR@$1$ is much lower than R@$1$, and the gap between them gradually decreases with the increase of IoU threshold. Interestingly, we find that the naive Bias-based baseline achieves even better results than SCDM and 2D-TAN methods in the R<EMAIL_ADDRESS>metric, while reversely in the dR<EMAIL_ADDRESS>metric. These results indicate that recall values under small IoU thresholds are unreliable and overrated: although some moment predictions meet the IoU requirement, they still have a great discrepancy to the ground-truth moments. Instead, our proposed dR@$n$,IoU@$m$ metric can alleviate this problem since it can discount the recall value based on the temporal distance between the predicted and ground- truth moment temporal locations. When the prediction meets the larger IoU requirements, the discount will be smaller, i.e., the dR@$n$,IoU@$m$ values and R@$n$,IoU@$m$ values will be closer to each other. Therefore, our predicted dR@$n$,IoU@$m$ metric is more stable on different IoU thresholds, and it can suppress some inflating results (such as Bias-based or PredictAll baselines) caused by the moment annotation biases in the datasets. Meanwhile, these results further reveal that it is more reliable to report the grounding accuracy on large IoUs. Figure 8. Performance (%) comparisons of SOTA TSGV methods between original metric (R@$1$,IoU@$m$) and proposed metric (dR@$1$,IoU@$m$). All results come from the test set of ActivityNet Captions. ## 6\. Conclusion In this paper, we take a closer look at the existing evaluation protocol of the Temporal Sentence Grounding in Videos (TSGV) task, and we find that both the prevailing dataset splits and evaluation metric are the devils to cause the unreliable benchmarking: the datasets have obvious moment annotation biases and the metric is prone to overrating the model performance. To solve these problems, we propose to re-split the current Charades-STA and ActivityNet Captions datasets by making the ground-truth moment annotation distributions different in the training and test set. Meanwhile, we propose a new evaluation metric to alleviate the inflating evaluations caused by dataset annotation biases such as overlong ground-truth moments. The proposed data splits and metric serve as a promising test-bed to monitor the progress in TSGV. We also thoroughly evaluate eight state-of-the-art TSGV methods with the new evaluation protocol, opening the door for future research. ## 7\. Acknowledgments This work was supported by the National Key Research and Development Program of China under Grant No.2020AAA0106301, National Natural Science Foundation of China No.62050110 and Tsinghua GuoQiang Research Center Grant 2020GQG1014. ## References * (1) * Anne Hendricks et al. (2017) Lisa Anne Hendricks, Oliver Wang, Eli Shechtman, Josef Sivic, Trevor Darrell, and Bryan Russell. 2017. Localizing moments in video with natural language. In _ICCV_. * Carreira and Zisserman (2017) Joao Carreira and Andrew Zisserman. 2017. Quo vadis, action recognition? a new model and the kinetics dataset. In _CVPR_. * Chen et al. (2018) Jingyuan Chen, Xinpeng Chen, Lin Ma, Zequn Jie, and Tat-Seng Chua. 2018. Temporally grounding natural sentence in video. In _EMNLP_. * Chen et al. (2020) Long Chen, Chujie Lu, Siliang Tang, Jun Xiao, Dong Zhang, Chilie Tan, and Xiaolin Li. 2020. Rethinking the Bottom-Up Framework for Query-Based Video Localization.. In _AAAI_. * Duan et al. (2018) Xuguang Duan, Wenbing Huang, Chuang Gan, Jingdong Wang, Wenwu Zhu, and Junzhou Huang. 2018\. Weakly supervised dense event captioning in videos. In _NeurIPS_. * Gao et al. (2017) Jiyang Gao, Chen Sun, Zhenheng Yang, and Ram Nevatia. 2017\. Tall: Temporal activity localization via language query. In _ICCV_. * Gao et al. (2019) Mingfei Gao, Larry Davis, Richard Socher, and Caiming Xiong. 2019\. WSLLN: Weakly Supervised Natural Language Localization Networks. In _EMNLP_. * Ge et al. (2019) Runzhou Ge, Jiyang Gao, Kan Chen, and Ram Nevatia. 2019\. Mac: Mining activity concepts for language-based temporal localization. In _WACV_. * Hahn et al. (2019) Meera Hahn, Asim Kadav, James M Rehg, and Hans Peter Graf. 2019\. Tripping through time: Efficient localization of activities in videos. In _arXiv_. * He et al. (2019) Dongliang He, Xiang Zhao, Jizhou Huang, Fu Li, Xiao Liu, and Shilei Wen. 2019\. Read, watch, and move: Reinforcement learning for temporally grounding natural language descriptions in videos. In _AAAI_. * Hendricks et al. (2018) Lisa Anne Hendricks, Oliver Wang, Eli Shechtman, Josef Sivic, Trevor Darrell, and Bryan Russell. 2018. Localizing moments in video with temporal language. In _EMNLP_. * Jiang et al. (2019) Bin Jiang, Xin Huang, Chao Yang, and Junsong Yuan. 2019\. Cross-modal video moment retrieval with spatial and language-temporal attention. In _ICMR_. * Krishna et al. (2017) Ranjay Krishna, Kenji Hata, Frederic Ren, Li Fei-Fei, and Juan Carlos Niebles. 2017. Dense-captioning events in videos. In _ICCV_. * Liu et al. (2018a) Meng Liu, Xiang Wang, Liqiang Nie, Xiangnan He, Baoquan Chen, and Tat-Seng Chua. 2018a. Attentive moment retrieval in videos. In _SIGIR_. * Liu et al. (2018b) Meng Liu, Xiang Wang, Liqiang Nie, Qi Tian, Baoquan Chen, and Tat-Seng Chua. 2018b. Cross-modal moment localization in videos. In _ACM MM_. * Lu et al. (2019) Chujie Lu, Long Chen, Chilie Tan, Xiaolin Li, and Jun Xiao. 2019. Debug: A dense bottom-up grounding approach for natural language video localization. In _EMNLP_. * Mithun et al. (2019) Niluthpol Chowdhury Mithun, Sujoy Paul, and Amit K Roy-Chowdhury. 2019\. Weakly supervised video moment retrieval from text queries. In _CVPR_. * Otani et al. (2020) Mayu Otani, Yuta Nakashima, Esa Rahtu, and Janne Heikkilä. 2020. Uncovering Hidden Challenges in Query-Based Video Moment Retrieval. In _BMVC_. * Pennington et al. (2014) Jeffrey Pennington, Richard Socher, and Christopher D Manning. 2014. Glove: Global vectors for word representation. In _EMNLP_. * Regneri et al. (2013) Michaela Regneri, Marcus Rohrbach, Dominikus Wetzel, Stefan Thater, Bernt Schiele, and Manfred Pinkal. 2013. Grounding action descriptions in videos. _TACL_ (2013). * Shou et al. (2016) Zheng Shou, Dongang Wang, and Shih-Fu Chang. 2016. Temporal action localization in untrimmed videos via multi-stage cnns. In _CVPR_. * Sigurdsson et al. (2016) Gunnar A Sigurdsson, Gül Varol, Xiaolong Wang, Ali Farhadi, Ivan Laptev, and Abhinav Gupta. 2016. Hollywood in homes: Crowdsourcing data collection for activity understanding. In _ECCV_. * Song and Han (2018) Xiaomeng Song and Yahong Han. 2018. Val: Visual-attention action localizer. In _PCM_. * Song et al. (2020) Yijun Song, Jingwen Wang, Lin Ma, Zhou Yu, and Jun Yu. 2020. Weakly-supervised multi-level attentional reconstruction network for grounding textual queries in videos. In _arXiv_. * Tan et al. (2019) Reuben Tan, Huijuan Xu, Kate Saenko, and Bryan A Plummer. 2019\. wman: Weakly-supervised moment alignment network for text-based video segment retrieval. In _arXiv_. * Tran et al. (2015) Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, and Manohar Paluri. 2015. Learning spatiotemporal features with 3d convolutional networks. In _ICCV_. * Wang et al. (2016) Limin Wang, Yuanjun Xiong, Zhe Wang, Yu Qiao, Dahua Lin, Xiaoou Tang, and Luc Van Gool. 2016. Temporal segment networks: Towards good practices for deep action recognition. In _ECCV_. * Wang et al. (2019) Weining Wang, Yan Huang, and Liang Wang. 2019. Language-driven temporal activity localization: A semantic matching reinforcement learning model. In _CVPR_. * Wu et al. (2020) Jie Wu, Guanbin Li, Si Liu, and Liang Lin. 2020\. Tree-Structured Policy based Progressive Reinforcement Learning for Temporally Language Grounding in Video. In _AAAI_. * Xiao et al. (2021) Shaoning Xiao, Long Chen, Songyang Zhang, Wei Ji, Jian Shao, Lu Ye, and Jun Xiao. 2021. Boundary Proposal Network for Two-Stage Natural Language Video Localization. In _AAAI_. * Xu et al. (2019) Huijuan Xu, Kun He, Bryan A Plummer, Leonid Sigal, Stan Sclaroff, and Kate Saenko. 2019\. Multilevel language and vision integration for text-to-clip retrieval. In _AAAI_. * Yuan et al. (2019a) Yitian Yuan, Lin Ma, Jingwen Wang, Wei Liu, and Wenwu Zhu. 2019a. Semantic conditioned dynamic modulation for temporal sentence grounding in videos. In _NeurIPS_. * Yuan et al. (2019b) Yitian Yuan, Tao Mei, and Wenwu Zhu. 2019b. To find where you talk: Temporal sentence localization in video with attention based location regression. In _AAAI_. * Zeng et al. (2020) Runhao Zeng, Haoming Xu, Wenbing Huang, Peihao Chen, Mingkui Tan, and Chuang Gan. 2020\. Dense regression network for video grounding. In _CVPR_. * Zhang et al. (2019a) Da Zhang, Xiyang Dai, Xin Wang, Yuan-Fang Wang, and Larry S Davis. 2019a. Man: Moment alignment network for natural language moment retrieval via iterative graph adjustment. In _CVPR_. * Zhang et al. (2020) Songyang Zhang, Houwen Peng, Jianlong Fu, and Jiebo Luo. 2020\. Learning 2D Temporal Adjacent Networks forMoment Localization with Natural Language. In _AAAI_. * Zhang et al. (2019b) Zhu Zhang, Zhijie Lin, Zhou Zhao, and Zhenxin Xiao. 2019b. Cross-modal interaction networks for query-based moment retrieval in videos. In _SIGIR_.
Checking Robustness Between Weak Transactional Consistency ModelsThis work is supported in part by the European Research Council (ERC) under the European Union's Horizon 2020 research and innovation programme (grant agreement No 678177). Sidi Mohamed Beillahi() Ahmed Bouajjani Constantin Enea S.M. Beillahi, A. Bouajjani, and C. Enea. Université de Paris, IRIF, CNRS, Paris, France, Concurrent accesses to databases are typically encapsulated in transactions in order to enable isolation from other concurrent computations and resilience to failures. Modern databases provide transactions with various semantics corresponding to different trade-offs between consistency and availability. Since a weaker consistency model provides better performance, an important issue is investigating the weakest level of consistency needed by a given program (to satisfy its specification). As a way of dealing with this issue, we investigate the problem of checking whether a given program has the same set of behaviors when replacing a consistency model with a weaker one. This property known as robustness generally implies that any specification of the program is preserved when weakening the consistency. We focus on the robustness problem for consistency models which are weaker than standard serializability, namely, causal consistency, prefix consistency, and snapshot isolation. We show that checking robustness between these models is polynomial time reducible to a state reachability problem under serializability. We use this reduction to also derive a pragmatic proof technique based on Lipton's reduction theory that allows to prove programs robust. We have applied our techniques to several challenging applications drawn from the literature of distributed systems and databases. § INTRODUCTION Concurrent accesses to databases are typically encapsulated in transactions in order to enable isolation from other concurrent computations and resilience to failures. Modern databases provide transactions with various semantics corresponding to different tradeoffs between consistency and availability. The strongest consistency level is achieved with serializable transactions [41] whose outcome in concurrent executions is the same as if the transactions were executed atomically in some order. Since serializability () carries a significant penalty on availability, modern databases often provide weaker consistency models, e.g., causal consistency () [37], prefix consistency () [21, 24], and snapshot isolation () [11]. Causal consistency requires that if a transaction $\tr_1$ “affects” another transaction $\tr_2$, e.g., $\tr_1$ executes before $\tr_2$ in the same session or $\tr_2$ reads a value written by $\tr_1$, then the updates in these two transactions are observed by any other transaction in this order. Concurrent transactions, which are not causally related to each other, can be observed in different orders, leading to behaviors that are not possible under . Prefix consistency requires that there is a total commit order between all the transactions such that each transaction observes all the updates in a prefix of this sequence ( is stronger than ). Two transactions can observe the same prefix, which leads to behaviors that are not admitted by . Snapshot isolation further requires that two different transactions observe different prefixes if they both write to a common variable. Since a weaker consistency model provides better performance, an important issue is identifying the weakest level of consistency needed by a program (to satisfy its specification). One way to tackle this issue is checking whether a program $P$ designed under a consistency model $S$ has the same behaviors when run under a weaker consistency model $W$. This property of a program is generally known as robustness against substituting $S$ with $W$. It implies that any specification of $P$ is preserved when weakening the consistency model (from $S$ to $W$). Preserving any specification is convenient since specifications are rarely present in practice. The problem of checking robustness for a given program has been investigated in several recent works, but only when the stronger model ($S$) is , e.g., [9, 10, 18, 25, 12, 39], or sequential consistency in the non-transactional case, e.g. [35, 14, 28]. there is a large class of specifications that can be implemented even in the presence of “anomalies”, i.e., behaviors which are not admitted under (see [45] for a discussion). In this context, an important question is whether a certain implementation (program) is robust against substituting a weak consistency model, e.g., , with a weaker one, e.g., . In this paper, we consider the sequence of increasingly strong consistency models mentioned above, , , and , and investigate the problem of checking robustness for a given program against weakening the consistency model to one in this range. We study the asymptotic complexity of this problem and propose effective techniques for establishing robustness based on abstraction. There are two important cases to consider: robustness against substituting with and with , respectively. Robustness against substituting with can be obtained as the conjunction of these two cases. In the first case ( vs ), checking robustness for a program $P$ is reduced to a reachability (assertion checking) problem in a composition of $P$ under with a monitor that checks whether a behavior is an “anomaly”, i.e., admitted by $P$ under , but not under . This approach raises two non-trivial challenges: (1) defining a monitor for detecting vs anomalies that uses a minimal amount of auxiliary memory (to remember past events), and (2) determining the complexity of checking if the composition of $P$ with the monitor reaches a specific control location[We assume that the monitor goes to an error location when detecting an anomaly.] under the (weaker) model . Interestingly enough, we address these two challenges by studying the relationship between these two weak consistency models, and , and serializability. The construction of the monitor is based on the fact that the vs anomalies can be defined as roughly, the difference between the vs and vs anomalies (investigated in previous work [12]), and we show that the reachability problem under can be reduced to a reachability problem under . These results lead to a polynomial-time reduction of this robustness problem (for arbitrary programs) to a reachability problem under , which is important from a practical point of view since the semantics (as opposed to the or semantics) can be encoded easily in existing verification tools (using locks to guard the isolation of transactions). These results also enable a precise characterization of the complexity class of this problem. Checking robustness against substituting with is reduced to the problem of checking robustness against substituting with . The latter has been shown to be polynomial-time reducible to reachability under in [10]. This surprising result relies on the reduction from reachability to reachability mentioned above. This reduction shows that a given program $P$ reaches a certain control location under iff a transformed program $P'$, where essentially, each transaction is split in two parts, one part containing all the reads, and one part containing all the writes, reaches the same control location under . Since this reduction preserves the structure of the program, vs anomalies of a program $P$ correspond to vs anomalies of the transformed program $P'$. Beyond enabling these reductions, the characterization of classes of anomalies or the reduction from the semantics to the semantics are also important for a better understanding of these weak consistency models and the differences between them. We believe that these results can find applications beyond robustness checking, e.g., verifying conformance to given specifications. As a more pragmatic approach for establishing robustness, which avoids a non-reachability proof under , we have introduced a proof methodology that builds on Lipton's reduction theory [38] and the concept of commutativity dependency graph introduced in [9], which represents mover type dependencies between the transactions in a program. We give sufficient conditions for robustness in all the cases mentioned above, which characterize the commutativity dependency graph associated to a given program. We tested the applicability of these verification techniques on a benchmark containing seven challenging applications extracted from previous work [29, 33, 18]. These techniques are precise enough for proving or disproving the robustness of all these applications, for all combinations of the consistency models. keywords = assume, select, return, stepnumber=1,numberblanklines=false,mathescape=true Process 1 CreateEvent(v, e1, 3): [ Tickets[v][e1] := 3 ] [ r := $\sum\limits_{\mbox{e}}$Tickets[v][e] ] Process 2 CreateEvent(v, e2, 3): [ Tickets[v][e2] := 3 ] [ r := $\sum\limits_{\mbox{e}}$Tickets[v][e] ] [->,>=stealth',shorten >=1pt,auto,node distance=4cm, semithick, transform shape,every text node part/.style=align=left] [shape=rectangle ,draw=none,font=] at (-4,0) (m) CreateEvent(v,e1,3); [shape=rectangle ,draw=none,font=, label=right://r=3] at (-4.3,-2) (n) CountTickets(v); [shape=rectangle ,draw=none,font=, ] at (0,0) (p)CreateEvent(v,e2,3); [shape=rectangle ,draw=none,font=, label=right://r=3] at (0,-2) (l) CountTickets(v); [ every edge/.style=draw=black,very thick] [->] (m.212) edge[left] node $\hbo$ (n.119); [->] (m.230) edge[right,dashed] node $\po$ (n.80); [->] (n) edge[below] node $\hbo$ (p); [->] (p.240) edge[left] node $\hbo$ (l.119); [->] (p.280) edge[right,dashed] node $\po$ (l.80); [->] (l) edge[above] node $\hbo$ (m); A trace of FusionTicket. Process 1 Register(u, p1): [ r := RegisteredUsers[u] assume r == 0 RegisteredUsers[u] := 1 Password[u] := p1 ] Process 2 Register(u, p2): [ r := RegisteredUsers[u] assume r == 0 RegisteredUsers[u] := 1 Password[u] := p2 ] [->,>=stealth',shorten >=1pt,auto,node distance=4cm, semithick, transform shape,every text node part/.style=align=left] [shape=rectangle ,draw=none,font=] at (-3.5,0) (m) Register(u,p1); [shape=rectangle ,draw=none,font=] at (0,0) (p)Register(u,p2); [ every edge/.style=draw=black,very thick] [->] (m) edge[bend right,above] node $\hbo$ (p); [->] (p) edge[bend right,above] node $\hbo$ (m); A and trace of Twitter. Process 1 RegisterRd(u, p1): [ r := RegisteredUsers[u] assume r == 0 ] RegisterWr(u, p1): [ RegisteredUsers[u] := 1 Password[u] := p1 ] Process 2 RegisterRd(u, p2): [ r := RegisteredUsers[u] assume r == 0 ] RegisterWr(u, p2): [ RegisteredUsers[u] := 1 Password[u] := p2 ] Transformed Twitter. [->,>=stealth',shorten >=1pt,auto,node distance=4cm, semithick, transform shape,every text node part/.style=align=left] [shape=rectangle ,draw=none,font=] at (-4,0) (m) RegisterRd(u,p1); [shape=rectangle ,draw=none,font=] at (-4,-2) (n) RegisterWr(u,p1); [shape=rectangle ,draw=none,font=] at (0,0) (p)RegisterRd(u,p2); [shape=rectangle ,draw=none,font=] at (0,-2) (l)RegisterWr(u,p2); [ every edge/.style=draw=black,very thick] [->] (m) edge[above] node $\hbo$ (l); [->] (m.235) edge[left] node $\hbo$ (n.120); [->] (m.275) edge[right,dashed] node $\po$ (n.80); [->] (n.360) edge[above] node $\hbo$ (l.180); [->] (p) edge[below] node $\hbo$ (n); [->] (p.235) edge[left] node $\hbo$ (l.120); [->] (p.275) edge[right,dashed] node $\po$ (l.80); A and trace of transformed Twitter. Process 1 [ assume time < TIMEOUT Bets[1] := 2 ] Process 2 [ assume time < TIMEOUT Bets[2] := 3 ] Process 3 [Bets' := Bets n := Bets'.Length assume time > TIMEOUT n > 0 select i s.t. Bets'[i] $\neq$ $\bot$ return := Bets'[i] ] [->,>=stealth',shorten >=1pt,auto,node distance=4cm, semithick, transform shape,every text node part/.style=align=left] [shape=rectangle ,draw=none,font=] at (-6,0) (m) PlaceBet(1,2); [shape=rectangle ,draw=none,font=] at (-3,0) (p) PlaceBet(2,3); [shape=rectangle ,draw=none,font=, label=right:// return=2] at (0.5,0) (q)SettleBet(); [ every edge/.style=draw=black,very thick] [->] (m) edge[bend right,above] node $\hbo$ (q); [->] (q) edge[above] node $\hbo$ (p); A and trace of Betting. [shape=rectangle ,draw=none,font=] (A) at (0,0) [] PlaceBet(1,2); [shape=rectangle ,draw=none,font=] (B) at (2.5,0) [] SettleBet(); [shape=rectangle ,draw=none,font=] (C) at (5,0) [] PlaceBet(2,3); [ every edge/.style=draw=black,very thick] [->] (A) edge [bend left] node [above,font=] (B); [->] (B) edge[bend left] node [above,font=] (A); [->] (C) edge [bend right] node [above,font=] (B); [->] (B) edge[bend right] node [above,font=] (C); Commutativity dependency graph of Betting. Transactional programs and traces under different consistency models. § OVERVIEW We give an overview of the robustness problems investigated in this paper, discussing first the case vs. , and then vs . We end with an example that illustrates the robustness checking technique based on commutativity arguments. Robustness vs . We illustrate the robustness against substituting with using the FusionTicket and the Twitter programs in Figure <ref> and Figure <ref>, respectively. FusionTicket manages tickets for a number of events, each event being associated with a venue. Its state consists of a two-dimensional map that stores the number of tickets for an event in a given venue ($r$ is a local variable, and the assignment in is interpreted as a read of the shared state). The program has two processes and each process contains two transactions. The first transaction creates an event $\mbox{e}$ in a venue $\mbox{v}$ with a number of tickets $\mbox{n}$, and the second transaction computes the total number of tickets for all the events in a venue $\mbox{v}$. A possible candidate for a specification of this program is that the values computed in are monotonically increasing since each such value is computed after creating a new event. Twitter provides a transaction for registering a new user with a given username and password, which is executed by two parallel processes. Its state contains two maps that record whether a given username has been registered (0 and 1 stand for non-registered and registered, respectively) and the password for a given username. Each transaction first checks whether a given username is free (see the statement). The intended specification is that the user must be registered with the given password when the registration transaction succeeds. A program is robust against substituting with if its set of behaviors under the two models coincide. We model behaviors of a given program as traces, which record standard control-flow and data-flow dependencies between transactions, e.g., the order between transactions in the same session and whether a transaction reads the value written by another (read-from). The transitive closure of the union of all these dependency relations is called happens-before. Figure <ref> pictures a trace of FusionTicket where the concrete values which are read in a transaction are written under comments. In this trace, each process registers a different event but in the same venue and with the same number of tickets, and it ignores the event created by the other process when computing the sum of tickets in the venue. Figure <ref> pictures a trace of FusionTicket under , which is a witness that FusionTicket is not robust against substituting with . This trace is also a violation of the intended specification since the number of tickets is not increasing (the sum of tickets is $3$ in both processes). The happens-before dependencies (pictured with $\hbo$ labeled edges) include the program-order $\po$ (the order between transactions in the same process), and read-write dependencies, since an instance of $\mbox{CountTickets(v)}$ does not observe the value written by the $\mbox{CreateEvent}$ transaction in the other process (the latter overwrites some value that the former reads). This trace is allowed under because the transaction $\mbox{CreateEvent(v, e1, 3)}$ executes concurrently with the transaction $\mbox{CountTickets(v)}$ in the other process, and similarly for $\mbox{CreateEvent(v, e2, 3)}$. However, it is not allowed under since it is impossible to define a total commit order between $\mbox{CreateEvent(v, e1, 3)}$ and $\mbox{CreateEvent(v, e2, 3)}$ that justifies the reads of both $\mbox{CountTickets(v)}$ transactions (these reads should correspond to the updates in a prefix of this order). For instance, assuming that $\mbox{CreateEvent(v, e1, 3)}$ commits before $\mbox{CreateEvent(v, e2, 3)}$, $\mbox{CountTickets(v)}$ in the second process must observe the effect of $\mbox{CreateEvent(v, e1, 3)}$ as well since it observes the effect of $\mbox{CreateEvent(v, e2, 3)}$. However, this contradicts the fact that $\mbox{CountTickets(v)}$ computes the sum of tickets as being $3$. On the other hand, Twitter is robust against substituting with . For instance, Figure <ref> pictures a trace of Twitter under , where the in both transactions pass. In this trace, the transactions $\mbox{Register(u,p1)}$ and $\mbox{Register(u,p2)}$ execute concurrently and are unaware of each other's writes (they are not causally related). The $\hbo$ dependencies include write-write dependencies since both transactions write on the same location (we consider the transaction in Process 2 to be the last one writing to the map), and read-write dependencies since each transaction reads that is written by the other. This trace is also allowed under since the commit order can be defined such that $\mbox{Register(u,p1)}$ is ordered before $\mbox{Register(u,p2)}$, and then both transactions read from the initial state (the empty prefix). Note that this trace has a cyclic happens-before which means that it is not allowed under serializability. Checking robustness vs . We reduce the problem of checking robustness against substituting with to the robustness problem against substituting with (the latter reduces to a reachability problem under  [10]). This reduction relies on a syntactic program transformation that rewrites behaviors of a given program $P$ to $\serc{}$ behaviors of another program $P'$. The program $P'$ is obtained by splitting each transaction $\atr$ of $P$ into two transactions: the first transaction performs all the reads in $\atr$ and the second performs all the writes in $\atr$ (the two are related by program order). Figure <ref> shows this transformation applied on Twitter. The trace in Figure <ref> is a serializable execution of the transformed Twitter which is “observationally” equivalent to the trace in Figure <ref> of the original Twitter, i.e., each read of the shared state returns the same value and the writes on the shared state are applied in the same order (the acyclicity of the happens-before shows that this is a serializable trace). The transformed FusionTicket coincides with the original version because it contains no transaction that both reads and writes on the shared state. We show that behaviors and behaviors of the original and transformed program, respectively, are related by a bijection. In particular, we show that any vs. robustness violation of the original program manifests as a vs. robustness violation of the transformed program, and vice-versa. For instance, the trace of the original Twitter in Figure <ref> corresponds to the trace of the transformed Twitter in Figure <ref>, and the acyclicity of the latter (the fact that it is admitted by ) implies that the former is admitted by the original Twitter under . On the other hand, the trace in Figure <ref> is also a of the transformed FusionTicket and its cyclicity implies that it is not admitted by FusionTicket under , and thus, it represents a robustness violation. Robustness vs . We illustrate the robustness against substituting with using Twitter and the Betting program in Figure <ref>. Twitter is not robust against substituting with , the trace in Figure <ref> being a witness violation. This trace is also a violation of the intended specification since one of the users registers a password that is overwritten in a concurrent transaction. This trace is not possible under because $\mbox{Register(u,p1)}$ and $\mbox{Register(u,p2)}$ observe the same prefix of the commit order (i.e., an empty prefix), but they write to a common memory location $\mbox{Password[u]}$ which is not allowed under . On the other hand, the Betting program in Figure <ref>, which manages a set of bets, is robust against substituting with . The first two processes execute one transaction that places a bet of a value $\mbox{v}$ with a unique bet identifier $\mbox{id}$, assuming that the bet expiration time is not yet reached (bets are recorded in the map ). The third process contains a single transaction that settles the betting assuming that the bet expiration time was reached and at least one bet has been placed. This transaction starts by taking a snapshot of the map into a local variable , and then selects a random non-null value (different from $\bot$) in the map to correspond to the winning bet. The intended specification of this program is that the winning bet corresponds to a genuine bet that was placed. Figure <ref> pictures a trace of Betting where $\mbox{SettleBet}$ observes only the bet of the first process $\mbox{PlaceBet(1,2)}$. The $\hbo$ dependency towards the second process denotes a read-write dependency ($\mbox{SettleBet}$ reads a cell of the map which is overwritten by the second process). This trace is allowed under because no two transactions write to the same location. Checking robustness vs . We reduce robustness against substituting with to a reachability problem under . This reduction is based on a characterization of happens-before cycles[Traces with an acyclic happens-before are not robustness violations because they are admitted under serializability, which implies that they are admitted under the weaker model as well.] that are possible under but not , and the transformation described above that allows to simulate the semantics of a program on top of . The former is used to define an instrumentation (monitor) for the transformed program that reaches an error state iff the original program is not robust. Therefore, we show that the happens-before cycles in traces that are not admitted by must contain a transaction that (1) overwrites a value written by another transaction in the cycle and (2) reads a value overwritten by another transaction in the cycle. For instance, the trace of Twitter in Figure <ref> is not allowed under because $\mbox{Register(u,p2)}$ overwrites a value written by $\mbox{Register(u,p1)}$ (the password) and reads a value overwritten by $\mbox{Register(u,p1)}$ (checking whether the username $u$ is registered). The trace of Betting in Figure <ref> is allowed under because its happens-before is acyclic. Checking robustness using commutativity arguments. Based on the reductions above, we propose an approximated method for proving robustness based on the concept of mover in Lipton's reduction theory [38]. A transaction is a left (resp., right) mover if it commutes to the left (resp., right) of another transaction (by a different process) while preserving the computation. We use the notion of mover to characterize the data-flow dependencies in the happens-before. Roughly, there exists a data-flow dependency between two transactions in some execution if one doesn't commute to the left/right of the other one. We define a commutativity dependency graph which summarizes the happens-before dependencies in all executions of a transformed program (obtained by splitting the transactions of the original program as explained above), and derive a proof method for robustness which inspects paths in this graph. Two transactions $\atr_1$ and $\atr_2$ are linked by a directed edge iff $\atr_1$ cannot move to the right of $\atr_2$ (or $\atr_2$ cannot move to the left of $\atr_1$), or if they are related by the program order. Moreover, two transactions $\atr_1$ and $\atr_2$ are linked by an undirected edge iff they are the result of splitting the same transaction. A program is robust against substituting with if roughly, its commutativity dependency graph does not contain a simple cycle of directed edges with two distinct transactions $\atr_1$ and $\atr_2$, such that $\atr_1$ does not commute left because of another transaction $\atr_3$ in the cycle that reads a variable that $\atr_1$ writes to, and $\atr_2$ does not commute right because of another transaction $\atr_4$ in the cycle ($\atr_3$ and $\atr_4$ can coincide) that writes to a variable that $\atr_2$ either reads from or writes to[The transactions $\atr_1$, $\atr_2$, $\atr_3$, and $\atr_4$ correspond to $\atr_1$, $\atr_i$, $\atr_n$, and $\atr_{i+1}$, respectively, in Theorem <ref>.]. For instance, Figure <ref> shows the commutativity dependency graph of the transformed Betting program, which coincides with the original Betting because $\mbox{PlaceBet(1,2)}$ and $\mbox{PlaceBet(2,3)}$ are write-only transactions and $\mbox{SettleBet()}$ is a read-only transaction. Both simple cycles in Figure <ref> contain just two transactions and therefore do not meet the criterion above which requires at least 3 transactions. Therefore, Betting is robust against substituting with . A program is robust against substituting with , if roughly, its commutativity dependency graph does not contain a simple cycle with two successive transactions $\atr_1$ and $\atr_2$ that are linked by an undirected edge, such that $\atr_1$ does not commute left because of another transaction $\atr_3$ in the cycle that writes to a variable that $\atr_1$ writes to, and $\atr_2$ does not commute right because of another transaction $\atr_4$ in the cycle ($\atr_3$ and $\atr_4$ can coincide) that writes to a variable that $\atr_2$ reads from[The transactions $\atr_1$, $\atr_2$, $\atr_3$, and $\atr_4$ correspond to $\atr_1$, $\atr_2$, $\atr_n$, and $\atr_3$, respectively, in Theorem <ref>.]. Betting is also robust against substituting with for the same reason (simple cycles of size 2). § CONSISTENCY MODELS <prog> ::= program <process>$^{*}$ <process> ::= process <pid> regs <reg>$^{*}$ <txn>$^{*}$ <txn> ::= begin <read>$^{*}$ <test>$^{*}$ <write>$^{*}$ commit <read> ::= <label>":" <reg> ":=" <var>";" goto <label>";" <test> ::= <label>":" assume <bexpr>";" goto <label>";" <write> ::= <label>":" <var> ":=" <reg-expr>";" goto <label>";" The syntax of our programming language. $a^{*}$ indicates zero or more occurrences of $a$. $\langle pid\rangle$, $\langle reg\rangle$, $\langle label \rangle$, and $\langle var\rangle$ represent a process identifier, a register, a label, and a shared variable, respectively. $\langle reg\text{-}expr \rangle$ is an expression over registers while $\langle bexpr \rangle$ is a Boolean expression over registers, or the non-deterministic choice $*$. We present our results in the context of the simple programming language, defined in Figure <ref>, where a program is a parallel composition of processes distinguished using a set of identifiers $\mathbb{P}$. A process is a sequence of transactions and each transaction is a sequence of labeled instructions. A transaction starts with a begin instruction and finishes with a commit instruction. Instructions include assignments to a process-local register from a set $\mathbb{R}$ or to a shared variable from a set $\mathbb{V}$, or an assume. The assignments use values from a data domain $\mathbb{D}$. An assignment to a register $\langle reg\rangle := \langle var\rangle$ is called a read of the shared-variable $\langle var\rangle$ and an assignment to a shared variable $\langle var\rangle := \langle reg\rangle$ is called a write to the shared-variable $\langle var\rangle$. The assume $\langle bexpr\rangle$ blocks the process if the Boolean expression $\langle bexpr\rangle$ over registers is false. It can be used to model conditionals. The goto statement transfers the control to the program location (instruction) specified by a given label. Since multiple instructions can have the same label, goto statements can be used to mimic imperative constructs like loops and conditionals inside transactions. We assume w.l.o.g. that every transaction is written as a sequence of reads or assume statements followed by a sequence of writes (a single goto statement from the sequence of read/assume instructions transfers the control to the sequence of writes). In the context of the consistency models we study in this paper, every program can be equivalently rewritten as a set of transactions of this form. To simplify the technical exposition, programs contain a bounded number of processes and each process executes a bounded number of transactions. A transaction may execute an unbounded number of instructions but these instructions concern a bounded number of variables, which makes it impossible to model SQL (select/update) queries that may access tables with a statically unknown number of rows. Our results can be extended beyond these restrictions as explained in Remark <ref> and Remark <ref>. We describe the semantics of a program under four consistency models, i.e., causal consistency[We consider a variation known as causal convergence [19, 15]] (), prefix consistency (), snapshot isolation (), and serializability (). In the semantics of a program under , shared variables are replicated across each process, each process maintaining its own local valuation of these variables. During the execution of a transaction in a process, its writes are stored in a transaction log that can be accessed only by the process executing the transaction and that is broadcasted to all the other processes at the end of the transaction. To read a shared variable $\anaddr$, a process $\apr$ first accesses its transaction log and takes the last written value on $\anaddr$, if any, and then its own valuation of the shared variable, if $\anaddr$ was not written during the current transaction. Transaction logs are delivered to every process in an order consistent with the causal relation between transactions, i.e., the transitive closure of the union of the program order (the order in which transactions are executed by a process), and the read-from relation (a transaction $\atr_1$ reads-from a transaction $\atr_2$ iff $\atr_1$ reads a value that was written by $\atr_2$). When a process receives a transaction log, it immediately applies it on its shared-variable valuation. In the semantics of a program under and , shared variables are stored in a central memory and each process keeps a local valuation of these variables. When a process starts a new transaction, it fetches a consistent snapshot of the shared variables from the central memory and stores it in its local valuation of these variables. During the execution of a transaction in a process, writes to shared variables are stored in the local valuation of these variables, and in a transaction log. To read a shared variable, a process takes its own valuation of the shared variable. A process commits a transaction by applying the updates in the transaction log on the central memory in an atomic way (to make them visible to all processes). Under , when a process applies the writes in a transaction log on the central memory, it must ensure that there were no concurrent writes that occurred after the last fetch from the central memory to a shared variable that was written during the current transaction. Otherwise, the transaction is aborted and its effects discarded. In the semantics of a program under , we adopt a simple operational model where we keep a single shared-variable valuation in a central memory (accessed by all processes) with the standard interpretation of read and write statements. Transactions execute serially, one after another. We use a standard model of executions of a program called trace. A trace represents the order between transactions in the same process, and the data-flow in an execution using standard happens-before relations between transactions. We assume that each transaction in a program is identified uniquely using a transaction identifier from a set $\mathbb{T}$. Also, $\amap: \mathbb{T} \rightarrow 2^{\mathbb{S}}$ is a mapping that associates each transaction in $\mathbb{T}$ with a sequence of read and write events from the set \begin{align*} \mathbb{S} = \{\readact(\atr,\anaddr,\aval), \writeact(\atr,\anaddr,\aval): \atr\in \mathbb{T}, \anaddr\in \mathbb{V}, \aval\in \mathbb{D}\} \end{align*} where $\readact(\atr,\anaddr,\aval)$ is a read of $\anaddr$ returning $\aval$, and $\writeact(\atr,\anaddr,\aval)$ is a write of $\aval$ to $\anaddr$. A trace is a tuple $\atrace = (\rho,\amap,\tor,\po,\rfo,\sto,\cfo)$ where $\rho\subseteq \mathbb{T}$ is a set of transaction identifiers, and * $\tor$ is a mapping giving the order between events in each transaction, i.e., it associates each transaction $\atr$ in $\rho$ with a total order $\tor(\atr)$ on $\amap(\atr) \times \amap(\atr)$. * $\po$ is the program order relation, a strict partial order on $\rho \times \rho$ that orders every two transactions issued by the same process. * $\rfo$ is the read-from relation between distinct transactions $(\atr1, \atr2) \in \rho \times \rho$ representing the fact that $\atr2$ reads a value written by $\atr1$. * $\sto$ is the store order relation on $\rho \times \rho$ between distinct transactions that write to the same shared variable. * $\cfo$ is the conflict order relation between distinct transactions, defined by $\cfo = \rfo^{-1};\sto$ ($;$ denotes the sequential composition of two relations). For simplicity, for a trace $\atrace = (\rho,\amap,\tor,\po,\rfo,\sto,\cfo)$, we write $t\in \atrace$ instead of $t\in\rho$. We also assume that each trace contains a fictitious transaction that writes the initial values of all shared variables, and which is ordered before any other transaction in program order. Also, $\tracesconf_{\textsf{X}}(\aprog)$ is the set of traces representing executions of program $\aprog$ under a consistency model $\textsf{X}$. For each $\textsf{X}\in \{\ccc{},\pcc{},\sic{},\serc{}\}$, the set of traces $\tracesconf_{\textsf{X}}(\aprog)$ can be described using the set of properties in Table <ref>. A trace $\atrace$ is possible under causal consistency iff there exist two relations $\viso$ a partial order (causal order) and $\arbo$ a total order (arbitration order) that includes $\viso$, such that the properties $\axpoco$, $\axcoarb$, and $\axretval$ hold [26, 15]. $\axpoco$ guarantees that the program order and the read-from relation are included in the causal order, and $\axcoarb$ guarantees that the causal order and the store order are included in the arbitration order. $\axretval$ guarantees that a read returns the value written by the last write in the last transaction that contains a write to the same variable and that is ordered by $\viso$ before the read's transaction. We use $\axcc$ to denote the conjunction of these three properties. A trace $\atrace$ is possible under prefix consistency iff there exist a causal order $\viso$ and an arbitration order $\arbo$ such that $\axcc$ holds and the property $\axprefix$ holds as well [26]. $\axprefix$ guarantees that every transaction observes a prefix of transactions that are ordered by $\arbo$ before it. We use $\axpc$ to denote the conjunction of $\axcc$ and $\axprefix$. A trace $\atrace$ is possible under snapshot isolation iff there exist a causal order $\viso$ and an arbitration order $\arbo$ such that $\axpc$ holds and the property $\axconflict$ holds [26]. $\axconflict$ guarantees that if two transactions write to the same variable then one of them must observe the other. We use $\axsi$ to denote the conjunction of $\axpc$ and $\axconflict$. A trace $\atrace$ is serializable iff there exist a causal order $\viso$ and an arbitration order $\arbo$ such that the property $\axser$ holds which implies that the two relations $\viso$ and $\arbo$ coincide. Note that for any given program $\aprog$, $\tracesconf_{\serc{}}(\aprog)\subseteq \tracesconf_{\sic{}}(\aprog)\subseteq \tracesconf_{\pcc{}}(\aprog)\subseteq \tracesconf_{\ccc{}}(\aprog)$. Also, the four consistency models we consider disallow anomalies such as dirty and phantom reads. $\axpoco$ $\viso_{0}^{+} \subseteq \viso$ $\axcoarb$ $\arbo_{0}^+ \subseteq \arbo$ $\axcc$ $\axretval \wedge \axpoco \wedge \axcoarb$ $\axprefix$ $\arbo ; \viso \subseteq \viso$ $\axpc$ $ \axprefix \wedge \axcc$ $\axconflict$ $\sto \subseteq \viso$ $\axsi$ $\axconflict \wedge \axpc$ $\axser$ $\axretval \wedge \axpoco \wedge \axcoarb \wedge \viso = \arbo $ $\viso_{0} = \po \cup \rfo$ and $\arbo_{0} = \po \cup \rfo \cup \sto$ $\axretval$ = $\forall\ t\in \atrace.\ \forall\ \readact(\atr,\anaddr,\aval) \in \amap(\atr)$ we have that * there exist a transaction $\atr_0 = Max_{\arbo}(\{\atr' \in \atrace\ |\ (\atr',\atr) \in \viso \wedge \exists\ \writeact(\atr',\anaddr,\cdot)\in\amap(\atr')\})$ and an event $\writeact(\atr_0,\anaddr,\aval) = Max_{\tor(\atr_0)}(\{\writeact(\atr_0,\anaddr,\cdot) \in \amap(\atr_0)\})$. Declarative definitions of consistency models. For an order relation $\leq$, $a = Max_{\leq}(A)$ iff $a \in A \wedge \forall\ b \in A.\ b \leq a$. For a given trace $\atrace=(\rho,\amap,\tor, \po, \rfo, \sto, \cfo)$, the happens before order is the transitive closure of the union of all the relations in the trace, i.e., $\hbo = (\po \cup \rfo \cup \sto \cup \cfo)^{+}$. A classic result states that a trace $\atrace$ is serializable iff $\hbo$ is acyclic [2, 46]. Note that $\hbo$ is acyclic implies that $\sto$ is a total order between transactions that write to the same variable, and $(\po \cup \rfo)^{+}$ and $(\po \cup \rfo \cup \sto)^{+}$ are acyclic. §.§ Robustness In this work, we investigate the problem of checking whether a program $\aprog$ under a semantics $\textsf{Y} \in \{\pcc{},\ \sic{}\}$ produces the same set of traces as under a weaker semantics $\textsf{X} \in \{\ccc{},\ \pcc{}\}$. When this holds, we say that $\aprog$ is robust against $\textsf{X}$ relative to $\textsf{Y}$. A program $\aprog$ is called robust against a semantics $\textsf{X} \in \{\ccc{},\ \pcc{},\ \sic{}\}$ relative to a semantics $\textsf{Y} \in \{\pcc{},\ \sic{},\ \serc{}\}$ such that $\textsf{Y}$ is stronger than $\textsf{X}$ iff $\tracesconf_{\textsf{X}}(\aprog)=\tracesconf_{\textsf{Y}}(\aprog)$. If $\aprog$ is not robust against $\textsf{X}$ relative to $\textsf{Y}$ then there must exist a trace $\atrace \in \tracesconf_{\textsf{X}}(\aprog) \setminus \tracesconf_{\textsf{Y}}(\aprog)$. We say that $\atrace$ is a robustness violation trace. [->,>=stealth',shorten >=1pt,auto,node distance=4cm, semithick, transform shape,every text node part/.style=align=left] [shape=rectangle ,draw=none,font=, label=left:$\atr_1$] at (-2.5,0) (m) $[x := 1]$; [shape=rectangle ,draw=none,font=, label=left:$\atr_2$] at (-2.5,-2) (n) $[r1 := y]\ //0$; [shape=rectangle ,draw=none,font=, label=right:$\atr_3$] at (0,0) (p)$[y := 1]$; [shape=rectangle ,draw=none,font=, label=right:$\atr_4$] at (0,-2) (l) $[r2 := x]\ //0$; [ every edge/.style=draw=black,very thick] [->] (m) edge[left] node $\po$ (n); [->] (n) edge[left] node $\cfo$ (p); [->] (p) edge[right] node $\po$ (l); [->] (l) edge[right] node $\cfo$ (m); Store Buffering ($\mathsf{SB}$). [->,>=stealth',shorten >=1pt,auto,node distance=4cm, semithick, transform shape,every text node part/.style=align=left] [shape=rectangle ,draw=none,font=, label=left:$\atr_1$] at (-3,0) (m) $[r1 := x\ \ //0$ $\ x := r1 + 1]$; [shape=rectangle ,draw=none,font=, label=right:$\atr_2$] at (0,0) (p)$[r2 := x\ \ //0$ $\ x := r2 + 1]$; [ every edge/.style=draw=black,very thick] [->] (m) edge[bend right,above] node $\sto$ (p); [->] (p) edge[bend right,above] node $\cfo$ (m); Lost Update ($\mathsf{LU}$). [->,>=stealth',shorten >=1pt,auto,node distance=4cm, semithick, transform shape,every text node part/.style=align=left] [shape=rectangle ,draw=none,font=, label=left:$\atr_1$] at (-3,0) (m) $[r1 := x\ \ //0$ $\ y := 1]$; [shape=rectangle ,draw=none,font=, label=right:$\atr_2$] at (0,0) (p)$[r2 := y\ \ //0$ $\ x := 1]$; [ every edge/.style=draw=black,very thick] [->] (m) edge[bend right,above] node $\cfo$ (p); [->] (p) edge[bend right,above] node $\cfo$ (m); Write Skew (WS). [->,>=stealth',shorten >=1pt,auto,node distance=4cm, semithick, transform shape,every text node part/.style=align=left] [shape=rectangle ,draw=none,font=, label=left:$\atr_1$] at (-2.5,0) (m) $[x := 1]$; [shape=rectangle ,draw=none,font=, label=left:$\atr_2$] at (-2.5,-2) (n) $[y := 1]$; [shape=rectangle ,draw=none,font=, label=right:$\atr_3$] at (0,0) (p)$[r1 := y]\ //1$; [shape=rectangle ,draw=none,font=, label=right:$\atr_4$] at (0,-2) (l) $[r2 := x]\ //1$; [ every edge/.style=draw=black,very thick] [->] (m) edge[left] node $\po$ (n); [->] (n) edge[left] node $\rfo$ (p); [->] (p) edge[right] node $\po$ (l); [->] (m) edge[right] node $\rfo$ (l); Message Passing (MP). Litmus programs We illustrate the notion of robustness on the programs in Figure <ref>, which are commonly used in the literature. In all programs, transactions of the same process are aligned vertically and ordered from top to bottom. Each read instruction is commented with the value it reads in some execution. The store buffering ($\mathsf{SB}$) program in Figure <ref> contains four transactions that are issued by two distinct processes. We emphasize an execution where $\atr_2$ reads $0$ from $y$ and $\atr_4$ reads $0$ from $x$. This execution is allowed under since the two writes by $\atr_1$ and $\atr_3$ are not causally dependent. Thus, $\atr_2$ and $\atr_4$ are executed without seeing the writes from $\atr_3$ and $\atr_1$, respectively. However, this execution is not feasible under (which implies that it is not feasible under both and ). In particular, we cannot have neither $(\atr_1,\atr_3) \in \arbo$ nor $(\atr_3,\atr_1) \in \arbo$ which contradicts the fact that $\arbo$ is total order. For example, if $(\atr_1,\atr_3) \in \arbo$, then $(\atr_1,\atr_4) \in \viso$ (since $\arbo;\viso \subset \viso$) which contradicts the fact that $\atr_4$ does not see $\atr_1$. Similarly, $(\atr_3,\atr_1) \in \arbo$ implies that $(\atr_3,\atr_2) \in \viso$ which contradicts the fact that $\atr_2$ does not see $\atr_3$. Thus, $\mathsf{SB}$ is not robust against relative to . The lost update ($\mathsf{LU}$) program in Figure <ref> has two transactions that are issued by two distinct processes. We highlight an execution where both transactions read $0$ from $x$. This execution is allowed under since both transactions are not causally dependent and can be executed in parallel by the two processes. However, it is not allowed under since both transactions write to a common variable (i.e., $x$). Thus, they cannot be executed in parallel and one of them must see the write of the other. Thus, $\mathsf{SB}$ is not robust against relative to . The write skew ($\mathsf{WS}$) program in Figure <ref> has two transactions that are issued by two distinct processes. We highlight an execution where $\atr_1$ reads $0$ from $x$ and $\atr_2$ reads $0$ from $y$. This execution is allowed under since both transactions are not causally dependent, do not write to a common variable, and can be executed in parallel by the two processes. However, this execution is not allowed under since one of the two transactions must see the write of the other. Thus, $\mathsf{WS}$ is not robust against relative to . The message passing ($\mathsf{MP}$) program in Figure <ref> has four transactions issued by two processes. Because $\atr_1$ and $\atr_2$ are causally dependent, under any semantics $\textsf{X} \in \{\ccc{},\ \pcc{},\ \sic{},\ \serc{}\}$ we only have three possible executions of $\mathsf{MP}$, which correspond to either $\atr_3$ and $\atr_4$ not observing the writes of $\atr_1$ and $\atr_2$, or $\atr_3$ and $\atr_4$ observe the writes of both $\atr_1$ and $\atr_2$, or $\atr_4$ observes the write of $\atr_1$ (we highlight the values read in the second case in Figure <ref>). Therefore, the executions of this program under the four consistency models coincide. Thus, $\mathsf{MP}$ is robust against relative to any other model. § ROBUSTNESS AGAINST RELATIVE TO We show that checking robustness against relative to can be reduced to checking robustness against relative to . The crux of this reduction is a program transformation that allows to simulate the semantics of a program $\aprog$ using the semantics of a program $\aprog_\pcinstr$. Checking robustness against relative to can be reduced in polynomial time to reachability under  [10]. Given a program $\aprog$ with a set of transactions $\trsaprog{\aprog}$, we define a program $\aprog_\pcinstr$ that every transaction $\atr\in \trsaprog{\aprog}$ is split into a transaction $\atrrd{\atr}$ that contains all the read/assume statements in $\atr$ (in the same order) and another transaction $\atrwr{\atr}$ that contains all the write statements in $\atr$ (in the same order). In the following, we establish the following result: A program $\aprog$ is robust against relative to iff $\aprog_\pcinstr$ is robust against relative to . Intuitively, under , processes can execute concurrent transactions that fetch the same consistent snapshot of the shared variables from the central memory and subsequently commit their writes. Decoupling the read part of a transaction from the write part allows to simulate such behaviors even under . The proof of this theorem relies on several intermediate results concerning the relationship between traces of $\aprog$ and $\aprog_\pcinstr$. Let $\atrace= (\rho, \po, \rfo, \sto, \cfo) \in \tracesconf_{\textsf{X}}(\aprog)$ be a trace of a program $\aprog$ under a semantics $\textsf{X}$. We define the trace $\atrace_\pcinstr= (\rho_\pcinstr, \po_\pcinstr, \rfo_\pcinstr, \sto_\pcinstr, \cfo_\pcinstr)$ where every transaction $\atr \in \atrace$ is split into two transactions $\atrrd{\atr}\in \atrace_\pcinstr$ and $\atrwr{\atr} \in \atrace_\pcinstr$, and the dependency relations are straightforward adaptations, i.e., * $\po_\pcinstr$ is the smallest transitive relation that includes $(\atrrd{\atr},\atrwr{\atr})$ for every $\atr$, and $(\atrwr{\atr},\atrrd{\atr'})$ if $(\atr,\atr')\in \po$, $(\atrwr{\atr'},\atrrd{\atr}) \in \rfo_\pcinstr$, $(\atrwr{\atr'},\atrwr{\atr}) \in \sto_\pcinstr$, and $(\atrrd{\atr'},\atrwr{\atr}) \in \cfo_\pcinstr$ if $(\atr',\atr) \in \rfo$, $(\atr',\atr) \in \sto$, and $(\atr',\atr) \in \cfo$, respectively. [->,>=stealth',shorten >=1pt,auto,node distance=4cm, semithick, transform shape,every text node part/.style=align=left] [shape=rectangle ,draw=none,font=, label=left:$\atrrd{\atr_1}$] at (-3,0) (m) $[r1 = x]\ //0$; [shape=rectangle ,draw=none,font=, label=left:$\atrwr{\atr_1}$] at (-3,-1.5) (m1) $[x = r1 + 1]$; [shape=rectangle ,draw=none,font=, label=right:$\atrrd{\atr_2}$] at (0,0) (p)$[r2 = x]\ //0$; [shape=rectangle ,draw=none,font=, label=right:$\atrwr{\atr_2}$] at (0,-1.5) (p1)$[x = r2 + 1]$; [ every edge/.style=draw=black,very thick] [->] (m1) edge[below] node $\sto$ (p1); [->] (m) edge[left] node $\po$ (m1); [->] (m) edge[below] node $\cfo$ (p1); [->] (p) edge[above] node $\cfo$ (m1); [->] (p) edge[left] node $\po$ (p1); A trace of the transformed LU program ($\mathsf{LU}_{\pcinstr}$). For instance, Figure <ref> pictures the trace $\atrace_\pcinstr$ for the $\mathsf{LU}$ trace $\atrace$ given in Figure <ref>. For traces $\atrace$ of programs that contain singleton transactions, e.g., $\mathsf{SB}$ in Figure <ref>, $\atrace_\pcinstr$ coincides with $\atrace$. Conversely, for a given trace $\atrace_\pcinstr= (\rho_\pcinstr, \po_\pcinstr, \rfo_\pcinstr, \sto_\pcinstr, \cfo_\pcinstr) \in \tracesconf_{\textsf{X}}(\aprog_\pcinstr)$ of a program $\aprog_\pcinstr$ under a semantics $\textsf{X}$, we define the trace $\atrace= (\rho, \po, \rfo, \sto, \cfo)$ where every two components $\atrrd{\atr}$ and $\atrwr{\atr}$ are merged into a transaction $\atr \in \atrace$. The dependency relations are defined in a straightforward way, e.g., if $(\atrwr{\atr'},\atrwr{\atr}) \in \sto_\pcinstr$ then $(\atr',\atr) \in \sto$. The following lemma shows that for any semantics $\textsf{X} \in \{\ccc,\ \pcc{},\ \sic{}\}$, if $\atrace \in \tracesconf_{\textsf{X}}(\aprog)$ for a program $\aprog$, then $\atrace_\pcinstr$ is a valid trace of $\aprog_\pcinstr$ under $\textsf{X}$, i.e., $\atrace_\pcinstr \in \tracesconf_{\textsf{X}}(\aprog_\pcinstr)$. Intuitively, this lemma shows that splitting transactions in a trace and defining dependency relations appropriately cannot introduce cycles in these relations and preserves the validity of the different consistency axioms. The proof of this lemma relies on constructing a causal order $\viso_\pcinstr$ and an arbitration order $\arbo_\pcinstr$ for the trace $\atrace_\pcinstr$ starting from the analogous relations in $\atrace$. In the case of $\ccc$, these are the smallest transitive relations such that: * $\po_\pcinstr\subseteq \viso_\pcinstr\subseteq \arbo_\pcinstr$, and * if $(\atr_{1},\atr_{2}) \in \viso$ then $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \viso_\pcinstr$, and if $(\atr_{1},\atr_{2}) \in \arbo$ then $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \arbo_\pcinstr$. For and , $\viso_\pcinstr$ must additionally satisfy: if $(\atr_{1},\atr_{2}) \in \arbo$, then $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \viso_\pcinstr$. This is required in order to satisfy the axiom $\axprefix$, i.e., $\arbo_\pcinstr;\viso_\pcinstr \subset \viso_\pcinstr$, when $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \arbo_\pcinstr$ and $(\atrrd{\atr_{2}},\atrwr{\atr_{2}}) \in \viso_\pcinstr$. This construction ensures that $\viso_\pcinstr$ is a partial order and $\arbo_\pcinstr$ is a total order because $\viso$ is a partial order and $\arbo$ is a total order. Also, based on the above rules, we have that: if $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \viso_\pcinstr$ then $(\atr_{1},\atr_{2}) \in \viso$, and similarly, if $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \arbo_\pcinstr$ then $(\atr_{1},\atr_{2}) \in \arbo$. If $\atrace \in \tracesconf_{\textsf{X}}(\aprog)$, then $\atrace_\pcinstr \in \tracesconf_{\textsf{X}}(\aprog_\pcinstr)$. We start with the case $\textsf{X} = \ccc$. We first show that $\atrace_\pcinstr$ satisfies $\axpoco$ and $\axcoarb$. For $\axpoco$, let $\atr_{1}' \in \{\atrrd{\atr_{1}},\atrwr{\atr_{1}}\}$ and $\atr_{2}' \in \{\atrrd{\atr_{2}},\atrwr{\atr_{2}}\}$, such that $(\atr_{1}',\atr_{2}') \in (\po_\pcinstr\cup\rfo_\pcinstr)^{+}$. By the definition of $\viso_\pcinstr$, we have that either $(\atr_{1}'=\atrrd{\atr_{1}},\atr_{2}'=\atrwr{\atr_{2}}) \in \po_\pcinstr$ and $\atr_1 = \atr_2$ or $(\atr_{1},\atr_2) \in (\po\cup\rfo)^{+}$, which implies that $(\atr_{1},\atr_2) \in \viso$. In both cases we obtain that $(\atr_{1}',\atr_{2}') \in \viso_\pcinstr$. The axiom $\axpoco$ can be proved in a similar way. Next, we show that $\atrace_\pcinstr$ satisfies the property $\axretval$. Let $\atr$ be a transaction in $\atrace$ that contains a read event $\readact(\atr,\anaddr,\aval)$. Let $\atr_0$ be the transaction in $\atrace$ such that $$\atr_0 = Max_{\arbo}(\{\atr' \in \atrace\ |\ (\atr',\atr) \in \viso \wedge \exists\ \writeact(\atr',\anaddr,\cdot)\in\amap(\atr')\}).$$ The read value $\aval$ must have been written by $\atr_0$ since $\atrace$ satisfies $\axretval$. Thus, the read $\readact(\atr,\anaddr,\aval)$ in $\atrrd{\atr}$ of $\atrace_\pcinstr$ must return the value written by $\atrwr{\atr_0}$. From the definitions of $\viso_\pcinstr$ and $\arbo_\pcinstr$, we get $$\atrwr{\atr_1}\in\{\atrwr{\atr'} \in \atrace_\pcinstr\ |\ (\atrwr{\atr'},\atrrd{\atr}) \in \viso_\pcinstr \wedge \exists\ \writeact(\atrwr{\atr'},\anaddr,\cdot)\in\amap(\atrwr{\atr'})\}$$ $$\atr_1 \in \{\atr' \in \atrace\ |\ (\atr',\atr) \in \viso \wedge \exists\ \writeact(\atr',\anaddr,\cdot)\in\amap(\atr')\}$$ because $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \viso_\pcinstr$ implies $(\atr_{1},\atr_{2}) \in \viso$. Since $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \arbo_\pcinstr$ implies $(\atr_{1},\atr_{2}) \in \arbo$, we also obtain that $$\atrwr{\atr_0} = Max_{\arbo_\pcinstr}(\{\atrwr{\atr'} \in \atrace_\pcinstr\ |\ (\atrwr{\atr'},\atrrd{\atr}) \in \viso_\pcinstr \wedge \exists\ \writeact(\atrwr{\atr'},\anaddr,\cdot)\in\amap(\atrwr{\atr'})\})$$ and since the read $\readact(\atr,\anaddr,\aval)$ in $\atrrd{\atr}$ of $\atrace_\pcinstr$ returns the value written by $\atrwr{\atr_0}$, $\atrace_\pcinstr$ satisfies $\axretval$. For the case $\textsf{X} = \pcc$, we show that $\atrace_\pcinstr$ satisfies the property $\axprefix$ (the other axioms are proved as in the case of $\ccc$). Suppose we have $(\atr_{1}',\atr_{2}') \in \arbo_\pcinstr$ and $(\atr_{2}',\atr_{3}') \in \viso_\pcinstr$ where $\atr_{1}' \in \{\atrrd{\atr_{1}},\atrwr{\atr_{1}}\}$, $\atr_{2}' \in \{\atrrd{\atr_{2}},\atrwr{\atr_{2}}\}$, and $\atr_{3}' \in \{\atrrd{\atr_{3}},\atrwr{\atr_{3}}\}$. The are five cases to be discussed: * $(\atr_{1}'=\atrrd{\atr_{1}},\atr_{2}'=\atrwr{\atr_{2}}) \in \po_\pcinstr$ and $\atr_1 = \atr_2$ and $(\atr_{2},\atr_{3}) \in \viso$, * $(\atr_{1},\atr_{2}) \in \arbo$ and $(\atr_{2},\atr_{3}) \in \viso$, * $(\atr_{1},\atr_{2}) \in \arbo$ and $(\atr_{2}'=\atrrd{\atr_{2}},\atr_{3}'=\atrwr{\atr_{3}}) \in \po_\pcinstr$ and $\atr_2 = \atr_3$, * $(\atr_{1}'=\atrrd{\atr_{1}},\atr_{2}'=\atrwr{\atr_{2}}) \in \po_\pcinstr$ and $\atr_1 = \atr_2$ and $(\atr_{2},\atr_{3}) \in \arbo$ and $\atr_{3}'= \atrwr{\atr_{3}}$, * $(\atr_{1},\atr_{2}) \in \arbo$ and $(\atr_{2},\atr_{3}) \in \arbo$ and $\atr_{3}'= \atrwr{\atr_{3}}$. Cases (a) and (b) imply that $(\atr_{1},\atr_{3}) \in \viso$ since $\arbo;\viso \subset \arbo$, which implies that $(\atr_{1}',\atr_{3}') \in \viso_\pcinstr$. Cases (c), (d), and (e) imply that $(\atr_{1},\atr_{3}) \in \arbo$ and $\atr_{3}'= \atrwr{\atr_{3}}$ then we get that $(\atrwr{\atr_{1}},\atrwr{\atr_{3}}) \in \viso_\pcinstr$ and $\atr_{3}'= \atrwr{\atr_{3}}$ which means that $(\atr_{1}',\atr_{3}') \in \viso_\pcinstr$. Note that the rule $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \viso_\pcinstr$ if $(\atr_{1},\atr_{2}) \in \arbo$ cannot change the fact that $$\atrwr{\atr_1}\in\{\atrwr{\atr'} \in \atrace_\pcinstr\ |\ (\atrwr{\atr'},\atrrd{\atr}) \in \viso_\pcinstr \wedge \exists\ \writeact(\atrwr{\atr'},\anaddr,\cdot)\in\amap(\atrwr{\atr'})\}$$ $$\atr_1 \in \{\atr' \in \atrace\ |\ (\atr',\atr) \in \viso \wedge \exists\ \writeact(\atr',\anaddr,\cdot)\in\amap(\atr')\}$$ Thus, the proof of $\axretval$ follows as in the previous case. For the case $\textsf{X} = \sic$, we show that $\atrace_\pcinstr$ satisfies $\axconflict$. If $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \sto_\pcinstr$, then $(\atr_{1},\atr_{2}) \in \sto \subset \viso$, which implies that $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \viso_\pcinstr$. Therefore, $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \viso_\pcinstr$, which concludes the proof. The axiom $\axretval$ can be proved as in the previous cases. Before presenting a strengthening of Lemma <ref> when $\textsf{X}$ is , we give an important characterization of traces. This characterization is stated in terms of acyclicity properties. $\atrace$ is a trace under iff $\arbo_{0}^{+}$ and $\viso_{0}^{+};\cfo$ are acyclic ($\arbo_{0}$ and $\viso_{0}$ are defined in Table <ref>). ($\Rightarrow$) Let $\atrace$ be a trace under . From $\axpoco$ and $\axcoarb$ we get that $\arbo_{0}^{+} \subset \arbo$, and $\arbo_{0}^{+}$ is acyclic because $\arbo$ is total order. Assume by contradiction that $\viso_{0}^{+};\cfo$ is cyclic which implies that $\viso;\cfo$ is cyclic since $\viso_{0}^{+} \subset \viso$, which means that there exist $\atr_1$ and $\atr_2$ such that $(\atr_1, \atr_2) \in \viso$ and $(\atr_2, \atr_1) \in \cfo$. $(\atr_2, \atr_1) \in \cfo$ implies that there exists $\atr_3$ such that $(\atr_3, \atr_1) \in \sto$ and $(\atr_3, \atr_2) \in \rfo$. Based on the definition of $\axretval$, $\atr_3$ has two possible instances: * $\atr_3$ corresponds to the "fictional" transaction that wrote the initial values which cannot be the case when $(\atr_1, \atr_2) \in \viso$ and $\atr_1$ writes to the same variable that $\atr_2$ reads from, * $\atr_3$ is the last transaction that occurs before $\atr_2$ that writes the value read by $\atr_2$, which means that $(\atr_1,\atr_3) \in \arbo$ which contradicts the fact that $(\atr_3, \atr_1) \in \sto$ since $\sto \subset \arbo$. ($\Leftarrow$) Let $\atrace$ be a trace such that $\arbo_{0}^{+}$ and $\viso_{0}^{+};\cfo$ are acyclic. Then, we define the relations $\viso$ and $\arbo$ such that $\viso = \viso_{0}^{+}$ and $\arbo$ is any total order that includes $\arbo_{0}^{+}$. Then, we obtain that $(\viso \cup \sto)^{+} \subset \arbo$ and $\viso;\cfo$ is acyclic. Thus, $\atrace$ satisfies the properties $\axpoco$ and $\axcoarb$. Next, we will show that $\atrace$ satisfies $\axretval$. Let $\atr$ be a transaction in $\atrace$ that contains a read event $\readact(\atr,\anaddr,\aval)$. Let $\atr_0$ be transaction in $\atrace$ such that $$\atr_0 = Max_{\arbo}(\{\atr' \in \atrace\ |\ (\atr',\atr) \in \viso \wedge \exists\ \writeact(\atr',\anaddr,\cdot)\in\amap(\atr')\})$$ then the read must return a value written by $\atr_0$. Assume by contradiction that there exists some other transaction $\atr_1 \neq \atr_0$ such that $(\atr_1,\atr) \in \rfo$. Then, we get that $(\atr_1,\atr_0) \in \arbo$ and both write to $\anaddr$, therefore, $(\atr_1,\atr_0) \in \sto$ since $\sto \subset \arbo$. Combining $(\atr_1,\atr) \in \rfo$ and $(\atr_1,\atr_0) \in \sto$ we obtain $(\atr,\atr_0) \in \cfo$ and since $(\atr_0,\atr) \in \viso$ then we obtain that $(\atr,\atr) \in \viso;\cfo$ which contradicts the fact that $\viso;\cfo$ is acyclic. Therefore, the read value was written by $\atr_0$ and $\atrace$ satisfies $\axretval$. Next we show that a trace $\atrace$ of a program $\aprog$ is iff the corresponding trace $\atrace_\pcinstr$ of $\aprog_\pcinstr$ is as well. This result is based on the observation that cycles in $\arbo_{0}^{+}$ or $\viso_{0}^{+};\cfo$ cannot be broken by splitting transactions. A trace $\atrace$ of $\aprog$ is iff the corresponding trace $\atrace_\pcinstr$ of $\aprog_\pcinstr$ is . The only-if direction follows from Lemma <ref>. For the if direction: consider a trace $\atrace_\pcinstr$ which is . We prove by contradiction that $\atrace$ must be as well. Assume that $\atrace$ is not then it must contain a cycle in either $\arbo_{0}^{+}$ or $\viso_{0}^{+};\cfo$ (based on Lemma <ref>). In the rest of the proof when we mention a cycle we implicitly refer to a cycle in either $\arbo_{0}^{+}$ or $\viso_{0}^{+};\cfo$. Splitting every transaction $\atr \in \atrace$ in a trace to a pair of transactions $\atrrd{\atr}$ and $\atrwr{\atr}$ that occur in this order might not maintain a cycle of $\atrace$. However, we prove that this is not possible and our splitting conserves the cycle. Assume we have a vertex $\atr$ as part of the cycle. We show that $\atr$ can be split into two transactions $\atrrd{\atr}$ and $\atrwr{\atr}$ while maintaining the cycle. Note that $\atr$ is part of a cycle iff either * $(\atr_{1},\atr) \in \arbo_{0}$ and $(\atr,\atr_{2})\in \arbo_{0}$ or * $(\atr_{1},\atr) \in \viso_{0}$ and $(\atr,\atr_{2})\in \viso_{0}$ or * $(\atr_{1},\atr) \in \viso_{0}$ and $(\atr,\atr_{2})\in \cfo$ or * $(\atr_{1},\atr) \in \cfo$ and $(\atr,\atr_{2})\in \viso_{0}$ where $\atr_{1}$ and $\atr_{2}$ might refer to the same transaction. Thus, by splitting $\atr$ to $\atrrd{\atr}$ and $\atrwr{\atr}$, the above four cases imply that: * if $(\atr_{1},\atr) \in \viso_{0}$ and $(\atr,\atr_{2})\in \arbo_{0}$ then $(\atr_{1}',\atrrd{\atr}) \in (\po_\pcinstr \cup \rfo_\pcinstr)$ and $(\atrwr{\atr},\atr_{2}')\in (\po_\pcinstr \cup \rfo_\pcinstr \cup \sto_\pcinstr)$ where $\atr_{1}' \in \{\atrrd{\atr_{1}},\atrwr{\atr_{1}}\}$ and $\atr_{2}' \in \{\atrrd{\atr_{2}},\atrwr{\atr_{2}}\}$. This maintains the vertices $\atr_{1}'$ and $\atr_{2}'$ connected in the cycle formed by the dependency relations of $\atrace_\pcinstr$ since $(\atrrd{\atr},\atrwr{\atr}) \in \po_\pcinstr$; * if $(\atr_{1},\atr) \in \sto$ and $(\atr,\atr_{2})\in \arbo_{0}$ then $(\atr_{1}',\atrwr{\atr}) \in \sto_\pcinstr$ and $(\atrwr{\atr},\atr_{2}')\in (\po_\pcinstr \cup \rfo_\pcinstr \cup \sto_\pcinstr)$ which maintains the vertices $\atr_{1}'$ and $\atr_{2}'$ connected in the cycle formed by the dependency relations of $\atrace_\pcinstr$; * $(\atr_{1},\atr) \in \viso_{0}$ and $(\atr_{2},\atr) \in \cfo$ then $(\atr_{1}',\atrrd{\atr}) \in (\po_\pcinstr \cup \rfo_\pcinstr)$ and $(\atrrd{\atr},\atr_{2}')\in \cfo_\pcinstr$ maintains the vertices $\atr_{1}'$ and $\atr_{2}'$ connected in the cycle formed by the dependency relations of $\atrace_\pcinstr$; * $(\atr_{1},\atr) \in \cfo$ and $(\atr_{2},\atr) \in \viso_{0}$ then $(\atr_{1}',\atrwr{\atr}) \in \cfo_\pcinstr$ and $(\atrwr{\atr},\atr_{2}')\in (\po_\pcinstr \cup \rfo_\pcinstr)$ which maintains the vertices $\atr_{1}'$ and $\atr_{2}'$ connected in the cycle formed by the dependency relations of $\atrace_\pcinstr$ as well. Therefore, doing the splitting creates a cycle in either $(\po_\pcinstr \cup \rfo_\pcinstr \cup \sto_\pcinstr)^{+}$ or $(\po_\pcinstr \cup \rfo_\pcinstr)^{+};\cfo_\pcinstr$ which implies that $\atrace_\pcinstr$ is not , a contradiction. The following lemma shows that a trace $\atrace$ is iff the corresponding trace $\atrace_\pcinstr$ is . The if direction in the proof is based on constructing a causal order $\viso$ and an arbitration order $\arbo$ for the trace $\atrace$ from the arbitration order $\arbo_\pcinstr$ in $\atrace_\pcinstr$ (since $\atrace_\pcinstr$ is a trace under serializability $\viso_\pcinstr$ and $\arbo_\pcinstr$ coincide). These are the smallest transitive relations such that: * if $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \arbo_\pcinstr$ then $(\atr_{1},\atr_{2}) \in \viso$, * if $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \arbo_\pcinstr$ then $(\atr_{1},\atr_{2}) \in \arbo$[If $\atrwr{\atr_{1}}$ is empty ($\atr_1$ is read-only), then we set $(\atr_{1},\atr_{2}) \in \arbo$ if $(\atrrd{\atr_{1}},\atrwr{\atr_{2}}) \in \viso_\pcinstr$. If $\atrwr{\atr_{2}}$ is empty, then $(\atr_{1},\atr_{2}) \in \arbo$ if $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \viso_\pcinstr$. If both $\atrwr{\atr_{1}}$ and $\atrwr{\atr_{2}}$ are empty, then $(\atr_{1},\atr_{2}) \in \arbo$ if $(\atrrd{\atr_{1}},\atrrd{\atr_{2}}) \in \viso_\pcinstr$.]. The only-if direction is based on the fact that any cycle in the dependency relations of $\atrace$ that is admitted under (characterized in Lemma <ref>) is “broken” by splitting transactions. Also, splitting transactions cannot introduce new cycles that do not originate in $\atrace$. A trace $\atrace$ is iff $\atrace_\pcinstr$ is ($\Leftarrow$) Assume that $\atrace_\pcinstr$ is . We will show that $\atrace$ is . Notice that if $(\atr_{1},\atr_{2}) \in \viso_{0}^{+}$ then $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \arbo_\pcinstr$ which implies that $(\atr_{1},\atr_{2}) \in \viso$. Similarly, if $(\atr_{1},\atr_{2}) \in \arbo_{0}^{+}$ then $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \arbo_\pcinstr$ or $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \arbo_\pcinstr$ which implies that $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \arbo_\pcinstr$ which in both cases implies that $(\atr_{1},\atr_{2}) \in \arbo$. Thus, $\atrace$ satisfies the properties $\axpoco$ and $\axcoarb$. Now assume that $(\atr_{1},\atr_{2})\in \arbo$ and $(\atr_{2},\atr_{3})\in \viso$. We show that $(\atr_{1},\atr_{3})\in \viso$. The assumption implies that $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \arbo_\pcinstr$ and $(\atrwr{\atr_{2}},\atrrd{\atr_{3}}) \in \arbo_\pcinstr$, which means that $(\atrwr{\atr_{1}},\atrrd{\atr_{3}}) \in \arbo_\pcinstr$. Therefore, $(\atr_{1},\atr_{3}) \in \viso$ and $\atrace$ satisfies the property $\axconflict$. Concerning $\axretval$, let $\atr$ be a transaction in $\atrace$ that contains a read event $\readact(\atr,\anaddr,\aval)$. Let $\atr_0$ be transaction in $\atrace$ such that $$\atr_0 = Max_{\arbo}(\{\atr' \in \atrace\ |\ (\atr',\atr) \in \viso \wedge \exists\ \writeact(\atr',\anaddr,\cdot)\in\amap(\atr')\}).$$ We show that the read must return a value written by $\atr_0$. The definitions of $\viso$ and $\arbo$ imply that $$\atrwr{\atr_1}\in\{\atrwr{\atr'} \in \atrace_\pcinstr\ |\ (\atrwr{\atr'},\atrrd{\atr}) \in \arbo_\pcinstr \wedge \exists\ \writeact(\atrwr{\atr'},\anaddr,\cdot)\in\amap(\atrwr{\atr'})\}$$ $$\atr_1 \in \{\atr' \in \atrace\ |\ (\atr',\atr) \in \viso \wedge \exists\ \writeact(\atr',\anaddr,\cdot)\in\amap(\atr')\}$$ because $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \arbo_\pcinstr$ implies $(\atr_{1},\atr_{2}) \in \viso$. Then, we obtain that $$\atrwr{\atr_0} = Max_{\arbo_\pcinstr}(\{\atrwr{\atr'} \in \atrace_\pcinstr\ |\ (\atrwr{\atr'},\atrrd{\atr}) \in \arbo_\pcinstr \wedge \exists\ \writeact(\atrwr{\atr'},\anaddr,\cdot)\in\amap(\atrwr{\atr'})\})$$ and since $\atrace_\pcinstr$ is we know that the read must return the value written by $\atrwr{\atr_0}$. Thus, the read returns the value written by $\atr_0$, which implies that $\atrace$ satisfies $\axretval$ holds. Therefore, $\atrace$ is . ($\Rightarrow$) Assume that $\atrace$ is . We show that $\atrace_\pcinstr$ is . Since $\atrace_\pcinstr$ is the result of splitting transactions, a cycle in its dependency relations can only originate from a cycle in $\atrace$. Therefore, it is sufficient to show that any happens-before cycle in $\atrace$ is broken in $\atrace_\pcinstr$. From Lemma <ref>, we have that $\atrace$ either does not admit a happens-before cycle or any (simple) happens-before cycle in $\atrace$ must have either two successive $\cfo$ dependencies or a $\sto$ dependency followed by a $\cfo$ dependency. If $\atrace$ does not admit a happens-before cycle then it is , and $\atrace_\pcinstr$ is trivially (since splitting transactions cannot introduce new cycles). [shape=rectangle ,draw=none,font=] (A0) at (0,0) [] $\atr_1$ ; [shape=rectangle ,draw=none,font=] (A1) at (2,0) [] $\atr_2$; [shape=rectangle ,draw=none,font=] (B1) at (4,0) [] $\atr_3$; [shape=rectangle ,draw=none,font=] (B2) at (5,0) [] $\Longrightarrow$; [shape=rectangle ,draw=none,font=] (C1) at (6,0) [] $\atrrd{\atr_{1}}$; [shape=rectangle ,draw=none,font=] (D1) at (8,0) [] $\atrwr{\atr_{1}}$; [shape=rectangle ,draw=none,font=] (D2) at (10,0) [] $\atrrd{\atr_{2}}$; [shape=rectangle ,draw=none,font=] (D0) at (12,0) [] $\atrwr{\atr_{2}}$; [shape=rectangle ,draw=none,font=] (E0) at (14,0) [] $\atrrd{\atr_{3}}$ ; [shape=rectangle ,draw=none,font=] (E1) at (16,0) [] $\atrwr{\atr_{3}}$; [ every edge/.style=draw=black,very thick] [->] (A0) edge[] node [above,font=] $\sto \cup \cfo$ (A1); [->] (A1) edge[] node [above,font=] $\cfo$ (B1); [->] (B1) edge[bend left] node [above,font=] $\hbo$ (A0); [->] (C1) edge[] node [above,font=] $\po_\pcinstr$ (D1); [->] (D1) edge[bend left] node [below,font=] $\sto_\pcinstr$ (D0); [->] (C1) edge[bend right] node [above,font=] $\cfo_\pcinstr$ (D0); [->] (D2) edge[] node [above,font=] $\po_\pcinstr$ (D0); [->] (D2) edge[bend right] node [above,font=] $\cfo_\pcinstr$ (E1); [->] (E0) edge[] node [above,font=] $\po_\pcinstr$ (E1); Otherwise, if $\atrace$ admits a happens-before cycle like above, then $\atrace$ must contain three transactions $\atr_{1}$, $\atr_{2}$, and $\atr_{3}$ such that $(\atr_{1},\atr_{2}) \in \sto \cup \cfo$, $(\atr_{2},\atr_{3}) \in \cfo$, and $(\atr_{3},\atr_{1}) \in \hbo$ (like in the picture above). Then, by splitting transactions we obtain that $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \sto_\pcinstr$ or $(\atrrd{\atr_{1}},\atrwr{\atr_{2}}) \in \cfo_\pcinstr$, and $(\atrrd{\atr_{2}},\atrwr{\atr_{3}}) \in \cfo_\pcinstr$. Since, we have $(\atrrd{\atr_{2}},\atrwr{\atr_{2}}) \in \po_\pcinstr$ (and not $(\atrwr{\atr_{2}},\atrrd{\atr_{2}}) \in \po_\pcinstr$), this cannot lead to a cycle in $\atrace_\pcinstr$, which concludes the proof that $\atrace_\pcinstr$ is The lemmas above are used to prove Theorem <ref> as follows: Proof of Theorem <ref>: For the if direction, assume by contradiction that $\aprog$ is not robust against relative to . Then, there must exist a trace $\atrace \in \tracesconf_{\ccc{}}(\aprog) \setminus \tracesconf_{\pcc{}}(\aprog)$. Lemmas <ref> and <ref> imply that the corresponding trace $\atrace_\pcinstr$ of $\aprog_\pcinstr$ is and not . Thus, $\aprog_\pcinstr$ is not robust against relative to . The only-if direction is proved similarly. Robustness against relative to has been shown to be reducible in polynomial time to the reachability problem under  [10]. Given a program $\aprog$ and a control location $\ell$, the reachability problem under asks whether there exists an execution of $\aprog$ under that reaches $\ell$. Therefore, as a corollary of Theorem <ref>, we obtain the following: Checking robustness against relative to is reducible to the reachability problem under in polynomial time. In the following we discuss the complexity of this problem in the case of finite-state programs (bounded data domain). The upper bound follows from Corollary <ref> and standard results about the complexity of the reachability problem under sequential consistency, which extend to , with a bounded [34] or parametric number of processes [44]. For the lower bound, given an instance $(\aprog,\ell)$ of the reachability problem under sequential consistency, we construct a program $\aprog'$ where each statement $s$ of $\aprog$ is executed in a different transaction that guards[That is, the transaction is of the form [lock; $s$; unlock]] the execution of $s$ using a global lock (the lock can be implemented in our programming language as usual, e.g., using a busy wait loop for locking), and where reaching the location $\ell$ enables the execution of a “gadget” that corresponds to the $\mathsf{SB}$ program in Figure <ref>. Executing each statement under a global lock ensures that every execution of $\aprog'$ under $\ccc$ is serializable, and faithfully represents an execution of $\aprog$ under sequential consistency. Moreover, $\aprog$ reaches $\ell$ iff $\aprog'$ contains a robustness violation, which is due to the $\mathsf{SB}$ execution. Checking robustness of a program with a fixed number of variables and bounded data domain against relative to is PSPACE-complete when the number of processes is bounded and EXPSPACE-complete, otherwise. § ROBUSTNESS AGAINST RELATIVE TO In this section, we show that checking robustness against relative to can be reduced in polynomial time to a reachability problem under the semantics. We reuse the program transformation from the previous section that allows to simulate behaviors on top of , and additionally, we provide a characterization of traces that distinguish the semantics from . We use this characterization to define an instrumentation (monitor) that is able to detect if a program under admits such traces. We show that the happens-before cycles in a robustness violation (against relative to ) must contain a $\sto$ dependency followed by a $\cfo$ dependency, and they should not contain two successive $\cfo$ dependencies. This follows from the fact that every happens-before cycle in a trace must contain either two successive $\cfo$ dependencies, or a $\sto$ dependency followed by a $\cfo$ dependency. Otherwise, the happens-before cycle would imply a cycle in the arbitration order. Then, any trace under where all its simple happens-before cycles contain two successive $\cfo$ dependencies is possible under . For instance, the trace of the non-robust $\mathsf{LU}$ execution in Figure <ref> contains $\sto$ dependency followed by a $\cfo$ dependency and does not contain two successive $\cfo$ dependencies which is disallowed , while the trace of the robust $\mathsf{WS}$ execution in Figure <ref> contains two successive $\cfo$ dependencies. As a first step, we prove the following theorem characterizing traces that are allowed under both and . A program $\aprog$ is robust against relative to iff every happens-before cycle in a trace of $\aprog$ under contains two successive $\cfo$ dependencies. Before giving the proof of the above theorem, we state several intermediate results that characterize cycles in or traces. First, we show that every trace in which all simple happens-before cycles contain two successive $\cfo$ is also a trace. If a trace $\atrace$ is and all happens-before cycles in $\atrace$ contain two successive $\cfo$ dependencies, then $\atrace$ is . Let $\arbo_{1}$ be a total order that includes $\arbo_{0}^{+}$ and $\arbo_{0}^{+};\cfo;\arbo_{0}^{*}$ ($\arbo_{0}^*$ is the reflexive closure of $\arbo_{0}$). This is well defined because there exists no cycle between tuples in these two relations. Indeed, if $(\atr_{1},\atr_{2}) \in \arbo_{0}^{+}$ and there exist $\atr_{3}$ and $\atr_{4}$ such that $(\atr_{2}, \atr_{3}) \in \arbo_{0}^{+}$, $(\atr_{3}, \atr_{4}) \in \cfo$, and $(\atr_{4}, \atr_{1}) \in \arbo_{0}^{*}$, then we have a cycle in $\arbo_{0}^{+};\cfo$ that does not contain two successive $\cfo$ dependencies, which contradicts the hypothesis. Also, for every pair of transactions $(\atr_{1},\atr_{2})$ there cannot exist $\atr_{3}$ and $\atr_{4}$ such that $$(\atr_{2}, \atr_{3}) \in \arbo_{0}^{+},\ (\atr_{3}, \atr_{4}) \in \cfo\mbox{ and }(\atr_{4}, \atr_{1}) \in \arbo_{0}^{*}$$ $\atr_{3}'$ and $\atr_{4}'$ such that $$(\atr_{1}, \atr_{3}') \in \arbo_{0}^{+},\ (\atr_{3}', \atr_{4}') \in \cfo\mbox{ and }(\atr_{4}', \atr_{2}) \in \arbo_{0}^{*}$$ This will imply a cycle in $\arbo_{0}^{+};\cfo;\arbo_{0}^{+};\cfo$ which again contradicts the hypothesis. Also, let $\viso_{1}$ be the smallest transitive relation that includes $\arbo_{0}^{+}$ and $\arbo_{1};\arbo_{0}^{+}$. We show that $\viso_{1}$ and $\arbo_{1}$ are causal and arbitration orders of $\atrace$ that satisfy all the axioms of . $\axpoco$ and $\axcoarb$ hold trivially. Since $\sto \subseteq \viso_{1}$, $\axconflict$ holds as well. $\axpc$ holds because $\arbo_{1} ; \viso_{1} = \arbo_{1};(\arbo_{0}^{+} \cup \arbo_{1};\arbo_{0}^{+})^+ = \arbo_{1};\arbo_{0}^{+} \subset \viso_{1}$. The axiom $\axretval$ is equivalent to the acyclicity of $\viso_{1};\cfo$ when $\axpoco$ and $\axcoarb$ hold. Assume by contradiction that $\viso_1;\cfo$ is cyclic. From the definition of $\viso_1$ and the fact that $\arbo_{1}$ is total order we obtain that either: * $\arbo_{0}^{+};\cfo$ is cyclic, which implies that there exists a happens-before cycle that does not contain two successive $\cfo$, which contradicts the hypothesis, or * $\arbo_{1};\arbo_{0}^{+};\cfo$ is cyclic, which implies that there exist $\atr_{1}$, $\atr_{2}$, and $\atr_{3}$ such that $(\atr_{2}, \atr_{3}) \in \arbo_{0}^{+}$, $(\atr_{3}, \atr_{1}) \in \cfo$ and $(\atr_{1},\atr_{2}) \in \arbo_{1}$. This contradicts the fact that $(\atr_{2}, \atr_{3}) \in \arbo_{0}^{+}$ and $(\atr_{3}, \atr_{1}) \in \cfo$ implies $(\atr_{2},\atr_{1}) \in \arbo_{1}$. Therefore, $\atrace$ satisfies $\axretval$ for $\viso_1$ and $\arbo_1$, which concludes the proof. The proof of Theorem <ref> also relies on the following lemma that characterizes happens-before cycles permissible under . [22, 12] If a trace $\atrace$ is , then all its happens-before cycles must contain two successive $\cfo$ dependencies. Proof of Theorem <ref>: For the only-if direction, if $\aprog$ is robust against relative to then every trace $\atrace$ of $\aprog$ under is as well. Therefore, by Lemma <ref>, all cycles in $\atrace$ contain two successive $\cfo$ which concludes the proof of this direction. For the reverse, let $\atrace$ be a trace of $\aprog$ under such that all its happens-before cycles contain two successive $\cfo$. Then, by Lemma <ref>, we have that $\atrace$ is . Thus, every trace $\atrace$ of $\aprog$ under is . Next, we present an important lemma that characterizes happens before cycles possible under the semantics. This is a strengthening of a result in [12] which shows that all happens before cycles under must have two successive dependencies in $\{\cfo,\sto\}$ and at least one $\cfo$. We show that the two successive dependencies cannot be $\cfo$ followed $\sto$, or two successive $\sto$. If a trace $\atrace$ is then all happens-before cycles in $\atrace$ must contain either two successive $\cfo$ dependencies or a $\sto$ dependency followed by a $\cfo$ dependency. It was shown in [12] that all happens-before cycles under must contain two successive dependencies in $\{\cfo,\sto\}$ and at least one $\cfo$. Assume by contradiction that there exists a cycle with $\cfo$ dependency followed by $\sto$ dependency or two successive $\sto$ dependencies. This cycle must contain at least one additional dependency. Otherwise, the cycle would also have a $\sto$ dependency followed by a $\cfo$ dependency, or it would imply a cycle in $\sto$, which is not possible (since $\sto \subset \arbo$ and $\arbo$ is a total order). Then, we get that the dependency just before $\cfo$ is either $\po$ or $\rfo$ (i.e., $\viso_0$) since we cannot have $\cfo$ or $\sto$ followed by $\cfo$. Also, the relation after $\sto$ is either $\po$ or $\rfo$ or $\sto$ (i.e., $\arbo_0$) since we cannot have $\sto$ followed by $\cfo$. Thus, the cycle has the following shape: [shape=rectangle ,draw=none,font=] (A0) at (0,0) [] $\atr_1$ ; [shape=rectangle ,draw=none,font=] (A1) at (1.3,0) [] $\atr_2$; [shape=rectangle ,draw=none,font=] (B1) at (2.6,0) [] $\atr_3$; [shape=rectangle ,draw=none,font=] (B2) at (3.9,0) [] $\atr_4$; [shape=rectangle ,draw=none,font=] (C0) at (4.5,0) [] $\cdots$ ; [shape=rectangle ,draw=none,font=] (C1) at (5.1,0) [] $\atr_i$; [shape=rectangle ,draw=none,font=] (D1) at (6.4,0) [] $\atr_{i+1}$; [shape=rectangle ,draw=none,font=] (D2) at (7.9,0) [] $\atr_{i+2}$; [shape=rectangle ,draw=none,font=] (D0) at (9.4,0) [] $\atr_{i+3}$; [shape=rectangle ,draw=none,font=] (E0) at (10.2,0) [] $\cdots$ ; [shape=rectangle ,draw=none,font=] (E1) at (11,0) [] $\atr_{n-4}$; [shape=rectangle ,draw=none,font=] (F1) at (12.5,0) [] $\atr_{n-3}$; [shape=rectangle ,draw=none,font=] (F2) at (14,0) [] $\atr_{n-2}$; [shape=rectangle ,draw=none,font=] (F3) at (15.5,0) [] $\atr_{n-1}$; [shape=rectangle ,draw=none,font=] (F4) at (17,0) [] $\atr_{n}$; [ every edge/.style=draw=black,very thick] [->] (A0) edge[] node [above,font=] $\cfo$ (A1); [->] (A1) edge[] node [above,font=] $\sto$ (B1); [->] (B1) edge[] node [above,font=] $\arbo_0$ (B2); [->] (C1) edge[] node [above,font=] $\viso_0$ (D1); [->] (D1) edge[] node [above,font=] $\cfo$ (D2); [->] (D2) edge[] node [above,font=] $\sto$ (D0); [->] (E1) edge[] node [above,font=] $\viso_0$ (F1); [->] (F1) edge[] node [above,font=] $\cfo$ (F2); [->] (F2) edge[] node [above,font=] $\sto$ (F3); [->] (F3) edge[] node [above,font=] $\arbo_0$ (F4); [->] (F4) edge[bend left=11] node [above,font=] $\viso_0$ (A0); Since $\viso_0;\cfo\subseteq \arbo$ is a consequence of the axioms [26], we get that $(\atr_{n}, \atr_2) \in \arbo$, $(\atr_{i}, \atr_{i+2}) \in \arbo$ and $(\atr_{n-4}, \atr_{n-2}) \in \arbo$, which allows to “short-circuit” the cycle. Using the fact that $\sto \subset \arbo$, $\viso_0 \subset \arbo$, and $\arbo_0 \subset \arbo$, and applying the short-circuiting process multiple times, we obtain a cycle in the arbitration order $\arbo$ which contradicts the fact that $\arbo$ is a total order. Combining the results of Theorem <ref> and Lemmas <ref> and <ref>, we obtain the following characterization of traces which violate robustness against relative to . A program $\aprog$ is not robust against relative to iff there exists a trace $\atrace_\pcinstr$ of $\aprog_\pcinstr$ under such that the trace $\atrace$ obtained by merging[This transformation has been defined at the beginning of Section <ref>.] read and write transactions in $\atrace_\pcinstr$ contains a happens-before cycle that does not contain two successive $\cfo$ dependencies, and it contains a $\sto$ dependency followed by a $\cfo$ dependency. The results above enable a reduction from checking robustness against relative to to a reachability problem under the semantics. For a program $\aprog$, we define an instrumentation denoted by $\sem{\aprog}$, such that $\aprog$ is not robust against relative to iff $\sem{\aprog}$ violates an assertion under . The instrumentation consists in rewriting every transaction of $\aprog$ as shown in Figure <ref>. [->,>=stealth',shorten >=1pt,auto,node distance=4cm, semithick, transform shape,every text node part/.style=align=left] [shape=rectangle ,draw=none,font=] at (0,0) (m) $\alpha$; [shape=rectangle ,draw=none,font=] at (1.5,0) (n) $\atr_{\instr}$; [shape=rectangle ,draw=none,font=] at (3,0) (n1) $\beta$; [shape=rectangle ,draw=none,font=] at (4.5,0) (n2) $\atr_{0}$; [shape=rectangle ,draw=none,font=] at (6,0) (n3) $\gamma$; [shape=rectangle ,draw=none,font=] at (7.5,0) (n4) $\atr$; [ every edge/.style=draw=black,very thick] [->] (n) edge[bend left=17] node $\cfo$ (n2); [->] (n2) edge[bend left=17] node $\hbo$ (n4); [->] (n4) edge[bend left=23, above] node $\sto$ (n); Execution simulating a violation to robustness against relative to . The instrumentation $\sem{\aprog}$ running under $\serc$ simulates the semantics of $\aprog$ using the same idea of decoupling the execution of the read part of a transaction from the write part. It violates an assertion when it simulates a trace containing a happens-before cycle as in Theorem <ref>. The execution corresponding to this trace has the shape given in Figure <ref>, where $\atr_{\instr}$ is the transaction that occurs between the $\sto$ and the $\cfo$ dependencies, and every transaction executed after $\atr_{\instr}$ (this can be a full transaction in $\aprog$, or only the read or write part of a transaction in $\aprog$) is related by a happens-before path to $\atr_{\instr}$ (otherwise, the execution of this transaction can be reordered to occur before $\atr_{\instr}$). A transaction in $\aprog$ can have its read part included in $\alpha$ and the write part included in $\beta$ or $\gamma$. Also, $\beta$ and $\gamma$ may contain transactions in $\aprog$ that executed only their read part. It is possible that $\atr_{0} = \atr$, $\beta=\gamma=\epsilon$, and $\alpha = \epsilon$ (the $\mathsf{LU}$ program shown in Figure <ref> is an example where this can happen). The instrumentation uses auxiliary variables to track happens-before dependencies, which are explained below. \begin{figure}[!ht] \footnotesize \begin{minipage}{\linewidth} Transaction ``\plog{begin} $\langle$read$\rangle^{*}$ $\langle$test$\rangle^{*}$ $\langle$write$\rangle^{*}$ \plog{commit}'' is rewritten to: \end{minipage} \begin{minipage}{0.655\linewidth} \begin{lstlisting} if ( !done$_\#$ ) if ( * ) begin <read>$^{*}$ <test>$^{*}$ commit $\label{ln:part11}$ if ( !done$_\#$ ) begin <write>$^{*}$ commit $\label{ln:part12}$ $\mathcal{I}$(begin) ($\mathcal{I}$(<write>))$^{*}$ $\mathcal{I}$(commit)$\label{ln:part31}$ begin ($\mathcal{I}_\#$(<read>))$^{*}$ <test>$^{*}$ ($\mathcal{I}_\#$(<write>))$^{*}$ $\mathcal{I}_\#$(commit)$\label{ln:part21}$ assume false; else if ( * ) rdSet' := $\emptyset$; wrSet' := $\emptyset$; $\mathcal{I}$(begin) ($\mathcal{I}$(<read>))$^{*}$ <test>$^{*}$ $\mathcal{I}$(commit)$\label{ln:part32}$ $\mathcal{I}$(begin) ($\mathcal{I}$(<write>))$^{*}$ $\mathcal{I}$(commit)$\label{ln:part33}$ \end{lstlisting} % $\mathcal{I}$(begin) ($\mathcal{I}$(<read>))$^{*}$ <test>$^{*}$ ($\mathcal{I}$(<write>))$^{*}$ $\mathcal{I}$(commit) \end{minipage}\hfill \begin{minipage}{0.305\linewidth} % \lstset{numbers=none} $\mathcal{I}_\#$( r := x ): \begin{lstlisting}[xleftmargin=2mm,firstnumber=16] r := x; $\label{ln:delay1}$ hbR['x'] := 0; rdSet := rdSet $\cup$ { 'x' }; \end{lstlisting} $\mathcal{I}_\#$( x := e ): \begin{lstlisting}[xleftmargin=2mm,firstnumber=19] if ( varW == $\bot$ and * ) varW := 'x'; \end{lstlisting} $\mathcal{I}_\#$( commit ): \begin{lstlisting}[xleftmargin=2mm,firstnumber=21] assume ( varW != $\bot$ ) done$_\#$ := true $\label{ln:delay2}$ \end{lstlisting} \end{minipage} \vspace{1mm} \begin{minipage}{0.5\linewidth} $\mathcal{I}$( begin ): \begin{lstlisting}[xleftmargin=3mm,firstnumber=23] hb := $\bot$ if ( hbP != $\bot$ and hbP < 2 ) hb := 0; else if ( hbP = 2 ) hb := 2; \end{lstlisting} $\mathcal{I}$( commit ): \begin{lstlisting}[xleftmargin=3mm,firstnumber=29] assume ( hb != $\bot$ ) $\label{ln:assume}$ assert ( hb == 2 or varW $\not\in$ wrSet' ); $\label{ln:assert}$ if ( hbP == $\bot$ or hbP > hb ) $\label{ln:hbupdates1}$ hbP = hb; for each 'x' $\in$ wrSet' if ( hbW['x'] == $\bot$ or hbW['x'] > hb ) hbW['x'] = hb; for each 'x' $\in$ rdSet' if ( hbR['x'] == $\bot$ or hbR['x'] > hb ) hbR['x'] = hb; $\label{ln:hbupdates2}$ rdSet := rdSet $\cup$ rdSet'; $\label{ln:rdSetUpdate}$ wrSet := wrSet $\cup$ wrSet'; $\label{ln:wrSetUpdate}$ \end{lstlisting} \end{minipage} \hfill \begin{minipage}{0.45\linewidth} $\mathcal{I}$( r := x ): \begin{lstlisting}[xleftmargin=3mm,firstnumber=42] r := x; rdSet' := rdSet' $\cup$ { 'x' }; if ( 'x' $\in$ wrSet ) $\label{ln:rdHBCont}$ if ( hbW['x'] != 2 ) hb := 0 $\label{ln:rdHBConthb1}$ else if ( hb == $\bot$ ) hb := hbW['x'] $\label{ln:rdHBConthb2}$ \end{lstlisting} $\mathcal{I}$( x := e ): \begin{lstlisting}[xleftmargin=3mm,firstnumber=49] x := e; wrSet' := wrSet' $\cup$ { 'x' }; if ( 'x' $\in$ wrSet ) $\label{ln:wrHBCont1}$ if ( hbW['x'] != 2 ) hb := 0 else if ( hb == $\bot$ ) hb := hbW['x'] if ( 'x' $\in$ rdSet ) $\label{ln:wrHBCont2}$ if ( hb = $\bot$ or hb > hbR['x'] + 1 ) hb := min(hbR['x'] + 1,2) \end{lstlisting} \end{minipage} \vspace{-10pt} \normalsize \caption{A program instrumentation for checking robustness against \pcc{} relative to \sic{}. The auxiliary variables used by the instrumentation are shared variables, except for \texttt{hbP}, \texttt{rdSet'}, and \texttt{wrSet'}, which are process-local variables, and they are initially set to $\bot$. This instrumentation uses program constructs which can be defined as syntactic sugar from the syntax presented in Section~\ref{sec:consistency}, e.g., if-then-else statements (outside transactions).} \label{Figure:Instr} \vspace{-10pt} \end{figure} % \scriptsize % \begin{minipage}{0.30\linewidth} % \begin{eqnarray} % %%%%%%%%%%% % &&\semidlerpcinstr{\thetransitionnotogo{\atr}{[\rds;\ \wrs]}} = \notag\\ % &&\thetransitionnotogo{\atrrd{\atr}}{[\rds]}\notag\\ % &&\thetransitionnotogo{\atrwr{\atr}}{[\wrs]}\notag\\[4mm] % %%%%%%%%%%% % &&\semidlermove{\thetransitionnotogo{\atr}{[\rds;\ \wrs]}} = \notag\\ % &&\thetransitionnolabel{\thecondition{ a_{\wrflag} = \perp \land\ \apr_{\wrflag} = \perp}}\notag\\ % &&\thetransitionnolabel{\lit*{Foreach}\ \ensuremath{\rdO}\in\ensuremath{\rds}}\notag\\ % &&\thetransitionnolabel{\ \ \ \sem{\rdO}_{\instr}}\notag\\ % &&\thetransitionnolabel{\lit*{Foreach}\ \ensuremath{\wrO}\in\ensuremath{\wrs}}\notag\\ % &&\thetransitionnolabel{\ \ \ \sem{\wrO}_{\instr}}\notag\\ % &&\thetransitionnolabel{\thecondition{ a_{\wrflag} \neq \perp \land\ \apr_{\wrflag} = \apr}}\notag\\[4mm] % %%%%%%%%%%% % %%%%%%%%%%% % &&\semidlermove{\thetransitionnotogo{\rdO}{\theload{\areg}{\anaddr}}} =\notag\\ % &&\thetransitionnolabel{\theload{\areg}{\anaddr'}}\notag\\ % &&\thetransitionnolabel{\theassign{\rdaddr{\anaddr}}{0}}\notag\\ % &&\thetransitionnolabel{\theassign{\rdSet}{\rdSet \cup\ \{\anaddr\}}}\notag\\[4mm] % %%%%%%%%%% % &&\semidlermove{\thetransitionnotogo{\wrO}{\thestore{\anaddr}{e}}} = \notag\\ % &&\thetransitionnolabel{\thestore{\anaddr'}{e}}\notag\\ % &&\thetransitionnolabel{\theifcondition{ a_{\wrflag} = \perp \land\ \apr_{\wrflag} = \perp \land\ *}}\notag\\ % &&\thetransitionnolabel{\ \ \ \theassign{a_{\wrflag}}{\anaddr}}\notag\\ % &&\thetransitionnolabel{\ \ \ \theassign{\apr_{\wrflag}}{\apr}}\notag % %%%%%%%%%%% % %%%%%%%%%%% % \end{eqnarray} % \end{minipage}\hfill % \begin{minipage}{0.30\linewidth} % \begin{eqnarray} % %%%%%%%%%%% % %%%%%%%%%%% % &&\semidler{\thetransitionnotogo{\atr}{[\rds;\ \wrs]}} = \notag\\ % &&\thetransitionnolabel{\thecondition{ a_{\wrflag} \neq \perp \land\ \apr \neq \apr_{\wrflag}}}\notag\\ % &&\thetransitionnolabel{\theifcondition{\rwMap{\apr} \neq \perp \land\ \rwMap{\apr} < 2 }}\notag\\ % &&\thetransitionnolabel{\ \ \ \theassign{\rw}{0}}\notag\\ % &&\thetransitionnolabel{\theelseifcondition{\rwMap{\apr} = 2}}\notag\\ % &&\thetransitionnolabel{\ \ \ \theassign{\rw}{2}}\notag\\ % &&\thetransitionnolabel{\lit*{Foreach}\ \ensuremath{\rdO}\in\ensuremath{\rds}}\notag\\ % &&\thetransitionnolabel{\ \ \ \sem{\rdO}_{\textsf{H}}}\notag\\ % &&\thetransitionnolabel{\lit*{Foreach}\ \ensuremath{\wrO}\in\ensuremath{\wrs}}\notag\\ % &&\thetransitionnolabel{\ \ \ \sem{\wrO}_{\textsf{H}}}\notag\\ % &&\thetransitionnolabel{\thecondition{\rw \neq \perp}}\notag\\ % &&\thetransitionnolabel{\theifcondition{\rw \neq 2 \land\ (\{a_{\wrflag}\} \cap\ \wrSet') \neq \perp}}\notag\\ % &&\thetransitionnolabel{\ \ \ \lit*{assert}\ \myfalse\lit*{;}}\notag\\ % &&\thetransitionnolabel{\theifcondition{\rwMap{\apr} = \perp \lor\ \rwMap{\apr} > \rw}}\notag\\ % &&\thetransitionnolabel{\ \ \ \theassign{\rwMap{\apr}}{\rw}}\notag\\ % &&\thetransitionnolabel{\lit*{Foreach}\ \ensuremath{\anaddr}\in\ensuremath{\wrSet'}}\notag\\ % &&\thetransitionnolabel{\ \ \ \theifcondition{\wraddr{\anaddr} = \perp \lor\ \wraddr{\anaddr} > \rw}}\notag\\ % &&\thetransitionnolabel{\ \ \ \ \ \ \theassign{\wraddr{\anaddr}}{\rw}}\notag\\ % &&\thetransitionnolabel{\lit*{Foreach}\ \ensuremath{\anaddr}\in\ensuremath{\rdSet'}}\notag\\ % &&\thetransitionnolabel{\ \ \ \theifcondition{\rdaddr{\anaddr} = \perp \lor\ \rdaddr{\anaddr} > \rw}}\notag\\ % &&\thetransitionnolabel{\ \ \ \ \ \ \theassign{\rdaddr{\anaddr}}{\rw}}\notag\\ % &&\thetransitionnolabel{\theassign{\wrSet}{\wrSet \cup\ \wrSet'}}\notag\\ % &&\thetransitionnolabel{\theassign{\rdSet}{\rdSet \cup\ \rdSet'}}\notag % %%%%%%%%%%% % %%%%%%%%%%% % \end{eqnarray} % \end{minipage}\hfill % \begin{minipage}{0.30\linewidth} % \begin{eqnarray} % %%%%%%%%%%% % &&\semidler{\thetransitionnotogo{\rdO}{\theload{\areg}{\anaddr}}} = \notag\\ % &&\thetransitionnolabel{\theload{\areg}{\anaddr}}\notag\\ % &&\thetransitionnolabel{\theassign{\rdSet'}{\rdSet' \cup\ \{\anaddr\}}}\notag\\ % &&\thetransitionnolabel{\theifcondition{(\wrSet \cap\ \{\anaddr\}) \neq \perp}}\notag\\ % &&\thetransitionnolabel{\ \ \ \theifcondition{\wraddr{\anaddr} \neq 2}}\notag\\ % &&\thetransitionnolabel{\ \ \ \ \ \ \theassign{\rw}{0}}\notag\\ % &&\thetransitionnolabel{\ \ \ \theelseifcondition{\rw = \perp}}\notag\\ % &&\thetransitionnolabel{\ \ \ \ \ \ \theassign{\rw}{\wraddr{\anaddr}}}\notag\\[4mm] % %%%%%%%%%%% % &&\semidler{\thetransitionnotogo{\wrO}{\thestore{\anaddr}{e}}} = \notag\\ % &&\thetransitionnolabel{\thestore{\anaddr}{e}}\notag\\ % &&\thetransitionnolabel{\theassign{\wrSet'}{\wrSet' \cup\ \{\anaddr\}}}\notag\\ % &&\thetransitionnolabel{\theifcondition{(\wrSet \cap\ \{\anaddr\}) \neq \perp}}\notag\\ % &&\thetransitionnolabel{\ \ \ \theifcondition{\wraddr{\anaddr} \neq 2}}\notag\\ % &&\thetransitionnolabel{\ \ \ \ \ \ \theassign{\rw}{0}}\notag\\ % &&\thetransitionnolabel{\ \ \ \theelseifcondition{\rw = \perp}}\notag\\ % &&\thetransitionnolabel{\ \ \ \ \ \ \theassign{\rw}{\wraddr{\anaddr}}}\notag\\ % &&\thetransitionnolabel{\theifcondition{(\rdSet \cap\ \{\anaddr\}) \neq \perp}}\notag\\ % &&\thetransitionnolabel{\ \ \ \theifcondition{\rw = \perp \vee\ \rw > \rdaddr{\anaddr} + 1}}\notag\\ % &&\thetransitionnolabel{\ \ \ \ \ \ \theassign{\rw}{\minOf{\rdaddr{\anaddr} + 1}{2}}}\notag % %%%%%%%%%%% % %%%%%%%%%%% % \end{eqnarray} % \end{minipage} % \normalsize % \caption{Program Instrumentation Rules. $min(a,b)$ is the smallest of $a$ and $b$.} % \label{Figure:Instr} % \end{figure} The instrumentation executes (incomplete) transactions without affecting the auxiliary variables (without tracking happens-before dependencies) (lines~\ref{ln:part11} and \ref{ln:part12}) until a non-deterministically chosen point in time when it declares the current transaction as the candidate for $_$ (line~\ref{ln:part21}). Only one candidate for $_$ can be chosen during the execution. This transaction executes only its reads and it chooses non-deterministically a variable that it could write as a witness for the $$ dependency (see lines~\ref{ln:delay1}-\ref{ln:delay2}). The name of this variable is stored in a global variable \texttt{varW} (see the definition of $ℐ_#$( x := e )). %The id of the process executing the current transaction is also recorded, in the variable \texttt{pW} (see the definition of $\mathcal{I}_\#$( x := e )). The writes are \emph{not} applied on the shared memory. Intuitively, $_$ should be thought as a transaction whose writes are delayed for later, after transaction $$ in Figure~\ref{fig:violationInstr} executed. The instrumentation checks that $_$ and $$ can be connected by some happens-before path that includes the $$ and $$ dependencies, and that does not contain two consecutive $$ dependencies. If it is the case, it violates an assertion at the commit point of $$. Since the write part of $_$ is intuitively delayed to execute after $$, the process executing $_$ is disabled all along the execution (see the \texttt{assume false}). After choosing the candidate for $_$, the instrumentation uses the auxiliary variables for tracking happens-before dependencies. Therefore, \texttt{rdSet} and \texttt{wrSet} record variables read and written, respectively, by transactions that are connected by a happens-before path to $_$ (in a trace of $$). This is ensured by the assume at line~\ref{ln:assume}. During the execution, the variables read or written by a transaction\footnote{These are stored in the local variables \texttt{rdSet'} and \texttt{wrSet'} while the transaction is running.} that writes a variable in \texttt{rdSet} (see line~\ref{ln:wrHBCont2}), or reads or writes a variable in \texttt{wrSet} (see lines~\ref{ln:rdHBCont} and~\ref{ln:wrHBCont1}), will be added to these sets (see lines~\ref{ln:rdSetUpdate} and~\ref{ln:wrSetUpdate}). %The \texttt{assume} at line~\ref{ln:assume} will pass anytime such a conflicting access happens. Since the variables that $_$ writes in $$ are not recorded in \texttt{wrSet}, these happens-before paths must necessarily start with a $$ dependency (from $_$). When the assertion fails (line~\ref{ln:assert}), the condition \texttt{varW} $∈$ \texttt{wrSet}' ensures that the current transaction has a $$ dependency towards the write part of $_$ (the current transaction plays the role of $$ in Figure~\ref{fig:violationInstr}). %A subtle point of the instrumentation is that after $\atr_{\instr}$, transactions can execute in their entirety (as in $\aprog$) or in two phases, first their read part and then the write part (to simulate the \pcc{} semantics). The two scenarios lead to different effects on the auxiliary variables tracking happens-before dependencies. The reads of a transaction that executes in two phases are not recorded in \texttt{rdSet} if none of them is conflicting with a write in a previous transaction, as opposed to the case when the transaction executes in its entirety (this transaction could be connected by happens-before to $\atr_{\instr}$ because of its writes). This is complete because a transaction $\atr'$ that must execute in two phases is related by $\cfo$ to another transaction $\atr''$ which is in happens before itself (more precisely before its write part). Then, either the happens-before from $\atr''$ to $\atr'$ is $\sto$ and this scenario would be captured when $\atr'$ is chosen as a candidate for $\atr_{\instr}$, or it is $\cfo$ and the path from $\atr_{\instr}$ to $\atr'$ contains two consecutive The rest of the instrumentation checks that there exists a happens-before path from $_$ to $$ that does not include two consecutive $$ dependencies, called a \sic{}$_$ path. This check is based on the auxiliary variables whose name is prefixed by \texttt{hb} and which take values in the domain ${,0,1,2}$ ($$ represents the initial value). \begin{itemize}[topsep=3pt] \item \texttt{hbR['x']} (resp., \texttt{hbW['x']}) is 0 iff there exists a transaction $\atr'$ that reads \texttt{x} (resp., writes to \texttt{x}), such that there exists a \sic{}$_{\neg}$ path from $\atr_{\instr}$ to $\atr'$ that ends with a dependency which is \emph{not} $\cfo$, \item \texttt{hbR['x']} (resp., \texttt{hbW['x']}) is 1 iff there exists a transaction $\atr'$ that reads \texttt{x} (resp., writes to \texttt{x}) that is connected to $\atr_{\instr}$ by a \sic{}$_{\neg}$ path, and \emph{every} \sic{}$_{\neg}$ path from $\atr_{\instr}$ to a transaction $\atr''$ that reads \texttt{x} (resp., writes to \texttt{x}) ends with an $\cfo$ dependency, \item \texttt{hbR['x']} (resp., \texttt{hbW['x']}) is 2 iff there exists no \sic{}$_{\neg}$ path from $\atr_{\instr}$ to a transaction $\atr'$ that reads \texttt{x} (resp., writes to \texttt{x}). \end{itemize} The local variable \texttt{hbP} has the same interpretation, except that $'$ and $”$ are instantiated over transactions in the same process (that already executed) instead of transactions that read or write a certain variable. Similarly, the variable \texttt{hb} is a particular case where $'$ and $”$ are instantiated to the current transaction. The violation of the assertion at line~\ref{ln:assert} implies that \texttt{hb} is 0 or 1, which means that there exists a \sic{}$_$ path from $_$ to $$. During each transaction that executes after $_$, the variable \texttt{hb} characterizing happens-before paths that end in this transaction is updated every time a new happens-before dependency is witnessed (using the values of the other variables). For instance, when witnessing a $$ dependency (line~\ref{ln:rdHBCont}), if there exists a \sic{}$_$ path to a transaction that writes to \texttt{x}, then the path that continues with the $$ dependency towards the current transaction is also a \sic{}$_$ path, and the last dependency of this path is not $$. Therefore, \texttt{hb} is set to 0 (see line~\ref{ln:rdHBConthb1}). Otherwise, if every path to a transaction that writes to \texttt{x} is not a \sic{}$_$ path, then every path that continues to the current transaction (by taking the $$ dependency) remains a non \sic{}$_$ path, and \texttt{hb} is set to the value of \texttt{hbW[`x`]}, which is 2 in this case (see line~\ref{ln:rdHBConthb2}). Before ending a transaction, the value of \texttt{hb} can be used to modify the \texttt{hbR}, \texttt{hbW}, and \texttt{hbP} variables, but only if those variables contain bigger values (see lines~\ref{ln:hbupdates1}--\ref{ln:hbupdates2}). %TODO I STOPPED HERE %Every transaction in $\aprog_\siinstr$ is constructed through the rewriting of the corresponding transaction %from $\aprog_\pcinstr$ where we use auxiliary flags to store the accessed locations and build the happens %before relation between $\atr_{\instr}$ and $\atr$. %The main ideas of the instrumentation consists executing the transactions of the transformed program as in %$\semidlerpcinstr{\atr}$ of Figure \ref{Figure:Instr} which constitute the elements of the sequence $\alpha$ in $\atrace_{\instr}$. Until reaching the transaction $\atr_{\instr}$ which is instrumented as in $\semidlermove{\atr_{\instr}}$ of Figure \ref{Figure:Instr} where we uses a copy of the variables in the original program denoted $\anaddr'$. Then, the instrumented transaction $\semidlermove{\atr_{\instr}}$ will write only to $\anaddr'$ and read only from $\anaddr'$. The writes made by all the other transactions that occur in $\alpha$ are applied to both $\anaddr'$ and $\anaddr$. Also, all of these transactions cannot read from $\anaddr'$. %We use the flag $a_{\wrflag}$ to store the name of variable that generated the dependency $(\atr,\atr_{\instr}) \in \sto$. %Also, we use the flag $\apr_{\wrflag}$ to store the identifier of the process that executed $\atr_{\instr}$. %Afterward, the transactions in $\beta \cdot \atr_{0} \cdot \gamma$ contribute to the happens-before relation between $\atr_{\instr}$ and $\atr$. The instrumentation only simulate traces such that $\beta \cdot \atr_{0} \cdot \gamma$ doesn't contain any transaction from the process that executed $\atr_{\instr}$, which is sound and complete. %Then, we for every transaction $\atr_1$ in $\beta \cdot \atr_{0} \cdot \gamma$ either we apply directly $\semidler{\atr_1}$ of Figure \ref{Figure:Instr} on $\atr_1$ or $\semidler{\semidlerpcinstr{\atr_1}}$ on the two transactions resulting from the splitting of $\atr_1$. %Note that in the instrumentation $\semidler{\atr_1}$ one of the sets $\ensuremath{\wrs}$ or $\ensuremath{\rds}$ can be empty. %To establish a happens-before relation between $\atr_{\instr}$ and $\atr$ in $\atrace_{\instr}$, we look whether a transaction can extend a happens-before started by $\atr_{\instr}$ (elements of $\beta \cdot \atr_{0} \cdot \gamma$). In order for a transaction $\atr_1$ to extend a happens-before relation, it has to satisfy one of the following conditions: % \begin{itemize}[topsep=0pt] % \item the transaction is from a process that has already another transaction in the happens-before. Thus, we ensure the continuity of the happens-before relation through $\po$ relation. % \item the transaction is reading from a variable that was written to by a previous transaction in the happens-before. Hence, we ensure the continuity of the happens-before relation through $\rfo$ relation. % \item the transaction writes to a variable that was written to by a previous transaction in the happens-before. Hence, we ensure the continuity of the happens-before relation through $\sto$ relation. % \item the transaction writes to a variable that was read by a previous transaction in the happens-before. Hence, we ensure the continuity of the happens-before relation through $\cfo$ relation. %Furthermore, as per Theorem \ref{corol:pcsi}, we must ensure that in the happens-before relation between $\atr_{\instr}$ and $\atr$ we don't have two successive $\cfo$ relations. %Therefore, we use two variables $\wrSet$ and $\rdSet$ to track the variables read and written by transactions in $\beta \cdot \atr_{0} \cdot \gamma$. %Also, for each variable, we introduce two flags $\wraddr{\anaddr}$ and $\rdaddr{\anaddr}$ for each variable $\anaddr$ to say that % \item if $\wraddr{\anaddr} = 0$ (resp., $\rdaddr{\anaddr} = 0$) there exists a transaction $\atr_1$ in the happens-before that writes to (resp., reads from) $\anaddr$ such that the happens-before relation between $\atr_{\instr}$ and $\atr_1$ doesn't contain two successive $\cfo$ relations and the last dependency relation in this happens-before is not $\cfo$. % \item if $\wraddr{\anaddr} = 1$ (resp., $\rdaddr{\anaddr} = 1$) there exists a transaction $\atr_1$ in the happens-before that writes to (resp., reads from) $\anaddr$ such that the happens-before relation between $\atr_{\instr}$ and $\atr_1$ doesn't contain two successive $\cfo$ relations, however, the last dependency relation in this happens-before is $\cfo$. % \item if $\wraddr{\anaddr} = 2$ (resp., $\rdaddr{\anaddr} = 2$) there exists a transaction $\atr_1$ in the happens-before that writes to (resp., reads from) $\anaddr$ such that the happens-before relation between $\atr_{\instr}$ and $\atr_1$ contains two successive $\cfo$ relations. %Similarly, we introduce a local flag $\rwMap{\apr}$ for each process $\apr$ to say that % \item if $\rwMap{\apr} = 0$, there exists a transaction $\atr_1$ from $\apr$ in the happens-before such that the happens-before relation between $\atr_{\instr}$ and $\atr_1$ doesn't contain two successive $\cfo$ relations and the last dependency relation in this happens-before is not $\cfo$. % \item if $\rwMap{\apr} = 1$, every transaction $\atr_1$ from $\apr$ in the happens-before where the happens-before relation between $\atr_{\instr}$ and $\atr_1$ doesn't contain two successive $\cfo$ relations, we have that the last dependency relation in this happens-before is $\cfo$. % \item if $\rwMap{\apr} = 2$, for every transaction $\atr_1$ from $\apr$ in the happens-before, we have that the happens-before relation between $\atr_{\instr}$ and $\atr_1$ contains two successive $\cfo$ relations. %Each transaction $\atr_1$ which is trying to join the happens-before is equipped with a local flag $\rw$ %that is initialized to null ($\perp$) at the start of $\atr_1$ and must contain a value that's different than null %when reaching the end of the transaction (meaning that $\atr_1$ satisfied one of the four conditions required in order to join the happens-before) such that % \item $\rw = 0$ means that there exists a happens-before relation between $\atr_{\instr}$ and $\atr_1$ that doesn't contain two successive $\cfo$ relations and the last dependency relation in this happens-before is not $\cfo$. % \item $\rw = 1$ means that in every happens-before relation between $\atr_{\instr}$ and $\atr_1$ that doesn't contain two successive $\cfo$ relations we have that the last dependency relation in this happens-before is $\cfo$. % \item $\rw = 2$ means that every happens-before relation between $\atr_{\instr}$ and $\atr_1$ contains two successive $\cfo$ relations. %Otherwise, the execution is blocked. %Also, note that at the start of the execution, all flags are initialized to null. %In general, whether a transaction is splitted and executed without instrumentation $\semidlerpcinstr{\atr}$, or instrumented as in $\semidlermove{\atr}$, or instrumented as in $\semidler{\atr}$, or splitted and each of the resulting transactions is instrumented as in $\semidler{\semidlerpcinstr{\atr}}$ is set non-deterministically and can vary from execution to execution. %In the instrumentation $\semidler{\atr_1}$ of a transaction $\atr_1$ in the happens before, we check whether we reached an error state if $\atr_1$ writes to the variable that was stored in $a_{\wrflag}$ and $\rw \neq 2$. %The final instrumentation of a given program $\aprog$, denoted by $\aprog_\siinstr$, is obtained by replacing each transaction $\atr$ with the concatenation of the four possible instrumentations, i.e., %\atr_\siinstr ::= \semidlerpcinstr{\atr} \hspace{0.17cm} \semidlermove{\atr} \hspace{0.17cm} \semidler{\atr} \hspace{0.17cm} \semidler{\semidlerpcinstr{\atr}} The correctness of the instrumentation is stated in the following theorem. \begin{theorem}\label{them:RobPcSiInstr} A program $\aprog$ is robust against \pcc{} relative to \sic{} iff the instrumentation in Figure~\ref{Figure:Instr} does not violate an assertion when executed under \serc{}. \end{theorem} %We give the proof of Theorem \ref{them:RobPcSiInstr} in the supplementary materials. Theorem~\ref{them:RobPcSiInstr} implies the following complexity result for finite-state programs. The lower bound is proved similarly to the case \ccc{} vs \pcc{}. %Similar to before, we obtain an upper bound of the robustness problem against \pcc{} relative to \sic{} based on the reachability problem under \serc{}. For the lower bound, we construct a program $\aprog'$ from $\aprog$ in the same manner as before just instead of the $\mathsf{SB}$ program, we use the $\mathsf{LU}$ program. Then, we obtain that $\aprog$ reaches a state $\astate$ iff $\aprog'$ is not robust against \pcc{} relative to \sic{}. \begin{corollary}\label{corol:SIRobcomplexity} Checking robustness of a program with a fixed number of variables and bounded data domain against \pcc{} relative to \sic{} is PSPACE-complete when the number of processes is bounded and EXPSPACE-complete, otherwise. % Checking robustness of a program with a fixed number of variables and bounded data domain against \pcc{} relative to \sic{} is PSPACE-complete when the number of processes is fixed and EXPSPACE-complete, otherwise. \end{corollary} Checking robustness against \ccc{} relative to \sic{} can be also shown to be reducible (in polynomial time) to a reachability problem under \serc{} by combining the results of checking robustness against \ccc{} relative to \pcc{} and \pcc{} relative to \sic{}. \begin{theorem} \label{them:RobCcSi} A program $\aprog$ is robust against \ccc{} relative to \sic{} iff $\aprog$ is robust against \ccc{} relative to \pcc{} and $\aprog$ is robust against \pcc{} relative to \sic{}. \end{theorem} \begin{remark} \label{rem:robustness} Our reductions of robustness checking to reachability apply to an extension of our programming language where the number of processes is unbounded and each process can execute an arbitrary number of times a statically known set of transactions. This holds because the instrumentation in Figure~\ref{Figure:Instr} and the one in~[10] (for the case \ccc{} vs. \serc{}) consist in adding a set of instructions that manipulate a fixed set of process-local or shared variables, which do not store process or transaction identifiers. These reductions extend also to SQL queries that access unbounded size tables. Rows in a table can be interpreted as memory locations (identified by primary keys in unbounded domains, e.g., integers), and SQL queries can be interpreted as instructions that read/write a set of locations in one shot. These possibly unbounded sets of locations can be represented symbolically using the conditions in the SQL queries (e.g., the condition in the WHERE part of a SELECT). The instrumentation in Figure 6 needs to be adapted so that read and write sets are updated by adding sets of locations for a given instruction (represented symbolically as mentioned above). \end{remark} % \begin{multicols}{2} % \begin{algorithmic}[1] % %\REQUIRE $n \geq 0 \vee x \neq 0$ % %\ENSURE $y = x^n$ % \STATE $y \Leftarrow 1$ % %\IF{$\hb{} = \perp$} % %\STATE $X \Leftarrow 1 / x$ % %\STATE $N \Leftarrow -n$ % %\ELSE % %\STATE $X \Leftarrow x$ % %\STATE $N \Leftarrow n$ % %\ENDIF % %\WHILE{$N \neq 0$} % %\IF{$N$ is even} % %\STATE $X \Leftarrow X \times X$ % %\STATE $N \Leftarrow N / 2$ % %\ELSE[$N$ is odd] % %\STATE $y \Leftarrow y \times X$ % %\STATE $N \Leftarrow N - 1$ % %\ENDIF % %\ENDWHILE % \end{algorithmic} % \end{multicols} % \caption{x} % \label{alg1} %!TEX root = draft.tex \vspace{-15pt} \section{Proving Robustness Using Commutativity Dependency Graphs} \label{sec:commutativitygraph} We describe an approximated technique for proving robustness, which leverages the concept of left/right mover in Lipton's reduction theory~[38]. This technique reasons on the \emph{commutativity dependency graph}~[9] associated to the transformation $_$ of an input program $$ that allows to simulate the \pcc{} semantics under serializability (we use a slight variation of the original definition of this class of graphs). We characterize robustness against \ccc{} relative to \pcc{} and \pcc{} relative to \sic{} in terms of certain properties that (simple) cycles in this graph must satisfy. We recall the concept of movers and the definition of commutativity dependency graphs. Given a program $$ and a trace $= _1·…·_n ∈_()$ of $$ under serializability, we say that $_i ∈$ \emph{moves right (resp., left)} in $$ if $_1·…·_i-1·_i+1·_i·_i+2·…·_n$ (resp., $_1·…·_i-2·_i·_i-1·_i+1·…·_n$) is also a valid execution of $$, $_i$ and $_i+1$ (resp., $_i-1$) are executed by distinct processes, and both traces reach the same end state. A transaction $∈$ is not a right (resp., left) mover iff there exists a trace $∈_()$ such that $∈$ and $$ doesn't move right (resp., left) in $$. Thus, when a transaction $$ is \emph{not} a right mover then there must exist another transaction $' ∈$ which caused $$ to not be permutable to the right (while preserving the end state). Since $$ and $'$ do not commute, then this must be because of either a write-read, write-write, or a read-write dependency relation between the two transactions. We say that $$ is not a right mover because of $'$ and a dependency relation that is either write-read, write-write, or read-write. Notice that when $$ is not a right mover because of $'$ then $'$ is not a left mover because of $$. We define $$ as a binary relation between transactions such that $(,') ∈$ when $$ is \emph{not} a right mover because of $'$ and a write-read dependency ($t'$ reads some value written by $t$). We define the relations $$ and $$ corresponding to write-write and read-write dependencies in a similar way. %We denote $(\atr,\atr') \in \msto$ when $\atr$ is not a right mover because $\atr'$ and write-write dependency. Similarly, we denote $(\atr,\atr') \in \mcfo$ when $\atr$ is not a right mover because $\atr'$ and a read-write dependency between the two transactions. We call $$, $$, and $$, \emph{non-mover} relations. The \emph{commutativity dependency graph} of a program $$ is a graph where vertices represent transactions in $$. Two vertices are linked by a program order edge if the two transactions are executed by the same process. The other edges in this graph represent the ``non-mover'' relations $$, $$, and $$. Two vertices that represent the two components $$ and $$ of the same transaction $$ (already linked by $$ edge) are also linked by an undirected edge labeled by $$ (same-transaction relation). %To model an arbitrary instantiation of the same transaction $\atr$, we take a symbolic integer $m > 0$ and add a sequence of $m$ vertices representing $m$ instantiations of $\atr$ in the commutativity graph. We relate every two successive vertices by a program order edge. Also, similar to before for every vertex in the sequence, we draw the incoming and outgoing edges using the ``non-mover'' relations $\mrfo$, $\msto$, and $\mcfo$. \begin{wrapfigure}{r}{0.53\textwidth} \vspace{-25pt} \lstset{basicstyle=\ttfamily\scriptsize} \centering \resizebox{!}{2.2cm}{ \begin{tikzpicture}[->,>=stealth',shorten >=1pt,auto,node distance=4cm, semithick, transform shape,every text node part/.style={align=left}] \node[shape=rectangle ,draw=none,font=\large, label={left:$\atrwr{\atr 1}$}] at (-3.2,0) (m) {$[x = 1]$}; \node[shape=rectangle ,draw=none,font=\large, label={left:$\atrwr{\atr 2}$}] at (-3.2,-2) (m1) {$[y = 1]$}; \node[shape=rectangle ,draw=none,font=\large, label={right:$\atrrd{\atr 3}$}] at (0,0) (p){$[r1 = y]$}; \node[shape=rectangle ,draw=none,font=\large, label={right:$\atrrd{\atr 4}$}] at (0,-2) (p1){$[r2 = x]$}; \begin{scope}[ every edge/.style={draw=black,very thick}] \path[->] (m) edge[left] node {$\po$} (m1); \path[->] (p1) edge[bend left, above] node[xshift=2.9mm,yshift=-1.3mm] {$\mcfo$} (m); \path[->] (m) edge[bend left, below] node {$\mrfo$} (p1); \path[->] (m1) edge[bend left, below] node[xshift=2.5mm,yshift=0.5mm] {$\mrfo$} (p); \path[->] (p) edge[bend left, above] node[xshift=-1.2mm,yshift=0.3mm] {$\mcfo$} (m1); \path[->] (p) edge[right] node {$\po$} (p1); \end{scope} \end{tikzpicture}} \vspace{-3pt} \caption{The commutativity dependency graph of the $\mathsf{MP}_{\pcinstr}$ program.} \label{fig:litmus4CDG} \vspace{-15pt} \end{wrapfigure} Our results about the robustness of a program $$ are stated over a slight variation of the commutativity dependency graph of $_$ (where a transaction is either read-only or write-only). This graph contains additional undirected edges that link every pair of transactions $$ and $$ of $_$ that were originally components of the same transaction $$ in $$. Given such a commutativity dependency graph, the robustness of $$ is implied by the absence of cycles of specific shapes. These cycles can be seen as an abstraction of potential robustness violations for the respective semantics (see Theorem~\ref{them:MovRobCcPc} and Theorem~\ref{them:MovRobPcSi}). % we ask whether this graph has cycles of some specific shapes based the characterization of robustness violation traces for the respective semantics. %The corresponding program is robust when the graph doesn't have these types of cycles Figure \ref{fig:litmus4CDG} pictures the commutativity dependency graph for the $𝖬𝖯$ program. Since every transaction in $𝖬𝖯$ is singleton, the two programs $𝖬𝖯$ and $𝖬𝖯_$ coincide. Using the characterization of robustness violations against \ccc{} relative to \serc{} from~[10] and the reduction in Theorem~\ref{them:RobCcPc}, we obtain the following result concerning the robustness against \ccc{} relative to \pcc{}. \begin{theorem} \label{them:MovRobCcPc} Given a program $\aprog$, if the commutativity dependency graph of the program $\aprog_\pcinstr$ does not contain a simple cycle formed by $\atr_1$ $\cdots$ $\atr_i$ $\cdots$ $\atr_n$ such that: \begin{itemize}[topsep=3pt] \item $(\atr_n,\atr_1) \in \mcfo$; \item $(\atr_j, \atr_{j+1}) \in (\po \cup \rfo)^{*}$, for $j \in [1,i-1]$; \item $(\atr_i,\atr_{i+1}) \in (\mcfo \cup \msto)$; \item $(\atr_j,\atr_{j+1}) \in (\mcfo \cup \msto \cup \mrfo \cup \po)$, for $j \in [i+1,n-1]$. \end{itemize} then $\aprog$ is robust against \ccc{} relative to \pcc{}. \end{theorem} \begin{comment} \begin{proof} It is enough to show: if $\aprog$ is not robust against \ccc{} relative to \pcc{} then we have a simple cycle in the commutativity dependency graph of $\aprog_\pcinstr$ of the form above. Assume $\aprog$ is not robust against \ccc{} relative to \pcc{}. Then, from Theorem \ref{them:RobCcPc}, we obtain $\aprog_\pcinstr$ is not robust against \ccc{} relative to \serc{}. Also it was shown in [10] that if a program is not robust then there must exist a robustness violation trace (\ccc{} relative to \serc{}) $\atrace_\pcinstr$ of the shape $\atrace_\pcinstr = \alpha \cdot \atr_1 \cdot \beta \cdot \atr_i \cdot \atr_{i+1} \cdot \gamma \cdot \atr_n$ where $(\atr_1,\atr_i) \in (\po \cup \rfo)^{+}$, $(\atr_{i},\atr_{i+1}) \in (\sto \cup \cfo)$, $(\atr_{i+1},\atr_n) \in \hbo$, and $(\atr_n,\atr_1) \in \cfo$. Note that since transactions in the trace $\atrace_\pcinstr$ can either be read-only or write-only. Then, $(\atr_{i},\atr_{i+1}) \in (\sto \cup \cfo)$ and $(\atr_{n},\atr_1) \in \cfo$ imply that $\atr_1$ and $\atr_{i+1}$ must be a write-only transactions and $\atr_{n}$ must be a read-only transaction. Note that we may have $\beta = \gamma = \epsilon$ as the case for the trace of the $\mathsf{SB}$ program given in Figure \ref{fig:litmus1}. %In the trace $\atrace_\pcinstr$, we let $\atr_1$ to be $\atr_1$, $\atr_i$ to be $\atr_2$, $\atr_3$ to be $\atr_{i+1}$, and $\atr_4$ to be $\atr_n$ of Theorem \ref{them:MovRobCcPc}. We consider first the general case when $\atr_1 \not\equiv \atr_2$. The other case can be proved in the same way. Consider the prefix $\atrace_{p}$ of $\atrace_\pcinstr$: $\atrace_{p} = \alpha \cdot \atr_1 \cdot \beta \cdot \atr_{i}$ where $(\atr_1,\atr_{i}) \in (\po \cup \rfo)^{+}$ which is a \serc{} trace of $\aprog_{\pcinstr}$. Then, we have a sequence of transactions from $\atr_1$ to $\atr_{i}$ that are related by either $\po$ or $\rfo$. In the case two transactions are only related by $\rfo$, then the first transaction is not a right mover because of the second transaction reads from a write in the first transaction. Thus, we can relate the two transactions using the relation $\mrfo$ in the commutativity dependency graph. Similarly consider the following trace $\atrace_{s}$ extracted from $\atrace_\pcinstr$: $\atrace _{s} = \alpha \cdot \atr_{i+1} \cdot \gamma \cdot \atr_n$ where $(\atr_{i+1},\atr_n) \in \hbo$ which is a \serc{} trace of $\aprog_{\pcinstr}$. Similar to before, we have a sequence of transactions from $\atr_{i+1}$ to $\atr_n$ that are related by either $\po$, $\rfo$, $\sto$, or $\cfo$. For any two transactions that are related only by either $\rfo$, $\sto$, or $\cfo$, this implies that the first transaction is not a right mover because of the second transaction and a write-read, write-write, or read-write dependency between the two, respectively. Thus, we can relate the two transactions using either $\mrfo$, $\msto$, or $\mcfo$, respectively. Now consider the following trace $\atrace_{1}$ extracted from $\atrace_\pcinstr$: $\atrace_{1} = \alpha \cdot \atr_1 \cdot \beta \cdot \atr_{i} \cdot \atr_{i+1}$ where $(\atr_{i},\atr_{i+1}) \in (\sto \cup \cfo)$ is a \serc{} trace of $\aprog_{\pcinstr}$. Because $\atr_{i}$ and $\atr_{i+1}$ are related by either $\sto$ or $\cfo$, then $\atr_{i}$ is not a right mover because of $\atr_{i+1}$ and a write-write or read-write dependency between the two, respectively. Thus, we can relate the two transactions using either $\msto$ or $\mcfo$, respectively. Finally, consider the following trace $\atrace_{2}$ extracted from $\atrace_\pcinstr$: $\atrace_{2} = \alpha \cdot \atr_{i+1} \cdot \gamma \cdot \atr_n \cdot \atr_1$ where $(\atr_n,\atr_1) \in \cfo$ is a \serc{} trace of $\aprog_{\pcinstr}$. Because $\atr_{n}$ and $\atr_{1}$ are related by $\cfo$, then $\atr_{n}$ is not a right mover because of $\atr_{1}$ and a read-write dependency between the two. Thus, we can relate the two transactions using $\mcfo$. \end{proof} \end{comment} Next we give the characterization of commutativity dependency graphs required for proving robustness against \pcc{} relative to \sic{}. \begin{theorem} \label{them:MovRobPcSi} Given a program $\aprog$, if the commutativity dependency graph of the program $\aprog_\pcinstr$ does not contain a simple cycle formed by $\atr_1$ $\cdots$ $\atr_n$ such that: \begin{itemize}[topsep=3pt] \item $(\atr_n,\atr_1) \in \msto$, $(\atr_1,\atr_2) \in \sametro$, and $(\atr_2,\atr_3) \in \mcfo$; \item $(\atr_j,\atr_{j+1}) \in (\mcfo \cup \msto \cup \mrfo \cup \po \cup \sametro)^{*}$, for $j \in [3,n-1]$; \item $\forall\ j \in [2,n-2].$ \begin{itemize} \item $\mbox{if }(\atr_j,\atr_{j+1}) \in \mcfo\mbox{ then }(\atr_{j+1},\atr_{j+2}) \in (\mrfo \cup \po \cup \msto)$; \item $\mbox{if }(\atr_{j+1},\atr_{j+2}) \in \mcfo\mbox{ then }(\atr_{j},\atr_{j+1}) \in (\mrfo \cup \po)$. \end{itemize} \item $\forall\ j \in [3,n-3]. \mbox{ if }(\atr_{j+1},\atr_{j+2}) \in \sametro\mbox{ and }(\atr_{j+2},\atr_{j+3}) \in \mcfo \mbox{ then }(\atr_{j},\atr_{j+1}) \in \msto$. \end{itemize} then $\aprog$ is robust against \pcc{} relative to \sic{}. \end{theorem} \begin{comment} \begin{proof} Similar to before it is enough to show: if $\aprog$ is not robust against \pcc{} relative to \sic{} then we have a simple cycle in the commutativity dependency graph of $\aprog_\pcinstr$ of the form above. Assume $\aprog$ is not robust against \pcc{} relative to \sic{}. Then, from Theorem \ref{them:RobPcSiInstr}, we obtain that if $\sem{\aprog}$ reaches an error state under \serc{} then we will have the following trace $\atrace$ under \serc: $\atrace = \alpha \cdot \atrrd{\atr_{\instr}} \cdot \atr_3 \cdot \beta \cdot \atr_n \cdot \atrwr{\atr_{\instr}}$\footnote{For simplicity, we assume here that after reaching the error state we execute the writes of $\atr_{\instr}$, i.e., $\atrwr{\atr_{\instr}}$} where $(\atrrd{\atr_{\instr}},\atr_3) \in \cfo$, $(\atr_3,\atr_n) \in \hbo$, $(\atr_n,\atrwr{\atr_{\instr}}) \in \sto$, and we don't have two successive $\cfo$ in the happens before between $\atr_3$ and $\atr_n$. In $\atrace$, $\atrwr{\atr_{\instr}}$ (resp., $\atrrd{\atr_{\instr}}$) represents $\atr_1$ (resp., $\atr_2$) in the theorem statement. Note that we may have $\alpha = \beta = \epsilon$ as is the case of the transformed $\mathsf{LU}$ program given in Figure \ref{fig:litmus2Instr}. The construction of the cycle in the commutativity dependency graph follows the same procedure taken in the proof of Theorem \ref{them:MovRobCcPc}. The only difference is that for every two transactions of $\atrace$ that are part of the happens before between $\atr_3$ and $\atr_n$, if the two are not connected by either $\po$, $\rfo$, $\sto$, or $\cfo$ then they must be the reads and writes of the same original transaction in $\aprog$. In this case, in the commutativity dependency graph we have the two transactions related by $\sametro$. %Notice that we abused notation by using the two components $\atrrd{\atr 1}$ and $\atrwr{\atr_1}$ of $\atr_1$ in $\atrace$ to denote that $\atr 1$ writes were not immediately written to the shared variables. \end{proof} \end{comment} In Figure \ref{fig:litmus4CDG}, we have three simple cycles in the graph: \begin{itemize}[topsep=3pt] \item $(\atrwr{\atr 1}, \atrrd{\atr 4}) \in \mrfo$ and $(\atrrd{\atr 4}, \atrwr{\atr 1}) \in \mcfo$, \item $(\atrwr{\atr 2}, \atrrd{\atr 3}) \in \mrfo$ and $(\atrrd{\atr 3}, \atrwr{\atr 2}) \in \mcfo$, \item $(\atrwr{\atr 1}, \atrwr{\atr 2}) \in \po$, $(\atrwr{\atr 2}, \atrrd{\atr 3}) \in \mrfo$, $(\atrrd{\atr 3}, \atrrd{\atr 4}) \in \po$, and $(\atrrd{\atr 4}, \atrwr{\atr 1}) \in \mcfo$. \end{itemize} Notice that none of the cycles satisfies the properties in Theorems \ref{them:MovRobCcPc} and \ref{them:MovRobPcSi}. Therefore, $𝖬𝖯$ is robust against \ccc{} relative to \pcc{} and against \pcc{} relative to \sic{}. \begin{remark} \label{rem:comgraph} For programs that contain an unbounded number of processes, an unbounded number of instantiations of a fixed number of process ``templates'', or unbounded loops with bodies that contain entire transactions, a sound robustness check consists in applying Theorem~\ref{them:MovRobCcPc} and Theorem~\ref{them:MovRobPcSi} to (bounded) programs that contain two copies of each process template, and where each loop is unfolded exactly two times. This holds because the mover relations are ``static'', they do not depend on the context in which the transactions execute, and each cycle requiring more than two process instances or more than two loop iterations can be short-circuited to a cycle that exists also in the bounded program. Every outgoing edge from a third instance/iteration can also be taken from the second instance/iteration. Two copies/iterations are necessary in order to discover cycles between instances of the same transaction (the cycles in Theorem~\ref{them:MovRobCcPc} and Theorem~\ref{them:MovRobPcSi} are simple and cannot contain the same transaction twice). These results extend easily to SQL queries as well because the notion of mover is independent of particular classes of programs or instructions. %Note that the notion of mover used in commutativity dependency graphs is quite generic and independent of particular classes of programs or instructions. For instance, a mover check between two transactions that access sets of locations defined using symbolic expressions corresponds to checking whether the conjunction of the two symbolic expressions is satisfiable. \end{remark} \input{experiments.tex} %!TEX root = draft.tex \vspace{-5pt} \section{Related Work} \label{sec:related} \vspace{-5pt} The consistency models in this paper were studied in several recent works~[20, 19, 24, 42, 15, 43, 13]. Most of them focused on their operational and axiomatic formalizations. % of programs semantics under the weak consistency models both operationally and declaratively. The formal definitions we use in this paper are based on those given in~[24, 15]. Biswas and Enea~[13] shows that checking whether an execution is \ccc{} is polynomial time while checking whether it is \pcc{} or \sic{} is NP-complete. %In this paper we tackle trace-based robustness problem. In the literature, the decidability and complexity of trace-based The robustness problem we study in this paper has been investigated in the context of weak memory models, but only relative to sequential consistency, against Release/Aquire (RA), TSO and Power~[35, 16, 14, 28]. Checking robustness against \ccc{} and \sic{} relative to \serc{} has been investigated in~[9, 10]. %All of these work study the robustness between a weak consistency model (e.g., RA, TSO, or \ccc{}) and the strong semantics model, i.e., sequential consistency and serialisability. %On the other hand, In this work, we study the robustness problem between two weak consistency models, which poses different non-trivial challenges. In particular, previous work proposed reductions to reachability under sequential consistency (or \serc{}) that relied on a concept of minimal robustness violations (w.r.t. an operational semantics), which does not apply in our case. % an important technique that was commonly used in previous work which consists of using borderline violations %cannot be applied here. Since, these violations defined such that removing the last action in the violation results in serialisable execution which is not valid in our case. %However, we have reduced the robustness against \ccc{} relative to \pcc{} to the robustness against \ccc{} relative to \serc{} which allowed us to use the results that were proved in [12, 10]. %Note that our reduction from \pcc{} to \serc{} is similar in spirit to the one shown in [43], however, %in our case execution traces include the store order dependency relation in order to construct the happens before relation. The relationship between \pcc{} and \serc{} is similar in spirit to the one given by Biswas and Enea~[13] in the context of checking whether an execution is \pcc{}. However, that relationship was proven in the context of a ``weaker'' notion of trace (containing only program order and read-from), and it does not extend to our notion of trace. For instance, that result does not imply preserving $$ dependencies which is crucial in our case. Some works describe various over- or under-approximate analyses for checking robustness relative to \serc{}. The works in~[12, 17, 18, 25, 39] propose static analysis techniques based on computing an abstraction of the set of computations, which is used for proving robustness. In particular, [18, 39] encode program executions under the weak consistency model using FOL formulas to describe the dependency relations between actions in the executions. These approaches may return false alarms due to the abstractions they consider in their encoding. Note that in this paper, we prove a strengthening of the results of [12] with regard to the shape of happens before cycles allowed under \pcc{}. An alternative to {\em trace-based} robustness, is {\em state-based} robustness which requires that a program is robust if the sets of reachable states under two semantics coincide. While state-robustness is the necessary and sufficient concept for preserving state-invariants, its verification, which amounts in computing the set of reachable states under the weak semantics models is in general a hard problem. The decidability and the complexity of this problem has been investigated in the context of relaxed memory models such as TSO and Power, and it has been shown that it is either decidable but highly complex (non-primitive recursive), or undecidable [5, 6]. Automatic procedures for approximate reachability/invariant checking have been proposed using either abstractions or bounded analyses, e.g., [7, 4, 27, 1]. Proof methods have also been developed for verifying invariants in the context of weakly consistent models such as [36, 31, 40, 3]. These methods, however, do not provide decision procedures. % \section{Conclusion} \label{sec:conclusion} We have proposed a reduction of the problem of checking robustness between weak consistency models to reachability that increases the size of the program only linearly, and therefore, we showed that checking robustness is in principle as hard as checking reachability. Then, we gave a pragmatic technique for proving robustness based on the notion of non-mover relations that can be constructed automatically using SMT solvers. We tested our techniques on realistic programs that model the most intricate parts of distributed applications that are obtained from the standard OLTP Benchmark and open source Github projects. \bibliographystyle{splncs04} \begin{thebibliography}{10} \providecommand{\url}[1]{\texttt{#1}} \providecommand{\urlprefix}{URL } \providecommand{\doi}[1]{https://doi.org/#1} [1] Abdulla, P.A., Atig, M.F., Bouajjani, A., Ngo, T.P.: Context-bounded analysis for {POWER}. In: Legay, A., Margaria, T. (eds.) Tools and Algorithms for the Construction and Analysis of Systems - 23rd International Conference, {TACAS} 2017, Held as Part of the European Joint Conferences on Theory and Practice of Software, {ETAPS} 2017, Uppsala, Sweden, April 22-29, 2017, Proceedings, Part {II}. Lecture Notes in Computer Science, vol. 10206, pp. 56--74 (2017). \doi{10.1007/978-3-662-54580-5\_4}, \url{https://doi.org/10.1007/978-3-662-54580-5\_4} [2] Adya, A.: Weak consistency: A generalized theory and optimistic implementations for distributed transactions. Ph.D. thesis (1999) [3] Alglave, J., Cousot, P.: Ogre and pythia: an invariance proof method for weak consistency models. In: Castagna, G., Gordon, A.D. (eds.) Proceedings of the 44th {ACM} {SIGPLAN} Symposium on Principles of Programming Languages, {POPL} 2017, Paris, France, January 18-20, 2017. pp. 3--18. {ACM} (2017), \url{http://dl.acm.org/citation.cfm?id=3009883} [4] Alglave, J., Kroening, D., Tautschnig, M.: Partial orders for efficient bounded model checking of concurrent software. In: Sharygina, N., Veith, H. (eds.) Computer Aided Verification - 25th International Conference, {CAV} 2013, Saint Petersburg, Russia, July 13-19, 2013. Proceedings. Lecture Notes in Computer Science, vol.~8044, pp. 141--157. Springer (2013). \doi{10.1007/978-3-642-39799-8\_9}, \url{https://doi.org/10.1007/978-3-642-39799-8\_9} [5] Atig, M.F., Bouajjani, A., Burckhardt, S., Musuvathi, M.: On the verification problem for weak memory models. In: Hermenegildo, M.V., Palsberg, J. (eds.) Proceedings of the 37th {ACM} {SIGPLAN-SIGACT} Symposium on Principles of Programming Languages, {POPL} 2010, Madrid, Spain, January 17-23, 2010. pp. 7--18. {ACM} (2010). \doi{10.1145/1706299.1706303}, \url{https://doi.org/10.1145/1706299.1706303} [6] Atig, M.F., Bouajjani, A., Burckhardt, S., Musuvathi, M.: What's decidable about weak memory models? In: Seidl, H. (ed.) Programming Languages and Systems - 21st European Symposium on Programming, {ESOP} 2012, Held as Part of the European Joint Conferences on Theory and Practice of Software, {ETAPS} 2012, Tallinn, Estonia, March 24 - April 1, 2012. Proceedings. Lecture Notes in Computer Science, vol.~7211, pp. 26--46. Springer (2012). \doi{10.1007/978-3-642-28869-2\_2}, \url{https://doi.org/10.1007/978-3-642-28869-2\_2} [7] Atig, M.F., Bouajjani, A., Parlato, G.: Getting rid of store-buffers in {TSO} analysis. In: Gopalakrishnan, G., Qadeer, S. (eds.) Computer Aided Verification - 23rd International Conference, {CAV} 2011, Snowbird, UT, USA, July 14-20, 2011. Proceedings. Lecture Notes in Computer Science, vol.~6806, pp. 99--115. Springer (2011). \doi{10.1007/978-3-642-22110-1\_9}, \url{https://doi.org/10.1007/978-3-642-22110-1\_9} [8] Barnett, M., Chang, B.E., DeLine, R., Jacobs, B., Leino, K.R.M.: Boogie: {A} modular reusable verifier for object-oriented programs. In: de~Boer, F.S., Bonsangue, M.M., Graf, S., de~Roever, W.P. (eds.) Formal Methods for Components and Objects, 4th International Symposium, {FMCO} 2005, Amsterdam, The Netherlands, November 1-4, 2005, Revised Lectures. Lecture Notes in Computer Science, vol.~4111, pp. 364--387. Springer (2005). \doi{10.1007/11804192\_17}, \url{https://doi.org/10.1007/11804192\_17} [9] Beillahi, S.M., Bouajjani, A., Enea, C.: Checking robustness against snapshot isolation. In: Dillig, I., Tasiran, S. (eds.) Computer Aided Verification - 31st International Conference, {CAV} 2019, New York City, NY, USA, July 15-18, 2019, Proceedings, Part {II}. Lecture Notes in Computer Science, vol. 11562, pp. 286--304. Springer (2019). \doi{10.1007/978-3-030-25543-5\_17}, \url{https://doi.org/10.1007/978-3-030-25543-5\_17} [10] Beillahi, S.M., Bouajjani, A., Enea, C.: Robustness against transactional causal consistency. In: Fokkink, W.J., van Glabbeek, R. (eds.) 30th International Conference on Concurrency Theory, {CONCUR} 2019, August 27-30, 2019, Amsterdam, the Netherlands. LIPIcs, vol.~140, pp. 30:1--30:18. Schloss Dagstuhl - Leibniz-Zentrum f{\"{u}}r Informatik (2019). \doi{10.4230/LIPIcs.CONCUR.2019.30}, \url{https://doi.org/10.4230/LIPIcs.CONCUR.2019.30} [11] Berenson, H., Bernstein, P.A., Gray, J., Melton, J., O'Neil, E.J., O'Neil, P.E.: A critique of {ANSI} {SQL} isolation levels. In: Carey, M.J., Schneider, D.A. (eds.) Proceedings of the 1995 {ACM} {SIGMOD} International Conference on Management of Data, San Jose, California, USA, May 22-25, 1995. pp. 1--10. {ACM} Press (1995). \doi{10.1145/223784.223785}, \url{https://doi.org/10.1145/223784.223785} [12] Bernardi, G., Gotsman, A.: Robustness against consistency models with atomic visibility. In: Desharnais, J., Jagadeesan, R. (eds.) 27th International Conference on Concurrency Theory, {CONCUR} 2016, August 23-26, 2016, Qu{\'{e}}bec City, Canada. LIPIcs, vol.~59, pp. 7:1--7:15. Schloss Dagstuhl - Leibniz-Zentrum f{\"{u}}r Informatik (2016). \doi{10.4230/LIPIcs.CONCUR.2016.7}, \url{https://doi.org/10.4230/LIPIcs.CONCUR.2016.7} [13] Biswas, R., Enea, C.: On the complexity of checking transactional consistency. Proc. {ACM} Program. Lang. \textbf{3}({OOPSLA}), 165:1--165:28 (2019). \doi{10.1145/3360591}, \url{https://doi.org/10.1145/3360591} [14] Bouajjani, A., Derevenetc, E., Meyer, R.: Checking and enforcing robustness against {TSO}. In: Felleisen, M., Gardner, P. (eds.) Programming Languages and Systems - 22nd European Symposium on Programming, {ESOP} 2013, Held as Part of the European Joint Conferences on Theory and Practice of Software, {ETAPS} 2013, Rome, Italy, March 16-24, 2013. Proceedings. Lecture Notes in Computer Science, vol.~7792, pp. 533--553. Springer (2013). \doi{10.1007/978-3-642-37036-6\_29}, \url{https://doi.org/10.1007/978-3-642-37036-6\_29} [15] Bouajjani, A., Enea, C., Guerraoui, R., Hamza, J.: On verifying causal consistency. In: Castagna, G., Gordon, A.D. (eds.) Proceedings of the 44th {ACM} {SIGPLAN} Symposium on Principles of Programming Languages, {POPL} 2017, Paris, France, January 18-20, 2017. pp. 626--638. {ACM} (2017), \url{http://dl.acm.org/citation.cfm?id=3009888} [16] Bouajjani, A., Meyer, R., M{\"{o}}hlmann, E.: Deciding robustness against total store ordering. In: Aceto, L., Henzinger, M., Sgall, J. (eds.) Automata, Languages and Programming - 38th International Colloquium, {ICALP} 2011, Zurich, Switzerland, July 4-8, 2011, Proceedings, Part {II}. Lecture Notes in Computer Science, vol.~6756, pp. 428--440. Springer (2011). \doi{10.1007/978-3-642-22012-8\_34}, \url{https://doi.org/10.1007/978-3-642-22012-8\_34} [17] Brutschy, L., Dimitrov, D., M{\"{u}}ller, P., Vechev, M.T.: Serializability for eventual consistency: criterion, analysis, and applications. In: Castagna, G., Gordon, A.D. (eds.) Proceedings of the 44th {ACM} {SIGPLAN} Symposium on Principles of Programming Languages, {POPL} 2017, Paris, France, January 18-20, 2017. pp. 458--472. {ACM} (2017), \url{http://dl.acm.org/citation.cfm?id=3009895} [18] Brutschy, L., Dimitrov, D., M{\"{u}}ller, P., Vechev, M.T.: Static serializability analysis for causal consistency. In: Foster, J.S., Grossman, D. (eds.) Proceedings of the 39th {ACM} {SIGPLAN} Conference on Programming Language Design and Implementation, {PLDI} 2018, Philadelphia, PA, USA, June 18-22, 2018. pp. 90--104. {ACM} (2018). \doi{10.1145/3192366.3192415}, \url{https://doi.org/10.1145/3192366.3192415} [19] Burckhardt, S.: Principles of eventual consistency. Found. Trends Program. Lang. \textbf{1}(1-2), 1--150 (2014). \doi{10.1561/2500000011}, \url{https://doi.org/10.1561/2500000011} [20] Burckhardt, S., Gotsman, A., Yang, H., Zawirski, M.: Replicated data types: specification, verification, optimality. In: Jagannathan, S., Sewell, P. (eds.) The 41st Annual {ACM} {SIGPLAN-SIGACT} Symposium on Principles of Programming Languages, {POPL} '14, San Diego, CA, USA, January 20-21, 2014. pp. 271--284. {ACM} (2014). \doi{10.1145/2535838.2535848}, \url{https://doi.org/10.1145/2535838.2535848} [21] Burckhardt, S., Leijen, D., Protzenko, J., F{\"{a}}hndrich, M.: Global sequence protocol: {A} robust abstraction for replicated shared state. In: Boyland, J.T. (ed.) 29th European Conference on Object-Oriented Programming, {ECOOP} 2015, July 5-10, 2015, Prague, Czech Republic. LIPIcs, vol.~37, pp. 568--590. Schloss Dagstuhl - Leibniz-Zentrum f{\"{u}}r Informatik (2015). \doi{10.4230/LIPIcs.ECOOP.2015.568}, \url{https://doi.org/10.4230/LIPIcs.ECOOP.2015.568} [22] Cahill, M.J., R{\"{o}}hm, U., Fekete, A.D.: Serializable isolation for snapshot databases. {ACM} Trans. Database Syst. \textbf{34}(4), 20:1--20:42 (2009). \doi{10.1145/1620585.1620587}, \url{https://doi.org/10.1145/1620585.1620587} [23] Cassandra-lock: \url{https://github.com/dekses/cassandra-lock} [24] Cerone, A., Bernardi, G., Gotsman, A.: A framework for transactional consistency models with atomic visibility. In: Aceto, L., de~Frutos{-}Escrig, D. (eds.) 26th International Conference on Concurrency Theory, {CONCUR} 2015, Madrid, Spain, September 1.4, 2015. LIPIcs, vol.~42, pp. 58--71. Schloss Dagstuhl - Leibniz-Zentrum f{\"{u}}r Informatik (2015). \doi{10.4230/LIPIcs.CONCUR.2015.58}, \url{https://doi.org/10.4230/LIPIcs.CONCUR.2015.58} [25] Cerone, A., Gotsman, A.: Analysing snapshot isolation. J. {ACM} \textbf{65}(2), 11:1--11:41 (2018). \doi{10.1145/3152396}, \url{https://doi.org/10.1145/3152396} [26] Cerone, A., Gotsman, A., Yang, H.: Algebraic laws for weak consistency. In: Meyer, R., Nestmann, U. (eds.) 28th International Conference on Concurrency Theory, {CONCUR} 2017, September 5-8, 2017, Berlin, Germany. LIPIcs, vol.~85, pp. 26:1--26:18. Schloss Dagstuhl - Leibniz-Zentrum f{\"{u}}r Informatik (2017). \doi{10.4230/LIPIcs.CONCUR.2017.26}, \url{https://doi.org/10.4230/LIPIcs.CONCUR.2017.26} [27] Dan, A.M., Meshman, Y., Vechev, M.T., Yahav, E.: Effective abstractions for verification under relaxed memory models. Comput. Lang. Syst. Struct. \textbf{47}, 62--76 (2017). \doi{10.1016/j.cl.2016.02.003}, \url{https://doi.org/10.1016/j.cl.2016.02.003} [28] Derevenetc, E., Meyer, R.: Robustness against power is pspace-complete. In: Esparza, J., Fraigniaud, P., Husfeldt, T., Koutsoupias, E. (eds.) Automata, Languages, and Programming - 41st International Colloquium, {ICALP} 2014, Copenhagen, Denmark, July 8-11, 2014, Proceedings, Part {II}. Lecture Notes in Computer Science, vol.~8573, pp. 158--170. Springer (2014). \doi{10.1007/978-3-662-43951-7\_14}, \url{https://doi.org/10.1007/978-3-662-43951-7\_14} [29] Difallah, D.E., Pavlo, A., Curino, C., Cudr{\'{e}}{-}Mauroux, P.: Oltp-bench: An extensible testbed for benchmarking relational databases. Proc. {VLDB} Endow. \textbf{7}(4), 277--288 (2013). \doi{10.14778/2732240.2732246}, \url{http://www.vldb.org/pvldb/vol7/p277-difallah.pdf} [30] Experiments: \url{https://github.com/relative-robustness/artifact} [31] Gotsman, A., Yang, H., Ferreira, C., Najafzadeh, M., Shapiro, M.: 'cause i'm strong enough: reasoning about consistency choices in distributed systems. In: Bod{\'{\i}}k, R., Majumdar, R. (eds.) Proceedings of the 43rd Annual {ACM} {SIGPLAN-SIGACT} Symposium on Principles of Programming Languages, {POPL} 2016, St. Petersburg, FL, USA, January 20 - 22, 2016. pp. 371--384. {ACM} (2016). \doi{10.1145/2837614.2837625}, \url{https://doi.org/10.1145/2837614.2837625} [32] Hawblitzel, C., Petrank, E., Qadeer, S., Tasiran, S.: Automated and modular refinement reasoning for concurrent programs. In: Kroening, D., Pasareanu, C.S. (eds.) Computer Aided Verification - 27th International Conference, {CAV} 2015, San Francisco, CA, USA, July 18-24, 2015, Proceedings, Part {II}. Lecture Notes in Computer Science, vol.~9207, pp. 449--465. Springer (2015). \doi{10.1007/978-3-319-21668-3\_26}, \url{https://doi.org/10.1007/978-3-319-21668-3\_26} [33] Holt, B., Bornholt, J., Zhang, I., Ports, D.R.K., Oskin, M., Ceze, L.: Disciplined inconsistency with consistency types. In: Aguilera, M.K., Cooper, B., Diao, Y. (eds.) Proceedings of the Seventh {ACM} Symposium on Cloud Computing, Santa Clara, CA, USA, October 5-7, 2016. pp. 279--293. {ACM} (2016). \doi{10.1145/2987550.2987559}, \url{https://doi.org/10.1145/2987550.2987559} [34] Kozen, D.: Lower bounds for natural proof systems. In: 18th Annual Symposium on Foundations of Computer Science, Providence, Rhode Island, USA, 31 October - 1 November 1977. pp. 254--266. {IEEE} Computer Society (1977). \doi{10.1109/SFCS.1977.16}, \url{https://doi.org/10.1109/SFCS.1977.16} [35] Lahav, O., Margalit, R.: Robustness against release/acquire semantics. In: McKinley, K.S., Fisher, K. (eds.) Proceedings of the 40th {ACM} {SIGPLAN} Conference on Programming Language Design and Implementation, {PLDI} 2019, Phoenix, AZ, USA, June 22-26, 2019. pp. 126--141. {ACM} (2019). \doi{10.1145/3314221.3314604}, \url{https://doi.org/10.1145/3314221.3314604} [36] Lahav, O., Vafeiadis, V.: Owicki-gries reasoning for weak memory models. In: Halld{\'{o}}rsson, M.M., Iwama, K., Kobayashi, N., Speckmann, B. (eds.) Automata, Languages, and Programming - 42nd International Colloquium, {ICALP} 2015, Kyoto, Japan, July 6-10, 2015, Proceedings, Part {II}. Lecture Notes in Computer Science, vol.~9135, pp. 311--323. Springer (2015). \doi{10.1007/978-3-662-47666-6\_25}, \url{https://doi.org/10.1007/978-3-662-47666-6\_25} [37] Lamport, L.: Time, clocks, and the ordering of events in a distributed system. Commun. {ACM} \textbf{21}(7), 558--565 (1978). \doi{10.1145/359545.359563}, \url{https://doi.org/10.1145/359545.359563} [38] Lipton, R.J.: Reduction: {A} method of proving properties of parallel programs. Commun. {ACM} \textbf{18}(12), 717--721 (1975). \doi{10.1145/361227.361234}, \url{https://doi.org/10.1145/361227.361234} [39] Nagar, K., Jagannathan, S.: Automated detection of serializability violations under weak consistency. In: Schewe, S., Zhang, L. (eds.) 29th International Conference on Concurrency Theory, {CONCUR} 2018, September 4-7, 2018, Beijing, China. LIPIcs, vol.~118, pp. 41:1--41:18. Schloss Dagstuhl - Leibniz-Zentrum f{\"{u}}r Informatik (2018). \doi{10.4230/LIPIcs.CONCUR.2018.41}, \url{https://doi.org/10.4230/LIPIcs.CONCUR.2018.41} [40] Najafzadeh, M., Gotsman, A., Yang, H., Ferreira, C., Shapiro, M.: The {CISE} tool: proving weakly-consistent applications correct. In: Alvaro, P., Bessani, A. (eds.) Proceedings of the 2nd Workshop on the Principles and Practice of Consistency for Distributed Data, PaPoC@EuroSys 2016, London, United Kingdom, April 18, 2016. pp. 2:1--2:3. {ACM} (2016). \doi{10.1145/2911151.2911160}, \url{https://doi.org/10.1145/2911151.2911160} [41] Papadimitriou, C.H.: The serializability of concurrent database updates. J. {ACM} \textbf{26}(4), 631--653 (1979). \doi{10.1145/322154.322158}, \url{https://doi.org/10.1145/322154.322158} [42] Perrin, M., Most{\'{e}}faoui, A., Jard, C.: Causal consistency: beyond memory. In: Asenjo, R., Harris, T. (eds.) Proceedings of the 21st {ACM} {SIGPLAN} Symposium on Principles and Practice of Parallel Programming, PPoPP 2016, Barcelona, Spain, March 12-16, 2016. pp. 26:1--26:12. {ACM} (2016). \doi{10.1145/2851141.2851170}, \url{https://doi.org/10.1145/2851141.2851170} [43] Raad, A., Lahav, O., Vafeiadis, V.: On the semantics of snapshot isolation. In: Enea, C., Piskac, R. (eds.) Verification, Model Checking, and Abstract Interpretation - 20th International Conference, {VMCAI} 2019, Cascais, Portugal, January 13-15, 2019, Proceedings. Lecture Notes in Computer Science, vol. 11388, pp. 1--23. Springer (2019). \doi{10.1007/978-3-030-11245-5\_1}, \url{https://doi.org/10.1007/978-3-030-11245-5\_1} [44] Rackoff, C.: The covering and boundedness problems for vector addition systems. Theor. Comput. Sci. \textbf{6}, 223--231 (1978). \doi{10.1016/0304-3975(78)90036-1}, \url{https://doi.org/10.1016/0304-3975(78)90036-1} [45] Shapiro, M., Ardekani, M.S., Petri, G.: Consistency in 3d. In: Desharnais, J., Jagadeesan, R. (eds.) 27th International Conference on Concurrency Theory, {CONCUR} 2016, August 23-26, 2016, Qu{\'{e}}bec City, Canada. LIPIcs, vol.~59, pp. 3:1--3:14. Schloss Dagstuhl - Leibniz-Zentrum f{\"{u}}r Informatik (2016). \doi{10.4230/LIPIcs.CONCUR.2016.3}, \url{https://doi.org/10.4230/LIPIcs.CONCUR.2016.3} [46] Shasha, D.E., Snir, M.: Efficient and correct execution of parallel programs that share memory. {ACM} Trans. Program. Lang. Syst. \textbf{10}(2), 282--312 (1988). \doi{10.1145/42190.42277}, \url{https://doi.org/10.1145/42190.42277} [47] Trade: \url{https://github.com/Haiyan2/Trade} [48] Twitter: \url{https://github.com/edmundophie/cassandra-twitter} \end{thebibliography} %%%%% To display Open Access text and logo, Please add below text and copy the cc_by_4-0.eps in the manuscript package %%% \vfill {\small\medskip\noindent{\bf Open Access} This chapter is licensed under the terms of the Creative Commons\break Attribution 4.0 International License (\url{http://creativecommons.org/licenses/by/4.0/}), which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made.} {\small \spaceskip .28em plus .1em minus .1em The images or other third party material in this chapter are included in the chapter's Creative Commons license, unless indicated otherwise in a credit line to the material.~If material is not included in the chapter's Creative Commons license and your intended\break use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.} \medskip\noindent\includegraphics{cc.pdf} %% Appendix \appendix \newpage %!TEX root = draft.tex \section{Proofs for Section \ref{sec:CCPCrobustness}} \label{sec:CCPCrobustnessProofs} \begin{proof}[Proof of Lemma \ref{lem:Transform}] We start with the case $\textsf{X} = \ccc$. We first show that $\atrace_\pcinstr$ satisfies $\axpoco$ and $\axcoarb$. For $\axpoco$, let $\atr_{1}' \in \{\atrrd{\atr_{1}},\atrwr{\atr_{1}}\}$ and $\atr_{2}' \in \{\atrrd{\atr_{2}},\atrwr{\atr_{2}}\}$, such that $(\atr_{1}',\atr_{2}') \in (\po_\pcinstr\cup\rfo_\pcinstr)^{+}$. By the definition of $\viso_\pcinstr$, we have that either $(\atr_{1}'=\atrrd{\atr_{1}},\atr_{2}'=\atrwr{\atr_{2}}) \in \po_\pcinstr$ and $\atr_1 = \atr_2$ or $(\atr_{1},\atr_2) \in (\po\cup\rfo)^{+}$, which implies that $(\atr_{1},\atr_2) \in \viso$. In both cases we obtain that $(\atr_{1}',\atr_{2}') \in \viso_\pcinstr$. The axiom $\axpoco$ can be proved in a similar way. % \item $\axcoarb$: the same proof steps as in $\axpoco$. % \end{itemize} Next, we show that $\atrace_\pcinstr$ satisfies the property $\axretval$. Let $\atr$ be a transaction in $\atrace$ that contains a read event $\readact(\atr,\anaddr,\aval)$. Let $\atr_0$ be the transaction in $\atrace$ such that $$\atr_0 = Max_{\arbo}(\{\atr' \in \atrace\ |\ (\atr',\atr) \in \viso \wedge \exists\ \writeact(\atr',\anaddr,\cdot)\in\amap(\atr')\}).$$ The read value $\aval$ must have been written by $\atr_0$ since $\atrace$ satisfies $\axretval$. Thus, the read $\readact(\atr,\anaddr,\aval)$ in $\atrrd{\atr}$ of $\atrace_\pcinstr$ must return the value written by $\atrwr{\atr_0}$. From the definitions of $\viso_\pcinstr$ and $\arbo_\pcinstr$, we get $$\atrwr{\atr_1}\in\{\atrwr{\atr'} \in \atrace_\pcinstr\ |\ (\atrwr{\atr'},\atrrd{\atr}) \in \viso_\pcinstr \wedge \exists\ \writeact(\atrwr{\atr'},\anaddr,\cdot)\in\amap(\atrwr{\atr'})\}$$ $$\atr_1 \in \{\atr' \in \atrace\ |\ (\atr',\atr) \in \viso \wedge \exists\ \writeact(\atr',\anaddr,\cdot)\in\amap(\atr')\}$$ because $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \viso_\pcinstr$ implies $(\atr_{1},\atr_{2}) \in \viso$. Since $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \arbo_\pcinstr$ implies $(\atr_{1},\atr_{2}) \in \arbo$, we also obtain that $$\atrwr{\atr_0} = Max_{\arbo_\pcinstr}(\{\atrwr{\atr'} \in \atrace_\pcinstr\ |\ (\atrwr{\atr'},\atrrd{\atr}) \in \viso_\pcinstr \wedge \exists\ \writeact(\atrwr{\atr'},\anaddr,\cdot)\in\amap(\atrwr{\atr'})\})$$ and since the read $\readact(\atr,\anaddr,\aval)$ in $\atrrd{\atr}$ of $\atrace_\pcinstr$ returns the value written by $\atrwr{\atr_0}$, $\atrace_\pcinstr$ satisfies $\axretval$. For the case $\textsf{X} = \pcc$, we show that $\atrace_\pcinstr$ satisfies the property $\axprefix$ (the other axioms are proved as in the case of $\ccc$). % In the previous case we already showed that $\axpoco$ and $\axcoarb$ hold. % Now we will show that $\atrace_\pcinstr$ satisfies the property $\axprefix$. Suppose we have $(\atr_{1}',\atr_{2}') \in \arbo_\pcinstr$ and $(\atr_{2}',\atr_{3}') \in \viso_\pcinstr$ where $\atr_{1}' \in \{\atrrd{\atr_{1}},\atrwr{\atr_{1}}\}$, $\atr_{2}' \in \{\atrrd{\atr_{2}},\atrwr{\atr_{2}}\}$, and $\atr_{3}' \in \{\atrrd{\atr_{3}},\atrwr{\atr_{3}}\}$. The are five cases to be discussed: \begin{enumerate} %[label=\alph*] \item $(\atr_{1}'=\atrrd{\atr_{1}},\atr_{2}'=\atrwr{\atr_{2}}) \in \po_\pcinstr$ and $\atr_1 = \atr_2$ and $(\atr_{2},\atr_{3}) \in \viso$, \item $(\atr_{1},\atr_{2}) \in \arbo$ and $(\atr_{2},\atr_{3}) \in \viso$, \item $(\atr_{1},\atr_{2}) \in \arbo$ and $(\atr_{2}'=\atrrd{\atr_{2}},\atr_{3}'=\atrwr{\atr_{3}}) \in \po_\pcinstr$ and $\atr_2 = \atr_3$, \item $(\atr_{1}'=\atrrd{\atr_{1}},\atr_{2}'=\atrwr{\atr_{2}}) \in \po_\pcinstr$ and $\atr_1 = \atr_2$ and $(\atr_{2},\atr_{3}) \in \arbo$ and $\atr_{3}'= \atrwr{\atr_{3}}$, \item $(\atr_{1},\atr_{2}) \in \arbo$ and $(\atr_{2},\atr_{3}) \in \arbo$ and $\atr_{3}'= \atrwr{\atr_{3}}$. \end{enumerate} Cases (a) and (b) imply that $(\atr_{1},\atr_{3}) \in \viso$ since $\arbo;\viso \subset \arbo$, which implies that $(\atr_{1}',\atr_{3}') \in \viso_\pcinstr$. Cases (c), (d), and (e) imply that $(\atr_{1},\atr_{3}) \in \arbo$ and $\atr_{3}'= \atrwr{\atr_{3}}$ then we get that $(\atrwr{\atr_{1}},\atrwr{\atr_{3}}) \in \viso_\pcinstr$ and $\atr_{3}'= \atrwr{\atr_{3}}$ which means that $(\atr_{1}',\atr_{3}') \in \viso_\pcinstr$. Note that the rule $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \viso_\pcinstr$ if $(\atr_{1},\atr_{2}) \in \arbo$ cannot change the fact that $$\atrwr{\atr_1}\in\{\atrwr{\atr'} \in \atrace_\pcinstr\ |\ (\atrwr{\atr'},\atrrd{\atr}) \in \viso_\pcinstr \wedge \exists\ \writeact(\atrwr{\atr'},\anaddr,\cdot)\in\amap(\atrwr{\atr'})\}$$ iff $$\atr_1 \in \{\atr' \in \atrace\ |\ (\atr',\atr) \in \viso \wedge \exists\ \writeact(\atr',\anaddr,\cdot)\in\amap(\atr')\}$$ Thus, the proof of $\axretval$ follows as in the previous case. For the case $\textsf{X} = \sic$, we show that $\atrace_\pcinstr$ satisfies $\axconflict$. If $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \sto_\pcinstr$, then $(\atr_{1},\atr_{2}) \in \sto \subset \viso$, which implies that $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \viso_\pcinstr$. Therefore, $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \viso_\pcinstr$, which concludes the proof. The axiom $\axretval$ can be proved as in the previous cases. %Thus, the property $\axconflict$ holds. Also, we prove in the same way as the previous two cases that $\atrace_\pcinstr$ satisfies the property $\axretval$. \end{proof} \begin{proof}[Proof of Lemma \ref{lem:cycles}] ($\Rightarrow$) Let $\atrace$ be a trace under \ccc{}. From $\axpoco$ and $\axcoarb$ we get that $\arbo_{0}^{+} \subset \arbo$, and $\arbo_{0}^{+}$ is acyclic because $\arbo$ is total order. Assume by contradiction that $\viso_{0}^{+};\cfo$ is cyclic which implies that $\viso;\cfo$ is cyclic since $\viso_{0}^{+} \subset \viso$, which means that there exist $\atr_1$ and $\atr_2$ such that $(\atr_1, \atr_2) \in \viso$ and $(\atr_2, \atr_1) \in \cfo$. $(\atr_2, \atr_1) \in \cfo$ implies that there exists $\atr_3$ such that $(\atr_3, \atr_1) \in \sto$ and $(\atr_3, \atr_2) \in \rfo$. Based on the definition of $\axretval$, $\atr_3$ has two possible instances: \begin{itemize} \item $\atr_3$ corresponds to the "fictional" transaction that wrote the initial values which cannot be the case when $(\atr_1, \atr_2) \in \viso$ and $\atr_1$ writes to the same variable that $\atr_2$ reads from, \item $\atr_3$ is the last transaction that occurs before $\atr_2$ that writes the value read by $\atr_2$, which means that $(\atr_1,\atr_3) \in \arbo$ which contradicts the fact that $(\atr_3, \atr_1) \in \sto$ since $\sto \subset \arbo$. \end{itemize} ($\Leftarrow$) Let $\atrace$ be a trace such that $\arbo_{0}^{+}$ and $\viso_{0}^{+};\cfo$ are acyclic. Then, we define the relations $\viso$ and $\arbo$ such that $\viso = \viso_{0}^{+}$ and $\arbo$ is any total order that includes $\arbo_{0}^{+}$. Then, we obtain that $(\viso \cup \sto)^{+} \subset \arbo$ and $\viso;\cfo$ is acyclic. Thus, $\atrace$ satisfies the properties $\axpoco$ and $\axcoarb$. Next, we will show that $\atrace$ satisfies $\axretval$. Let $\atr$ be a transaction in $\atrace$ that contains a read event $\readact(\atr,\anaddr,\aval)$. Let $\atr_0$ be transaction in $\atrace$ such that $$\atr_0 = Max_{\arbo}(\{\atr' \in \atrace\ |\ (\atr',\atr) \in \viso \wedge \exists\ \writeact(\atr',\anaddr,\cdot)\in\amap(\atr')\})$$ then the read must return a value written by $\atr_0$. Assume by contradiction that there exists some other transaction $\atr_1 \neq \atr_0$ such that $(\atr_1,\atr) \in \rfo$. Then, we get that $(\atr_1,\atr_0) \in \arbo$ and both write to $\anaddr$, therefore, $(\atr_1,\atr_0) \in \sto$ since $\sto \subset \arbo$. Combining $(\atr_1,\atr) \in \rfo$ and $(\atr_1,\atr_0) \in \sto$ we obtain $(\atr,\atr_0) \in \cfo$ and since $(\atr_0,\atr) \in \viso$ then we obtain that $(\atr,\atr) \in \viso;\cfo$ which contradicts the fact that $\viso;\cfo$ is acyclic. Therefore, the read value was written by $\atr_0$ and $\atrace$ satisfies $\axretval$. %%(2): The only-if direction: similar to (1) since $\ccc{} \subset \pcc$, we have that $\arbo_{0}^{+}$ is acyclic. We assume by contradiction that $\arbo_{0}^{+}?;\viso_{0};\cfo$ is cyclic. The property $\axprefix$ implies that $\sto;\viso \subset \viso$ since $\sto \subset \arbo$ and since $\viso_{0}^{+} \subset \viso$ then $\arbo_{0}^{+}?;\viso_{0} \subset \viso$. Thus, $\viso;\cfo$ is cyclic which results in a contradiction as in (1). %%(2): The if direction: we define the relations $\viso$ and $\arbo$ such that $\arbo_{0}^{+}\subset \arbo$ and $\viso = \arbo ; \viso_{0} \cup \viso_{0}$ which imply that $\viso_{0}^{+} \subset \viso$ and $\arbo;\viso \subset \viso$. Thus, $\atrace$ satisfies the properties $\axpoco$, $\axcoarb$, and $\axprefix$. Using the same proof steps as in (1) we can show that $\atrace$ satisfies $\axretval$. %%(3): The only-if direction: similar to above we have that $\arbo_{0}^{+}$ is acyclic. We assume by contradiction that $\arbo_{0}^{+};\cfo$ is cyclic. Combining the properties $\axconflict$ and $\axpoco$ implies that $\arbo_{0}^{+} \subset \viso$. Thus, $\viso;\cfo$ is cyclic which results in a contradiction as in (1). %%(3): The if direction: we can define the relations $\viso$ and $\arbo$ such that $\viso = \arbo_{0}^{+}$ and $\arbo_{0}^{+}\subset \arbo$ which imply the properties $\axpoco$, $\axcoarb$, and $\axconflict$. Using the same proof steps as in (1) we can show that $\atrace$ satisfies $\axretval$. \end{proof} \begin{proof}[Proof of Lemma~\ref{lem:CcCc}] The only-if direction follows from Lemma \ref{lem:Transform}. For the if direction: consider a trace $\atrace_\pcinstr$ which is \ccc{}. We prove by contradiction that $\atrace$ must be \ccc{} as well. Assume that $\atrace$ is not \ccc{} then it must contain a cycle in either $\arbo_{0}^{+}$ or $\viso_{0}^{+};\cfo$ (based on Lemma \ref{lem:cycles}). In the rest of the proof when we mention a cycle we implicitly refer to a cycle in either $\arbo_{0}^{+}$ or $\viso_{0}^{+};\cfo$. Splitting every transaction $\atr \in \atrace$ in a trace to a pair of transactions $\atrrd{\atr}$ and $\atrwr{\atr}$ that occur in this order might not maintain a cycle of $\atrace$. However, we prove that this is not possible and our splitting conserves the cycle. Assume we have a vertex $\atr$ as part of the cycle. We show that $\atr$ can be split into two transactions $\atrrd{\atr}$ and $\atrwr{\atr}$ while maintaining the cycle. Note that $\atr$ is part of a cycle iff either \begin{enumerate} \item $(\atr_{1},\atr) \in \arbo_{0}$ and $(\atr,\atr_{2})\in \arbo_{0}$ or \item $(\atr_{1},\atr) \in \viso_{0}$ and $(\atr,\atr_{2})\in \viso_{0}$ or \item $(\atr_{1},\atr) \in \viso_{0}$ and $(\atr,\atr_{2})\in \cfo$ or \item $(\atr_{1},\atr) \in \cfo$ and $(\atr,\atr_{2})\in \viso_{0}$ \end{enumerate} where $\atr_{1}$ and $\atr_{2}$ might refer to the same transaction. Thus, by splitting $\atr$ to $\atrrd{\atr}$ and $\atrwr{\atr}$, the above four cases imply that: \begin{enumerate} \item if $(\atr_{1},\atr) \in \viso_{0}$ and $(\atr,\atr_{2})\in \arbo_{0}$ then $(\atr_{1}',\atrrd{\atr}) \in (\po_\pcinstr \cup \rfo_\pcinstr)$ and $(\atrwr{\atr},\atr_{2}')\in (\po_\pcinstr \cup \rfo_\pcinstr \cup \sto_\pcinstr)$ where $\atr_{1}' \in \{\atrrd{\atr_{1}},\atrwr{\atr_{1}}\}$ and $\atr_{2}' \in \{\atrrd{\atr_{2}},\atrwr{\atr_{2}}\}$. This maintains the vertices $\atr_{1}'$ and $\atr_{2}'$ connected in the cycle formed by the dependency relations of $\atrace_\pcinstr$ since $(\atrrd{\atr},\atrwr{\atr}) \in \po_\pcinstr$; \item if $(\atr_{1},\atr) \in \sto$ and $(\atr,\atr_{2})\in \arbo_{0}$ then $(\atr_{1}',\atrwr{\atr}) \in \sto_\pcinstr$ and $(\atrwr{\atr},\atr_{2}')\in (\po_\pcinstr \cup \rfo_\pcinstr \cup \sto_\pcinstr)$ which maintains the vertices $\atr_{1}'$ and $\atr_{2}'$ connected in the cycle formed by the dependency relations of $\atrace_\pcinstr$; \item $(\atr_{1},\atr) \in \viso_{0}$ and $(\atr_{2},\atr) \in \cfo$ then $(\atr_{1}',\atrrd{\atr}) \in (\po_\pcinstr \cup \rfo_\pcinstr)$ and $(\atrrd{\atr},\atr_{2}')\in \cfo_\pcinstr$ maintains the vertices $\atr_{1}'$ and $\atr_{2}'$ connected in the cycle formed by the dependency relations of $\atrace_\pcinstr$; \item $(\atr_{1},\atr) \in \cfo$ and $(\atr_{2},\atr) \in \viso_{0}$ then $(\atr_{1}',\atrwr{\atr}) \in \cfo_\pcinstr$ and $(\atrwr{\atr},\atr_{2}')\in (\po_\pcinstr \cup \rfo_\pcinstr)$ which maintains the vertices $\atr_{1}'$ and $\atr_{2}'$ connected in the cycle formed by the dependency relations of $\atrace_\pcinstr$ as well. \end{enumerate} Therefore, doing the splitting creates a cycle in either $(\po_\pcinstr \cup \rfo_\pcinstr \cup \sto_\pcinstr)^{+}$ or $(\po_\pcinstr \cup \rfo_\pcinstr)^{+};\cfo_\pcinstr$ which implies that $\atrace_\pcinstr$ is not \ccc{}, a contradiction. \end{proof} \begin{proof}[Proof of Lemma \ref{lem:PcSer}] ($\Leftarrow$) Assume that $\atrace_\pcinstr$ is \serc{}. We will show that $\atrace$ is \pcc{}. Notice that if $(\atr_{1},\atr_{2}) \in \viso_{0}^{+}$ then $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \arbo_\pcinstr$ which implies that $(\atr_{1},\atr_{2}) \in \viso$. Similarly, if $(\atr_{1},\atr_{2}) \in \arbo_{0}^{+}$ then $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \arbo_\pcinstr$ or $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \arbo_\pcinstr$ which implies that $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \arbo_\pcinstr$ which in both cases implies that $(\atr_{1},\atr_{2}) \in \arbo$. Thus, $\atrace$ satisfies the properties $\axpoco$ and $\axcoarb$. Now assume that $(\atr_{1},\atr_{2})\in \arbo$ and $(\atr_{2},\atr_{3})\in \viso$. We show that $(\atr_{1},\atr_{3})\in \viso$. The assumption implies that $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \arbo_\pcinstr$ and $(\atrwr{\atr_{2}},\atrrd{\atr_{3}}) \in \arbo_\pcinstr$, which means that $(\atrwr{\atr_{1}},\atrrd{\atr_{3}}) \in \arbo_\pcinstr$. Therefore, $(\atr_{1},\atr_{3}) \in \viso$ and $\atrace$ satisfies the property $\axconflict$. Concerning $\axretval$, let $\atr$ be a transaction in $\atrace$ that contains a read event $\readact(\atr,\anaddr,\aval)$. Let $\atr_0$ be transaction in $\atrace$ such that $$\atr_0 = Max_{\arbo}(\{\atr' \in \atrace\ |\ (\atr',\atr) \in \viso \wedge \exists\ \writeact(\atr',\anaddr,\cdot)\in\amap(\atr')\}).$$ We show that the read must return a value written by $\atr_0$. The definitions of $\viso$ and $\arbo$ imply that $$\atrwr{\atr_1}\in\{\atrwr{\atr'} \in \atrace_\pcinstr\ |\ (\atrwr{\atr'},\atrrd{\atr}) \in \arbo_\pcinstr \wedge \exists\ \writeact(\atrwr{\atr'},\anaddr,\cdot)\in\amap(\atrwr{\atr'})\}$$ $$\atr_1 \in \{\atr' \in \atrace\ |\ (\atr',\atr) \in \viso \wedge \exists\ \writeact(\atr',\anaddr,\cdot)\in\amap(\atr')\}$$ because $(\atrwr{\atr_{1}},\atrrd{\atr_{2}}) \in \arbo_\pcinstr$ implies $(\atr_{1},\atr_{2}) \in \viso$. Then, we obtain that $$\atrwr{\atr_0} = Max_{\arbo_\pcinstr}(\{\atrwr{\atr'} \in \atrace_\pcinstr\ |\ (\atrwr{\atr'},\atrrd{\atr}) \in \arbo_\pcinstr \wedge \exists\ \writeact(\atrwr{\atr'},\anaddr,\cdot)\in\amap(\atrwr{\atr'})\})$$ and since $\atrace_\pcinstr$ is \serc{} we know that the read must return the value written by $\atrwr{\atr_0}$. Thus, the read returns the value written by $\atr_0$, which implies that $\atrace$ satisfies $\axretval$ holds. Therefore, $\atrace$ is \pcc{}. ($\Rightarrow$) Assume that $\atrace$ is \pcc{}. We show that $\atrace_\pcinstr$ is \serc{}. Since $\atrace_\pcinstr$ is the result of splitting transactions, a cycle in its dependency relations can only originate from a cycle in $\atrace$. Therefore, it is sufficient to show that any happens-before cycle in $\atrace$ is broken in $\atrace_\pcinstr$. From Lemma \ref{lem:pccycles}, we have that $\atrace$ either does not admit a happens-before cycle or any (simple) happens-before cycle in $\atrace$ must have either two successive $\cfo$ dependencies or a $\sto$ dependency followed by a $\cfo$ dependency. If $\atrace$ does not admit a happens-before cycle then it is \serc{}, and $\atrace_\pcinstr$ is trivially \serc{} (since splitting transactions cannot introduce new cycles). \scalebox{0.67} \begin{tikzpicture} \node[shape=rectangle ,draw=none,font=\large] (A0) at (0,0) [] {$\atr_1$ }; \node[shape=rectangle ,draw=none,font=\large] (A1) at (2,0) [] {$\atr_2$}; \node[shape=rectangle ,draw=none,font=\large] (B1) at (4,0) [] {$\atr_3$}; \node[shape=rectangle ,draw=none,font=\large] (B2) at (5,0) [] {$\Longrightarrow$}; %\node[shape=rectangle ,draw=none,font=\large] (C0) at (8,0) [] {$\cdots$ }; \node[shape=rectangle ,draw=none,font=\large] (C1) at (6,0) [] {$\atrrd{\atr_{1}}$}; \node[shape=rectangle ,draw=none,font=\large] (D1) at (8,0) [] {$\atrwr{\atr_{1}}$}; \node[shape=rectangle ,draw=none,font=\large] (D2) at (10,0) [] {$\atrrd{\atr_{2}}$}; \node[shape=rectangle ,draw=none,font=\large] (D0) at (12,0) [] {$\atrwr{\atr_{2}}$}; \node[shape=rectangle ,draw=none,font=\large] (E0) at (14,0) [] {$\atrrd{\atr_{3}}$ }; \node[shape=rectangle ,draw=none,font=\large] (E1) at (16,0) [] {$\atrwr{\atr_{3}}$}; \begin{scope}[ every edge/.style={draw=black,very thick}] \path [->] (A0) edge[] node [above,font=\small] {$\sto \cup \cfo$} (A1); \path [->] (A1) edge[] node [above,font=\small] {$\cfo$} (B1); \path [->] (B1) edge[bend left] node [above,font=\small] {$\hbo$} (A0); \path [->] (C1) edge[] node [above,font=\small] {$\po_\pcinstr$} (D1); \path [->] (D1) edge[bend left] node [below,font=\small] {$\sto_\pcinstr$} (D0); \path [->] (C1) edge[bend right] node [above,font=\small] {$\cfo_\pcinstr$} (D0); \path [->] (D2) edge[] node [above,font=\small] {$\po_\pcinstr$} (D0); \path [->] (D2) edge[bend right] node [above,font=\small] {$\cfo_\pcinstr$} (E1); \path [->] (E0) edge[] node [above,font=\small] {$\po_\pcinstr$} (E1); \end{scope} \end{tikzpicture}} Otherwise, if $\atrace$ admits a happens-before cycle like above, then $\atrace$ must contain three transactions $\atr_{1}$, $\atr_{2}$, and $\atr_{3}$ such that $(\atr_{1},\atr_{2}) \in \sto \cup \cfo$, $(\atr_{2},\atr_{3}) \in \cfo$, and $(\atr_{3},\atr_{1}) \in \hbo$ (like in the picture above). Then, by splitting transactions we obtain that $(\atrwr{\atr_{1}},\atrwr{\atr_{2}}) \in \sto_\pcinstr$ or $(\atrrd{\atr_{1}},\atrwr{\atr_{2}}) \in \cfo_\pcinstr$, and $(\atrrd{\atr_{2}},\atrwr{\atr_{3}}) \in \cfo_\pcinstr$. Since, we have $(\atrrd{\atr_{2}},\atrwr{\atr_{2}}) \in \po_\pcinstr$ (and not $(\atrwr{\atr_{2}},\atrrd{\atr_{2}}) \in \po_\pcinstr$), this cannot lead to a cycle in $\atrace_\pcinstr$, which concludes the proof that $\atrace_\pcinstr$ is \serc{} %Thus, in all cases $\atrace_\pcinstr$ doesn't contain cycles. Therefore, $\atrace_\pcinstr$ is \serc{}. \end{proof} %!TEX root = draft.tex \section{Proofs for Section~\ref{sec:robustness}} \label{sec:robustnessProofs} \begin{proof}[Proof of Lemma~\ref{lem:pcsicycles}] %Since $\atrace$ is \pcc{}, there exists a causal order $\viso$ and an arbitration order $\arbo$, such that $\viso\subseteq \arbo$ and $\axpc$ hold. Let $\arbo_{1}$ be a total order that includes $\arbo_{0}^{+}$ and $\arbo_{0}^{+};\cfo;\arbo_{0}^{*}$ ($\arbo_{0}^*$ is the reflexive closure of $\arbo_{0}$). This is well defined because there exists no cycle between tuples in these two relations. Indeed, if $(\atr_{1},\atr_{2}) \in \arbo_{0}^{+}$ and there exist $\atr_{3}$ and $\atr_{4}$ such that $(\atr_{2}, \atr_{3}) \in \arbo_{0}^{+}$, $(\atr_{3}, \atr_{4}) \in \cfo$, and $(\atr_{4}, \atr_{1}) \in \arbo_{0}^{*}$, then we have a cycle in $\arbo_{0}^{+};\cfo$ that does not contain two successive $\cfo$ dependencies, which contradicts the hypothesis. Also, for every pair of transactions $(\atr_{1},\atr_{2})$ there cannot exist $\atr_{3}$ and $\atr_{4}$ such that $$(\atr_{2}, \atr_{3}) \in \arbo_{0}^{+},\ (\atr_{3}, \atr_{4}) \in \cfo\mbox{ and }(\atr_{4}, \atr_{1}) \in \arbo_{0}^{*}$$ and $\atr_{3}'$ and $\atr_{4}'$ such that $$(\atr_{1}, \atr_{3}') \in \arbo_{0}^{+},\ (\atr_{3}', \atr_{4}') \in \cfo\mbox{ and }(\atr_{4}', \atr_{2}) \in \arbo_{0}^{*}$$ This will imply a cycle in $\arbo_{0}^{+};\cfo;\arbo_{0}^{+};\cfo$ which again contradicts the hypothesis. %if $(\atr_{1},\atr_{2}) \not\in \arbo_{0}^{+}$ and there exist $\atr_{3}$ and $\atr_{4}$ such that $(\atr_{2}, \atr_{3}) \in \arbo_{0}^{+}$, $(\atr_{3}, \atr_{4}) \in \cfo$, and $(\atr_{4}, \atr_{1}) \in \arbo_{0}^{*}$ ($\arbo_{0}^*$ denotes the reflexive closure of $\arbo_{0}$), then $(\atr_{2},\atr_{1}) \in \arbo_{1}$. % and $\arbo_{1}$ is transitive\footnote{Note that .}. Also, let $\viso_{1}$ be the smallest transitive relation that includes $\arbo_{0}^{+}$ and $\arbo_{1};\arbo_{0}^{+}$. We show that $\viso_{1}$ and $\arbo_{1}$ are causal and arbitration orders of $\atrace$ that satisfy all the axioms of \sic{}. $\axpoco$ and $\axcoarb$ hold trivially. Since $\sto \subseteq \viso_{1}$, $\axconflict$ holds as well. %We show first that $\arbo_{1}$ is a well defined total order. Note that because $\axcoarb$, $\arbo_{0}^{+} \subset \arbo$ is acyclic. % the fact that all happens-before cycles in $\atrace$ contain two successive $\cfo$. Thus, $\arbo_{1}$ is a well defined total order. %Also, since $\arbo_{0}^+ \subset \arbo_{1}$ then $\axcoarb$ holds. %Similar to above $\viso_{1}$ is a well defined partial order. %Note that $\viso_{1} \subset \arbo_{1}$ as well. Also, since $\viso_{0}^+ \subset \viso_{1}$ then $\axpoco$ holds. $\axpc$ holds because $\arbo_{1} ; \viso_{1} = \arbo_{1};(\arbo_{0}^{+} \cup \arbo_{1};\arbo_{0}^{+})^+ = \arbo_{1};\arbo_{0}^{+} \subset \viso_{1}$. The axiom $\axretval$ is equivalent to the acyclicity of $\viso_{1};\cfo$ when $\axpoco$ and $\axcoarb$ hold. Assume by contradiction that $\viso_1;\cfo$ is cyclic. From the definition of $\viso_1$ and the fact that $\arbo_{1}$ is total order we obtain that either: %$\arbo_{0}^{+};\cfo$ or $\arbo_{1};\arbo_{0}^{+};\cfo$ (since ) is cyclic: \begin{itemize} \item $\arbo_{0}^{+};\cfo$ is cyclic, which implies that there exists a happens-before cycle that does not contain two successive $\cfo$, which contradicts the hypothesis, or \item $\arbo_{1};\arbo_{0}^{+};\cfo$ is cyclic, which implies that there exist $\atr_{1}$, $\atr_{2}$, and $\atr_{3}$ such that $(\atr_{2}, \atr_{3}) \in \arbo_{0}^{+}$, $(\atr_{3}, \atr_{1}) \in \cfo$ and $(\atr_{1},\atr_{2}) \in \arbo_{1}$. This contradicts the fact that $(\atr_{2}, \atr_{3}) \in \arbo_{0}^{+}$ and $(\atr_{3}, \atr_{1}) \in \cfo$ implies $(\atr_{2},\atr_{1}) \in \arbo_{1}$. %by definition if $(\atr_{1},\atr_{2}) \not\in \arbo_{0}^{+}$, otherwise, we get $\arbo_{0}^{+};\arbo_{0}^{+};\cfo = \arbo_{0}^{+};\cfo$ is cyclic which leads to the previous case. \end{itemize} Therefore, $\atrace$ satisfies $\axretval$ for $\viso_1$ and $\arbo_1$, which concludes the proof. \end{proof} \medskip Next, we present an important lemma that characterizes happens before cycles possible under the \pcc{} semantics. This is a strengthening of a result in~[12] which shows that all happens before cycles under \pcc{} must have two successive dependencies in ${,}$ and at least one $$. We show that the two successive dependencies cannot be $$ followed $$, or two successive $$. \begin{proof}[Proof of Lemma~\ref{lem:pccycles}] It was shown in [12] that all happens-before cycles under \pcc{} must contain two successive dependencies in $\{\cfo,\sto\}$ and at least one $\cfo$. Assume by contradiction that there exists a cycle with $\cfo$ dependency followed by $\sto$ dependency or two successive $\sto$ dependencies. This cycle must contain at least one additional dependency. Otherwise, the cycle would also have a $\sto$ dependency followed by a $\cfo$ dependency, or it would imply a cycle in $\sto$, which is not possible (since $\sto \subset \arbo$ and $\arbo$ is a total order). % it does this implies a cycle in $\sto \subset \arbo$ which is a contradiction. Then, we get that the dependency just before $\cfo$ is either $\po$ or $\rfo$ (i.e., $\viso_0$) since we cannot have $\cfo$ or $\sto$ followed by $\cfo$. Also, the relation after $\sto$ is either $\po$ or $\rfo$ or $\sto$ (i.e., $\arbo_0$) since we cannot have $\sto$ followed by $\cfo$. Thus, the cycle has the following shape: \medskip \scalebox{0.65} \begin{tikzpicture} \node[shape=rectangle ,draw=none,font=\large] (A0) at (0,0) [] {$\atr_1$ }; \node[shape=rectangle ,draw=none,font=\large] (A1) at (1.3,0) [] {$\atr_2$}; \node[shape=rectangle ,draw=none,font=\large] (B1) at (2.6,0) [] {$\atr_3$}; \node[shape=rectangle ,draw=none,font=\large] (B2) at (3.9,0) [] {$\atr_4$}; \node[shape=rectangle ,draw=none,font=\large] (C0) at (4.5,0) [] {$\cdots$ }; \node[shape=rectangle ,draw=none,font=\large] (C1) at (5.1,0) [] {$\atr_i$}; \node[shape=rectangle ,draw=none,font=\large] (D1) at (6.4,0) [] {$\atr_{i+1}$}; \node[shape=rectangle ,draw=none,font=\large] (D2) at (7.9,0) [] {$\atr_{i+2}$}; \node[shape=rectangle ,draw=none,font=\large] (D0) at (9.4,0) [] {$\atr_{i+3}$}; \node[shape=rectangle ,draw=none,font=\large] (E0) at (10.2,0) [] {$\cdots$ }; \node[shape=rectangle ,draw=none,font=\large] (E1) at (11,0) [] {$\atr_{n-4}$}; \node[shape=rectangle ,draw=none,font=\large] (F1) at (12.5,0) [] {$\atr_{n-3}$}; \node[shape=rectangle ,draw=none,font=\large] (F2) at (14,0) [] {$\atr_{n-2}$}; \node[shape=rectangle ,draw=none,font=\large] (F3) at (15.5,0) [] {$\atr_{n-1}$}; \node[shape=rectangle ,draw=none,font=\large] (F4) at (17,0) [] {$\atr_{n}$}; \begin{scope}[ every edge/.style={draw=black,very thick}] \path [->] (A0) edge[] node [above,font=\small] {$\cfo$} (A1); \path [->] (A1) edge[] node [above,font=\small] {$\sto$} (B1); \path [->] (B1) edge[] node [above,font=\small] {$\arbo_0$} (B2); \path [->] (C1) edge[] node [above,font=\small] {$\viso_0$} (D1); \path [->] (D1) edge[] node [above,font=\small] {$\cfo$} (D2); \path [->] (D2) edge[] node [above,font=\small] {$\sto$} (D0); \path [->] (E1) edge[] node [above,font=\small] {$\viso_0$} (F1); \path [->] (F1) edge[] node [above,font=\small] {$\cfo$} (F2); \path [->] (F2) edge[] node [above,font=\small] {$\sto$} (F3); \path [->] (F3) edge[] node [above,font=\small] {$\arbo_0$} (F4); \path [->] (F4) edge[bend left=11] node [above,font=\small] {$\viso_0$} (A0); \end{scope} \end{tikzpicture}} \medskip Since $\viso_0;\cfo\subseteq \arbo$ is a consequence of the \pcc{} axioms~[26], we get that $(\atr_{n}, \atr_2) \in \arbo$, $(\atr_{i}, \atr_{i+2}) \in \arbo$ and $(\atr_{n-4}, \atr_{n-2}) \in \arbo$, which allows to ``short-circuit'' the cycle. Using the fact that $\sto \subset \arbo$, $\viso_0 \subset \arbo$, and $\arbo_0 \subset \arbo$, and applying the short-circuiting process multiple times, we obtain a cycle in the arbitration order $\arbo$ which contradicts the fact that $\arbo$ is a total order. \end{proof} \begin{proof}[Proof of Theorem~\ref{them:RobCcSi}] For the only-if direction: assume that $\aprog$ is robust against \ccc{} relative to \sic{}. Then, the set of traces of $\aprog$ under the two consistency models coincide. Since the set of traces under \sic{} is subset of the one under \pcc{}, then the set of traces under \ccc{} is subset of the one under \pcc{}. This implies that $\aprog$ is robust against \ccc{} relative to \pcc{}. % because the set of traces under \pcc{} is a subset of the set of executions under \ccc{}. Thus, we obtain that the set of traces of $\aprog$ under the three consistency models coincide. Therefore, $\aprog$ is robust against \pcc{} relative to \sic{} as well. For the if direction: assume that $\aprog$ is robust against \ccc{} relative to \pcc{} and $\aprog$ is robust against \pcc{} relative to \sic{}. Then, the set of traces of $\aprog$ under the three consistency models coincide. Thus, we obtain that $\aprog$ is robust against \ccc{} relative to \sic{}. \end{proof} %!TEX root = draft.tex \section{Proofs for Section~\ref{sec:commutativitygraph}} \label{sec:commutativitygraphProofs} \begin{proof}[Proof of Theorem~\ref{them:MovRobCcPc}] It is enough to show: if $\aprog$ is not robust against \ccc{} relative to \pcc{} then we have a simple cycle in the commutativity dependency graph of $\aprog_\pcinstr$ of the form above. Assume $\aprog$ is not robust against \ccc{} relative to \pcc{}. Then, from Theorem \ref{them:RobCcPc}, we obtain $\aprog_\pcinstr$ is not robust against \ccc{} relative to \serc{}. Also it was shown in [10] that if a program is not robust then there must exist a robustness violation trace (\ccc{} relative to \serc{}) $\atrace_\pcinstr$ of the shape $\atrace_\pcinstr = \alpha \cdot \atr_1 \cdot \beta \cdot \atr_i \cdot \atr_{i+1} \cdot \gamma \cdot \atr_n$ where $(\atr_1,\atr_i) \in (\po \cup \rfo)^{+}$, $(\atr_{i},\atr_{i+1}) \in (\sto \cup \cfo)$, $(\atr_{i+1},\atr_n) \in \hbo$, and $(\atr_n,\atr_1) \in \cfo$. Note that since transactions in the trace $\atrace_\pcinstr$ can either be read-only or write-only. Then, $(\atr_{i},\atr_{i+1}) \in (\sto \cup \cfo)$ and $(\atr_{n},\atr_1) \in \cfo$ imply that $\atr_1$ and $\atr_{i+1}$ must be a write-only transactions and $\atr_{n}$ must be a read-only transaction. Note that we may have $\beta = \gamma = \epsilon$ as the case for the trace of the $\mathsf{SB}$ program given in Figure \ref{fig:litmus1}. %In the trace $\atrace_\pcinstr$, we let $\atr_1$ to be $\atr_1$, $\atr_i$ to be $\atr_2$, $\atr_3$ to be $\atr_{i+1}$, and $\atr_4$ to be $\atr_n$ of Theorem \ref{them:MovRobCcPc}. We consider first the general case when $\atr_1 \not\equiv \atr_2$. The other case can be proved in the same way. Consider the prefix $\atrace_{p}$ of $\atrace_\pcinstr$: $\atrace_{p} = \alpha \cdot \atr_1 \cdot \beta \cdot \atr_{i}$ where $(\atr_1,\atr_{i}) \in (\po \cup \rfo)^{+}$ which is a \serc{} trace of $\aprog_{\pcinstr}$. Then, we have a sequence of transactions from $\atr_1$ to $\atr_{i}$ that are related by either $\po$ or $\rfo$. In the case two transactions are only related by $\rfo$, then the first transaction is not a right mover because of the second transaction reads from a write in the first transaction. Thus, we can relate the two transactions using the relation $\mrfo$ in the commutativity dependency graph. Similarly consider the following trace $\atrace_{s}$ extracted from $\atrace_\pcinstr$: $\atrace _{s} = \alpha \cdot \atr_{i+1} \cdot \gamma \cdot \atr_n$ where $(\atr_{i+1},\atr_n) \in \hbo$ which is a \serc{} trace of $\aprog_{\pcinstr}$. Similar to before, we have a sequence of transactions from $\atr_{i+1}$ to $\atr_n$ that are related by either $\po$, $\rfo$, $\sto$, or $\cfo$. For any two transactions that are related only by either $\rfo$, $\sto$, or $\cfo$, this implies that the first transaction is not a right mover because of the second transaction and a write-read, write-write, or read-write dependency between the two, respectively. Thus, we can relate the two transactions using either $\mrfo$, $\msto$, or $\mcfo$, respectively. Now consider the following trace $\atrace_{1}$ extracted from $\atrace_\pcinstr$: $\atrace_{1} = \alpha \cdot \atr_1 \cdot \beta \cdot \atr_{i} \cdot \atr_{i+1}$ where $(\atr_{i},\atr_{i+1}) \in (\sto \cup \cfo)$ is a \serc{} trace of $\aprog_{\pcinstr}$. Because $\atr_{i}$ and $\atr_{i+1}$ are related by either $\sto$ or $\cfo$, then $\atr_{i}$ is not a right mover because of $\atr_{i+1}$ and a write-write or read-write dependency between the two, respectively. Thus, we can relate the two transactions using either $\msto$ or $\mcfo$, respectively. Finally, consider the following trace $\atrace_{2}$ extracted from $\atrace_\pcinstr$: $\atrace_{2} = \alpha \cdot \atr_{i+1} \cdot \gamma \cdot \atr_n \cdot \atr_1$ where $(\atr_n,\atr_1) \in \cfo$ is a \serc{} trace of $\aprog_{\pcinstr}$. Because $\atr_{n}$ and $\atr_{1}$ are related by $\cfo$, then $\atr_{n}$ is not a right mover because of $\atr_{1}$ and a read-write dependency between the two. Thus, we can relate the two transactions using $\mcfo$. \end{proof} \begin{proof}[Proof of Theorem~\ref{them:MovRobPcSi}] Similar to before it is enough to show: if $\aprog$ is not robust against \pcc{} relative to \sic{} then we have a simple cycle in the commutativity dependency graph of $\aprog_\pcinstr$ of the form above. Assume $\aprog$ is not robust against \pcc{} relative to \sic{}. Then, from Theorem \ref{them:RobPcSiInstr}, we obtain that if $\sem{\aprog}$ reaches an error state under \serc{} then we will have the following trace $\atrace$ under \serc: $\atrace = \alpha \cdot \atrrd{\atr_{\instr}} \cdot \atr_3 \cdot \beta \cdot \atr_n \cdot \atrwr{\atr_{\instr}}$\footnote{For simplicity, we assume here that after reaching the error state we execute the writes of $\atr_{\instr}$, i.e., $\atrwr{\atr_{\instr}}$.} where $(\atrrd{\atr_{\instr}},\atr_3) \in \cfo$, $(\atr_3,\atr_n) \in \hbo$, $(\atr_n,\atrwr{\atr_{\instr}}) \in \sto$, and we don't have two successive $\cfo$ in the happens before between $\atr_3$ and $\atr_n$. In $\atrace$, $\atrwr{\atr_{\instr}}$ (resp., $\atrrd{\atr_{\instr}}$) represents $\atr_1$ (resp., $\atr_2$) in the theorem statement. Note that we may have $\alpha = \beta = \epsilon$ as is the case of the transformed $\mathsf{LU}$ program given in Figure \ref{fig:litmus2Instr}. The construction of the cycle in the commutativity dependency graph follows the same procedure taken in the proof of Theorem \ref{them:MovRobCcPc}. The only difference is that for every two transactions of $\atrace$ that are part of the happens before between $\atr_3$ and $\atr_n$, if the two are not connected by either $\po$, $\rfo$, $\sto$, or $\cfo$ then they must be the reads and writes of the same original transaction in $\aprog$. In this case, in the commutativity dependency graph we have the two transactions related by $\sametro$. %Notice that we abused notation by using the two components $\atrrd{\atr 1}$ and $\atrwr{\atr_1}$ of $\atr_1$ in $\atrace$ to denote that $\atr 1$ writes were not immediately written to the shared variables. \end{proof} \end{document}
# A quantum Boltzmann equation for strongly correlated electrons Antonio Picano Department of Physics, University of Erlangen-Nürnberg, 91058 Erlangen, Germany Jiajun Li Department of Physics, University of Erlangen- Nürnberg, 91058 Erlangen, Germany Martin Eckstein Department of Physics, University of Erlangen-Nürnberg, 91058 Erlangen, Germany ###### Abstract Collective orders and photo-induced phase transitions in quantum matter can evolve on timescales which are orders of magnitude slower than the femtosecond processes related to electronic motion in the solid. Quantum Boltzmann equations can potentially resolve this separation of timescales, but are often constructed within a perturbative framework. Here we derive a quantum Boltzmann equation which only assumes a separation of timescales (taken into account through the gradient approximation for convolutions in time), but is based on a non-perturbative scattering integral, and makes no assumption on the spectral function such as the quasiparticle approximation. In particular, a scattering integral corresponding to non-equilibrium dynamical mean-field theory is evaluated in terms of an Anderson impurity model in a non- equilibrium steady state with prescribed distribution functions. This opens the possibility to investigate dynamical processes in correlated solids with quantum impurity solvers designed for the study of non-equilibrium steady states. ## I Introduction One of the biggest challenges in the theoretical description of quantum many- particle systems is to predict their non-equilibrium dynamics at long times after a perturbation. This would be essential for the understanding of non- equilibrium phenomena in complex solids,Basov _et al._ (2017); Giannetti _et al._ (2016) including photo-induced metal-insulator transitions and hidden phases with spin, orbital, charge, or superconducting order.Ichikawa _et al._ (2011); Fausti _et al._ (2011); Beaud _et al._ (2014); Wegkamp _et al._ (2014); Stojchevska _et al._ (2014); Mor _et al._ (2017); Budden _et al._ (2020) The evolution of the electronic structure in these situations is often intertwined with the dynamics of the crystal lattice, collective orders, or slow electronic variables such as non-thermal band occupations, which is orders of magnitude slower than intrinsic electronic processes such as the electron tunnelling between atoms. Moreover, a large timescale separation becomes apparent in the thermalization of pre-thermal states,Polkovnikov _et al._ (2011); Berges _et al._ (2004); Moeckel and Kehrein (2008) where approximate conservation laws provide a dynamical constraint.Kollar _et al._ (2011); Langen _et al._ (2016) A major goal is therefore to devise an approach that can explore the dynamics on the slower timescale, while still taking into account accurately the fast degrees of freedom. Within the Keldysh formalism, non-equilibrium quantum many-particle systems can be described in terms of time and frequency- dependent spectral functions $A_{\bm{k}}(\omega,t)$ and distribution functions $F_{\bm{k}}(\omega,t)$. (For simplicity, spin and orbital indices in addition to momentum $\bm{k}$ are not shown here.) A large separation can become evident between the variation of the functions with time $t$, and the intrinsic timescales related to the linewidth of relevant spectral features. If these timescales are well separated, one can cast the full many-body dynamics into a differential equation known as the quantum Boltzmann equation (QBE).Kadanoff and Baym (1962); Kamenev (2011) In abstract form, the QBE defines a scattering contribution to the evolution of the distribution functions, $\displaystyle\big{[}\partial_{t}F_{\bm{k}}(\omega,t)\big{]}_{scatt}=I[F,A],$ (1) where the so-called scattering integral $I$ depends on the spectrum and distribution function at the same time. The full time-dependence is determined by additional contributions from the coherent single-particle propagation, and a separate equation for the evolution of the spectrum in terms of the distribution function. While the applicability of the QBE is in principle only controlled by the time-scale separation, the formalism is used predominantly for semiconductors or Fermi liquids with well-defined quasiparticles,Haug (1962) where one can make use of two additional and in practice rather important simplifications: (i) The quasiparticle approximation assumes the spectra $A_{\bm{k}}(\omega,t)$ to be sharply peaked at energies $\omega=\epsilon_{\bm{k}}$, and therefore allows to evaluate the QBE on-shell, essentially trading the frequency- dependent distribution function for quasiparticle occupations $n_{\bm{k}}(t)$. Moreover, (ii), the scattering kernel is often evaluated in a perturbative manner. In strongly correlated systems, both of these approximations are challenged. For example, doped Mott insulators show strange metallic behaviors without well-defined Fermi liquid quasiparticles in a wide parameter regime Georges _et al._ (2013); Deng _et al._ (2013), and similar behavior is observed in photo-doped Mott insulators.Eckstein and Werner (2013); Sayyad and Eckstein (2016); Dasari _et al._ (2020); Petersen _et al._ (2017); Sahota _et al._ (2019) Furthermore, the electronic structure in correlated systems depends strongly on the non-equilibrium distribution, as most clearly demonstrated through the possibility of photo-induced metal insulator transitions. For that reason, the dynamics of correlated systems has been mostly discussed within the formally exact non-equilibrium Green’s function (NEGF) techniques. In the NEGF formalism, the dynamics is described in terms of two-time Green’s functions $G_{\bm{k}}(t,t^{\prime})$, which are related to spectra and occupation functions through a Fourier transform with respect to relative time $t-t^{\prime}$. A two-time self-energy acts as a memory kernel in a non- Markovian propagation of the Green’s functions, the so-called Kadanoff-Baym equation. NEGF techniques can be combined with different diagrammatic approximations,Golež _et al._ (2016); Babadi _et al._ (2015); Rameau _et al._ (2016); Schlünzen _et al._ (2017) including in particular dynamical mean-field theory (DMFT),Aoki _et al._ (2014); Georges _et al._ (1996) and they do not rely on a quasiparticle approximation for the spectrum. On the other hand, they also do not make use of the time-scale separation, and therefore imply a high numerical cost: The effort scales like $\mathcal{O}(t_{\text{max}}^{3})$ with the simulation time $t_{\text{max}}$, as compared to $\mathcal{O}(t_{\text{max}})$ for the QBE. For weakly interacting systems, the perturbatively controlled generalized Kadanoff-Baym Ansatz (GKBA) Lipavský _et al._ (1986) has recently been set up to reach $\mathcal{O}(t_{\text{max}})$ scaling of the computational effort.Schlünzen _et al._ (2020) For strongly correlated systems, a systematic truncationSchüler _et al._ (2018) or compact compressionKaye and Golež (2020) of the memory kernel in the Kadanoff-Baym equations provide interesting perspectives, but so far the investigation of many fundamental questions has remained out of reach because of the $\mathcal{O}(t_{\text{max}}^{3})$ scaling. It would therefore be desirable to formulate a QBE which incorporates the simplifications due to the time-scale separation, but does not rely on quasiparticle or perturbative approximations. For example, if the equilibrium state of the system is described well by means of DMFT, the steady state fixed point of the QBE should be identical to this DMFT solution. A previous work has successfully employed a QBE without the quasiparticle approximation for a Mott insulator,Wais _et al._ (2018, 2020) assuming a rigid density of states and a renormalized second-order scattering integral. Here we show how such a scattering integral can be obtained from an auxiliary non-equilibrium steady state formalism. This allows to consistently combine the QBE with non- perturbative methods which have been developed to study true non-equilibrium steady states within DMFT.Joura _et al._ (2008); Li _et al._ (2015); Titvinidze _et al._ (2018); Matthies _et al._ (2018); Scarlatella _et al._ (2020); Li _et al._ (2020); Panas _et al._ (2019) The paper is organized as follows: In section II, we present the formulation of a non-perturbative QBE which is consistent with non-equilibrium DMFT. In Sec. III we compare its solution to a non-equilibrium DMFT simulation for the thermalization in a correlated metal. Section IV gives a conclusion and outlook. ## II Quantum Boltzmann equation ### II.1 General setting We will derive the QBE for a generic model, $\displaystyle H=\sum_{\bm{k},a,b}h_{\bm{k},ab}(t)c^{\dagger}_{\bm{k},a}c_{\bm{k},b}+H_{\text{int}},$ (2) where $c_{\bm{k},a}$ ($c_{\bm{k},a}^{\dagger}$) denotes the annihilation (creation) operator for a fermion with spin and orbital indices $a$ and momentum $\bm{k}$, and $H_{\text{int}}$ is an arbitrary two-particle interaction, and $h_{\bm{k},ab}(t)$ incorporates all single-particle terms. We assume that the system of interest is initially prepared in thermal equilibrium at temperature $T$, and driven out of equilibrium for times $t>0$ by external fields and a coupling to external heat and/or particle reservoirs. The description of this situation within many-body theory is based on contour- ordered Green’s functions, $\displaystyle G_{\bm{k},ab}(t,t^{\prime})=-i\langle T_{\mathcal{C}}c_{{\bm{k}},a}(t)c_{{\bm{k}},b}^{\dagger}(t^{\prime})\rangle,$ (3) with time arguments $t$ and $t^{\prime}$ on the Keldysh contour $\mathcal{C}$ that runs from 0 to time $t_{\text{max}}$ (the largest time of interest) on the real time axis, back to 0, and finally to $-i\beta$ along the imaginary time axis. (For an introduction to the Keldysh formalism and the notation, see, e.g., Ref. Aoki _et al._ , 2014.) Spin and orbital indices will be no longer shown in the following for simplicity; all Green’s functions, self- energies, dispersion functions $h_{\bm{k}}$ are matrices in these indices. From the contour-ordered function (3), one derives real and imaginary time Green’s functions, of which the retarded, lesser, and greater components are most important in the following. The retarded Green’s function (with real time arguments) $\displaystyle G^{R}_{\bm{k}}(t,t^{\prime})=-i\theta(t-t^{\prime})\langle[c_{\bm{k}}(t),c^{\dagger}_{\bm{k}}(t^{\prime})]_{+}\rangle,$ (4) is related to the spectral function of the system, while the occupied and unoccupied density of states are extracted from the lesser Green’s function, $\displaystyle G^{<}_{\bm{k}}(t,t^{\prime})=+i\langle c^{\dagger}_{\bm{k}}(t^{\prime})c_{\bm{k}}(t)\rangle.$ (5) $\displaystyle G^{>}_{\bm{k}}(t,t^{\prime})=-i\langle c_{\bm{k}}(t)c^{\dagger}_{\bm{k}}(t^{\prime})\rangle,$ (6) so that $\displaystyle G^{R}_{\bm{k}}(t,t^{\prime})=\theta(t-t^{\prime})[G^{>}_{\bm{k}}(t,t^{\prime})-G^{<}_{\bm{k}}(t,t^{\prime})].$ (7) In equilibrium, or in any time-translationally invariant state, all two-time correlation functions depend only on the relative time $t-t^{\prime}$. By taking the Fourier transform of $G^{R}$ with respect to this time difference, one obtains the spectral function $A$: $\displaystyle A_{\bm{k}}(\omega)=-\frac{1}{\pi}\imaginary G^{R}_{\bm{k}}(\omega+i0),$ (8) which is related to the lesser and greater Green’s functions through a fluctuation-dissipation theorem $\displaystyle G^{<}_{\bm{k}}(\omega)$ $\displaystyle=2\pi iA_{\bm{k}}(\omega)f_{\beta}(\omega),$ (9) $\displaystyle G^{>}_{\bm{k}}(\omega)$ $\displaystyle=-2\pi iA_{\bm{k}}(\omega)[1-f_{\beta}(\omega)],$ (10) where $f_{\beta}(\omega)$ is the Fermi distribution function, $f_{\beta}(\omega)=1/(e^{\beta\omega}+1)$. In a non-equilibrium steady-state, one can thus define the distribution function as the ratio $\displaystyle F_{\bm{k}}(\omega)=\frac{G^{<}_{\bm{k}}(\omega)}{2\pi iA_{\bm{k}}(\omega)}.$ (11) This is an energy distribution function, which is defined even in the absence of well-defined quasi-particles. The QBE provides an equation of motion for its time-dependent generalization, as introduced in the following. ### II.2 The QBE For every two-time quantity $X(t,t^{\prime})$ one can introduce the Wigner transform, $\displaystyle X(\omega,t)=\int ds\,e^{i\omega s}\,X(t+s/2,t-s/2),$ (12) where $t$ is the average time and $s=t-t^{\prime}$ is the relative time. In particular, this can be used to define a time-dependent spectrum and occupation function $F_{\bm{k}}(\omega,t)$ in analogy to Eqs. (7), (8), and (11), $\displaystyle A_{\bm{k}}(\omega,t)$ $\displaystyle=-\frac{1}{\pi}\imaginary G^{R}_{\bm{k}}(\omega+i0,t),$ (13) $\displaystyle=[G^{>}_{\bm{k}}(\omega,t)-G^{<}_{\bm{k}}(\omega,t)]/(-2\pi i),$ (14) $\displaystyle F_{\bm{k}}(\omega,t)$ $\displaystyle=G^{<}_{\bm{k}}(\omega,t)/[2\pi iA_{\bm{k}}(\omega,t)],$ (15) where $G^{R,<,>}_{\bm{k}}(\omega,t)$ are given by the Wigner transform. While Eqs. (13) to (15) always provide a valid mathematical definition, the functions gain a physical significance in particular in the limit in which there is a well-defined separation of timescales. Let us assume that there are scales $\delta\omega$ and $\delta t$ on which $G(\omega,t)$ varies in frequency and time, such that $\displaystyle\Big{|}\frac{\partial_{\omega}G_{\bm{k}}(\omega,t)}{G_{\bm{k}}(\omega,t)}\Big{|}<1/\delta\omega,\,\,\,\,\Big{|}\frac{\partial_{t}G_{\bm{k}}(\omega,t)}{G_{\bm{k}}(\omega,t)}\Big{|}<1/\delta t,$ (16) for lesser, greater, or retarded component. The scale $\delta\omega$ measures the relevant internal energy differences in the system, such as the linewidth or relevant spectral features, and $\delta t$ sets the scale for the time- evolution, with $\delta t\to\infty$ in a steady state. The QBE will be derived in the limit where these timescale are well separated, $\displaystyle\delta t\gg 1/\delta\omega.$ (17) This is also the limit in which the spectral and occupation functions gain their usual meaning in terms of a density of states: One can always approximate $G(\omega,t)$ by the average $\displaystyle G(\omega,t)\\!\approx\\!\\!\int\frac{dt^{\prime}d\omega^{\prime}}{\pi\Omega\tau}e^{-(\frac{t^{\prime}}{\tau})^{2}-(\frac{\omega^{\prime}}{\Omega})^{2}}G(\omega+\omega^{\prime},t+t^{\prime})$ (18) over a time interval $\tau\ll\delta t$ and a frequency interval $\Omega\ll\delta\omega$ on which the function varies weakly. With a sufficiently large time-scale separation (17), it is possible to choose $\Omega=1/\tau$ without violating the conditions $\tau\ll\delta t$ and $\Omega\ll\delta\omega$. With this, the average (18), with $G$ replaced by $-iG^{<}$, is the expression for the time-resolved photoemission spectrum Freericks _et al._ (2009); Eckstein and Kollar (2008) computed with a Gaussian probe pulse of duration $\tau$, and therefore has a well-defined interpretation in terms of an occupied density of states. In addition, this implies that the expression is real and positive, which can be proven by casting Eq. (18) in the form of a complete square using a Lehmann representation for the Green’s function. In the same way, $iG^{>}(\omega,t)$ can be interpreted as the unoccupied density of states (electron addition spectrum), and the spectral function $A(\omega,t)=[G^{>}(\omega,t)-G^{<}(\omega,t)]/(-2\pi i)$ has the usual meaning of a single-particle density of states in the many-body system. The QBE provides an equation of motion for the spectral and occupation functions (13) and (15) in the limit of well separated times.Kamenev (2011) Most importantly, the limit (17) allows for the simplification of the convolution $[A\ast B](t,t^{\prime})=\int d\bar{t}A(t,\bar{t})B(\bar{t},t^{\prime})$ of two real-time functions $A$ and $B$. In mathematical terms, the Wigner transform of the convolution is given by the Moyal product $\displaystyle[A\ast B](\omega,t)=e^{\frac{i}{2}[\partial_{t}^{A}\partial_{\omega}^{B}-\partial_{t}^{B}\partial_{\omega}^{A}]}A(\omega,t)B(\omega,t).$ (19) If Eqs. (16) and (17) hold for $A$ and $B$, the Moyal product can be simplified by considering only the leading term $\displaystyle[A\ast B](\omega,t)\approx A(\omega,t)B(\omega,t),$ (20) because $|\partial_{t}A\>\partial_{\omega}B|\ll|AB|$. This is the so-called gradient approximation. In a time-evolving state, Eq. (9) is generalized to the ansatz $\displaystyle G_{\bm{k}}^{<}(t,t^{\prime})=[F_{\bm{k}}\ast G_{\bm{k}}^{A}](t,t^{\prime})-[G_{\bm{k}}^{R}\ast F_{\bm{k}}](t,t^{\prime}),$ (21) where $F_{\bm{k}}(t,t^{\prime})$ depends on two-times, and $G_{\bm{k}}^{A}(t,t^{\prime})=G_{\bm{k}}^{R}(t^{\prime},t)^{\dagger}$ is the advanced Green’s function. By applying the gradient approximation (20) to this ansatz, we obtain the factorization $\displaystyle G^{<}_{\bm{k}}(\omega,t)=2\pi iA_{\bm{k}}(\omega,t)F_{\bm{k}}(\omega,t),$ (22) equivalent to Eq. (15), using $G^{A}_{\bm{k}}(\omega,t)=G^{R}_{\bm{k}}(\omega,t)^{\dagger}$. In order to derive the QBE for the evolution of the distribution function $F_{\bm{k}}$, one can consider the equations of motion for the Green’s function. For a non-interacting system with Green’s function $\mathcal{G}_{\bm{k}}(t,t^{\prime})=-i\langle\mathcal{T}_{\mathcal{C}}c_{\bm{k}}(t)c_{\bm{k}}^{\dagger}(t^{\prime})\rangle$ this is written as $\displaystyle\\{\mathcal{G}^{-1}_{\bm{k}}\ast\mathcal{G}_{{\bm{k}}}\\}(t,t^{\prime})=\delta_{\mathcal{C}}(t,t^{\prime}),$ (23) $\displaystyle\mathcal{G}^{-1}_{\bm{k}}(t,t^{\prime})=[i\partial_{t}+\mu- h_{\bm{k}}(t)]\delta_{\mathcal{C}}(t,t^{\prime}),$ (24) where $\delta_{\mathcal{C}}(t,t^{\prime})$ represents the delta-function on the Keldysh contour and $\mu$ is the chemical potential of the system. In the following, it will be convenient to also include the Hartree and Fock self- energy into the dispersion $h_{\bm{k}}(t)$. To include correlations we take into account the contour-ordered self-energy $\Sigma(t,t^{\prime})$ and obtain the interacting Green’s function $G$ via the Dyson equation $\displaystyle\\{[(\mathcal{G}_{\bm{k}})^{-1}-\Sigma_{\bm{k}}]\ast G_{\bm{k}}\\}(t,t^{\prime})=\delta_{\mathcal{C}}(t,t^{\prime})$ (25) on the Keldysh contour. From the Dyson equation for the lesser component, [$(\mathcal{G}^{R}_{\bm{k}})^{-1}-\Sigma^{R}_{\bm{k}}]\ast G_{\bm{k}}^{<}=\Sigma^{<}_{\bm{k}}\ast G_{\bm{k}}^{A}$, and the ansatz (21), we get $\displaystyle(\mathcal{G}^{R}_{\bm{k}})^{-1}\ast F_{\bm{k}}-$ $\displaystyle F_{\bm{k}}\ast(\mathcal{G}^{A}_{\bm{k}})^{-1}=$ $\displaystyle=\Sigma^{<}_{\bm{k}}+\Sigma^{R}_{\bm{k}}\ast F_{\bm{k}}-F_{\bm{k}}\ast\Sigma^{A}_{\bm{k}}.$ (26) (Real-time time arguments are shown only where otherwise ambiguous.) We thus obtain the equation of motion for $F_{\bm{k}}(t,t^{\prime})$: $\displaystyle i(\partial_{t}+\partial_{t^{\prime}})F_{\bm{k}}(t,t^{\prime})=$ $\displaystyle h_{\bm{k}}(t)F_{\bm{k}}(t,t^{\prime})-F_{\bm{k}}(t,t^{\prime})h_{\bm{k}}(t^{\prime})$ $\displaystyle+\Sigma^{<}_{\bm{k}}+\Sigma^{R}_{\bm{k}}\ast F_{\bm{k}}-F_{\bm{k}}\ast\Sigma^{A}_{\bm{k}}.$ (27) Equation (II.2) is still exact. To obtain the QBE, we then use the gradient approximation (20) to rewrite Eq. (II.2) as $\displaystyle\partial_{t}F_{\bm{k}}(\omega,t)$ $\displaystyle=-i[h_{\bm{k}}(t),F_{\bm{k}}(\omega,t)]+I_{\bm{k}}(\omega,t),$ (28) $\displaystyle I_{\bm{k}}(\omega,t)$ $\displaystyle=-i\big{[}\Sigma_{\bm{k}}^{R}(\omega,t)F_{\bm{k}}(\omega,t)-F_{\bm{k}}(\omega,t)\Sigma_{\bm{k}}^{A}(\omega,t)$ $\displaystyle\,\,\,\,\,\,\,+\Sigma^{<}_{\bm{k}}(\omega,t)\big{]},$ (29) where $I_{\bm{k}}(\omega,t)$ is the scattering integral. This equation is completed by the Dyson equation for the retarded Green’s function to leading order in the gradient approximation, $\displaystyle G^{R}_{\bm{k}}(\omega,t)$ $\displaystyle=[\omega+i0+\mu- h_{\bm{k}}(t)-\Sigma_{\bm{k}}^{R}(\omega,t)]^{-1}.$ (30) This set of equations must be combined with a given expression for the self- energy. For example, a simple perturbative expression would be a second-order diagram in terms of a two-particle density-density interaction $v_{\bm{q}},$ $\displaystyle\Sigma_{\bm{k}}(t,t^{\prime})=\sum_{\bm{k}^{\prime},\bm{q}}v_{\bm{q}}^{2}G_{\bm{k}^{\prime}+\bm{q}}(t,t^{\prime})G_{\bm{k}^{\prime}}(t^{\prime},t)G_{\bm{k}^{\prime}-\bm{q}}(t,t^{\prime}).$ (31) Such an analytic perturbative expression for $\Sigma$ can then be evaluated in the gradient approximation, thus closing the equation. In the following, we discuss a strategy to incorporate a non-perturbative self-energy approximation like DMFT into the QBE formalism, in which an explicit analytical expression for $\Sigma$ is not given. ### II.3 Non-perturbative evaluation of the scattering integral In general, the self-energy includes contributions from the interaction, and a possible coupling to a noninteracting environment, which can be used to represent thermal and particle reservoirs Tsuji _et al._ (2009); Büttiker (1985); Aoki _et al._ (2014). In the following, we write $\Sigma=\Sigma_{\text{int}}+\Gamma$, where $\Sigma_{\text{int}}$ is the interaction contribution, and $\Gamma$ represents the noninteracting reservoirs. Evaluating the interaction self-energy is the main challenge. We assume that the interaction self-energy $\Sigma_{\text{int}}(t,t^{\prime})=\hat{\Sigma}_{\bm{k},t,t^{\prime}}^{\text{skel}}[G]$ is a functional of the full Green’s function $G$, as obtained in particular as the so-called skeleton expansion through derivatives of the Luttinger-Ward functionalLuttinger and Ward (1960) for any conserving approximation.Baym and Kadanoff (1961) Also DMFT and its extensions can be cast in this language.Georges _et al._ (1996) A simple perturbative example would be the second-order diagram Eq. (31). Let us now imagine a system which has the same interaction but general non-interacting reservoirs so that the system resides in a non-equilibrium steady state (NESS) with steady state spectrum $\bar{A}_{\bm{k}}(\omega)$, and the steady state distribution $\bar{F}_{\bm{k}}(\omega)$. Evaluation of the full skeleton functional $\hat{\Sigma}_{\bm{k},t,t^{\prime}}^{\text{skel}}[G]$ at the translationally invariant Green’s function $\bar{G}[\bar{A},\bar{F}\big{]}$ defines a non- equilibrium steady-state functional through the Wigner transform (12) $\displaystyle\hat{\Sigma}^{\text{ness- skel}}_{\bm{k},\omega}\big{[}\bar{A},\bar{F}\big{]}=\int ds\,e^{i\omega s}\,\hat{\Sigma}_{\bm{k},s/2,-s/2}^{\text{skel}}[\bar{G}].$ (32) This skeleton functional is universal in the sense that it parametrically depends only on the interaction,Potthoff (2003) but not on the single-particle part of the Hamiltonian, and hence the functional (32) is independent of the choice of the reservoirs. In order to write the equations below in a more compact form, we note that the self-consistent evaluation of the functional (32), together with the steady state Dyson equation for the retarded function $\displaystyle\bar{A}_{\bm{k}}(\omega)$ $\displaystyle=-\frac{1}{\pi}\text{Im}\frac{1}{\omega^{+}+\mu-\bar{h}_{\bm{k}}-\bar{\Gamma}^{R}_{\bm{k}}(\omega)-\bar{\Sigma}^{R}_{\text{int},\bm{k}}(\omega)}$ (33) and given $\bar{h}_{\bm{k}}$ and $\bar{\Gamma}^{R}_{\bm{k}}(\omega)$, implicitly defines a steady-state functional of the self-energy and the spectral function in terms of the distribution function only, which we will denote by $\displaystyle\hat{\Sigma}_{\bm{k},\omega}^{\text{ness}}\big{[}\bar{F};\bar{h}_{\bm{k}},\bar{\Gamma}^{R}_{\bm{k}}\big{]},\,\,\,\,\hat{A}_{\bm{k},\omega}^{\text{ness}}\big{[}\bar{F};\bar{h}_{\bm{k}},\bar{\Gamma}^{R}_{\bm{k}}\big{]}.$ (34) Back to the QBE, at each order of a diagrammatic expression, the two-time self-energy $\Sigma_{\text{int}}(t,t^{\prime})$ can be written as a sum of convolutions and products of the full Green’s function $G$. In each of these terms, one can consistently use the leading order of the gradient approximation, in combination with the factorization (22). This procedure would be the same as evaluating $\hat{\Sigma}_{\text{int},\bm{k},t,t^{\prime}}^{\text{skel}}[\bar{G}]$ with a time-translationally invariant function $\bar{G}$ with spectral function $\bar{A}_{\bm{k}}(\omega)=A_{\bm{k}}(\omega,t)$ and distribution function $\bar{F}_{\bm{k}}(\omega)=F_{\bm{k}}(\omega,t)$. Hence the self-energy in the gradient approximation amounts to evaluating the NESS functional (32) $\displaystyle\Sigma_{\text{int},\bm{k}}(\omega,t)=\hat{\Sigma}^{\text{skel- ness}}_{\bm{k},\omega}\big{[}A(\cdot,t),F(\cdot,t)\big{]}.$ (35) Here the notation $X(\cdot,t)$ of the functional arguments $X=A,F$ indicates that the latter are considered as function of all their arguments except for $t$, which is considered as a fixed parameter. With Eq. (34), the QBE is now formally written as $\displaystyle\partial_{t}F_{\bm{k}}(\omega,t)$ $\displaystyle=-i[h_{\bm{k}}(t),F_{\bm{k}}(\omega,t)]+I_{\bm{k},\omega}[F(\cdot,t)],$ (36) $\displaystyle I_{\bm{k},\omega}[F(\cdot,t)]$ $\displaystyle=-i\big{[}\Sigma_{\bm{k}}^{R}(\omega,t)F_{\bm{k}}(\omega,t)-F_{\bm{k}}(\omega,t)\Sigma_{\bm{k}}^{A}(\omega,t)$ $\displaystyle\,\,\,\,\,\,\,+\Sigma^{<}_{\bm{k}}(\omega,t)\big{]},$ (37) where in the second line $\Sigma=\Sigma_{\text{int}}+\Gamma$, with $\displaystyle\Sigma_{\text{int},\bm{k}}(\omega,t)$ $\displaystyle=\hat{\Sigma}_{\bm{k},\omega}^{\text{ness}}\big{[}F(\cdot,t);h_{\bm{k}}(t),\Gamma^{R}_{\bm{k}}(\cdot,t)\big{]}.$ (38) In addition, the spectral function is given by $\displaystyle A_{\bm{k}}(\omega,t)$ $\displaystyle=\hat{A}_{\bm{k},\omega}^{\text{ness}}\big{[}F(\cdot,t);h_{\bm{k}}(t),\Gamma^{R}_{\bm{k}}(\cdot,t)\big{]}.$ (39) Physically, the last equation (39) means that we allow the electronic distribution function to instantaneously influence the electronic structure of the material. We will therefore refer to Eq. (39) as the instantaneous response approximation. Equations (36) to (39) now provide a closed set of time-dependent equations. This implicit scheme allows a non-perturbative evaluation of the QBE, provided that an efficient numerical description of a NESS is available: To evaluate $\hat{\Sigma}_{\bm{k},\omega}^{\text{ness}}\big{[}F(\cdot,t),...\big{]}$ and $A_{\bm{k}}(\omega,t)=\hat{A}_{\bm{k},\omega}^{\text{ness}}\big{[}F(\cdot,t),...\big{]}$ for a given distribution function $\bar{F}$, we choose an auxiliary steady state system with reservoir self-energy $\bar{\Gamma}^{R}_{\bm{k}}(\omega)=\Gamma^{R}_{\bm{k}}(\omega,t)$, while the bath occupation function, and hence $\bar{\Gamma}^{<}_{\bm{k}}(\omega)$ is treated as a free parameter. The latter is chosen such that the solution $\bar{F}_{\bm{k}}(\omega)$ gives the prescribed $F_{\bm{k}}(\omega,t)$, after which the outcomes $\bar{A}_{\bm{k}}(\omega)$ and $\bar{\Sigma}_{\text{int},\bm{k}}(\omega)$ are used to evaluate (38) and (39). In particular, within non-equilibrium DMFT, where only local self-energies need to be evaluated in a quantum impurity model, several promising non- perturbative techniques are available that can directly target such non- equilibrium states (see discussion in Sec. IV). Once Eqs. (38) and (39) can be evaluated for a given $F$, the QBE Eq. (36) can be solved as any differential equation. (In the implementation below, we use a simple Runge-Kutta algorithm.) In the following two sections, we will adapt the general formalism to the non- equilibrium DMFT framework. Before that, we conclude this section with a side remark: It is known even in equilibrium that the self-consistent solution of the Dyson equation with a skeleton self energy functional can have multiple unphysical solutions.Kozik _et al._ (2015) However, a possible multi- valuedness of the functional (34) will not be a problem here. The functions $A_{\bm{k}}(\omega,t)$, $F_{\bm{k}}(\omega,t)$, and $\Sigma_{\bm{k}}(\omega,t)$ evolve continuously as a function of time, so that even if unphysical steady-state solutions exist for a given distribution function, the physical solution is always selected by the requirement of continuity and the initial condition. On the other hand, if the system would evolve as a function of time into a branching point where multiple solutions of Eq. (34) meet, this would hint at a rather unconventional dynamical behavior. For example, in equilibrium it is known that the multi-valuedness of self-consistent perturbation theory is related to vertex singularities Schäfer _et al._ (2013), and in the Hubbard model these vertex singularities apparently fall together with the dynamical critical point found in Ref. Eckstein _et al._ , 2009. ### II.4 Scattering integral in DMFT In the following, we adapt the general QBE framework to non-equilibrium DMFT. Within DMFT, one maps the lattice model (2) onto an effective single-site impurity model. The impurity site has the same interaction as a site in the lattice, and its coupling to the environment is described by the so-called hybridization function $\Delta(t,t^{\prime})$, which is self-consistently determined such that the local ($\bm{k}$-averaged) lattice Green’s function $\displaystyle G_{\text{loc}}(t,t^{\prime})=\sum_{\bm{k}}G_{\bm{k}}(t,t^{\prime})$ (40) coincides with the impurity Green’s function. The key approximation of DMFT is that the lattice self-energy is local in space (independent of $\bm{k}$), and one requires the local lattice self-energy to be identical to the impurity self energy. In detail, the impurity model is defined by an action $\displaystyle\mathcal{S}=-i\int_{\mathcal{C}}dt\,H_{loc}(t)-i\int_{\mathcal{C}}dtdt^{\prime}\sum_{\sigma}c_{\sigma}^{\dagger}(t)\Delta(t,t^{\prime})c_{\sigma}(t^{\prime}),$ (41) in terms of the self-consistent hybridization function. The non-interacting Green’s function $\mathcal{G}$ is determined by the Dyson equation $\displaystyle\mathcal{G}^{-1}(t,t^{\prime})=[i\partial_{t}+\mu-h(t)]\delta_{\mathcal{C}}(t,t^{\prime})-\Delta(t,t^{\prime}),$ (42) where $h(t)$ is the single particle Hamiltonian in the impurity model. The interacting impurity Green’s function is given by $\displaystyle G_{\text{imp}}^{-1}=\mathcal{G}^{-1}-\Sigma_{\text{imp}},$ (43) and the self-consistency requires $\displaystyle G_{\text{imp}}=G_{\text{loc}},\,\,\,\,\Sigma_{\text{imp}}=\Sigma.$ (44) The self-consistent impurity model provides an implicit way to evaluate a non- perturbative expression $\hat{\Sigma}_{\text{int}}[G_{\text{loc}}]$ for a local self-energy in terms of a local Green’s functions. Along the line of the previous section, we can therefore use an impurity model in a NESS to construct the steady state functional (38) for the local self-energy. An impurity model in the steady state simply implies that the hybridization function itself is translationally invariant in time, and specified through its retarded and lesser components, $\Delta^{R}(\omega)$ and $\Delta^{<}(\omega)$. The evaluation of the functionals (38) and (39) within DMFT, for a given distribution function $\bar{F}_{\bm{k}}(\omega)$, depends on the type of impurity solver. Below we exemplify this for an impurity solver which determines the self energy from an expansion in terms of the noninteracting impurity Green’s function $\bar{\mathcal{G}}$, (such as weak-coupling Keldysh quantum Monte Carlo or iterated perturbation theory): * 1) Start with some guess for $\bar{\Sigma}_{\text{int}}^{R}(\omega)$ and $\bar{\Sigma}_{\text{int}}^{<}(\omega)$, and calculate the $\bm{k}$-dependent lattice Green’s functions [Eq. (30) with $\bm{k}$-independent self-energy] $\displaystyle\bar{G}_{\bm{k}}^{R}(\omega)=[\omega+\mu-\bar{h}_{\bm{k}}-\bar{\Gamma}_{\bm{k}}^{R}(\omega)-\bar{\Sigma}_{\text{int}}^{R}(\omega)]^{-1}.$ (45) and the spectrum $\bar{A}_{\bm{k}}(\omega)=-\frac{1}{\pi}\text{Im}G_{\bm{k}}^{R}(\omega+i0)$. * 2) Determine the lesser Green’s function from the given distribution function, $\displaystyle\bar{G}_{\bm{k}}^{<}(\omega)=2\pi i\bar{F}_{\bm{k}}(\omega)\bar{A}_{\bm{k}}(\omega).$ (46) * 3) Calculate the local lattice Green’s functions. $\displaystyle\bar{G}_{\text{loc}}^{R,<}(\omega)$ $\displaystyle=\sum_{\bm{k}}\bar{G}_{\bm{k}}^{R,<}(\omega,t).$ (47) * 4) Express the noninteracting Green’s function $\mathcal{G}$ of the impurity model in terms of $\Sigma_{\text{imp}}$ of $G_{\text{imp}}$ using the Dyson equation for the impurity model [Eqs. (42) and (43)] in the steady state. For example, this can be written as $\displaystyle\mathcal{G}^{R}(\omega)=[G_{\text{imp}}^{R}(\omega)^{-1}+\Sigma_{\text{imp}}^{R}(\omega)]^{-1},$ (48) $\displaystyle\Delta^{<}(\omega)=G^{R}_{\text{imp}}(\omega)^{-1}G^{<}_{\text{imp}}(\omega)G^{A}_{\text{imp}}(\omega)^{-1}-\Sigma^{<}_{\text{imp}}(\omega)$ $\displaystyle\mathcal{G}^{<}(\omega)=\mathcal{G}^{R}(\omega)\Delta^{<}\mathcal{G}^{A}(\omega),$ (49) Solve these equations for $\mathcal{G}(\omega)$ using the DMFT self- consistency for the lattice and impurity quantities, $\Sigma_{\text{imp}}(\omega)=\bar{\Sigma}_{\text{int}}(\omega)$ and $G_{\text{imp}}(\omega)=\bar{G}_{\text{loc}}(\omega)$. * 5) Calculate a new $\Sigma_{\text{imp}}$ by using an expansion in $\mathcal{G}^{R}(\omega)$. * 6) Set $\bar{\Sigma}^{R,<}_{\text{int}}(\omega)=\Sigma^{R,<}_{\text{imp}}(\omega)$, and iterate Step 2) to 5) until convergence. This iteration is basically a steady-state non-equilibrium DMFT simulation where the distribution function of the system is prescribed and the distribution of the reservoirs is determined, in contrast to conventional steady-state DMFT where the distribution function of the system of the system is determined by reservoirs with a given distribution function. ## III Comparison to the full DMFT simulation ### III.1 Model As a first test case for the methodology, we study the particle-hole symmetric single-band Hubbard model $\displaystyle\hat{H}=-t_{h}\sum_{\langle i,j\rangle,\sigma}c^{\dagger}_{i\sigma}c_{j\sigma}+U\sum_{j}\big{(}\hat{n}_{j\uparrow}-\tfrac{1}{2}\big{)}\big{(}\hat{n}_{j\downarrow}-\tfrac{1}{2}\big{)}.$ (50) Here $c_{j,\sigma}$ denotes the annihilation operator for a Fermion with spin $\sigma\in\\{\uparrow,\downarrow\\}$ at lattice site $j$, $\hat{n}_{j\sigma}=c^{\dagger}_{j\sigma}c_{j\sigma}$ is the particle number operator, $t_{h}$ the hopping matrix element between nearest neighbour sites, and $U$ the on-site interaction strength. The actual simulations assume a semi-elliptic local density of states $D(\epsilon)=\sqrt{4-\epsilon^{2}}/(2\pi)$ for the noninteracting model with bandwidth $4$, corresponding to a Bethe lattice with hopping $t_{h}=1$. The latter sets the unit of energy, and its inverse defines the unit of time ($\hbar=1$). The system is studied in the metallic regime, where $U$ is smaller than the bandwidth. Initially, the system is in equilibrium with a inverse temperature $\beta$. Within a short time interval, we then create a non-thermal population of electrons and holes similar to a photo-excited population (the precise protocol is given below). This non-thermal population will then relax under the influence of the electron-electron interaction and the coupling to a phonon bath, and we compare a simulation of this relaxation dynamics within the full non-equilibrium DMFT simulation and the QBE. For the excitation, we shortly couple a fermionic reservoir with density of states $\displaystyle A_{\text{bath}}(\omega)=A(\omega-2.5)+A(\omega+2.5)$ (51) consisting of two smooth bands with bandwidth $W_{\text{bath}}=6$ around the energies $\omega=\pm 2.5$; we choose $A(\omega)=\frac{1}{\pi}\cos[2](\pi\omega/W_{\text{bath}})$ in the interval $[-W_{\text{bath}}/2,W_{\text{bath}}/2]$, see dashed line at the bottom of Fig 1c for $A_{bath}(\omega)$. Choosing a population inversion in this reservoir will lead to a rapid transfer of electrons from the system into the negative energy part of the reservoir, and of electrons from the positive energy part of the bath to the system, thus generating an electron transfer similar to a photo-excitation process. The bath adds a local contribution $\Gamma(t,t^{\prime})$ to the self-energy (as obtained by integrating out the bath), $\displaystyle\Gamma(t,t^{\prime})=V(t)G_{\text{bath}}(t,t^{\prime})V(t^{\prime})^{*},$ (52) where $V(t)$ is the time-profile of the coupling, and $G_{\text{bath}}(t,t^{\prime})$ is the bath Green’s function, $\displaystyle G_{\text{bath}}^{R}(t,t^{\prime})$ $\displaystyle=-i\theta(t-t^{\prime})\int d\omega\,e^{-i\omega(t-t^{\prime})}A_{\text{bath}}(\omega),$ (53) $\displaystyle G_{\text{bath}}^{<}(t,t^{\prime})$ $\displaystyle=i\int d\omega\,e^{-i\omega(t-t^{\prime})}f_{\text{bath}}(\omega)A_{\text{bath}}(\omega).$ (54) The bath occupation $f_{\text{bath}}(\omega)=f_{-\beta}(\omega)$ is taken to be, during the whole time-evolution of the system, a negative temperature Fermi-Dirac distribution (population inversion) , and the switching profile $V(t)=0.75\sin[2](\pi/5(t-t_{0}))$ is centred around an early time $t_{0}=27.5$ with a duration of just five inverse hoppings. In general, the QBE is expected to describe the evolution of the system only on timescales much longer than the inverse hopping, so that these details of the excitation protocol are not important for the present study. The coupling to the bosonic bath is included via a local electron-phonon self- energy $\Sigma_{\text{ph}}$. In order for the bosons to act as heath bath, we need to neglect the back-action of the electrons on the phonons, and we take $\Sigma_{\text{ph}}$ to be the simple first-order diagram of a local electron- phonon interaction, $\displaystyle\Sigma_{\text{ph}}(t,t^{\prime})$ $\displaystyle=g^{2}G(t,t^{\prime})D_{\text{ph}}(t,t^{\prime}),$ (55) where $G$ is the fully interacting local electron Green’s function of the system, $g$ measures the electron-phonon coupling strength, and $D_{\text{ph}}$ is the propagator for free bosons with an Ohmic density of states $\frac{\omega}{4\omega_{\text{ph}}^{2}}\exp(-\omega/\omega_{\text{ph}})$ with exponential cutoff $\omega_{\text{ph}}=0.2$. The occupation function of bosons is kept in equilibrium with inverse temperature $\beta$. The temperature of the heat bath is the same as the initial one of the system in equilibrium, such that the system will eventually thermalize back to its initial temperature long after the excitation. ### III.2 Full DMFT solution For the semi-elliptic density of states, the DMFT self-consistency can be formulated in closed form, and the hybridization of the impurity model is simply given by Georges _et al._ (1996); Aoki _et al._ (2014) $\displaystyle\Delta(t,t^{\prime})=G(t,t^{\prime})+\Gamma(t,t^{\prime})$ (56) in terms of the local Green’s function $G$. With the non-interacting Green’s function of the impurity model [Eq. (42)], the Dyson equation for the impurity model reads $\displaystyle G^{-1}(t,t^{\prime})=\mathcal{G}^{-1}(t,t^{\prime})-\Sigma_{\text{int}}(t,t^{\prime}).$ (57) Here $\displaystyle\Sigma_{\text{int}}(t,t^{\prime})=\Sigma_{U}(t,t^{\prime})+\Sigma_{\text{ph}}(t,t^{\prime})$ (58) is the interaction self-energy due to the electron phonon interaction and the Hubbard interaction. The latter is determined using the iterated perturbation theory (IPT) impurity solver, i.e., a second-order expansion in terms of $\mathcal{G}$, $\displaystyle\Sigma_{U}(t,t^{\prime})=U^{2}\mathcal{G}(t,t^{\prime})\mathcal{G}(t,t^{\prime})\mathcal{G}(t^{\prime},t).$ (59) In addition, the local energy $h(t)$ in Eq. (42) is the Hartree self-energy, $h(t)=Un_{\sigma}(t)$ with the density $n_{\sigma}(t)$ per spin. In the present case we study a half-filled system, so that $\mu=U/2$ and $\mu+h(t)=0$. The self-consistent solution of the system of Eq. (56) to (59) together with the excitation and phonon self energies Eq. (55) and Eq. (52) determines the time evolution of the physical system. The equations are solved on the Keldysh contour using the NESSi simulation package.Schüler _et al._ (2020) For the comparison with the QBE, the local spectral function and distribution function are then extracted from the Wigner transform of the local Green’s function $\displaystyle A(\omega,t)$ $\displaystyle=-\frac{1}{\pi}\text{Im}G^{R}(\omega+i0,t),$ (60) $\displaystyle F(\omega,t)$ $\displaystyle=\frac{G^{<}(\omega,t)}{2\pi iA(\omega,t)}.$ (61) Furthermore, we compute the total energy as: $\displaystyle E_{\text{DMFT}}=-2i(\Delta*G)^{<}(t,t)-i(\Sigma_{\text{int}}*G)^{<}(t,t)$ (62) The first and second term represent the kinetic and interaction energy, respectively, with a factor two in the kinetic energy for the summation over spin components. ### III.3 QBE formulation For the present model, for which a closed set of equations is given in terms of local (momentum-averaged) quantities, the QBE can be derived directly for the local quantities. Instead of deriving Eq. (28) and (29) from the lattice Dyson equation (25), one can perform an analogous argument directly for the Dyson equation of the DMFT impurity model [Eq. (57)]. This leads to a local QBE $\displaystyle\partial_{t}F(\omega,t)=$ $\displaystyle\,I[F(\cdot)],$ (63) $\displaystyle I[F(\cdot)]=$ $\displaystyle-i\big{(}\Sigma^{<}(\omega,t)+[\Sigma^{R}(\omega,t)+\Delta^{R}(\omega,t)]F(\omega,t)+$ $\displaystyle-F(\omega,t)[\Sigma^{A}(\omega,t)+\Delta^{A}(\omega,t)]\big{)},$ (64) where again $\Sigma=\Gamma+\Sigma_{\text{int}}$, and $\displaystyle\Sigma_{\text{int}}(\omega,t)=$ $\displaystyle\Sigma_{\omega}^{\text{ness}}[F(\cdot,t)],\,\,A(\omega,t)=A_{\omega}^{\text{ness}}[F(\cdot,t)],$ (65) $\Sigma(\omega,t)$ and the spectrum $A(\omega,t)$ are understood in terms of an auxiliary steady state impurity model with given prescribed distribution function $\bar{F}(\omega)=F(\omega,t)$. The evaluation of these functionals is again done iteratively: 1. 1) Start from a guess for $\bar{\Sigma}_{\text{int}}(\omega)$. Solve the steady state variant of Eq. (57) for $\bar{G}^{R}(\omega)$, $\displaystyle\bar{G}^{R}(\omega)=[\omega+\mu-\bar{h}-\Delta^{R}(\omega)-\bar{\Sigma}_{\text{int}}^{R}(\omega)]^{-1}.$ (66) and determine $\bar{A}(\omega)=-\frac{1}{\pi}\bar{G}^{R}(\omega+i0)$. 2. 2) Determine the lesser Green’s function from the given distribution function, $\bar{G}^{<}(\omega)=2\pi i\bar{F}(\omega)\bar{A}(\omega)$. 3. 3) Use the self-consistency Eq. (56) to fix the hybridization function of the effective steady state impurity model, $\Delta(\omega)=\bar{G}(\omega)+\Gamma(\omega)$. 4. 4) Solve the impurity model. With IPT as an impurity solver, we first determine $\mathcal{G}(\omega)$ from $\Delta(\omega)$, $\displaystyle\mathcal{G}^{R}(\omega)=[\omega+\mu-h(t)-\Delta^{R}(\omega)]^{-1},$ (67) $\displaystyle\mathcal{G}^{<}(\omega)=\mathcal{G}^{R}(\omega)\Delta^{<}(\omega)\mathcal{G}^{A}(\omega),$ (68) transform to real time, evaluate Eq. (59), and transform back to frequency space to obtain $\Sigma_{U}^{R,<}(\omega)$. Similarly, $\Sigma_{\text{ph}}^{R,<}(\omega)$ is evaluated. 5. 5) Set $\bar{\Sigma}_{\text{int}}(\omega)=\Sigma_{U}(\omega)+\Sigma_{\text{ph}}(\omega)$, and iterate step 2) to 5) until convergence. The iteration serves as a way to evaluate $\Sigma^{\text{ness}}[F(\cdot,t)]$. The differential equation (63) is then solved using a Runge-Kutta algorithm. In addition to the spectral and distribution functions, we then compute the total energy $\displaystyle E_{\text{QBE}}=-2i[\Delta(\omega)G(\omega)]^{<}-i[\Sigma_{\text{int}}(\omega)G(\omega)]^{<}$ (69) in order to compare with the full solution (62). ### III.4 Results and Discussion Figure 1: a) Energy $-E_{\text{DMFT}}$ obtained from the full DMFT solution (interaction $U=3$, initial inverse temperature $\beta=20$, electron phonon coupling $g^{2}=0.5$). Coloured dots indicate the energies obtained from the auxiliary steady state $A_{\omega}^{\text{ness}}[F]$ [Eq. (65)], for different initial times $t_{0}$ at which the distribution functions $F(\omega,t_{0})$ is taken from the DMFT solution and copied in the auxiliary steady state problem. b) Distribution functions $F(\omega,t_{0})$ obtained from the full DMFT solution, at times $t_{0}$ corresponding to the dots in a), and copied in the auxiliary steady state problem. c) Dashed lines show the spectrum $A(\omega,t_{0})$ at various initial times, obtained from the full DMFT solution. Solid lines show the spectra obtained from the auxiliary steady state $A_{\omega}^{\text{ness}}[F]$ [Eq. (65)], evaluated with the distribution functions $F(\omega,t_{0})$ in b) taken from the DMFT solution. The dotted line at the bottom of c) shows the (rescaled) spectral function $A_{\text{bath}}(\omega)$ [Eq. (51)] for the excitation bath (the shaded orange area shows the occupied density of states for the bath), and the shaded area in a) the time window over which this bath is coupled to the system. In this subsection, we compare the QBE description with the full solution of the KB equations for the setting introduced in Sec. III.1. Figure 1a) shows the evolution of the energy in the full DMFT solution, which increases during the short excitation window, and subsequently relaxes back to the initial state due to electron thermalization and the electron-phonon interaction. Figure 1b) and c) then show the spectra and distribution functions at some points in time. In the initial and final state the spectrum has a central peak, representing a band of renormalized quasiparticles, which coexists with two Hubbard bands around $\omega=\pm U/2$. In equilibrium, with increasing $T$, the quasiparticle peak would be replaced by a dip in the spectral function, indicating that the high-temperature state is a bad-metal without coherent quasiparticles. After the excitation, the distribution function is highly non-thermal, and the quasiparticle band is strongly suppressed. With time, $F(\omega,t)$ approaches back the shape of an approximate Fermi distribution (electron thermalization), and simultaneously the effective temperature of this distribution relaxes back to the initial $1/\beta$. Together with this evolution of the distribution function, the quasiparticle peak in the spectrum is reformed. Before computing the time evolution generated by the QBE, we can independently evaluate the quality of the auxiliary steady-state representation of the spectra at each given time, i.e., the accuracy of the functional $A_{\omega}^{\text{ness}}[F]$, Eq. (65): We take the distribution function $F(\omega,t_{0})$ from the full solution at a given time $t_{0}$, evaluate $A_{\omega}^{\text{ness}}[\bar{F}]$ with $\bar{F}(\omega)=F(\omega,t_{0})$ as described below Eq. (65) to compute a steady state spectrum $\bar{A}(\omega)$, and compare the result with the full solution $A(\omega,t_{0})$. In Fig. 1c, dashed lines correspond to the DMFT solution $A(\omega,t_{0})$, while solid lines show the corresponding $\bar{A}(\omega)$. The comparison is perfect, even for relatively early times. Only for times immediately after the ultrafast excitation ($t=30$), where the gradient approximation is not supposed to work, can one observe a failure of the auxiliary steady state representation. We can therefore affirm that the density of states can be very accurately obtained as a steady state functional of the distribution function, even in the correlated metallic regime. For smaller values of $U$, the agreement is as good (not shown here). Furthermore, not only the density of states can be very accurately obtained as a steady state functional of the distribution function, but the whole Green’s function and self-energy: The energy values represented by coloured dots in Fig. 1a), calculated with Eq. (69), exactly match the ones of the full DMFT code at the same time, calculated with Eq. (62). Figure 2: Time-evolution of the total energy for $U=1$ (a), $U=2$ (b), and $U=3$ (c) (initial inverse temperature $\beta=20$, electron phonon coupling $g^{2}=0.5$). The black dashed lines show the energy $-E_{\text{DMFT}}$ obtained from the full DMFT evolution, solid lines show the energy $-E_{\text{QBE}}$ obtained from the QBE. The QBE is started at different times $t_{0}$ (indicated by the dots at the beginning of the dashed lines), taking the distribution function $F_{\text{DMFT}}(\omega,t_{0})$ as an initial state for a solution of the QBE at times $t>t_{0}$. In passing, we note that a non-equilibrium spectral function $A(\omega,t)$ defined by the Wigner transform (12) is real (hermitian) by construction, but not necessarily positive, while a steady-state fermionic spectral function is always positive. Moreover, for numerical reasons, for short times the integral in the Wigner transform (12) is truncated, possibly leading to small artefacts. In practice, the relation $F(\omega,t)=\bar{F}(\omega)$ will therefore not be enforced exactly, but as a best fit. It should be noted, however, that the positivity of $A(\omega,t)$ and $F(\omega,t)$ is indeed satisfied wherever the gradient approximation is accurate, as discussed in connection with Eq. (18). In particular, as one can see from Fig. 1b), the distribution functions are already positive in the relevant time interval for the present case. Next, we compare the relaxation dynamics of the system in the two descriptions. For this, we simply take the distribution function $F(\omega,t_{0})$ at a given time $t_{0}$ from the full DMFT solution as an initial state for a solution of the QBE for $t>t_{0}$. The time-evolution of the energy is shown in Fig. 2 for three different values of $U$, and different starting times $t_{0}$ of the QBE simulation. For small values of $U$ ($U=1$ and $U=2$ in Fig. 2a) and b), respectively), the energy relaxation rate obtained from the QBE is almost identical to the one from full DMFT. For $U=3$ (Fig. 2c), one can observe a difference in the magnitude of the time-constants related to the relaxation of the total energy in the two approaches. In particular, the QBE presents an artificially faster relaxation with respect to the full DMFT solution. This indicates that the gradient approximation is less justifies for $U=3$, which could be related to the existence of a more narrow quasiparticle band. As the starting point $t_{0}$ of the Boltzmann code shifts forward in time, the difference between the time evolution of the energies becomes less pronounced. If one decreases the coupling $g^{2}$ with the phonon bath (not shown), the relaxation dynamics of the system is slowed down, the gradient approximation is more justified, and the difference in the energy relaxation rate in the two approaches is less pronounced. Figure 3: Distribution function (upper panels) and spectral function (lower panels) obtained from the full DMFT solution (left panels) and the QBE (right panels) one at $U=3$. The QBE takes the DMFT distribution function $F(\omega,t_{0})$ at time $t_{0}=32$ as initial state for the evolution at $t>t_{0}$ (initial inverse temperature $\beta=20$, electron phonon coupling $g^{2}=0.5$). Although the relaxation rate for the energy in the QBE seems to be overestimated for larger values of $U$, Fig. 3 shows that the spectra and distribution functions obtained from the full DMFT and the QBE follow the same qualitative behavior, i.e., a relaxation of $F(\omega,t)$ to a Fermi function together with an evolution of the temperature in this Fermi function towards the initial temperature. ## IV Conclusion In conclusion, we developed a kinetic equation which works without the need to assume the existence of quasiparticles with well-defined dispersion, $\epsilon_{\bm{k}}$ and, above all, evaluates the scattering integral in a non-perturbative manner. In particular, a scattering integral which is consistent with DMFT is obtained by extracting self-energies from a quantum impurity model in an auxiliary non-equilibrium steady state. Most importantly, this guaranties that the final state of the evolution is a proper description of the fully interacting state of the correlated electron system, which makes the present formalism unique with respect to conventional quantum kinetic approaches based on perturbative scattering integrals or certain assumptions on the spectral function, such as assuming a rigid density of states or the quasiparticle approximation. While for full non-equilibrium Green’s function simulations the numerical effort for the propagation over a time interval $t_{\text{max}}$ scales with $\mathcal{O}(t_{\text{max}}^{3})$, and the required memory scales with $\mathcal{O}(t_{\text{max}}^{2})$, in the QBE the numerical effort is linear with $t_{\text{max}}$ and the memory required is independent of $t_{\text{max}}$. We have tested the framework on the relaxation of the electronic state in a correlated metal after a population transfer that simulates a photo- excitation. One assumption of the QBE, i.e., that the spectra at the correlated system can be obtained from an auxiliary steady state, is found to be satisfied with remarkable accuracy. Moreover, the relaxation dynamics for both spectral functions and distribution functions within the full non- equilibrium DMFT simulation and the QBE are consistent. Quantitatively, the gradient approximation underlying the QBE leads to a slight overestimation of the relaxation rate. Whether this can be corrected by higher order expansions of the gradient approximation is left for future investigations. The success of the QBE approach for the present setting motivates an application to different models. In particular this includes symmetry-broken states where interesting long-time phenomena have been observed,Picano and Eckstein (2020) and the evolution of the Mott phase, where already a QBE with an ad-hoc scattering integral has shown relative success.Wais _et al._ (2018) Possible applications of the formalism include the evolution of the density of states in correlated systems, in particular multi-orbital systems where a pronounced effect of the redistribution of weight has already been discussed using quasiparticle kinetic equations.He and Millis (2016) In this context, the method can be combined with GW Wegkamp _et al._ (2014) or DMFT+GWGolež _et al._ (2019), which have demonstrated again a pronounced dependence of the spectra on the distribution. Finally, another interesting perspective of the approach is that there are several promising numerical approaches to study non-equilibrium steady states within DMFT. This includes variants of the strong-coupling expansion Scarlatella2019; Li and Eckstein (2020), matrix product states,Schwarz _et al._ (2018) auxiliary master equations,Arrigoni _et al._ (2013) or Quantum Monte Carlo.Profumo _et al._ (2015); Bertrand _et al._ (2019) The QBE formalism would allow these non-perturbative techniques to access not only true steady states, but also non-equilibrium states of correlated electrons on the picosecond timescale relevant for photo-induced phase transition and collective orders. ###### Acknowledgements. We acknowledge Philipp Werner for useful discussions, and Nagamalleswararao Dasari for discussions as well as his contribution to the implementation of the Ohmic bath. This work was supported by the ERC Starting Grant No. 716648. The calculations have been done at the RRZE of the University Erlangen- Nuremberg. ## References * Basov _et al._ (2017) D. N. Basov, R. D. Averitt, and D. Hsieh, Nature Materials 16, 1077 (2017). * Giannetti _et al._ (2016) C. Giannetti, M. Capone, D. Fausti, M. Fabrizio, F. Parmigiani, and D. Mihailovic, Advances in Physics 65, 58 (2016). * Ichikawa _et al._ (2011) H. Ichikawa, S. Nozawa, T. Sato, A. Tomita, K. Ichiyanagi, M. Chollet, L. Guerin, N. Dean, A. Cavalleri, S.-i. Adachi, T.-h. Arima, H. Sawa, Y. Ogimoto, M. Nakamura, R. Tamaki, K. Miyano, and S.-y. Koshihara, Nature Materials 10, 101 (2011). * Fausti _et al._ (2011) D. Fausti, R. I. Tobey, N. Dean, S. Kaiser, a. Dienst, M. C. Hoffmann, S. Pyon, T. Takayama, H. Takagi, and A. Cavalleri, Science (New York, N.Y.) 331, 189 (2011). * Beaud _et al._ (2014) P. Beaud, A. Caviezel, S. O. Mariager, L. Rettig, G. Ingold, C. Dornes, S.-W. Huang, J. A. Johnson, M. Radovic, T. Huber, T. Kubacka, A. Ferrer, H. T. Lemke, M. Chollet, D. Zhu, J. M. Glownia, M. Sikorski, A. Robert, H. Wadati, M. Nakamura, M. Kawasaki, Y. Tokura, S. L. Johnson, and U. Staub, Nature Materials 13, 923 (2014). * Wegkamp _et al._ (2014) D. Wegkamp, M. Herzog, L. Xian, M. Gatti, P. Cudazzo, C. L. McGahan, R. E. Marvel, R. F. Haglund, A. Rubio, M. Wolf, and J. Stähler, Phys. Rev. Lett. 113, 216401 (2014). * Stojchevska _et al._ (2014) L. Stojchevska, I. Vaskivskyi, T. Mertelj, P. Kusar, D. Svetin, S. Brazovskii, and D. Mihailovic, Science 344, 177 (2014). * Mor _et al._ (2017) S. Mor, M. Herzog, D. Golež, P. Werner, M. Eckstein, N. Katayama, M. Nohara, H. Takagi, T. Mizokawa, C. Monney, and J. Stähler, Phys. Rev. Lett. 119, 086401 (2017). * Budden _et al._ (2020) M. Budden, T. Gebert, M. Buzzi, G. Jotzu, E. Wang, T. Matsuyama, G. Meier, Y. Laplace, D. Pontiroli, M. Riccò, _et al._ , arXiv preprint arXiv:2002.12835 (2020). * Polkovnikov _et al._ (2011) A. Polkovnikov, K. Sengupta, A. Silva, and M. Vengalattore, Rev. Mod. Phys. 83, 863 (2011). * Berges _et al._ (2004) J. Berges, S. Borsányi, and C. Wetterich, Phys. Rev. Lett. 93, 142002 (2004). * Moeckel and Kehrein (2008) M. Moeckel and S. Kehrein, Phys. Rev. Lett. 100, 175702 (2008). * Kollar _et al._ (2011) M. Kollar, F. A. Wolf, and M. Eckstein, Phys. Rev. B 84, 054304 (2011). * Langen _et al._ (2016) T. Langen, T. Gasenzer, and J. Schmiedmayer, Journal of Statistical Mechanics: Theory and Experiment 2016, 064009 (2016). * Kadanoff and Baym (1962) L. P. Kadanoff and G. Baym, _Quantum Statistical Mechanics: Green’s Function Methods in Equilibrium and Nonequilibrium Problems._ (Taylor and Francis, 1962). * Kamenev (2011) A. Kamenev, _Field Theory of Non-Equilibrium Systems._ (Cambridge University Press, 2011). * Haug (1962) J. A.-P. Haug, Hartmut, _Quantum Kinetics in Transport and Optics of Semiconductors_ (Springer-Verlag Berlin Heidelberg, 1962). * Georges _et al._ (2013) A. Georges, L. d. Medici, and J. Mravlje, Annual Review of Condensed Matter Physics 4, 137 (2013). * Deng _et al._ (2013) X. Deng, J. Mravlje, R. Žitko, M. Ferrero, G. Kotliar, and A. Georges, Physical Review Letters 110, 086401 (2013). * Eckstein and Werner (2013) M. Eckstein and P. Werner, Physical Review Letters 110, 126401 (2013). * Sayyad and Eckstein (2016) S. Sayyad and M. Eckstein, Phys. Rev. Lett. 117, 096403 (2016). * Dasari _et al._ (2020) N. Dasari, J. Li, P. Werner, and M. Eckstein, “A photo-induced strange metal with electron and hole quasi-particles,” (2020), arXiv:2010.04095 [cond-mat.str-el] . * Petersen _et al._ (2017) J. C. Petersen, A. Farahani, D. G. Sahota, R. Liang, and J. S. Dodge, Phys. Rev. B 96, 115133 (2017). * Sahota _et al._ (2019) D. G. Sahota, R. Liang, M. Dion, P. Fournier, H. A. Dabkowska, G. M. Luke, and J. S. Dodge, Phys. Rev. Research 1, 033214 (2019). * Golež _et al._ (2016) D. Golež, P. Werner, and M. Eckstein, Phys. Rev. B 94, 035121 (2016). * Babadi _et al._ (2015) M. Babadi, E. Demler, and M. Knap, Phys. Rev. X 5, 041005 (2015). * Rameau _et al._ (2016) J. D. Rameau, S. Freutel, A. F. Kemper, M. A. Sentef, J. K. Freericks, I. Avigo, M. Ligges, L. Rettig, Y. Yoshida, H. Eisaki, J. Schneeloch, R. D. Zhong, Z. J. Xu, G. D. Gu, P. D. Johnson, and U. Bovensiepen, Nature Communications 7, 13761 (2016). * Schlünzen _et al._ (2017) N. Schlünzen, J.-P. Joost, F. Heidrich-Meisner, and M. Bonitz, Phys. Rev. B 95, 165139 (2017). * Aoki _et al._ (2014) H. Aoki, N. Tsuji, M. Eckstein, M. Kollar, T. Oka, and P. Werner, Reviews of Modern Physics 86, 779 (2014). * Georges _et al._ (1996) A. Georges, G. Kotliar, W. Krauth, and M. J. Rozenberg, Rev. Mod. Phys. 68, 13 (1996). * Lipavský _et al._ (1986) P. Lipavský, V. Špička, and B. Velický, Phys. Rev. B 34, 6933 (1986). * Schlünzen _et al._ (2020) N. Schlünzen, J.-P. Joost, and M. Bonitz, Phys. Rev. Lett. 124, 076601 (2020). * Schüler _et al._ (2018) M. Schüler, M. Eckstein, and P. Werner, Phys. Rev. B 97, 245129 (2018). * Kaye and Golež (2020) J. Kaye and D. Golež, arXiv e-prints , arXiv:2010.06511 (2020), arXiv:2010.06511 [cond-mat.str-el] . * Wais _et al._ (2018) M. Wais, M. Eckstein, R. Fischer, P. Werner, M. Battiato, and K. Held, Phys. Rev. B 98, 134312 (2018). * Wais _et al._ (2020) M. Wais, J. Kaufmann, M. Battiato, and K. Held, “Comparing scattering rates from boltzmann and dynamical mean-field theory,” (2020), arXiv:2012.11257 [cond-mat.str-el] . * Joura _et al._ (2008) A. V. Joura, J. K. Freericks, and T. Pruschke, Phys. Rev. Lett. 101, 196401 (2008). * Li _et al._ (2015) J. Li, C. Aron, G. Kotliar, and J. E. Han, Phys. Rev. Lett. 114, 226403 (2015). * Titvinidze _et al._ (2018) I. Titvinidze, M. E. Sorantin, A. Dorda, W. von der Linden, and E. Arrigoni, Phys. Rev. B 98, 035146 (2018). * Matthies _et al._ (2018) A. Matthies, J. Li, and M. Eckstein, Phys. Rev. B 98, 180502 (2018). * Scarlatella _et al._ (2020) O. Scarlatella, A. A. Clerk, R. Fazio, and M. Schiró, arXiv preprint arXiv:2008.02563 (2020). * Li _et al._ (2020) J. Li, D. Golez, P. Werner, and M. Eckstein, Phys. Rev. B 102, 165136 (2020). * Panas _et al._ (2019) J. Panas, M. Pasek, A. Dhar, T. Qin, A. Geißler, M. Hafez-Torbati, M. E. Sorantin, I. Titvinidze, and W. Hofstetter, Phys. Rev. B 99, 115125 (2019). * Freericks _et al._ (2009) J. K. Freericks, H. R. Krishnamurthy, and T. Pruschke, Phys. Rev. Lett. 102, 136401 (2009). * Eckstein and Kollar (2008) M. Eckstein and M. Kollar, Phys. Rev. B 78, 245113 (2008). * Tsuji _et al._ (2009) N. Tsuji, T. Oka, and H. Aoki, Phys. Rev. Lett. 103, 047403 (2009). * Büttiker (1985) M. Büttiker, Phys. Rev. B 32, 1846 (1985). * Luttinger and Ward (1960) J. M. Luttinger and J. C. Ward, Phys. Rev. 118, 1417 (1960). * Baym and Kadanoff (1961) G. Baym and L. P. Kadanoff, Phys. Rev. 124, 287 (1961). * Potthoff (2003) M. Potthoff, The European Physical Journal B - Condensed Matter and Complex Systems 32, 429 (2003). * Kozik _et al._ (2015) E. Kozik, M. Ferrero, and A. Georges, Phys. Rev. Lett. 114, 156402 (2015). * Schäfer _et al._ (2013) T. Schäfer, G. Rohringer, O. Gunnarsson, S. Ciuchi, G. Sangiovanni, and A. Toschi, Phys. Rev. Lett. 110, 246405 (2013). * Eckstein _et al._ (2009) M. Eckstein, M. Kollar, and P. Werner, Phys. Rev. Lett. 103, 056403 (2009). * Schüler _et al._ (2020) M. Schüler, D. Golež, Y. Murakami, N. Bittner, A. Herrmann, H. U. Strand, P. Werner, and M. Eckstein, Computer Physics Communications , 107484 (2020). * Picano and Eckstein (2020) A. Picano and M. Eckstein, “Accelerated gap collapse in a slater antiferromagnet,” (2020), arXiv:2009.04961 [cond-mat.str-el] . * He and Millis (2016) Z. He and A. J. Millis, Phys. Rev. B 93, 115126 (2016). * Golež _et al._ (2019) D. Golež, L. Boehnke, M. Eckstein, and P. Werner, Phys. Rev. B 100, 041111 (2019). * Li and Eckstein (2020) J. Li and M. Eckstein, “Nonequilibrium steady-state theory of photodoped mott insulators,” (2020), arXiv:2007.12511 [cond-mat.str-el] . * Schwarz _et al._ (2018) F. Schwarz, I. Weymann, J. von Delft, and A. Weichselbaum, Phys. Rev. Lett. 121, 137702 (2018). * Arrigoni _et al._ (2013) E. Arrigoni, M. Knap, and W. von der Linden, Physical Review Letters 110, 086403 (2013). * Profumo _et al._ (2015) R. E. V. Profumo, C. Groth, L. Messio, O. Parcollet, and X. Waintal, Phys. Rev. B 91, 245154 (2015). * Bertrand _et al._ (2019) C. Bertrand, S. Florens, O. Parcollet, and X. Waintal, Phys. Rev. X 9, 041008 (2019).
# A Decentralized Analysis of Multiparty Protocols Bas van den Heuvel and Jorge A. Pérez ###### Abstract Protocols provide the unifying glue in concurrent and distributed software today; verifying that message-passing programs conform to such governing protocols is important but difficult. Static approaches based on multiparty session types (MPST) use protocols as types to avoid protocol violations and deadlocks in programs. An elusive problem for MPST is to ensure _both_ protocol conformance _and_ deadlock freedom for implementations with interleaved and delegated protocols. We propose a decentralized analysis of multiparty protocols, specified as global types and implemented as interacting processes in an asynchronous $\pi$-calculus. Our solution rests upon two novel notions: _router processes_ and _relative types_. While router processes use the global type to enable the composition of participant implementations in arbitrary process networks, relative types extract from the global type the intended interactions and dependencies between _pairs_ of participants. In our analysis, processes are typed using APCP, a type system that ensures protocol conformance and deadlock freedom with respect to _binary_ protocols, developed in prior work. Our decentralized, router-based analysis enables the sound and complete transference of protocol conformance and deadlock freedom from APCP to multiparty protocols. ###### Contents 1. 1 Introduction 2. 2 APCP: Asynchronous Processes, Deadlock Free by Typing 3. 3 Global Types and Relative Projection 1. 3.1 Relative Types 2. 3.2 Relative Projection and Well-Formedness 4. 4 Analyzing Global Types using Routers 1. 4.1 Synthesis of Routers 2. 4.2 Types for the Router’s Channels 1. 4.2.1 The Channels between Routers and Implementations 2. 4.2.2 The Channels between Pairs of Routers 3. 4.3 Networks of Routed Implementations 1. 4.3.1 The Typability of Routers 2. 4.3.2 Transference of Results (Operational Correspondence) 4. 4.4 Routers Strictly Generalize Centralized Orchestrators 1. 4.4.1 Synthesis of Orchestrators 2. 4.4.2 Orchestrators and Centralized Compositions of Routers are Behaviorally Equivalent 5. 5 Routers in Action 1. 5.1 Delegation and Interleaving 2. 5.2 Another Example of Delegation 3. 5.3 The Authorization Protocol in Action 6. 6 Related Work 7. 7 Conclusion 8. A Comparing Merge-based Well-formedness and Relative Well-formedness 1. A.1 Relative Well-Formed, Not Merge Well-Formed 2. A.2 Merge Well-Formed, Not Relative Well-Formed ## 1 Introduction This paper presents a new approach to the analysis of the _protocols_ that pervade concurrent and distributed software. Such protocols provide an essential unifying glue between communicating programs; ensuring that communicating programs implement protocols correctly, avoiding protocol violations and deadlocks, is an important but difficult problem. Here, we study _multiparty session types (MPST)_ [36], an approach to correctness in message-passing programs that uses governing multiparty protocols as types in program verification. As a motivating example, let us consider a _recursive authorization protocol_ , adapted from an example by Scalas and Yoshida [48]. It involves three participants: a Client, a Server, and an Authorization service. Intuitively, the protocol proceeds as follows. The Server requests the Client either to _login_ or to _quit_ the protocol. In the case of login, the Client sends a password to the Authorization service, which then may authorize the login with the Server; subsequently, the protocol can be performed again: this is useful when, e.g., clients must renew their authorization privileges after some time. In the case of quit, the protocol ends. MPST use _global types_ to specify multiparty protocols. The authorization protocol just described can be specified by the following global type between Client (‘$c$’), Server (‘$s$’), and Authorization service (‘$a$’): $G_{\mathsf{auth}}=\mu X\mathbin{.}s\mathbin{\twoheadrightarrow}c\left\\{\begin{array}[]{l}\mathsf{login}\mathbin{.}c\mathbin{\twoheadrightarrow}a\big{\\{}\mathsf{passwd}\langle\mathsf{str}\rangle\mathbin{.}a\mathbin{\twoheadrightarrow}s\\{\mathsf{auth}\langle\mathsf{bool}\rangle\mathbin{.}X\\}\big{\\}},\\\ \mathsf{quit}\mathbin{.}c\mathbin{\twoheadrightarrow}a\\{\mathsf{quit}\mathbin{.}\bullet\\}\end{array}\right\\}$ (1) After declaring a recursion on the variable $X$ (‘$\mu X$’), the global type $G_{\mathsf{auth}}$ stipulates that $s$ sends to $c$ (‘$s\mathbin{\twoheadrightarrow}c$’) a label login or quit. The rest of the protocol depends on this choice by $s$. In the login-branch, $c$ sends to $a$ a label passwd along with a string value (‘$\langle\mathsf{str}\rangle$’) and $a$ sends to $s$ a label auth and a boolean value, after which the protocol loops to the beginning of the recursion (‘$X$’). In the quit-branch, $c$ sends to $a$ a label quit after which the protocol ends (‘$\bullet$’). In MPST, participants are implemented as distributed processes that communicate asynchronously. Each process must correctly implement its corresponding portion of the protocol; these individual guarantees ensure that the interactions between processes conform to the given global type. Correctness follows from _protocol fidelity_ (processes interact as stipulated by the protocol), _communication safety_ (no errors or mismatches in messages), and _deadlock freedom_ (processes never get stuck waiting for each other). Ensuring that implementations satisfy these properties is a challenging problem, which is further compounded by two common and convenient features in interacting processes: _delegation_ and _interleaving_. We motivate them in the context of our example: * • _Delegation,_ or higher-order channel passing, can effectively express that the Client transparently reconfigures its involvement by asking another participant (say, a Password Manager) to act on its behalf; * • _Interleaving_ arises when a single process implements more than one role, as in, e.g., an implementation of both the Server and the Authorization service in a sequential process. Note that while delegation is explicitly specified in a global type, interleaving arises in its implementation as interacting processes, not in its specification. MPST have been widely studied from foundational and applied angles [7, 18, 37, 47, 5, 6, 48, 19, 39, 42]. The original theory by Honda _et al._ [35] defines a behavioral type system [38, 3] for a $\pi$-calculus, which exploits linearity to ensure protocol fidelity and communication safety; most derived works retain this approach and target the same properties. Deadlock freedom is hard to ensure by typing when implementations feature delegation and interleaving. In simple scenarios without interleaving and/or delegation, deadlock freedom is easy, as it concerns a single-threaded protocol. In contrast, deadlock freedom for processes running multiple, interleaved protocols (possibly delegated) is a much harder problem, addressed only by some advanced type systems [7, 44, 21]. In this paper, we tackle the problem of ensuring that networks of interacting processes correctly implement a given global type in a deadlock free manner, while supporting delegation and interleaving. Our approach is informed by the differences between _orchestration_ and _choreography_ , two salient approaches to the coordination and organization of interacting processes in service-oriented paradigms [45]: * • In _orchestration_ -based approaches, processes interact through a coordinator process which ensures that they all follow the protocol as intended. Quoting Van der Aalst, in an orchestration “the conductor tells everybody in the orchestra what to do and makes sure they all play in sync” [52]. * • In _choreography_ -based approaches, processes interact directly following the protocol without external coordination. Again quoting Van der Aalst, in a choreography “dancers dance following a global scenario without a single point of control” [52]. Specification and analysis techniques based on MPST fall under the choreography-based approach. The global type provides the protocol’s specification; based on the global type, implementations for each participant interact directly with each other, without an external coordinator. As we will see, the contrast between orchestration and choreography is relevant here because it induces a different _network topology_ for interacting processes. In an orchestration, the resulting process network is _centralized_ : all processes must connect to a central orchestrator process. In a choreography, the process network is _decentralized_ , as processes can directly connect to each other. ##### Contributions We develop a new decentralized analysis of multiparty protocols. * • Here ‘analysis’ refers to (i) ways of specifying such protocols as interacting processes _and_ (ii) techniques to verify that those processes satisfy the intended correctness properties. * • Also, aligned with the above discussion, ‘decentralized’ refers to the intended network topology for processes, which does not rely on an external coordinator. Our decentralized analysis of global types enforces protocol fidelity, communication safety, and deadlock freedom for process implementations, while uniformly supporting delegation, interleaving, and asynchronous communication. $P$router$Q$router$R$router $\begin{array}[]{c}\text{medium}\\\ \text{or}\\\ \text{arbiter}\end{array}$ $P$$Q$$R$ Figure 1: Given processes $P$, $Q$, and $R$ implementing the roles of $c$, $s$, and $a$, respectively, protocol $G_{\mathsf{auth}}$ can be realized as a choreography of routed implementations (our approach, left) and as an orchestration of implementations, with a medium or arbiter process (previous works, right). The _key idea_ of our analysis is to exploit global types to generate _router processes_ (simply _routers_) that enable participant implementations to communicate directly. There is a router per participant; it is intended to serve as a “wrapper” for an individual participant’s implementation. The composition of an implementation with its corresponding router is called a _routed implementation_. Collections of routed implementations can then be connected in arbitrary _process networks_ that correctly realize the multiparty protocol, subsuming centralized and decentralized topologies. Routers are _synthesized_ from global types, and do not change the behavior of the participant implementations they wrap; they merely ensure that networks of routed implementations correctly behave as described by the given multiparty protocol. Returning to Van der Aalst’s analogies quoted above, we may say that in our setting participant implementations are analogous to skilled but barefoot dancers, and that routers provide them with the appropriate shoes to dance without a central point of control. To make this analogy a bit more concrete, Figure 1 (left) illustrates the decentralized process network formed by routed implementations of the participants of $G_{\mathsf{auth}}$: once wrapped by an appropriate router, implementations $P$, $Q$, and $R$ can be composed directly in a decentralized process network. A central technical challenge in our approach is to ensure that compositions of routed implementations conform to their global type. The channels that enable the arbitrary composition of routed implementations need to be typed in accordance with the given multiparty protocol. Unfortunately, the usual notion of projection in MPST, which obtains a single participant’s perspective from a global type, does not suffice: we need a local perspective that is relative to the _two participants_ that the connected routed implementations represent. To this end, we introduce a new notion, _relative projection_ , which isolates the exchanges of the global type that relate to pairs of participants. In the case of $G_{\mathsf{auth}}$, for instance, we need three relative types, describing the protocol for $a$ and $c$, for $a$ and $s$, and for $c$ and $s$. A derived challenge is that when projecting a global type onto a pair of participants, it is possible to encounter _non-local choices_ : choices by other participants that affect the protocol between the two participants involved in the projection. To handle this, relative projection explicitly records non-local choices in the form of _dependencies_ , which inform the projection’s participants that they need to coordinate on the results of the non-local choices. To summarize, our decentralized analysis of global types relies on three intertwined novel notions: * • Routers that wrap participant implementations in order to enable their composition in arbitrary network topologies, whilst guaranteeing that the resulting process networks correctly follow the given global type in a deadlock free manner. * • Relative Types that type the channels between routed implementations, obtained by means of a new notion of projection of global types onto pairs of participants. * • Relative Projection and Dependencies that make it explicit in relative types that participants need to coordinate on non-local choices. The key ingredients of our decentralized analysis for $G_{\mathsf{auth}}$ are jointly depicted in Figure 2. With respect to prior analyses of multiparty protocols, a distinguishing feature of our work is its natural support of decentralized process networks, as expected in a choreography-based approach. Caires and Pérez [12] connect participant implementations through a central coordinator, called _medium process_. This medium process is generated from a global type, and intervenes in all exchanges to ensure that the participant implementations follow the multiparty protocol. The composition of the medium with the participant implementations can then be analyzed using a type system for binary sessions. In a similar vein, Carbone _et al._ [16] define a type system in which they use global types to validate choreographies of participant implementations. Their analysis of protocol implementations—in particular, deadlock freedom—relies on encodings into another type system where participant implementations connect to a central coordinator, called the _arbiter process_. Similar to mediums, arbiters are generated from the global type to ensure that participant implementations follow the protocol as intended. Both these approaches are clear examples of orchestration, and thus do not support decentralized network topologies. To highlight the differences between our decentralized analysis and prior approaches, compare the choreography of routed implementations in Figure 1 (left) with an implementation of $G_{\mathsf{auth}}$ in the style of Caires and Pérez and of Carbone _et al._ , given in Figure 1 (right). These prior works rely on orchestration because the type systems they use for verifying process implementations restrict connections between processes: they only admit a form of process composition that makes it impossible to simultaneously connect three or more participant implementations [24]. In this paper, we overcome this obstacle by relying on APCP (Asynchronous Priority-based Classical Processes) [51], a type system that allows for more general forms of process composition. By using annotations on types, APCP prevents _circular dependencies_ , i.e., cyclically connected processes that are stuck waiting for each other. This is how our approach supports networks of routed participants in both centralized and decentralized topologies, thus subsuming choreography and orchestration approaches. $G_{\mathsf{auth}}$$L_{c}$local projection (§ 4.2.1)$R_{cs},R_{ca}$relative projection (§ 3.2)$P$type check in APCP (§ 2)$\mathcal{R}_{c}$router synthesis (§ 4.1)$G_{\mathsf{auth}}$$P\mathbin{|}\mathcal{R}_{c}$clients$Q\mathbin{|}\mathcal{R}_{s}$serverl$R\mathbin{|}\mathcal{R}_{a}$authorization servicesnetwork of routed implementations of $G_{\mathsf{auth}}$ (Def. 25)routed implementation of $c$ (Def. 24) Figure 2: Decentralized analysis of $G_{\mathsf{auth}}$ into a network of routed implementations. The definition of $G_{\mathsf{auth}}$ contains message types. Focusing on the client $c$ (on the left), $L_{c}$ denotes a session type, whereas $R_{cs}$ and $R_{ca}$ are relative types with respect to the server and the authorization service, respectively. ##### Outline This paper is structured as follows. Next, Section 2 recalls APCP as introduced by Van den Heuvel and Pérez [51] and summarizes the correctness properties for asynchronous processes derived from typing. The following three sections develop and illustrate our contributions: * • Section 3 introduces relative types and relative projection, and defines well- formed global types, a class of global types that includes protocols with non- local choices. * • Section 4 introduces the synthesis of routers. A main result is their typability in APCP (Theorem 11). We establish deadlock freedom for networks of routed implementations (Theorem 18), which we transfer to multiparty protocols via an operational correspondence result (Theorems 19 and 23). Moreover, we show that our approach strictly generalizes prior analyses based on centralized topologies (Theorem 27). * • Section 5 demonstrates our contributions in action, with a full development of the routed implementations for $G_{\mathsf{auth}}$, and an example of the flexible support for _delegation_ and _interleaving_ enabled by our router- based approach and APCP. We discuss further related works in § 6 and conclude the paper in § 7. We use colors to improve readability. ## 2 APCP: Asynchronous Processes, Deadlock Free by Typing We recall APCP as defined by Van den Heuvel and Pérez [51]. APCP is a type system for asynchronous $\pi$-calculus processes (with non-blocking outputs) [34, 9], with support for recursion and cyclic connections. In this type system, channel endpoints are assigned linear types that represent two-party (binary) _session types_ [33]. Well-typed APCP processes preserve typing (Theorem 2) and are deadlock free (Theorem 5). At its basis, APCP combines Dardha and Gay’s Priority-based Classical Processes (PCP) [22] with DeYoung _et al._ ’s continuation-passing semantics for asynchrony [27], and adds recursion, inspired by the work of Toninho _et al._ [49]. We refer the interested reader to the work by Van den Heuvel and Pérez [51] for a motivation of design choices and proofs of results. ##### Process Syntax Process syntax: $\displaystyle P,Q::=$ $\displaystyle\leavevmode\nobreak\ x[y,z]$ output $\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \;\mbox{\large{$\mid$}}\;\leavevmode\nobreak\ \leavevmode\nobreak\ x(y,z)\mathbin{.}P$ input $\mid$ $\displaystyle\leavevmode\nobreak\ x[z]\triangleleft i$ selection $\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \;\mbox{\large{$\mid$}}\;\leavevmode\nobreak\ \leavevmode\nobreak\ x(z)\triangleright\\{i{:}\leavevmode\nobreak\ P\\}_{i\in I}$ branching $\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \;\mbox{\large{$\mid$}}\;\leavevmode\nobreak\ \leavevmode\nobreak\ (\bm{\nu}xy)P$ restriction $\mid$ $\displaystyle\leavevmode\nobreak\ P\mathbin{|}Q$ parallel $\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \;\mbox{\large{$\mid$}}\;\leavevmode\nobreak\ \leavevmode\nobreak\ \bm{0}$ inaction $\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \;\mbox{\large{$\mid$}}\;\leavevmode\nobreak\ \leavevmode\nobreak\ x\mathbin{\leftrightarrow}y$ forwarder $\mid$ $\displaystyle\leavevmode\nobreak\ \mu X(\tilde{z})\mathbin{.}P$ recursive loop $\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \;\mbox{\large{$\mid$}}\;\leavevmode\nobreak\ \leavevmode\nobreak\ X{\langle\tilde{z}\rangle}$ recursive call . Structural congruence: $\displaystyle P\equiv_{\alpha}P^{\prime}\implies{}$ $\displaystyle P$ $\displaystyle\equiv P^{\prime}$ $\displaystyle x\mathbin{\leftrightarrow}y$ $\displaystyle\equiv y\mathbin{\leftrightarrow}x$ $\displaystyle P\mathbin{|}Q$ $\displaystyle\equiv Q\mathbin{|}P$ $\displaystyle(\bm{\nu}xy)x\mathbin{\leftrightarrow}y$ $\displaystyle\equiv\bm{0}$ $\displaystyle P\mathbin{|}\bm{0}$ $\displaystyle\equiv P$ $\displaystyle P\mathbin{|}(Q\mathbin{|}R)$ $\displaystyle\equiv(P\mathbin{|}Q)\mathbin{|}R$ $\displaystyle x,y\notin\mathrm{fn}(P)\implies{}$ $\displaystyle P\mathbin{|}(\bm{\nu}xy)Q$ $\displaystyle\equiv(\bm{\nu}xy)(P\mathbin{|}Q)$ $\displaystyle(\bm{\nu}xy)\bm{0}$ $\displaystyle\equiv\bm{0}$ $\displaystyle|\tilde{z}|=|\tilde{y}|\implies{}$ $\displaystyle\mu X(\tilde{z})\mathbin{.}P$ $\displaystyle\equiv P\big{\\{}(\mu X(\tilde{y})\mathbin{.}P\\{\tilde{y}/\tilde{z}\\})/X{\langle\tilde{y}\rangle}\big{\\}}$ $\displaystyle(\bm{\nu}xy)P$ $\displaystyle\equiv(\bm{\nu}yx)P$ $\displaystyle(\bm{\nu}xy)(\bm{\nu}zw)P$ $\displaystyle\equiv(\bm{\nu}zw)(\bm{\nu}xy)P$ . Reduction: $\displaystyle\beta_{\text{Id}}$ $\displaystyle z,y\neq x\implies{}$ $\displaystyle(\bm{\nu}yz)(x\mathbin{\leftrightarrow}y\mathbin{|}P)$ $\displaystyle\longrightarrow P\\{x/z\\}$ $\displaystyle\beta_{\mathbin{\otimes}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}}$ $\displaystyle(\bm{\nu}xy)(x[a,b]\mathbin{|}y(v,z)\mathbin{.}P)$ $\displaystyle\longrightarrow P\\{a/v,b/z\\}$ $\displaystyle\beta_{{\oplus}\&}$ $\displaystyle j\in I\implies{}$ $\displaystyle(\bm{\nu}xy)(x[b]\triangleleft j\mathbin{|}y(z)\triangleright\\{i{:}\leavevmode\nobreak\ P_{i}\\}_{i\in I})$ $\displaystyle\longrightarrow P_{j}\\{b/z\\}$ $\displaystyle\kappa_{\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}}$ $\displaystyle x\notin\tilde{v},\tilde{w}\implies{}$ $\displaystyle(\bm{\nu}\tilde{v}\tilde{w})(x(y,z)\mathbin{.}P\mathbin{|}Q)$ $\displaystyle\longrightarrow x(y,z)\mathbin{.}(\bm{\nu}\tilde{v}\tilde{w})(P\mathbin{|}Q)$ $\displaystyle\kappa_{\&}$ $\displaystyle x\notin\tilde{v},\tilde{w}\implies{}$ $\displaystyle(\bm{\nu}\tilde{v}\tilde{w})(x(z)\triangleright\\{i{:}\leavevmode\nobreak\ P_{i}\\}_{i\in I}\mathbin{|}Q)$ $\displaystyle\longrightarrow x(z)\triangleright\\{i{:}\leavevmode\nobreak\ (\bm{\nu}\tilde{v}\tilde{w})(P_{i}\mathbin{|}Q)\\}_{i\in I}$ $(P\equiv P^{\prime})\wedge(P^{\prime}\longrightarrow Q^{\prime})\wedge(Q^{\prime}\equiv Q)$ $\shortrightarrow_{\equiv}$ $P\longrightarrow Q$ $\raisebox{9.0pt}{}P\longrightarrow Q$ $\shortrightarrow_{\nu}$ $(\bm{\nu}xy)P\longrightarrow(\bm{\nu}xy)Q$ $\raisebox{9.0pt}{}P\longrightarrow Q$ $\shortrightarrow_{\mathbin{|}}$ $P\mathbin{|}R\longrightarrow Q\mathbin{|}R$ Figure 3: Definition of the process language of APCP. We write $x,y,z,\ldots$ to denote (channel) _endpoints_ (also known as _names_), and write $\tilde{x},\tilde{y},\tilde{z},\ldots$ to denote sequences of endpoints. Also, we write $i,j,k,\ldots$ to denote _labels_ for choices and $I,J,K,\ldots$ to denote sets of labels. We write $X,Y,\ldots$ to denote _recursion variables_ , and $P,Q,\ldots$ to denote processes. Figure 3 (top) gives the syntax of processes, which communicate asynchronously by following a continuation-passing style. The output action ‘$x[y,z]$’ denotes the sending of endpoints $y$ and $z$ along $x$: while the former is the message, the latter is the protocol’s continuation; both $y$ and $z$ are free. The input prefix ‘$x(y,z)\mathbin{.}P$’ blocks until a message $y$ and a continuation endpoint $z$ are received on $x$, binding $y$ and $z$ in $P$. The selection action ‘$x[z]\mathbin{\triangleleft}i$’ sends a label $i$ and a continuation endpoint $z$ along $x$. The branching prefix ‘$x(z)\mathbin{\triangleright}\\{i{:}\leavevmode\nobreak\ P_{i}\\}_{i\in I}$’ blocks until it receives a label $i\in I$ and a continuation endpoint $z$ on $x$, binding $z$ in each $P_{i}$. Restriction ‘$(\bm{\nu}xy)P$’ binds $x$ and $y$ in $P$, thus declaring them as the two endpoints of the same channel and enabling communication, as in Vasconcelos [53]. The process ‘$\mkern-1.0muP\mathbin{|}Q\mkern 1.0mu$’ denotes the parallel composition of $P$ and $Q$. The process ‘$\bm{0}$’ denotes inaction. The forwarder process ‘$x\mathbin{\leftrightarrow}y$’ is a primitive copycat process that links together $x$ and $y$. The prefix ‘$\mu X(\tilde{z})\mathbin{.}P$’ defines a recursive loop, where $\mu$ binds any free occurrences of $X$ in $P$ and the endpoints $\tilde{z}$ form a context for $P$. The recursive call ‘$X{\langle\tilde{z}\rangle}$’ loops to its corresponding $\mu X$, providing the endpoints $\tilde{z}$ as context. We only consider contractive recursion, disallowing processes with subexpressions of the form ‘$\mu X_{1}(\tilde{z})\ldots\mu X_{n}(\tilde{z})\mathbin{.}X_{1}{\langle\tilde{z}\rangle}$’. Endpoints and recursion variables are free unless they are bound somewhere. We write ‘$\mathrm{fn}(P)$’ and ‘$\mathrm{frv}(P)$’ for the sets of free names and free recursion variables of $P$, respectively. Also, we write ‘$P\\{x/y\\}$’ to denote the capture-avoiding substitution of the free occurrences of $y$ in $P$ for $x$. The notation ‘$P\big{\\{}(\mu X(\tilde{y})\mathbin{.}Q)/X{\langle\tilde{y}\rangle}\big{\\}}$’ denotes the substitution of occurrences of recursive calls ‘$X{\langle\tilde{y}\rangle}$’ for any sequence of names $\tilde{y}$ in $P$ with the recursive loop ‘$\mu X(\tilde{y})\mathbin{.}Q$’, which we call _unfolding_ recursion. We write sequences of substitutions ‘$P\\{x_{1}/y_{1}\\}\ldots\\{x_{n}/y_{n}\\}$’ as ‘$P\\{x_{1}/y_{1},\ldots,x_{n}/y_{n}\\}$’. In an output ‘$x[y,z]$’, both $y$ and $z$ are free, as mentioned above; they can be bound to a continuation process using parallel composition and restriction, as in, e.g., $(\bm{\nu}ya)(\bm{\nu}zb)(x[y,z]\mathbin{|}P_{a,b})$. The same applies to selection ‘$x[z]\mathbin{\triangleleft}i$’. We introduce useful notations that elide the restrictions and continuation endpoints: ###### Notation 1 (Derivable Actions and Prefixes). We use the following syntactic sugar: $\displaystyle\overline{x}[y]\cdot P$ $\displaystyle:=(\bm{\nu}ya)(\bm{\nu}zb)(x[a,b]\mathbin{|}P\\{z/x\\})$ $\displaystyle\overline{x}\mathbin{\triangleleft}\ell\cdot P$ $\displaystyle:=(\bm{\nu}zb)(x[b]\mathbin{\triangleleft}\ell\mathbin{|}P\\{z/x\\})$ $\displaystyle x(y)\mathbin{.}P$ $\displaystyle:=x(y,z)\mathbin{.}P\\{z/x\\}$ $\displaystyle x\mathbin{\triangleright}\\{i{:}\leavevmode\nobreak\ P_{i}\\}_{i\in I}$ $\displaystyle:=x(z)\mathbin{\triangleright}\\{i{:}\leavevmode\nobreak\ P_{i}\\{z/x\\}\\}_{i\in I}$ Note the use of ‘${}\cdot{}$’ instead of ‘${}\mathbin{.}{}$’ in output and selection actions to stress that they are non-blocking. ##### Operational Semantics We define a reduction relation for processes ($P\longrightarrow Q$) that formalizes how complementary actions on connected endpoints may synchronize. As usual for $\pi$-calculi, reduction relies on _structural congruence_ ($P\equiv Q$), which equates the behavior of processes with minor syntactic differences; it is the smallest congruence relation satisfying the axioms in Figure 3 (center). Structural congruence defines the following properties of our process language. Processes are equivalent up to $\alpha$-equivalence. Parallel composition is associative and commutative, with unit ‘$\bm{0}$’. The forwarder process is symmetric, and equivalent to inaction if both endpoints are bound together through restriction. A parallel process may be moved into or out of a restriction as long as the bound channels do not appear free in the moved process: this is _scope inclusion_ and _scope extrusion_ , respectively. Restrictions on inactive processes may be dropped, and the order of endpoints in restrictions and of consecutive restrictions does not matter. Finally, a recursive loop is equivalent to its unfolding, replacing any recursive calls with copies of the recursive loop, where the call’s endpoints are pairwise substituted for the contextual endpoints of the loop. We define the reduction relation by the axioms and closure rules in Figure 3 (bottom). Axioms labeled ‘$\beta$’ are _synchronizations_ and those labeled ‘$\kappa$’ are _commuting conversions_ , which allow pulling prefixes on free channels out of restrictions; they are not necessary for deadlock freedom, but they are usually presented in Curry-Howard interpretations of linear logic as session types [14, 54, 22, 27]. Rule $\beta_{\text{Id}}$ implements the forwarder as a substitution. Rule $\beta_{\mathbin{\otimes}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}}$ synchronizes an output and an input on connected endpoints and substitutes the message and continuation endpoint. Rule $\beta_{{\oplus}\&}$ synchronizes a selection and a branch: the received label determines the continuation process, substituting the continuation endpoint appropriately. Rule $\kappa_{\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}}$ (resp. $\kappa_{\&}$) pulls an input (resp. a branching) prefix on free channels out of enclosing restrictions. Rules $\rightarrow_{\equiv}$, $\rightarrow_{\nu}$, and $\rightarrow_{\mathbin{|}}$ close reduction under structural congruence, restriction, and parallel composition, respectively. ###### Notation 2 (Reductions). We write ‘$\longrightarrow_{\beta}$’ for reductions derived from $\beta$-axioms, and ‘$\longrightarrow^{\ast}$’ for the reflexive, transitive closure of ‘$\longrightarrow$’. Also, we write ‘$P\longrightarrow^{\star}Q$’ if $P\longrightarrow^{\ast}Q$ in a finite number of steps, and ‘$P{\centernot\longrightarrow}^{\ast}Q$’ for the non-existence of a series of reductions from $P$ to $Q$. ##### Session Types Session Type Endpoint Behavior $A\mathbin{\otimes}^{\mathsf{o}}B$ output an endpoint of type $A$, then behave as $B$ $A\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}}B$ input an endpoint of type $A$, then behave as $B$ ${{\oplus}}^{\mathsf{o}}\\{i{:}\leavevmode\nobreak\ A_{i}\\}_{i\in I}$ select a label $i\in I$, then behave as $A_{i}$ $\&^{\mathsf{o}}\\{i{:}\leavevmode\nobreak\ A_{i}\\}_{i\in I}$ receive a choice for a label $i\in I$, then behave as $A_{i}$ $\bullet$ closed session; no behavior Table 1: Session types and their associated endpoint behaviors (cf. Definition 1). The type system assigns session types to channel endpoints. We present session types as linear logic propositions following, e.g., Wadler [54], Caires and Pfenning [13], and Dardha and Gay [22]. We extend these propositions with recursion and _priority_ annotations on connectives. Intuitively, actions typed with lower priority should be performed before those with higher priority. We write $\mathsf{o},\kappa,\pi,\rho,\ldots$ to denote priorities, and ‘$\omega$’ to denote the ultimate priority that is greater than all other priorities and cannot be increased further. That is, $\forall t\in\mathbb{N}.\leavevmode\nobreak\ \omega>t$ and $\forall t\in\mathbb{N}.\leavevmode\nobreak\ \omega+t=\omega$. ###### Definition 1 (Session Types). The following grammar defines the syntax of _session types_ $A,B$. Let $\mathsf{o}\in\mathbb{N}\cup\\{\omega\\}$. $\displaystyle A,B$ $\displaystyle::=A\mathbin{\otimes}^{\mathsf{o}}B\;\mbox{\large{$\mid$}}\;A\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}}B\;\mbox{\large{$\mid$}}\;{{\oplus}}^{\mathsf{o}}\\{i:A\\}_{i\in I}\;\mbox{\large{$\mid$}}\;\&^{\mathsf{o}}\\{i:A\\}_{i\in I}\;\mbox{\large{$\mid$}}\;\bullet\;\mbox{\large{$\mid$}}\;\mu X\mathbin{.}A\;\mbox{\large{$\mid$}}\;X$ Table 1 gives _session types_ and the behavior that is expected of an endpoint with each type (recursive types entail a communication behavior only after unfolding). Note that ‘$\bullet$’ does not require a priority, as closed endpoints do not exhibit behavior and thus are non-blocking. We define ‘$\bullet$’ as a single, self-dual type for closed endpoints (cf. Caires [11] and Atkey _et al._ [4]). Type ‘$\mu X\mathbin{.}A$’ denotes a recursive type, in which $A$ may contain occurrences of the recursion variable ‘$X$’. As customary, ‘$\mu$’ is a binder: it induces the standard notions of $\alpha$-equivalence, substitution (denoted ‘$A\\{B/X\\}$’), and free recursion variables (denoted ‘$\mathrm{frv}(A)$’). We work with tail-recursive, contractive types, disallowing types of the form ‘$\mu X_{1}\ldots\mu X_{n}\mathbin{.}X_{1}$’. We postpone the formalization of the unfolding of recursive types, as it requires additional definitions to ensure consistency of priorities in types. _Duality_ , the cornerstone of session types and linear logic, ensures that the two endpoints of a channel have matching actions. Furthermore, dual types must have matching priority annotations. The following inductive definition of duality suffices for our tail-recursive types (cf. Gay _et al._ [31]). ###### Definition 2 (Duality). The _dual_ of session type $A$, denoted ‘$\overline{A}$’, is defined inductively as follows: $\displaystyle\overline{A\mathbin{\otimes}^{\mathsf{o}}B}$ $\displaystyle:=\overline{A}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}}\overline{B}$ $\displaystyle\overline{{{\oplus}}^{\mathsf{o}}\\{i:A_{i}\\}_{i\in I}}$ $\displaystyle:=\&^{\mathsf{o}}\\{i:\overline{A_{i}}\\}_{i\in I}$ $\displaystyle\overline{\bullet}$ $\displaystyle:=\bullet$ $\displaystyle\overline{\mu X\mathbin{.}A}$ $\displaystyle:=\mu X\mathbin{.}\overline{A}$ $\displaystyle\overline{A\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}}B}$ $\displaystyle:=\overline{A}\mathbin{\otimes}^{\mathsf{o}}\overline{B}$ $\displaystyle\overline{\&^{\mathsf{o}}\\{i:A_{i}\\}_{i\in I}}$ $\displaystyle:={{\oplus}}^{\mathsf{o}}\\{i:\overline{A_{i}}\\}_{i\in I}$ $\displaystyle\overline{X}$ $\displaystyle:=X$ The priority of a type is determined by the priority of the type’s outermost connective: ###### Definition 3 (Priorities). For session type $A$, ‘$\mathsf{pr}(A)$’ denotes its _priority_ : $\displaystyle\mathsf{pr}(A\mathbin{\otimes}^{\mathsf{o}}B):=\mathsf{pr}(A\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}}B)$ $\displaystyle:=\mathsf{o}$ $\displaystyle\mathsf{pr}(\mu X\mathbin{.}A)$ $\displaystyle:=\mathsf{pr}(A)$ $\displaystyle\mathsf{pr}({\oplus}^{\mathsf{o}}\\{i{:}\leavevmode\nobreak\ A_{i}\\}_{i\in I}):=\mathsf{pr}(\&^{\mathsf{o}}\\{i{:}\leavevmode\nobreak\ A_{i}\\}_{i\in I})$ $\displaystyle:=\mathsf{o}$ $\displaystyle\mathsf{pr}(\bullet):=\mathsf{pr}(X)$ $\displaystyle:=\omega$ The priority of ‘$\bullet$’ and ‘$X$’ is $\omega$: they denote “final”, non- blocking actions of protocols. Although ‘$\mathbin{\otimes}$’ and ‘${\oplus}$’ also denote non-blocking actions, their priority is not constant: duality ensures that the priority for ‘$\mathbin{\otimes}$’ (resp. ‘${\oplus}$’) matches the priority of a corresponding ‘$\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}$’ (resp. ‘$\&$’), which denotes a blocking action. Having defined the priority of types, we now turn to formalizing the unfolding of recursive types. Recall the intuition that actions typed with lower priority should be performed before those with higher priority. Based on this rationale, we observe that unfolding should increase the priorities of the unfolded type. This is because the actions related to the unfolded recursion should be performed _after_ the prefix. The following definition _lifts_ priorities in types: ###### Definition 4 (Lift). For proposition $A$ and $t\in\mathbb{N}$, we define ‘$\mkern 2.0mu{\uparrow^{t}}A\mkern-3.0mu$’ as the _lift_ operation: $\displaystyle{\uparrow^{t}}(A\mathbin{\otimes}^{\mathsf{o}}B)$ $\displaystyle:=({\uparrow^{t}}A)\mathbin{\otimes}^{\mathsf{o}+t}({\uparrow^{t}}B)$ $\displaystyle{\uparrow^{t}}({\oplus}^{\mathsf{o}}\\{i{:}\leavevmode\nobreak\ A_{i}\\}_{i\in I})$ $\displaystyle:={\oplus}^{\mathsf{o}+t}\\{i{:}\leavevmode\nobreak\ {\uparrow^{t}}A_{i}\\}_{i\in I}$ $\displaystyle{\uparrow^{t}}\bullet$ $\displaystyle:=\bullet$ $\displaystyle{\uparrow^{t}}(A\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}}B)$ $\displaystyle:=({\uparrow^{t}}A)\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}+t}({\uparrow^{t}}B)$ $\displaystyle{\uparrow^{t}}(\&^{\mathsf{o}}\\{i{:}\leavevmode\nobreak\ A_{i}\\}_{i\in I})$ $\displaystyle:=\&^{\mathsf{o}+t}\\{i{:}\leavevmode\nobreak\ {\uparrow^{t}}A_{i}\\}_{i\in I}$ $\displaystyle{\uparrow^{t}}(\mu X\mathbin{.}A)$ $\displaystyle:=\mu X\mathbin{.}{\uparrow^{t}}(A)$ $\displaystyle{\uparrow^{t}}X$ $\displaystyle:=X$ ###### Definition 5. The _unfolding_ of ‘$\mu X\mathbin{.}A$’ is ‘$A\\{\mu X\mathbin{.}({\uparrow^{t}}A)/X\\}$’, denoted ‘$\mathrm{unfold}^{t}(\mu X\mathbin{.}A)$’, where $t\in\mathbb{N}$. When unfolding $\mu X\mathbin{.}A$ as $\mathrm{unfold}^{t}(\mu X\mathbin{.}A)$, the “lifter” $t$ will depend on the highest priority of the types appearing in a typing context. The highest priority of a type is defined as follows: ###### Definition 6 (Highest Priority). For session type $A$, ‘$\max_{\mathsf{pr}}(A)$’ denotes its _highest priority_ : $\displaystyle\max_{\mathsf{pr}}(A\mathbin{\otimes}^{\mathsf{o}}B):=\max_{\mathsf{pr}}(A\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}}B)$ $\displaystyle:=\max(\max_{\mathsf{pr}}(A),\max_{\mathsf{pr}}(B),\mathsf{o})$ $\displaystyle\max_{\mathsf{pr}}(\mu X\mathbin{.}A)$ $\displaystyle:=\max_{\mathsf{pr}}(A)$ $\displaystyle\max_{\mathsf{pr}}({\oplus}^{\mathsf{o}}\\{i:A_{i}\\}_{i\in I}):=\max_{\mathsf{pr}}(\&^{\mathsf{o}}\\{i:A_{i}\\}_{i\in I})$ $\displaystyle:=\max(\max_{i\in I}(\max_{\mathsf{pr}}(A_{i})),\mathsf{o})$ $\displaystyle\max_{\mathsf{pr}}(\bullet):=\max_{\mathsf{pr}}(X)$ $\displaystyle:=0$ Notice how, in contrast to Definition 3, the highest priority of ‘$\bullet$’ and ‘$X$’ is 0: this is because they do not contribute to the increase in priority needed for unfolding recursive types. ##### Type Checking The typing (or, type checking) rules of APCP enforce that channel endpoints implement their ascribed session types, while ensuring that actions with lower priority are performed before those with higher priority (cf. Dardha and Gay [22]). They enforce the following laws: 1. 1. an action with priority $\mathsf{o}$ must be prefixed only by inputs and branches with priority strictly smaller than $\mathsf{o}$—this law only applies to inputs and branches, because outputs and selections are not prefixes; 2. 2. dual actions leading to synchronizations must have equal priorities (cf. Def. 1). Judgments are of the form ‘$P\vdash\Omega;\Gamma$’: * • $P$ is a process; * • $\Gamma$ is a context that assigns types to channels (‘$x{:}\leavevmode\nobreak\ A$’); * • $\Omega$ is a context that assigns tuples of types to recursion variables (‘$X{:}\leavevmode\nobreak\ (A,B,\ldots)$’). A judgment ‘$P\vdash\Omega;\Gamma$’ then means that $P$ can be typed in accordance with the type assignments for names recorded in $\Gamma$ and the recursion variables in $\Omega$. Intuitively, the recursive context $\Omega$ ensures that the context endpoints concur between recursive definitions and calls. Both contexts $\Gamma$ and $\Omega$ obey _exchange_ : assignments may be silently reordered. $\Gamma$ is _linear_ , disallowing _weakening_ (i.e., all assignments must be used) and _contraction_ (i.e., assignments may not be duplicated). $\Omega$ allows weakening and contraction, because a recursive definition may be called _zero or more_ times. The empty context is written ‘$\emptyset$’. We write ‘${\uparrow^{t}}\Gamma$’ to denote the component-wise extension of lift (Definition 4) to typing contexts. Also, we write ‘$\mathsf{pr}(\Gamma)$’ to denote the least priority of all types in $\Gamma$ (Definition 3). An assignment ‘$\tilde{z}{:}\leavevmode\nobreak\ \tilde{A}$’ means ‘$z_{1}{:}\leavevmode\nobreak\ A_{1},\ldots,z_{k}{:}\leavevmode\nobreak\ A_{k}$’. Empty $\bm{0}\vdash\Omega;\emptyset$ $P\vdash\Omega;\Gamma$ $\bullet$ $P\vdash\Omega;\Gamma,x{:}\leavevmode\nobreak\ \bullet$ Id $x\mathbin{\leftrightarrow}y\vdash\Omega;x{:}\leavevmode\nobreak\ \overline{A},y{:}\leavevmode\nobreak\ A$ $P\vdash\Omega;\Gamma$ $Q\vdash\Omega;\Delta$ Mix $P\mathbin{|}Q\vdash\Omega;\Gamma,\Delta$ $P\vdash\Omega;\Gamma,x{:}\leavevmode\nobreak\ A,y{:}\leavevmode\nobreak\ \overline{A}$ Cycle $(\bm{\nu}xy)P\vdash\Omega;\Gamma$ $\mathbin{\otimes}$ $x[y,z]\vdash\Omega;x{:}\leavevmode\nobreak\ A\mathbin{\otimes}^{\mathsf{o}}B,y{:}\leavevmode\nobreak\ \overline{A},z{:}\leavevmode\nobreak\ \overline{B}$ $P\vdash\Omega;\Gamma,y{:}\leavevmode\nobreak\ A,z{:}\leavevmode\nobreak\ B$ $\mathsf{o}<\mathsf{pr}(\Gamma)$ $\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}$ $x(y,z)\mathbin{.}P\vdash\Omega;\Gamma,x{:}\leavevmode\nobreak\ A\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}}B$ $j\in I\vphantom{P_{i}\mathsf{pr}(\Gamma)}$ ${\oplus}$ $x[z]\triangleleft j\vdash\Omega;x{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}}\\{i:A_{i}\\}_{i\in I},z{:}\leavevmode\nobreak\ \overline{A_{j}}\vphantom{j}$ $\forall i\in I.\leavevmode\nobreak\ P_{i}\vdash\Omega;\Gamma,z{:}\leavevmode\nobreak\ A_{i}$ $\mathsf{o}<\mathsf{pr}(\Gamma)$ $\&$ $x(z)\triangleright\\{i:P_{i}\\}_{i\in I}\vdash\Omega;\Gamma,x{:}\leavevmode\nobreak\ \&^{\mathsf{o}}\\{i:A_{i}\\}_{i\in I}$ $t\in\mathbb{N}>\max_{\mathsf{pr}}(\tilde{A})$ $P\vdash\Omega,X{:}\leavevmode\nobreak\ {\tilde{A}};\tilde{z}{:}\leavevmode\nobreak\ \tilde{U}$ where each $U_{i}=\mathrm{unfold}^{t}(\mu X\mathbin{.}A_{i})$ $\forall A_{i}\in\tilde{A}.\leavevmode\nobreak\ A_{i}\neq X$ Rec $\mu X(\tilde{z})\mathbin{.}P\vdash\Omega;\tilde{z}{:}\leavevmode\nobreak\ \widetilde{\mu X\mathbin{.}A}$ Var $X{\langle\tilde{z}\rangle}\vdash\Omega,X{:}\leavevmode\nobreak\ \tilde{A};\tilde{z}{:}\leavevmode\nobreak\ \widetilde{\mu X\mathbin{.}A}$ . $P\vdash\Omega;\Gamma,y{:}\leavevmode\nobreak\ A,x{:}\leavevmode\nobreak\ B$ $\mathbin{\otimes}^{\star}$ $\overline{x}[y]\cdot P\vdash\Omega;\Gamma,x{:}\leavevmode\nobreak\ A\mathbin{\otimes}^{\mathsf{o}}B$ $P\vdash\Omega;\Gamma,x{:}\leavevmode\nobreak\ A_{j}$ $j\in I$ ${\oplus}^{\star}$ $\overline{x}\triangleleft j\cdot P\vdash\Omega;\Gamma,x{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}}\\{i:A_{i}\\}_{i\in I}$ $P\vdash\Omega;\Gamma$ $t\in\mathbb{N}$ Lift $P\vdash\Omega;{\uparrow^{t}}\Gamma$ Figure 4: The typing rules of APCP (top) and admissible rules (bottom). Figure 4 (top) gives the typing rules. In typing rules, we often write ‘$\Gamma,x{:}\leavevmode\nobreak\ A$’ (or similarly for $\Omega$) to denote _disjoint union_ , i.e. $x\notin\mathsf{dom}(\Gamma)$. Some type-preserving transformations of typing derivations correspond to process reductions (cf. Theorem 2). Other such transformations correspond to structural congruences (cf. Figure 3 (middle)); we sometimes use this explicitly in typing derivations in the form of a rule ‘$\equiv$’. If $P\equiv Q$ and $P\vdash\Omega;\Gamma$ and $Q\vdash\Omega;\Gamma^{\prime}$ where $\Gamma$ and $\Gamma^{\prime}$ are equal up to the unfolding of recursive types, then we say that $P\vdash\Omega;\Gamma^{\prime}$ and $Q\vdash\Omega;\Gamma$; in the context of a typing derivation, we equate recursive types and their unfoldings. We describe the typing rules from a _bottom-up_ perspective. Axiom ‘Empty’ types an inactive process with no endpoints. Rule ‘$\bullet$’ silently removes a closed endpoint to the typing context. Axiom ‘Id’ types forwarding between endpoints of dual type. Rule ‘Mix’ types the parallel composition of two processes that do not share assignments on the same endpoints. Rule ‘Cycle’ types a restriction, where the two restricted endpoints must be of dual type. Note that a single application of ‘Mix’ followed by ‘Cycle’ coincides with the usual rule ‘Cut’ in type systems based on linear logic [14, 54]. Axiom ‘$\mathbin{\otimes}$’ types an output action; this rule does not have premises to provide a continuation process, leaving the free endpoints to be bound to a continuation process using ‘Mix’ and ‘Cycle’. Similarly, axiom ‘${\oplus}$’ types an unbounded selection action. Priority checks are confined to rules ‘$\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}$’ and ‘$\&$’, which type an input and a branching prefix, respectively. In both cases, the used endpoint’s priority must be lower than the priorities of the other types in the continuation’s typing context. Rule ‘Rec’ types a recursive definition by introducing a recursion variable to the recursion context whose tuple of types concurs with the contents of the recursive types in the typing context, where contractivity is guaranteed by requiring that the eliminated recursion variable may not appear unguarded in each of the context’s types. At the same time, the recursive types in the context are unfolded, and their priorities are lifted by a common value, denoted $t$ in the rule, that must be greater than the highest priority appearing in the original types (cf. Definition 6). Using a “common lifter”, i.e., lifting the priorities of all types by the same amount is crucial: it maintains the relation between the priorities of the types in the context. Axiom ‘Var’ types a recursive call on a variable in the recursive context. The rule requires that all the types in the context are recursive on the recursion variable called, and that the types inside the recursive definitions concur with the respective types assigned to the recursion varialbe in the recursive context. As mentioned before, the types associated to the introduced and consequently eliminated recursion variable is crucial in ensuring that a recursion is called with endpoints of the same type as required by its definition. The binding of output and selection actions to continuation processes (1) is derivable in APCP. The corresponding typing rules in Figure 4 (bottom) are admissible using ‘Mix’ and ‘Cycle’ (cf. [51]). Figure 4 (bottom) also includes an admissible rule ‘Lift’ that lifts a process’ priorities. The following result assures that, given a type, we can construct a process with an endpoint typable with the given type: ###### Proposition 1. Given a type $A$, there exists a $P$ such that $P\vdash\Omega;x{:}\leavevmode\nobreak\ A$. ###### Proof. We inductively define a function ‘$\mathrm{char}^{x}(A)$’ that, given a type $A$ and an endpoint $x$, constructs a process that performs the behavior described by $A$: $\displaystyle\mathrm{char}^{x}(A\mathbin{\otimes}^{\mathsf{o}}B)$ $\displaystyle:=\overline{x}[y]\cdot(\mathrm{char}^{y}(A)\mathbin{|}\mathrm{char}^{x}(B))$ $\displaystyle\mathrm{char}^{x}({\oplus}^{\mathsf{o}}\\{i:A_{i}\\}_{i\in I})$ $\displaystyle:=\overline{x}\mathbin{\triangleleft}j\cdot\mathrm{char}^{x}(A_{j})\quad\text{[any $j\in I$]}$ $\displaystyle\mathrm{char}^{x}(A\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}}B)$ $\displaystyle:=x(y)\mathbin{.}(\mathrm{char}^{y}(A)\mathbin{|}\mathrm{char}^{x}(B))$ $\displaystyle\mathrm{char}^{x}(\&^{\mathsf{o}}\\{i:A_{i}\\}_{i\in I})$ $\displaystyle:=x\mathbin{\triangleright}\\{i:\mathrm{char}^{x}(A_{i})\\}_{i\in I}$ $\displaystyle\mathrm{char}^{x}(\bullet)$ $\displaystyle:=\bm{0}\qquad\qquad\mathrm{char}^{x}(\mu X\mathbin{.}A):=\mu X(x)\mathbin{.}\mathrm{char}^{x}(A)\qquad\qquad\mathrm{char}^{x}(X):=X{\langle x\rangle}$ For finite types, we have: $\mathrm{char}^{x}(A)\vdash\emptyset;x{:}\leavevmode\nobreak\ A$. For simplicity, we omit details about recursive types, which require unfolding. For closed, recursive types, we have: $\mathrm{char}^{x}(\mu X\mathbin{.}A)\vdash\emptyset;x{:}\leavevmode\nobreak\ \mu X\mathbin{.}A$. ∎ ##### Type Preservation Well-typed processes satisfy protocol fidelity, communication safety, and deadlock freedom. The first two properties follow directly from _type preservation_ (also known as _subject reduction_), which ensures that reduction preserves typing. In contrast to Caires and Pfenning [14] and Wadler [54], where type preservation corresponds to the elimination of (top-level) applications of rule Cut, in APCP it corresponds to the more general elimination of (top-level) applications of rule Cycle. ###### Theorem 2 (Type Preservation [51]). If $P\vdash\Omega;\Gamma$ and $P\longrightarrow Q$, then $Q\vdash\Omega;{\uparrow^{t}}\Gamma$ for $t\in\mathbb{N}$. ##### Deadlock Freedom The deadlock freedom result for APCP adapts that for PCP [22]. As mentioned before, binding asynchronous outputs and selections to continuations involves additional, low-level uses of Cycle, which we cannot eliminate through process reduction. Therefore, top-level deadlock freedom holds for _live processes_ (Theorem 4). A process is live if it is equivalent to a restriction on _active names_ that perform unguarded actions. This way, e.g., in ‘$x[y,z]$’ the name $x$ is active, but $y$ and $z$ are not. ###### Definition 7 (Active Names). The _set of active names_ of $P$, denoted ‘$\mathrm{an}(P)\mkern-2.0mu$’, contains the (free) names that are used for unguarded actions (output, input, selection, branching): $\displaystyle\mathrm{an}(x[y,z])$ $\displaystyle:=\\{x\\}$ $\displaystyle\mathrm{an}(x(y,z)\mathbin{.}P)$ $\displaystyle:=\\{x\\}$ $\displaystyle\mathrm{an}(\bm{0})$ $\displaystyle:=\emptyset$ $\displaystyle\mathrm{an}(x[z]\mathbin{\triangleleft}j)$ $\displaystyle:=\\{x\\}$ $\displaystyle\mathrm{an}(x(z)\mathbin{\triangleright}\\{i{:}\leavevmode\nobreak\ P_{i}\\}_{i\in I})$ $\displaystyle:=\\{x\\}$ $\displaystyle\mathrm{an}(x\mathbin{\leftrightarrow}y)$ $\displaystyle:=\\{x,y\\}$ $\displaystyle\mathrm{an}(P\mathbin{|}Q)$ $\displaystyle:=\mathrm{an}(P)\cup\mathrm{an}(Q)$ $\displaystyle\mathrm{an}(\mu X(\tilde{x})\mathbin{.}P)$ $\displaystyle:=\mathrm{an}(P)$ $\displaystyle\mathrm{an}((\bm{\nu}xy)P)$ $\displaystyle:=\mathrm{an}(P)\setminus\\{x,y\\}$ $\displaystyle\mathrm{an}(X{\langle\tilde{x}\rangle})$ $\displaystyle:=\emptyset$ ###### Definition 8 (Live Process). A process $P$ is _live_ , denoted ‘$\mkern 2.0mu\mathrm{live}(P)\mkern-3.0mu$’, if there are names $x,y$ and process $P^{\prime}$ such that $P\equiv(\bm{\nu}xy)P^{\prime}$ with $x,y\in\mathrm{an}(P^{\prime})$. We additionally need to account for recursion: as recursive definitions do not entail reductions, we must fully unfold them before eliminating Cycles. ###### Lemma 3 (Unfolding). If $P\vdash\Omega;\Gamma$, then there is a process $P^{\star}$ such that $P^{\star}\equiv P$ and $P^{\star}$ is not of the form ‘$\mu X(\tilde{z});P^{\prime}\mkern-2.0mu$’ and $P^{\star}\vdash\Omega;\Gamma$. Deadlock freedom, given next, states that typable processes that are live can reduce. It follows from an analysis of the priorities in the typing of the process, which makes it possible to find a pair of non-blocked, parallel, dual actions on connected endpoints, such that a communication can occur. The analysis also considers the possibility that a blocking action is on an endpoint which is not connected (i.e., the endpoint is free), in which case a commuting conversion can be performed. Confer the full proof by Van den Heuvel and Pérez [51, Theorem 5] for more details. ###### Theorem 4 (Deadlock Freedom). If $P\vdash\emptyset;\Gamma$ and $\mathrm{live}(P)$, then there is process $Q$ such that $P\longrightarrow Q$. We now state the deadlock freedom result formalized by Van den Heuvel and Pérez [51]. Following, e.g., Caires and Pfenning [14] and Dardha and Gay [22], it concerns processes typable under empty contexts. This way, the reduction guaranteed by Theorem 4 corresponds to a synchronization ($\beta$-rule), rather than a commuting conversion ($\kappa$-rule). ###### Theorem 5 (Deadlock Freedom for Processes Typable under Empty Contexts [51]). If $P\vdash\emptyset;\emptyset$, then either $P\equiv\bm{0}$ or $P\longrightarrow_{\beta}Q\mkern 2.0mu$ for some $Q$. ##### Fairness Processes typable under empty contexts are not only deadlock free, they are _fair_ : for each endpoint in the process, we can eventually observe a reduction involving that endpoint. To formalize this property, we define _labeled reductions_ , which expose details about a communication: ###### Definition 9 (Labeled Reductions). Consider the labels $\displaystyle\alpha::=x\mathbin{\leftrightarrow}y\leavevmode\nobreak\ \;\mbox{\large{$\mid$}}\;\leavevmode\nobreak\ x\rangle y{:}a\leavevmode\nobreak\ \;\mbox{\large{$\mid$}}\;\leavevmode\nobreak\ x\rangle y{:}\ell\qquad\qquad\text{(forwarding, output/input, selection/branching)}$ where each label has subjects $x$ and $y$. The _labeled reduction_ ‘$\mkern 1.0muP\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color<EMAIL_ADDRESS>is defined by the following rules: $\displaystyle(\bm{\nu}yz)(x\mathbin{\leftrightarrow}y\mathbin{|}P)\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}x\mathbin{\leftrightarrow}y}}}}P\\{x/z\\}\qquad(\bm{\nu}xy)(x[a,b]\mathbin{|}y(v,z)\mathbin{.}P)\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}x\rangle y{:}a}}}}P\\{a/v,b/z\\}$ $\displaystyle(\bm{\nu}xy)(x[b]\mathbin{\triangleleft}j\mathbin{|}y(z)\mathbin{\triangleright}\\{i{:}\leavevmode\nobreak\ P_{i}\\}_{i\in I})\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}x\rangle y{:}j}}}}P_{j}\\{b/z\\}\quad\text{(if $j\in I$)}$ $(P\equiv P^{\prime})\wedge(P^{\prime}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q^{\prime})\wedge(Q^{\prime}\equiv Q)$ $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q$ $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q$ $(\bm{\nu}xy)P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}(\bm{\nu}xy)Q$ $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q$ $P\mathbin{|}R\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q\mathbin{|}R$ ###### Proposition 6. For any $P$ and $P^{\prime}$, $P\longrightarrow_{\beta}P^{\prime}$ if and only if there exists a label $\alpha$ such that $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}P^{\prime}$. ###### Proof. Immediate by definition, for each $\beta$-reduction in Figure 3 (bottom) corresponds to a labeled reduction, and vice versa. ∎ Our fairness result states that processes typable under empty contexts have at least one finite reduction sequence (‘$\longrightarrow^{\star}$’) that enables a labeled reduction involving a _pending_ endpoint—an endpoint that occurs as the subject of an action, and is not bound by input or branching (see below). Clearly, the typed process may have other reduction sequences, not necessarily finite. ###### Definition 10 (Pending Names). Given a process $P$, we define the set of _pending names_ of $P$, denoted ‘$\mkern 1.0mu\mathrm{pn}(P)\mkern-3.0mu$’, as follows: $\displaystyle\mathrm{pn}(x[y,z])$ $\displaystyle:=\\{x\\}$ $\displaystyle\mathrm{pn}(x(y,z).P)$ $\displaystyle:=\\{x\\}\cup(\mathrm{pn}(P)\setminus\\{y,z\\})$ $\displaystyle\mathrm{pn}(\bm{0})$ $\displaystyle:=\emptyset$ $\displaystyle\mathrm{pn}(x[z]\mathbin{\triangleleft}j)$ $\displaystyle:=\\{x\\}$ $\displaystyle\mathrm{pn}(x(z)\mathbin{\triangleright}\\{i:P_{i}\\}_{i\in I})$ $\displaystyle:=\\{x\\}\cup({\mathchoice{\textstyle}{}{}{}\bigcup}_{i\in I}\mathrm{pn}(P_{i})\setminus\\{z\\})$ $\displaystyle\mathrm{pn}(x\mathbin{\leftrightarrow}y)$ $\displaystyle:=\\{x,y\\}$ $\displaystyle\mathrm{pn}(P\mathbin{|}Q)$ $\displaystyle:=\mathrm{pn}(P)\cup\mathrm{pn}(Q)$ $\displaystyle\mathrm{pn}(\mu X(\tilde{x})\mathbin{.}P)$ $\displaystyle:=\mathrm{pn}(P)$ $\displaystyle\mathrm{pn}((\bm{\nu}xy)P)$ $\displaystyle:=\mathrm{pn}(P)$ $\displaystyle\mathrm{pn}(X{\langle\tilde{x}\rangle})$ $\displaystyle:=\emptyset$ ###### Theorem 7 (Fairness). Suppose given a process $P\vdash\emptyset;\emptyset$. Then, for every $x\in\mathrm{pn}(P)$ there exists a process $P^{\prime}$ such that $P\longrightarrow^{\star}P^{\prime}$ and $P^{\prime}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}\,Q$, for some process $Q$ and label $\alpha$ with subject $x$. ###### Proof. Take any $x\in\mathrm{pn}(P)$. Because $P$ is typable under empty contexts, $x$ is bound to some $y\in\mathrm{pn}(P)$ by restriction. By typing, in $P$ there is exactly one action on $x$ and one action on $y$ (they may also appear in forwarder processes). Following the restrictions on priorities in the typing of $x$ and $y$ in $P$, the actions on $x$ and $y$ cannot appear sequentially in $P$ (cf. the proof by Van den Heuvel and Pérez [51] for details on this reasoning). By typability, the action on $y$ is dual to the action on $x$. We apply induction on the number of inputs, branches, and recursive definitions in $P$ blocking the actions on $x$ and $y$, denoted $n$ and $m$, respectively. Because $P$ is typable under empty contexts, the blocking inputs and branches that are on names in $\mathrm{pn}(P)$ also have to be bound to pending names by restriction. The actions on these connected names may also be prefixed by inputs, branches, and recursive definitions, so we may need to unblock those actions as well. Since there can only be a finite number of names in any given process, we also apply induction on the number of prefixes blocking these connected actions. * • If $n=0$ and $m=0$, then the actions on $x$ and $y$ occur at the top-level; because they do not appear sequentially, the communication between $x$ and $y$ can take place immediately. Hence, $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q$ where $x$ and $y$ are the subjects of $\alpha$. This proves the thesis, with $P^{\prime}=P$. * • If $n>0$ or $m>0$, the analysis depends on the foremost prefix of the actions on $x$ and $y$. If the foremost prefix of either action is a recursive definition (‘$\mu X(\tilde{y})$’), we unfold the recursion. Because a corresponding recursive call (‘$X{\langle\tilde{z}\rangle}$’) cannot occur as a prefix, the effect of unfolding either (i) triggers actions that occur in parallel to those on $x$ and $y$, or (ii) the actions on $x$ or $y$ prefix the unfolded recursive call. In either case, the number of prefixes decreases, and the thesis follows from the IH. Otherwise, if neither foremost prefix is a recursive definition, then the foremost prefixes must be actions on names in $\mathrm{pn}(P)$. Consider the action that is typable with the least priority. W.l.o.g. assume that this is the foremost prefix of $x$. Suppose this action is on some endpoint $w$ connected to another endpoint $z\in\mathrm{pn}(P)$ by restriction. By typability, the priority of $w$ is less than that of $x$ and all of the prefixes in between. This means that the number of prefixes blocking the action on $z$ strictly decreases. Hence, by the IH, $P\longrightarrow^{\star}P^{\prime\prime}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}}}}}\,Q^{\prime}$ in a finite number of steps, where $w$ and $z$ are the subjects of $\alpha^{\prime}$. The communication between $w$ and $z$ can be performed, and $n$ decreases. By Type Preservation (Theorem 2), $Q^{\prime}\vdash\emptyset;\emptyset$. The thesis then follows from the IH: $P\longrightarrow^{\star}P^{\prime\prime}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}}}}}\,Q^{\prime}\longrightarrow^{\star}P^{\prime}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}\,Q$ in finite steps, where $x$ and $y$ are the subjects of $\alpha$. ∎ ### Examples To illustrate APCP processes and their session types, we give implementations of the three participants in $G_{\mathsf{auth}}$ in Section 1. ###### Example 1. Processes $P$, $Q$, and $R$ are typed implementations for participants $c$, $s$, and $a$, respectively, where each process uses a single channel to perform the actions described by $G_{\mathsf{auth}}$. $\displaystyle P$ $\displaystyle:=\mu X({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})\mathbin{.}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\mathbin{\triangleright}\left\\{\begin{array}[]{ll}\mathsf{login}{:}&{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}(u)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}\mathbin{\triangleleft}\mathsf{passwd}\cdot\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}[\bm{logmein345}]\cdot X{\langle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle},\\\ \mathsf{quit}{:}&{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}(w)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}\mathbin{\triangleleft}\mathsf{quit}\cdot\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}[z]\cdot\bm{0}\end{array}\right\\}$ $\displaystyle\vdash{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ \mu X\mathbin{.}\&^{2}\left\\{\begin{array}[]{ll}\mathsf{login}{:}&\bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{3}{\oplus}^{4}\\{\mathsf{passwd}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{5}X\\},\\\ \mathsf{quit}{:}&\bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{3}{\oplus}^{4}\\{\mathsf{quit}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{5}\bullet\\}\end{array}\right\\}$ $\displaystyle Q$ $\displaystyle:=\mu X({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}\mathbin{\triangleleft}\mathsf{login}\cdot\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}[u]\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\mathbin{\triangleright}\\{\mathsf{auth}{:}\leavevmode\nobreak\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}(v)\mathbin{.}X{\langle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle}\\}$ $\displaystyle\vdash{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ \mu X\mathbin{.}{\oplus}^{0}\\{\mathsf{login}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{1}\&^{10}\\{\mathsf{auth}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{11}X\\},\mathsf{quit}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{1}\bullet\\}$ $\displaystyle R$ $\displaystyle:=\mu X({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})\mathbin{.}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\mathbin{\triangleright}\left\\{\begin{array}[]{ll}\mathsf{login}{:}&{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\mathbin{\triangleright}\\{\mathsf{passwd}{:}\leavevmode\nobreak\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}(u)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}\mathbin{\triangleleft}\mathsf{auth}\cdot\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}[v]\cdot X{\langle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle}\\},\\\ \mathsf{quit}{:}&{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\mathbin{\triangleright}\\{\mathsf{quit}{:}\leavevmode\nobreak\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}(w)\mathbin{.}\bm{0}\\}\end{array}\right\\}$ $\displaystyle\vdash{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ \mu X\mathbin{.}\&^{2}\left\\{\begin{array}[]{ll}\mathsf{login}{:}&\&^{6}\\{\mathsf{passwd}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{7}{\oplus}^{8}\\{\mathsf{auth}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{9}X\\}\\},\\\ \mathsf{quit}{:}&\&^{6}\\{\mathsf{quit}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{7}\bullet\\}\end{array}\right\\}$ Process $P$ is a specific implementation for $c$, where we use ‘$\bm{logmein345}$’ to denote a closed channel endpoint representing the password string “logmein345”. Similarly, $Q$ is a specific implementation for $s$ that continuously chooses the login branch. Note that the processes above cannot be directly connected to each other to implement $G_{\mathsf{auth}}$. Our goal is to enable the composition of (typed) implementations such as $P$, $Q$, and $R$ in a correct and deadlock free manner. We shall proceed as follows. After setting up the routers that enable the composition of these processes according to $G_{\mathsf{auth}}$ (Section 4), we will return to this example in Section 5. At that point, it will become clear that the priorities in the types of $P$, $Q$, and $R$ were chosen to ensure the correct composition with their respective routers. ## 3 Global Types and Relative Projection We analyze multiparty protocols specified as _global types_. We consider a standard syntax, with session delegation and recursion, subsuming the one given in the seminal paper by Honda _et al._ [36]. In the following, we write $p,q,r,s,\ldots$ to denote _(protocol) participants_. ###### Definition 11 (Types). _Global types_ $G$ and _message types_ $S,T$ are defined as: $\displaystyle G$ $\displaystyle::=p\mathbin{\twoheadrightarrow}q\\{i\langle S\rangle\mathbin{.}G\\}_{i\in I}\;\mbox{\large{$\mid$}}\;\mu X\mathbin{.}G\;\mbox{\large{$\mid$}}\;X\;\mbox{\large{$\mid$}}\;\bullet\;\mbox{\large{$\mid$}}\;\mathsf{skip}\mathbin{.}G$ $\displaystyle S,T$ $\displaystyle::={!}T\mathbin{.}S\;\mbox{\large{$\mid$}}\;{?}T\mathbin{.}S\;\mbox{\large{$\mid$}}\;{{\oplus}}\\{i{:}\leavevmode\nobreak\ S\\}_{i\in I}\;\mbox{\large{$\mid$}}\;\&\\{i{:}\leavevmode\nobreak\ S\\}_{i\in I}\;\mbox{\large{$\mid$}}\;\bullet$ We include basic types (e.g., unit, bool, int), which are all syntactic sugar for $\bullet$. The type ‘$p\mathbin{\twoheadrightarrow}q\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}$’ specifies a direct exchange from participant $p$ to participant $q$, which precedes protocol $G_{i}$: $p$ chooses a label $i\in I$ and sends it to $q$ along with a message of type $S_{i}$. Message exchange is _asynchronous_ : the protocol can continue as $G_{i}$ before the message has been received by $q$. The type ‘$\mu X\mathbin{.}G$’ defines a recursive protocol: whenever a path of exchanges in $G$ reaches the recursion variable $X$, the protocol continues as ‘$\mu X\mathbin{.}G$’. The type ‘$\bullet$’ denotes the completed protocol. For technical convenience, we introduce the construct ‘$\mathsf{skip}\mathbin{.}G$’, which denotes an unobservable step that precedes $G$. Recursive definitions bind recursion variables, so recursion variables not bound by a recursive definition are free. We write ‘$\mathrm{frv}(G)$’ to denote the set of free recursion variables of $G$, and say $G$ is _closed_ if $\mathrm{frv}(G)=\emptyset$. Recursion in global types is tail-recursive and _contractive_ (i.e. they contain no subexpressions of the form ‘$\mu X_{1}\ldots\mu X_{n}\mathbin{.}X_{1}$’). As for the session types in Section 2, we define the unfolding of a recursive global type by substituting copies of the recursive definition for recursive calls, i.e. ‘$\mu X\mathbin{.}G$’ unfolds to ‘$G\\{\mu X\mathbin{.}G/X\\}$’. In approaches based on MPST, the grammar of global types specifies multiparty protocols but does not ensure their correct implementability; such guarantees are given in terms of _well-formedness_ , defined as projectability onto all participants (cf. § 3.2). Message types $S,T$ define binary protocols, not to be confused with the types in § 2. Type ‘${!}T\mathbin{.}S$’ (resp. ‘${?}T\mathbin{.}S$’) denotes the output (resp. input) of a message of type $T$ followed by the continuation $S$. Type ‘${{\oplus}}\\{i{:}\leavevmode\nobreak\ S_{i}\\}_{i\in I}$’ denotes _selection_ : the output of choice for a label $i\in I$ followed by the continuation $S_{i}$. Type ‘$\&\\{i{:}\leavevmode\nobreak\ S_{i}\\}_{i\in I}$’ denotes _branching_ : the input of a label $i\in I$ followed by the continuation $S_{i}$. Type ‘$\bullet$’ denotes the end of the protocol. Note that, due to the tail-recursiveness of session and global types, there are no recursive message types. It is useful to obtain the set of participants of a global type: ###### Definition 12 (Participants). We define the _set of participants_ of global type $G$, denoted ‘$\mkern 1.0mu\mathsf{prt}(G)\mkern-2.0mu$’: $\displaystyle\mathsf{prt}(p\mathbin{\twoheadrightarrow}q\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I})$ $\displaystyle:=\\{p,q\\}\cup({\mathchoice{\textstyle}{}{}{}\bigcup}_{i\in I}\leavevmode\nobreak\ \mathsf{prt}(G_{i}))$ $\displaystyle\mathsf{prt}(\mathsf{skip}\mathbin{.}G)$ $\displaystyle:=\mathsf{prt}(G)$ $\displaystyle\mathsf{prt}(\bullet)$ $\displaystyle:=\emptyset$ $\displaystyle\mathsf{prt}(\mu X\mathbin{.}G)$ $\displaystyle:=\mathsf{prt}(G)$ $\displaystyle\mathsf{prt}(X)$ $\displaystyle:=\emptyset$ ### 3.1 Relative Types While a global type such as $G_{\mathsf{auth}}$ (1) describes a protocol from a vantage point, we introduce _relative types_ that describe the interactions between _pairs_ of participants. This way, relative types capture the peer-to- peer nature of multiparty protocols. We develop projection from global types onto relative types (cf. § 3.2) and use it to establish a new class of _well- formed_ global types. A choice between participants in a global type is _non-local_ if it influences future exchanges between other participants. Our approach uses _dependencies_ to expose these non-local choices in the relative types of these other participants. Relative types express interactions between two participants. Because we obtain a relative type through projection of a global type, we know which participants are involved. Therefore, a relative type only mentions the sender of each exchange; we implicitly know that the recipient is the other participant. ###### Definition 13 (Relative Types). _Relative types_ $R$ are defined as follows, where the $S_{i}$ are message types (cf. Def. 11): $R::=p\\{i\langle S_{i}\rangle\mathbin{.}R\\}_{i\in I}\;\mbox{\large{$\mid$}}\;p{?}r\\{i\mathbin{.}R\\}_{i\in I}\;\mbox{\large{$\mid$}}\;p{!}r\\{i\mathbin{.}R\\}_{i\in I}\;\mbox{\large{$\mid$}}\;\mu X\mathbin{.}R\;\mbox{\large{$\mid$}}\;X\;\mbox{\large{$\mid$}}\;\bullet\;\mbox{\large{$\mid$}}\;\mathsf{skip}\mathbin{.}R$ We detail the syntax above, given participants $p$ and $q$. * • Type ‘$p\\{i\langle S_{i}\rangle\mathbin{.}R_{i}\\}_{i\in I}$’ specifies that $p$ must choose a label $i\in I$ and send it to $q$ along with a message of type $S_{i}$ after which the protocol continues with $R_{i}$. * • Given an $r$ which is _not_ involved in the relative type (i.e., $p\neq r,q\neq r$), type ‘$p{?}r\\{i\mathbin{.}R_{i}\\}_{i\in I}$’ expresses a dependency: a non-local choice between $p$ and $r$ which influences the protocol between $p$ and $q$. Here, the dependency indicates that after $p$ receives from $r$ the chosen label, $p$ must forward it to $q$, determining the protocol between $p$ and $q$. * • Similarly, type ‘$p{!}r\\{i\mathbin{.}R_{i}\\}_{i\in I}$’ expresses a dependency, which indicates that after $p$ sends to $r$ the chosen label, $p$ must forward it to $q$. * • Types ‘$\mu X\mathbin{.}R$’ and ‘$X$’ define recursion, just as their global counterparts. * • The type ‘$\bullet$’ specifies the end of the protocol between $p$ and $q$. * • The type ‘$\mathsf{skip}\mathbin{.}R$’ denotes an unobservable step that precedes $R$. ###### Definition 14 (Participants of Relative Types). We define the _set of participants_ of relative type $R$, denoted ‘$\mkern 1.0mu\mathsf{prt}(R)\mkern-2.0mu$’: $\displaystyle\mathsf{prt}(p\\{i\langle S_{i}\rangle\mathbin{.}R_{i}\\}_{i\in I})$ $\displaystyle:=\\{p\\}\cup({\mathchoice{\textstyle}{}{}{}\bigcup}_{i\in I}\leavevmode\nobreak\ \mathsf{prt}(R_{i}))$ $\displaystyle\mathsf{prt}(\mathsf{skip}\mathbin{.}R)$ $\displaystyle:=\mathsf{prt}(R)$ $\displaystyle\mathsf{prt}(\bullet)$ $\displaystyle:=\emptyset$ $\displaystyle\mathsf{prt}(p{?}r\\{i\mathbin{.}R_{i}\\}_{i\in I})$ $\displaystyle:=\\{p\\}\cup({\mathchoice{\textstyle}{}{}{}\bigcup}_{i\in I}\leavevmode\nobreak\ \mathsf{prt}(R_{i}))$ $\displaystyle\mathsf{prt}(\mu X\mathbin{.}R)$ $\displaystyle:=\mathsf{prt}(R)$ $\displaystyle\mathsf{prt}(X)$ $\displaystyle:=\emptyset$ $\displaystyle\mathsf{prt}(p{!}r\\{i\mathbin{.}R_{i}\\}_{i\in I})$ $\displaystyle:=\\{p\\}\cup({\mathchoice{\textstyle}{}{}{}\bigcup}_{i\in I}\leavevmode\nobreak\ \mathsf{prt}(R_{i}))$ We introduce some useful notation: ###### Notation 3. * • We write ‘$\mkern 1.0mup\mathbin{\twoheadrightarrow}q{:}i\langle S\rangle\mathbin{.}G\mkern-3.0mu$’ for a global type with a single branch ‘$\mkern 1.0mup\mathbin{\twoheadrightarrow}q\\{i\langle S\rangle\mathbin{.}G\\}\mkern-3.0mu$ (and similarly for exchanges and dependencies in relative types). * • We omit ‘unit’ message types from global and relative types, writing ‘$\mkern 1.0mui\mathbin{.}G\mkern-2.0mu$’ for ‘$\mkern 1.0mui\langle\mathsf{unit}\rangle\mathbin{.}G\mkern-2.0mu$’. * • Given $k>1$, we write ‘$\mkern 1.0mu\mathsf{skip}^{k}\mkern-3.0mu$’ for a sequence of $k$ $\mathsf{skip}$s. ### 3.2 Relative Projection and Well-Formedness $\displaystyle\mathrm{ddep}((p,q),s\mathbin{\twoheadrightarrow}r\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}):=\begin{cases}\mathsf{skip}\mathbin{.}(G_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))\leavevmode\nobreak\ \text{[any $i^{\prime}\in I$]}&\text{if $\forall i,j.$}\\\ &\text{$G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=G_{j}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$}\\\ p{!}r\\{i\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))\\}_{i\in I}&\text{if $p=s$}\\\ q{!}r\\{i\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))\\}_{i\in I}&\text{if $q=s$}\\\ p{?}s\\{i\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))\\}_{i\in I}&\text{if $p=r$}\\\ q{?}s\\{i\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))\\}_{i\in I}&\text{if $q=r$}\end{cases}$ . $\displaystyle(s\mathbin{\twoheadrightarrow}r\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I})\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ $\displaystyle:=\begin{cases}p\\{i\langle S_{i}\rangle\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))\\}_{i\in I}&\text{if $p=s$ and $q=r$}\\\ q\\{i\langle S_{i}\rangle\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))\\}_{i\in I}&\text{if $q=s$ and $p=r$}\\\ \mathrm{ddep}((p,q),s\mathbin{\twoheadrightarrow}r\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I})&\text{otherwise}\end{cases}$ $\displaystyle(\mu X\mathbin{.}G)\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ $\displaystyle:=\begin{cases}\mu X\mathbin{.}(G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))&\text{if $G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ defined and contractive on $X$}\\\ \bullet&\text{otherwise}\end{cases}$ $\displaystyle X\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q):=X\qquad\bullet\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q):=\bullet\qquad(\mathsf{skip}\mathbin{.}G)\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q):=\mathsf{skip}\mathbin{.}(G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))$ Above, ‘$\mathsf{skip}^{\ast}$’ denotes a sequence of zero or more $\mathsf{skip}$. Figure 5: Dependency Detection (top), and Relative Projection (bottom, cf. Definition 16). When a side-condition does not hold, either is undefined. We define _relative projection_ for global types. We want relative projection to fail when it would return a non-contractive recursive type. To this end, we define a notion of contractiveness on relative types: ###### Definition 15 (Contractive Relative Types). Given a relative type $R$ and a recursion variable $X$, we say _$R$ is contractive on $X$_ if either of the following holds: * • $R$ contains an exchange, or * • $R$ ends in a recursive call on a variable other than $X$. Relative projection then relies on the contractiveness of relative types. It also relies on an auxiliary function to determine if a dependency message is needed and possible. ###### Definition 16 (Relative Projection). Given a global type $G$, we define its relative projection onto a pair of participants $p$ and $q$, denoted ‘ $G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$​’, by induction on the structure of $G$ as given in Figure 5 (bottom), using the auxiliary function $\mathrm{ddep}$ (cf. Figure 5, top). We discuss how Definition 16 projects global types onto a pair of participants $(p,q)$, as per Figure 5 (bottom). The most interesting case is the projection of a direct exchange ‘$s\mathbin{\twoheadrightarrow}r\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}$’. When the exchange involves both $p$ and $q$, the projection yields an exchange between $p$ and $q$ with the appropriate sender. Otherwise, the projection relies on the function ‘$\mathrm{ddep}$’ in Figure 5 (top), which determines whether the exchange is a non-local choice for $p$ and $q$ and yields an appropriate projection accordingly: * • If the projections of all branches are equal, the exchange is not a non-local choice and $\mathrm{ddep}$ yields the unobservable step ‘$\mathsf{skip}$’ followed by the projection of any branch. * • If there are branches with different projections, the exchange is a non-local choice, so $\mathrm{ddep}$ yields a dependency if possible. If $p$ or $q$ is involved in the exchange, $\mathrm{ddep}$ yields an appropriate dependency (e.g., ‘$p{!}r$’ if $p$ is the sender, or ‘$q{?}s$’ if $q$ is the recipient). If neither $p$ nor $q$ are involved, then $\mathrm{ddep}$ cannot yield a dependency and projection is thus undefined. The projection of ‘$\mu X\mathbin{.}G^{\prime}$’ considers the projection of the body ‘$G^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$’ to see whether $p$ and $q$ interact in $G^{\prime}$. If $G^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ is a (possibly empty) sequence of $\mathsf{skip}$s followed by $\bullet$ or $X$, then $p$ and $q$ do not interact and the projection yields $\bullet$. Otherwise, $p$ and $q$ do interact and projection preserves the recursive definition. Note that Definition 15 (contractiveness) is key here: e.g., $G^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=\mathsf{skip}\mathbin{.}\mu Y\mathbin{.}\mathsf{skip}\mathbin{.}X$ is not contractive on $X$, so $(\mu X\mathbin{.}G^{\prime})\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=\bullet$. The projection of a recursive call ‘$X$’ is simply ‘$X$’. The projection of ‘$G_{1}\mathbin{|}G_{2}$’ is standard [35]: it ensures that $G_{1}$ and $G_{2}$ do not share participants and only continues with either global type if both $p$ and $q$ are participants. The projections of ‘$\bullet$’ and ‘$\mathsf{skip}$’ are homomorphic. ###### Example 2 (Projections of $G_{\mathsf{auth}}$). To demonstrate relative projection, let us consider again $G_{\mathsf{auth}}$: $\displaystyle G_{\mathsf{auth}}=\mu X\mathbin{.}s\mathbin{\twoheadrightarrow}c\left\\{\begin{array}[]{@{}l@{}}\mathsf{login}\mathbin{.}c\mathbin{\twoheadrightarrow}a{:}\mathsf{passwd}\langle\mathsf{str}\rangle\mathbin{.}a\mathbin{\twoheadrightarrow}s{:}\mathsf{auth}\langle\mathsf{bool}\rangle\mathbin{.}X,\\\ \mathsf{quit}\mathbin{.}c\mathbin{\twoheadrightarrow}a{:}\mathsf{quit}\mathbin{.}\bullet\end{array}\right\\}$ The relative projection onto $(s,c)$ is straightforward, as there are no non- local choices to consider: $\displaystyle G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(s,c)=\mu X\mathbin{.}s\left\\{\begin{array}[]{@{}l@{}}\mathsf{login}\mathbin{.}\mathsf{skip}^{2}\mathbin{.}X,\\\ \mathsf{quit}\mathbin{.}\mathsf{skip}\mathbin{.}\bullet\end{array}\right\\}$ However, compare the projection of the initial login branch onto $(s,a)$ and $(c,a)$ with the projection of the quit branch: they are different. Therefore, the initial exchange between $s$ and $c$ is a non-local choice in the protocols relative to $(s,a)$ and $(c,a)$. Since $s$ is involved in this exchange, the non-local choice is detected by ‘$\mkern 1.0mu\mathrm{ddep}\mkern-3.0mu$’: $\displaystyle\mathrm{ddep}((s,a),s\mathbin{\twoheadrightarrow}c\\{\mathsf{login}\ldots,\quad\mathsf{quit}\ldots\\})=s{!}c\\{\mathsf{login}\ldots,\quad\mathsf{quit}\ldots\\}$ Hence, this non-local choice can be included in the relative projection onto $(s,a)$ as a dependency: $\displaystyle G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(s,a)=\mu X\mathbin{.}s{!}c\left\\{\begin{array}[]{@{}l@{}}\mathsf{login}\mathbin{.}\mathsf{skip}\mathbin{.}a{:}\mathsf{auth}\langle\mathsf{bool}\rangle\mathbin{.}X,\\\ \mathsf{quit}\mathbin{.}\mathsf{skip}\mathbin{.}\bullet\end{array}\right\\}$ Similarly, $c$ is involved in the initial exchange, so the non-local choice can also be included in the relative projection onto $(c,a)$ as a dependency: $\displaystyle G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(c,a)=\mu X\mathbin{.}c{?}s\left\\{\begin{array}[]{@{}l@{}}\mathsf{login}\mathbin{.}c{:}\mathsf{passwd}\langle\mathsf{str}\rangle\mathbin{.}\mathsf{skip}\mathbin{.}X,\\\ \mathsf{quit}\mathbin{.}c{:}\mathsf{quit}\langle\mathsf{unit}\rangle\mathbin{.}\bullet\end{array}\right\\}$ Since relative types are relative to pairs of participants, the input order of participants for projection does not matter: ###### Proposition 8. Suppose a global type $G$ and distinct participants $p,q\in\mathsf{prt}(G)$. * • If $G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ is defined, then $G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(q,p)$ and $\mathsf{prt}(G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))\subseteq\\{p,q\\}$; * • $G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ is undefined if and only if $G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(q,p)$ is undefined. ##### Well-formed Global Types We may now define _well-formedness_ for global types. Unlike usual MPST approaches, our definition relies exclusively on (relative) projection (Def. 16), and does not appeal to external notions such as merge and subtyping [37, 55]. ###### Definition 17 (Relative Well-Formedness). A global type $G$ is _relative well-formed_ if, for every distinct $p,q\in\mathsf{prt}(G)$, the projection $G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ is defined. The following contrasts our new notion of relative well-formedness with notions of well-formedness based on the usual notion of local types [35, 26]. ###### Example 3. Consider the following global type involving participants $p,q,r,s$: $G_{3}:=p\mathbin{\twoheadrightarrow}q\left\\{\begin{array}[]{l}1\langle S_{a}\rangle\mathbin{.}p\mathbin{\twoheadrightarrow}r{:}1\langle S_{b}\rangle\mathbin{.}p\mathbin{\twoheadrightarrow}s{:}1\langle S_{c}\rangle\mathbin{.}q\mathbin{\twoheadrightarrow}r{:}1\langle S_{d}\rangle\mathbin{.}q\mathbin{\twoheadrightarrow}s{:}1\langle S_{e}\rangle\mathbin{.}\bullet,\\\ 2\langle S_{f}\rangle\mathbin{.}r\mathbin{\twoheadrightarrow}p{:}2\langle S_{g}\rangle\mathbin{.}s\mathbin{\twoheadrightarrow}p{:}2\langle S_{h}\rangle\mathbin{.}r\mathbin{\twoheadrightarrow}q{:}2\langle S_{i}\rangle\mathbin{.}s\mathbin{\twoheadrightarrow}q{:}2\langle S_{j}\rangle\mathbin{.}\bullet\end{array}\right\\}$ The initial exchange between $p$ and $q$ is a non-local choice influencing the protocols between other pairs of participants. Well-formedness as in [35, 26] forbids non-local choices. In contrast, $G_{3}$ is relative well-formed: $p$ and $q$ must both forward the selected label to both $r$ and $s$. The dependencies in the following relative projections express precisely this: $\displaystyle G_{3}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r)$ $\displaystyle=p{!}q\\{1\mathbin{.}p{:}1\langle S_{b}\rangle\mathbin{.}\mathsf{skip}^{3}\mathbin{.}\bullet,\quad 2\mathbin{.}r{:}2\langle S_{g}\rangle\mathbin{.}\mathsf{skip}^{3}\mathbin{.}\bullet\\}$ $\displaystyle G_{3}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s)$ $\displaystyle=p{!}q\\{1\mathbin{.}\mathsf{skip}\mathbin{.}p{:}1\langle S_{c}\rangle\mathbin{.}\mathsf{skip}^{2}\mathbin{.}\bullet,\quad 2\mathbin{.}\mathsf{skip}\mathbin{.}s{:}2\langle S_{h}\rangle\mathbin{.}\mathsf{skip}^{2}\mathbin{.}\bullet\\}$ $\displaystyle G_{3}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(q,r)$ $\displaystyle=q{?}p\\{1\mathbin{.}\mathsf{skip}^{2}\mathbin{.}q{:}1\langle S_{d}\rangle\mathbin{.}\mathsf{skip}\mathbin{.}\bullet,\quad 2\mathbin{.}\mathsf{skip}^{2}\mathbin{.}r{:}2\langle S_{i}\rangle\mathbin{.}\mathsf{skip}\mathbin{.}\bullet\\}$ $\displaystyle G_{3}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(q,s)$ $\displaystyle=q{?}p\\{1\mathbin{.}\mathsf{skip}^{3}\mathbin{.}q{:}1\langle S_{e}\rangle\mathbin{.}\bullet,\quad 2\mathbin{.}\mathsf{skip}^{3}\mathbin{.}s{:}2\langle S_{j}\rangle\mathbin{.}\bullet\\}$ Dependencies in relative types follow the non-local choices in the given global type: by implementing such choices, dependencies ensure correct projectability. They induce additional messages, but in our view this is an acceptable price to pay for an expressive notion of well-formedness based only on projection. It is easy to see that in a global type with $n$ participants, the number of messages per communication is $\mathcal{O}(n)$—an upper-bound following from the worst-case scenario in which both sender and recipient have to forward a label to $n-2$ participants due to dependencies, as in the example above. However, in practice, sender and recipient will rarely both have to forward labels, let alone both to all participants. ## 4 Analyzing Global Types using Routers $P$${\llbracket G_{\mathsf{auth}}\rrbracket}_{c}^{\\{s,a\\}}$${}\in\mathrm{ri}(G_{\mathsf{auth}},\\{c\\})$${\llbracket G_{\mathsf{auth}}\rrbracket}_{s}^{\\{a,s\\}}$$Q$${}\in\mathrm{ri}(G_{\mathsf{auth}},\\{s\\})$${\llbracket G_{\mathsf{auth}}\rrbracket}_{a}^{\\{s,c\\}}$$R$${}\in\mathrm{ri}(G_{\mathsf{auth}},\\{a\\})$ $P$${\llbracket G_{\mathsf{auth}}\rrbracket}_{c}^{\\{s,a\\}}$${}\in\mathrm{ri}(G_{\mathsf{auth}},\\{c\\})$${\llbracket G_{\mathsf{auth}}\rrbracket}_{s}^{\\{a,s\\}}$${\llbracket G_{\mathsf{auth}}\rrbracket}_{a}^{\\{s,c\\}}$$S$${}\in\mathrm{ri}(G_{\mathsf{auth}},\\{s,a\\})$ Figure 6: Two different networks of routed implementations for $G_{\mathsf{auth}}$ (1), without interleaving (left) and with interleaving (right). For participants $p$ and $\tilde{q}$, Definition 19 gives the router process $\mkern 1.0mu{\llbracket G\rrbracket}_{p}^{\tilde{q}}\mkern-3.0mu$ and Definition 24 gives the set $\mathrm{ri}(G,\tilde{q})$. Lines indicate channels and boxes are local compositions of processes. In this section, we develop our decentralized analysis of multiparty protocols (§ 3) using relative types (§ 3.1) and APCP (§ 2). The intended setup is as follows. Each participant’s role in a global type $G$ is implemented by a process, which is connected to a _router_ : a process that orchestrates the participant’s interactions in $G$. The resulting _routed implementations_ (Def. 24) can then directly connect to each other to form a decentralized _network of routed implementations_ that implements $G$. This way we realize the scenario sketched in Figure 1 (left), which is featured in more detail in Figure 6 (left). Key in our analysis is the _synthesis_ of a participant’s router from a global type (§ 4.1). To assert well-typedness—and thus deadlock freedom—of networks of routed implementations (Theorem 11), we extract binary session types from the global type and its associated relative types (§ 4.2): * • from the global type we extract types for channels between implementations and routers; * • from the relative types we extract types for channels between pairs of routers. After defining routers and showing their typability, we set up networks of routed implementations of global types (§ 4.3). To enable the transference of deadlock freedom APCP to multiparty protocols, we then establish an operational correspondence between global types and networks of routed implementations (Theorems 19 and 23). Finally, to show that our routed approach strictly generalizes the prior centralized analyses [12, 16], we define an orchestrated analysis of global types and show that it is behaviorally equivalent to a centralized composition of routers (§ 4.4). In the following section (§ 5), we will show routers in action. ### 4.1 Synthesis of Routers We synthesize routers by decomposing each exchange in the global type into four sub-steps, which we motivate by considering the initial exchange from $s$ to $c$ in $G_{\mathsf{auth}}$ 1: $s\mathbin{\twoheadrightarrow}c\\{\mathsf{login}\ldots,\quad\mathsf{quit}\ldots\\}$. As explained in Example 2, this exchange induces a dependency in the relative projections of $G_{\mathsf{auth}}$ onto $(s,a)$ and $(c,a)$. We decompose this initial exchange as follows, where $P$, $Q$, and $R$ are the implementations of $c$, $s$, and $a$, respectively (given in Example 1) and $\mathcal{R}_{x}$ stands for the router of each $x\in\\{s,c,a\\}$. Below, multiple actions in one step happen concurrently: 1. 1. $Q$ sends $\ell\in\\{\mathsf{login},\mathsf{quit}\\}$ to $\mathcal{R}_{s}$. 2. 2. $\mathcal{R}_{s}$ sends $\ell$ to $\mathcal{R}_{c}$ (recipient) and $\mathcal{R}_{a}$ (output dependency). $Q$ sends unit value $v$ to $\mathcal{R}_{s}$. 3. 3. $\mathcal{R}_{c}$ sends $\ell$ to $P$ and $\mathcal{R}_{a}$ (input dependency). $\mathcal{R}_{s}$ forwards $v$ to $\mathcal{R}_{c}$. 4. 4. $\mathcal{R}_{c}$ forwards $v$ to $P$. $\mathcal{R}_{a}$ sends $\ell$ to $R$. In Section 4.2, we follow this decomposition to assign to each consecutive step a consecutive priority: this ensures the consistency of priority checks required to establish the deadlock freedom of networks of routed implementations. We define router synthesis by means of an algorithm that returns a _router process_ for a given global type and participant. More precisely: given $G$, a participant $p$, and $\tilde{q}=\mathsf{prt}(G)\setminus\\{p\\}$, the algorithm generates a process, denoted ‘${\llbracket G\rrbracket}_{p}^{\tilde{q}}$’, which connects with a process implementing $p$’s role in $G$ on channel ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$; we shall write such channels in pink. This router for $p$ connects with the routers of the other participants in $G$ ($q_{i}\in\tilde{q}$) on channels ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q_{1}}},\ldots,{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q_{n}}}$; we shall write such channels in purple. (This convention explains the colors of the lines in Figure 6.) The router synthesis algorithm relies on relative projection to detect non- local choices; this way, the router can synchronize with the participant’s implementation and with other routers appropriately. To this end, we define the predicate ‘$\mathrm{hdep}$’, which is true for an exchange and a pair of participants if the exchange induces a dependency for either participant. Recall that relative projection produces a ‘$\mathsf{skip}$’ when an exchange is not non-local (cf. Figure 5). Thus, ‘$\mathrm{hdep}$’ only holds true if relative projection does not produce a ‘$\mathsf{skip}$’. ###### Definition 18. The predicate ‘$\mkern 2.0mu\mathrm{hdep}(q,p,G)\mkern-3.0mu$’ is true if and only if * • $G=s\mathbin{\twoheadrightarrow}r\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}$ and $q\notin\\{s,r\\}$ and $p\in\\{s,r\\}$, and * • $\mathrm{ddep}((p,q),G)\neq\mathsf{skip}\mathbin{.}R$ for all relative types $R$, where $\mathrm{ddep}$ is as in Fig. 5 (top). ###### Example 4. Consider the global type $G_{\mathsf{h}}:=p\mathbin{\twoheadrightarrow}q\\{\mathsf{a}\mathbin{.}p\mathbin{\twoheadrightarrow}r{:}\mathsf{a}\mathbin{.}\bullet,\quad\mathsf{b}\mathbin{.}r\mathbin{\twoheadrightarrow}p{:}\mathsf{b}\mathbin{.}\bullet\\}$. We have that $\mathrm{hdep}(q,p,G_{\mathsf{h}})$ is false because the initial exchange in $G_{\mathsf{h}}$ is not a dependency for $p$ and $q$, but $\mathrm{hdep}(r,p,G_{\mathsf{h}})$ is true because the initial exchange in $G_{\mathsf{h}}$ is indeed a dependency for $p$ and $r$. 1 def _${\llbracket G\rrbracket}_{p}^{\tilde{q}}$_ as 2 switch _$G$_ do 3 case _$s\mathbin{\twoheadrightarrow}r\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}$_ do 4 $\mathsf{deps}:=\\{q\in\tilde{q}\mid\mathrm{hdep}(q,p,G)\\}$ 5 6 if _$p=s$_ then return ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\mathbin{\triangleright}\big{\\{}i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}}\mathbin{\triangleleft}i\cdot{(\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}}\mathbin{\triangleleft}i)}_{q\in\mathsf{deps}}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}(v)\mathbin{.}\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}}[w]\cdot(v\mathbin{\leftrightarrow}w\mathbin{|}{\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}})\big{\\}}_{i\in I}$ 7 8 else if _$p=r$_ then return ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\mathbin{\triangleright}\big{\\{}i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i\cdot{(\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}}\mathbin{\triangleleft}i)}_{q\in\mathsf{deps}}\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}(v)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[w]\cdot(v\mathbin{\leftrightarrow}w\mathbin{|}{\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}})\big{\\}}_{i\in I}$ 9 10 else if _$p\notin\\{s,r\\}$_ then 11 $\mathsf{depon}_{s}:=(s\in\tilde{q}\wedge\mathrm{hdep}(p,s,G))$ 12 $\mathsf{depon}_{r}:=(r\in\tilde{q}\wedge\mathrm{hdep}(p,r,G))$ 13 14 if _$\mathsf{depon}_{s}$ and $\neg\mathsf{depon}_{r}$_ then return ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\mathbin{\triangleright}\big{\\{}i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i\cdot{\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\big{\\}}_{i\in I}$ 15 16 else if _$\mathsf{depon}_{r}$ and $\neg\mathsf{depon}_{s}$_ then return ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}\mathbin{\triangleright}\big{\\{}i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i\cdot{\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\big{\\}}_{i\in I}$ 17 18 else if _$\mathsf{depon}_{s}$ and $\mathsf{depon}_{r}$_ then return ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\mathbin{\triangleright}\big{\\{}i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}\mathbin{\triangleleft}\\{i{:}\leavevmode\nobreak\ {\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\\}\big{\\}}_{i\in I}$ 19 20 else return ${\llbracket G_{j}\rrbracket}_{p}^{\tilde{q}}$ for any $j\in I$ 21 22 23 24 case _$\mu X\mathbin{.}G^{\prime}$_ do 25 $\tilde{q}^{\prime}:=\\{q\in\tilde{q}\mid G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)\neq\bullet\\}$ 26 if _$\tilde{q}^{\prime}\neq\emptyset$_ then return $\mu X({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}},{({{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}})}_{q\in\tilde{q}^{\prime}})\mathbin{.}{\llbracket G^{\prime}\rrbracket}_{p}^{\tilde{q}^{\prime}}$ 27 else return $\bm{0}$ 28 29 30 case _$X$_ do return $X{\langle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}},{({{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}})}_{q\in\tilde{q}}\rangle}$ 31 32 case _$\mathsf{skip}\mathbin{.}G^{\prime}$_ do return ${\llbracket G^{\prime}\rrbracket}_{p}^{\tilde{q}}$ 33 34 case _$\bullet$_ do return $\bm{0}$ 35 Algorithm 1 Synthesis of Router Processes (Def. 19). ###### Definition 19 (Router Synthesis). Given a global type $G$, a participant $p$, and participants $\tilde{q}$, Algorithm 1 defines the synthesis of a _router process_ , denoted ‘$\mkern 1.0mu{\llbracket G\rrbracket}_{p}^{\tilde{q}}\mkern-3.0mu$’, that interfaces the interactions of $p\mkern-2.0mu$ with the other protocol participants according to $G$. We often write ‘$\mathcal{R}_{p}$’ for ‘${\llbracket G\rrbracket}_{p}^{\mathsf{prt}(G)\setminus\\{p\\}}$’ when $G$ is clear from the context. Algorithm 1 distinguishes six cases depending on the syntax of $G$ (Def. 11). The key case is ‘${s\mathbin{\twoheadrightarrow}r\\{i\langle U_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}}$’ (algorithm 1). First, the algorithm computes a set $\mathsf{deps}$ of participants that depend on the exchange using $\mathrm{hdep}$ (cf. Def. 18). Then, the algorithm considers the three possibilities for $p$: 1. 1. If $p=s$ then $p$ is the sender (algorithm 1): the algorithm returns a process that receives a label $i\in I$ over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$; sends $i$ over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}$ and over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ for every $q\in\mathsf{deps}$; receives a channel $v$ over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$; forwards $v$ as $w$ over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}$; and continues as ‘${\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}$’. 2. 2. If $p=r$ then $p$ is the recipient (algorithm 1): the algorithm returns a process that receives a label $i\in I$ over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$; sends $i$ over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ and over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ for every $q\in\mathsf{deps}$; receives a channel $v$ over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$; forwards $v$ as $w$ over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$; and continues as ‘${\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}$’. 3. 3. Otherwise, if $p$ is not involved (algorithm 1), we use ‘$\mathrm{hdep}$’ to determine whether $p$ depends on an output from $s$, an input from $r$, or on both (algorithms 1 and 1). If $p$ only depends on the output from $s$, the algorithm returns a process that receives a label $i\in I$ over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$; sends $i$ over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$; and continues as ‘${\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}$’ (algorithm 1). If $p$ only depends on an input from $r$, the returned process is similar; the only difference is that $i$ is received over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}$ (algorithm 1). When $p$ depends on _both_ the output from $s$ and on the input from $r$ (algorithm 1), the algorithm returns a process that receives a label $i\in I$ over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$; sends $i$ over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$; receives the label $i$ over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}$; and continues as ‘${\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}$’. If there are no dependencies, the returned process is ‘${\llbracket G_{j}\rrbracket}_{p}^{\tilde{q}}$’, for arbitrary $j\in I$ (algorithm 1). In case ‘$\mu X\mathbin{.}G^{\prime}$’ (algorithm 1), the algorithm stores in ‘$\tilde{q}^{\prime}$’ those $q\in\tilde{q}$ that interact with $p$ in $G^{\prime}$ (i.e. $\mu X\mathbin{.}G^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)\neq\bullet$). Then, if $\tilde{q}^{\prime}$ is non-empty (algorithm 1), the algorithm returns a recursive definition with as context the channels ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ for $q\in\tilde{q}^{\prime}$ and ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$. Otherwise, the algorithm returns ‘$\bm{0}$’ (algorithm 1). In case ‘$X$’ (algorithm 1), the algorithm returns a recursive call with as context the channels ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ for $q\in\tilde{q}$ and ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$. In case ‘$\mathsf{skip}\mathbin{.}G^{\prime}$’ (algorithm 1), it continues with ‘$G^{\prime}$’ immediately. Finally, in case ‘$\bullet$’ (algorithm 1), the algorithm returns ‘$\bm{0}$’. Considering the number of steps required to return a process, the complexity of Algorithm 1 is linear in the size of the given global type (defined as the sum of the number of communications over all branches). ### 4.2 Types for the Router’s Channels Here, we obtain session types (cf. Def. 1) for (i) the channels between routers and implementations (§ 4.2.1) and for (ii) the channels between pairs of routers (§ 4.2.2). While the former are extracted from global types, the latter are extracted from relative types. #### 4.2.1 The Channels between Routers and Implementations We begin with the session types for the channels between routers and implementations (given in pink), which we extract directly from the global type. A participant’s implementation performs on this channel precisely those actions that the participant must perform as per the global type. Hence, we define this extraction as a form of _local projection_ of the global type onto a _single participant_. The resulting session type may used as a guidance for specifying a participant implementation, which can then connect to the router’s dually typed channel endpoint. Below, $\mathsf{o}\in\mathbb{N}$ is arbitrary: $\displaystyle{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}\bullet{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}$ $\displaystyle:=\bullet$ $\displaystyle{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}{!}T\mathbin{.}S{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}$ $\displaystyle:={{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}T{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}}{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}$ $\displaystyle{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}{{\oplus}}\\{i{:}\leavevmode\nobreak\ S_{i}\\}_{i\in I}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}$ $\displaystyle:={\oplus}^{\mathsf{o}}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\\}_{i\in I}$ $\displaystyle{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}{?}T\mathbin{.}S{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}$ $\displaystyle:={{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}T{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}}{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}$ $\displaystyle{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}\&\\{i{:}\leavevmode\nobreak\ S_{i}\\}_{i\in I}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}$ $\displaystyle:=\&^{\mathsf{o}}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\\}_{i\in I}$ . If $G=s\mathbin{\twoheadrightarrow}r\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}$, $G\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}p:=\begin{cases}{{\oplus}}^{\mathsf{o}}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}+1}(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}p)\\}_{i\in I}&\text{if $p=s$}\\\ \&^{\mathsf{o}+2}\\{i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}+3}(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}p)\\}_{i\in I}&\text{if $p=r$}\\\ \&^{\mathsf{o}+2}\\{i{:}\leavevmode\nobreak\ (G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}p)\\}_{i\in I}&\text{if $p\notin\\{s,r\\}$ and $\mathrm{hdep}(p,s,G)$}\\\ \&^{\mathsf{o}+3}\\{i{:}\leavevmode\nobreak\ (G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}p)\\}_{i\in I}&\text{if $p\notin\\{s,r\\}$ and $\neg\mathrm{hdep}(p,s,G)$ and $\mathrm{hdep}(p,r,G)$}\\\ G_{i^{\prime}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}p\leavevmode\nobreak\ \text{[any $i^{\prime}\in I$]}&\text{otherwise}\end{cases}$ Otherwise, $\displaystyle\bullet\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}p:=\bullet\qquad(\mathsf{skip}\mathbin{.}G^{\prime})\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}p:=G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}p\qquad X\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}p:=X$ $\displaystyle(\mu X\mathbin{.}G^{\prime})\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}p$ $\displaystyle:=\begin{cases}\mu X\mathbin{.}(G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}p)&\text{if $G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}p$ defined and contractive on $X$}\\\ \bullet&\text{otherwise}\end{cases}$ Figure 7: Extracting Session Types from Message Types (top), and Local Projection: Extracting Session Types from a Global Type (bottom, cf. Definition 22). Global types contain message types (Def. 11), so we must first define how we extract session types from message types. This is a straightforward definition, which leaves priorities unspecified: they do not matter for the typability of routers, which forward messages between implementations and other routers. Note that one must still specify these priorities when type- checking implementations, making sure they concur between sender and recipient. ###### Definition 20 (From Message Types to Session Types). We define the extraction of a session type from message type $S$, denoted ‘$\mkern 1.0mu{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S\mkern 1.0mu{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mkern-3.0mu$’, by induction on the structure of $S$ as in Figure 7 (top). We now define local projection. To deal with non-local choices, local projection incorporates dependencies by relying on the dependency detection of relative projection (cf. Def. 16). Also similar to relative projection, local projection relies on a notion of contractiveness for session types. ###### Definition 21 (Contractive Session Types). Given a session type $A$ and a recursion variable $X$, we say _$A$ is contractive on $X$_ if either of the following holds: * • $A$ contains a connective in $\\{\mathbin{\otimes},\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}},{\oplus},\&\\}$, or * • $A$ is a recursive call on a variable other than $X$. ###### Definition 22 (Local Projection: From Global Types to Session Types). We define the local projection of global type $G$ onto participant $p$ with priority $\mathsf{o}$, denoted ‘$\mkern 1.0muG\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}p\mkern-2.0mu$’, by induction on the structure of $G$ as in Figure 7 (bottom), relying on message type extraction (Def. 20) and the predicate ‘$\mkern 1.0mu\mathrm{hdep}\mkern-3.0mu$’ (Def. 18). We consider the local projection of an exchange in a global type onto a participant $p$ with priority $\mathsf{o}$. The priorities in local projection reflect the four sub-steps into which we decompose exchanges in global types (cf. Section 4.1). There are three possibilities, depending on the involvement of $p$ in the exchange: 1. 1. If $p$ is the sender, local projection specifies a choice (${\oplus}$) between the exchange’s labels at priority $\mathsf{o}$ and an output ($\mathbin{\otimes}$) of the associated message type at priority $\mathsf{o}+1$, followed by the projection of the chosen branch at priority $\mathsf{o}+4$. 2. 2. If $p$ is the recipient, local projection specifies a branch ($\&$) on the exchange’s labels at priority $\mathsf{o}+2$ and an input ($\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}$) of the associated message type at priority $\mathsf{o}+3$, followed by the projection of the chosen branch at priority $\mathsf{o}+4$. 3. 3. If $p$ is neither sender nor recipient, local projection uses the predicate ‘$\mathrm{hdep}$’ (Def. 18) to detect a dependency on the sender’s output or the recipient’s input. If there is a dependency on the output, local projection specifies a branch on the exchange’s labels at priority $\mathsf{o}+2$. If there is a dependency on the input, local projection specifies a branch at priority $\mathsf{o}+3$. Otherwise, when there is no dependency at all, local projection simply continues with the projection of any branch at priority $\mathsf{o}+4$. Projection only preserves recursive definitions if they contain actual behavior (i.e. the projection of the recursive loop is contractive, cf. Definition 21). The projections of ‘$\bullet$’ and recursion variables are homomorphic. The projection of ‘$\mathsf{skip}$’ simply projects the skip’s continuation, at priority $\mathsf{o}+4$ to keep the priority aligned with the priorities of the other types of the router. #### 4.2.2 The Channels between Pairs of Routers $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mus\\{i\langle S_{i}\rangle\mathbin{.}R_{i}\\}_{i\in I}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}}$ $\displaystyle:=\begin{cases}{{\oplus}}^{\mathsf{o}+1}\left\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}+2}{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muR_{i}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}+4}\right\\}_{i\in I}&\text{if $p=s$}\\\\[6.0pt] \&^{\mathsf{o}+1}\left\\{i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}+2}{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muR_{i}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}+4}\right\\}_{i\in I}&\text{if $q=s$}\end{cases}$ $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mur{?}s\\{i\mathbin{.}R_{i}\\}_{i\in I}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}}$ $\displaystyle:=\begin{cases}{{\oplus}}^{\mathsf{o}+2}\left\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muR_{i}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}+4}\right\\}_{i\in I}&\text{if $p=r$}\\\\[6.0pt] \&^{\mathsf{o}+2}\left\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muR_{i}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}+4}\right\\}_{i\in I}&\text{if $q=r$}\end{cases}$ $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mus{!}r\\{i\mathbin{.}R_{i}\\}_{i\in I}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}}$ $\displaystyle:=\begin{cases}{{\oplus}}^{\mathsf{o}+1}\left\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muR_{i}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}+4}\right\\}_{i\in I}&\text{if $p=s$}\\\\[6.0pt] \&^{\mathsf{o}+1}\left\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muR_{i}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}+4}\right\\}_{i\in I}&\text{if $q=s$}\end{cases}$ $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mu\bullet{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}}:=\bullet\qquad{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mu\mathsf{skip}\mathbin{.}R{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}}:={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muR{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}+4}\qquad{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mu\mu X\mathbin{.}R{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}}:=\mu X\mathbin{.}{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muR{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}}\qquad{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muX{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}}:=X$ Figure 8: Extracting Session Types from Relative Types (cf. Definition 23). For the channels between pairs of routers (given in purple), we extract session types from relative types (Def. 13). Considering a relative type that describes the protocol between $p$ and $q$, this entails decomposing it into a type for $p$ and a dual type for $q$. ###### Definition 23 (From Relative Types to Session Types). We define the extraction of a session type from relative type $R$ between $p$ and $q$ at $p$’s perspective with priority $\mathsf{o}$, denoted ‘$\mkern 1.0mu{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muR{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}}\mkern-3.0mu$’, by induction on the structure of $R$ as in Figure 8. Here, extraction is _directional_ : in ‘${{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muR{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}}$’, the annotation ‘$p\rangle q$’ says that the session type describes the perspective of $p$’s router with respect to $q$’s. Messages with sender $p$ are decomposed into selection (${\oplus}$) at priority $\mathsf{o}+1$ followed by output ($\mathbin{\otimes}$) at priority $\mathsf{o}+2$. Dependencies on messages recieved by $p$ become selection types (${\oplus}$) at priority $\mathsf{o}+1$, and dependencies on messages sent by $p$ become selection types (${\oplus}$) at priority $\mathsf{o}+2$. Messages from $q$ and dependencies on $q$ yield dual types. Extraction from ‘$\bullet$’ and recursion is homomorphic, and extraction from ‘$\mathsf{skip}$’ simply extracts from the skip’s continuation at priority $\mathsf{o}+4$. This way, the channel endpoint of $p$’s router that connects to $q$’s router will be typed ‘${{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}}$’, i.e. the session type extracted from the relative projection of $G$ onto $p,q$ at $p$’s perspective. Similarly, the endpoint of this channel at $q$’s router will have the type ‘${{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{q{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}p}^{\mathsf{o}}$’, i.e. the same relative projection but at $q$’s perspective. Clearly, these session types must be dual. ###### Theorem 9. Given a relative well-formed global type $G$ and $p,q\in\mathsf{prt}(G)$, ${{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}}=\overline{{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{q{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}p}^{\mathsf{o}}}.$ ###### Proof. By construction from Definition 16 and Definition 23. ∎ ### 4.3 Networks of Routed Implementations Having defined routers and types for their channels, we now turn to defining _networks of routed implementations_ , i.e., process networks of routers and implementations that correctly represent a given multiparty protocol. Then, we appeal to the types obtained in § 4.2 to establish the typability of routers (Theorem 11). Finally, we show that all networks of routed implementations of well-formed global types are deadlock free (Theorem 18), and that networks of routed implementations behave as depicted by the global types from which they are generated (Theorems 19 and 23). We begin by defining routed implementations, which connect implementations of subsets of protocol participants with routers: ###### Definition 24 (Routed Implementations). Given a closed, relative well-formed global type $G$, for participants $\tilde{p}\subseteq\mathsf{prt}(G)$, the _set of routed implementations_ of $\tilde{p}$ in $G$ is defined as follows (cf. Def. 22 for local projection ‘$\mkern 1.0mu\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}\mkern-3.0mu$’ and Def. 19 for router synthesis ‘$\mkern 1.0mu{\llbracket\ldots\rrbracket}\mkern-3.0mu$’): $\mathrm{ri}(G,\tilde{p}):=\left\\{(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})_{p\in\tilde{p}}\,(Q\mathbin{|}{\mathchoice{\textstyle}{}{}{}\prod}_{p\in\tilde{p}}\mathcal{R}_{p})\mathrel{}\middle|\mathrel{}\begin{array}[]{l}Q\vdash\emptyset;\Gamma,{({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ G\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}p)}_{p\in\tilde{p}}\\\ {}\wedge\forall p\in\tilde{p}.\leavevmode\nobreak\ \mathcal{R}_{p}={\llbracket G\rrbracket}_{p}^{\mathsf{prt}(G)\setminus\\{p\\}}\end{array}\right\\}$ We write $\mathcal{N}_{\tilde{p}},\mathcal{N}^{\prime}_{\tilde{p}},\ldots$ to denote elements of $\mathrm{ri}(G,\tilde{p})$. Thus, the composition of a collection of routers and an implementation $Q$ is a routed implementation as long as $Q$ can be typed in a context that includes the corresponding projected types. Note that the parameter $\tilde{p}$ indicates the presence of _interleaving_ : when $\tilde{p}$ is a singleton, the set $\mathrm{ri}(G,\tilde{p})$ contains processes in which there is a single router and the implementation $Q$ is single-threaded (non-interleaved); more interestingly, when $\tilde{p}$ includes two or more participants, the set $\mathrm{ri}(G,\tilde{p})$ consists of processes in which the implementation $Q$ interleaves the roles of the multiple participants in $\tilde{p}$. A network of routed implementations of a global type, or simply a _network_ , is then the composition of any combination of routed implementations that together account for all the protocol’s participants. Hence, we define sets of networks, quantified over all possible combinations of sets of participants and their respective routed implementations. The definition relies on _complete partitions_ of the participants of a global type, i.e., a split of $\mathsf{prt}(G)$ into non-empty, disjoint subsets whose union yields $\mathsf{prt}(G)$. ###### Definition 25 (Networks). Suppose given a closed, relative well-formed global type $G$. Let $\mathbb{P}_{G}$ be the set of all complete partitions of $\mathsf{prt}(G)$ with elements $\pi,\pi^{\prime},\ldots$. The _set of networks_ of $G$ is defined as $\mathrm{net}(G):=\big{\\{}(\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}})_{p,q\in\mathsf{prt}(G)}({\mathchoice{\textstyle}{}{}{}\prod}_{\tilde{p}\in\pi}\mathcal{N}_{\tilde{p}})\leavevmode\nobreak\ \big{|}\leavevmode\nobreak\ \pi\in\mathbb{P}_{G}\wedge\forall\tilde{p}\in\pi.\leavevmode\nobreak\ \mathcal{N}_{\tilde{p}}\in\mathrm{ri}(G,\tilde{p})\big{\\}}.$ We write $\mathcal{N},\mathcal{N}^{\prime},\ldots$ to denote elements of $\mathrm{net}(G)$. ###### Example 5. Figure 6 depicts two networks in $\mathrm{net}(G_{\mathsf{auth}})$ related to different partitions of $\mathsf{prt}(G_{\mathsf{auth}})$, namely $\big{\\{}\\{a\\},\\{s\\},\\{c\\}\big{\\}}$ (non-interleaved) on the left and $\big{\\{}\\{a,s\\},\\{c\\}\big{\\}}$ (interleaved) on the right. Because a network $\mathcal{N}$ may not be typable under the empty typing context, we have the following definition to “complete” networks. ###### Definition 26 (Completable Networks). Suppose given a network $\mathcal{N}$ such that $\mathcal{N}\vdash\emptyset;\Gamma$. We say that $\mathcal{N}$ is _completable_ if (i) $\Gamma$ is empty or (ii) there exist $\tilde{v},\tilde{w}$ such that $(\bm{\nu}\tilde{v}\tilde{w})\mathcal{N}\vdash\emptyset;\emptyset$. When $\mathcal{N}$ is completable, we write ‘$\mkern 1.0mu\mathcal{N}^{\circlearrowright}\mkern-3.0mu$’ to stand for $\mathcal{N}$ (if $\mathcal{N}\vdash\emptyset;\emptyset$) or $(\bm{\nu}\tilde{v}\tilde{w})\mathcal{N}$ (otherwise). ###### Proposition 10. For any closed, relative well-formed global type $G$, there exists at least one completable network $\mathcal{N}\in\mathrm{net}(G)$. ###### Proof. To construct a completable network in $\mathrm{net}(G)$, we construct a routed implementation (Def. 24) for every $p\in\mathsf{prt}(G)$. Given a $p\in\mathsf{prt}(G)$, by Proposition 1, there exists $Q\vdash\emptyset;{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ G\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}p$. Composing each such characteristic implementation process with routers, and then composing the routed implementations, we obtain a network $\mathcal{N}\in\mathrm{net}(G)$, where $\mathcal{N}\vdash\emptyset;\emptyset$. Hence, $\mathcal{N}$ is completable. ∎ $P$implementation${\llbracket G_{\mathsf{auth}}\rrbracket}_{c}^{\\{s,a\\}}$router${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{:}\leavevmode\nobreak\ G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}c$${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{:}\leavevmode\nobreak\ \overline{G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}c}$${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(c,s){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{c{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}s}^{\mathsf{o}}$${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(c,a){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{c{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}a}^{\mathsf{o}}$Definition 22Definition 19Definition 16Definition 23 Figure 9: Overview of Theorem 11, with the definitions and notations for synthesizing and typing routers, using participant $c$ of $G_{\mathsf{auth}}$ implemented as $P$ (cf. Example 1). Lines indicate channels and boxes indicate processes. #### 4.3.1 The Typability of Routers We wish to establish that the networks of a global type are deadlock free. This result, formalized by Theorem 18 (Theorem 18), hinges on the typability of routers, which we address next. Figure 9 gives an overview of the definitions and notations involved in this theorem’s statement. ###### Theorem 11. Suppose given a closed, relative well-formed global type $G$, and a $p\in\mathsf{prt}(G)$. Then, ${\llbracket G\rrbracket}_{p}^{\mathsf{prt}(G)\setminus\\{p\\}}\vdash\emptyset;\leavevmode\nobreak\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{G\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}p},\leavevmode\nobreak\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{0}\big{)}}_{q\in\mathsf{prt}(G)\setminus\\{p\\}}.$ This result is a corollary of Theorem 16 (Theorem 16), which we show next. We give a full proof on Section 4.3.1, after the proof of Theorem 16. ##### Alarm Processes We focus on networks of routed implementations—compositions of synthesized routers and well-typed processes. However, in order to establish the typability of routers we must account for an edge case that goes beyond these assumptions, namely when a routed implementation is connected to some undesirable implementation, not synthesized by Algorithm 1. Consider the following example: ###### Example 6. Consider again the global type $G_{\mathsf{auth}}$, which, for the purpose of this example, we write as follows: $\displaystyle G_{\mathsf{auth}}=s\mathbin{\twoheadrightarrow}c\left\\{\begin{array}[]{@{}l@{}}\mathsf{login}{:}\leavevmode\nobreak\ G_{\mathsf{login}},\\\ \mathsf{quit}{:}\leavevmode\nobreak\ G_{\mathsf{quit}}\end{array}\right\\}$ As established in Example 2, the initial exchange between $s$ and $c$ determines a dependency for the interactions of $a$ with both $s$ and $c$. Therefore, the implementation of $a$ needs to receive the choice between login and quit from the implementations of both $s$ and $c$. An undesirable implementation for $c$, without a router, could be for instance as follows: $\displaystyle R^{\prime}:={{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\mathbin{\triangleright}\left\\{\begin{array}[]{@{}l@{}}\mathsf{login}{:}\leavevmode\nobreak\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}\mathbin{\triangleleft}\mathsf{quit}\cdot\ldots,\\\ \mathsf{quit}{:}\leavevmode\nobreak\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}\mathbin{\triangleleft}\mathsf{quit}\cdot\ldots\end{array}\right\\}$ Notice how $R^{\prime}$ always sends to $a$ the label quit, even if the choice made by $s$ (and sent to $c$) is login. Now, if $s$ chooses login, the router of $a$ is in limbo: on the one hand, it expects $s$ to behave as specified in $G_{\mathsf{login}}$; on the other hand, it expects $c$ to behave as specified in $G_{\mathsf{quit}}$. Clearly, the router of $a$ is in an inconsistent state due to $c$’s implementation. Because routers always forward the chosen label correctly, this kind of undesirable behavior never occurs in the networks of Definition 25—we state this formally in § 4.3.2 (Theorem 17). Still, in order to prove that our routers are well-typed, we must accommodate the possibility that a router ends up in an undesirable state due to inconsistent forwarding. For this, we extend APCP with an _alarm process_ that signals an inconsistency on a given set of channel endpoints. ###### Definition 27 (Alarm Process). Given channel endpoints $\tilde{x}=x_{1},\ldots,x_{n}$, we write ‘$\mkern 2.0mu{\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({\tilde{x}})}\mkern-3.0mu$’ to denote an inconsistent state on those endpoints. In a way, ${\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({\tilde{x}})}$ is closer to an observable action (a “barb”) than to an actual process term: ${\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({\tilde{x}})}$ does not have reductions, and no process from Figure 3 (top) can reduce to ${\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({\tilde{x}})}$. We assume that ${\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({\tilde{x}})}$ does not occur in participant implementations (cf. $Q$ in Definition 24); we treat it as a process solely for the purpose of refining the router synthesis algorithm (Algorithm 1) with the possibility of inconsistent forwarding. The refinement concerns the process on algorithm 1: $\displaystyle{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\mathbin{\triangleright}\big{\\{}i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}\mathbin{\triangleleft}\\{i{:}\leavevmode\nobreak\ {\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\\}\big{\\}}_{i\in I}$ We extend it with additional branches, as follows: $\displaystyle{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\mathbin{\triangleright}\big{\\{}i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}\mathbin{\triangleleft}\left(\begin{array}[]{@{}l@{}}\\{i{:}\leavevmode\nobreak\ {\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\\}\\\\[5.0pt] \leavevmode\nobreak\ \underline{\cup\leavevmode\nobreak\ \\{i^{\prime}{:}\leavevmode\nobreak\ {\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}},{({{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}})}_{q\in\tilde{q}}})}\\}_{i^{\prime}\in I\setminus\\{i\\}}}\nobreak\leavevmode\nobreak\leavevmode\end{array}\right)\big{\\}}_{i\in I}$ (4) This new process for algorithm 1 captures the kind of inconsistency illustrated by Example 6, which occurs when a label $i\in I$ is received over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$ after which a label $i^{\prime}\in I\setminus\\{i\\}$ is received over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}$. We account for this case by using the underlined alarm processes. Routers are then made of processes as in Figure 3 (top), selectively extended with alarms as just described. Because ${\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({\tilde{x}})}$ merely acts as an observable that signals undesirable behavior, we find it convenient to type it using the following axiom: Alarm ${\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({x_{1},\ldots,x_{n}})}\vdash\Omega;x_{1}{:}\leavevmode\nobreak\ A_{1},\ldots,x_{n}{:}\leavevmode\nobreak\ A_{n}$ where the recursive context $\Omega$ and types $A_{1},\ldots,A_{n}$ are arbitrary. ##### Context-based Typability Considering the refinement of Algorithm 1 with alarm processes, we prove Theorem 16 on Theorem 16, from which Theorem 11 follows as a corollary. It relies on some additional auxiliary definitions and results. To type the router for a participant at any point in the protocol, we need the definition of the entire protocol. It is not enough to only consider the current (partial) protocol at such points: we need information about bound recursion variables in order to perform unfolding in types. To this end, we define _global contexts_ , that allow us to look at part of a protocol while retaining definitions that concern the entire protocol. ###### Definition 28 (Global Contexts). _Global contexts_ $\mathcal{C}$ are given by the following grammar: $\displaystyle C::=p\mathbin{\twoheadrightarrow}q\left(\begin{array}[]{@{}l@{}}\\{i\langle S\rangle\mathbin{.}G\\}_{i\in I}\\\ {}\cup\\{i^{\prime}\langle S\rangle\mathbin{.}C\\}_{i^{\prime}\notin I}\end{array}\right)\;\mbox{\large{$\mid$}}\;\mathsf{skip}\mathbin{.}C\;\mbox{\large{$\mid$}}\;\mu X\mathbin{.}C\;\mbox{\large{$\mid$}}\;[]$ We often simply write ‘context’ when it is clear that we are referring to a global context. Given a context $C$ and a global type $G$, we write ‘$\mkern 1.0muC[G]\mkern-3.0mu$’ to denote the global type obtained by replacing the hole ‘$\mkern 1.0mu[]\mkern-3.0mu$’ in $C$ with $G$. If $G=C[G_{s}]$ for some context $C$ and global type $G_{s}$, then we write ‘$\mkern 1.0muG_{s}\leq_{C}G\mkern-3.0mu$’. As mentioned before, a context captures information about the recursion variables that are bound at any given point in a global type. Our goal is to obtain a _context-based_ typability result for routers. The order in which recursive variables are bound is important to correctly unfold types: ###### Example 7. Consider the following global type with three nested recursive definitions: $\displaystyle G_{\mathsf{rec}}=\mu X\mathbin{.}a\mathbin{\twoheadrightarrow}b:1\mathbin{.}\mu Y\mathbin{.}a\mathbin{\twoheadrightarrow}b:2\mathbin{.}\mu Z\mathbin{.}a\mathbin{\twoheadrightarrow}b\\{\mathsf{x}:X,\quad\mathsf{y}:Y,\quad\mathsf{z}:Z\\}$ To type the router for, e.g., $a$ at the final exchange between $a$ and $b$, we need to be aware of the unfolding of recursion. The recursion on $X$, $Y$, and $Z$ have all to be unfolded, and the recursion on $Z$ must include first the unfolding of $X$ and then the unfolding of $Y$, which must in turn include the prior unfolding of $X$. To account for nested recursions, the following definition gives the bound variables of a context exactly in the order in which they appear: ###### Definition 29 (Recursion Binders of Contexts). Given a global context $C$, the _sequence of recursion binders to the hole_ of $C$, denoted ‘$\mkern 1.0mu\mathrm{ctxbind}(C)\mkern-3.0mu$’, is defined as follows: $\displaystyle\mathrm{ctxbind}(\mu X\mathbin{.}C):=(X,\mathrm{ctxbind}(C))\qquad\mathrm{ctxbind}(\mathsf{skip}\mathbin{.}C):=\mathrm{ctxbind}(C)\qquad\mathrm{ctxbind}([]):=()$ $\displaystyle\mathrm{ctxbind}(p\mathbin{\twoheadrightarrow}q\left(\begin{array}[]{@{}l@{}}\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}\\\ {}\cup\\{i^{\prime}\langle S_{i^{\prime}}\rangle\mathbin{.}C\\}_{i^{\prime}\notin I}\end{array}\right))$ $\displaystyle:=\mathrm{ctxbind}(C)$ Given $G_{s}\leq_{C}G$, the sequence of recursion binders of $G_{s}$, denoted ‘$\mkern 1.0mu\mathrm{subbind}(G_{s},G)\mkern-3.0mu$’, is defined as $\mathrm{ctxbind}(C)$. The following retrieves the body of a recursive definition from a global context, informing us on how to unfold types: ###### Definition 30 (Recursion Extraction). The function ‘$\mkern 1.0mu\mathrm{recdef}(X,G)\mkern-3.0mu$’ extracts the recursive definition on $X$ from $G$, i.e. $\mathrm{recdef}(X,G)=G^{\prime}$ if $\mu X\mathbin{.}G^{\prime}\leq_{C}G$ for some context $C$. Also, ‘$\mkern 1.0mu\mathrm{recCtx}(X,G)\mkern-3.0mu$’ extracts the context of the recursive definition on $X$ in $G$, i.e. $\mathrm{recCtx}(X,G)=C$ if $\mu X\mathbin{.}\mathrm{recdef}(X,G)\leq_{C}G$. When unfolding bound recursion variables, we need the priorities of the unfolded types. The following definition gives a priority that is expected at the hole in a context, as well as the priority expected at any recursive definition in a global type: ###### Definition 31 (Absolute Priorities of Contexts). Given a context $C$ and $\mathsf{o}\in\mathbb{N}$, we define $\mathrm{ctxpri}^{\mathsf{o}}(C)$ as follows: $\displaystyle\mathrm{ctxpri}^{\mathsf{o}}([]):=\mathsf{o}\qquad\mathrm{ctxpri}^{\mathsf{o}}(\mathsf{skip}\mathbin{.}C):=\mathrm{ctxpri}^{\mathsf{o}+4}(C)\qquad\mathrm{ctxpri}^{\mathsf{o}}(\mu X\mathbin{.}C):=\mathrm{ctxpri}^{\mathsf{o}}(C)$ $\displaystyle\mathrm{ctxpri}^{\mathsf{o}}(p\mathbin{\twoheadrightarrow}q\left(\begin{array}[]{@{}l@{}}\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}\\\ {}\cup\\{i^{\prime}\langle S_{i^{\prime}}\mathbin{.}C\\}_{i^{\prime}\notin I}\end{array}\right))$ $\displaystyle:=\mathrm{ctxpri}^{\mathsf{o}+4}(C)$ Then, the _absolute priority_ of $C$, denoted ‘$\mkern 1.0mu\mathrm{ctxpri}(C)\mkern-3.0mu$’, is defined as $\mathrm{ctxpri}^{0}(C)$. The absolute priority of $X$ in $G$, denoted ‘$\mkern 1.0mu\mathrm{varpri}(X,G)\mkern-3.0mu$’, is defined as $\mathrm{ctxpri}(C)$ for some context $C$ such that $\mu X\mathbin{.}\mathrm{recdef}(X,G)\leq_{C}G$. To avoid non-contractive recursive types, relative projection (cf. Figure 5) closes a type when the participants do not interact inside a recursive definition. Hence, when typing a router for a recursive definition, we must determine which pairs of participants are “active” at any given point in a protocol, and close the connections with the “inactive” participants. ###### Example 8. Consider the following global type, where a client (‘$c$’) requests two independent, infinite Fibonacci sequences (‘$f_{1}$’ and ‘$f_{2}$’): $\displaystyle G_{\mathsf{fib}}=c\mathbin{\twoheadrightarrow}f_{1}:\mathsf{init}\langle\mathsf{int}\times\mathsf{int}\rangle\mathbin{.}c\mathbin{\twoheadrightarrow}f_{2}:\mathsf{init}\langle\mathsf{int}\times\mathsf{int}\rangle\mathbin{.}\underbrace{\mu X\mathbin{.}f_{1}\mathbin{\twoheadrightarrow}c:\mathsf{next}\langle\mathsf{int}\rangle\mathbin{.}f_{2}\mathbin{\twoheadrightarrow}c:\mathsf{next}\langle\mathsf{int}\rangle\mathbin{.}X}_{G^{\prime}_{\mathsf{fib}}}$ Participants $f_{1}$ and $f_{2}$ do not interact with each other in the body of the recursion, as formalized by their relative projection: $\displaystyle\mathrm{recdef}(X,G_{\mathsf{fib}})\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(f_{1},f_{2})=\mathsf{skip}\mathbin{.}\mathsf{skip}\mathbin{.}X$ Hence, $G^{\prime}_{\mathsf{fib}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(f_{1},f_{2})=\bullet$, and $f_{1}$ and $f_{2}$ do not form an active pair of participants for the recursion in $G_{\mathsf{fib}}$. Therefore, $f_{1}$’s router closes its connection with $f_{2}$’s router at the start of the recursion on $X$, and vice versa. The following definition uses relative projection to determine the pairs of active participants at the hole of a context, as well as at any recursive definition in a global type. We consider pairs of participants $(p,q)$ and $(q,p)$ to be equivalent. ###### Definition 32 (Active Participants). Suppose given a relative well-formed global type $G$. The following mutually defined functions compute sets of _pairs of active participants_ for recursive definitions and contexts, denoted ‘$\mkern 1.0mu\mathrm{recactive}(X,G)\mkern-3.0mu$’ and ‘$\mkern 1.0mu\mathrm{active}(C,G)\mkern-3.0mu$’, respectively. $\displaystyle\mathrm{recactive}(X,G)$ $\displaystyle:=\\{(p,q)\in\mathrm{active}(\mathrm{recCtx}(X,G),G)\mid(\mu X\mathbin{.}\mathrm{recdef}(X,G))\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)\neq\bullet\\}$ $\displaystyle\mathrm{active}(C,G)$ $\displaystyle:=\begin{cases}\mathrm{recactive}(Y,G)&\text{if $\mathrm{ctxbind}(C)=(\tilde{X},Y)$}\\\ \mathsf{prt}(G)^{2}&\text{otherwise}\end{cases}$ The interdependency between ‘$\mkern 1.0mu\mathrm{recactive}(X,G)\mkern-3.0mu$’ and ‘$\mkern 1.0mu\mathrm{active}(C,G)\mkern-3.0mu$’ is well-defined: the former function considers the active participants of a context, which contains less recursive definitions. When typing a router for a given protocol, we have to keep track of assignments in the recursive context at any point in the protocol. The following two lemmas ensure that the active participants of recursive definitions are consistent with the active participants of their bodies. ###### Lemma 12. Suppose given a closed, relative well-formed global type $G$, and a global type $G_{s}$ and context $C$ such that $G_{s}\leq_{C}G$. For any $Z\in\mathrm{ctxbind}(C)$, $\mathrm{active}(C,G)\subseteq\mathrm{recactive}(Z,G)$. ###### Proof. Take any $Z\in\mathrm{ctxbind}(C)$. Then $\mathrm{ctxbind}(C)=(\tilde{X},Y)$. By definition, ${\mathrm{active}(C,G)=\mathrm{recactive}(Y,G)}$. If $Y=Z$, the thesis is proven. Otherwise, by definition, ${\mathrm{recactive}(Y,G)\subseteq\mathrm{active}(\mathrm{recCtx}(Y,G),G)}$. Since the recursive definition on $Z$ appears in $\mathrm{recCtx}(Y,G)$, it follows by induction on the size of $\tilde{X}$ that $\mathrm{active}(\mathrm{recCtx}(Y,G),G)\subseteq\mathrm{recactive}(Z,G)$. This proves the thesis. ∎ The following lemma ensures that when typing a recursive call, the endpoints given as context for the recursive call concur with the endpoints in the recursive context: ###### Lemma 13. Suppose given a closed, relative well-formed global type $G$, a recursion variable $Z$, and a context $C$ such that $Z\leq_{C}G$. Then, $\mathrm{active}(C,G)=\mathrm{recactive}(Z,G)$. ###### Proof. Because $G=C[Z]$ and $G$ is closed (i.e. $\mathrm{frv}(G)=\emptyset$), there is a recursive definition on $Z$ in $G$. Hence, $\mathrm{ctxbind}(C)\neq\emptyset$, i.e. $\mathrm{ctxbind}(C)=(\tilde{X},Y)$ and $\mathrm{active}(C,G)=\mathrm{recactive}(Y,G)$. If $Y=Z$, the thesis is proven. Otherwise, the recursive definition on $Y$ in $G$ appears somewhere inside the recursive definition on $Z$. Suppose, for contradiction, that $\mathrm{active}(C,G)\neq\mathrm{recactive}(Z,G)$. There are two cases: there exists $(p,q)\in{\mathsf{prt}(G)}^{2}$ s.t. (i) $(p,q)\in\mathrm{active}(C,G)$ and $(p,q)\notin\mathrm{recactive}(Z,G)$, or (ii) $(p,q)\in\mathrm{recactive}(Z,G)$ and $(p,q)\notin\mathrm{active}(C,G)$. Case (i) contradicts Lemma 12. In case (ii), $(\mu Z\mathbin{.}\mathrm{recdef}(Z,G))\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)\neq\bullet$ and $(\mu Y\mathbin{.}\mathrm{recdef}(Y,G))\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=\bullet$. The recursive call on $Z$ in $G$ appears somewhere inside the recursive definition on $Y$, and hence $\mathrm{recdef}(Y,G)\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ contains the recursive call on $Z$. This means that $\mathrm{recdef}(Y,G)\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ is contractive on $Y$ (Def. 15), and hence $(\mu Y\mathbin{.}\mathrm{recdef}(Y,G))\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)\neq\bullet$, contradicting the assumption. ∎ Our typability result for routers relies on relative and local projection. Hence, we need to guarantee that all the projections we need at any given point of a protocol are defined. The following result shows a form of compositionality for relative and local projection, guaranteeing the definedness of projections for all active participants of a given context: ###### Proposition 14. Suppose given a closed, relative well-formed global type $G$, and a global type $G_{s}$ such that $G_{s}\leq_{C}G$. Then, for every $(p,q)\in\mathrm{active}(C,G)$, the relative projection $G_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ is defined. Also, for every $p\in\\{p\in\mathsf{prt}(G)\mid\exists q\in\mathsf{prt}(G).\leavevmode\nobreak\ (p,q)\in\mathrm{active}(C,G)\\}$, the local projection $G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}p$ is defined for any priority $\mathsf{o}$. ###### Proof. Suppose that, for contradiction, $G_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ is undefined. We show by induction on the structure of $C$ that this means that $G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ is undefined, contradicting the relative well-formedness of $G$. * • Hole: $C=[]$. We have $G_{s}=G$, and the thesis follows immediately. * • Exchange: $C=r\mathbin{\twoheadrightarrow}s\left(\begin{array}[]{@{}l@{}}\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}\\\ {}\cup\\{i^{\prime}\langle S_{i^{\prime}}\rangle\mathbin{.}C^{\prime}\\}_{i^{\prime}\notin I}\end{array}\right)$. By the IH, $C^{\prime}[G_{s}]\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ is undefined. Since the relative projection of an exchange relies on the relative projection of each of the exchange’s branches, $G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ is undefined. * • Skip: $C=\mathsf{skip}\mathbin{.}C^{\prime}$. By the IH, $C^{\prime}[G_{s}]\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ is undefined. Since the relative projection of a skip relies on the relative projection of the skip’s continuation, $G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ is undefined. * • Recursive definition: $C=\mu X\mathbin{.}C^{\prime}$. It follows from Lemma 12 that $\mathrm{active}(C,G)\subseteq\mathrm{recactive}(X,G)$. Hence, $(p,q)\in\mathrm{recactive}(X,G)$, and thus $(\mu X\mathbin{.}\mathrm{recdef}(X,G))\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=(\mu X\mathbin{.}C^{\prime}[G_{s}])\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)\neq\bullet$, which means that $C^{\prime}[G_{s}]\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ is defined. This contradicts the IH. The proof for the definedness of local projection is analogous. ∎ Recall Example 7, where nested recursive definitions in a protocol require nested unfolding of recursive types. The following definition gives us a concise way of writing such nested (or _deep_) unfoldings: ###### Definition 33 (Deep Unfolding). Suppose given a sequence of tuples $\tilde{U}$, with each tuple consisting of a recursion variable $X_{i}$, a lift $t_{i}\in\mathbb{N}$, and a type $B_{i}$. The _deep unfolding_ of the type $A$ with $\tilde{U}$, denoted ‘$\mkern 1.0mu\mathrm{deepUnfold}(A,\tilde{U})\mkern-3.0mu$’, is the type defined as follows: $\displaystyle\mathrm{deepUnfold}(A,())$ $\displaystyle:=A$ $\displaystyle\mathrm{deepUnfold}(A,(\tilde{U},(X,t,B)))$ $\displaystyle:=\mathrm{deepUnfold}(A,\tilde{U})\\{\big{(}\mu X\mathbin{.}({\uparrow^{t}}\mathrm{deepUnfold}(B,\tilde{U}))\big{)}/X\\}$ When typing a router’s recursive call, the types of the router’s endpoints are unfoldings of the types in the recursive context. However, because of the deep unfolding in types, this is far from obvious. The following result connects a particular form of deep unfolding with regular unfolding (cf. Definition 5). ###### Proposition 15. Suppose given a type $A$ and a sequence of tuples $\tilde{U}$ consisting of a recursion variable, a lift, and a substitution type. Then, $\displaystyle\mathrm{deepUnfold}(A,(\tilde{U},(X,t,A)))$ $\displaystyle=\mathrm{unfold}^{t}(\mu X\mathbin{.}\mathrm{deepUnfold}(A,\tilde{U})).$ ###### Proof. By Definition 33: $\displaystyle\mathrm{deepUnfold}(A,(\tilde{U},(X,t,A)))$ $\displaystyle=\mathrm{deepUnfold}(A,\tilde{U})\\{\big{(}\mu X\mathbin{.}({\uparrow^{t}}\mathrm{deepUnfold}(A,\tilde{U}))\big{)}/X\\}$ $\displaystyle=\mathrm{unfold}^{t}(\mu X\mathbin{.}\mathrm{deepUnfold}(A,\tilde{U}))\qed$ Armed with these definitions and results, we can finally state our context- based typability result for routers: ###### Theorem 16. Suppose given a closed, relative well-formed global type $G$. Also, suppose given a global type $G_{s}$ such that $G_{s}\leq_{C}G$, and a $p\in\mathsf{prt}(G)$ for which there is a $q\in\mathsf{prt}(G)$ such that $(p,q)\in\mathrm{active}(C,G)$. Consider: * • the participants with whom $p$ interacts in $G_{s}$: $\tilde{q}=\\{q\in\mathsf{prt}(G)\mid(p,q)\in\mathrm{active}(C,G)\\}$, * • the absolute priority of $G_{s}$: $\mathsf{o}_{C}=\mathrm{ctxpri}(C)$, * • the sequence of bound recursion variables of $G_{s}$: $\widetilde{X_{C}}=\mathrm{ctxbind}(C)$, * • for every $X\in\widetilde{X_{C}}$: * – the body of the recursive definition on $X$ in $G$: $G_{X}=\mathrm{recdef}(X,G)$, * – the participants with whom $p$ interacts in $G_{X}$: $\tilde{q}_{X}=\\{q\in\mathsf{prt}(G)\mid(p,q)\in\mathrm{recactive}(X,G)\\}$, * – the absolute priority of $G_{X}$: $\mathsf{o}_{X}=\mathrm{varpri}(X,G)$, * – the sequence of bound recursion variables of $G_{X}$ excluding $X$: $\widetilde{Y_{X}}=\mathrm{subbind}(\mu X\mathbin{.}G_{X},G)$, * – the type required for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ for a recursive call on $X$: $\displaystyle A_{X,p}=\mathrm{deepUnfold}(\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p},{(Y,t_{Y},\overline{G_{Y}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Y}}p})}_{Y\in\widetilde{Y_{X}}}),$ * – the type required for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ for a recursive call on $X$: $\displaystyle B_{X,q}=\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}},{(Y,t_{Y},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{Y}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{Y}})}_{Y\in\widetilde{Y_{X}}}),$ * – the minimum lift for typing a recursive definition on $X$: $t_{X}=\max_{\mathsf{pr}}\left(A_{X},{(B_{X,q})}_{q\in\tilde{q}_{X}}\right)+1$, * • the type expected for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ for $p$’s router for $G_{s}$: $\displaystyle D_{p}=\mathrm{deepUnfold}(\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{X_{C}}}),$ * • the type expected for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ for $p$’s router for $G_{s}$: $\displaystyle E_{q}=\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},{(X,t_{X},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}})}_{X\in\widetilde{X_{C}}}).$ Then, we have: $\displaystyle{\llbracket G_{s}\rrbracket}_{p}^{\tilde{q}}\vdash{\Big{(}X{:}\leavevmode\nobreak\ \big{(}A_{X},{(B_{X,q})}_{q\in\tilde{q}_{X}}\big{)}\Big{)}}_{X\in\widetilde{X_{C}}};\leavevmode\nobreak\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ D_{p},\leavevmode\nobreak\ {({{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ E_{q})}_{q\in\tilde{q}}$ ###### Proof. We apply induction on the structure of $G_{s}$, with six cases as in Algorithm 1. We only detail the cases of exchange and recursion. Axiom Alarm is used in only one sub-case (case 3(c), cf. Figure 11 below). * • _Exchange_ : $G_{s}=s\mathbin{\twoheadrightarrow}r\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}$ (algorithm 1). In this case, we add connectives to the types obtained from the IH. Since we do not introduce any recursion variables to these types, the substitutions in the types from the IH are not affected. Hence, we can omit these substitutions from the types. Also, for each $i\in I$, we have $\mathrm{frv}(G_{i})\subseteq\mathrm{frv}(G_{s})$, i.e. the recursive context remains untouched in this derivation, so we also omit the recursive context. Let $\mathsf{deps}:=\\{q\in\tilde{q}\mid\mathrm{hdep}(q,p,G_{s})\\}$ (as on algorithm 1). There are three cases depending on the involvement of $p$. 1. 1. If $p=s$, then $p$ is the sender (algorithm 1). Let us consider the relative projections onto $p$ and the participants in $\tilde{q}$. For the recipient $r$, $G_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r)=p\\{i\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r))\\}_{i\in I}.$ (5) For each $q\in\mathsf{deps}$, by Definition 18, $\mathrm{ddep}((q,p),G)\neq\mathsf{skip}\mathbin{.}R$ for some $R$. That is, since $p$ is the sender of the exchange, for each $q\in\mathsf{deps}$, by the definitions in Figure 5, $G_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=p{!}r\\{i\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))\\}_{i\in I}.$ (6) On the other hand, for each $q\in\tilde{q}\setminus\mathsf{deps}\setminus\\{r\\}$, $G_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=\mathsf{skip}\mathbin{.}(G_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))$ (7) for any $i^{\prime}\in I$, because for each $i,j\in I$, $G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=G_{j}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q).$ (8) Let us take stock of the types we expect for each of the router’s channels. For ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ we expect $\displaystyle\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p}$ $\displaystyle=\overline{{{\oplus}}^{\mathsf{o}_{C}}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}_{C}+1}(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p)\\}_{i\in I}}$ $\displaystyle=\&^{\mathsf{o}_{C}}\\{i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}_{C}+1}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p)}\\}_{i\in I}.$ (9) For ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}$ we expect $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}}$ $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mup\\{i\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r))\\}_{i\in I}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}}$ $\displaystyle={{\oplus}}^{\mathsf{o}_{C}+1}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}_{C}+2}{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}+4}\\}_{i\in I}.$ (10) For each $q\in\mathsf{deps}$, for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ we expect $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}}$ $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mup{!}r\\{i.\leavevmode\nobreak\ (G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))\\}_{i\in I}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}}$ $\displaystyle={{\oplus}}^{\mathsf{o}_{C}+1}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\\}_{i\in I}.$ (11) For each $q\in\tilde{q}\setminus\mathsf{deps}\setminus\\{r\\}$, for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ we expect $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}}$ $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mu\mathsf{skip}\mathbin{.}(G_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}}$ $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\leavevmode\nobreak\ \text{for any $i^{\prime}\in I$.}$ (12) Let us now consider the process returned by Algorithm 1, with each prefix marked with a number: ${\llbracket G_{s}\rrbracket}_{p}^{\tilde{q}}=\underbrace{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\mathbin{\triangleright}\big{\\{}i{:}\big{.}}_{1}\leavevmode\nobreak\ \underbrace{\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}}\mathbin{\triangleleft}i}_{2_{i}}\cdot\underbrace{{(\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}}\mathbin{\triangleleft}i)}_{q\in\mathsf{deps}}}_{3_{i}}\cdot\underbrace{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}(v)}_{4_{i}}\mathbin{.}\underbrace{\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}}[w]}_{5_{i}}\cdot(v\mathbin{\leftrightarrow}w\mathbin{|}{\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}})\big{.}\big{\\}}_{i\in I}$ For each $i^{\prime}\in I$, let $C_{i^{\prime}}:=C[s\mathbin{\twoheadrightarrow}r(\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I\setminus\\{i^{\prime}\\}}\cup\\{i^{\prime}\langle S_{i^{\prime}}\rangle\mathbin{.}[]\\})]$. Clearly, $G_{i^{\prime}}\leq_{C_{i^{\prime}}}G$. Also, because we are not adding recursion binders, the current value of $\tilde{q}$ is appropriate for the IH. With this context $C_{i^{\prime}}$ and $\tilde{q}$, we apply the IH to obtain the typing of ${\llbracket G_{i^{\prime}}\rrbracket}_{p}^{\tilde{q}}$, where priorities start at $\mathrm{ctxpri}(C_{i^{\prime}})=\mathrm{ctxpri}(C)+4=\mathsf{o}_{C}+4$ (cf. Def. 31). Following these typings, Figure 10 gives the typing of ${\llbracket G_{s}\rrbracket}_{p}^{\tilde{q}}$, referring to parts of the process by the number marking its foremost prefix above. Clearly, the priorities in the derivation of Figure 10 meet all requirements. The order of the applications of ${\oplus}^{\star}$ for each $q\in\mathsf{deps}$ does not matter, since the selection actions are asynchronous. Id $\forall i\in I.\leavevmode\nobreak\ v\mathbin{\leftrightarrow}w\vdash v{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}},w{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}$ $\forall i\in I.\leavevmode\nobreak\ {\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\vdash{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p)},{\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}}$ Mix $\forall i\in I.\leavevmode\nobreak\ v\mathbin{\leftrightarrow}w\mathbin{|}{\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\vdash\begin{array}[t]{@{}lr@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p)},v{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}},w{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}},\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}}\end{array}$ $\mathbin{\otimes}^{\star}$ $\forall i\in I.\leavevmode\nobreak\ 5_{i}\vdash\begin{array}[t]{@{}lr@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p)},v{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}_{C}+2}{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}+4},\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}\setminus\\{r\\}}\end{array}$ $\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}$ $\forall i\in I.\leavevmode\nobreak\ 4_{i}\vdash\begin{array}[t]{@{}lr@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}_{C}+1}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p)},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}_{C}+2}{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}+4},\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}\setminus\\{r\\}}\end{array}$ $\forall q\in\mathsf{deps}.\leavevmode\nobreak\ {\oplus}^{\star}$ $\forall i\in I.\leavevmode\nobreak\ 3_{i}\vdash\begin{array}[t]{@{}lr@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}_{C}+1}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p)},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}_{C}+2}{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}+4},\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}_{C}+1}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\\}_{i\in I}\big{)}}_{q\in\mathsf{deps}},\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}\setminus\mathsf{deps}}\end{array}$ ${\oplus}^{\star}$ $\forall i\in I.\leavevmode\nobreak\ 2_{i}\vdash\begin{array}[t]{@{}lr@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}_{C}+1}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p)},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}_{C}+1}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}_{C}+2}{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}+4}\\}_{i\in I},\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}_{C}+1}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\\}_{i\in I}\big{)}}_{q\in\mathsf{deps}},\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}\setminus\mathsf{deps}}&\text{(cf.\ \eqref{eq:outSkipSame})}\end{array}$ $\&$ ${\llbracket G_{s}\rrbracket}_{p}^{\tilde{q}}=1\vdash\begin{array}[t]{@{}lr@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \&^{\mathsf{o}_{C}}\\{i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}_{C}+1}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p)}\\}_{i\in I},&\text{(cf.\ \eqref{eq:outCiType})}\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}_{C}+1}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}_{C}+2}{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}+4}\\}_{i\in I},&\text{(cf.\ \eqref{eq:outCrtRecvType})}\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}_{C}+1}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\\}_{i\in I}\big{)}}_{q\in\mathsf{deps}},&\text{(cf.\ \eqref{eq:outCrtDepsType})}\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}\setminus\mathsf{deps}}&\text{(cf.\ \eqref{eq:outCrtSkipType})}\end{array}$ Figure 10: Typing derivation used in the proof of Theorem 11. 2. 2. If $p=r$, then $p$ is the recipient (algorithm 1). This case is analogous to the previous one. 3. 3. If $p\notin\\{r,s\\}$ (algorithm 1), then further analysis depends on whether the exchange is a dependency for $p$. Let $\displaystyle\mathsf{depon}_{s}$ $\displaystyle:=(s\in\tilde{q}\wedge\mathrm{hdep}(p,s,G))$ (as on algorithm 1), and $\displaystyle\mathsf{depon}_{r}$ $\displaystyle:=(r\in\tilde{q}\wedge\mathrm{hdep}(p,r,G))$ (as on algorithm 1). To see what the truths of $\mathsf{depon}_{s}$ and $\mathsf{depon}_{r}$ mean, we follow Definition 18 and the definitions in Figure 5. $\displaystyle G_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s)$ $\displaystyle=\begin{cases}s{!}r\\{i\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s))\\}_{i\in I}&\text{if $\mathsf{depon}_{s}$ is true}\\\ \mathsf{skip}\mathbin{.}(G_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s))\leavevmode\nobreak\ \text{for any $i^{\prime}\in I$}&\text{otherwise}\end{cases}$ (13) $\displaystyle G_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r)$ $\displaystyle=\begin{cases}r{?}s\\{i\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r))\\}_{i\in I}&\text{if $\mathsf{depon}_{r}$ is true}\\\ \mathsf{skip}\mathbin{.}(G_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r))\leavevmode\nobreak\ \text{for any $i^{\prime}\in I$}&\text{otherwise}\end{cases}$ (14) Let us also consider the relative projections onto $p$ and the participants in $\tilde{q}$ besides $r$ and $s$, which follow by the relative well-formedness of $G_{s}$. For each $q\in\tilde{q}\setminus\\{r,s\\}$, $G_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=\mathsf{skip}\mathbin{.}(G_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q))$ (15) for any $i^{\prime}\in I$. The rest of the analysis depends on the truth of $\mathsf{depon}_{s}$ and $\mathsf{depon}_{r}$. There are four cases. 1. (a) If $\mathsf{depon}_{s}$ is true and $\mathsf{depon}_{r}$ is false (algorithm 1), let us take stock of the types we expect for each of the router’s channels. For ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ we expect $\displaystyle\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p}$ $\displaystyle=\overline{\&^{\mathsf{o}_{C}+2}\\{i{:}\leavevmode\nobreak\ (G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p)\\}_{i\in I}}$ $\displaystyle={{\oplus}}^{\mathsf{o}_{C}+2}\\{i{:}\leavevmode\nobreak\ \overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p)}\\}_{i\in I}.$ (16) For ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$ we expect $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}s}^{\mathsf{o}_{C}}$ $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mus{!}r\\{i\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s))\\}_{i\in I}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}s}^{\mathsf{o}_{C}}$ (cf. (13)) $\displaystyle=\&^{\mathsf{o}_{C}+1}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}s}^{\mathsf{o}_{C}+4}\\}_{i\in I}.$ (17) For each $q\in\tilde{q}\setminus\\{s\\}$, for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ we expect $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}}$ $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mu\mathsf{skip}\mathbin{.}(G_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}}$ (cf. (14) and (15)) $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\leavevmode\nobreak\ \text{for any $i^{\prime}\in I$}.$ (18) Similar to case (1), we apply the IH to obtain the typing of ${\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}$ for each $i\in I$, starting at priority $\mathsf{o}_{C}+4$. We derive the typing of ${\llbracket G_{s}\rrbracket}_{p}^{\tilde{q}}$: $\forall i\in I.\leavevmode\nobreak\ {\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\vdash{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p},{\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}}$ ${\oplus}^{\star}$ $\forall i\in I.\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i\cdot{\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\vdash{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}_{C}+2}\\{i{:}\leavevmode\nobreak\ \overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p}\\}_{i\in I},{\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}}$ $\&$ ${\llbracket G_{s}\rrbracket}_{p}^{\tilde{q}}={{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\mathbin{\triangleright}\\{i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i\cdot{\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\\}_{i\in I}\vdash\begin{array}[t]{@{}lr@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}_{C}+2}\\{i{:}\leavevmode\nobreak\ \overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p}\\}_{i\in I},&\text{(cf.\ \eqref{eq:rDepSCiType})}\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}{:}\leavevmode\nobreak\ \&^{\mathsf{o}_{C}+1}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\\}_{i\in I},&\text{(cf.\ \eqref{eq:rDepSCrtSType})}\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}}&\text{(cf.\ \eqref{eq:rDepSCrtOtherType})}\end{array}$ 2. (b) The case where $\mathsf{depon}_{s}$ is false and $\mathsf{depon}_{r}$ is true (algorithm 1) is analogous to the previous one. 3. (c) If both $\mathsf{depon}_{s}$ and $\mathsf{depon}_{r}$ are true (algorithm 1 and (4)), let us once again take stock of the types we expect for each of the router’s channels. For ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ we expect $\displaystyle\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p}$ $\displaystyle=\overline{\&^{\mathsf{o}_{C}+2}\\{i{:}\leavevmode\nobreak\ (G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p)\\}_{i\in I}}$ $\displaystyle={{\oplus}}^{\mathsf{o}_{C}+2}\\{i{:}\leavevmode\nobreak\ \overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p)}\\}_{i\in I}$ (19) For ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$ we expect $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}s}^{\mathsf{o}_{C}}$ $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mus{!}r\\{i\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s))\\}_{i\in I}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}s}^{\mathsf{o}_{C}}$ (cf. (13)) $\displaystyle=\&^{\mathsf{o}_{C}+1}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}s}^{\mathsf{o}_{C}+4}\\}_{i\in I}$ (20) For ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}$ we expect $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}}$ $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mur{?}s\\{i\mathbin{.}(G_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r))\\}_{i\in I}{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}}$ (cf. (14)) $\displaystyle=\&^{\mathsf{o}_{C}+2}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}+4}\\}_{i\in I}$ (21) For each $q\in\tilde{q}\setminus\\{s,r\\}$, for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ we expect $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}}$ $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mu\mathsf{skip}\mathbin{.}(G_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}}$ (cf. (15)) $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\leavevmode\nobreak\ \text{for any $i^{\prime}\in I$}$ (22) It is clear from (20) and (21) that the router will receive label $i\in I$ first on ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$ and then $i^{\prime}\in I$ on ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}$. We rely on alarm processes (Definition 27) to handle the case $i^{\prime}\neq i$. Similar to case (1), we apply the IH to obtain the typing of ${\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}$ for each $i\in I$, starting at priority $\mathsf{o}_{C}+4$. Figure 11 gives the typing of ${\llbracket G_{s}\rrbracket}_{p}^{\tilde{q}}$. $\begin{array}[b]{@{}l@{}}\forall i\in I.\\\ {\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\vdash\begin{array}[t]{@{}l@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p},\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}}\end{array}\end{array}$ Alarm $\begin{array}[b]{@{}l@{}}\forall i\in I.\\\ \forall i^{\prime}\in I\setminus\\{i\\}.\leavevmode\nobreak\ {\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({\mathsf{chs}})}\vdash\begin{array}[t]{@{}l@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}s}^{\mathsf{o}_{C}+4},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}+4},\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}\setminus\\{s,r\\}}\end{array}\end{array}$ $\&$ $\forall i\in I.\leavevmode\nobreak\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}\mathbin{\triangleright}\\{i{:}\leavevmode\nobreak\ {\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\\}\cup\\{i^{\prime}{:}\leavevmode\nobreak\ {\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({\mathsf{chs}})}\\}_{i^{\prime}\in I\setminus\\{i\\}}\vdash\begin{array}[t]{@{}l@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}s}^{\mathsf{o}_{C}+4},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}{:}\leavevmode\nobreak\ \&^{\mathsf{o}_{C}+2}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}+4}\\}_{i\in I},\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}\setminus\\{s,r\\}}\end{array}$ ${\oplus}^{\star}$ $\forall i\in I.\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}\mathbin{\triangleright}\\{i{:}\leavevmode\nobreak\ {\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\\}\cup\\{i^{\prime}{:}\leavevmode\nobreak\ {\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({\mathsf{chs}})}\\}_{i^{\prime}\in I\setminus\\{i\\}}\vdash\begin{array}[t]{@{}l@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}_{C}+2}\\{i{:}\leavevmode\nobreak\ \overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p}\\}_{i\in I},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}s}^{\mathsf{o}_{C}+4},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}{:}\leavevmode\nobreak\ \&^{\mathsf{o}_{C}+2}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}+4}\\}_{i\in I},\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}\setminus\\{s,r\\}}\end{array}$ $\&$ $\underbrace{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\mathbin{\triangleright}\\{i{:}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}\mathbin{\triangleright}\\{i{:}{\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}\\}{\cup}\\{i^{\prime}{:}{\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({\mathsf{chs}})}\\}_{i^{\prime}\in I\setminus\\{i\\}}\\}_{i\in I}}_{{\llbracket G_{s}\rrbracket}_{p}^{\tilde{q}}}\vdash\begin{array}[t]{@{}lr@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}_{C}+2}\\{i{:}\leavevmode\nobreak\ \overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p}\\}_{i\in I},&\text{(cf.\ \eqref{eq:rDepCiType})}\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}{:}\leavevmode\nobreak\ \&^{\mathsf{o}_{C}+1}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,s){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}s}^{\mathsf{o}_{C}+4}\\}_{i\in I},&\text{(cf.\ \eqref{eq:rDepCrtSType})}\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}{:}\leavevmode\nobreak\ \&^{\mathsf{o}_{C}+2}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,r){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}r}^{\mathsf{o}_{C}+4}\\}_{i\in I},&\text{(cf.\ \eqref{eq:rDepCrtRType})}\\\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\big{)}}_{q\in\tilde{q}\setminus\\{s,r\\}}&\text{(cf.\ \eqref{eq:rDepCrtOtherType})}\end{array}$ Figure 11: Typing derivation used in the proof of Theorem 11, where $\mathsf{chs}=\\{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\\}\cup\\{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}\mid q\in\tilde{q}\\}$. 4. (d) If both $\mathsf{depon}_{s}$ and $\mathsf{depon}_{r}$ are false, let us again take stock of the types we expect for each of the router’s channels. For ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ we expect $\displaystyle\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p}$ $\displaystyle=\overline{G_{i^{\prime}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}+4}p}\leavevmode\nobreak\ \text{for any $i^{\prime}\in I$}.$ For each $q\in\tilde{q}$, for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ we expect $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}}$ $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mu\mathsf{skip}\mathbin{.}(G_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}}$ (cf. (13), (14) and (15)) $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{i^{\prime}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}+4}\leavevmode\nobreak\ \text{for any $i^{\prime}\in I$.}$ Similar to case (1), we apply the IH to obtain the typing of ${\llbracket G_{i^{\prime}}\rrbracket}_{p}^{\tilde{q}}$, starting at priority $\mathsf{o}_{C}+4$. This directly proves the thesis. * • _Recursive definition_ : $G_{s}=\mu Z\mathbin{.}G^{\prime}$ (algorithm 1). Let $\displaystyle\tilde{q}^{\prime}:=\\{q\in\tilde{q}\mid G_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)\neq\bullet\\}$ (23) (as on algorithm 1). We consider the relative projections onto $p$ and the participants in $\tilde{q}$. For each $q\in\tilde{q}^{\prime}$, we know $G_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)\neq\bullet$, while for each $q\in\tilde{q}\setminus\tilde{q}^{\prime}$, we know $G_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=\bullet$. More precisely, by Definition 16, for each $q\in\tilde{q}^{\prime}$, $\displaystyle G_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ $\displaystyle=(\mu Z\mathbin{.}G^{\prime})\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=\mu Z\mathbin{.}(G^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)).$ (24) and thus $\displaystyle G^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ $\displaystyle\neq\mathsf{skip}^{\ast}\mathbin{.}\bullet\leavevmode\nobreak\ \text{and}\leavevmode\nobreak\ G^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)\neq\mathsf{skip}^{\ast}\mathbin{.}Z.$ For each $q\in\tilde{q}\setminus\tilde{q}^{\prime}$, $\displaystyle G_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ $\displaystyle=(\mu Z\mathbin{.}G^{\prime})\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=\bullet,$ (25) and thus $\displaystyle G^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ $\displaystyle=\mathsf{skip}^{\ast}\mathbin{.}\bullet\leavevmode\nobreak\ \text{or}\leavevmode\nobreak\ G^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)=\mathsf{skip}^{\ast}\mathbin{.}Z.$ Further analysis depends on whether $\tilde{q}^{\prime}=\emptyset$ or not. We thus examine two cases: * – If $\tilde{q}^{\prime}=\emptyset$ (algorithm 1), let us consider the local projection $G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p$. We prove that $G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p=\bullet$. Suppose, for contradiction, that $G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p\neq\bullet$. Then, by the definitions in Figure 7, $G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p\neq X$ and $G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p\neq\bullet$. That is, $G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p$ contains communication actions or some recursion variable other than $Z$. However, communication actions in $G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p$ originate from exchanges in $G^{\prime}$, either involving $p$ and some $q\in\tilde{q}$, or as a dependency on an exchange involving some $q\in\tilde{q}$. Moreover, recursion variables in $G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p$ originate from recursion variables in $G^{\prime}$. But this would mean that for this $q$, $G^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)$ contains interactions or recursion variables, contradicting (25). Therefore, it cannot be the case that $G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p\neq\bullet$. Let us take stock of the types we expect for each of the router’s channels. For now, we omit the substitutions in the types. For ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ we expect $\displaystyle\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p}$ $\displaystyle=\overline{\bullet}=\bullet.$ For each $q\in\tilde{q}$, for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ we expect $\displaystyle{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}}$ $\displaystyle={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mu\bullet{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}}=\bullet.$ (cf. (25)) Because all expected types are $\bullet$, the substitutions do not affect the types, so we can omit them altogether. First we apply Empty, giving us an arbitrary recursive context, and thus the recursive context we need. Then, we apply $\bullet$ for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ and for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ for each $q\in\tilde{q}$, and obtain the typing of ${\llbracket G_{s}\rrbracket}_{p}^{\tilde{q}}$ (omitting the recursive context): ${\llbracket G_{s}\rrbracket}_{p}^{\tilde{q}}=\bm{0}\vdash{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \bullet,{({{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ \bullet)}_{q\in\tilde{q}}$ * – If $\tilde{q}^{\prime}\neq\emptyset$ (algorithm 1), then, following similar reasoning as in the previous case, ${G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p=\mu Z\mathbin{.}(G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p)}$. We take stock of the types we expect for each of the router’s channels. Note that, because of the recursive definition on $Z$ in $G_{s}$, there cannot be another recursive definition in the context $C$ capturing the recursion variable $Z$. Therefore, by Definition 29, $Z\notin\widetilde{X_{C}}$. For ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ we expect $\displaystyle\mathrm{deepUnfold}(\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},\ldots)$ $\displaystyle{}=\mathrm{deepUnfold}(\overline{\mu Z\mathbin{.}(G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p)},\ldots)$ $\displaystyle{}=\mathrm{deepUnfold}(\mu Z\mathbin{.}\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},\ldots)$ $\displaystyle{}=\mu Z\mathbin{.}\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{X_{C}}}).$ (26) For each $q\in\tilde{q}^{\prime}$, for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ we expect $\displaystyle\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},\ldots)$ $\displaystyle{}=\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mu\mu Z\mathbin{.}(G^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q)){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},\ldots)$ $\displaystyle{}=\mathrm{deepUnfold}(\mu Z\mathbin{.}{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},\ldots)$ (cf. (24)) $\displaystyle=\mu Z\mathbin{.}\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},{(X,t_{X},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}})}_{X\in\widetilde{X_{C}}}).$ (27) For each $q\in\tilde{q}\setminus\tilde{q}^{\prime}$, for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ we expect $\displaystyle\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},\ldots)$ $\displaystyle{}=\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0mu\bullet{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},\ldots)$ $\displaystyle{}=\mathrm{deepUnfold}(\bullet,\ldots)$ (cf. (25)) $\displaystyle{}=\bullet.$ (28) We also need an assignment in the recursive context for every $X\in\widetilde{X_{C}}$, but not for $Z$. Let $C^{\prime}=C[\mu Z\mathbin{.}[]]$. Clearly, $G^{\prime}\leq_{C^{\prime}}G$. Let us first establish some facts about the recursion binders, priorities, and active participants related to $C^{\prime}$, $G^{\prime}$, and $Z$: * * $\widetilde{X_{C^{\prime}}}=\mathrm{ctxbind}(C^{\prime})=(\mathrm{ctxbind}(C),Z)=(\widetilde{X_{C}},Z)$ (cf. Def. 29). * * $G_{Z}=\mathrm{recdef}(Z,G)=G^{\prime}$, as proven by the context $C^{\prime}$ (cf. Def. 30). * * $\widetilde{Y_{Z}}=\mathrm{subbind}(\mu Z\mathbin{.}G_{Z},G)=\mathrm{ctxbind}(C)=\widetilde{X_{C}}$. * * $\mathsf{o}_{C^{\prime}}=\mathrm{ctxpri}(C^{\prime})=\mathrm{ctxpri}(C)=\mathsf{o}_{C}$, and $\mathsf{o}_{Z}=\mathrm{varpri}(Z,G)=\mathrm{ctxpri}(C)=\mathsf{o}_{C}$, and hence $\mathsf{o}_{C^{\prime}}=\mathsf{o}_{Z}$ (cf. Def. 31). * * $\tilde{q}_{Z}=\tilde{q}^{\prime}$ (cf. Def. 32 and (23)). Because $\widetilde{X_{C^{\prime}}}=(\widetilde{X_{C}},Z)$ and $\tilde{q}^{\prime}=\tilde{q}_{Z}$, $\tilde{q}^{\prime}$ is appropriate for the IH. We apply the IH on $C^{\prime}$, $G^{\prime}$, and $\tilde{q}^{\prime}$ to obtain a typing for ${\llbracket G^{\prime}\rrbracket}_{p}^{\tilde{q}^{\prime}}$, where we immediately make use of the facts established above. We give the assignment to $Z$ in the recursive context separate from those for the recursion variables in $\widetilde{X_{C}}$. Also, by Proposition 15, we can write the final unfolding on $Z$ in the types separately. For example, the type for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ is $\displaystyle\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C^{\prime}}}p},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{X_{C^{\prime}}}})$ $\displaystyle=\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in(\widetilde{X_{C}},Z)})$ $\displaystyle=\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},\big{(}{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{X_{C}}},(Z,t_{Z},\overline{G_{Z}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Z}}p})\big{)})$ $\displaystyle=\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},\big{(}{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{X_{C}}},(Z,t_{Z},\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p})\big{)})$ $\displaystyle=\mathrm{unfold}^{t_{Z}}\big{(}\mu Z\mathbin{.}\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{X_{C}}})\big{)}.$ The resulting typing is as follows: $\displaystyle{\llbracket G^{\prime}\rrbracket}_{p}^{\tilde{q}^{\prime}}\vdash\begin{array}[t]{@{}l@{}}{\left(X{:}\leavevmode\nobreak\ \left(\begin{array}[]{@{}l@{}}\mathrm{deepUnfold}(\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p},{(Y,t_{Y},\overline{G_{Y}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Y}}p})}_{Y\in\widetilde{Y_{X}}}),\\\\[4.0pt] {\Big{(}\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}},{(Y,t_{Y},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{Y}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{Y}})}_{Y\in\widetilde{Y_{X}}})\Big{)}}_{q\in\tilde{q}_{X}}\end{array}\right)\right)}_{X\in\widetilde{X_{C}}},\\\\[16.0pt] Z{:}\leavevmode\nobreak\ \left(\begin{array}[]{@{}l@{}}\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{X_{C}}}),\\\\[4.0pt] {\Big{(}\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},{(X,t_{X},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}})}_{X\in\widetilde{X_{C}}})\Big{)}}_{q\in\tilde{q}^{\prime}}\end{array}\right);\\\\[16.0pt] {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \mathrm{unfold}^{t_{Z}}\big{(}\mu Z\mathbin{.}\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{X_{C}}})\big{)},\\\\[6.0pt] {\Big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ \mathrm{unfold}^{t_{Z}}\big{(}\mu Z\mathbin{.}\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},{(X,t_{X},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}})}_{X\in\widetilde{X_{C}}})\big{)}\Big{)}}_{q\in\tilde{q}^{\prime}}\end{array}$ By assumption, we have $\displaystyle t_{Z}$ $\displaystyle=\max_{\mathsf{pr}}\left(\begin{array}[]{@{}l@{}}\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{X_{C}}})\\\\[4.0pt] {\Big{(}\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},{(X,t_{X},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}})}_{X\in\widetilde{X_{C}}})\Big{)}}_{q\in\tilde{q}^{\prime}}\end{array}\right)+1,$ so $t_{Z}$ is clearly greater than the maximum priority appearing in the types before unfolding. Hence, we can apply Rec to eliminate $Z$ from the recursive context, and to fold the types, giving the typing of ${\llbracket G_{s}\rrbracket}_{p}^{\tilde{q}}=\mu Z({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}},{({{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}})}_{q\in\tilde{q}^{\prime}})\mathbin{.}{\llbracket G^{\prime}\rrbracket}_{p}^{\tilde{q}^{\prime}}$: $\displaystyle{\llbracket G_{s}\rrbracket}_{p}^{\tilde{q}}\vdash\begin{array}[t]{@{}l@{}}{\left(X{:}\leavevmode\nobreak\ \left(\begin{array}[]{@{}l@{}}\mathrm{deepUnfold}(\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p},{(Y,t_{Y},\overline{G_{Y}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Y}}p})}_{Y\in\widetilde{Y_{X}}}),\\\\[4.0pt] {\Big{(}\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}},{(Y,t_{Y},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{Y}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{Y}})}_{Y\in\widetilde{Y_{X}}})\Big{)}}_{q\in\tilde{q}_{X}}\end{array}\right)\right)}_{X\in\widetilde{X_{C}}};\\\\[16.0pt] {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \mu Z\mathbin{.}\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{X_{C}}}),\\\\[6.0pt] {\Big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ \mu Z\mathbin{.}\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG^{\prime}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},{(X,t_{X},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}})}_{X\in\widetilde{X_{C}}})\Big{)}}_{q\in\tilde{q}^{\prime}}\end{array}$ In this typing, the type for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ concurs with (26), and, for every $q\in\tilde{q}^{\prime}$, the type for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ concurs with (27). For every $q\in\tilde{q}\setminus\tilde{q}^{\prime}$, we can add the type for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ in (28) by applying $\bullet$. This proves the thesis. * • _Recursive call_ : $G_{s}=Z$ (algorithm 1). Clearly, because $G$ is closed (i.e. $\mathrm{frv}(G)=\emptyset$), $Z\in\widetilde{X_{C}}$. More precisely, $\widetilde{X_{C}}=(\tilde{X}_{1},Z,\tilde{X}_{2})$. Note that the recursive definitions on the variables in $\tilde{X}_{1}$ appear in $G$ after the recursive definitions on the variables in $(Z,\tilde{X}_{2})$. Because the unfoldings of $(Z,\tilde{X}_{2})$ occur before the unfoldings of $\tilde{X}_{1}$, the recursive definitions on the variables in $\tilde{X}_{1}$ are renamed in order to avoid capturing these variables when performing the unfoldings of $(Z,\tilde{X}_{2})$. So, after the unfoldings of $(Z,\tilde{X}_{2})$, there are no recursive calls on the variables in $\tilde{X}_{1}$ anymore, so the unfoldings on $\tilde{X}_{1}$ do not have any effect on the types. Also, note that $\tilde{X}_{2}=\widetilde{Y_{Z}}$ (cf. Def. 29). Let us take stock of the types we expect for our router’s channels. For ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ we expect $\displaystyle\mathrm{deepUnfold}(\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},\ldots)$ $\displaystyle{}=\mathrm{deepUnfold}(\overline{Z\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},\ldots)$ $\displaystyle{}=\mathrm{deepUnfold}(\overline{Z},\ldots)$ $\displaystyle{}=\mathrm{deepUnfold}(Z,{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in(\tilde{X}_{1},Z,\widetilde{Y_{Z}})})$ $\displaystyle{}=\mathrm{deepUnfold}(Z,{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in(Z,\widetilde{Y_{Z}})})$ $\displaystyle{}=\mu Z\mathbin{.}({\uparrow^{t_{Z}}}\mathrm{deepUnfold}(\overline{G_{Z}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Z}}p},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{Y_{Z}}}))$ (29) For each $q\in\tilde{q}$, for ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ we expect $\displaystyle\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},\ldots)$ $\displaystyle{}=\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muZ{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},\ldots)$ $\displaystyle{}=\mathrm{deepUnfold}(Z,{(X,t_{X},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}})}_{X\in(\tilde{X}_{1},Z,\widetilde{Y_{Z}})})$ $\displaystyle{}=\mathrm{deepUnfold}(Z,{(X,t_{X},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}})}_{X\in(Z,\widetilde{Y_{Z}})})$ $\displaystyle{}=\mu Z\mathbin{.}({\uparrow^{t_{Z}}}\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{Z}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{Z}},{(X,t_{X},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}})}_{X\in\widetilde{Y_{Z}}}))$ (30) Also, we need an assignment in the recursive context for every $X\in\widetilde{X_{C}}$. By Lemma 13, $\tilde{q}=\tilde{q}_{Z}$. Hence, for $Z$, the assignment should be as follows: $\displaystyle Z{:}\leavevmode\nobreak\ \left(\begin{array}[]{@{}l@{}}\mathrm{deepUnfold}(\overline{G_{Z}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Z}}p},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{Y_{Z}}}),\\\ {\Big{(}\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{Z}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{Z}},{(X,t_{X},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}})}_{X\in\widetilde{Y_{Z}}})\Big{)}}_{q\in\tilde{q}}\end{array}\right)$ (33) We apply Var to obtain the typing of ${\llbracket G_{s}\rrbracket}_{\tilde{q}}^{p}$, where we make us the rule’s allowance for an arbitrary recursive context up to the assignment to $Z$. Var is applicable, because the types are recursive definitions on $Z$, concurring with the types assigned to $Z$, and lifted by a common lifter $t_{Z}$. Var $\begin{array}[]{@{}l@{}}{\llbracket G_{s}\rrbracket}_{\tilde{q}}^{p}=X{\langle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}},{({{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}})}_{q\in\tilde{q}}\rangle}\\\\[6.0pt] {}\vdash\begin{array}[t]{@{}l@{}}{\left(X{:}\leavevmode\nobreak\ \left(\begin{array}[]{@{}l@{}}\mathrm{deepUnfold}(\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p},{(Y,t_{Y},\overline{G_{Y}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Y}}p})}_{Y\in\widetilde{Y_{X}}}),\\\\[4.0pt] {\Big{(}\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}},{(Y,t_{Y},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{Y}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{Y}})}_{Y\in\widetilde{Y_{X}}})\Big{)}}_{q\in\tilde{q}_{X}}\end{array}\right)\right)}_{X\in\widetilde{X_{C}}\setminus(Z)},\\\\[8.0pt] Z{:}\leavevmode\nobreak\ \left(\begin{array}[]{@{}l@{}}\mathrm{deepUnfold}(\overline{G_{Z}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Z}}p},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{Y_{Z}}}),\\\ {\Big{(}\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{Z}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{Z}},{(X,t_{X},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}})}_{X\in\widetilde{Y_{Z}}})\Big{)}}_{q\in\tilde{q}}\end{array}\right);\\\\[8.0pt] {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \mu Z\mathbin{.}({\uparrow^{t_{Z}}}\mathrm{deepUnfold}(\overline{G_{Z}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Z}}p},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{Y_{Z}}})),\\\\[6.0pt] {\left({{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ \mu Z\mathbin{.}({\uparrow^{t_{Z}}}\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{Z}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{Z}},{(X,t_{X},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}})}_{X\in\widetilde{Y_{Z}}}))\right)}_{q\in\tilde{q}}\end{array}\end{array}$ In this typing, the type of ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ concurs with the expected type in (29), the types of ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ for each $q\in\tilde{q}$ concur with the expected types in (30), and the assignment to $Z$ in the recursive context concurs with (33). This proves the thesis. ∎ Now, we can prove Theorem 11 as a corollary of Theorem 16: ###### Proof of Theorem 11 on Theorem 11. We have been given a closed, relative well-formed global type $G$, and a participant $p\in\mathsf{prt}(G)$. Let $C:=[]$ and $G_{s}:=G$. Clearly, $G_{s}\leq_{C}G$. By Definition 32, ${\mathrm{active}(C,G)={\mathsf{prt}(G)}^{2}}$. For $p$ to be a participant of $G$, there must be an exchange involving $p$ and some other participant $q$, i.e. there exists a $q\in\mathsf{prt}(G)$ such that $(p,q)\in\mathrm{active}(C,G)$. Moreover, $\tilde{q}$ as defined in Theorem 16 is $\\{q\in\mathsf{prt}(G)\mid(p,q)\in\mathrm{active}(C,G)\\}=q\in\mathsf{prt}(G)\setminus\\{p\\}$. Hence, Theorem 16 allows us to find a typing for ${\llbracket G\rrbracket}_{p}^{\mathsf{prt}(G)\setminus\\{p\\}}$. Let us consider the precise values of the ingredients of Theorem 16 in our application: 1. 1. $\mathsf{o}_{C}=\mathrm{ctxpri}(C)=0$, 2. 2. $\widetilde{X_{C}}=\mathrm{ctxbind}(C)=()$, 3. 3. $\begin{array}[t]{@{}rll@{}}D_{p}&=\mathrm{deepUnfold}(\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}p},{(X,t_{X},\overline{G_{x}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}p})}_{X\in\widetilde{X_{C}}})\\\ &=\overline{G\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}p}&\text{(cf.\ \lx@cref{creftypecap~refnum}{d:deepUnfold}),}\end{array}$ 4. 4. $\begin{array}[t]{@{}rll@{}}E_{q}&=\mathrm{deepUnfold}({{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{s}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{C}},{(X,t_{X},{{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{X}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{\mathsf{o}_{X}})}_{X\in\widetilde{X_{C}}})\\\ &={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{0}&\text{(cf.\ \lx@cref{creftypecap~refnum}{d:deepUnfold}).}\end{array}$ Finally, the result of Theorem 16 is as follows: $\displaystyle{\llbracket G_{s}\rrbracket}_{p}^{\tilde{q}}\vdash{\Big{(}X{:}\leavevmode\nobreak\ \big{(}A_{X},{(B_{X,q})}_{q\in\tilde{q}_{X}}\big{)}\Big{)}}_{X\in\widetilde{X_{C}}};\leavevmode\nobreak\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ D_{p},\leavevmode\nobreak\ {({{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ E_{q})}_{q\in\tilde{q}}$ Applying (1)–(4) above, we get the following: $\displaystyle{\llbracket G\rrbracket}_{p}^{\mathsf{prt}(G)\setminus\\{p\\}}\vdash\emptyset;\leavevmode\nobreak\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{G\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}p},\leavevmode\nobreak\ {\big{(}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{:}\leavevmode\nobreak\ {{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{p{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}q}^{0}\big{)}}_{q\in\mathsf{prt}(G)\setminus\\{p\\}}$ This coincides exactly with the result of Theorem 11. ∎ #### 4.3.2 Transference of Results (Operational Correspondence) Given a global type $G$, we now formalize the transference of correctness properties such as deadlock freedom from ‘$\mathrm{net}(G)$’ (cf. Definition 25) to ‘$G$’. Here, we define an _operational correspondence_ between networks and global types, in both directions. That is, we show that a network performs interactions between implementations and routers and between pairs of routers if and only if that communication step is stipulated in the corresponding global type (Theorems 19 and 23). Before formalizing the operational correspondence, we show that networks of routed implementations never reduce to alarm processes. To be precise, because alarm processes only can occur in routers (not in implementations), we show that none of the routers of a network reduces to an alarm process, formalized using evaluation contexts: ###### Definition 34 (Evaluation Context). We define an _evaluation context_ as a process with a single hole ‘$\mkern 1.0mu[\,]\mkern-3.0mu$’, not prefixed by input or branching: $\displaystyle E::=(\bm{\nu}xy)\,E\;\mbox{\large{$\mid$}}\;P\mathbin{|}E\;\mbox{\large{$\mid$}}\;\mu X(\tilde{z})\mathbin{.}E\;\mbox{\large{$\mid$}}\;[\,]$ Given an evaluation context $E$, we write ‘$\mkern 1.0muE[P]\mkern-3.0mu$’ to denote the process obtained by replacing the hole in $E$ with $P$. ###### Theorem 17. Given a relative well-formed global type $G$ and a network of routed implementations $\mathcal{N}\in\mathrm{net}(G)$, then $\mathcal{N}{\centernot\longrightarrow}^{\ast}E[{\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({\tilde{x}})}],$ for any evaluation context $E$ and set of endpoints $\tilde{x}$. ###### Proof. By definition (Definition 25), $\mathcal{N}$ consists only of routers (Definition 19) and well-typed processes not containing the alarm process (cf. the assumption below Definition 27). Suppose, for contradiction, that there are $E\in\mathcal{E}$ and $\tilde{x}$ such that $\mathcal{N}\longrightarrow^{\ast}E[{\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({\tilde{x}})}]$. Since only routers can contain the alarm process, there is a router $\mathcal{R}_{p}$ in $\mathcal{N}$ for participant $p\in\mathsf{prt}(G)$ that reduces to the alarm process. Since it is the only possibility for a router synthesized by Algorithm 1 to contain the alarm process, it must contain the process in (4). This process is synthesized on algorithm 1 of Algorithm 1, so there is an exchange in $G$ with sender $s\in\mathsf{prt}(G)\setminus\\{p\\}$ and recipient $r\in\mathsf{prt}(G)\setminus\\{p\\}$ that is a dependency for the interactions of $p$ with both $s$ and $r$. For this exchange, the router $\mathcal{R}_{s}$ for $s$ contains the process returned on algorithm 1 of Algorithm 1, and the router $\mathcal{R}_{r}$ for $r$ contains the process returned on algorithm 1. Suppose $s$ has a choice between the labels in $I$, and the implementation of $s$ chooses $i\in I$. Then, $\mathcal{R}_{s}$ sends $i$ to $\mathcal{R}_{r}$ and $\mathcal{R}_{p}$. Now, for $\mathcal{R}_{p}$ to reduce to the alarm process, it has to receive from $\mathcal{R}_{r}$ a label $i^{\prime}\in I\setminus\\{i\\}$. However, this contradicts algorithm 1 of Algorithm 1, which clearly defines $\mathcal{R}_{r}$ to send $i$ to $\mathcal{R}_{p}$. Hence, $\mathcal{N}{\centernot\longrightarrow}^{\ast}E[{\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({\tilde{x}})}]$. ∎ It follows from this and the typability of routers (Theorem 11) that networks of routed implementations are deadlock free: ###### Theorem 18. For relative well-formed global type $G$, every $\mathcal{N}\in\mathrm{net}(G)$ is deadlock free. ###### Proof. By the typability of routers (Theorem 11) and the duality of the types of router channels (Theorem 9), $\mathcal{N}\vdash\emptyset;\emptyset$. Hence, by Theorem 5, $\mathcal{N}$ is deadlock free, and by Theorem 17, $\mathcal{N}$ never reduces to the alarm process. ∎ To formalize our operational correspondence result, we apply the labeled reductions for processes ‘$\mkern 1.0muP\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q$’ (cf. Definition 9) and define a labeled transition system (LTS) for global types. ###### Definition 35 (LTS for Global Types). We define the relation ‘$\mkern 1.0muG\xrightarrow{\vspace{-.8ex}\alpha}G^{\prime}\mkern-2.0mu$’, with labels ‘$\mkern 1.0mu\beta\mkern-3.0mu$’ of the form ‘$\mkern 1.0mup\rangle q{:}\ell\langle S\rangle\mkern-3.0mu$’ (sender, recipient, label, and message type), by the following rules: $\raisebox{8.0pt}{}j\in I$ $\mkern-6.0mup\mkern 1.0mu{\mathbin{\twoheadrightarrow}}\mkern 1.0muq\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}\mkern 1.0mu{\xrightarrow{\vspace{-.8ex}p\rangle q{:}j\langle S_{j}\rangle}}\mkern 1.0muG_{j}\mkern-6.0mu$ $G\xrightarrow{\vspace{-.8ex}\alpha}G^{\prime}$ $\mkern-6.0mu\raisebox{12.0pt}{}\mathsf{skip}\mathbin{.}G\mkern 1.0mu{\xrightarrow{\vspace{-.8ex}\alpha}}\mkern 1.0muG^{\prime}\mkern-6.0mu$ $G\\{\mu X\mathbin{.}G/X\\}\xrightarrow{\vspace{-.8ex}\alpha}G^{\prime}$ $\mu X\mathbin{.}G\xrightarrow{\vspace{-.8ex}\alpha}G^{\prime}$ Intuitively, operational correspondence states: 1. 1. every transition of a global type is mimicked by a precise sequence of labeled reductions originating from an associated completable network (_completeness_ ; Theorem 19), and 2. 2. for every labeled reduction originated in a completable network there is a corresponding global type transition (_soundness_ ; Theorem 23). We write ‘$\rho_{1}\rho_{2}$’ for the composition of relations ‘$\rho_{1}$’ and ‘$\rho_{2}$’. Recall that the notation ‘$\longrightarrow^{\star}$’ stands for finite sequences of reductions, as defined in 2. ###### Theorem 19 (Operational Correspondence: Completeness). Suppose given a relative well-formed global type $G$. Also, suppose given $p,q\in\mathsf{prt}(G)$ and a set of labels $J$ such that $j\in J$ if and only if $G\xrightarrow{\vspace{-.8ex}p\rangle q:j\langle S_{j}\rangle}G_{j}$ for some $S_{j}$. Then, 1. 1. for any completable $\mathcal{N}\in\mathrm{net}(G)$, there exists a $j^{\prime}\in J$ such that $\mathcal{N}^{\circlearrowright}\longrightarrow^{\star}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}:j^{\prime}}}}}\,\mathcal{N}_{0}$; 2. 2. for any $j^{\prime}\in J$, there exists a completable $\mathcal{N}\in\mathrm{net}(G)$ such that $\mathcal{N}^{\circlearrowright}\longrightarrow^{\star}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}:j^{\prime}}}}}\,\mathcal{N}_{0}$; 3. 3. for any completable $\mathcal{N}\in\mathrm{net}(G)$ and any $j^{\prime}\in J$, if $\mathcal{N}^{\circlearrowright}\longrightarrow^{\star}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}:j^{\prime}}}}}\,\mathcal{N}_{0}$, then there exists a completable $\mathcal{N}_{j^{\prime}}\in\mathrm{net}(G_{j^{\prime}})$ such that, $\displaystyle\mathcal{N}_{0}\,\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}\rangle{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}{:}j^{\prime}}}}}\longrightarrow^{\star}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}j^{\prime}}}}}\longrightarrow^{\star}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}v}}}}\,\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}\rangle{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}{:}w}}}}\,\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}v\mathbin{\leftrightarrow}w}}}}\longrightarrow^{\star}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}w}}}}\,\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}v\mathbin{\leftrightarrow}w}}}}\,\mathcal{N}_{j^{\prime}}^{\circlearrowright}.$ ###### Proof. By the labelled transitions of global types (Def. 35) and relative well- formedness, $G$ is a sequence of $\mathsf{skip}$s followed by an exchange from $p$ to $q$ over the labels in $J$. Since the $\mathsf{skip}$s do not influence the behavior of routers, let us assume simply that $\displaystyle G=p\mathbin{\twoheadrightarrow}q\\{j\langle S_{j}\rangle\mathbin{.}G_{j}\\}_{j\in J}.$ We prove each Subitem separately. 1. (a) Take any completable $\mathcal{N}\in\mathrm{net}(G)$. By definition (Def. 26), $\mathcal{N}^{\circlearrowright}\vdash\emptyset;\emptyset$. By the construction of networks of routed implementations (Def. 25), ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\in\mathrm{bn}(\mathcal{N}^{\circlearrowright})$, and ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ is connected to ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$. Also by construction, the type of ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ in the typing derivation of $\mathcal{N}^{\circlearrowright}$ is $\displaystyle G\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}p={\oplus}^{0}\\{j:{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{j}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{1}(G_{j}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{4}p)\\})_{j\in J}.$ By the well-typedness of $\mathcal{N}^{\circlearrowright}$, we can infer the kind of action that is defined on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$: a selection, or a forwarder. By induction on the number of connected forwarders (which is finite by the finiteness of process terms), eventually a forwarder has to be connected to a selection. So, after reducing the forwarders, we have a selection on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$, of some $j^{\prime}\in J$. Hence, by Fairness (Theorem 7), after a finite number of steps, we can observe a communication of the label $j^{\prime}$ from ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ to ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$. This proves the thesis: $\mathcal{N}^{\circlearrowright}\longrightarrow^{\star}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}:j^{\prime}}}}}\,\mathcal{N}_{0}$. 2. (b) Following the proof of the existence of completable networks (Proposition 10), we can generate an implementation process for all of $G$’s participants from local projections (cf. Proposition 1). Take any $j^{\prime}\in J$. For the implementation process of $p$, we specifically generate an implementation process that sends the label $j^{\prime}$. These implementation processes allow us to construct $\mathcal{N}$, which by construction is in $\mathrm{net}(G)$ and is completable. Following the reasoning as in Subitem (a), ${\mathcal{N}^{\circlearrowright}\longrightarrow^{\star}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}:j^{\prime}}}}}\,\mathcal{N}_{0}}$. 3. (c) By definition (Def. 26), $\mathcal{N}^{\circlearrowright}\vdash\emptyset;\emptyset$. Hence, by Fairness (Theorem 7), for any of the pending names of $\mathcal{N}^{\circlearrowright}$, we can observe a communication after a finite number of steps. By construction (Def. 25), the endpoints that we are required to observe by thesis are bound in $\mathcal{N}^{\circlearrowright}$. From the shape of $G$, the definition of routed implementations (Def. 24), and the typability of routers (Theorem 11), we know the types of all the required endpoints in $\mathcal{N}^{\circlearrowright}$. We can deduce the required labeled reductions following the reasoning as in Subitem (a). Let us summarize the origin of each of the network’s steps: 1. 1. $\mathcal{N}^{\circlearrowright}\longrightarrow^{\star}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}:j^{\prime}}}}}\,\mathcal{N}_{0}$: The implementation of $p$ selects label $j^{\prime}$ with $p$’s router. 2. 2. $\mathcal{N}_{0}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}\rangle{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}:j^{\prime}}}}}\mathcal{N}_{1}$: The router of $p$ forwards $j^{\prime}$ to $q$’s router. 3. 3. $\mathcal{N}_{1}\longrightarrow^{\star}\mathcal{N}_{2}$: The router of $p$ forwards $j^{\prime}$ to the routers of the participant that depend on the output by $p$, and these routers forward $j^{\prime}$ to their respective implementations. 4. 4. $\mathcal{N}_{2}\,\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}:j^{\prime}}}}}\longrightarrow^{\star}\mathcal{N}_{3}$: The router of $q$ forwards $j^{\prime}$ to $q$’s implementation, and to the routers of the participants that depend on the input by $q$, and these routers forward $j^{\prime}$ to their respective implementation (if they have not done so already for the output dependency on $p$). 5. 5. $\mathcal{N}_{3}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}:v}}}}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}\rangle{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}:w}}}}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}v\mathbin{\leftrightarrow}w}}}}\mathcal{N}_{4,v}$: The implementation of $p$ sends an endpoint $v$ to $p$’s router, which sends a fresh endpoint $w$ to $q$’s router, and $v$ is forwarded to $w$. 6. 6. $\mathcal{N}_{4,v}\longrightarrow^{\star}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}:w}}}}\,\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}v\mathbin{\leftrightarrow}w}}}}\,\mathcal{N}_{j^{\prime}}^{\circlearrowright}$: The router of $q$ sends a fresh endpoint $w$ to $q$’s implementations, and $v$ is forwarded to $w$. In $\mathcal{N}_{j^{\prime}}^{\circlearrowright}$, all routers have transitioned to routers for $G_{j^{\prime}}$. Moreover, by Type Preservation (Theorem 2), $\mathcal{N}_{j^{\prime}}^{\circlearrowright}\vdash\emptyset;\emptyset$. By isolating restrictions on endpoints that belong only to implementation processes, we can find $\mathcal{N}_{j^{\prime}}\in\mathrm{net}(G_{j^{\prime}})$ such that $\mathcal{N}_{j^{\prime}}^{\circlearrowright}$ is its completion. This proves the thesis. Note that $G$ can also contain recursive definitions before the initial exchange; this case can be dealt with by unfolding. ∎ Our soundness result, given below as Theorem 23, will capture the notion that after any sequence of reductions from the network of a global type $G$, a network of another global type $G^{\prime}$ can be reached. Crucially, $G^{\prime}$ can be reached from $G$ through a series of transitions. Networks are inherently concurrent, whereas global types are built out of sequential compositions; as a result, the network could have enabled (asynchronous) actions that correspond to exchanges that are not immediately enabled in the global type. For example, consider the global types $G=a\mathbin{\twoheadrightarrow}b\big{\\{}1\langle S_{1}\rangle.c\mathbin{\twoheadrightarrow}d\\{1\langle S^{\prime}\rangle.\mathsf{end}\\},2\langle S_{2}\rangle.c\mathbin{\twoheadrightarrow}d\\{1\langle S^{\prime}\rangle.\mathsf{end}\\}\big{\\}}$ and $G^{\prime}=a\mathbin{\twoheadrightarrow}b\big{\\{}1\langle S\rangle.b\mathbin{\twoheadrightarrow}c\\{1\langle S^{\prime}\rangle.\mathsf{end}\\}\big{\\}}$. Clearly, the initial exchange in $G$ between $a$ and $b$ is not a dependency for the following exchange between $c$ and $d$. The routers of $c$ and $d$ synthesized from $G$ thus start with their exchange, without awaiting the initial exchange between $a$ and $b$ to complete. Hence, in a network of $G$, both exchanges in $G$ may be enabled simultaneously. We further refer to exchanges that may be simultaneously enabled in networks as _independent (global) exchanges_. While all exchanges appearing in $G$ are independent, the two exchanges in $G^{\prime}$ are not. In the proof of soundness, we may encounter in a network reductions related to independent exchanges, so we have to be able to identify the independent exchanges in the global type to which the network belongs. Lemma 21 states that independent exchanges related to observed reductions in a network of a global type $G$ can be reached from $G$ after any sequence of transitions in a finite number of steps. The proof of this lemma relies on Lemma 20, which ensures that if a participant does not depend on a certain exchange, then the routers synthesized at each of the branches of the exchange are equal. ###### Lemma 20. Suppose given a relative well-formed global type $G=s\mathbin{\twoheadrightarrow}r\\{i\langle S_{i}\rangle.G_{i}\\}_{i\in I}$, and take any ${p\in\mathsf{prt}(G)\setminus\\{s,r\\}}$ and $\tilde{q}\subseteq\mathsf{prt}(G)\setminus\\{p\\}$. If neither $\mathrm{hdep}(p,s,G)$ nor $\mathrm{hdep}(p,r,G)$ holds, then ${{\llbracket G_{i}\rrbracket}_{p}^{\tilde{q}}={\llbracket G_{j}\rrbracket}_{p}^{\tilde{q}}}$ for every $i,j\in I$. ###### Proof. The analysis proceeds by cases on the structure of $G$. As a representative case we consider ${G=s\mathbin{\twoheadrightarrow}r\\{1\langle S_{1}\rangle.G_{1},2\langle S_{2}\rangle.G_{2}\\}}$. Towards a contradiction, we assume ${\llbracket G_{1}\rrbracket}_{p}^{\tilde{q}}\neq{\llbracket G_{2}\rrbracket}_{p}^{\tilde{q}}$. There are many cases where Algorithm 1 generates differents routers for $p$ at $G_{1}$ and at $G_{2}$. We discuss the interesting case where ${\llbracket G_{1}\rrbracket}_{p}^{\tilde{q}}={{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\mathbin{\triangleright}\ldots$ (algorithm 1) and ${\llbracket G_{2}\rrbracket}_{p}^{\tilde{q}}={{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q_{2}}}\mathbin{\triangleright}\ldots$ (algorithm 1). Then $G_{1}=p\mathbin{\twoheadrightarrow}q_{1}\\{\ldots\\}$ and $G_{2}=q_{2}\mathbin{\twoheadrightarrow}p\\{\ldots\\}$. We have $G_{1}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q_{1})=p\\{\ldots\\}$ and $G_{2}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q_{1})=\mathsf{skip}\ldots$ or $G_{2}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q_{1})=p{?}q_{2}\\{\ldots\\}$ (w.l.o.g., assume the former). Since $G$ is relative well-formed, the projection $G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q_{1})$ must exist. Hence, since $p\notin\\{s,r\\}$ and $G_{1}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q_{1})\neq G_{2}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q_{1})$, it must be the case that $q_{1}\in\\{s,r\\}$—w.l.o.g., assume $q_{1}=s$. Then $G\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(p,q_{1})=q_{1}{!}r\\{1.p\\{\ldots\\},2.\mathsf{skip}\ldots\\}$, and thus $\mathrm{hdep}(p,q_{1},G)=\mathrm{hdep}(p,s,G)$ is true. This contradicts the assumption that $\mathrm{hdep}(p,s,G)$ is false. ∎ ###### Lemma 21. Suppose given a relative well-formed global type $G$ and a completable $\mathcal{N}\in\mathrm{net}(G)$ such that $\mathcal{N}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}:\ell}}}}$, for some $c\in\mathsf{prt}(G)$. For every $G^{\prime}$ and $\beta_{1},\ldots,\beta_{n}$ ($n\geq 0$) such that $G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{n}}G^{\prime}$ where $c$ is not involved in any $\beta_{k}$ (with $G=G^{\prime}$ if $n=0$), there exist $G^{\prime\prime}$, $d\in\mathsf{prt}(G)$, and $\beta^{\prime}_{1},\ldots,\beta^{\prime}_{m}$ ($m\geq 0$) such that ${G^{\prime}\xrightarrow{\vspace{-.8ex}\beta^{\prime}_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta^{\prime}_{m}}G^{\prime\prime}=c\mathbin{\twoheadrightarrow}d\\{i\langle S_{i}\rangle.G_{i}\\}_{i\in I}}$ where $c$ is not involved in any $\beta^{\prime}_{k}$ (with $G^{\prime\prime}=c\mathbin{\twoheadrightarrow}d\\{i\langle S_{i}\rangle.G_{i}\\}_{i\in I}$ if $m=0$). ###### Proof. By induction on $n$ (IH1). We first observe that the behavior on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}$ in $\mathcal{N}^{\circlearrowright}$ can only arise from the router generated for $c$ at $G$, following Algorithm 1 (algorithm 1) after finitely many passes through lines 1 (no dependency) and 1 (skip); for simplicity, assume only algorithm 1 applies. * • Case $n=0$. Let $x\geq 0$ denote the number of passes through algorithm 1 to generate the router for $c$ at $G$. We apply induction on $x$ (IH2): * – Case $x=0$. The router for $c$ at $G$ is generated through algorithm 1, so $G=c\mathbin{\twoheadrightarrow}d\\{i\langle S_{i}\rangle.G_{i}\\}_{i\in I}$, proving the thesis. * – Case $x=x^{\prime}+1$. Then $G=a\mathbin{\twoheadrightarrow}b\\{i\langle S_{i}\rangle.G_{i}\\}_{i\in I}$ and algorithm 1 returns the router for $c$ at $G_{j}$ for any $j\in I$. We have $G\xrightarrow{\vspace{-.8ex}a\rangle b:j\langle S_{j}\rangle}G_{j}$. Given the same implementation process for $c$ as in $\mathcal{N}$, we can construct a completable $\mathcal{M}\in\mathrm{net}(G_{j})$ such that $\mathcal{M}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}:\ell}}}}$. Hence, the thesis follows from IH2. * • Case $n=n^{\prime}+1$. By assumption, $G\xrightarrow{\vspace{-.8ex}\beta_{1}}G^{\prime}_{1}$ where $c$ is not the sender or recipient in $\beta_{1}$. Hence, $G=a\mathbin{\twoheadrightarrow}b\\{i.\langle S_{i}\rangle.G_{i}\\}_{i\in I}$ where $G^{\prime}_{1}=G_{j}$ for some $j\in I$. The router for $c$ at $G$ is thus generated through algorithm 1 of Algorithm 1. It follows from Lemma 20 that this router is equal to the router for $c$ at $G^{\prime}_{1}$, but with one less pass through algorithm 1. Given the same implementation process for $c$ as in $\mathcal{N}$, we can construct a completable $\mathcal{M}\in\mathrm{net}(G^{\prime}_{1})$ such that $\mathcal{M}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}:\ell}}}}$. Hence, the thesis follows from IH1. ∎ The proof of soundness relies on Proposition 22: if different reductions are enabled for a given process, then they do not exclude each other. That is, the same process is reached no matter the order in which those reductions are executed. We refer to simultaneously enabled reductions as _independent reductions_. ###### Proposition 22 (Independent Reductions). Suppose given a process $P\vdash\Omega;\Gamma$ and reduction labels $\alpha$ and $\alpha^{\prime}_{1},\ldots,\alpha^{\prime}_{n}$ ($n\geq 1$) where $\alpha\notin\\{\alpha^{\prime}_{1},\ldots,\alpha^{\prime}_{n}\\}$ (cf. Definition 9). If $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}$ and $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{1}}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{n}}}}}$, then there exists a process $Q$ such that $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{1}}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{n}}}}}Q$ and $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{1}}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{n}}}}}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}\leavevmode\nobreak\ Q$. ###### Proof. By induction on $n$: * • $n=1$. By assumption, $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}$ and $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{1}}}}}$. The proof proceeds by considering all possible combinations of shapes for $\alpha$ and $\alpha^{\prime}_{1}$ (forwarder, output/input, and selection/branching). Consider the case where ${\alpha=x\rangle y:a}$ and $\alpha^{\prime}_{1}=w\rangle z:b$. Because $P$ is well-typed, we infer that there are evaluation contexts $E_{1}$ and $E_{2}$ such that $P\equiv E_{1}[(\bm{\nu}xy)(x[a,c]\mathbin{|}y(a,c).P_{1})]\equiv E_{2}[(\bm{\nu}wz)(w[b,d]\mathbin{|}z(b,d).P_{2}])$ (Definition 34). Since the reductions labeled $\alpha$ and $\alpha^{\prime}_{1}$ are both enabled in $P$, it cannot be the case that $x,y\in\mathrm{fn}(P_{2})$ and $w,z\in\mathrm{fn}(P_{1})$. Hence, there exists an evaluation context $E_{3}$ such that $P\equiv E_{3}[(\bm{\nu}xy)(x[a,c]\mathbin{|}y(a,c).P_{1})\mathbin{|}(\bm{\nu}wz)(w[b,d]\mathbin{|}z(b,d).P_{2}])$. Then $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q_{1}\equiv E_{3}[P_{1}\mathbin{|}(\bm{\nu}wz)(w[b,d]\mathbin{|}z(b,d).P_{2}])$ and $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{1}}}}}Q_{2}\equiv E_{3}[(\bm{\nu}xy)(x[a,c]\mathbin{|}y(a,c).P_{1})\mathbin{|}P_{2}]$. Let $Q=E_{3}[P_{1}\mathbin{|}P_{2}]$; then $Q_{1}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{1}}}}}Q$ and $Q_{2}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q$. Hence, $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{1}}}}}Q$ and $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{1}}}}}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q$. All other cases proceed similarly. Note that when one of the reductions (say, $\alpha$) has a selection/branching label, such a reduction would discard some branches and thus possible behaviors. This is not an issue for establishing the thesis, because typability guarantees that the sub-process that enables the $\alpha^{\prime}$-labeled reduction does not appear under the to-be- discarded branches. Hence, the execution of $\alpha$ will not jeopardize the $\alpha^{\prime}$-labeled reduction. * • $n=n^{\prime}+1$ for $n^{\prime}\geq 1$. By the IH, ${P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{1}}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{n^{\prime}}}}}}Q^{\prime}_{1}}$ and $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{1}}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{n^{\prime}}}}}}P^{\prime}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q^{\prime}_{1}$. By assumption, $P^{\prime}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{n}}}}}Q^{\prime}_{2}$. Since $P$ is well-typed, by Theorem 2 (Subject Reduction), $P^{\prime}$ is well-typed. Since $P^{\prime}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q^{\prime}_{1}$ and $P^{\prime}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{n}}}}}Q^{\prime}_{2}$, we can follow the same argumentation as in the base case to show that $Q^{\prime}_{1}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{n}}}}}Q$ and $Q^{\prime}_{2}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q$. Hence, $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{1}}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{n^{\prime}}}}}}Q^{\prime}_{1}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{n}}}}}Q$ and $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{1}}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{n^{\prime}}}}}}P^{\prime}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}_{n}}}}}Q^{\prime}_{2}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha}}}}Q$. ∎ To understand the proof of soundness and the rôle of independent reductions therein, consider the following example. We first introduce some notation which we also use in the proof of soundness: given an ordered sequence of reduction labels $A=(\alpha_{1},\ldots,\alpha_{k})$, we write $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A}}}}Q$ to denote $P\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha_{1}}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha_{k}}}}}Q$. ###### Example 9. The recursive global type $G=\mu X.a\mathbin{\twoheadrightarrow}b:1\langle S\rangle.c\mathbin{\twoheadrightarrow}d:1\langle S\rangle.X$ features two independent exchanges. Consider a network $\mathcal{N}\in\mathrm{net}(G)$. Let $A$ denote the sequence of labeled reductions necessary to complete the exchange in $G$ between $a$ and $b$, and $C$ similarly for the exchange between $c$ and $d$. Assuming that communication with routers is not blocked by implementation processes, we have $\mathcal{N}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A}}}}$ and $\mathcal{N}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}C}}}}$, because the exchanges are independent. Now, suppose that from $\mathcal{N}^{\circlearrowright}$ we observe $m$ times the sequence of $C$ reductions: $\mathcal{N}^{\circlearrowright}\underbrace{\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}C}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}C}}}}}_{\text{$m$ times}}N^{\prime}$. We see that $N^{\prime}$ is not a network of a global type reachable from $G$: there are still $m$ exchanges between $a$ and $b$ pending. Still, we can exhibit a series of transitions from $G$ that includes $m$ times the exchange between $c$ and $d$: $G\underbrace{\xrightarrow{\vspace{-.8ex}a\rangle b:1\langle S\rangle}\xrightarrow{\vspace{-.8ex}c\rangle d:1\langle S\rangle}\ldots\xrightarrow{\vspace{-.8ex}a\rangle b:1\langle S\rangle}\xrightarrow{\vspace{-.8ex}c\rangle d:1\langle S\rangle}}_{\text{$m$ times}}G$ Following these transitions, we can exhibit a corresponding sequence of reductions from $\mathcal{N}^{\circlearrowright}$ that includes $m$ times the sequence $C$ and ends up in another network $\mathcal{M}\in\mathrm{net}(G)$: $\mathcal{N}^{\circlearrowright}\underbrace{\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A}}}}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}C}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A}}}}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}C}}}}}_{\text{$m$ times}}\mathcal{M}^{\circlearrowright}$ At this point it is crucial that from $\mathcal{N}^{\circlearrowright}$ the sequences of reductions $A$ and $C$ can be performed independently. Hence, by Proposition 22, $\mathcal{N}^{\circlearrowright}\underbrace{\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}C}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}C}}}}}_{\text{$m$ times}}N^{\prime}\underbrace{\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A}}}}}_{\text{$m$ times}}\mathcal{M}^{\circlearrowright}$. In the proof of soundness, whenever we assure that certain reductions are independent, we refer to those assurances as _independence facts_ (IFacts). Also, in the proof we consider labeled reductions, and distinguish between _protocol_ and _implementation_ reductions: the former are reductions with labels that indicate any interaction with a router, and the latter are any other reductions (which, by the definition of networks, can only occur within participant implementation processes). By a slight abuse of notation, given ordered sequences of reduction labels $A$ and $A^{\prime}$, we write $A^{\prime}\subseteq A$ to denote that $A^{\prime}$ is a subsequence of $A$, where the labels in $A^{\prime}$ appear in the same order in $A$ but not necessarily in sequence (and similarly for $A^{\prime}\subset A$). With $A\setminus A^{\prime}$ we denote the sequence obtained from $A$ by removing all the labels in $A^{\prime}$, and $A\cup A^{\prime}$ denotes the sequence obtained by adding the labels from $A^{\prime}$ to the end of $A$. ###### Theorem 23 (Operational Correspondence: Soundness). Suppose given a relative well-formed global type $G$ and a completable $\mathcal{N}\in\mathrm{net}(G)$. For every ordered sequence of $k\geq 0$ reduction labels $A=(\alpha_{1},\ldots,\alpha_{k})$ and $N^{\prime}$ such that $\mathcal{N}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A}}}}N^{\prime}$, there exist $G^{\prime}$ and $\beta_{1},\ldots,\beta_{n}$ (with $n\geq 0$) such that (i) $G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{n}}G^{\prime}$ and (ii) $N^{\prime}\longrightarrow^{\ast}\mathcal{M}^{\circlearrowright}$, with $\mathcal{M}\in\mathrm{net}(G^{\prime})$. ###### Proof. By induction on the structure of $G$; we detail the interesting cases of labeled exchanges with implicitly unfolded recursive definitions. We exhibit transitions $G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{n}}G^{\prime}$ and establish a corresponding sequence of reductions $\mathcal{N}^{\circlearrowright}\longrightarrow^{\ast}\mathcal{M}^{\circlearrowright}$ that includes all the labels in $A$, with $\mathcal{M}\in\mathrm{net}(G^{\prime})$. During this step, we assure the independence between the observed reductions $A$ and the reductions we establish (IFacts). Using these independence assurances, we show that also $N^{\prime}\longrightarrow^{\ast}\mathcal{M}^{\circlearrowright}$. We apply induction on the size of $A$ (IH1) to show the existence of (i) $G^{\prime}$ and $\beta_{1},\ldots,\beta_{n}$ such that (i) $G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{n}}G^{\prime}$ and (ii) $\mathcal{N}^{\circlearrowright}\longrightarrow^{\ast}\mathcal{M}^{\circlearrowright}$ including all reductions in $A$, with $\mathcal{M}\in\mathrm{net}(G^{\prime})$: * • Base case: then $A$ is empty, and the thesis holds trivially, with $G^{\prime}=G$ and $\mathcal{M}=\mathcal{N}$. * • Inductive case: then $A$ is non-empty. By the definition of networks (Definition 25), we know that reductions starting at $\mathcal{N}^{\circlearrowright}$ are protocol reductions related to an independent exchange in $G$, or implementation reductions. Every protocol reduction in $A$ is related to some exchange in $G$, and so we can group sequences of protocol reductions related to the same exchange. By construction, every such sequence of protocol reductions $A_{\ast}\subseteq A$ starts with an implementation sending a label to a router, i.e., with a label of the form $\alpha_{\ast}={{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}:\ell$. For each such $\alpha_{\ast}$, the router in $\mathcal{N}$ of the sender $c$ has been synthesized from $G$ in a finite number of inductive steps. We take the $\alpha_{\ast}$ that originates from the router synthesized in the least number of steps. This gives us the $A_{\ast}$ starting with $\alpha_{\ast}$ that relates to an exchange in $G$ which is not prefixed by exchanges relating to any of the other $A^{\prime}_{\ast}\subseteq A\setminus A_{\ast}$. Networks are well-typed by definition. None of the reductions in $A_{\ast}$ are blocked by protocol reductions appearing earlier in $A$ (IFact 1): they originate from exchanges in $G$ appearing after the exchange related to $A_{\ast}$, and the priorities in their related types are thus higher than those in the types related to $A_{\ast}$, i.e., blocking by input or branching would contradict the well-typedness of $\mathcal{N}^{\circlearrowright}$. However, it may be that some implementation reductions $A_{+}\subseteq A\setminus A_{\ast}$ do block the reductions in $A_{\ast}$; they are also not blocked by any prior protocol reductions due to priorities (IFact 2). Hence, from $\mathcal{N}^{\circlearrowright}$ we can perform the implementation reductions in $A_{+}$. By Subject Reduction (Theorem 2), this results in another completed network $\mathcal{N}_{0}^{\circlearrowright}$ of $G$. This establishes the reduction sequence $\mathcal{N}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A_{+}}}}}\mathcal{N}_{0}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A_{\ast}}}}}$. By Lemma 21, there are $m\geq 0$ transitions $G\xrightarrow{\vspace{-.8ex}\beta_{1}}G_{1}\ldots\xrightarrow{\vspace{-.8ex}\beta_{m}}G_{m}$ where the initial prefix of $G_{m}$ corresponds to the labeled choice by the implementation of $c$: $G_{m}=c\mathbin{\twoheadrightarrow}d\\{i\langle S_{i}\rangle.G^{\prime}_{i}\\}_{i\in I}$, with $\ell\in I$. Additionally, $G_{m}$ contains exchanges related to every sequence of protocol reductions in $A\setminus A_{+}\setminus A_{\ast}$: all these sequences start with a selection from implementation to router, and thus the involved participants do not depend on any of the exchanges between $G$ and $G_{m}$, such that Lemma 20 applies. To establish a sequence of reductions from $\mathcal{N}_{0}^{\circlearrowright}$ to the completion of a network $\mathcal{N}_{m}\in\mathrm{net}(G_{m})$, we apply induction on $m$ (IH2): * – The base case where $m=0$ is trivial, with $G_{m}=G$ and thus $\mathcal{N}_{m}^{\circlearrowright}=\mathcal{N}_{0}^{\circlearrowright}$. * – In the inductive case, following the same approach as in the proof of completeness (Theorem 19), we reduce $\mathcal{N}_{0}^{\circlearrowright}\longrightarrow^{\ast}\mathcal{N}_{1}^{\circlearrowright}$ such that $\mathcal{N}_{1}\in\mathrm{net}(G_{1})$. Then, by IH2, $\mathcal{N}_{1}^{\circlearrowright}\longrightarrow^{\ast}\mathcal{N}_{m}^{\circlearrowright}$ where $\mathcal{N}_{m}\in\mathrm{net}(G_{m})$. Note that these reductions may require implementation reductions to unblock protocol reductions, and these implementation reductions may appear in $A$. None of the reductions from $\mathcal{N}_{0}^{\circlearrowright}$ to $\mathcal{N}_{m}^{\circlearrowright}$ can be blocked by any of the other protocol reductions in $A$, following again from priorities in types; hence, the leftover reductions in $A$ are independent from these reductions (IFact 3). Additionally, the sequence of protocol reductions $A_{\ast}$ was already enabled from $\mathcal{N}_{0}^{\circlearrowright}$, so those reductions are also independent (IFact 4). We know that $\mathcal{N}_{m}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\rangle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}:\ell}}}}$ and $G_{m}\xrightarrow{\vspace{-.8ex}c\rangle d:\ell\langle S_{\ell}\rangle}G^{\prime}_{\ell}$. From $\mathcal{N}_{m}^{\circlearrowright}$, we again follow the proof of completeness to show that $\mathcal{N}_{m}^{\circlearrowright}\longrightarrow^{\ast}\mathcal{M}_{\ell}^{\circlearrowright}$, where $\mathcal{M}_{\ell}\in\mathrm{net}(G^{\prime}_{\ell})$. Given the definition of routers, it must be that all the reductions in $A_{\ast}$ appear in this sequence of reductions. Let $A^{\prime}\subset A$ denote the leftover reductions from $A$ (i.e., $A$ except all reductions that occurred between $\mathcal{N}^{\circlearrowright}$ and $\mathcal{M}_{\ell}^{\circlearrowright}$, including $A_{\ast}$ and $A_{+}$). By IFacts 1–4, $\mathcal{M}_{\ell}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A^{\prime}}}}}M^{\prime}$. Then by IH1, there exist $G^{\prime}$ and $\beta_{m+2},\ldots,\beta_{n}$ (with $n\geq m+1$) such that (i) $G^{\prime}_{\ell}\xrightarrow{\vspace{-.8ex}\beta_{m+2}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{n}}G^{\prime}$ and (ii) $\mathcal{M}_{\ell}^{\circlearrowright}\longrightarrow^{\ast}\mathcal{M}^{\circlearrowright}$ including all reductions in $A^{\prime}$, with $\mathcal{M}\in\mathrm{net}(G^{\prime})$. Let $\beta_{m+1}=c\rangle d:\ell\langle S_{\ell}\rangle$. We have shown the existence of $G^{\prime}$ and $\beta_{1},\ldots,\beta_{n}$ such that (i) $G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{m}}G_{m}\xrightarrow{\vspace{-.8ex}\beta_{m+1}}G^{\prime}_{\ell}\xrightarrow{\vspace{-.8ex}\beta_{m+2}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{n}}G^{\prime}$ and (ii) $\mathcal{N}^{\circlearrowright}\longrightarrow^{\ast}\mathcal{N}_{m}^{\circlearrowright}\longrightarrow^{\ast}\mathcal{M}_{\ell}^{\circlearrowright}\longrightarrow^{\ast}\mathcal{M}^{\circlearrowright}$ including all reductions in $A$, with $\mathcal{M}\in\mathrm{net}(G^{\prime})$. We are left to show that from $\mathcal{N}^{\circlearrowright}\longrightarrow^{\ast}\mathcal{M}^{\circlearrowright}$ and $\mathcal{N}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A}}}}N^{\prime}$, we can conclude that $N^{\prime}\longrightarrow^{\ast}\mathcal{M}^{\circlearrowright}$. We apply induction on the size of $A$ (IH3), using IFacts 1–4 and Proposition 22: * • Base case: Then $A$ is empty, there is nothing to do, and the thesis is proven. * • Inductive case: Then $A=A^{\prime}\cup(\alpha^{\prime})$. By IH3, $\mathcal{N}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A^{\prime}}}}}N^{\prime\prime}\longrightarrow^{\ast}N^{\prime\prime\prime}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}}}}}\longrightarrow^{\ast}\mathcal{M}^{\circlearrowright}$. Moreover, by assumption, $N^{\prime\prime}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}}}}}N^{\prime}$. IFacts 1–4 show that the $\alpha^{\prime}$-labeled reduction is independent from the reductions between $N^{\prime\prime}$ and $N^{\prime\prime\prime}$. Hence, by Proposition 22, we have $\mathcal{N}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A^{\prime}}}}}N^{\prime\prime}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\alpha^{\prime}}}}}N^{\prime}\longrightarrow^{\ast}\mathcal{M}^{\circlearrowright}$. That is, $\mathcal{N}^{\circlearrowright}\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A}}}}N^{\prime}\longrightarrow^{\ast}\mathcal{M}^{\circlearrowright}$, proving the thesis. ∎ In the light of Theorem 23, let us revisit Example 9: ###### Example 10 (Revisiting Example 9). Recall the global type $G=\mu X.a\mathbin{\twoheadrightarrow}b:1\langle S\rangle.c\mathbin{\twoheadrightarrow}d:1\langle S\rangle.X$ from Example 9, with two independent exchanges. We take some $\mathcal{N}\in\mathrm{net}(G)$ such that $\mathcal{N}^{\circlearrowright}\underbrace{\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}C}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}C}}}}}_{\text{$m$ times}}N^{\prime}$, where $C$ denotes the sequence of reduction labels corresponding to the exchange between $c$ and $d$. By Theorem 23, there indeed are $G^{\prime}$ and $\beta_{1},\ldots,\beta_{n}$ such that $G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{n}}G^{\prime}$ and $N^{\prime}\longrightarrow^{\ast}\mathcal{M}^{\circlearrowright}$, with $\mathcal{M}\in\mathrm{net}(G^{\prime})$. To be precise, following Theorem 23, indeed $G\underbrace{\xrightarrow{\vspace{-.8ex}a\rangle b:1\langle S\rangle}\xrightarrow{\vspace{-.8ex}c\rangle d:1\langle S\rangle}\ldots\xrightarrow{\vspace{-.8ex}a\rangle b:1\langle S\rangle}\xrightarrow{\vspace{-.8ex}c\rangle d:1\langle S\rangle}}_{\text{$m$ times}}G\quad\text{ and }\quad\mathcal{N}^{\circlearrowright}\underbrace{\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}C}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}C}}}}}_{\text{$m$ times}}N^{\prime}\underbrace{\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A}}}}\ldots\mathbin{{\color[rgb]{0.7578125,0.38671875,0.89453125}\definecolor[named]{pgfstrokecolor}{rgb}{0.7578125,0.38671875,0.89453125}\xrightharpoondown{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}A}}}}}_{\text{$m$ times}}\mathcal{M}^{\circlearrowright}$ where $A$ is the sequence of reduction labels corresponding to the exchange between $a$ and $b$ and ${\mathcal{M}\in\mathrm{net}(G)}$. ### 4.4 Routers Strictly Generalize Centralized Orchestrators Unlike our decentralized analysis, previous analyses of global types using binary session types rely on centralized orchestrators (called mediums [12] or arbiters [16]). Here, we show that our approach strictly generalizes these centralized approaches. Readers interested in our decentralized approach in action may safely skip this section and go directly to Section 5. We introduce an algorithm that synthesizes an orchestrator—a single process that orchestrates the interactions between a protocol’s participants (§ 4.4.1). We show that the composition of this orchestrator with a context of participant implementations is behaviorally equivalent to the specific case in which routed implementations are organized in a _centralized composition_ (Theorem 27 in § 4.4.2). #### 4.4.1 Synthesis of Orchestrators 1 def _${\mathsf{O}}_{\tilde{q}}[G]$_ as 2 switch _$G$_ do 3 case _$s\mathbin{\twoheadrightarrow}r\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}$_ do 4 $\mathsf{deps}:=\\{q\in\tilde{q}\mid\mathrm{hdep}(q,s,G)\vee\mathrm{hdep}(q,r,G)\\}$ 5 return ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}\triangleright\\{i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}}\triangleleft i\cdot\underline{{(\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}}\triangleleft i)}_{q\in\mathsf{deps}}}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}(v)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}}[w]\cdot(v\mathbin{\leftrightarrow}w\mathbin{|}{\mathsf{O}}_{\tilde{q}}[G_{i}])\\}_{i\in I}$ 6 7 case _$\mu X\mathbin{.}G^{\prime}$_ do 8 $\tilde{q}^{\prime}:=\\{q\in\tilde{q}\mid G\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}q\neq\bullet\\}$ 9 if _$\tilde{q}^{\prime}\neq\emptyset$_ then return $\mu X({({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}})}_{q\in\tilde{q}^{\prime}})\mathbin{.}{\mathsf{O}}_{\tilde{q}^{\prime}}[G^{\prime}]$ 10 else return $\bm{0}$ 11 12 13 case _$X$_ do return $X{\langle{({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}})}_{q\in\tilde{q}}\rangle}$ 14 15 case _$\mathsf{skip}\mathbin{.}G^{\prime}$_ do return ${\mathsf{O}}_{\tilde{q}}[G^{\prime}]$ 16 17 case _$\mathsf{end}$_ do return $\bm{0}$ 18 Algorithm 2 Synthesis of Orchestrator Processes (Def. 36). We define the synthesis of an orchestrator from a global type. The orchestrator of $G$ will have a channel endpoint ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p_{i}}}$ for connecting to the process implementation of every $p_{i}\in\mathsf{prt}(G)$. ###### Definition 36 (Orchestrator). Given a global type $G$ and participants $\tilde{q}$, Algorithm 2 defines the synthesis of an _orchestrator process_ , denoted ‘$\mkern 1.0mu{\mathsf{O}}_{\tilde{q}}[G]\mkern-3.0mu$’, that orchestrates interactions according to $G$. Algorithm 2 follows a similar structure as the router synthesis algorithm (Algorithm 1). The input parameter ‘$\tilde{q}$’ keeps track of active participants, making sure recursions are well-defined; it should be initialized as ‘$\mathsf{prt}(G)$’. We briefly discuss how the orchestrator process is generated. The interesting case is an exchange ‘$p\mathbin{\twoheadrightarrow}q\\{i\langle U_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}$’ (algorithm 2), where the algorithm combines the several cases of the router’s algorithm (that depend on the involvement of the router’s participant). First, the sets of participants ‘$\mathsf{deps}$’ that depend on the sender and on the recipient are computed (algorithm 2) using the auxiliary predicate ‘$\mathrm{hdep}$’ (cf. Def. 18). Then, the algorithm returns a process (algorithm 2) that receives a label $i\in I$ over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}$; forwards it over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}$ and over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ for all $q\in\mathsf{deps}$; receives a channel over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}$; forwards it over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}$; and continues as ‘${\mathsf{O}}_{\tilde{q}}[G_{i}]$’. The synthesis of a recursive definition ‘$\mu X\mathbin{.}G^{\prime}$’ (algorithm 2) requires care, as the set of active participants $\tilde{q}$ may change. In order to decide which $q\in\tilde{q}$ are active in $G^{\prime}$, the algorithm computes the local projection of $G$ onto each $q\in\tilde{q}$ to determine the orchestrator’s future behavior on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$, creating a new set $\tilde{q}^{\prime}$ with those $q\in\tilde{q}$ for which the projection is different from ‘$\bullet$’ (algorithm 2). Then, the algorithm returns a recursive process with as context the channel endpoints ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ for $q\in\tilde{q}^{\prime}$, with ‘${\mathsf{O}}_{\tilde{q}^{\prime}}[G^{\prime}]$’ as the body. The synthesis of a recursive call ‘$X$’ (algorithm 2) yields a recursive call with as context the channels ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ for $q\in\tilde{q}$. Finally, for ‘$\mathsf{skip}\mathbin{.}G^{\prime}$’ (algorithm 2) the algorithm returns the orchestrator for $G^{\prime}$, and for ‘$\bullet$’ (algorithm 2) the algorithm returns ‘$\bm{0}$’. There is a minor difference between the orchestrators synthesized by Algorithm 2 and the mediums defined by Caires and Pérez [12]. The difference is in the underlined portion in algorithm 2, which denotes explicit messages (obtained via dependency detection) needed to deal with non-local choices. The mediums by Caires and Pérez do not include such communications, as their typability is based on local types, which rely on a merge operation at projection time. The explicit actions in algorithm 2 make the orchestrator compatible with participant implementations that connect with routers. Aside from these actions, our concept of orchestrator is essentially the same as that of the mediums by Caires and Pérez. Crucially, orchestrators can be typed using local projection (cf. Def. 22) similar to the typing of routers using relative projection (cf. Theorem 11). This result follows by construction: ###### Theorem 24. Given a closed, relative well-formed global type $G$, $\displaystyle{\mathsf{O}}_{\mathsf{prt}(G)}[G]\vdash\emptyset;{({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{(G\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}p)})}_{p\in\mathsf{prt}(G)}.$ ###### Proof. We prove a more general statement. Suppose given a closed, relative well- formed global type $G$. Also, suppose given a global type $G_{s}\leq_{C}G$. Consider: * • the participants that are active in $G_{s}$: $\tilde{q}=\\{q\in\mathsf{prt}(G)\mid\exists p\in\mathsf{prt}(G).\leavevmode\nobreak\ (p,q)\in\mathrm{active}(C,G)\\}$, * • the absolute priority of $G_{s}$: $\mathsf{o}_{C}=\mathrm{ctxpri}(C)$, * • the sequence of bound recursion variables of $G_{s}$: $\widetilde{X_{C}}=\mathrm{ctxbind}(C)$, * • for every $X\in\widetilde{X_{C}}$: * – the body of the recursive definition on $X$ in $G$: $G_{X}=\mathrm{recdef}(X,G)$, * – the participants that are active in $G_{X}$: $\tilde{q}_{X}=\\{q\in\mathsf{prt}(G)\mid\exists p\in\mathsf{prt}(G).\leavevmode\nobreak\ (p,q)\in\mathrm{recactive}(X,G)\\}$, * – the absolute priority of $G_{X}$: $\mathsf{o}_{X}=\mathrm{varpri}(X,G)$, * – the sequence of bound recursion variables of $G_{X}$ excluding $X$: $\widetilde{Y_{X}}=\mathrm{subbind}(\mu X\mathbin{.}G_{X},G)$, * – the type required for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ for a recursive call on $X$: $\displaystyle A_{X,q}=\mathrm{deepUnfold}(\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q},{(Y,t_{Y},\overline{G_{Y}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Y}}q})}_{Y\in\widetilde{Y_{X}}})$ * – the minimum lift for typing a recursive definition on $X$: $t_{X}=\max_{\mathsf{pr}}\left({(A_{X,q})}_{q\in\tilde{q}_{X}}\right)+1$, * • the type expected for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ for the orchestrator for $G_{s}$: $\displaystyle D_{q}=\mathrm{deepUnfold}(\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}q},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q})}_{X\in\widetilde{X_{C}}}).$ Then, we have: $\displaystyle{\mathsf{O}}_{\tilde{q}}[G_{s}]\vdash{\left(X{:}\leavevmode\nobreak\ {\big{(}A_{X,q}\big{)}}_{q\in\tilde{q}_{X}}\right)}_{X\in\widetilde{X_{C}}};\leavevmode\nobreak\ {\left({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ D_{q}\right)}_{q\in\tilde{q}}$ Similar to how Theorem 11 follows from Theorem 16, the thesis follows as a corollary from this more general statement (cf. the proof of Theorem 11 on Section 4.3.1). We apply induction on the structure of $G_{s}$, with six cases as in Algorithm 2. We only detail the cases of exchange and recursion. * • _Exchange_ : $G_{s}=s\mathbin{\twoheadrightarrow}r\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}$ (algorithm 2). Following similar reasoning as in the case for exchange in the proof of Theorem 16, we can omit the unfoldings on types, as well as the recursive context. Let $\mathsf{deps}_{s}:=\\{q\in\tilde{q}\mid\mathrm{hdep}(q,s,G_{s})\\}$ and $\mathsf{deps}_{r}:=\\{q\in\tilde{q}\setminus\mathsf{deps}_{s}\mid\mathrm{hdep}(q,r,G_{s})\\}$. Note that $\mathsf{deps}_{s}\cup\mathsf{deps}_{r}$ coincides with $\mathsf{deps}$ as defined on algorithm 2 and that $s,r\notin\mathsf{deps}_{s}\cup\mathsf{deps}_{r}$. Let us take stock of the types we expect for each of the orchestrator’s channels. For ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}$ we expect $\displaystyle\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}s}$ $\displaystyle=\overline{{{\oplus}}^{\mathsf{o}}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}+1}(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}s)\\}_{i\in I}}$ $\displaystyle=\&^{\mathsf{o}}\\{i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}+1}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}s)}\\}_{i\in I}.$ (34) For ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}$ we expect $\displaystyle\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}r}$ $\displaystyle=\overline{{\&}^{\mathsf{o}+2}\\{i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}+3}(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}r)\\}_{i\in I}}$ $\displaystyle={{\oplus}}^{\mathsf{o}+2}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}+3}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}r)}\\}_{i\in I}.$ (35) For each $q\in\mathsf{deps}_{s}$, for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ we expect $\displaystyle\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}q}$ $\displaystyle=\overline{\&^{\mathsf{o}+2}\\{i{:}\leavevmode\nobreak\ (G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q)\\}_{i\in I}}$ $\displaystyle={{\oplus}}^{\mathsf{o}+2}\\{i{:}\leavevmode\nobreak\ \overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q)}\\}_{i\in I}.$ (36) For each $q\in\mathsf{deps}_{r}$, for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ we expect $\displaystyle\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}q}$ $\displaystyle=\overline{\&^{\mathsf{o}+3}\\{i{:}\leavevmode\nobreak\ (G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q)\\}_{i\in I}}$ $\displaystyle={{\oplus}}^{\mathsf{o}+3}\\{i{:}\leavevmode\nobreak\ \overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q)}\\}_{i\in I}.$ (37) For each $q\in\tilde{q}\setminus\mathsf{deps}_{s}\setminus\mathsf{deps}_{r}\setminus\\{s,r\\}$, for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ we expect $\displaystyle\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}q}$ $\displaystyle=\overline{G_{i^{\prime}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q}\leavevmode\nobreak\ \text{for any $i^{\prime}\in I$}.$ (38) Let us now consider the process returned by Algorithm 1, with each prefix marked with a number. ${\mathsf{O}}_{\tilde{q}}[G]=\underbrace{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}\triangleright\\{i{:}}_{1}\leavevmode\nobreak\ \underbrace{\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}}\triangleleft i}_{2_{i}}\cdot\underbrace{{(\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}}\triangleleft i)}_{q\in\mathsf{deps}}}_{3_{i}}\cdot\underbrace{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}(v)}_{4_{i}}\mathbin{.}\underbrace{\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}}[w]}_{5_{i}}\cdot(v\mathbin{\leftrightarrow}w\mathbin{|}{\mathsf{O}}_{\tilde{q}}[G_{i}])\\}_{i\in I}$ For each $i^{\prime}\in I$, let $C_{i^{\prime}}:=C[s\mathbin{\twoheadrightarrow}r(\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I\setminus\\{i^{\prime}\\}}\cup\\{i^{\prime}\langle S_{i^{\prime}}\rangle\mathbin{.}[]\\})]$. Clearly, $G_{i^{\prime}}\leq_{C_{i^{\prime}}}G$. Also, because we are not adding recursion binders, the current value of $\tilde{q}$ is appropriate for the IH. With $C_{i^{\prime}}$ and $\tilde{q}$, we apply the IH to obtain the typing of ${\mathsf{O}}_{\tilde{q}}[G_{i^{\prime}}]$, where priorities start at $\mathrm{ctxpri}(C_{i^{\prime}})=\mathrm{ctxpri}(C)+4$ (cf. Def. 31). Following these typings, Figure 12 gives the typing of ${\mathsf{O}}_{\tilde{q}}[G_{s}]$, referring to parts of the process by the number marking its foremost prefix above. Clearly, the priorities in the derivation in Figure 12 meet all requirements. The order of the applications of ${\oplus}^{\star}$ for each $q\in\mathsf{deps}_{s}\cup\mathsf{deps}_{r}$ does not matter, since the selection actions are asynchronous. Id $\forall i\in I.\leavevmode\nobreak\ v\mathbin{\leftrightarrow}w\vdash\begin{array}[t]{@{}l@{}}v{:}\leavevmode\nobreak\ \overline{S_{i}},w{:}\leavevmode\nobreak\ S_{i}\end{array}$ $\forall i\in I.\leavevmode\nobreak\ {\mathsf{O}}_{\tilde{q}}[G_{i}]\vdash\begin{array}[t]{@{}l@{}}{({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ \overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q})}_{q\in\tilde{q}}\end{array}$ Mix $\forall i\in I.\leavevmode\nobreak\ v\mathbin{\leftrightarrow}w\mathbin{|}{\mathsf{O}}_{\tilde{q}}[G_{i}]\vdash\begin{array}[t]{@{}l@{}}v{:}\leavevmode\nobreak\ \overline{S_{i}},w{:}\leavevmode\nobreak\ S_{i},\\\ {({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ \overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}}q)}_{q\in\tilde{q}}\end{array}$ $\mathbin{\otimes}^{\star}$ $\forall i\in I.\leavevmode\nobreak\ 5_{i}\vdash\begin{array}[t]{@{}l@{}}v{:}\leavevmode\nobreak\ \overline{S_{i}},\\\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}+3}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}r)},\\\ {({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ \overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q})}_{q\in\tilde{q}\setminus\\{r\\}}\end{array}$ $\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}$ $\forall i\in I.\leavevmode\nobreak\ 4_{i}\vdash\begin{array}[t]{@{}l@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}+1}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}s)},\\\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}+3}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}r)},\\\ {({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ \overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q})}_{q\in\tilde{q}\setminus\\{s,r\\}}\end{array}$ $\forall q\in\mathsf{deps}_{s}\cup\mathsf{deps}_{r}.\leavevmode\nobreak\ {\oplus}^{\star}$ $\forall i\in I.\leavevmode\nobreak\ 3_{i}\vdash\begin{array}[t]{@{}l@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}+1}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}s)},\\\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}+3}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}r)},\\\ {({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}+2}\\{i{:}\leavevmode\nobreak\ \overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q)}\\}_{i\in I})}_{q\in\mathsf{deps}_{s}}\\\ {({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}+3}\\{i{:}\leavevmode\nobreak\ \overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q)}\\}_{i\in I})}_{q\in\mathsf{deps}_{r}}\\\ {({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ \overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q})}_{q\in\tilde{q}\setminus\mathsf{deps}_{s}\setminus\mathsf{deps}_{r}\setminus\\{s,r\\}}\end{array}$ ${\oplus}^{\star}$ $\forall i\in I.\leavevmode\nobreak\ 2_{i}\vdash\begin{array}[t]{@{}l@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}+1}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}s)},\\\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}+2}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}+3}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}r)}\\}_{i\in I},\\\ {({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}+2}\\{i{:}\leavevmode\nobreak\ \overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q)}\\}_{i\in I})}_{q\in\mathsf{deps}_{s}}\\\ {({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}+3}\\{i{:}\leavevmode\nobreak\ \overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q)}\\}_{i\in I})}_{q\in\mathsf{deps}_{r}}\\\ {({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ \overline{G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q})}_{q\in\tilde{q}\setminus\mathsf{deps}_{s}\setminus\mathsf{deps}_{r}\setminus\\{s,r\\}}\end{array}$ $\&$ ${\mathsf{O}}_{\tilde{q}}[G_{s}]=1\vdash\begin{array}[t]{@{}lr@{}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}{:}\leavevmode\nobreak\ \&^{\mathsf{o}}\\{i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}+1}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}s)}\\}_{i\in I},&\text{(cf.\ \eqref{eq:mSType})}\\\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}+2}\\{i{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S_{i}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{\mathsf{o}+3}\overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}r)}\\}_{i\in I},&\text{(cf.\ \eqref{eq:mRType})}\\\ {({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}+2}\\{i{:}\leavevmode\nobreak\ \overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q)}\\}_{i\in I})}_{q\in\mathsf{deps}_{s}}&\text{(cf.\ \eqref{eq:mDepSType})}\\\ {({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ {{\oplus}}^{\mathsf{o}+3}\\{i{:}\leavevmode\nobreak\ \overline{(G_{i}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q)}\\}_{i\in I})}_{q\in\mathsf{deps}_{r}}&\text{(cf.\ \eqref{eq:mDepRType})}\\\ {({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ \overline{G_{i^{\prime}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}+4}q})}_{q\in\tilde{q}\setminus\mathsf{deps}_{s}\setminus\mathsf{deps}_{r}\setminus\\{s,r\\}}&\text{(cf.\ \eqref{eq:mDepOtherType})}\end{array}$ Figure 12: Typing derivation used in the proof of Theorem 24. * • _Recursive definition_ : $G_{s}=\mu Z\mathbin{.}G^{\prime}$ (algorithm 2). Let $\displaystyle\tilde{q}^{\prime}:=\\{q\in\tilde{q}\mid G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}q\neq\bullet\\}$ (39) (as on algorithm 2). The analysis depends on whether $\tilde{q}^{\prime}=\emptyset$ or not. * – If $\tilde{q}^{\prime}=\emptyset$ (algorithm 2), let us take stock of the types expected for each of the orchestrator’s channels. For now, we omit the substitutions in the types. For each $q\in\tilde{q}$, for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ we expect $\displaystyle\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}q}$ $\displaystyle=\bullet.$ (40) Because all expected types are $\bullet$, the substitutions do not affect the types, so we can omit them altogether. First we apply Empty, giving us an arbitrary recursive context, thus the recursive context we need. Then, we apply $\bullet$ for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ for each $q\in\tilde{q}$ (cf. (40)), and obtain the typing of ${\mathsf{O}}_{\tilde{q}}[G_{s}]$ (omitting the recursive context): ${\mathsf{O}}_{\tilde{q}}[G_{s}]=\bm{0}\vdash{({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ \bullet)}_{q\in\tilde{q}}.$ * – If $\tilde{q}^{\prime}\neq\emptyset$ (algorithm 2), let us take stock of the types expected for each of the orchestrator’s channels. Note that, because of the recursive definition on $Z$ in $G_{s}$, there cannot be another recursive definition in the context $C$ capturing the recursion variable $Z$. Therefore, by Definition 29, $Z\notin\widetilde{X_{C}}$. For each $q\in\tilde{q}^{\prime}$, for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ we expect $\displaystyle\mathrm{deepUnfold}(\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}q},\ldots)$ $\displaystyle=\mathrm{deepUnfold}(\overline{\mu Z\mathbin{.}(G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}q)},\ldots)$ $\displaystyle=\mathrm{deepUnfold}(\mu Z\mathbin{.}\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}q},\ldots)$ $\displaystyle=\mu Z\mathbin{.}\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}q},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q})}_{X\in\widetilde{X_{C}}}).$ (41) For each $q\in\tilde{q}\setminus\tilde{q}^{\prime}$, for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ we expect $\displaystyle\mathrm{deepUnfold}(\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}q},\ldots)$ $\displaystyle=\mathrm{deepUnfold}(\bullet,\ldots)=\bullet.$ (42) We also need an assignment in the recursive context for every $X\in\widetilde{X_{C}}$, but not for $Z$. Let $C^{\prime}=C[\mu Z\mathbin{.}[]]$. Clearly, $G^{\prime}\leq_{C^{\prime}}G$. Let us establish some facts about the recursion binders, priorities, and active participants related to $C^{\prime}$, $G^{\prime}$, and $Z$: * * $\widetilde{X_{C^{\prime}}}=\mathrm{ctxbind}(C^{\prime})=(\mathrm{ctxbind}(C),Z)=(\widetilde{X_{C}},Z)$ (cf. Def. 29). * * $G_{Z}=\mathrm{recdef}(Z,G)=G^{\prime}$, as proven by the context $C^{\prime}$ (cf. Def. 30). * * $\widetilde{Y_{Z}}=\mathrm{subbind}(\mu Z\mathbin{.}G_{Z},G)=\mathrm{ctxbind}(C)=\widetilde{X_{C}}$. * * $\mathsf{o}_{C^{\prime}}=\mathrm{ctxpri}(C^{\prime})=\mathrm{ctxpri}(C)=\mathsf{o}_{C}$, and $\mathsf{o}_{Z}=\mathrm{varpri}(Z,G)=\mathrm{ctxpri}^{C}(=)\mathsf{o}_{C}$, and hence $\mathsf{o}_{C^{\prime}}=\mathsf{o}_{Z}$ (cf. Def. 31). * * $\tilde{q}_{Z}=\tilde{q}^{\prime}$ (cf. Def. 32 and (39)). Because $\widetilde{X_{C^{\prime}}}=(\widetilde{X_{C}},Z)$ and $\tilde{q}^{\prime}=\tilde{q}_{Z}$, $\tilde{q}^{\prime}$ is appropriate for the IH. We apply the IH on $C^{\prime}$, $G^{\prime}$, and $\tilde{q}^{\prime}$ to obtain a typing for ${\mathsf{O}}_{\tilde{q}^{\prime}}[G^{\prime}]$, where we immediately make use of the facts established above. We given the assignment to $Z$ in the recursive context separate from those for the recursion variables in $\widetilde{X_{C}}$. Also, by Proposition 15, we can write the final unfolding on $Z$ in the types separately. $\displaystyle{\mathsf{O}}_{\tilde{q}^{\prime}}[G^{\prime}]\vdash\begin{array}[t]{@{}l@{}}{\left(X{:}\leavevmode\nobreak\ {\big{(}\mathrm{deepUnfold}(\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q},{(Y,t_{Y},\overline{G_{Y}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Y}}q})}_{Y\in\widetilde{Y_{X}}})\big{)}}_{q\in\tilde{q}_{X}}\right)}_{X\in\widetilde{X_{C}}},\\\\[6.0pt] Z{:}\leavevmode\nobreak\ {\big{(}\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}q},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q})}_{X\in\widetilde{X_{C}}})\big{)}}_{q\in\tilde{q}^{\prime}};\\\\[6.0pt] {\left({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ \mathrm{unfold}^{t_{Z}}(\mu Z\mathbin{.}\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}q},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q})}_{X\in\widetilde{X_{C}}}))\right)}_{q\in\tilde{q}^{\prime}}\end{array}$ By assumption, we have $\displaystyle t_{Z}$ $\displaystyle=\max_{\mathsf{pr}}{\left(\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}q},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q})}_{X\in\widetilde{X_{C}}})\right)}_{q\in\tilde{q}^{\prime}}+1,$ so $t_{Z}$ is clearly bigger than the maximum priority appearing in the types before unfolding. Hence, we can apply Rec to eliminate $Z$ from the recursive context, and to fold the types, giving the typing of ${\mathsf{O}}_{\tilde{q}}[G_{s}]=\mu Z({({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}})}_{q\in\tilde{q}^{\prime}})\mathbin{.}{\mathsf{O}}_{\tilde{q}^{\prime}}[G^{\prime}]$: $\displaystyle{\mathsf{O}}_{\tilde{q}}[G_{s}]\vdash\begin{array}[t]{@{}l@{}}{\left(X{:}\leavevmode\nobreak\ {\big{(}\mathrm{deepUnfold}(\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q},{(Y,t_{Y},\overline{G_{Y}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Y}}q})}_{Y\in\widetilde{Y_{X}}})\big{)}}_{q\in\tilde{q}_{X}}\right)}_{X\in\widetilde{X_{C}}};\\\\[6.0pt] {\left({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ \mu Z\mathbin{.}\mathrm{deepUnfold}(\overline{G^{\prime}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}q},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q})}_{X\in\widetilde{X_{C}}})\right)}_{q\in\tilde{q}^{\prime}}\end{array}$ In this typing, the type for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ for every $q\in\tilde{q}^{\prime}$ concurs with (41). For every $q\in\tilde{q}\setminus\tilde{q}^{\prime}$, we can add the type for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ in (42) by applying $\bullet$. This proves the thesis. * • _Recursive call_ : $G_{s}=Z$ (algorithm 2). Following similar reasoning as in the case of recursive call in the proof of Theorem 16, let us take stock of the types we expect for our orchestrator’s channels. For each $q\in\tilde{q}$, for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ we expect $\displaystyle\mathrm{deepUnfold}(\overline{G_{s}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}q},\ldots)$ $\displaystyle=\mathrm{deepUnfold}(\overline{Z\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{C}}q},\ldots)$ $\displaystyle=\mathrm{deepUnfold}(\overline{Z},\ldots)$ $\displaystyle=\mathrm{deepUnfold}(Z,{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q})}_{X\in(\tilde{X}_{1},Z,\widetilde{Y_{Z}})})$ $\displaystyle=\mathrm{deepUnfold}(Z,{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q})}_{X\in(Z,\widetilde{Y_{Z}})})$ $\displaystyle=\mu Z\mathbin{.}({\uparrow^{t_{Z}}}\mathrm{deepUnfold}(\overline{G_{Z}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Z}}q},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q})}_{X\in\widetilde{Y_{Z}}}))$ (43) Also, we need an assignment in the recursive context for every $X\in\widetilde{X_{C}}$. By Lemma 13, $\tilde{q}=\tilde{q}_{Z}$. Hence, for $Z$, the assignment should be as follows: $\displaystyle Z{:}\leavevmode\nobreak\ {\left(\mathrm{deepUnfold}(\overline{G_{Z}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Z}}q},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q)}_{X\in\widetilde{Y_{Z}}}})\right)}_{q\in\tilde{q}}$ (44) We apply Var to obtain the typing of ${\mathsf{O}}_{\tilde{q}}[G_{s}]$, where we make us the rule’s allowance for an arbitrary recursive context up to the assignment to $Z$. Var is applicable, because the types are recursive definitions on $Z$, concurring with the types assigned to $Z$, and lifted by a common lifter $t_{Z}$. Var ${\mathsf{O}}_{\tilde{q}}[G_{s}]=X{\langle{({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}})}_{q\in\tilde{q}}\rangle}\vdash\begin{array}[t]{@{}l@{}}{\left(X{:}\leavevmode\nobreak\ {\big{(}\mathrm{deepUnfold}(\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q},{(Y,t_{Y},\overline{G_{Y}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Y}}q})}_{Y\in\widetilde{Y_{X}}})\big{)}}_{q\in\tilde{q}_{X}}\right)}_{X\in\widetilde{X_{C}}\setminus(Z)},\\\ Z{:}\leavevmode\nobreak\ {\left(\mathrm{deepUnfold}(\overline{G_{Z}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Z}}q},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q)}_{X\in\widetilde{Y_{Z}}}})\right)}_{q\in\tilde{q}};\\\ {\big{(}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ \mu Z\mathbin{.}({\uparrow^{t_{Z}}}\mathrm{deepUnfold}(\overline{G_{Z}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{Z}}q},{(X,t_{X},\overline{G_{X}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}_{X}}q})}_{X\in\widetilde{Y_{Z}}}))\big{)}}_{q\in\tilde{q}}\end{array}$ In this typing, the types of ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}$ for each $q\in\tilde{q}$ concur with the expected types in (43), and the assignment to $Z$ in the recursive context concurs with (44). This proves the thesis. ∎ #### 4.4.2 Orchestrators and Centralized Compositions of Routers are Behaviorally Equivalent First, we formalize what we mean with a centralized composition of routers, which we call a _hub of routers_. A hub of routers is just a specific composition of routers, formalized as the centralized composition of the routers of all a global type’s participants synthesized from the global type: ###### Definition 37 (Hub of a Global Type). Given global type $G$, we define the _hub of routers_ of $G$ as follows: $\mathcal{H}_{G}:=(\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}})_{p,q\in\mathsf{prt}(G)}\big{(}{\mathchoice{\textstyle}{}{}{}\prod}_{p\in\mathsf{prt}(G)}\mathcal{R}_{p}\big{)}$ Hubs of routers can be typed using local projection (cf. Def. 22), identical to the typing of orchestrators (cf. Theorem 24): ###### Theorem 25. For relative well-formed global type $G$ and priority $\mathsf{o}$, $\mathcal{H}_{G}\vdash\emptyset;{({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{(G\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}p)})}_{p\in\mathsf{prt}(G)}.$ ###### Proof. By the typability of routers (Theorem 11) and the duality of the types of the endpoints connecting pairs of routers (Theorem 9). ∎ out $x[a,b]\xrightarrow{\vspace{-.8ex}x[a,b]}\bm{0}$ $P\xrightarrow{\vspace{-.8ex}x[a,b]}P^{\prime}$ out-open $(\bm{\nu}ya)(\bm{\nu}zb)P\xrightarrow{\vspace{-.8ex}(\bm{\nu}ya)(\bm{\nu}zb)x[a,b]}P^{\prime}$ in $x(v,w)\mathbin{.}P\xrightarrow{\vspace{-.8ex}x(v,w)}P$ $P\xrightarrow{\vspace{-.8ex}(\bm{\nu}ya)(\bm{\nu}zb)x[a,b]}P^{\prime}$ $Q\xrightarrow{\vspace{-.8ex}x(v,w)}Q^{\prime}$ out-close $P\mathbin{|}Q\xrightarrow{\vspace{-.8ex}\tau}(\bm{\nu}yv)(\bm{\nu}zw)(P^{\prime}\mathbin{|}Q^{\prime})$ sel $x[b]\triangleleft j\xrightarrow{\vspace{-.8ex}x[b]\triangleleft j}\bm{0}$ $P\xrightarrow{\vspace{-.8ex}x[b]\triangleleft j}P^{\prime}$ sel-open $(\bm{\nu}zb)P\xrightarrow{\vspace{-.8ex}(\bm{\nu}zb)x[b]\triangleleft j}P^{\prime}$ $j\in I\raisebox{16.79158pt}{}$ bra $x(w)\triangleright\\{i:P_{i}\\}_{i\in I}\xrightarrow{\vspace{-.8ex}x(w)\triangleright j}P_{j}$ $P\xrightarrow{\vspace{-.8ex}(\bm{\nu}zb)x[b]\triangleleft j}P^{\prime}$ $Q\xrightarrow{\vspace{-.8ex}x(w)\triangleright j}Q^{\prime}$ sel-close $P\mathbin{|}Q\xrightarrow{\vspace{-.8ex}\tau}(\bm{\nu}zw)(P^{\prime}\mathbin{|}Q^{\prime})$ $P\xrightarrow{\vspace{-.8ex}\alpha}Q$ $\mathrm{bn}(\alpha)\cap\mathrm{fn}(R)=\emptyset$ par-L $P\mathbin{|}R\xrightarrow{\vspace{-.8ex}\alpha}Q\mathbin{|}R$ $P\xrightarrow{\vspace{-.8ex}\alpha}Q$ $\mathrm{bn}(\alpha)\cap\mathrm{fn}(R)=\emptyset$ par-R $R\mathbin{|}P\xrightarrow{\vspace{-.8ex}\alpha}R\mathbin{|}Q$ id $(\bm{\nu}yz)(x\mathbin{\leftrightarrow}y\mathbin{|}P)\xrightarrow{\vspace{-.8ex}\tau}P\\{x/z\\}$ $P\xrightarrow{\vspace{-.8ex}\alpha}Q$ $\\{y,y^{\prime}\\}\cap\mathrm{fn}(\alpha)=\emptyset$ res $(\bm{\nu}yy^{\prime})P\xrightarrow{\vspace{-.8ex}\alpha}(\bm{\nu}yy^{\prime})Q$ Figure 13: Labeled transition system for APCP (cf. Definition 38). In order to state the behavioral equivalence of orchestrators and hubs of routers, we first define the specific behavioral equivalence we desire. To this end, we first define a labeled transition system (LTS) for APCP: ###### Definition 38 (LTS for APCP). We define the labels $\alpha$ for transitions for processes as follows: $\displaystyle\alpha::=$ $\displaystyle\leavevmode\nobreak\ \tau$ communication $\mid$ $\displaystyle\leavevmode\nobreak\ x[a,b]$ output $\displaystyle\qquad\;\mbox{\large{$\mid$}}\;(\bm{\nu}ya)(\bm{\nu}zb)x[a,b]$ bound output $\mid$ $\displaystyle\leavevmode\nobreak\ x[b]\triangleleft j$ selection $\displaystyle\qquad\;\mbox{\large{$\mid$}}\;(\bm{\nu}zb)x[b]\triangleleft j$ bound selection $\mid$ $\displaystyle\leavevmode\nobreak\ x(v,w)$ input $\displaystyle\qquad\;\mbox{\large{$\mid$}}\;x(w)\triangleright j$ branch The relation _labeled transition_ ($P\xrightarrow{\vspace{-.8ex}\alpha}Q$) is then defined by the rules in Figure 13. ###### Proposition 26. $P\longrightarrow_{\beta}Q$ if and only if $P\xrightarrow{\vspace{-.8ex}\tau}Q$. As customary, we write ‘$\Rightarrow$’ for the reflexive, transitive closure of $\xrightarrow{\vspace{-.8ex}\tau}$, and we write ‘$\xRightarrow{\alpha}$’ for $\Rightarrow\xrightarrow{\vspace{-.8ex}\alpha}\Rightarrow$ if $\alpha\neq\tau$ and for $\Rightarrow$ otherwise. We can now define the behavioral equivalence we desire: ###### Definition 39 (Weak bisimilarity). A binary relation $\mathbb{B}$ on processes is a _weak bisimulation_ if whenever $(P,Q)\in\mathbb{B}$, * • $P\xrightarrow{\vspace{-.8ex}\alpha}P^{\prime}$ implies that there is $Q^{\prime}$ such that $Q\xRightarrow{\alpha}Q^{\prime}$ and $(P^{\prime},Q^{\prime})\in\mathbb{B}$, and * • $Q\xrightarrow{\vspace{-.8ex}\alpha}Q^{\prime}$ implies that there is $P^{\prime}$ such that $P\xRightarrow{\alpha}P^{\prime}$ and $(P^{\prime},Q^{\prime})\in\mathbb{B}$. Two processes $P$ and $Q$ are _weakly bisimilar_ if there exists a weak bisimulation $\mathbb{B}$ such that $(P,Q)\in\mathbb{B}$. Our equivalence result shall relate the behavior of an orchestrator and a hub on a single but arbitrary channel. More specifically, our result will demonstrate that both settings exhibit the same actions on a channel endpoint connect to a particular participant’s implementation. In order to isolate such a channel, we place the orchestrator and hub of routers in an evaluation context consisting of restrictions and parallel compositions with arbitrary processes, such that it connects all but one of the orchestrator’s or hub’s channels. For example, given a global type $G$ and implementations $P_{q}\vdash\emptyset;{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{:}\leavevmode\nobreak\ G\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}q$ for every participant $q\in\mathsf{prt}(G)\setminus\\{p\\}$, we could use the following evaluation context: $\displaystyle E:=(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})_{q\in\mathsf{prt}(G)\setminus\\{p\\}}\big{(}{{\mathchoice{\textstyle}{}{}{}\prod}}_{q\in\mathsf{prt}(G)\setminus\\{p\\}}P_{q}\mathbin{|}[\,]\big{)}$ Replacing the hole in this evaluation context with the orchestrator or hub of routers of $G$ leaves one channel free: the channel ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ for the implementation of $p$. Now, we can observe the behavior of these two processes on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$. In what follows we write $\mathcal{O}_{G}^{\tilde{q}}$ instead of ${\mathsf{O}}_{\tilde{q}}[G]$. When we appeal to router and orchestrator synthesis, we often omit the parameter $\tilde{q}$. That is, we write ${\llbracket G\rrbracket}_{p}$ instead of ${\llbracket G\rrbracket}_{p}^{\tilde{q}}$, and $\mathcal{O}_{G}$ instead of $\mathcal{O}_{G}^{\tilde{q}}$. ###### Theorem 27. Suppose given a relative well-formed global type $G$. Let $\mathcal{H}_{G}$ be the hub of routers of $G$ (Def. 37) and take the orchestrator $\mathcal{O}_{G}^{\mathsf{prt}(G)}$ of $G$ (Def. 36). Let $p\in\mathsf{prt}(G)$, and let $E$ be an evaluation context such that $E[\mathcal{H}_{G}]\vdash\emptyset;{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}{:}\leavevmode\nobreak\ \overline{(G\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{\mathsf{o}}p)}$. Then, $E[\mathcal{H}_{G}]$ and $E[\mathcal{O}_{G}]$ are weakly bisimilar (Def. 39). We first give an intuition for the proof of Theorem 27 and its ingredients, after which we give the proof using these ingredients; then, we detail the ingredients. The proof is by coinduction, i.e., by exhibiting a weak bisimulation $\mathbb{B}$ that contains the pair $(E[\mathcal{H}_{G}],E[\mathcal{O}_{G}])$. To construct $\mathbb{B}$ and prove that it is a weak bisimulation we require the following: * • We define a function that, given a global type $G$ and a _starting relation_ $\mathbb{B}_{0}$, computes a corresponding _candidate relation_. This function is denoted $\mathbb{B}(G,\mathbb{B}_{0})$ (Def. 40). * • Suppose $G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime}$, with $k\geq 0$. Given some starting relation $\mathbb{B}_{0}$, we want to show that the relation obtained from $\mathbb{B}(G^{\prime},\mathbb{B}_{0})$ is a weak bisimulation, for which we need to assert that $\mathbb{B}_{0}$ is an appropriate starting relation. To this end, we define a function that computes a _consistent_ starting relation for a bisimulation relation, given a pair $(P,Q)$ of processes and a participant $p$ of $G$. This function is denoted $\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime},(P,Q),p\rangle$ (Def. 41). * • The property that processes in such a consistent starting relation follow a pattern of specific labeled transitions, passing through a context containing the router of $p$ or the orchestrator (Lemma 28). * • The property that the relation obtained from $\mathbb{B}(G^{\prime},\mathbb{B}_{0},p)$ is a weak bisimulation, given the consistent starting relation $\mathbb{B}_{0}=\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime},(E[\mathcal{H}_{G}],E[\mathcal{O}_{G}]),p\rangle$ (Lemma 29). Theorem 27 follows from these definitions and results: ###### Proof of Theorem 27. Let $\mathbb{B}=\mathbb{B}(G,\mathbb{B}_{0})$, where $\mathbb{B}_{0}=\langle G,(E[\mathcal{H}_{G}],E[\mathcal{O}_{G}]),p\rangle$. By Lemma 29, $\mathbb{B}$ is a weak bisimulation. Because $(E[\mathcal{H}_{G}],E[\mathcal{O}_{G}])\in\mathbb{B}_{0}\subseteq\mathbb{B}$, it then follows that $E[\mathcal{H}_{G}]$ and $E[\mathcal{O}_{G}]$ are weakly bisimilar. ∎ We setup some notations: ###### Notation 4. We adopt the following notational conventions. * • We write $\mathsf{Proc}$ to denote the set of all typable APCP processes. * • In the LTS for APCP (Def. 38), we simplify labels: we write an overlined variant for output and selection (e.g., for $(\bm{\nu}ab){{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}[a]\mathbin{\triangleleft}\ell$ we write $\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}\ell$), and omit continuation channels for input and branching (e.g., for ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}(a)\mathbin{\triangleright}\ell$ we write ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\mathbin{\triangleright}\ell$). * • Also, we write $P\xRightarrow{\alpha_{1}\ldots\alpha_{n}}Q$ rather than $P\xRightarrow{\alpha_{1}}P_{1}\xRightarrow{\alpha_{2}}P_{2}\ldots\xRightarrow{\alpha_{n}}Q$. * • We write $\tilde{\alpha}$ to denote a sequence of labels, e.g., if $\tilde{\alpha}=\alpha_{1}\ldots\alpha_{n}$ then ${\xRightarrow{\tilde{\alpha}}}={\xRightarrow{\alpha_{1}\ldots\alpha_{n}}}$. If $\tilde{\alpha}=\epsilon$ (empty sequence), then ${\xRightarrow{\tilde{\alpha}}}={\Rightarrow}$. The following function defines a relation on processes, which we will use as the weak bisimulation between $E[\mathcal{H}_{G}]$ and $E[\mathcal{O}_{G}]$: ###### Definition 40 (Candidate Relation). Let $G$ be a global type and let $p$ be a participant of $G$. Also, let $\mathbb{B}_{0}\subseteq\mathsf{Proc}\times\mathsf{Proc}$ denote a relation on processes. We define a _candidate relation_ for a weak bisimulation of the hub and orchestrator of $G$ observed on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ starting at $\mathbb{B}_{0}$, by abuse of notation denoted $\mathbb{B}(G,\mathbb{B}_{0},p)$. The definition is inductive on the structure of $G$: * • $G=\bullet$. Then $\mathbb{B}(G,\mathbb{B}_{0},p)=\mathbb{B}_{0}$. * • $G=s\mathbin{\twoheadrightarrow}r\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}$. We distinguish four cases, depending on the involvement of $p$: * – $p=s$. For every $i\in I$, let $\displaystyle\mathbb{B}_{1}^{i}$ $\displaystyle=\\{(P_{1},Q_{1})\mid\exists(P_{0},Q_{0})\in\mathbb{B}_{0}\leavevmode\nobreak\ \text{s.t.}\leavevmode\nobreak\ P_{0}\xrightarrow{\vspace{-.8ex}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\mathbin{\triangleright}i}\Rightarrow P_{1}\leavevmode\nobreak\ \text{and}\leavevmode\nobreak\ Q_{0}\xrightarrow{\vspace{-.8ex}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\mathbin{\triangleright}i}\Rightarrow Q_{1}\\};$ $\displaystyle\mathbb{B}_{2}^{i}$ $\displaystyle=\\{(P_{2},Q_{2})\mid\exists(P_{1},Q_{1})\in\mathbb{B}_{1}^{i}\leavevmode\nobreak\ \text{s.t.}\leavevmode\nobreak\ P_{1}\xrightarrow{\vspace{-.8ex}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}(y)}\Rightarrow P_{2}\leavevmode\nobreak\ \text{and}\leavevmode\nobreak\ Q_{1}\xrightarrow{\vspace{-.8ex}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}(y)}\Rightarrow Q_{2}\\}$ Then $\mathbb{B}(G,\mathbb{B}_{0},p)=\mathbb{B}_{0}\cup{\mathchoice{\textstyle}{}{}{}\bigcup}_{i\in I}(\mathbb{B}_{1}^{i}\cup\mathbb{B}(G_{i},\mathbb{B}_{2}^{i},p)).$ * – $p=r$. For every $i\in I$, let $\displaystyle\mathbb{B}_{1}^{i}$ $\displaystyle=\\{(P_{1},Q_{1})\mid\exists(P_{0},Q_{0})\in\mathbb{B}_{0}\leavevmode\nobreak\ \text{s.t.}\leavevmode\nobreak\ P_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i}\Rightarrow P_{1}\leavevmode\nobreak\ \text{and}\leavevmode\nobreak\ Q_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i}\Rightarrow Q_{1}\\};$ $\displaystyle\mathbb{B}_{2}^{i}$ $\displaystyle=\\{(P_{2},Q_{2})\mid\exists(P_{1},Q_{1})\in\mathbb{B}_{1}\leavevmode\nobreak\ \text{and}\leavevmode\nobreak\ y\leavevmode\nobreak\ \text{s.t.}\leavevmode\nobreak\ P_{1}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}\Rightarrow P_{2}\leavevmode\nobreak\ \text{and}\leavevmode\nobreak\ Q_{1}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}\Rightarrow Q_{2}\\}.$ Then $\mathbb{B}(G,\mathbb{B}_{0},p)=\mathbb{B}_{0}\cup{\mathchoice{\textstyle}{}{}{}\bigcup}_{i\in I}(\mathbb{B}_{1}^{i}\cup\mathbb{B}(G_{i},\mathbb{B}_{2}^{i},p)).$ * – $p\notin\\{s,r\\}$ and $\mathrm{hdep}(p,s,G)$ or $\mathrm{hdep}(p,r,G)$. For every $i\in I$, let $\mathbb{B}_{1}^{i}=\\{(P_{1},Q_{1})\mid\exists(P_{0},Q_{0})\in\mathbb{B}_{0}\leavevmode\nobreak\ \text{s.t.}\leavevmode\nobreak\ P_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i}\Rightarrow P_{1}\leavevmode\nobreak\ \text{and}\leavevmode\nobreak\ Q_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i}\Rightarrow Q_{1}\\}$ Then $\mathbb{B}(G,\mathbb{B}_{0},p)=\mathbb{B}_{0}\cup{\mathchoice{\textstyle}{}{}{}\bigcup}_{i\in I}\mathbb{B}(G_{i},\mathbb{B}_{1}^{i},p).$ * – $p\notin\\{s,r\\}$ and neither $\mathrm{hdep}(p,s,G)$ nor $\mathrm{hdep}(p,r,G)$. Then $\mathbb{B}(G,\mathbb{B}_{0},p)=\mathbb{B}(G_{j},\mathbb{B}_{0},p)$ for any $j\in I$. * • $G=\mu X\mathbin{.}G^{\prime}$. Then $\mathbb{B}(G,\mathbb{B}_{0},p)=\mathbb{B}(G^{\prime}\\{\mu X\mathbin{.}G^{\prime}/X\\},\mathbb{B}_{0},p)$. * • $G=\mathsf{skip}\mathbin{.}G^{\prime}$. Then $\mathbb{B}(G,\mathbb{B}_{0},p)=\mathbb{B}(G^{\prime},\mathbb{B}_{0},p)$. The function $\mathbb{B}(G,\mathbb{B}_{0},p)$ constructs a relation between processes by following labeled transitions on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ that concur with the expected behavior of $p$’s router and the orchestrator depending on the shape of $G$. For example, for $G=s\mathbin{\twoheadrightarrow}p\\{i\langle S_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I}$, for each $i\in I$, the function constructs $\mathbb{B}_{1}^{i}$ containing the processes reachable from $\mathbb{B}_{0}$ through a transition labeled $\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i$ (selection of the label chosen by $s$), and $\mathbb{B}_{2}^{i}$ containing the processes reachable from $\mathbb{B}_{0}$ through a transition labeled $\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]$ (output of the endpoint sent by $s$); the resulting relation then consists of $\mathbb{B}_{0}$ and, for each $i\in I$, $\mathbb{B}_{1}^{i}$ and $\mathbb{B}(G_{i},\mathbb{B}_{2}^{i},p)$ (i.e., the candidate relation for $G_{i}$ starting with $B_{2}^{i}$). Since we are interested in a _weak_ bisimulation, the $\tau$-transitions of one process do not need to be simulated by related processes. Hence, e.g., if $(P,Q)\in\mathbb{B}_{0}$ and $P\xrightarrow{\vspace{-.8ex}\tau}P^{\prime}$ and $Q\xrightarrow{\vspace{-.8ex}\tau}Q^{\prime}$, then $\\{(P,Q),(P^{\prime},Q),(P,Q^{\prime}),(P^{\prime},Q^{\prime})\\}\subseteq\mathbb{B}(G,\mathbb{B}_{0},p)$. This way, we only _synchronize_ related processes when they can both take the same labeled transition. We intend to show that, if $G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime}$, the function $\mathbb{B}(G^{\prime},\mathbb{B}_{0},p)$ constructs a weak bisimulation. However, for this to hold, the starting relation $\mathbb{B}_{0}$ cannot be arbitrary: the pairs of processes in $\mathbb{B}_{0}$ have to be reachable from $E[\mathcal{H}_{G}]$ and $E[\mathcal{O}_{G}]$ through labeled transitions that concur with the transitions from $G$ to $G^{\prime}$. Moreover, the processes must have “passed through” evaluation contexts containing the router for $p$ at $G^{\prime}$ and the orchestrator at $G^{\prime}$. The following defines a _consistent starting relation_ , parametric on $k$, that satisfies these requirements. Note that for constructing the relation $\mathbb{B}$, we only need the following definition for $k=0$. However, in the proof that $\mathbb{B}$ is a weak bisimulation we need to generalize it to $k\geq 0$ to assure that the starting relation of coinductive steps is consistent. ###### Definition 41 (Consistent Starting Relation). Let $G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime}$ (with $k\geq 0$) be a sequence of labeled transitions from $G$ to $G^{\prime}$ including the intermediate global types (cf. Definition 35) and let $p$ be a participant of $G$. Also, let $(P,Q)$ be a pair of initial processes. We define the _consistent starting relation_ for observing the hub and orchestrator of $G^{\prime}$ on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ starting at $(P,Q)$ after the transitions from $G$ to $G^{\prime}$, denoted $\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime},(P,Q),p\rangle$. The definition is inductive on the number $k$ of transitions: * • $k=0$. Then $\langle G,(P,Q),p\rangle=\\{(P^{\prime},Q^{\prime})\mid P\Rightarrow P^{\prime}\leavevmode\nobreak\ \text{and}\leavevmode\nobreak\ Q\Rightarrow Q^{\prime}\\}$. * • $k=k^{\prime}+1$. Then $\displaystyle\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k^{\prime}}}G_{k^{\prime}}\xrightarrow{\vspace{-.8ex}\beta_{k}}G_{k},(P,Q),p\rangle={}$ $\displaystyle\quad\\{(P_{k},Q_{k})\mid\begin{array}[t]{@{}l@{}}\exists(P_{k^{\prime}},Q_{k^{\prime}})\in\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k^{\prime}}}G_{k^{\prime}},(P,Q),p\rangle\\\ \text{s.t.}\leavevmode\nobreak\ (\begin{array}[t]{@{}l@{}l@{}}&\leavevmode\nobreak\ (\exists C\leavevmode\nobreak\ \text{s.t.}\leavevmode\nobreak\ P_{k^{\prime}}\xRightarrow{\tilde{\alpha}}C[{\llbracket G_{k}\rrbracket}_{p}]\Rightarrow P_{k})\\\ \text{and}&\leavevmode\nobreak\ (\exists D\leavevmode\nobreak\ \text{s.t.}\leavevmode\nobreak\ Q_{k^{\prime}}\xRightarrow{\tilde{\alpha}}D[\mathcal{O}_{G_{k}}]\Rightarrow Q_{k}))\\},\end{array}\end{array}$ where $\tilde{\alpha}$ depends on $\beta_{k}=s\rangle r:j\langle S_{j}\rangle$ and $G_{k^{\prime}}$ (in unfolded form if $G_{k^{\prime}}=\mu X\mathbin{.}G^{\prime}_{k^{\prime}}$): * – If $p=s$, then $\tilde{\alpha}={{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\mathbin{\triangleright}j\,{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}(y)$. * – If $p=r$, then $\tilde{\alpha}=\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j\,\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]$. * – If $p\notin\\{s,r\\}$ and $\mathrm{hdep}(p,s,G_{k})$ or $\mathrm{hdep}(p,r,G_{k})$, then $\tilde{\alpha}=\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j$. * – If $p\notin\\{s,r\\}$ and neither $\mathrm{hdep}(p,s,G_{k})$ nor $\mathrm{hdep}(p,r,G_{k})$, then $\tilde{\alpha}=\epsilon$. ###### Lemma 28. Let $G$ be a relative well-formed global type such that $G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime}$ for $k\geq 0$ and let $p$ be a participant of $G$. Also, let $E$ be an evaluation context such that $\mathrm{fn}(E)=\\{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\\}$. Then there exists $\tilde{\alpha}$ such that, for every $(P,Q)\in\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime},(E[\mathcal{H}_{G}],E[\mathcal{O}_{G}]),p\rangle$, * • $E[\mathcal{H}_{G}]\xRightarrow{\tilde{\alpha}}C\big{[}{\llbracket G^{\prime}\rrbracket}_{p}\big{]}\Rightarrow P$ where $C$ is an evaluation context without an output or selection on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$; and * • $E[\mathcal{O}_{G}]\xRightarrow{\tilde{\alpha}}D\big{[}\mathcal{O}_{G^{\prime}}\big{]}\Rightarrow Q$ where $D$ is an evaluation context without an output or selection on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$. ###### Proof. By induction on $k$. In the base case ($k=0$), we have $G=G^{\prime}$, so $E[\mathcal{H}_{G}]=C[{\llbracket G^{\prime}\rrbracket}_{p}]\Rightarrow P$ and $E[\mathcal{O}]=D[\mathcal{O}_{G^{\prime}}]\Rightarrow Q$. For the inductive case ($k=k^{\prime}+1$), we detail the representative case where $G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k^{\prime}}}G_{k^{\prime}}=p\mathbin{\twoheadrightarrow}s\\{i\langle S_{i}\rangle\mathbin{.}G^{\prime}_{i}\\}_{i\in I}\xrightarrow{\vspace{-.8ex}p\rangle s:i^{\prime}\langle S_{i^{\prime}}\rangle}G^{\prime}$ for some $i^{\prime}\in I$. By the IH, for every $(P_{k^{\prime}},Q_{k^{\prime}})\in\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k^{\prime}}}G_{k^{\prime}},(E[\mathcal{H}_{G}],E[\mathcal{O}_{G}]),p\rangle$, there exists $\tilde{\alpha}^{\prime}$ such that $E[\mathcal{H}_{G}]\xRightarrow{\tilde{\alpha}^{\prime}}C^{\prime}[{\llbracket G_{k^{\prime}}\rrbracket}^{p}]\Rightarrow P_{k^{\prime}}$ and $E[\mathcal{O}_{G}]\xRightarrow{\tilde{\alpha}^{\prime}}D^{\prime}[\mathcal{O}_{G_{k^{\prime}}}]\Rightarrow Q_{k^{\prime}}$ where $C^{\prime}$ and $D^{\prime}$ are without output or selection on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$. Take any $(P,Q)\in\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k^{\prime}}}G_{k^{\prime}}\xrightarrow{\vspace{-.8ex}s\rangle p:i^{\prime}\langle S_{i^{\prime}}\rangle}G^{\prime},(E[\mathcal{H}_{G}],E[\mathcal{O}_{G}]),p\rangle$. By definition, there exists $(P_{k^{\prime}},Q_{k^{\prime}})\in\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k^{\prime}}}G_{k^{\prime}},(E[\mathcal{H}_{G}],E[\mathcal{O}_{G}]),p\rangle$ such that $P_{k^{\prime}}\xRightarrow{\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i^{\prime}\,\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}C[{\llbracket G^{\prime}\rrbracket}^{p}]\Rightarrow P\text{ and }Q_{k^{\prime}}\xRightarrow{\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i^{\prime}\,\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}D[\mathcal{O}_{G^{\prime}}]\Rightarrow Q$ where there are no outputs or selection on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ in $C$ and $D$. Let $\tilde{\alpha}=\tilde{\alpha}^{\prime}\,\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\mathbin{\triangleleft}i^{\prime}}\,\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]$. Then ${E[\mathcal{H}_{G}]\xRightarrow{\tilde{\alpha}}C[{\llbracket G^{\prime}\rrbracket}_{p}]\Rightarrow P}$ and $E[\mathcal{O}_{G}]\xRightarrow{\tilde{\alpha}}D[\mathcal{O}_{G^{\prime}}]\Rightarrow Q$. ∎ ###### Lemma 29. Let $G$ be a relative well-formed global type such that $G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime}$ (with $k\geq 0$) and let $p$ be a participant of $G$. Also, let $E$ be an evaluation context such that $\mathrm{fn}(E)=\\{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\\}$. Then the relation $\mathbb{B}(G^{\prime},\mathbb{B}_{0})$, with $\mathbb{B}_{0}=\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime},(E[\mathcal{H}_{G}],E[\mathcal{O}_{G}]),p\rangle$, is a weak bisimulation (cf. Definition 39). ###### Proof. By coinduction on the structure of $G^{\prime}$; there are four cases (communication, recursion, $\mathsf{skip}$, and $\bullet$). We only detail the interesting case of communication, which is the only case which involves transitions with labels other than $\tau$. There are four subcases depending on the involvement of $p$ in the communication ($p$ is sender, $p$ is recipient, $p$ depends on the communication, or $p$ does not depend on the communication). In each subcase, the proof follows the same pattern, so as a representative case, we detail when $p$ is the recipient of the communication, i.e., $G^{\prime}=s\mathbin{\twoheadrightarrow}p\\{i\langle S_{i}\rangle\mathbin{.}G^{\prime}_{i}\\}_{i\in I}$. Recall $\displaystyle{\llbracket G^{\prime}\rrbracket}_{p}$ $\displaystyle={{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\mathbin{\triangleright}\big{\\{}i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}i\cdot{(\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}}\mathbin{\triangleleft}i)}_{q\in\mathsf{deps}}\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}(v)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[w]\cdot(v\mathbin{\leftrightarrow}w\mathbin{|}{\llbracket G^{\prime}_{i}\rrbracket}_{p})\big{\\}}_{i\in I},$ (Algorithm 1 algorithm 1) $\displaystyle{\llbracket G^{\prime}\rrbracket}_{s}$ $\displaystyle={{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\mathbin{\triangleright}\big{\\{}i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}}\mathbin{\triangleleft}i\cdot{(\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}}\mathbin{\triangleleft}i)}_{q\in\mathsf{deps}}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}(v)\mathbin{.}\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}}[w]\cdot(v\mathbin{\leftrightarrow}w\mathbin{|}{\llbracket G^{\prime}_{i}\rrbracket}_{s})\big{\\}}_{i\in I},$ (Algorithm 1 algorithm 1) $\displaystyle\mathcal{O}_{G^{\prime}}$ $\displaystyle={{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}\triangleright\\{i{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\triangleleft i\cdot{(\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}}\triangleleft i)}_{q\in\mathsf{deps}}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}(v)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[w]\cdot(v\mathbin{\leftrightarrow}w\mathbin{|}\mathcal{O}_{G^{\prime}_{i}})\\}_{i\in I}.$ (Algorithm 2 algorithm 2) Let $\mathbb{B}=\mathbb{B}(G^{\prime},\mathbb{B}_{0})$. We have $\mathbb{B}=\mathbb{B}_{0}\cup{\mathchoice{\textstyle}{}{}{}\bigcup}_{i\in I}(\mathbb{B}_{1}^{i}\cup\mathbb{B}(G^{\prime}_{i},\mathbb{B}_{2}^{i}))$ with $\mathbb{B}_{1}^{i}$ and $\mathbb{B}_{2}^{i}$ as defined above. Take any $(P,Q)\in\mathbb{B}$; we distinguish cases depending on the subset of $\mathbb{B}$ to which $(P,Q)$ belongs: * • $(P,Q)\in\mathbb{B}_{0}$. By Lemma 28, we have $E[\mathcal{H}_{G}]\xRightarrow{\tilde{\alpha}}C[{\llbracket G^{\prime}\rrbracket}_{p}]\Rightarrow P$ and $E[\mathcal{O}]\xRightarrow{\tilde{\alpha}}D[\mathcal{O}_{G^{\prime}}]\Rightarrow Q$, where $C$ and $D$ do not contain an output or selection on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$. Suppose $P\xrightarrow{\vspace{-.8ex}\alpha}P^{\prime}$; we need to exhibit a matching weak transition from $Q$. By assumption, there are no outputs or selections on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ in $C$ and $D$. Since there are no outputs or selections on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ in $C$, by definition of ${\llbracket G^{\prime}\rrbracket}_{p}$, we need only consider two cases for $\alpha$: * – $\alpha=\tau$. We have $Q\Rightarrow Q$, so $Q\xRightarrow{\tau}Q$. Since $C[{\llbracket G^{\prime}\rrbracket}_{p}]\Rightarrow P^{\prime}$ and $D[\mathcal{O}_{G^{\prime}}]\Rightarrow Q$, we have $(P^{\prime},Q)\in\mathbb{B}_{0}\subseteq\mathbb{B}$. * – $\alpha=\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j$ for some $j\in I$. To enable this transition, which originates from $p$’s router, somewhere in the $\tau$-transitions between $C[{\llbracket G^{\prime}\rrbracket}_{p}]$ and $P$ the label $j$ was received on ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$, sent by the router of $s$ on ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}$. For this to happen, the label $j$ was received on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}$, sent from the context on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$. Since $\mathcal{H}_{G}$ and $\mathcal{O}$ are embedded in the same context, the communication of $j$ between ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ and ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}$ can also take place after a number of $\tau$-transitions from $D[\mathcal{O}_{G^{\prime}}]$, after which the selection of $j$ on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ becomes enabled. Hence, since there are no outputs or selection on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ in $D$, we have $Q\Rightarrow Q_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}Q^{\prime}$. We have $D[\mathcal{O}_{G^{\prime}}]\Rightarrow Q_{0}$, so $(P,Q_{0})\in\mathbb{B}_{0}$. Since $P\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow P^{\prime}$ and $Q_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow Q^{\prime}$, we have $(P^{\prime},Q^{\prime})\in\mathbb{B}_{1}^{j}\subseteq\mathbb{B}^{\prime}$. Now suppose $Q\xrightarrow{\vspace{-.8ex}\alpha}Q^{\prime}$; we need to exhibit a matching weak transition from $P$. Again, we need only consider two cases for $\alpha$: * – $\alpha=\tau$. Analogous to the similar case above. * – $\alpha=\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j$ for some $j\in I$. To enable this transition, which originates from the orchestrator, somewhere in the $\tau$-transitions between $D[\mathcal{O}_{G^{\prime}}]$ and $Q$ the label $j$ was received on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}$, sent from the context on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$. Hence, this communication can also take place after a number of transitions from $E[\mathcal{H}_{G}]$, where the label is received by the router of $s$. After this, from $C[{\llbracket G^{\prime}\rrbracket}_{p}]$, the router of $s$ forwards $j$ to $p$’s router (communication between ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}$ and ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$), enabling the selection of $j$ on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ in $p$’s router. Hence, since there are no outputs or selections in $C$, we have $P\Rightarrow P_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}P^{\prime}$. We have $C[{\llbracket G^{\prime}\rrbracket}_{p}]\Rightarrow P_{0}$, so $(P_{0},Q)\in\mathbb{B}_{0}$. Since $P_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow P^{\prime}$ and $Q\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow Q^{\prime}$, we have $(P^{\prime},Q^{\prime})\in\mathbb{B}_{1}^{j}\subseteq\mathbb{B}$. * • $(P,Q)\in\mathbb{B}_{1}^{j}$ for some $j\in I$. We have $E[\mathcal{H}_{G}]\xRightarrow{\tilde{\alpha}}C[{\llbracket G^{\prime}\rrbracket}_{p}]\Rightarrow P_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow P$ and ${E[\mathcal{O}_{G}]\xRightarrow{\tilde{\alpha}}D[\mathcal{O}_{G^{\prime}}]\Rightarrow Q_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow Q}$ where $(P_{0},Q_{0})\in\mathbb{B}_{0}$. Since we have already observed the selection of $j$ on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ from both the hub and the orchestrator, we know that the routers of $p$ and $s$ are in branch $j$, and similarly the orchestrator is in branch $j$. Suppose $P\xrightarrow{\vspace{-.8ex}\alpha}P^{\prime}$. To exhibit a matching weak transition from $Q$ we only need to consider two cases for $\alpha$: * – $\alpha=\tau$. We have $Q\xRightarrow{\tau}Q$, and $P_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow P^{\prime}$ and $Q_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow Q$, so $(P^{\prime},Q)\in\mathbb{B}_{1}^{j}\subseteq\mathbb{B}$. * – $\alpha=\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]$ for some $y$. The observed output of some $y$ on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ must originate from $p$’s router. This output is only enabled after receiving some $v$ over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$, which must be sent by the router of $s$ over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}$. The output by the router of $s$ is only enabled after receiving some $v$ over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}$, sent by the context over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$. Since the hub and the orchestrator are embedded in the same context, the communication of $v$ from ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ to ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}$ can also occur (or has already occurred) for the orchestrator. After this, the output of $y$ over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ is enabled in the orchestrator, i.e., $Q\Rightarrow Q_{1}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}Q^{\prime}$. We have $Q_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow Q_{1}$, so $(P,Q_{1})\in\mathbb{B}_{1}^{j}$. Since $P\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}\Rightarrow P^{\prime}$ and $Q_{1}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}\Rightarrow Q^{\prime}$, we have $(P^{\prime},Q^{\prime})\in\mathbb{B}_{2}^{j}$. By definition, $\mathbb{B}_{2}^{j}\subseteq\mathbb{B}(G^{\prime}_{j},B_{2}^{j})\subseteq\mathbb{B}$, so $(P^{\prime},Q^{\prime})\in\mathbb{B}$. Now suppose $Q\xrightarrow{\vspace{-.8ex}\alpha}Q^{\prime}$. To exhibit a matching weak transition from $P$ we only need to consider two cases for $\alpha$: * – $\alpha=\tau$. Analogous to the similar case above. * – $\alpha=\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]$ for some $y$. The observed output of some $y$ on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ must originate from the orchestrator. This output is only enabled after receiving some $v$ over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}$, sent by the context of ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$. Since the hub and the orchestrator are embeded in the same context, the communication of $v$ from ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ to ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}$ can also occur (or has already occurred) for the router of $s$. After this, the router of $s$ sends another channel $v^{\prime}$ over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}$, received by $p$’s router on ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$. This enables the output of $y$ on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ by $p$’s router, i.e., $P\Rightarrow P_{1}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}P^{\prime}$. We have $P_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow P_{1}$, so $(P_{1},Q)\in\mathbb{B}_{1}^{j}$. Since $P\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}\Rightarrow P^{\prime}$ and $Q_{1}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}\Rightarrow Q^{\prime}$, we have $(P^{\prime},Q^{\prime})\in\mathbb{B}_{2}^{j}$. As above, this implies that $(P^{\prime},Q^{\prime})\in\mathbb{B}$. * • For some $j\in I$, $(P,Q)\in\mathbb{B}(G^{\prime}_{j},\mathbb{B}_{2}^{j})$. The thesis follows from proving that $\mathbb{B}(G^{\prime}_{j},\mathbb{B}_{2}^{j})$ is a weak bisimulation. For this, we want to appeal to the coinduction hypothesis, so we have to show that $\mathbb{B}_{2}^{j}=\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime}\xrightarrow{\vspace{-.8ex}s\rangle p:j\langle S_{j}\rangle}G^{\prime}_{j},(E[\mathcal{H}_{G}],E[\mathcal{O}]),p\rangle$. We prove that $(P_{2},Q_{2})\in\mathbb{B}_{2}^{j}$ if and only if $(P_{2},Q_{2})\in\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime}\xrightarrow{\vspace{-.8ex}s\rangle p:j\langle S_{j}\rangle}G^{\prime}_{j},(E[\mathcal{H}_{G}],E[\mathcal{O}]),p\rangle$, i.e., we prove both directions of the bi-implication: * – Take any $(P_{2},Q_{2})\in\mathbb{B}_{2}^{j}$. We have $E[\mathcal{H}_{G}]\xRightarrow{\tilde{\alpha}}C[{\llbracket G^{\prime}\rrbracket}_{p}]\Rightarrow P_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow P_{1}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}\Rightarrow P_{2}$ and $E[\mathcal{O}]\xRightarrow{\tilde{\alpha}}D[\mathcal{O}_{G^{\prime}}]\Rightarrow Q_{0}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow Q_{1}\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}\Rightarrow Q_{2}$, where $(P_{0},Q_{0})\in\mathbb{B}_{0}$ and $(P_{1},P_{1})\in\mathbb{B}_{1}^{j}$. By definition, somewhere during the transitions from $C[{\llbracket G^{\prime}\rrbracket}_{p}]$ to $P_{1}$, we find $C^{\prime}[{\llbracket G^{\prime}_{j}\rrbracket}_{p}]$, which may then further reduce by $\tau$-transitions towards $P_{2}$. As soon as we do find $C^{\prime}[{\llbracket G^{\prime}_{j}\rrbracket}_{p}]$, the output on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ is available, and the selection on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ has already occurred or is still available. Because they are asynchronous actions, we can observe the selection and output on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ as soon as they are available, before further reducing $p$’s router. Hence, we can observe $C[{\llbracket G^{\prime}\rrbracket}_{p}]\Rightarrow\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}\Rightarrow C^{\prime\prime}[{\llbracket G^{\prime}_{j}\rrbracket}_{p}]\Rightarrow P_{2}$, i.e., ${E[\mathcal{H}_{G}]\xRightarrow{\tilde{\alpha}}C[{\llbracket G^{\prime}\rrbracket}_{p}]\xRightarrow{{\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\,{\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}}C^{\prime\prime}[{\llbracket G^{\prime}_{j}\rrbracket}_{p}]\Rightarrow P_{2}}.$ By definition, ${\llbracket G^{\prime}_{j}\rrbracket}_{p}$ has no output or selection on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ available, so there are no outputs or selections on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ in $C^{\prime\prime}$. By a similar argument, we can observe $D[\mathcal{O}_{G^{\prime}}]\Rightarrow\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\Rightarrow\xrightarrow{\vspace{-.8ex}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}\Rightarrow D^{\prime\prime}[\mathcal{O}_{G^{\prime}_{j}}]\Rightarrow Q_{2}$, i.e., $E[\mathcal{O}]\xRightarrow{\tilde{\alpha}}D[\mathcal{O}_{G^{\prime}}]\xRightarrow{{\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j}\,{\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}}D^{\prime\prime}[\mathcal{O}_{G^{\prime}_{j}}]\Rightarrow Q_{2}$. Also in this case, there are no outputs or selections on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ in $D^{\prime\prime}$. By assumption and definition, $(C^{\prime\prime}[{\llbracket G^{\prime}\rrbracket}_{p}],D^{\prime\prime}[\mathcal{O}_{G^{\prime}}])\in\mathbb{B}_{0}=\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime},(E[\mathcal{H}_{G}],E[\mathcal{O}]),p\rangle.$ Hence, by definition, $(P_{2},Q_{2})\in\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime}\xrightarrow{\vspace{-.8ex}s\rangle p:j\langle S_{j}\rangle}G^{\prime}_{j},(E[\mathcal{H}_{G}],E[\mathcal{O}]),p\rangle$. * – Take any $(P,Q)\in\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime}\xrightarrow{\vspace{-.8ex}s\rangle p:j\langle S_{j}\rangle}G^{\prime}_{j},(E[\mathcal{H}_{G}],E[\mathcal{O}]),p\rangle$. By definition, there are $(P^{\prime},Q^{\prime})\in\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime},(E[\mathcal{H}_{G}],E[\mathcal{O}]),p\rangle$ such that $P^{\prime}\xRightarrow{\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j\,\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}C[{\llbracket G^{\prime}\rrbracket}_{p}]\Rightarrow P$ and $Q^{\prime}\xRightarrow{\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}\mathbin{\triangleleft}j\,\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}}[y]}D[\mathcal{O}_{G^{\prime}}]\Rightarrow Q$. Since, $\mathbb{B}_{0}=\langle G\xrightarrow{\vspace{-.8ex}\beta_{1}}\ldots\xrightarrow{\vspace{-.8ex}\beta_{k}}G^{\prime},(E[\mathcal{H}_{G}],E[\mathcal{O}]),p\rangle$, by definition $(P,Q)\in\mathbb{B}_{2}^{j}$. ∎ ## 5 Routers in Action We demonstrate our router-based analysis of global types by means of several examples. First, in § 5.1 and § 5.2 we consider two simple protocols: they illustrate the different components of our approach, and our support for delegation and interleaving. Then in § 5.3 we revisit the authorization protocol $G_{\mathsf{auth}}$ from Section 1 to illustrate how our analysis supports also more complex protocols featuring also non-local choices and recursion. ### 5.1 Delegation and Interleaving We illustrate our analysis by considering a global type with delegation and interleaving, based on an example by Toninho and Yoshida [50, Ex. 6.9]. Consider the global type: $\displaystyle G_{\mathsf{intrl}}:=p\mathbin{\twoheadrightarrow}q{:}1\langle{!}\mathsf{int}\mathbin{.}\bullet\rangle\mathbin{.}r\mathbin{\twoheadrightarrow}t{:}2\langle\mathsf{int}\rangle\mathbin{.}p\mathbin{\twoheadrightarrow}q{:}3\mathbin{.}\bullet$ Following Toninho and Yoshida [50], we define implementations of the roles of the four participants ($p,q,r,t$) of $G_{\mathsf{intrl}}$ using three processes ($P_{1}$, $P_{2}$, and $P_{3}$): $P_{2}$ and $P_{3}$ implement the roles of $q$ and $r$, respectively, and $P_{1}$ interleaves the roles of $p$ and $t$ by sending a channel $s$ to $q$ and receiving an int value $v$ from $r$, which it should forward to $q$ over $s$. $\displaystyle P_{1}$ $\displaystyle:=\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}\mathbin{\triangleleft}1\cdot\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}[s]\cdot({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}t}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\mathbin{\triangleright}\\{2{:}\leavevmode\nobreak\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}t}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}(v)\mathbin{.}\overline{s}[w]\cdot v\mathbin{\leftrightarrow}w\\}\mathbin{|}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}\mathbin{\triangleleft}3\cdot\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}[z]\cdot\bm{0})$ $\displaystyle\vdash{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ {\oplus}^{0}\big{\\{}1{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}{!}\mathsf{int}\mathbin{.}\bullet{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{1}{\oplus}^{8}\\{3{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{9}\bullet\\}\big{\\}},\leavevmode\nobreak\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}t}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ \&^{6}\\{2{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{7}\bullet\\}$ $\displaystyle P_{2}$ $\displaystyle:={{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\mathbin{\triangleright}\\{1{:}\leavevmode\nobreak\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}(y)\mathbin{.}y(x)\mathbin{.}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\mathbin{\triangleright}\\{3{:}\leavevmode\nobreak\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}(u)\mathbin{.}\bm{0}\\}\\}\vdash{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ \&^{2}\big{\\{}1{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}{!}\mathsf{int}\mathbin{.}\bullet{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{3}\&^{10}\\{3{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{11}\bullet\\}\big{\\}}$ $\displaystyle P_{3}$ $\displaystyle:=\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}\mathbin{\triangleleft}2\cdot\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}[\bm{33}]\cdot\bm{0}\vdash{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ {\oplus}^{4}\\{2{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{5}\bullet\\}$ where ‘$\bm{33}$’ denotes a closed channel endpoint representing the number “$33$”. To prove that $P_{1}$, $P_{2}$, and $P_{3}$ correctly implement $G_{\mathsf{intrl}}$, we compose them with the routers synthesized from $G_{\mathsf{intrl}}$. For example, the routers for $p$ and $t$, to which $P_{1}$ will connect, are as follows (omitting curly braces for branches on a single label): $\displaystyle\mathcal{R}_{p}$ $\displaystyle={{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\mathbin{\triangleright}1\mathbin{.}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}\mathbin{\triangleleft}1\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}(s)\mathbin{.}\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}}[s^{\prime}]\cdot(s\mathbin{\leftrightarrow}s^{\prime}\mathbin{|}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}\mathbin{\triangleright}3\mathbin{.}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}\mathbin{\triangleleft}3\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}(z)\mathbin{.}\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}}[z^{\prime}]\cdot(z\mathbin{\leftrightarrow}z^{\prime}\mathbin{|}\bm{0}))$ $\displaystyle\mathcal{R}_{t}$ $\displaystyle={{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}t}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}\mathbin{\triangleright}2\mathbin{.}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}t}}\mathbin{\triangleleft}2\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}t}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}(v)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}t}}}[v^{\prime}]\cdot(v\mathbin{\leftrightarrow}v^{\prime}\mathbin{|}\bm{0})$ We assign values to the priorities in ${{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}{!}\mathsf{int}\mathbin{.}\bullet{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}=\bullet\mathbin{\otimes}^{\mathsf{o}}\bullet$ to ensure that $P_{1}$ and $P_{2}$ are well-typed; assigning $\mathsf{o}=8$ works, because the output on $s$ in $P_{1}$ occurs after the input on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}t}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ (which has priority 6–7) and the input on $y$ in $P_{2}$ occurs before the second input on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ (which has priority 10–11). The types assigned to ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ and ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}t}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ in $P_{1}$ coincide with $(G_{\mathsf{intrl}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}p)$ and $(G_{\mathsf{intrl}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}t)$, respectively (cf. Def. 22). Therefore, by Theorem 11, the process $P_{1}$ connect to the routers for $p$ and $t$ $(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}})(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}t}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu t}})(P_{1}\mathbin{|}\mathcal{R}_{p}\mathbin{|}\mathcal{R}_{t})$ is well- typed. Similarly, $(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}})(P_{2}\mathbin{|}\mathcal{R}_{q})$ and $(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}})(P_{3}\mathbin{|}\mathcal{R}_{r})$ are well-typed. The composition of these routed implementations results in the following network: $\displaystyle N_{\mathsf{intrl}}:=\begin{array}[]{c}(\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}})(\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}})\\\ (\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}t}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}t}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}})(\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}})\\\ (\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}t}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}t}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}})(\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}t}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}t}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}r}})\end{array}\left(\begin{array}[]{l}\phantom{{}\mathbin{|}{}}(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}})(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}t}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}t}})(P_{1}\mathbin{|}\mathcal{R}_{p}\mathbin{|}\mathcal{R}_{t})\\\ {}\mathbin{|}(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}q}})(P_{2}\mathbin{|}\mathcal{R}_{q})\\\ {}\mathbin{|}(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}r}})(P_{3}\mathbin{|}\mathcal{R}_{r})\end{array}\right)$ We have $N_{\mathsf{intrl}}\in\mathrm{net}(G_{\mathsf{intrl}})$ (cf. Def. 25), so, by Theorem 18, $N_{\mathsf{intrl}}$ is deadlock free and, by Theorem 19 and Theorem 23, it correctly implements $G_{\mathsf{intrl}}$. ### 5.2 Another Example of Delegation Here, we further demonstrate our support for interleaving, showing how a participant can delegate the rest of its interactions in a protocol. The following global type formalizes a protocol in which a Client ($c$) asks an online Password Manager ($p$) to login with a Server ($s$): $\displaystyle G_{\mathsf{deleg}}:=c\mathbin{\twoheadrightarrow}p{:}\mathsf{login}\langle S\rangle\mathbin{.}G^{\prime}_{\mathsf{deleg}}$ where $\displaystyle S$ $\displaystyle:={!}({?}\mathsf{bool}\mathbin{.}\bullet)\mathbin{.}S^{\prime}$ $\displaystyle S^{\prime}$ $\displaystyle:=\&\\{\mathsf{passwd}{:}\leavevmode\nobreak\ {?}\mathsf{str}\mathbin{.}{\oplus}\\{\mathsf{auth}{:}\leavevmode\nobreak\ {!}\mathsf{bool}\mathbin{.}\bullet\\}\\}$ $\displaystyle G^{\prime}_{\mathsf{deleg}}$ $\displaystyle:=c\mathbin{\twoheadrightarrow}s{:}\mathsf{passwd}\langle\mathsf{str}\rangle\mathbin{.}s\mathbin{\twoheadrightarrow}c{:}\mathsf{auth}\langle\mathsf{bool}\rangle\mathbin{.}\bullet$ Here $S^{\prime}$ expresses the type of $\mathcal{R}_{c}$’s channel endpoint ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}$. This means that we can give implementations for $c$ and $p$ such that $c$ can send its channel endpoint ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ to $p$, after which $p$ logs in with $s$ in $c$’s place, forwarding the authorization boolean received from $s$ to $c$. Giving such implementations is relatively straightforward, demonstrating the flexibility of our global types and analysis using APCP and routers. Using local projection, we can compute a type for $c$’s implementation to safely connect with its router $\displaystyle G_{\mathsf{deleg}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}c={\oplus}^{0}\\{\mathsf{login}{:}\leavevmode\nobreak\ {{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}\mathbin{\otimes}^{1}(G^{\prime}_{\mathsf{deleg}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{4}c)\\}$ where $\displaystyle{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}$ $\displaystyle=(\bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\mathsf{o}}\bullet)\mathbin{\otimes}^{\kappa}{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S^{\prime}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}$ $\displaystyle{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S^{\prime}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}$ $\displaystyle=\&^{\pi}\\{\mathsf{passwd}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{\rho}{\oplus}^{\delta}\\{\mathsf{auth}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{\phi}\bullet\\}\\}$ $\displaystyle G^{\prime}_{\mathsf{deleg}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{4}c$ $\displaystyle={\oplus}^{4}\\{\mathsf{passwd}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{5}\&^{10}\\{\mathsf{auth}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{11}\bullet\\}\\}$ Notice how $\overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S^{\prime}{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}=G^{\prime}_{\mathsf{deleg}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{4}c$, given the assignments $\pi=4,\rho=5,\delta=10,\phi=11$. We can use these types to guide the design of a process implementation for $c$. Consider the process: $\displaystyle Q:=\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}\mathbin{\triangleleft}\mathsf{login}\cdot\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}[u]\cdot\overline{u}[v]\cdot(u\mathbin{\leftrightarrow}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\mathbin{|}v(a)\mathbin{.}\bm{0})\vdash\emptyset;{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ G_{\mathsf{deleg}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}c$ This implementation is interesting: after the first exchange in $G_{\mathsf{deleg}}$—sending a fresh channel $u$ (to $p$)—$c$ sends another fresh channel $v$ over $u$; then, $c$ delegates the rest of its exchanges in $G^{\prime}_{\mathsf{deleg}}$ by forwarding all traffic on ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ over $u$; in the meantime, $c$ awaits an authorization boolean over $v$. Again, using local projection, we can compute a type for $p$’s implementation to connect with its router: $\displaystyle G_{\mathsf{deleg}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}p=\&^{2}\\{\mathsf{login}{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}}\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{3}\bullet\\}$ We can then use it to type the following implementation for $p$: $\displaystyle P:={{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\mathbin{\triangleright}\left\\{\begin{array}[]{rl}\mathsf{login}{:}&{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})\mathbin{.}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}(v)\\\ &{}\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}\mathbin{\triangleleft}\mathsf{passwd}\cdot\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}}[\bm{pwd123}]\\\ &{}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}\mathbin{\triangleleft}\\{\mathsf{auth}{:}\leavevmode\nobreak\ {{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}(a)\mathbin{.}\overline{v}[a^{\prime}]\cdot a\mathbin{\leftrightarrow}a^{\prime}\\}\end{array}\right\\}\vdash\emptyset;{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ G_{\mathsf{deleg}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}p$ In this implementation, $p$ receives a channel ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ (from $c$) over which it first receives a channel $v$. Then, it behaves over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}$ according to $c$’s role in $G^{\prime}_{\mathsf{deleg}}$. Finally, $p$ forwards the authorization boolean received from $s$ over $v$, effectively sending the boolean to $c$. Given an implementation for $s$, say $S\vdash\emptyset;{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ G_{\mathsf{deleg}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}s$, what remains is to assign values to the remaining priorities in ${{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\llparenthesis}S{\color[rgb]{0.7421875,0.2890625,0}\definecolor[named]{pgfstrokecolor}{rgb}{0.7421875,0.2890625,0}\rrparenthesis}}$: assigning $\mathsf{o}=12,\kappa=4$ works. Now, we can compose the implementations $P$, $Q$ and $S$ with their respective routers and then compose these routed implementations together to form a deadlock free network of $G_{\mathsf{deleg}}$. This way, e.g., the router for $c$ is as follows (again, omitting curly braces for branches on a single label): $\displaystyle\mathcal{R}_{c}$ $\displaystyle={{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}\mathbin{\triangleright}\mathsf{login}\mathbin{.}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}\mathbin{\triangleleft}\mathsf{login}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}(u)\mathbin{.}\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}}[u^{\prime}]\cdot($ $\displaystyle\phantom{{}={}}\quad u\mathbin{\leftrightarrow}u^{\prime}\mathbin{|}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}\mathbin{\triangleright}\mathsf{passwd}\mathbin{.}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\mathbin{\triangleleft}\mathsf{passwd}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}(v)\mathbin{.}\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}}[v^{\prime}]\cdot($ $\displaystyle\phantom{{}={}}\qquad v\mathbin{\leftrightarrow}v^{\prime}\mathbin{|}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\mathbin{\triangleright}\mathsf{auth}\mathbin{.}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}\mathbin{\triangleleft}\mathsf{auth}\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}(w)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}}[w^{\prime}]\cdot(w\mathbin{\leftrightarrow}w^{\prime}\mathbin{|}\bm{0})))$ Interestingly, the router is agnostic of the fact that the endpoint $u$ it receives over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}$ is in fact the opposite endpoint of the channel formed by ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}$. ### 5.3 The Authorization Protocol in Action Let us repeat $G_{\mathsf{auth}}$ from Section 1: $\displaystyle G_{\mathsf{auth}}=\mu X\mathbin{.}s\mathbin{\twoheadrightarrow}c\left\\{\begin{array}[]{@{}l@{}}\mathsf{login}\mathbin{.}c\mathbin{\twoheadrightarrow}a{:}\mathsf{passwd}\langle\mathsf{str}\rangle\mathbin{.}a\mathbin{\twoheadrightarrow}s{:}\mathsf{auth}\langle\mathsf{bool}\rangle\mathbin{.}X,\\\ \mathsf{quit}\mathbin{.}c\mathbin{\twoheadrightarrow}a{:}\mathsf{quit}\mathbin{.}\bullet\end{array}\right\\}$ The relative projections of $G_{\mathsf{auth}}$ are as follows: $\displaystyle G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(s,a)$ $\displaystyle=\mu X\mathbin{.}s{!}c\left\\{\begin{array}[]{@{}l@{}}\mathsf{login}\mathbin{.}\mathsf{skip}\mathbin{.}a{:}\mathsf{auth}\langle\mathsf{bool}\rangle\mathbin{.}X,\\\ \mathsf{quit}\mathbin{.}\mathsf{skip}\mathbin{.}\bullet\end{array}\right\\}$ $\displaystyle G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(c,a)$ $\displaystyle=\mu X\mathbin{.}c{?}s\left\\{\begin{array}[]{@{}l@{}}\mathsf{login}\mathbin{.}c{:}\mathsf{passwd}\langle\mathsf{str}\rangle\mathbin{.}\mathsf{skip}\mathbin{.}X,\\\ \mathsf{quit}\mathbin{.}c{:}\mathsf{quit}\mathbin{.}\bullet\end{array}\right\\}$ $\displaystyle G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(s,c)$ $\displaystyle=\mu X\mathbin{.}s\left\\{\begin{array}[]{@{}l@{}}\mathsf{login}\mathbin{.}\mathsf{skip}^{2}\mathbin{.}X,\\\ \mathsf{quit}\mathbin{.}\mathsf{skip}\mathbin{.}\bullet\end{array}\right\\}$ $\displaystyle\mathcal{R}_{c}$ $\displaystyle=\mu X({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}})\mathbin{.}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\mathbin{\triangleright}\left\\{\begin{array}[]{ll}\mathsf{login}{:}&\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}}\mathbin{\triangleleft}\mathsf{login}\cdot\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}}\mathbin{\triangleleft}\mathsf{login}\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}(u)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}}[u^{\prime}]\\\ &{}\cdot(u\mathbin{\leftrightarrow}u^{\prime}\mathbin{|}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}\mathbin{\triangleright}\left\\{\begin{array}[]{ll}\mathsf{passwd}{:}&\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}}\mathbin{\triangleleft}\mathsf{passwd}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}(v)\mathbin{.}\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}}[v^{\prime}]\\\ &{}\cdot(v\mathbin{\leftrightarrow}v^{\prime}\mathbin{|}X{\langle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}\rangle})\end{array}\right\\}),\\\ \mathsf{quit}{:}&\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}}\mathbin{\triangleleft}\mathsf{quit}\cdot\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}}\mathbin{\triangleleft}\mathsf{quit}\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}(w)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}}[w^{\prime}]\\\ &{}\cdot(w\mathbin{\leftrightarrow}w^{\prime}\mathbin{|}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}\mathbin{\triangleright}\left\\{\begin{array}[]{ll}\mathsf{quit}{:}&\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}}\mathbin{\triangleleft}\mathsf{quit}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}(z)\mathbin{.}\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}}[z^{\prime}]\\\ &\cdot(z\mathbin{\leftrightarrow}z^{\prime}\mathbin{|}\bm{0})\end{array}\right\\})\end{array}\right\\}$ $\displaystyle\vdash\begin{array}[t]{l}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}{:}\leavevmode\nobreak\ \mu X\mathbin{.}{\oplus}^{2}\left\\{\begin{array}[]{ll}\mathsf{login}{:}&\bullet\mathbin{\otimes}^{3}\&^{4}\\{\mathsf{passwd}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{5}X\\},\\\ \mathsf{quit}{:}&\bullet\mathbin{\otimes}^{3}\&^{4}\\{\mathsf{quit}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{5}\bullet\\}\end{array}\right\\}=\overline{(G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}c)},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}{:}\leavevmode\nobreak\ \mu X\mathbin{.}\&^{1}\\{\mathsf{login}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{2}X,\mathsf{quit}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{2}\bullet\\}={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(c,s){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{c{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}s}^{0},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}{:}\leavevmode\nobreak\ \mu X\mathbin{.}{\oplus}^{2}\left\\{\begin{array}[]{ll}\mathsf{login}{:}&{\oplus}^{5}\\{\mathsf{passwd}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{6}X\\},\\\ \mathsf{quit}{:}&{\oplus}^{5}\\{\mathsf{quit}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{6}\bullet\\}\end{array}\right\\}={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(c,a){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{c{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}a}^{0}\end{array}$ $\displaystyle\mathcal{R}_{s}$ $\displaystyle=\mu X({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}})\mathbin{.}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}\mathbin{\triangleright}\left\\{\begin{array}[]{ll}\mathsf{login}{:}&\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}}\mathbin{\triangleleft}\mathsf{login}\cdot\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}}\mathbin{\triangleleft}\mathsf{login}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}(u)\mathbin{.}\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}}[u^{\prime}]\\\ &{}\cdot(u\mathbin{\leftrightarrow}u^{\prime}\mathbin{|}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}\mathbin{\triangleright}\left\\{\begin{array}[]{rl}\mathsf{auth}{:}&\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}}\mathbin{\triangleleft}\mathsf{auth}\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}(v)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}}[v^{\prime}]\\\ &{}\cdot(v\mathbin{\leftrightarrow}v^{\prime}\mathbin{|}X{\langle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}\rangle})\end{array}\right\\}),\\\ \mathsf{quit}{:}&\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}}\mathbin{\triangleleft}\mathsf{quit}\cdot\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}}\mathbin{\triangleleft}\mathsf{quit}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}(v)\mathbin{.}\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}}[v^{\prime}]\\\ &\cdot(v\mathbin{\leftrightarrow}v^{\prime}\mathbin{|}\bm{0})\end{array}\right\\}$ $\displaystyle\vdash\begin{array}[t]{l}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}{:}\leavevmode\nobreak\ \mu X\mathbin{.}\&^{0}\\{\mathsf{login}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{1}{\oplus}^{10}\\{\mathsf{auth}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{11}X\\},\mathsf{quit}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{1}\bullet\\}=\overline{(G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}s)},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}{:}\leavevmode\nobreak\ \mu X\mathbin{.}{\oplus}^{1}\\{\mathsf{login}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{2}X,\mathsf{quit}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{2}\bullet\\}={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(s,c){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{s{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}c}^{0},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}{:}\leavevmode\nobreak\ \mu X\mathbin{.}{\oplus}^{1}\\{\mathsf{login}{:}\leavevmode\nobreak\ \&^{9}\\{\mathsf{auth}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{10}X\\},\mathsf{quit}{:}\leavevmode\nobreak\ \bullet\\}={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(s,a){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{s{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}a}^{0}\end{array}$ $\displaystyle\mathcal{R}_{a}$ $\displaystyle=\mu X({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}})\mathbin{.}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\mathbin{\triangleright}\left\\{\begin{array}[]{l}\mathsf{login}{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}}\mathbin{\triangleleft}\mathsf{login}\\\ {}\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}\mathbin{\triangleright}\left\\{\begin{array}[]{l}\mathsf{login}{:}\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}\mathbin{\triangleright}\left\\{\begin{array}[]{l}\mathsf{passwd}{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}}\mathbin{\triangleleft}\mathsf{passwd}\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}(u)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}}[u^{\prime}]\\\ {}\cdot\left(\begin{array}[]{l}u\mathbin{\leftrightarrow}u^{\prime}\\\ {}\mathbin{|}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}\mathbin{\triangleright}\left\\{\begin{array}[]{l}\mathsf{auth}{:}\\\ \overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}}\mathbin{\triangleleft}\mathsf{auth}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}(v)\mathbin{.}\overline{{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}}[v^{\prime}]\\\ {}\cdot(v\mathbin{\leftrightarrow}v^{\prime}\mathbin{|}X{\langle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}\rangle})\end{array}\right\\}\end{array}\right)\end{array}\right\\},\\\ \mathsf{quit}{:}\leavevmode\nobreak\ {\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}})}\end{array}\right\\},\\\ \mathsf{quit}{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}}\mathbin{\triangleleft}\mathsf{quit}\\\ {}\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}\mathbin{\triangleright}\left\\{\begin{array}[]{l}\mathsf{login}{:}\leavevmode\nobreak\ {\mathchoice{\textstyle}{}{}{}\mathsf{alarm}({{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}})},\\\ \mathsf{quit}{:}\leavevmode\nobreak\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}\mathbin{\triangleright}\\{\mathsf{quit}{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}}\mathbin{\triangleleft}\mathsf{quit}\cdot{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}(w)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}}[w^{\prime}]\cdot(w\mathbin{\leftrightarrow}w^{\prime}\mathbin{|}\bm{0})\\},\end{array}\right\\}\end{array}\right\\}$ $\displaystyle\vdash\begin{array}[t]{l}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}{:}\leavevmode\nobreak\ \mu X\mathbin{.}{\oplus}^{2}\left\\{\begin{array}[]{ll}\mathsf{login}{:}&{\oplus}^{6}\\{\mathsf{passwd}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{7}\&^{8}\\{\mathsf{auth}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{9}X\\}\\},\\\ \mathsf{quit}{:}&{\oplus}^{6}\\{\mathsf{quit}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{7}\bullet\\}\end{array}\right\\}=\overline{(G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}a)},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}{:}\leavevmode\nobreak\ \mu X\mathbin{.}\&^{2}\left\\{\begin{array}[]{ll}\mathsf{login}{:}&\&^{5}\\{\mathsf{passwd}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{6}X\\},\\\ \mathsf{quit}{:}&\&^{5}\\{\mathsf{quit}{:}\leavevmode\nobreak\ \bullet\mathbin{\mathchoice{\rotatebox[origin={c}]{180.0}{$\displaystyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\textstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptstyle{\&}$}}{\rotatebox[origin={c}]{180.0}{$\scriptscriptstyle{\&}$}}}^{6}\bullet\\}\end{array}\right\\}={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(a,c){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{a{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}c}^{0},\\\ {{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}{:}\leavevmode\nobreak\ \mu X\mathbin{.}\&^{1}\\{\mathsf{login}{:}\leavevmode\nobreak\ {\oplus}^{9}\\{\mathsf{auth}{:}\leavevmode\nobreak\ \bullet\mathbin{\otimes}^{10}X\\},\mathsf{quit}{:}\leavevmode\nobreak\ \bullet\\}={{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\llparenthesis}\mkern-1.0muG_{\mathsf{auth}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(a,s){\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rrparenthesis}}_{a{\color[rgb]{0,0.56640625,0.61328125}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.56640625,0.61328125}\rangle}s}^{0}\end{array}$ Figure 14: Routers synthesized from $G_{\mathsf{auth}}$. The typed routers synthesized from $G_{\mathsf{auth}}$ are given in Figure 14. Let us explain the behavior of $\mathcal{R}_{a}$, the router of $a$. $\mathcal{R}_{a}$ is a recursive process on recursion variable $X$, using the endpoint for the implementation ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}$ and the endpoint for the other routers ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}$ and ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$ as context. The initial message in $G_{\mathsf{auth}}$ from $s$ to $c$ is a dependency for $a$’s interactions with both $s$ and $c$. Therefore, the router first branches on the first dependency with $s$: a label received over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$ (login or quit). Let us detail the login branch. Here, the router sends login over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}$. Then, the router branches on the second dependency with $c$: a label received over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}$ (again, login or quit). * • In the second login branch, the router receives the label passwd over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}$, which it then sends over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}$. The router then receives an endpoint (the password) over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}$, which it forwards over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}$. Finally, the router receives the label auth over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}$, which it sends over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$. Then, the router receives an endpoint (the authorization result) over ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}$, which it forwards over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$. The router then recurses to the beginning of the loop on the recursion variable $X$, passing the endpoints ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}$ as recursive context. * • In the quit branch, the router is in an inconsistent state, because it has received a label over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}$ which does not concur with the label received over ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}$. Hence, the router signals an alarm on its endpoints ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}},{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}$. Notice how the typing of the routers in Figure 14 follows Theorem 11: for each $p\in\\{c,s,a\\}$, the endpoint ${{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}p}}$ is typed with local projection (Def. 22), and for each $q\in\\{c,s,a\\}\setminus\\{p\\}$ the endpoint ${{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}p}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}q}}$ is typed with relative projection (Defs. 16 and 23). $\displaystyle{\mathsf{O}}_{\\{c,s,a\\}}[G_{\mathsf{auth}}]$ $\displaystyle\hskip 10.00002pt=\begin{array}[t]{l}\mu X({{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}},{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}},{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}})\\\ {}\mathbin{.}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}\mathbin{\triangleright}\left\\{\begin{array}[]{ll}\mathsf{login}{:}&\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}}\mathbin{\triangleleft}\mathsf{login}\cdot\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}}\mathbin{\triangleleft}\mathsf{login}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}(u)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}}[u^{\prime}]\\\ &{}\cdot(u\mathbin{\leftrightarrow}u^{\prime}\mathbin{|}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}\mathbin{\triangleright}\left\\{\begin{array}[]{ll}\mathsf{auth}{:}&\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}}\mathbin{\triangleleft}\mathsf{passwd}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}(v)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}}[v^{\prime}]\\\ &{}\cdot(v\mathbin{\leftrightarrow}v^{\prime}\mathbin{|}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}\mathbin{\triangleright}\left\\{\begin{array}[]{ll}\mathsf{auth}{:}&\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}}\mathbin{\triangleleft}\mathsf{auth}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}(w)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}}[w^{\prime}]\\\ &{}\cdot(w\mathbin{\leftrightarrow}w^{\prime}\mathbin{|}X{\langle{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}},{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}},{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}\rangle})\end{array}\right\\})\end{array}\right\\}),\\\ \\\ \mathsf{quit}{:}&\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}}\mathbin{\triangleleft}\mathsf{quit}\cdot\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}}\mathbin{\triangleleft}\mathsf{quit}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}(z)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}}[z^{\prime}]\\\ &{}\cdot(z\mathbin{\leftrightarrow}z^{\prime}\mathbin{|}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}\mathbin{\triangleright}\\{\mathsf{quit}{:}\leavevmode\nobreak\ \overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}}\mathbin{\triangleleft}\mathsf{quit}\cdot{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}(y)\mathbin{.}\overline{{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}}[y^{\prime}]\cdot(y\mathbin{\leftrightarrow}y^{\prime}\mathbin{|}\bm{0})\\})\end{array}\right\\}\end{array}$ $\displaystyle\hskip 10.00002pt\vdash{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}{:}\leavevmode\nobreak\ \overline{G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}c},{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}{:}\leavevmode\nobreak\ \overline{G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}s},{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}{:}\leavevmode\nobreak\ \overline{G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}a}$ Figure 15: Orchestrator synthesized from $G_{\mathsf{auth}}$ (cf. Def. 36). Consider again the participant implementations given in Example 1: $P$ implements the role of $c$, $Q$ the role of $s$, and $R$ the role of $a$. Notice that the types of the channels of these processes coincide with relative projections: $\displaystyle P$ $\displaystyle\vdash\emptyset;{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}c$ $\displaystyle Q$ $\displaystyle\vdash\emptyset;{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}s$ $\displaystyle R$ $\displaystyle\vdash\emptyset;{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}{:}\leavevmode\nobreak\ G_{\mathsf{auth}}\mathbin{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\downharpoonright}}^{0}a$ Let us explore how to compose these implementations with their respective routers. The order of composition determines the network topology. Decentralized By first composing each router with their respective implementation, and then composing the resulting routed implementations, we obtain a decentralized topology: $N_{\mathsf{auth}}^{\mathsf{decentralized}}:=\begin{array}[]{r}(\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}})\\\ (\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}})\\\ (\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}})\end{array}\left(\begin{array}[]{l}\phantom{{}\mathbin{|}{}}(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})\\\ {}\mathbin{|}(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})\\\ {}\mathbin{|}(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})\end{array}\begin{array}[]{l}(\mathcal{R}_{c}\mathbin{|}P)\\\ (\mathcal{R}_{s}\mathbin{|}Q)\\\ (\mathcal{R}_{a}\mathbin{|}R)\end{array}\right)$ This composition is in fact a network of routed implementations of $G$ (cf. Def. 25), so Theorems 19, 23 and 18 apply: we have $N_{\mathsf{auth}}^{\mathsf{decentralized}}\in\mathrm{net}(G_{\mathsf{auth}})$, so $N_{\mathsf{auth}}^{\mathsf{decentralized}}$ behaves as specified by $G_{\mathsf{auth}}$ and is deadlock free. Centralized By first composing the routers, and then composing the connected routers with each implementation, we obtain a centralized topology: $N_{\mathsf{auth}}^{\mathsf{centralized}}:=\begin{array}[]{c}(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})\\\ (\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})\\\ (\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})\end{array}\left(\begin{array}[]{l}\begin{array}[]{c}(\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}})\\\ (\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}c}})\\\ (\bm{\nu}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}a}}_{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}s}})\end{array}\left(\begin{array}[]{l}\phantom{{}\mathbin{|}{}}\mathcal{R}_{c}\\\ {}\mathbin{|}\mathcal{R}_{s}\\\ {}\mathbin{|}\mathcal{R}_{a}\end{array}\right)\begin{array}[]{l}{}\mathbin{|}P\\\ {}\mathbin{|}Q\\\ {}\mathbin{|}R\end{array}\end{array}\right)$ Note that the composition of routers is a _hub of routers_ (Def. 37). Consider the composition of $P$, $Q$ and $R$ with the orchestrator of $G_{\mathsf{auth}}$ (given in Figure 15): $N_{\mathsf{auth}}^{\mathsf{orchestrator}}:=\begin{array}[]{c}(\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}c}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})\\\ (\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}s}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})\\\ (\bm{\nu}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}a}}_{{\color[rgb]{0.84765625,0,0.4140625}\definecolor[named]{pgfstrokecolor}{rgb}{0.84765625,0,0.4140625}\mu}})\end{array}\left({\mathsf{O}}_{\\{c,s,a\\}}[G_{\mathsf{auth}}]\begin{array}[]{l}{}\mathbin{|}P\\\ {}\mathbin{|}Q\\\ {}\mathbin{|}R\end{array}\right)$ By Theorem 27, the hub of routers and the orchestrator of $G_{\mathsf{auth}}$ are weakly bisimilar (Def. 39). Hence, $N_{\mathsf{auth}}^{\mathsf{centralized}}$ and $N_{\mathsf{auth}}^{\mathsf{orchestrator}}$ behave the same. Since each of $N_{\mathsf{auth}}^{\mathsf{top}}$ with $\mathsf{top}\in\\{\mathsf{decentralized},\mathsf{centralized},\mathsf{orchestrator}\\}$ is typable in empty contexts, by Theorem 18, each of these compositions is deadlock free. Moreover, $N_{\mathsf{auth}}^{\mathsf{decentralized}}$ and $N_{\mathsf{auth}}^{\mathsf{centralized}}$ are structurally congruent, so, by Theorems 19 and 23, they behave as prescribed by $G_{\mathsf{auth}}$. Finally, by Theorem 27, $N_{\mathsf{auth}}^{\mathsf{centralized}}$ and $N_{\mathsf{auth}}^{\mathsf{orchestrator}}$ are bisimilar, and so $N_{\mathsf{auth}}^{\mathsf{orchestrator}}$ also behaves as prescribed by $G_{\mathsf{auth}}$. ## 6 Related Work ###### Types for Deadlock Freedom Our decentralized analysis of global types is related to type systems that ensure deadlock freedom for multiparty sessions with delegation and interleaving [7, 44, 21]. Unlike these works, we rely on a type system for _binary_ sessions which is simple and enables an expressive analysis of global types. Coppo _et al._ [7, 20, 21] give type systems for multiparty protocols, with asynchrony and support for interleaved sessions by tracking of mutual dependencies between them; as per Toninho and Yoshida [50], our example in Section 5.1 is typable in APCP but untypable in their system. Padovani _et al._ [44] develop a type system that enforces liveness properties for multiparty sessions, defined on top of a $\pi$-calculus with labeled communication. Rather than global types, their type structure follows approaches based on _conversation types_ [15]. Toninho and Yoshida [50] analyze binary sessions, leveraging on deadlock freedom results for multiparty sessions to extend Wadler’s CLL [54] with cyclic networks. Their process language is synchronous and uses replication rather than recursion. We note that their Examples 6.8 and 6.9 can be typed in APCP (cf. § 5.1); a detailed comparison between their extended CLL and APCP is interesting future work. ###### MPST and Binary Analyses of Global Types There are many works on MPST and their integration into programming languages; see [38, 3] for surveys. Triggered by flawed proofs of type safety and limitations of usual theories, Scalas and Yoshida [48] define a meta-framework of multiparty protocols based on local types, without global types and projection. Their work has been a source of inspiration for our developments; we address similar issues by adopting relative types, instead of cutting ties with global types. As already mentioned, Caires and Pérez [12] and Carbone _et al._ [16] reduce the analysis of global types to binary session type systems based on intuitionistic and classical linear logic, respectively. Our routers strictly generalize the centralized mediums of Caires and Pérez (cf. § 4.4). We substantially improve over the expressivity of the decentralized approach of Carbone _et al._ based on coherence, but reliant on encodings into centralized arbiters; for instance, their approach does not support the example from Toninho and Yoshida [50] we discuss in § 5.1. Also, Caires and Pérez support neither recursive global types nor asynchronous communication, and neither do Carbone _et al._. Scalas _et al._ [47] leverage on an encoding of binary session types into _linear types_ [23, 41] to reduce multiparty sessions to processes typable with linear types, with applications in Scala programming. Their analysis is decentralized but covers processes with synchronous communication only; also, their deadlock freedom result is limited with respect to ours: it does not support interleaving, such as in the example in § 5.1. ###### Monitoring through MPST Our work and the works discussed so far all consider the verification of implementations of multiparty protocols through static type checking. Bocchi _et al._ [8] use a _dynamic_ approach: communication between implementations is enacted by _monitors_ , which are derived from the global type to prevent protocol violations. In their approach, Bocchi _et al._ rely on the traditional workflow for MPST: projection onto binary session types based on the merge operation. Interestingly, Bocchi _et al._ ’s semantics relies on _routing_ , which is similar in spirit, but not in details, to our routers: their routing approach abstracts away from the actual network structure, while our routers enable the concrete realization of a decentralized network structure. We also note that Bocchi _et al._ ’s monitors, based on finite state machines, live on the level of semantics, while our routers, $\pi$-calculus processes, live on the same level as implementations. The theory by Bocchi _et al._ has resulted in the development of tools for a practical application of monitoring in Python [25], including an extension to real-time systems [43]. ###### Other Approaches to Multiparty Protocols In a broader context, Message Sequence Charts (MSCs) provide graphical specifications of multiparty protocols. Alur _et al._ [2] and Abdallah _et al._ [1] study the decidability of model-checking properties such as implementability of MSC Graphs and High-level MSCs (HMSCs) as Communicating FSMs (CFSMs). Genest _et al._ [32] study the synthesis of implementations of HMSCs as CFSMs; as we do, they use extra synchronization messages in some cases. We follow an entirely different research strand: our analysis is type- based and targets well-formed global types that are implementable by design. We note that the decidability of key notions for MPST (such as well-formedness and typability) has been addressed in [36]. Collaboration diagrams are another visual model for communicating processes (see, e.g. [10]). Salaün _et al._ [46] encode collaboration diagrams into the LOTOS process algebra [28] to enable model-checking [30], realizability checks for synchronous and asynchronous communication, and synthesis of participant implementations. Their implementation synthesis is reminiscent of our router synthesis, and also adds extra synchronization messages to realize otherwise unrealizable protocols with non-local choices. ## 7 Conclusion We have developed a new analysis of multiparty protocols specified as global types. One distinguishing feature of our analysis is that it accounts for multiparty protocols implemented by arbitrary process networks, which can be centralized (as in orchestration-based approaches) but also decentralized (as in choreography-based approaches). Another salient feature is that we can ensure both protocol conformance (protocol fidelity, communication safety) and deadlock freedom, which is notoriously hard to establish for protocols/implementations involving delegation and interleaving. To this end, we have considered asynchronous process implementations in APCP, the typed process language that we introduced in [51]. Our analysis enables the transference of correctness properties from APCP to multiparty protocols. We have illustrated these features using the authorization protocol $G_{\mathsf{auth}}$ adapted from Scalas and Yoshida [48] as a running example; additional examples further justify how our approach improves over previous analyses (cf. Section 5). Our analysis of multiparty protocols rests upon three key innovations: _routers_ , which enable global type analysis as decentralized networks; _relative types_ that capture protocols between pairs of participants; _relative projection_ , which admits global types with non-local choices. In our opinion, these notions are interesting on their own. In particular, relative types shed new light on more expressive protocol specifications than usual MPST, which are tied to notions of local types and merge/subtyping. There are several interesting avenues for future work. Comparing relative and merge-based well-formedness would continue the tread of new projections of global types (cf. App. A for initial findings). We would also like to develop a type system based on relative types, integrating the logic of routers into a static type checking that ensures deadlock freedom for processes. Finally, we are interested in developing practical tool support based on our findings. For this latter point, following [40], we would like to first formalize a theory of runtime monitoring based on routers, which can already be seen as an elementary form of _choreographed monitoring_ (cf. [29]). ##### Acknowledgments We are grateful to the anonymous reviewers for their constructive feedback and suggestions, which were enormously helpful to improve the presentation. Research partially supported by the Dutch Research Council (NWO) under project No. 016.Vidi.189.046 (Unifying Correctness for Communicating Software). ## References * [1] Rouwaida Abdallah, Loïc Hélouët, and Claude Jard. Distributed implementation of message sequence charts. Software & Systems Modeling, 14(2):1029–1048, May 2015. doi:10.1007/s10270-013-0357-1. * [2] Rajeev Alur, Kousha Etessami, and Mihalis Yannakakis. Realizability and verification of MSC graphs. Theoretical Computer Science, 331(1):97–114, February 2005. doi:10.1016/j.tcs.2004.09.034. * [3] Davide Ancona, Viviana Bono, Mario Bravetti, Joana Campos, Giuseppe Castagna, Pierre-Malo Deniélou, Simon J. Gay, Nils Gesbert, Elena Giachino, Raymond Hu, Einar Broch Johnsen, Francisco Martins, Viviana Mascardi, Fabrizio Montesi, Rumyana Neykova, Nicholas Ng, Luca Padovani, Vasco T. Vasconcelos, and Nobuko Yoshida. Behavioral Types in Programming Languages. Foundations and Trends® in Programming Languages, 3(2-3):95–230, July 2016. doi:10.1561/2500000031. * [4] Robert Atkey, Sam Lindley, and J. Garrett Morris. Conflation Confers Concurrency. In Sam Lindley, Conor McBride, Phil Trinder, and Don Sannella, editors, A List of Successes That Can Change the World: Essays Dedicated to Philip Wadler on the Occasion of His 60th Birthday, Lecture Notes in Computer Science, pages 32–55. Springer International Publishing, Cham, 2016. doi:10.1007/978-3-319-30936-1_2. * [5] Franco Barbanera and Mariangiola Dezani-Ciancaglini. Open Multiparty Sessions. Electronic Proceedings in Theoretical Computer Science, 304:77–96, September 2019. arXiv:1909.05972, doi:10.4204/EPTCS.304.6. * [6] Andi Bejleri, Elton Domnori, Malte Viering, Patrick Eugster, and Mira Mezini. Comprehensive Multiparty Session Types. The Art, Science, and Engineering of Programming, 3(3):6:1–6:59, February 2019. doi:10.22152/programming-journal.org/2019/3/6. * [7] Lorenzo Bettini, Mario Coppo, Loris D’Antoni, Marco De Luca, Mariangiola Dezani-Ciancaglini, and Nobuko Yoshida. Global Progress in Dynamically Interleaved Multiparty Sessions. In Franck van Breugel and Marsha Chechik, editors, CONCUR 2008 - Concurrency Theory, Lecture Notes in Computer Science, pages 418–433, Berlin, Heidelberg, 2008. Springer. doi:10.1007/978-3-540-85361-9_33. * [8] Laura Bocchi, Tzu-Chun Chen, Romain Demangeon, Kohei Honda, and Nobuko Yoshida. Monitoring Networks through Multiparty Session Types. Theoretical Computer Science, 669:33–58, March 2017. doi:10.1016/j.tcs.2017.02.009. * [9] Gérard Boudol. Asynchrony and the Pi-calculus. Research Report RR-1702, INRIA, 1992. * [10] Tevfik Bultan and Xiang Fu. Specification of realizable service conversations using collaboration diagrams. Service Oriented Computing and Applications, 2(1):27–39, April 2008\. doi:10.1007/s11761-008-0022-7. * [11] Luís Caires. Types and Logic, Concurrency and Non-Determinism. Technical Report MSR-TR-2014-104, In Essays for the Luca Cardelli Fest, Microsoft Research, September 2014. * [12] Luís Caires and Jorge A. Pérez. Multiparty Session Types Within a Canonical Binary Theory, and Beyond. In Elvira Albert and Ivan Lanese, editors, Formal Techniques for Distributed Objects, Components, and Systems, Lecture Notes in Computer Science, pages 74–95. Springer International Publishing, 2016. doi:10.1007/978-3-319-39570-8_6. * [13] Luís Caires and Jorge A. Pérez. Linearity, Control Effects, and Behavioral Types. In Hongseok Yang, editor, Programming Languages and Systems, Lecture Notes in Computer Science, pages 229–259, Berlin, Heidelberg, 2017. Springer. doi:10.1007/978-3-662-54434-1_9. * [14] Luís Caires and Frank Pfenning. Session Types as Intuitionistic Linear Propositions. In Paul Gastin and François Laroussinie, editors, CONCUR 2010 - Concurrency Theory, Lecture Notes in Computer Science, pages 222–236, Berlin, Heidelberg, 2010. Springer. doi:10.1007/978-3-642-15375-4_16. * [15] Luís Caires and Hugo Torres Vieira. Conversation types. Theoretical Computer Science, 411(51):4399–4440, December 2010\. doi:10.1016/j.tcs.2010.09.010. * [16] Marco Carbone, Sam Lindley, Fabrizio Montesi, Carsten Schürmann, and Philip Wadler. Coherence Generalises Duality: A Logical Explanation of Multiparty Session Types. In Josée Desharnais and Radha Jagadeesan, editors, 27th International Conference on Concurrency Theory (CONCUR 2016), volume 59 of Leibniz International Proceedings in Informatics (LIPIcs), pages 33:1–33:15, Dagstuhl, Germany, 2016. Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik. doi:10.4230/LIPIcs.CONCUR.2016.33. * [17] Marco Carbone, Nobuko Yoshida, and Kohei Honda. Asynchronous Session Types: Exceptions and Multiparty Interactions. In Marco Bernardo, Luca Padovani, and Gianluigi Zavattaro, editors, Formal Methods for Web Services: 9th International School on Formal Methods for the Design of Computer, Communication, and Software Systems, SFM 2009, Bertinoro, Italy, June 1-6, 2009, Advanced Lectures, Lecture Notes in Computer Science, pages 187–212. Springer, Berlin, Heidelberg, 2009. doi:10.1007/978-3-642-01918-0_5. * [18] Giuseppe Castagna, Mariangiola Dezani-Ciancaglini, and Luca Padovani. On Global Types and Multi-Party Session. Logical Methods in Computer Science, 8(1), March 2012. doi:10.2168/LMCS-8(1:24)2012. * [19] Ilaria Castellani, Mariangiola Dezani-Ciancaglini, Paola Giannini, and Ross Horne. Global types with internal delegation. Theoretical Computer Science, 807:128–153, February 2020. doi:10.1016/j.tcs.2019.09.027. * [20] Mario Coppo, Mariangiola Dezani-Ciancaglini, Luca Padovani, and Nobuko Yoshida. Inference of Global Progress Properties for Dynamically Interleaved Multiparty Sessions. In Rocco De Nicola and Christine Julien, editors, Coordination Models and Languages, Lecture Notes in Computer Science, pages 45–59, Berlin, Heidelberg, 2013. Springer. doi:10.1007/978-3-642-38493-6_4. * [21] Mario Coppo, Mariangiola Dezani-Ciancaglini, Nobuko Yoshida, and Luca Padovani. Global progress for dynamically interleaved multiparty sessions. Mathematical Structures in Computer Science, 26(2):238–302, February 2016. doi:10.1017/S0960129514000188. * [22] Ornela Dardha and Simon J. Gay. A New Linear Logic for Deadlock-Free Session-Typed Processes. In Christel Baier and Ugo Dal Lago, editors, Foundations of Software Science and Computation Structures, Lecture Notes in Computer Science, pages 91–109. Springer International Publishing, 2018\. doi:10.1007/978-3-319-89366-2_5. * [23] Ornela Dardha, Elena Giachino, and Davide Sangiorgi. Session types revisited. In Danny De Schreye, Gerda Janssens, and Andy King, editors, Principles and Practice of Declarative Programming, PPDP’12, Leuven, Belgium - September 19 - 21, 2012, pages 139–150. ACM, 2012. doi:10.1145/2370776.2370794. * [24] Ornela Dardha and Jorge A. Pérez. Comparing Deadlock-Free Session Typed Processes. Electronic Proceedings in Theoretical Computer Science, 190:1–15, August 2015. arXiv:1508.06707, doi:10.4204/EPTCS.190.1. * [25] Romain Demangeon, Kohei Honda, Raymond Hu, Rumyana Neykova, and Nobuko Yoshida. Practical interruptible conversations: Distributed dynamic verification with multiparty session types and Python. Formal Methods in System Design, 46(3):197–225, June 2015. doi:10.1007/s10703-014-0218-8. * [26] Pierre-Malo Deniélou and Nobuko Yoshida. Multiparty Compatibility in Communicating Automata: Characterisation and Synthesis of Global Session Types. In Fedor V. Fomin, Rūsiņš Freivalds, Marta Kwiatkowska, and David Peleg, editors, Automata, Languages, and Programming, Lecture Notes in Computer Science, pages 174–186, Berlin, Heidelberg, 2013. Springer. doi:10.1007/978-3-642-39212-2_18. * [27] Henry DeYoung, Luís Caires, Frank Pfenning, and Bernardo Toninho. Cut Reduction in Linear Logic as Asynchronous Session-Typed Communication. In Patrick Cégielski and Arnaud Durand, editors, Computer Science Logic (CSL’12) - 26th International Workshop/21st Annual Conference of the EACSL, volume 16 of Leibniz International Proceedings in Informatics (LIPIcs), pages 228–242, Dagstuhl, Germany, 2012. Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik. doi:10.4230/LIPIcs.CSL.2012.228. * [28] Ed Brinksma. LOTOS — A formal description technique based on the temporal ordering of observational behaviour. Technical Report ISO 8807:1989, International Organization for Standardization, February 1989. * [29] Adrian Francalanza, Jorge A. Pérez, and César Sánchez. Runtime Verification for Decentralised and Distributed Systems. In Ezio Bartocci and Yliès Falcone, editors, Lectures on Runtime Verification: Introductory and Advanced Topics, Lecture Notes in Computer Science, pages 176–210. Springer International Publishing, Cham, 2018. doi:10.1007/978-3-319-75632-5_6. * [30] Hubert Garavel, Radu Mateescu, Frédéric Lang, and Wendelin Serwe. CADP 2006: A Toolbox for the Construction and Analysis of Distributed Processes. In Werner Damm and Holger Hermanns, editors, Computer Aided Verification, Lecture Notes in Computer Science, pages 158–163, Berlin, Heidelberg, 2007. Springer. doi:10.1007/978-3-540-73368-3_18. * [31] Simon J. Gay, Peter Thiemann, and Vasco T. Vasconcelos. Duality of Session Types: The Final Cut. Electronic Proceedings in Theoretical Computer Science, 314:23–33, April 2020. arXiv:2004.01322, doi:10.4204/EPTCS.314.3. * [32] Blaise Genest, Anca Muscholl, Helmut Seidl, and Marc Zeitoun. Infinite-state high-level MSCs: Model-checking and realizability. Journal of Computer and System Sciences, 72(4):617–647, June 2006\. doi:10.1016/j.jcss.2005.09.007. * [33] Kohei Honda. Types for dyadic interaction. In Eike Best, editor, CONCUR’93, Lecture Notes in Computer Science, pages 509–523, Berlin, Heidelberg, 1993. Springer. doi:10.1007/3-540-57208-2_35. * [34] Kohei Honda and Mario Tokoro. An object calculus for asynchronous communication. In Pierre America, editor, ECOOP’91 European Conference on Object-Oriented Programming, Lecture Notes in Computer Science, pages 133–147, Berlin, Heidelberg, 1991. Springer. doi:10.1007/BFb0057019. * [35] Kohei Honda, Nobuko Yoshida, and Marco Carbone. Multiparty asynchronous session types. In Proceedings of the 35th Annual ACM SIGPLAN-SIGACT Symposium on Principles of Programming Languages, POPL ’08, pages 273–284, San Francisco, California, USA, January 2008. Association for Computing Machinery. doi:10.1145/1328438.1328472. * [36] Kohei Honda, Nobuko Yoshida, and Marco Carbone. Multiparty asynchronous session types. Journal of the ACM, 63(1), March 2016. doi:10.1145/2827695. * [37] Raymond Hu, Andi Bejleri, Nobuko Yoshida, and Pierre-Malo Denielou. Parameterised Multiparty Session Types. Logical Methods in Computer Science, Volume 8, Issue 4, October 2012\. doi:10.2168/LMCS-8(4:6)2012. * [38] Hans Hüttel, Ivan Lanese, Vasco T. Vasconcelos, Luís Caires, Marco Carbone, Pierre-Malo Deniélou, Dimitris Mostrous, Luca Padovani, António Ravara, Emilio Tuosto, Hugo Torres Vieira, and Gianluigi Zavattaro. Foundations of Session Types and Behavioural Contracts. ACM Comput. Surv., 49(1):3:1–3:36, April 2016. doi:10.1145/2873052. * [39] Keigo Imai, Rumyana Neykova, Nobuko Yoshida, and Shoji Yuen. Multiparty Session Programming With Global Protocol Combinators. In Robert Hirschfeld and Tobias Pape, editors, 34th European Conference on Object-Oriented Programming (ECOOP 2020), volume 166 of Leibniz International Proceedings in Informatics (LIPIcs), pages 9:1–9:30, Dagstuhl, Germany, 2020. Schloss Dagstuhl–Leibniz-Zentrum für Informatik. doi:10.4230/LIPIcs.ECOOP.2020.9. * [40] Limin Jia, Hannah Gommerstadt, and Frank Pfenning. Monitors and Blame Assignment for Higher-order Session Types. In Proceedings of the 43rd Annual ACM SIGPLAN-SIGACT Symposium on Principles of Programming Languages, POPL ’16, pages 582–594, New York, NY, USA, 2016. ACM. doi:10.1145/2837614.2837662. * [41] Naoki Kobayashi, Benjamin C. Pierce, and David N. Turner. Linearity and the pi-calculus. ACM Transactions on Programming Languages and Systems, 21(5):914–947, September 1999. doi:10.1145/330249.330251. * [42] Rupak Majumdar, Nobuko Yoshida, and Damien Zufferey. Multiparty motion coordination: From choreographies to robotics programs. Proceedings of the ACM on Programming Languages, 4(OOPSLA):134:1–134:30, November 2020. doi:10.1145/3428202. * [43] Rumyana Neykova, Laura Bocchi, and Nobuko Yoshida. Timed runtime monitoring for multiparty conversations. Formal Aspects of Computing, 29(5):877–910, September 2017. doi:10.1007/s00165-017-0420-8. * [44] Luca Padovani, Vasco Thudichum Vasconcelos, and Hugo Torres Vieira. Typing Liveness in Multiparty Communicating Systems. In Eva Kühn and Rosario Pugliese, editors, Coordination Models and Languages, Lecture Notes in Computer Science, pages 147–162, Berlin, Heidelberg, 2014. Springer. doi:10.1007/978-3-662-43376-8_10. * [45] C. Peltz. Web services orchestration and choreography. Computer, 36(10):46–52, October 2003. doi:10.1109/MC.2003.1236471. * [46] G. Salaün, T. Bultan, and N. Roohi. Realizability of Choreographies Using Process Algebra Encodings. IEEE Transactions on Services Computing, 5(3):290–304, Third 2012\. doi:10.1109/TSC.2011.9. * [47] Alceste Scalas, Ornela Dardha, Raymond Hu, and Nobuko Yoshida. A Linear Decomposition of Multiparty Sessions for Safe Distributed Programming. In Peter Müller, editor, 31st European Conference on Object-Oriented Programming (ECOOP 2017), volume 74 of Leibniz International Proceedings in Informatics (LIPIcs), pages 24:1–24:31, Dagstuhl, Germany, 2017. Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik. doi:10.4230/LIPIcs.ECOOP.2017.24. * [48] Alceste Scalas and Nobuko Yoshida. Less is more: Multiparty session types revisited. Proceedings of the ACM on Programming Languages, 3(POPL):30:1–30:29, January 2019. Revised, extended version at https://www.doc.ic.ac.uk/research/technicalreports/2018/DTRS18-6.pdf. doi:10.1145/3290343. * [49] Bernardo Toninho, Luis Caires, and Frank Pfenning. Corecursion and Non-divergence in Session-Typed Processes. In Matteo Maffei and Emilio Tuosto, editors, Trustworthy Global Computing, Lecture Notes in Computer Science, pages 159–175, Berlin, Heidelberg, 2014. Springer. doi:10.1007/978-3-662-45917-1_11. * [50] Bernardo Toninho and Nobuko Yoshida. Interconnectability of Session-Based Logical Processes. ACM Transactions on Programming Languages and Systems (TOPLAS), 40(4):17, December 2018. doi:10.1145/3242173. * [51] Bas van den Heuvel and Jorge A. Pérez. Deadlock Freedom for Asynchronous and Cyclic Process Networks (Extended Version). arXiv:2111.13091 [cs], November 2021. A short version appears in the Proceedings of ICE’21: arXiv:2110.00146. arXiv:2111.13091. * [52] W. M. P. van der Aalst. Orchestration. In Ling Liu and M. Tamer Özsu, editors, Encyclopedia of Database Systems, pages 2004–2005. Springer US, Boston, MA, 2009. doi:10.1007/978-0-387-39940-9_1197. * [53] Vasco T. Vasconcelos. Fundamentals of session types. Information and Computation, 217:52–70, August 2012. doi:10.1016/j.ic.2012.05.002. * [54] Philip Wadler. Propositions As Sessions. In Proceedings of the 17th ACM SIGPLAN International Conference on Functional Programming, ICFP ’12, pages 273–286, New York, NY, USA, 2012. ACM. doi:10.1145/2364527.2364568. * [55] Nobuko Yoshida and Lorenzo Gheri. A Very Gentle Introduction to Multiparty Session Types. In Dang Van Hung and Meenakshi D´Souza, editors, Distributed Computing and Internet Technology, Lecture Notes in Computer Science, pages 73–93, Cham, 2020. Springer International Publishing. doi:10.1007/978-3-030-36987-3_5. ## Appendix A Comparing Merge-based Well-formedness and Relative Well- formedness It is instructive to examine how the notion of well-formed global types induced by our relative projection compares to _merge-based_ well-formedness, the notion induced by (usual) local projection [37, 17]. Before we recall the definition of merge-based well-formedness, we define the projection of global types to local types. Local types express one particular participant’s perspective of a global protocol. Although $\mathsf{skip}$ is not part of standard definitions of local types, we include it to enable a fair comparison with relative types. ###### Definition 42 (Local types). _Local types_ $L$ are defined as follows, where the $S_{i}$ are the message types from Def. 11: $L::={?}p\\{i\langle S\rangle\mathbin{.}L\\}_{i\in I}\;\mbox{\large{$\mid$}}\;{!}p\\{i\langle S\rangle\mathbin{.}L\\}_{i\in I}\;\mbox{\large{$\mid$}}\;\mu X\mathbin{.}L\;\mbox{\large{$\mid$}}\;X\;\mbox{\large{$\mid$}}\;\bullet\;\mbox{\large{$\mid$}}\;\mathsf{skip}\mathbin{.}L$ The local types ${?}p\\{i\langle S_{i}\rangle\mathbin{.}L_{i}\\}_{i\in I}$ and ${!}p\\{i\langle S_{i}\rangle\mathbin{.}L_{i}\\}_{i\in I}$ represent receiving a choice from $p$ and sending a choice to $p$, respectively. All of $\bullet$, $\mu X\mathbin{.}L$, $X$, and $\mathsf{skip}$ are just as before. Instead of external dependencies, the projection onto local types relies on an operation on local types called _merge_. Intuitively, merge allows combining overlapping but not necessarily identical receiving constructs. This is one main difference with respect to our relative projection. ###### Definition 43 (Merge of Local Types). For local types $L_{1}$ and $L_{2}$, we define $L_{1}\sqcup L_{2}$ as the _merge_ of $L_{1}$ and $L_{2}$: $\displaystyle\mathsf{skip}\mathbin{.}L_{1}\sqcup\mathsf{skip}\mathbin{.}L_{2}$ $\displaystyle:=L_{1}\sqcup L_{2}$ $\displaystyle\bullet\sqcup\bullet$ $\displaystyle:=\bullet$ $\displaystyle\mu X\mathbin{.}L_{1}\sqcup\mu X\mathbin{.}L_{2}$ $\displaystyle:=\mu X\mathbin{.}(L_{1}\sqcup L_{2})$ $\displaystyle X\sqcup X$ $\displaystyle:=X$ $\displaystyle{!}p\\{i\langle S_{i}\rangle\mathbin{.}L_{i}\\}_{i\in I}\sqcup{!}p\\{i\langle S_{i}\rangle\mathbin{.}L_{i}\\}_{i\in I}$ $\displaystyle:={!}p\\{i\langle S_{i}\rangle\mathbin{.}L_{i}\\}_{i\in I}$ $\displaystyle{?}p\\{i\langle S_{i}\rangle\mathbin{.}L_{i}\\}_{i\in I}\sqcup{?}p\\{j\langle S^{\prime}_{j}\rangle\mathbin{.}L^{\prime}_{j}\\}_{j\in J}$ $\displaystyle:={?}p\left(\begin{array}[]{l}\phantom{{}\cup{}}\\{i\langle S_{i}\rangle\mathbin{.}L_{i}\\}_{i\in I\setminus J}\\\ {}\cup\\{j\langle S^{\prime}_{j}\rangle\mathbin{.}L^{\prime}_{j}\\}_{j\in J\setminus I}\\\ {}\cup\\{k\langle S_{k}\sqcup S^{\prime}_{k}\rangle\mathbin{.}(L_{k}\sqcup L^{\prime}_{k})\\}_{k\in I\cap J}\end{array}\right)$ The merge between message types $S_{1}\sqcup S_{2}$ corresponds to the identity function. If the local types do not match the above definition, their merge is undefined. We can now define local projection based on merge: ###### Definition 44 (Merge-based Local Projection). For global type $G$ and participant $p$, we define $G\mathbin{\upharpoonright}p$ as the _merge-based local projection_ of $G$ under $p$: $\displaystyle\bullet\mathbin{\upharpoonright}p$ $\displaystyle:=\bullet$ $\displaystyle(\mathsf{skip}\mathbin{.}G)\mathbin{\upharpoonright}p$ $\displaystyle:=\mathsf{skip}\mathbin{.}(G\mathbin{\upharpoonright}p)$ $\displaystyle X\mathbin{\upharpoonright}p$ $\displaystyle:=X$ $\displaystyle(\mu X\mathbin{.}G)\mathbin{\upharpoonright}p$ $\displaystyle:=\mathrlap{\begin{cases}\bullet&\text{if $G\mathbin{\upharpoonright}p=\mathsf{skip}^{\ast}\mathbin{.}\bullet$ or $G\mathbin{\upharpoonright}p=\mathsf{skip}^{\ast}\mathbin{.}X$}\\\ \mu X\mathbin{.}(G\mathbin{\upharpoonright}p)&\text{otherwise}\end{cases}}$ $\displaystyle(r\mathbin{\twoheadrightarrow}s\\{i\langle U_{i}\rangle\mathbin{.}G_{i}\\}_{i\in I})\mathbin{\upharpoonright}p$ $\displaystyle:=\mathrlap{\begin{cases}{?}r\\{i\langle U_{i}\rangle\mathbin{.}(G_{i}\mathbin{\upharpoonright}p)\\}_{i\in I}&\text{if $p=s$}\\\ {!}s\\{i\langle U_{i}\rangle\mathbin{.}(G_{i}\mathbin{\upharpoonright}p)\\}_{i\in I}&\text{if $p=r$}\\\ \mathsf{skip}\mathbin{.}(\sqcup_{i\in I}(G_{i}\mathbin{\upharpoonright}p))&\text{otherwise}\end{cases}}$ $\displaystyle(G_{1}\mathbin{|}G_{2})\mathbin{\upharpoonright}p$ $\displaystyle:=\mathrlap{\begin{cases}G_{1}\mathbin{\upharpoonright}p&\text{if $p\in\mathsf{prt}(G_{1})$ and $p\notin\mathsf{prt}(G_{2})$}\\\ G_{2}\mathbin{\upharpoonright}p&\text{if $p\in\mathsf{prt}(G_{2})$ and $p\notin\mathsf{prt}(G_{1})$}\\\ \bullet&\text{if $p\notin\mathsf{prt}(G_{1})\cup\mathsf{prt}(G_{2})$}\end{cases}}$ ###### Definition 45 (Merge Well-Formedness). A global type $G$ is _merge well-formed_ if, for every $p\in\mathsf{prt}(G)$, the merge-based local projection $G\mathbin{\upharpoonright}p$ is defined. The classes of relative and merge-based well-formed global types overlap: there are protocols that can be expressed using dependencies in relative types, as well as using merge in local types. Interestingly, the classes are _incomparable_ : some relative well-formed global types are not merge-based well-formed, and vice versa. We now explore these differences. ### A.1 Relative Well-Formed, Not Merge Well-Formed The merge of local types with outgoing messages of different labels is undefined. Therefore, if a global type has communications, e.g., from $s$ to $a$ with different labels across branches of a prior communication between $b$ and $a$, the global type is not merge well-formed. In contrast, such global types can be relative well-formed, because the prior communication may induce a dependency. Similarly, global types with communications with different participants across branches of a prior communication are never merge well- formed, but may be relative well-formed. The following example demonstrates a global type with messages of different labels across branches of a prior communication: ###### Example. We give an adaptation of the two-buyer-seller protocol in which Seller ($s$) tells Alice ($a$) to pay or not, depending on whether Bob ($b$) tells $a$ to buy or not. $\displaystyle G_{\mathsf{rwf}}:=b\mathbin{\twoheadrightarrow}a\left\\{\begin{array}[]{@{}l@{}}\mathsf{ok}\mathbin{.}s\mathbin{\twoheadrightarrow}a{:}\mathsf{pay}\langle\mathsf{int}\rangle\mathbin{.}\bullet,\\\ \mathsf{cancel}\mathbin{.}s\mathbin{\twoheadrightarrow}a{:}\mathsf{cancel}\mathbin{.}\bullet\end{array}\right\\}$ This protocol is relative well-formed, as the relative projections under every combination of participants are defined. Notice how there is a dependency in the relative projection under $s$ and $a$: $\displaystyle G_{\mathsf{rwf}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(s,a)=a{?}b\left\\{\begin{array}[]{@{}l@{}}\mathsf{ok}\mathbin{.}s{:}\mathsf{pay}\langle\mathsf{int}\rangle\mathbin{.}\bullet,\\\ \mathsf{cancel}\mathbin{.}s{:}\mathsf{cancel}\mathbin{.}\bullet\end{array}\right\\}$ However, we do not have merge well-formedness: the merge-based local projection under $s$ is not defined: $\displaystyle G_{\mathsf{rwf}}\mathbin{\upharpoonright}s=\mathsf{skip}\mathbin{.}({!}a{:}\mathsf{pay}\langle\mathsf{int}\rangle\mathbin{.}\bullet\sqcup{!}a{:}\mathsf{cancel}\mathbin{.}\bullet)$ ### A.2 Merge Well-Formed, Not Relative Well-Formed For a communication between, e.g., $a$ and $b$ to induce a dependency for subsequent communications between other participants, at least one of $a$ and $b$ must be involved. Therefore, global types where communications with participants other than $a$ and $b$ have different labels across branches of a prior communication between $a$ and $b$ are never relative well-formed. In contrast, merge can combine the reception of different labels, so such global types may be merge well-formed—as long as the sender is aware of which branch has been taken before. The following example demonstrates such a situation, and explains how such global types can be modified to be relative well-formed: ###### Example. Consider a variant of the two-buyer-seller protocol in which Seller ($s$) invokes a new participant, Mail-service ($m$), to deliver the requested product. In the following global type, Bob ($b$) tells Alice ($a$) of its decision to buy or not, after which $b$ sends the same choice to $s$, who then either invokes $m$ to deliver the product or not: $\displaystyle G_{\mathsf{mwf}}:=b\mathbin{\twoheadrightarrow}a\left\\{\begin{array}[]{@{}l@{}}\mathsf{ok}\mathbin{.}b\mathbin{\twoheadrightarrow}s{:}\mathsf{ok}\mathbin{.}s\mathbin{\twoheadrightarrow}m{:}\mathsf{deliver}\langle\mathsf{str}\rangle\mathbin{.}\bullet,\\\ \mathsf{quit}\mathbin{.}b\mathbin{\twoheadrightarrow}s{:}\mathsf{quit}\mathbin{.}s\mathbin{\twoheadrightarrow}m{:}\mathsf{quit}\mathbin{.}\bullet\end{array}\right\\}$ $G_{\mathsf{mwf}}$ is merge well-formed: the merge-based local projections under all participants are defined. Notice how the two different messages from $s$ are merged in the merge-based local projection under $m$: $\displaystyle G_{\mathsf{mwf}}\mathbin{\upharpoonright}m=\mathsf{skip}^{2}\mathbin{.}{?}s\left\\{\begin{array}[]{@{}l@{}}\mathsf{deliver}\langle\mathsf{str}\rangle\mathbin{.}\bullet,\\\ \mathsf{quit}\mathbin{.}\bullet\end{array}\right\\}$ $G_{\mathsf{mwf}}$ is not relative well-formed: the relative projection under $s$ and $m$ is not defined. The initial exchange between $b$ and $a$ cannot induce a dependency, since neither of $s$ and $m$ is involved. Hence, the relative projections of both branches must be identical, but they are not: $\displaystyle\mathsf{skip}\mathbin{.}s{:}\mathsf{deliver}\langle\mathsf{str}\rangle\mathbin{.}\bullet\neq\mathsf{skip}\mathbin{.}s{:}\mathsf{quit}\mathbin{.}\bullet$ We recover relative well-formedness by modifying $G_{\mathsf{mwf}}$: we give $s$ the same options to send to $m$ in both branches of the initial communication: $\displaystyle G^{\prime}_{\mathsf{mwf}}:=b\mathbin{\twoheadrightarrow}a\left\\{\begin{array}[]{@{}l@{}}\mathsf{ok}\mathbin{.}b\mathbin{\twoheadrightarrow}s{:}\mathsf{ok}\mathbin{.}s\mathbin{\twoheadrightarrow}m\\{\mathsf{deliver}\langle\mathsf{str}\rangle\mathbin{.}\bullet,\quad\mathsf{quit}\mathbin{.}\bullet\\},\\\ \mathsf{quit}\mathbin{.}b\mathbin{\twoheadrightarrow}s{:}\mathsf{quit}\mathbin{.}s\mathbin{\twoheadrightarrow}m\\{\mathsf{deliver}\langle\mathsf{str}\rangle\mathbin{.}\bullet,\quad\mathsf{quit}\mathbin{.}\bullet\\}\end{array}\right\\}$ The new protocol is still merge well-formed, but it is now relative well- formed too; the relative projection under $s$ and $m$ is defined: $\displaystyle G^{\prime}_{\mathsf{mwf}}\mathbin{{\color[rgb]{0.09765625,0,0.54296875}\definecolor[named]{pgfstrokecolor}{rgb}{0.09765625,0,0.54296875}\upharpoonright}}(s,m)=\mathsf{skip}^{2}\mathbin{.}s\left\\{\begin{array}[]{@{}l@{}}\mathsf{deliver}\langle\mathsf{address}\rangle\mathbin{.}\bullet,\\\ \mathsf{quit}\mathbin{.}\bullet\end{array}\right\\}$ This modification may not be ideal, though, because $s$ can quit the protocol even if $b$ has ok’ed the transaction, and that $s$ can still invoke a delivery even if $b$ has quit the transaction.
# Blast waves in a paraxial fluid of light Murad Abuzarli Tom Bienaimé Elisabeth Giacobino Alberto Bramati Quentin Glorieux<EMAIL_ADDRESS>ALaboratoire Kastler Brossel, Sorbonne Université, CNRS, ENS-Université PSL, Collège de France - Paris, France ###### Abstract We study experimentally blast wave dynamics on a weakly interacting fluid of light. The fluid density and velocity are measured in 1D and 2D geometries. Using a state equation arising from the analogy between optical propagation in the paraxial approximation and the hydrodynamic Euler’s equation, we access the fluid hydrostatic and dynamic pressure. In the 2D configuration, we observe a negative differential hydrostatic pressure after the fast expansion of a localized over-density, which is a typical signature of a blast wave for compressible gases. Our experimental results are compared to the Friedlander waveform hydrodynamical model[1]. Velocity measurements are presented in 1D and 2D configurations and compared to the local speed of sound, to identify supersonic region of the fluid. Our findings show an unprecedented control over hydrodynamic quantities in a paraxial fluid of light. ††preprint: APS/123-QED ## Introduction In classical hydrodynamics, a blast wave is characterized by an increased pressure and flow resulting from the rapid release of energy from a concentrated source [2]. The particular characteristics of a blast wave is that it is followed by a wind of negative pressure, which induces an attractive force back towards the origin of the shock. Typical blast waves occur after the detonation of trinitrotoluene [3, 4], nuclear fission [5], break of a pressurized container [6] or heating caused by a focused pulsed laser [7]. The sudden release of energy causes a rapid expansion, which in a three dimensional space is analogous to a spherical piston [8] and produces a compression wave in the ambient gas. For a fast enough piston, the compression wave develops into a shock wave which is characterized by the rapid increase of all the physical properties of the gas, namely, the hydrostatic pressure, density and particle velocity [9]. In 1946, Friedlander predicted that immediately after the shock front the physical properties at a given position in space decay exponentially [1, 10]. In this model, for 3-dimensional and 2-dimensional spaces the hydrostatic pressure and the density are expected to fall below the values of the ambient atmosphere leading to a blast wind [2]. Shock waves have been studied in several contexts in physics, including acoustics, plasma physics, ultra-cold atomic gases [11, 12, 13] and non-linear optics [14, 15, 16, 17, 6]. In optics, the hydrodynamics interpretation relies on the Madelung transforms which identify the light intensity to the fluid density and the phase gradient to the fluid velocity[18]. Recently several works have studied analytically shock wave formation in one and two dimensions [19, 20]. Optical systems allow for repeatable experiments and precise control of the experimental parameters. For example dispersive superfluid-like shock waves have been observed [14], as well as generation of solitons [16], shocks in non-local media [21, 15], shocks in disordered media [17], analogue dam break [6] and Riemann waves [22]. However, an experimental study of blast waves has not been done in atomic gases nor in non-linear optics systems. In this work, we demonstrate the generation of a blast wave in a fluid of light. Interestingly, the prediction of a blast wind with negative pressure and density holds in two dimensional space but not in 1 dimension [23]. Optical analogue systems allow for an experimental validation of this prediction. In this letter, we study the formation of blast waves in a paraxial fluid of light. We measure the time evolution of analogue physical properties such as the hydrostatic pressure, the density, the particle velocity and the dynamic pressure at a fixed point for 1 and 2-dimensional systems. We report the observation of a negative hydrostatic differential pressure after a shock wave in 2-dimensional system and we show that the Friedlander waveform describes quantitatively our experimental results for all physical parameters. This paper is organized as follows. We first introduce the analogy between the propagation equation of a laser beam through non-linear medium (a warm atomic vapor) and the hydrodynamics equation and derive the relevant analogue physical properties. In the second section of this work, we describe our experimental setup and present our results on the density and hydrostatic pressure measurements. We highlight the striking differences between 1 and 2-dimensional systems. Finally, we study the time evolution of the velocity and dynamic pressure. ## Theoretical Model We describe the propagation of a linearly polarized monochromatic beam in a local Kerr medium. We separate the electric field’s fast oscillating carrier from the slowly varying (with respect to the laser wavelength) envelope : $E=\mathcal{E}(\mathbf{r},z)$e${}^{i(kz-k_{0}ct)}+$ complex conjugate. Under the paraxial approximation, the propagation equation for the envelope $\mathcal{E}$ is the Non-Linear Schrödinger Equation (NLSE) [18]: $i\frac{\partial\mathcal{E}}{\partial z}=\left(-\frac{1}{2k}\nabla^{2}_{\perp}+g{\mid}\mathcal{E}{\mid}^{2}-\frac{i\alpha}{2}\right)\mathcal{E},$ (1) where $k$ is the laser wavevector in the medium, $\alpha$ is the extinction coefficient accounting for losses due to absorption, and the $g$ parameter is linked to the intensity dependent refractive index variation $\Delta n$ via: $g{\mid}\mathcal{E}{\mid}^{2}=-k_{0}\Delta n$ (with $k_{0}$ the laser wavevector in vacuum). The NLSE is analogous to a 2D Gross-Pitaevskii equation describing the dynamics of a quantum fluid in the mean-field approximation. This analogy is possible by mapping the envelope $\mathcal{E}$ to the quantum fluid many-body wavefunction and the axial coordinate $z$ to an effective evolution time. The non-linear refractive index variation plays then the role of a repulsive photon-photon interaction, since all measurements in this work are done in the self-defocusing regime i.e. $\Delta n<0$ and therefore $g>0$. Diffraction acts as kinetic energy with the effective mass emerging from the paraxial approximation and given by the laser wavevector $k=8.10^{6}$ m-1. Using the Madelung transformation: $\mathcal{E}=\sqrt{\rho}\textrm{e}^{i\phi}$, $\mathbf{v}=\frac{c}{k}\nabla_{\perp}\phi$ one can derive from the NLSE hydrodynamic equations [19, 14], linking the fluid’s density $\rho$ with its velocity $\mathbf{v}$: $\displaystyle\frac{\partial\rho}{\partial z}+\nabla_{\perp}.\left(\rho\frac{\mathbf{v}}{c}\right)=-\alpha\rho$ (2) $\displaystyle\frac{\partial\mathbf{v}}{\partial z}+\frac{1}{2c}\nabla_{\perp}\mathbf{v}^{2}=-\nabla_{\perp}\left(\frac{cg\rho}{k}-\frac{c}{2k^{2}\sqrt{\rho}}\nabla^{2}_{\perp}\sqrt{\rho}\right).$ (3) Eq. (2) is the continuity equation with a loss term accounting for photon absorption. Eq. (3) is similar to the Euler equation without viscosity, in which the driving force stems from interaction and the so-called quantum pressure term due to diffraction. Establishing the formal analogy requires, however, defining an analogue pressure $P$ to be able to re-express the right- hand side of Eq. (3) as: $-1/\rho\cdot\nabla_{\perp}P$. This is possible for the first term stemming from interactions. Using the identity: $-\nabla_{\perp}\rho=-1/(2\rho)\nabla_{\perp}\rho^{2}$ one can define the so called bulk hydrostatic pressure $P$ as: $\displaystyle P=\frac{c^{2}}{2}\frac{\rho^{2}g}{k}=\frac{1}{2}\rho c_{s}^{2},$ (4) where the last equality comes from $c_{s}^{2}=c^{2}\cdot g\rho/k$. Eq. (4) is the state equation linking the fluid hydrostatic pressure $P$ to its density if one neglects the quantum pressure term. It is the consequence of the mean- field formulation of the interaction. It also implies that the fluid of light is compressible with the compressibility equal to: $k/(c^{2}\rho^{2}g)$. One then gets the analogue Euler equation: $\displaystyle\frac{\partial\mathbf{v}}{\partial(z/c)}+\frac{1}{2}\nabla_{\perp}\mathbf{v}^{2}=-\frac{1}{\rho}\nabla_{\perp}P,$ (5) with a pressure $P$ of dimension [density]$\times$[speed]2. As already mentioned, the fluid dynamics can be studied by accessing its state at different $z$ positions, however this is not recommend practically since imaging inside a non-linear medium is highly challenging task. Alternatively, one can instead re-scale the effective time by incorporating fluid interaction [24, 20]. Fluid interaction can then be varied experimentally and the fluid dynamics can be studied while imaging only the state at the medium output plane. Re-scaling the time is based on defining following quantities: $\displaystyle z_{NL}=\frac{1}{g\rho(0,L)},\ \ \textrm{non-linear axial length}$ (6) $\displaystyle\xi=\sqrt{\frac{z_{NL}}{k},}\ \ \textrm{transverse healing length}$ (7) $\displaystyle c_{s}=\frac{c}{k\xi},\ \ \textrm{speed of sound}$ (8) $\displaystyle\psi=\frac{\mathcal{E}}{\sqrt{\rho(0,L)}},$ (9) and substituting the time and space variables as: $\tau=~{}z/z_{NL}$, $\tilde{\mathbf{r}}=\mathbf{r}/\xi$, $\tilde{\nabla}_{\perp}=\xi\nabla_{\perp}$. $L$ is the non-linear medium length. The propagation equation then reads: $i\frac{\partial\psi}{\partial\tau}=\left(-\frac{1}{2}\tilde{\nabla}^{2}_{\perp}+{\mid}\psi{\mid}^{2}\right)\psi.$ (10) One can note that dynamics of $\psi$ is not anymore dissipative, due to the normalization with respect to the exponentially decaying density: $\rho(0,L)=\rho(0,0)\textrm{exp}(-\alpha L)$, measured at at the medium exit plane. This formulation is necessary to describe accurately the experimental results of this work probing temporal dynamics of a fluid of light by varying the strength of the optical non-linearity and not the imaged $z$ plane. The effective time $\tau=|\Delta n(\mathbf{r}_{\perp}=0,L)|k_{0}L$ equals to the maximal accumulated non-linear phase. Rewriting the Madelung transformation with the new variables, we obtain: $\psi=\sqrt{\tilde{\rho}}\textrm{e}^{i\phi}=\sqrt{\frac{\rho}{\rho(0,L)}}\textrm{e}^{i\phi},\ \ \ \tilde{\mathbf{v}}=\frac{\mathbf{v}}{c_{s}}=\tilde{\nabla}_{\perp}\phi.$ (11) One gets dimensionless Euler and the continuity equations: $\displaystyle\frac{\partial\tilde{\rho}}{\partial\tau}+\tilde{\nabla_{\perp}}.\left(\tilde{\rho}\tilde{\mathbf{v}}\right)=0$ (12) $\displaystyle\frac{\partial\tilde{\mathbf{v}}}{\partial\tau}+\frac{1}{2}\tilde{\nabla_{\perp}}\tilde{\mathbf{v}}^{2}=-\tilde{\nabla_{\perp}}\left(\tilde{\rho}-\frac{1}{2\sqrt{\tilde{\rho}}}\tilde{\nabla^{2}_{\perp}}\sqrt{\tilde{\rho}}\right),$ (13) where the link between Eq. (13) and the Euler equation is made by neglecting the quantum pressure and defining the the dimensionless hydrostatic pressure as: $\tilde{P}=\frac{1}{2}\tilde{\rho}^{2}.$ (14) Finally, the dynamic pressure is defined as a vector quantity by: $\tilde{P_{d}}=\frac{1}{2}\tilde{\rho}\tilde{\mathbf{v}}|\tilde{\mathbf{v}}|,$ (15) The dynamic pressure is the fluid kinetic energy flux and accounts for the amount of pressure due to fluid motion. The impact force on an obstacle hit by a shockwave is proportional to its dynamic pressure. Expressed in dimensionless units, the dynamic pressure gives the strength of the convection term normalized by the pressure due to the interactions in the Eq. (13). It can be computed directly from the density and velocity measurements. ## Shock waves and blast wind In this work, we study the dynamics of a fluid of light disturbed by a localized Gaussian over-density $\delta\rho(\mathbf{r},0)=~{}\rho_{1}~{}\textrm{exp}\left(-2\mathbf{r}^{2}/\omega_{1}^{2}\right)$. $\rho_{1}$ is of the same magnitude as the background fluid density $\rho_{0}$ and $\omega_{1}$ quantifies the perturbation width. We can write $\rho(\mathbf{r},L)=\rho_{0}+\delta\rho(\mathbf{r},L).$ Normalizing the total density by its maximal undisturbed value one gets: $\tilde{\rho}(\mathbf{r},\tau)=\rho(\mathbf{r},L)/\rho_{0}(0,L)$. Extending this definition to $\rho_{0}$ and $\rho_{1}$, we obtain $\tilde{\rho}_{0}$ bound between 0 and 1 and having a Gaussian shape, and $\tilde{\rho}_{1}$ expressing the perturbation strength with respect to the fluid background density. To take into account the Gaussian profile of the density $\rho_{0}$, we define the over-pressure from the pressure difference between the case with and without perturbation: $\delta\tilde{P}(\mathbf{r},\tau)=\tilde{P}(\mathbf{r},\tau)-\tilde{P}_{0}(\mathbf{r},\tau).$ (16) To evaluate the differential pressure $\Delta\tilde{P}(\tau)$, showing the instantaneous difference in pressure between the perturbation center and the external undisturbed area, we define: $\Delta\tilde{P}(\tau)=\tilde{P}(0,\tau)-P_{0}(r_{ext},\tau).$ (17) The differential pressure $\Delta\tilde{P}(\tau)$ is the most important quantity we study in this work and we expect major differences in the non- linear perturbation dynamics between the 1D and the 2D geometries. Finally, the fluid velocity can be measured experimentally. It requires a measurement of the beam wavefront which is realized using off-axis interferometry. Calculating numerically the gradient of the phase, we obtain the background fluid velocity $\mathbf{v}_{0}$ and the perturbation velocity $\mathbf{v}_{1}$ by analyzing the images without and with the perturbation, respectively. Several studies have been performed in both $\rho_{1}\ll~{}\rho_{0}$ and $\rho_{1}\gg\rho_{0}$ regimes, observing the Bogoliubov dispersion of the linearized waves created by the perturbation [25, 26, 27], and the shock waves [14, 20], respectively. In this work we investigate the case $\rho_{1}\sim\rho_{0}$ by analyzing the fluid density, velocity and pressure both in the 1D and 2D geometries. The NLSE is known to give rise to sound-like dispersion to the low amplitude waves, governed by the Bogoliubov theory. Here, a perturbation of the same order (or larger) than the background results in the sound velocity variation following the local density inside the perturbation. This is the prerequisite for observing shock waves, a special type of waves changing their shape during propagation towards a steepening profile. In hydrodynamics, shock waves are usually reported as a time evolution measurement of a physical quantity (pressure, density…) at a fixed point in space. After the passage of a the shock wave front, a blast wind (a negative differential pressure) should be observed in 2 and 3 dimensional space. A direct physical consequence of this wind in classical hydrodynamics is observed for example after an explosion inside an edifice: the presence of glass pieces within the building is the signature of the blast wind . In the next section we report the time evolution as well as the time snapshots (spatial map of a physical quantity at fixed time) typically not accessible in classical hydrodynamics experiments. ## Experimental setup In our experiment, we investigate the propagation of a near-resonance laser beam through a warm rubidium vapor cell, which induces effective photon-photon interactions [28]. Two configurations are studied: the 2D geometry with a radially symmetric dynamics and the 1D geometry with a background much larger along $x$ than along $y$ which allows for a 1D description of non-linear wave dynamics [20]. A tapered amplified diode laser is split into a background, a reference and a perturbation beams (see supplementary materials for details). The background beam is enlarged with a telescope up to a waist of 2.5$\pm$0.5 mm along $x$ and 0.8$\pm$0.1 mm along $y$ in the 1D geometry, and 1.8$\pm$0.3 mm along the radial coordinate in the 2D configuration. The reference beam (for interferometric phase measurement) is matched to the same dimensions. The perturbation beam is focused to get the waist of 0.12$\pm$0.03 mm in the middle of the cell (the corresponding Rayleigh range is 55 mm). The background and perturbation are recombined with a 90R:10T beam splitter such that 90 % of the background beam power is reflected towards the cell. The second arm of the BS is sent through a 200 $\mu m$ diameter pinhole into a photodiode to stabilize the interferometer. The control is realized by locking on local minimum acting on a piezoelectric mirror mount with a RedPitaya hardware run by the PyRPL software [29]. Cell temperature is 149(2)° C leading to an atomic density of 8.3$\pm$0.8$\times 10^{13}$ cm-3. The cell output is imaged with a $\times$4.2 magnifying 4-f setup onto a camera. Sets of 4 images (background only, background with reference, background with perturbation and finally background with both perturbation and reference) at different input powers $\mathcal{P}$ ranging from 50 to 600 mW and different laser detunings $\Delta$ from the 85Rb D2 line $F=3\rightarrow F^{\prime}$ transition are taken (see supplementary materials for details). The reference beam is superimposed with other beams with an angle of 30 milli-radians, giving rise to interferogramms with vertical fringes of average periodicity of 25$\pm$1 $\mu$m. Figure 1: Density data: The left column corresponds to the 1D configuration and the right column to the 2D case. a) and b) are over-density maps at time $\tau=31$, obtained by subtracting the images with no perturbation from the ones with perturbation in the 1D and 2D geometry, respectively. c), d) are density profiles without (blue) and with (red) perturbation in the 1D and 2D geometry, respectively. The profiles are shifted vertically (spacing of 2) for better visibility. ## Density The density is an important physical parameter needed to compute the static and hydrodynamic pressure. It is directly given by the intensity measurement. In figure 1 a) and b), we present the experimental maps of the over-density $\delta\tilde{\rho}$ at time $\tau=31$, after subtracting the background fluid, in the 1D and 2D geometries respectively. By changing the laser intensity and detuning, we can modify the effective time $\tau$. The associated time $\tau$ is calculated from the nonlinear index $\Delta n$ via the off-axis interferometric measurement for each experimental configuration ($\mathcal{P},\Delta$) (see supplementary materials for details). Fig. 1 a) and b) show the spatio-temporal over-density diagrams. We present the corresponding density profiles at different times in figure 1 c) and d). The 1D density data are averaged over the $y$ direction for ${\mid}y{\mid}<0.1$ mm and the 2D images are radially averaged to get the background fluid density (blue curves) and the total fluid density including the perturbation (red curves). These results show two important effects. In the 1D geometry, a clear steepening of the perturbation front and the development of dispersive shock waves can be seen as an oscillating pattern developing in beyond the shock front with effective time $\tau$. In the 2D geometry, interestingly, the steepening of the shock front is less pronounced. Moreover, a density much lower than the background density is observed in the center of the 2D profiles for long time $\tau>20$, which is not the case in 1D. This negative differential density has a direct consequence on the differential pressure calculated using Eq. (17). ## Static pressure Figure 2: Pressure analysis: a),c): 1D & 2D over-pressure profiles evaluated at different effective times $\tau$. Each following profile shifted vertically by 2 for better visibility. b),d) show the 1D and 2D spatio-temporal diagrams of the over-pressure evolution, respectively. The dotted black lines show the trajectory of expansion at the speed of sound according to the parabolic equation with the prefactor given by $k/L=107$ mm-1 in both geometries. The blue dotted lines show the same trajectories shifted horizontally by 250 $\mu m$ and 200 $\mu m$ in 1D and 2D cases, respectively. It corresponds to external undisturbed area used for the measurement of the differential pressure. Dashed green rectangles around $\tau=40$ show the presence of a second shock due to an increasing differential pressure. Figure 3: Differential pressure calculated from Eq. (17) for the 1D (circular dots) and the 2D cases (square dots). The uncertainty bars correspond to the statistical analysis of multiple images. The pressure is normalized as described in the main text. Blue line is the ambient pressure outside of the shock. Black dashed line is the Friedlander model for a blast wave described in Eq. (18) with $P_{s}=1$ and $t^{*}=20$. To isolate the effect of the perturbation on the static pressure, we compute the over-pressure from images of the background with and without the bump taken at same effective times $\tau(\mathcal{P},\Delta)$, using Eq. (16) and (14). The over-pressure as a function of time $\tau$ is shown in Fig. (2) b) and d) and profiles averaged along $y$ in the 1D case and radially in the 2D case are presented in Fig. (2) a) and c) for various times. The trajectory of a density pulse spreading with no dispersion at the speed of sound can be expressed as follows: $r=c_{s}(\tau)\times(L/c)$. The coefficient can be calculated using the time dependence of the sound velocity: $c_{s}=c\sqrt{\tau/(kL)}$ obtained from Eqs. (6) and (8). It directly leads to $\tau=kr^{2}/L$ and knowing that: $L=75$ mm and $k=8\times 10^{3}$ mm-1, one gets: $\tau=107\times r^{2}$. The coefficient does not depend on the dimensionality of the system. In the pressure maps (Fig. (2) b) and d)), we have added a black dashed line following this trend: $\tau=107\times x^{2}$ (1D) and $\tau=107\times r^{2}$ (2D). As expected, this trajectory follows closely the shock front in the 1D geometry. The differential pressure is defined as the pressure difference between inside and outside of the shock as expressed in Eq. (17). The undisturbed pressure as function of time is evaluated along the same trend line $\tau=107\times(r_{ext}-r_{0})^{2}$, translated $r_{0}=250~{}\mu$m in 1D and $r_{0}=200~{}\mu$m in 2D, which corresponds to $\sim 1.5$ times the perturbation beam waist (blue dashed line). In 2D, the shock front expansion is slower than the calculated trajectory, as described in [20], and the blue dashed line can therefore still used to define the undisturbed pressure. The temporal evolution of the differential static pressure (at $x=0$) is presented in Fig. 3. 1D (red circles) and 2D (gray triangles) geometries are compared from $\tau=0$ to $\tau=45$. An important difference can be seen between the two geometries: in the 2D situation the differential pressure becomes negative at $\tau=20$ as it goes to zero in the 1D case. The observation of the negative pressure is the typical signature of a blast wind. This measurement reveals the dramatic impact of the geometry on blast wind in a fluid of light and exemplifies the analogy with classical hydrodynamics. To quantify this analogy, we use the Friedlander waveform model which is known to describe the dynamics of physical quantities in a free-field (i.e. in a open 3-dimensional space) blast wave [2]. In this model the differential pressure follows an exponential decay of the form: $\Delta\tilde{P}=P_{s}e^{-\tau/t^{*}}(1-\tau/t^{*}),$ (18) where $P_{s}$ and $t^{*}$ are two parameters which corresponds respectively to the peak differential pressure immediately behind the shock and to the time when the differential pressure becomes negative. The period when the hydrostatic pressure is above the ambient value is known as the positive phase, and the period when the properties are below the ambient value is the negative phase. We use $P_{s}=1$ (since the differential pressure is normalized) and $t^{*}=20$ and plot the corresponding model with a black dashed line in Fig. 3. An intriguing feature can also be seen in the 2D time evolution at $\tau=40$. Close to the minimum of the negative phase, a second peak of differential pressure is observed (the single point at $\tau=40$ Fig. 3 is the average of several realizations with errors bars indicating the standard deviation of the measurement) in our optical analogue which is reminiscent of the second shock observed in classical explosion. In classical blast wave dynamics, this second shock is believed to be a consequence of the expansion and subsequent implosion of the detonation products and source materials. Our results suggest that this second shock might be of more general nature than currently thought. ## Velocity Figure 4: Fluid velocities from the off-axis interferometry. a),b) Space-time evolution of the Mach number with respect to the background’s local speed of sound, in the 1D and 2D geometry, respectively. The dotted black line in a) shows the calculated trajectory of expansion at the speed of sound (see main text). c),d) show the background’s $\tilde{v}_{0}$ (blue) and total $\tilde{v}$ (red) Mach number profiles, at different times, for the 1D (x coordinate) and 2D geometry (radial coordinate), respectively. Each following profile shifted vertically (spacing of 1) for visibility. For blast waves, there are no simple thermodynamic relationships between the physical properties of the fluid at a fixed point [30]. This means that the temporal evolution of the static pressure measured at a fixed point is not sufficient to calculate the temporal evolution of the velocity or the dynamic pressure from that single measurement. To fully describe the physical properties of a fluid in a blast wave it is necessary to independently measure at least three of the physical properties, such as, the static pressure, the density and the fluid velocity or the dynamic pressure. In the last section of this work, we report the measurement of last two physical properties, which are vector quantities. The fluid velocity is calculated from its phase (see Eq. (11)) which is measured using off-axis interferometric imaging. The off-axis configuration consists in the tilted recombination of the signal beam with the reference beam on the camera plane. This results in the set of linear fringes evolving along the relative tilt direction and locally deformed (stretched or compressed) according to the beams relative curvature. Using a collimated Gaussian beam as the reference, the measured curvature is the one of the signal beam. The acquired interferogramm carries the information on the beam phase via its amplitude modulated term. This term shows spatial periodicity and in the Fourier space it translates to two peaks shifted by a distance proportional to the off-axis tilt angle, symmetric with respect to the origin. By numerically calculating the spatial spectrum and filtering one of these peaks, the inverse Fourier transform gives the beam complex envelope with a spatial resolution bound by the fringe wavelength. The measured phase is unwrapped and the contribution due to the relative tilt is removed by subtracting the phase ramp. The resulting phase is averaged and numerically differentiated to get the velocity map. Using this procedure, the off-axis interferograms of the background fluid and of the background fluid with the perturbation are analyzed to give access to $v_{0}(r,\tau)$ and $v(r,\tau)$, respectively. The difference of these quantities gives the perturbation velocity $v_{1}(r,\tau)$. The non-zero velocity $v_{0}$ of the background fluid arises from its finite size causing its expansion due to a non-zero pressure gradient. The knowledge of $v_{0}$ is essential to calculate the effective interaction $g$ and therefore the time $\tau$ and the sound velocity. Indeed, $\phi_{0}=\tau\tilde{\rho}_{0}$ can be accessed by integrating $v_{0}$ over the transverse coordinate and using the fact that $\phi(\tilde{r}\rightarrow\infty,\tau)\rightarrow 0$. Knowing $\tau$, the sound velocity is $c_{s}(\mathbf{r}_{\perp},\tau)=c\sqrt{\tau\tilde{\rho}_{0}(\mathbf{r}_{\perp},\tau)/(k_{0}L)}$. The velocity maps normalized by the local sound velocity (in Mach units) are presented in figure 4 a) and b) for the 1D and 2D configurations, respectively. Since velocity is a vector quantity, negative values correspond to a propagation along $-x$ direction. Figure 4 c) and d) show the corresponding profiles obtained for three specific times $\tau=2;\ 23$ and $45$. The maximal speed of sound at these times is 0.18, 0.62 and 0.86 percent of the speed of light in vacuum. Positive outward velocity, as well as zero velocity at the center is observed at all times both in the 1D and 2D cases. Whereas it is intuitively expected in the 1D geometry with the differential pressure never dropping to negative values, it also holds in the 2D case in which a negative phase for the differential pressure exists. A possible explanation lies in the fact that when the negative phase is reached for the differential pressure, the perturbation has already expanded enough such that the net resulting force is smaller due to a larger distance. It is also worth noting that the velocity is at least 2 times larger in the 1D geometry than in 2D, as seen by comparison of the y-axis scales in Figure 4 c) and d). Additionally, clear steepening of the velocity profiles is observed in the 1D case reaching a Mach number of 1 at the steepest position. Figure 5: Dynamic pressure analysis. a) and b) show the spatio-temporal evolution maps of the dynamic pressure profiles, for the 1D (the x component) and 2D geometry (the radial component), respectively. Below, the c) and d) panels show various superimposed dynamic pressure profiles at different times, in 1D and 2D geometry, respectively. ## Dynamic pressure Alternatively, we can measure the dynamic pressure to compute a third thermodynamic quantity: the total pressure. The dynamic pressure is also a vector quantity and can be obtained from a phase measurement similar to fluid velocity using Eq. (15). The dynamic pressure maps are presented in Figs. 5 a) and b). Once again Figs. 5 c) and d) show dynamic pressure profiles for three selected times. In 1D, the dynamic pressure forms a steep overpressure characteristic of the shock front which increases as function of time. In the 2D geometry, on the contrary the dynamic pressure reaches a plateau at the shock front without forming a steep overpressure peak. This behavior is in agreement with the velocity distributions presented previously. ## Conclusion Relying on detailed measurements of all thermodynamic quantities in a fluid of light blast wave, we have demonstrated for the first time the occurence of a blast wave in a fluid of light. We compare 1D and 2D geometry and report the observation of a negative phase during the blast only for the 2-dimensional case. The differential pressure in the 2D geometry is compared to the classical hydrodynamics of Friedlander blast-wave and we see a very good agreement with this model. Velocity maps and dynamic pressure are finally presented to complete the study. Our work opens the way to precise engineering of a fluid of light density and velocity distribution which will prove to be a valuable tool to design new experiments studying superfluid turbulence [31] or analogue gravity where an excitation of a fluid of light changes from a subsonic to a supersonic region. ###### Acknowledgements. The authors thank Ferdinand Claude and Samuel Deléglise for useful discussions for setting up PyRPL. This work is supported by the PhoQus project. ## References * Friedlander [1946] F. G. Friedlander, The diffraction of sound pulses i. diffraction by a semi-infinite plane, Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences 186, 322 (1946). * Dewey [2016] J. M. Dewey, Measurement of the physical properties of blast waves, in _Experimental Methods of Shock Wave Research_ (Springer, 2016) pp. 53–86. * Dewey [1964] J. M. Dewey, The air velocity in blast waves from tnt explosions, Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences 279, 366 (1964). * Reed [1977] J. W. Reed, Atmospheric attenuation of explosion waves, The Journal of the Acoustical Society of America 61, 39 (1977). * Taylor [1950] G. I. Taylor, The formation of a blast wave by a very intense explosion.-ii. the atomic explosion of 1945, Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences 201, 175 (1950). * Xu _et al._ [2017] G. Xu, M. Conforti, A. Kudlinski, A. Mussot, and S. Trillo, Dispersive dam-break flow of a photon fluid, Physical review letters 118, 254101 (2017). * Li _et al._ [2004] Z. Li, D. Zhang, B. Yu, and L. Guan, Characteristics of plasma shock waves in pulsed laser deposition process, The European Physical Journal Applied Physics 28, 205 (2004). * Hoefer _et al._ [2008] M. A. Hoefer, M. J. Ablowitz, and P. Engels, Piston dispersive shock wave problem, Phys. Rev. Lett. 100, 084504 (2008). * Dewey [2018] J. M. Dewey, The friedlander equations, in _Blast Effects_ (Springer, 2018) pp. 37–55. * Dewey [2010] J. M. Dewey, The shape of the blast wave: studies of the friedlander equation, in _Proceedings of the 21st International Symposium on Military Aspects of Blast and Shock_ (2010) pp. 1–9. * Hoefer _et al._ [2006] M. A. Hoefer, M. J. Ablowitz, I. Coddington, E. A. Cornell, P. Engels, and V. Schweikhard, Dispersive and classical shock waves in bose-einstein condensates and gas dynamics, Phys. Rev. A 74, 023623 (2006). * Meppelink _et al._ [2009] R. Meppelink, S. B. Koller, J. M. Vogels, P. van der Straten, E. D. van Ooijen, N. R. Heckenberg, H. Rubinsztein-Dunlop, S. A. Haine, and M. J. Davis, Observation of shock waves in a large bose-einstein condensate, Phys. Rev. A 80, 043606 (2009). * Chang _et al._ [2008] J. J. Chang, P. Engels, and M. A. Hoefer, Formation of dispersive shock waves by merging and splitting bose-einstein condensates, Phys. Rev. Lett. 101, 170404 (2008). * Wan _et al._ [2007] W. Wan, S. Jia, and J. W. Fleischer, Dispersive superfluid-like shock waves in nonlinear optics, Nature Physics 3, 46 (2007). * Vocke _et al._ [2015] D. Vocke, T. Roger, F. Marino, E. M. Wright, I. Carusotto, M. Clerici, and D. Faccio, Experimental characterization of nonlocal photon fluids, Optica 2, 484 (2015). * Conti _et al._ [2009] C. Conti, A. Fratalocchi, M. Peccianti, G. Ruocco, and S. Trillo, Observation of a gradient catastrophe generating solitons, Physical review letters 102, 083902 (2009). * Ghofraniha _et al._ [2012] N. Ghofraniha, S. Gentilini, V. Folli, E. DelRe, and C. Conti, Shock waves in disordered media, Physical review letters 109, 243902 (2012). * Carusotto [2014] I. Carusotto, Superfluid light in bulk nonlinear media, Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences 470, 20140320 (2014). * Isoard _et al._ [2019] M. Isoard, A. M. Kamchatnov, and N. Pavloff, Wave breaking and formation of dispersive shock waves in a defocusing nonlinear optical material, Phys. Rev. A 99, 053819 (2019). * Bienaimé _et al._ [2021] T. Bienaimé, M. Isoard, Q. Fontaine, A. Bramati, A. Kamchatnov, Q. Glorieux, and N. Pavloff, Controlled shock wave dynamics in a fluid of light, arXiv preprint arXiv:2101.00720 (2021). * Ghofraniha _et al._ [2007] N. Ghofraniha, C. Conti, G. Ruocco, and S. Trillo, Shocks in nonlocal media, Physical review letters 99, 043903 (2007). * Wetzel _et al._ [2016] B. Wetzel, D. Bongiovanni, M. Kues, Y. Hu, Z. Chen, S. Trillo, J. M. Dudley, S. Wabnitz, and R. Morandotti, Experimental generation of riemann waves in optics: a route to shock wave control, Physical review letters 117, 073902 (2016). * Sadot _et al._ [2018] O. Sadot, O. Ram, E. Nof, E. Kochavi, and G. Ben-Dor, Small-scale blast wave experiments by means of an exploding wire, in _Blast Effects_ (Springer, 2018) pp. 141–170. * Pavloff [2019] N. Pavloff, Optical hydrodynamics and nonlinear diffraction, in _Waves Côte d’Azur_ (2019). * Fontaine _et al._ [2018] Q. Fontaine, T. Bienaimé, S. Pigeon, E. Giacobino, A. Bramati, and Q. Glorieux, Observation of the bogoliubov dispersion in a fluid of light, Phys. Rev. Lett. 121, 183604 (2018). * Fontaine _et al._ [2020] Q. Fontaine, P.-É. Larré, G. Lerario, T. Bienaimé, S. Pigeon, D. Faccio, I. Carusotto, É. Giacobino, A. Bramati, and Q. Glorieux, Interferences between bogoliubov excitations in superfluids of light, Physical Review Research 2, 043297 (2020). * Piekarski _et al._ [2020] C. Piekarski, W. Liu, J. Steinhauer, E. Giacobino, A. Bramati, and Q. Glorieux, Short bragg pulse spectroscopy for a paraxial fluids of light, arXiv preprint arXiv:2011.12935 (2020). * Agha _et al._ [2011] I. H. Agha, C. Giarmatzi, Q. Glorieux, T. Coudreau, P. Grangier, and G. Messin, Time-resolved detection of relative-intensity squeezed nanosecond pulses in an 87rb vapor, New Journal of Physics 13, 043030 (2011). * Neuhaus _et al._ [2017] L. Neuhaus, R. Metzdorff, S. Chua, T. Jacqmin, T. Briant, A. Heidmann, P.Cohadon, and S. Deléglise, Pyrpl (python red pitaya lockbox) — an open-source software package for fpga-controlled quantum optics experiments, in _2017 CLEO/Europe-EQEC_ (2017). * Voronov _et al._ [1992] B. Voronov, A. Korobov, and O. V. Rudenko, Nonlinear acoustic waves in media with absorption and dispersion, Soviet Physics Uspekhi 35, 796 (1992). * Rodrigues _et al._ [2020] J. D. Rodrigues, J. T. Mendonça, and H. Terças, Turbulence excitation in counterstreaming paraxial superfluids of light, Physical Review A 101, 043810 (2020). * Siddons _et al._ [2008] P. Siddons, C. S. Adams, C. Ge, and I. G. Hughes, Absolute absorption on rubidium d lines: comparison between theory and experiment, Journal of Physics B: Atomic, Molecular and Optical Physics 41, 155004 (2008). * Weller _et al._ [2011] L. Weller, R. J. Bettles, P. Siddons, C. S. Adams, and I. G. Hughes, Absolute absorption on the rubidium d1line including resonant dipole–dipole interactions, Journal of Physics B: Atomic, Molecular and Optical Physics 44, 195006 (2011). Supplemental Materials: Blast waves in a paraxial fluid of light Experimental details The scheme of the experimental setup is shown on Figure S1. Toptica DLCpro 780 with TA was used for all measurements. The laser frequency was tuned around 780 nm and measured with a MogWave Multimeter LambdaMeter and calibrated with Saturable absorption spectroscopy (SAS). The laser beam was mode cleaned with a single mode fiber and then split into the Background, Bump and the Reference arms. The respective intensity ratio was fixed by the angles of the Half-Wave-Plates (HWP), placed before the Polarizing Beam Splitters (PBS), in agreement with experimental requirements: $\tilde{\rho}_{1}(\textbf{r}=0,\tau=0)\approx 3$ and minimal sufficient power into the reference beam to have noticeable fringe contrast. Since the Background and the Reference have the same polarization during recombination, their interference needs to be constructive at the cell input in order to create the desired input state for the fluid’s density. The beamsplitter’s unused arm’s power at the Background-Bump overlap area should then be minimal. This signal was measured with a 200 $\mu m$ diameter pinhole centered at the overlap area and a photodiode. The relative phase needs to be locked in order to minimize permanently this signal and make it insensitive to perturbations such as air currents. Therefore the photodiode signal was transformed into an error signal of a piezoelectric mirror mount controlling the relative phase. The error signal generation from the photodiode signal was realized with the PyRPL software running on a Red Pitaya FPGA [29]. The modulation frequency was around 2-3 kHz. Figure S1: Schematic visualization of the experimental setup. Diode laser frequency calibration was performed with Saturable Absorption Spectroscopy (SAS), and during the experiment the frequency measurement was performed with a MogWave Lambdameter. The laser was mode cleaned with a polarization maintaining single mode siber (PMSMF), before being split into the Background, Bump and the Reference. The Background-Bump interference arm complementary to the Rb vapor cell was cropped with a 200 $\mu m$ diameter pinhole (Ph) to measure the the power of the overlap area on a photodiode. This signal was minimized by controlling the relative phase via piezoelectric motion of a mirror mount to have permanently constructive interference on the vapor cell arm. The error signal was generated from the photodiode signal with the PyRPL lockbox software. Vapor Temperature Figure S2: Vapor’s transmission and its intensity dependent refractive index measurement. a) and b) show the maximal refractive index variation calculated from the off-axis interferograms of the backgroung beam with a reference, in 1D and 2D geometry, respectively. c) Background beam’s transmission spectrum with respect to 85Rb cooling transition measured at different input powers. Dashed line is the theory of a linear multilevel vapor at temperature 150 °C and 0.5 % the isotopic fraction of 87Rb inside the cell. Checking the ”Kerr” approximation: d) and e) show the variation of the refractive index with laser power at fixed laser detuning in both geometries. One of the useful knobs to control the light-matter interaction in hot vapor cells is the atomic density. The latter is directly linked to the vapor pressure via the ideal gas law (neglecting the atom-atom interactions). It equals the Rb vapor’s saturation pressure at thermal liquid-gas equilibrium and can be increased by several orders of magnitude when heating the cell from 50° C to 150° C. Keeping the vapor temperature constant during the experiment is therefore necessary to control the atomic susceptibility. In our experiment, several electric resistors were wound around the cell and connected in parallel to a DC power supply to heat up the cell. The vapor temperature was accessed by measuring the transmission spectrum around the Rb D2 line in the weak beam limit. The frequency calibration was performed via Saturable absorption spectroscopy, as shown on Figure S1. The experimental spectrum was fitted with the linear susceptibility model developped in [32] taking into account all hyperfine transitions of both isotopes and the collisional self-broadening due to resonant dipole-dipole interactions [33], with the atomic density and the number fraction of 87Rb isotope as free parameters. The temperature was measured before and after each experiment to prevent any temperature drift. Non-linear refractive index variation measurement The intensity dependent refractive index of our hot atomic vapor is the key parameter governing the fluid’s dynamics as it is linked to the effective evolution time: $\tau=\Delta nk_{0}L$ and its speed of sound: $c_{s}=c\sqrt{\Delta n}$. In this work it was measured using the off-axis interferometry which gives access to the transverse phase variations at the cell exit plane. The transverse phase profile of the Background beam is assumed to depend as follows on the beam’s intensity I(r): $\phi_{th}(\textbf{r},L)=k_{0}L\frac{n_{2}I(\textbf{r})}{1+I(\textbf{r})/I_{s}}+\phi_{0}$ (S1) Where $n_{2}$ is the Kerr index, $I_{s}$ the saturation intensity of the Kerr effect and $\phi_{0}$ a constant phase. The gradient of the phase, giving access to the fluid velocity, is numerically calculated and fitted with $\nabla\phi_{th}$ with $n_{2}$ and $I_{s}$ as free parameters. Figure S2 a) and b) show measured maximal variation of refractive index for different experimental configurations of the laser detuning $\Delta$ and power $P$. Each point corresponds to a processed image. c) Shows the transmission spectra through the cell for different input powers. No saturation of the absortpion can be evidenced. The black dashed line is the theoretical calculation of the linear susceptibility used for the measurement of the vapor’s temperature. Finally, d) and e) show the variation of the refractive index with intensity. The graphs show that the results of this work are obtained below the regime of the saturation of the Kerr effect. Background beam’s expansion Figure S3: Background fluid’s expansion. a) shows the Background’s expansion in the transverse y direction in the 1D geometry and b) shows the Background’s radial expansion in the 2D geometry. In the theoretical discussion developed in the main text and for the $\Delta n$ measurement it is assumed that the background fluid beam’s density is invariant with time. The experimental data to verify this hypothesis are shown in Figure S3. No expansion in the x direction of the 1D case was observed. The expansion is most pronounced in the transverse y direction of the 1D case. In the 2D case the background’s insignificant expansion is observed. Relevance of the Quantum Pressure Figure S4: Quantum pressure calculated from experimental density profiles. a) in the 1D geometry and b) in the 2D geometry. Same colormap is used for both graphs. As mentioned in the main text, the Quantum pressure was neglected in the theoretical description of the experimental data as we are interested in the fluid’s behavior in the long wavelength limit. This term is known to have a dispersive contribution to the shockwave profile which, upon steepening, becomes composed of an increased amount of various momentum components moving at different velocities. To evaluate the relevance of the Quantum Pressure in this work we calculated it from the experimental density profiles at different evolution times for both 1D and 2D geometry as: $\tilde{P}_{q}=\frac{1}{2\sqrt{\tilde{\rho}}}\tilde{\nabla}^{2}_{\perp}\sqrt{\tilde{\rho}}$ (S2) Depending on the dimensionality the Laplacian was calculated as: $\tilde{\nabla}^{2}_{\perp}=\xi^{2}\partial^{2}/\partial x^{2}$ in 1D or as: $\tilde{\nabla}^{2}_{\perp}=\xi^{2}[\partial^{2}/\partial r^{2}+(1/r)\times\partial/\partial r]$ in 2D using the radial symmetry. With this formulation the value of the dimensionless Quantum Pressure directly compares with the dimensionless density (stemming from interactions) in the right hand side of the Euler-like Madelung equation. The result is shown on Figure S4 a) for the 1D and b) for the 2D case. The Quantum Pressure seems to be most pronounced at the vicinity of the Shock front in 1D. In 2D it seems to decay with time. In both cases it does not exceed 0.1 for times $\tau>5$. This validates the theoretical approach chosen in work and consisting in neglecting the Quantum Pressure. For lower times the calculation seems inaccurate. This may be due to a large uncertainty on the healing length $\xi$ in this regime.
# Homotopy Methods for Eigenvector-Dependent Nonlinear Eigenvalue Problems ††thanks: The research was supported in part by the National Natural Science Foundation of China (11971092) Xuping Zhang School of Mathematical Sciences, Dalian University of Technology, Dalian, Liaoning 116025, P. R. China (zhangxp@dlut.edu.cn). Haimei Huo School of Mathematical Sciences, Dalian University of Technology, Dalian, Liaoning 116025, P. R. China (ab1234@mail.dlut.edu.cn). ###### Abstract Eigenvector-dependent nonlinear eigenvalue problems are considered which arise from the finite difference discretizations of the Gross-Pitaevskii equation. Existence and uniqueness of positive eigenvector for both one and two dimensional cases and existence of antisymmetric eigenvector for one dimensional case are proved. In order to compute eigenpairs corresponding to excited states as well as ground state, homotopies for both one and two dimensional problems are constructed respectively and the homotopy paths are proved to be regular and bounded. Numerical results are presented to verify the theories derived for both one and two dimensional problems. Key Words eigenvector-dependent nonlinear eigenvalue problem, Gross-Pitaevskii equation, homotopy continuation method Subject Classification(AMS):65H17, 65H20, 65N06, 65N25 ## 1 Introduction In this paper, we are concerned with the eigenvector-dependent nonlinear eigenvalue problems resulting from the finite difference discretizations of the Gross-Pitaevskii equation (GPE) describing Bose-Einstein condensates (BEC). BEC are clouds of ultracold alkali-metal atoms or molecules that occupy a single quantum state [1, 2]. The properties of a BEC at temperature $T$ much smaller than the critical condensation temperature $T_{c}$ are usually described by the nonlinear Schrödinger equation (NLS) for the macroscopic wave function known as the Gross-Pitaevskii equation $\begin{array}[]{lcl}{i\psi_{t}=-\frac{1}{2}\Delta\psi+V(x)\psi+\beta\ |\psi|^{2}\psi},&&{t>0,~{}x\in\Omega},\\\ {\psi(x,t)=0},&&{t\geq 0,~{}x\in\partial\Omega},\end{array}\\\ $ (1) where $\psi=\psi(x,t)$ is the macroscopic wave function of the BEC, $V(x)=\frac{1}{2}(x_{1}^{2}+x_{2}^{2}+\cdots+x_{N}^{2})$ is a typical trapping potential, $\Omega$ is a bounded domain in $\mathbb{R}^{N}$, $N\leq 3$, and $\beta$ positive or negative corresponds to the defocusing or focusing NLS. Two important invariants of GPE are the normalization of the wave function $\begin{array}[]{lcl}N(\psi)=\int_{\Omega}|\psi(x,t)|^{2}dx=1,&&t\geq 0,\end{array}$ (2) and the energy $\displaystyle E(\psi(x,t))=\int_{\Omega}\left[\frac{1}{2}|\nabla\psi|^{2}+V(x)|\psi|^{2}+\frac{\beta}{2}|\psi|^{4}\right]dx=E(\psi(x,0)).$ (3) To find stationary solution of (1), we substitute the formula $\psi(x,t)=e^{-i\lambda t}\phi(x)$ into (1) and (2) and obtain the time- independent Schrödinger equation with Dirichlet boundary condition and the normalized condition $\displaystyle\lambda\phi(x)$ $\displaystyle=-\frac{1}{2}\Delta\phi(x)+V(x)\phi(x)+\beta\phi^{3}(x),\quad{x\in\Omega},$ (4) $\displaystyle\phi(x)$ $\displaystyle=0,\quad x\in\partial\Omega,$ (5) $\displaystyle\int_{\Omega}|\phi(x)|^{2}dx$ $\displaystyle=1,$ (6) where $\lambda$ is the chemical potential of the condensate and $\phi(x)$ is a real function independent of $t$ [3]. (4)-(6) is a nonlinear eigenvalue problem. The eigenfunction corresponding to the minimum energy is called ground state and other eigenfunctions corresponding to larger energy are called excited states in the literature. There have been many theoretical studies as well as numerical studies for the time-independent Schrödinger equation. Bao and Cai [4] pointed out when $\beta>0$, the positive ground state is unique, and if $V(x)$ is radially symmetric in 2D, the positive ground state must be radially symmetric. Bao and Tang [1] proposed methods by directly minimizing the energy functional via finite element approximation to obtain the ground state and by continuation method to obtain excited states. Edwards and Burnett [5] presented a Runge- Kutta type method and employed it to solve the spherically symmetric time- independent GPE. Adhikari [6] used this approach to get the ground state solution of GPE in 2D with radial symmetry. Chang and Chien [7] and Chang, Chien and Jeng [8] investigated stationary state solutions of $(\ref{sec 1: transformed nonlinear problem})$ using numerical continuation method, where $\lambda$ was treated as a continuation parameter. The solution curves branching from the first few bifurcation points of $(\ref{sec 1: transformed nonlinear problem})$ were numerically traced using continuation method under the normalization condition $(\ref{sec 1:transformed constraint})$. Since nonlinearity rather than discretization method is our main concern and finite difference discretization will lead to a simpler nonlinear structure, finite difference discretization is adopted in this paper. The finite difference discretization of (4)-(6) is the following eigenvector-dependent nonlinear eigenvalue problem, $\begin{array}[]{c}D\varphi+\beta\varphi^{3}=\lambda\varphi,\\\ h\varphi^{\mathrm{T}}\varphi-1=0,\end{array}$ (7) where $D=\frac{1}{2}D_{1}+V$, $D_{1}$ is the coefficient matrix corresponding to $-\Delta$, $V$ is the diagonal matrix corresponding to the potential $V(x)$, $\lambda$ and $\varphi$ are the unknowns, and $h$ is a constant related to mesh size. $\varphi^{3}$ represents the vector with elements being the corresponding elements of $\varphi$ to the power 3. This convention will be used throughout this paper. With respect to the theoretical aspects of eigenvector-dependent nonlinear eigenvalue problem, [9] and [10] studied the following general nonlinear eigen-value problem $Ax+F(x)=\lambda x,$ (8) where $A$ is an $n\times n$ irreducible Stieltjes matrix, i.e, an irreducible symmetric positive definite matrix with off-diagonal entries nonpositive, $F(x)=(f_{1}(x_{1}),\ldots,f_{n}(x_{n}))^{\mathrm{T}}$ and $x=(x_{1},\cdots,x_{n})^{\mathrm{T}}$. The functions $f_{i}(x_{i})$ are assumed to have the property that $f_{i}(x_{i})>0$, when $x_{i}>0$, $i=1,\ldots,n$. It is shown that under certain conditions on $F(x)$, there exists a positive eigenvector $x(\lambda)$ if and only if $\lambda>\mu$, where $\mu$ is the smallest eigenvalue of $A$, and for every $\lambda>\mu$, the positive eigenvector is unique. Moreover, such a solution is a monotone increasing function of $\lambda$. The most popular numerical method to the eigenvector-dependent nonlinear eigenvalue problems is the self-consistent field (SCF) iteration, which is suitable for computing the ground state; for instance, see [11, 12] and the references therein. In [13], inverse iteration method was applied to solve eigenvector-dependent nonlinear eigenvalue problems. Most of the above papers concentrate on the ground state and the first excited state. As far as we know, there are only a few numerical works on other excited states, such as [1, 14, 15, 16]. The main purpose of this paper is to design algorithms for computing excited states of high energy. Homotopy method is one of the effective methods for solving eigenvalue problems. A great advantage of the homotopy method is that it is to a large degree parallel, in the sense that each eigenpath is traced independently of the others. There are several works on homotopy methods for linear eigenvalue problems. Remarkable numerical results have been obtained by using homotopy algorithm on eigenvalue problems of tridiagonal symmetric matrices [17, 18]. Solving eigenvalue problems of real nonsymmetric matrices with real homotopy was developed in [19, 20]. The homotopy method is also used to solve the generalized eigenvalue problem [21]. For eigenvalue-dependent nonlinear eigen- problems such as $\lambda$-matrix problems, a homotopy was given by Chu, Li and Sauer [22]. The major part of this paper is the construction of homotopy for computing many eigenpairs of the eigenvector-dependent nonlinear eigen-problem. Key issues encountered in constructing the homotopy are the selection of the homotopy parameter and that of an appropriate initial eigenvalue problem so that the homotopy paths determined by the homotopy equation are regular and the numerical work in following these paths is at reasonable cost. The parameter $\beta$ in the original problem seems to be a natural choice for the homotopy parameter. However, it seems difficult to prove that 0 is a regular value for such homotopy. In fact, 0 is probably not a regular value of the natural homotopy with parameter $\beta$. Instead, an artificial parameter $t$ is chosen as the homotopy parameter to connect a constructed initial eigenvalue problem and the target one. As for the selection of an initial eigenvalue problem, random matrix with certain sparse structure is designed, which guarantees that 0 is a regular value of the homopoty with probability one and which renders the initial problem and the target problem possess similar structures. The rest of this paper is organized as follows. In Section 2, the time- independent GPE Dirichlet problem (4)-(6) is discretized by finite difference method and existence of certain types of solution of the discretized problems is derived. In Section 3, homotopies for (7) are constructed with $\Omega\subset\mathbb{R}$ and $\Omega\subset\mathbb{R}^{2}$ respectively, and regularity and boundedness of the homotopy paths are proved. In Section 4, numerical results are presented to verify the theoretical results derived for $\Omega\subset\mathbb{R}$ and $\Omega\subset\mathbb{R}^{2}$ respectively. Conclusions are drawn in the last section. ## 2 Discretizations of the nonlinear eigenvalue problem ### 2.1 Finite difference discretizations For one dimensional problem (4)-(6) with $\Omega=[a,b]\subset\mathbb{R}$, the grid points are $x_{j}=a+jh$, $j=0,\ldots,n+1$, where $n\in\mathbb{N}^{+}$ and $h=\frac{b-a}{n+1}$ is the mesh size. The finite difference discretization of the differential equation and a simple quadrature of the normalization condition lead to the following system of algebraic equations, $\begin{array}[]{c}D\varphi+\beta\varphi^{3}-\lambda\varphi=0,\\\ \frac{1}{2}\left(\frac{1}{h}-\varphi^{\mathrm{T}}\varphi\right)=0,\end{array}$ (9) where $\varphi=(\varphi_{1},\cdots,\varphi_{n})^{\mathrm{T}}$, $\varphi_{j}$ are the approximations of $\phi(x_{j})$, $v_{j}=V(x_{j})$, $j=1,\ldots,n$, and $D=\frac{1}{2}D_{1}+V$ with $\displaystyle D_{1}=\frac{1}{h^{2}}\left(\begin{array}[]{cccc}2&-1&\\\ -1&2&\ddots&\\\ &\ddots&\ddots&-1\\\ &&-1&2\end{array}\right),\quad V=\left(\begin{array}[]{ccc}v_{1}&&\\\ &\ddots&\\\ &&v_{n}\end{array}\right).$ (17) The discretization of the normalized condition is rewritten so that the Jacobian matrix of the nonlinear mapping with respect to $(\varphi,\lambda)$ is symmetric, as will be seen below. For two dimensional problem (4)-(6) with $\Omega=[a,b]\times[c,d]$, the domain is divided into a $(m+1)\times(n+1)$ mesh with step size $h_{1}=\frac{b-a}{m+1}$ in $x$-direction, $h_{2}=\frac{d-c}{n+1}$ in $y$-direction. The grid points $(x_{i},y_{j})$ are $x_{i}=a+ih_{1}$, $i=0,\ldots,m+1$, and $y_{j}=c+jh_{2}$, $j=0,\ldots,n+1$. Using central difference, we get $\begin{array}[]{c}D\varphi+\beta\varphi^{3}-\lambda\varphi=0,\\\ \frac{1}{2}\left(\frac{1}{h_{1}h_{2}}-\varphi^{\mathrm{T}}\varphi\right)=0,\end{array}$ (18) where $\varphi=(\varphi_{11},\cdots,\varphi_{1n},\cdots,\varphi_{m1},\cdots,\varphi_{mn})^{\mathrm{T}}$, $\varphi_{ij}$ are the approximations of $\phi(x_{i},y_{j})$, $v_{ij}=V(x_{i},y_{j})$, $i=1,\ldots,m$, $j=1,\ldots,n$ and D is a block tridiagonal matrix $D=\frac{1}{2}\left(\begin{array}[]{cccc}D_{11}&D_{12}&\\\ D_{21}&D_{22}&\ddots&\\\ &\ddots&\ddots&D_{{m-1},m}\\\ &&D_{m,{m-1}}&D_{mm}\end{array}\right)+\left(\begin{array}[]{cccc}V_{1}&&\\\ &V_{2}&&\\\ &&\ddots&\\\ &&&V_{m}\end{array}\right),$ where $\displaystyle D_{ii}=\left(\begin{array}[]{cccc}\frac{2}{h_{1}^{2}}+\frac{2}{h_{2}^{2}}&-\frac{1}{h_{2}^{2}}&\\\ -\frac{1}{h_{2}^{2}}&\frac{2}{h_{1}^{2}}+\frac{2}{h_{2}^{2}}&\ddots&\\\ &\ddots&\ddots&-\frac{1}{h_{2}^{2}}\\\ &&-\frac{1}{h_{2}^{2}}&\frac{2}{h_{1}^{2}}+\frac{2}{h_{2}^{2}}\end{array}\right)\in\mathbb{R}^{n\times n},$ (23) $\displaystyle D_{i-1,i}=\left(\begin{array}[]{cccc}-\frac{1}{h_{1}^{2}}&&&\\\ &-\frac{1}{h_{1}^{2}}&&\\\ &&\ddots&\\\ &&&-\frac{1}{h_{1}^{2}}\end{array}\right)\in\mathbb{R}^{n\times n},\quad D_{{i-1},i}=D_{i,{i-1}},$ (28) $\displaystyle V_{i}=\left(\begin{array}[]{cccc}v_{i1}&&&\\\ &v_{i2}&&\\\ &&\ddots&\\\ &&&v_{in}\end{array}\right),\quad i=1,\ldots,m.$ (33) ###### Remark 2.1 For both one and two dimensional cases, D is an irreducible symmetric diagonal dominant matrix and the diagonal entries of D are all positive. Therefore D is positive definite. ### 2.2 Existence of certain types of solution In this subsection, we will study the existence of solution for the discretized nonlinear eigenvalue problem. From [4], we know for (4)-(6), when $\beta>0$, the positive ground state is unique, and if $V(x)$ is radially symmetric in 2D, the positive ground state must be radially symmetric. We will prove the existence of positive solution and the existence of antisymmetric solution for discretized nonlinear eigenvalue problem. For convenient reading, two underlying theorems from [9] are quoted as underlying lemmas. ###### Lemma 2.2 ([9]) Let A be an irreducible Stieltjes matrix and $\mu$ be the smallest positive eigenvalue of A. Let $\lambda>\mu$ and let $F(x)=\left(\begin{array}[]{c}f_{1}(x_{1})\\\ \vdots\\\ f_{n}(x_{n})\end{array}\right),$ (34) where for $i=1,\ldots,n$, $f_{i}(x):[0,\infty)\rightarrow[0,\infty)$ are $C^{1}$ functions satisfying the conditions: $\lim\limits_{t\rightarrow 0}\frac{f_{i}(t)}{t}=0,\qquad\lim\limits_{t\rightarrow\infty}\frac{f_{i}(t)}{t}=\infty.$ (35) Then $Ax+F(x)=\lambda x$ has a positive solution. If, in addition, for $i=1,\ldots,n$, $\frac{f_{i}(s)}{s}<\frac{f_{i}(t)}{t}$ (36) whenever $0<s<t$, then the solution is unique. ###### Lemma 2.3 ([9]) Let the conditions $(\ref{sec 2: limit conditon})$ and $(\ref{sec 2: inequality condition})$ of Lemma 2.2 be satisfied and let $x(\lambda)$ denote the unique positive eigenvector corresponding to $\lambda\in(\mu,\infty)$. Then: 1. (i) $x(\lambda_{1})<x(\lambda_{2})$, if $\mu<\lambda_{1}<\lambda_{2}<\infty$; 2. (ii) $x(\lambda)$ is continuous on $(\mu,\infty)$; 3. (iii) $\lim\limits_{\lambda\rightarrow\infty}{x_{i}(\lambda)}=\infty$, $i=1,\cdots,n$; 4. (iv) $\lim\limits_{\lambda\rightarrow\mu^{+}}{x_{i}(\lambda)}=0$, $i=1,\cdots,n$. ###### Remark 2.4 Lemma 2.3 indicates for any given normalization $r>0$, there exist a unique $\lambda>\mu$ and unique positive $x(\lambda)$ such that $||x(\lambda)||=r$. ###### Theorem 2.5 If $\beta>0$, there exist unique positive eigenvectors for problem $(\ref{sec 2: finite difference - one dimen})$ and $(\ref{sec 2: finite difference - two dimen})$ respectively. Proof. From Remark 2.1, we know $D$ in both $(\ref{sec 2: finite difference - one dimen})$ and $(\ref{sec 2: finite difference - two dimen})$ is an irreducible Stieltjes matrix. In addition it can be verified that $\beta\varphi^{3}$ in both $(\ref{sec 2: finite difference - one dimen})$ and $(\ref{sec 2: finite difference - two dimen})$ satisfies the conditions of $F(x)$ in Lemma 2.2. From Remark 2.4, the claim is proved. ###### Theorem 2.6 Let $\beta>0$ and $\Omega=[-a,a]$ for problem $(\ref{sec 2: finite difference - one dimen})$. 1. (i) When $n$ is odd and the grid points $x_{i}$, $i=1,\ldots,n$, satisfy $x_{1}=-x_{n},~{}x_{2}=-x_{n-1},~{}\cdots,~{}x_{\frac{n-1}{2}}=-x_{\frac{n+3}{2}},~{}x_{\frac{n+1}{2}}=0,$ there exists a unique solution $\varphi=(\varphi_{1},\varphi_{2},\cdots,\varphi_{n})^{\mathrm{T}}$ with $\varphi_{1}=-\varphi_{n}$, $\varphi_{2}=-\varphi_{n-1},\cdots$, $\varphi_{\frac{n-1}{2}}=-\varphi_{\frac{n+3}{2}}$, $\varphi_{\frac{n+1}{2}}=0$, $\varphi_{j}>0$, $j=1,\ldots,n$. 2. (ii) When $n$ is even and the grid points $x_{i}$, $i=1,\ldots,n$, satisfy $x_{1}=-x_{n},~{}x_{2}=-x_{n-1},~{}\cdots,~{}x_{\frac{n}{2}}=-x_{\frac{n}{2}+1},$ there exists a unique solution $\varphi=(\varphi_{1},\varphi_{2},\cdots,\varphi_{n})^{\mathrm{T}}$ with $\varphi_{1}=-\varphi_{n}$, $\varphi_{2}=-\varphi_{n-1}$, $\cdots$, $\varphi_{\frac{n}{2}}=-\varphi_{\frac{n}{2}+1}$, $\varphi_{j}>0$, $j=1,\ldots,n$. Proof. (i). When $n$ is odd, consider the following equations $\begin{array}[]{c}D_{2}\varphi+\beta\varphi^{3}=\lambda\varphi,\\\ \varphi^{\mathrm{T}}\varphi-\frac{1}{2h}=0,\end{array}$ (37) where $\displaystyle D_{2}=\left(\begin{array}[]{cccc}\frac{1}{h^{2}}+\frac{1}{2}x_{1}^{2}&-\frac{1}{2h^{2}}&&\\\ -\frac{1}{2h^{2}}&\frac{1}{h^{2}}+\frac{1}{2}x_{2}^{2}&\ddots&\\\ &\ddots&\ddots&-\frac{1}{2h^{2}}\\\ &&-\frac{1}{2h^{2}}&\frac{1}{h^{2}}+\frac{1}{2}x_{\frac{n-1}{2}}^{2}\end{array}\right)\in\mathbb{R}^{\frac{n-1}{2}\times\frac{n-1}{2}}.$ (42) Note that $D_{2}$ is an irreducible Stieltjes matrix and $\beta\varphi^{3}$ satisfies the conditions of $F(x)$ in Lemma 2.2. Therefore there exists a unique positive solution $\varphi=(\varphi_{1},\varphi_{2},\ldots,\varphi_{\frac{n-1}{2}})^{\mathrm{T}}$ for (37). Due to the relations $x_{1}=-x_{n},~{}x_{2}=-x_{n-1},~{}\cdots,~{}x_{\frac{n-1}{2}}=-x_{\frac{n+3}{2}},~{}x_{\frac{n+1}{2}}=0$, set $\varphi_{n}=-\varphi_{1}$, $\varphi_{n-1}=-\varphi_{2}$, $\cdots$, $\varphi_{\frac{n+3}{2}}=-\varphi_{\frac{n-1}{2}}$, $\varphi_{\frac{n+1}{2}}=0$. Then $\varphi=(\varphi_{1},\varphi_{2},\ldots,\varphi_{n})^{\mathrm{T}}$ is a solution of $(\ref{sec 2: finite difference - one dimen})$. (ii). When $n$ is even, consider the following equations $\begin{array}[]{c}D_{2}\varphi+\beta\varphi^{3}=\lambda\varphi,\\\ \varphi^{\mathrm{T}}\varphi-\frac{1}{2h}=0,\end{array}$ (43) where $\displaystyle D_{2}=\left(\begin{array}[]{cccc}\frac{1}{h^{2}}+\frac{1}{2}x_{1}^{2}&-\frac{1}{2h^{2}}&&\\\ -\frac{1}{2h^{2}}&\frac{1}{h^{2}}+\frac{1}{2}x_{2}^{2}&\ddots&\\\ &\ddots&\ddots&-\frac{1}{2h^{2}}\\\ &&-\frac{1}{2h^{2}}&\frac{3}{2h^{2}}+\frac{1}{2}x_{\frac{n}{2}}^{2}\end{array}\right)\in\mathbb{R}^{\frac{n}{2}\times\frac{n}{2}}.$ (48) Similarly there exists a unique positive solution $\varphi=(\varphi_{1},\varphi_{2},\ldots,\varphi_{\frac{n}{2}})^{\mathrm{T}}$ for (43). Due to the relations $x_{1}=-x_{n},~{}x_{2}=-x_{n-1},~{}\cdots,~{}x_{\frac{n}{2}}=-x_{\frac{n}{2}+1}$, set $\varphi_{n}=-\varphi_{1}$, $\varphi_{n-1}=-\varphi_{2}$, $\cdots$, $\varphi_{\frac{n}{2}+1}=-\varphi_{\frac{n}{2}}$. Then $\varphi=(\varphi_{1},\varphi_{2},\ldots,\varphi_{n})^{\mathrm{T}}$ is a solution of $(\ref{sec 2: finite difference - one dimen})$. ## 3 Homotopy methods In this section, in order to compute many eigenpairs, we construct homotopy equations for 1D discretized problem $(\ref{sec 2: finite difference - one dimen})$ and 2D discretized problem $(\ref{sec 2: finite difference - two dimen})$ respectively. We shall prove the regularity and boundedness of the homotopy paths. The regularity of homotopy paths can be usually obtained by random perturbations of appropriate parameters, so the most important feature of our construction is the choice of appropriate parameters. In addition, if the initial matrix can be chosen as close to the matrix $D$ as possible, then most of the homotopy paths are close to straight lines and will be easy to follow [19]. ### 3.1 One dimensional case For $(\ref{sec 2: finite difference - one dimen})$, a homotopy $H:\mathbb{R}^{n}\times\mathbb{R}\times[0,1]\rightarrow\mathbb{R}^{n}\times\mathbb{R}$ is defined as $H(\varphi,\lambda,t)=\left(\begin{array}[]{c}(1-t)A(K)\varphi+D\varphi+t\beta\varphi^{3}-\lambda\varphi\\\ \frac{1}{2}\left(\frac{1}{h}-\varphi^{\mathrm{T}}\varphi\right)\end{array}\right)=0,$ (49) where $A(K)=\mbox{diag}(a_{1},\cdots,a_{n})$ is a random diagonal matrix with $K=(a_{1},\cdots,a_{n})^{\mathrm{T}}\in\mathbb{R}^{n}$. At $t=0$, $H(\varphi,\lambda,0)$ corresponds to the linear eigenvalue problem $H(\varphi,\lambda,0)=\left(\begin{array}[]{c}A(K)\varphi+D\varphi-\lambda\varphi\\\ \frac{1}{2}\left(\frac{1}{h}-\varphi^{\mathrm{T}}\varphi\right)\end{array}\right)=0,$ (50) while at $t=1$, $H(\varphi,\lambda,1)=0$ corresponds to the problem (9). Assuming that the eigenpairs of $H(\varphi,\lambda,0)=0$ are $(\varphi^{(i)},\lambda_{i})$, $i=1,\ldots,n$, $\lambda_{1}\leq\cdots\leq\lambda_{n}$, we shall use these $n$ points $(\varphi^{(i)},\lambda_{i},0)$ as our initial points when tracing the homotopy curves leading to the desired solutions of $H(\varphi,\lambda,1)=0$. The choice of the initial matrix $A(K)+D$ provides some advantages. First, it makes sure that $\forall t\in[0,1)$, what we need to solve is a sparse nonlinear eigenvalue problem with constraint. Second, since $\forall K\in\mathbb{R}^{n}$, $\forall t\in[0,1)$ and $\varphi\in\mathbb{R}^{n}$, $(1-t)A(K)+D+t\beta\mbox{diag}(\varphi^{2})$ is a real symmetric matrix, the solution curves starting from the initial points at $t=0$ are real. Finally, since $A(K)+D$ is a tridiagonal matrix with all the subdiagonal and supdiagonal entries nonzero, all the eigenvalues of $A(K)+D$ are simple and the Jacobian matrix of $H$ at $(\varphi_{0},\lambda_{0},0)$ is nonsingular for $\forall(\varphi_{0},\lambda_{0})$ such that $H(\varphi_{0},\lambda_{0},0)=0$. Thus locally a unique curve around $(\varphi_{0},\lambda_{0},0)$ is guaranteed. The effectiveness of the homotopy is based on the following Parametrized Sard’s Theorem. ###### Theorem 3.1 (Parametrized Sard’s Theorem) Let $f:M\times P\subset\mathbb{R}^{m}\times\mathbb{R}^{q}\to\mathbb{R}^{n}$ be a $C^{k}$ mapping with $k>max(0,m-n)$, where $M$ and $P$ are open sets in $\mathbb{R}^{m}$ and $\mathbb{R}^{q}$ respectively. If $y$ is a regular value of f, then y is also a regular value of $f(\cdot,p)$ for almost all $p\in P$. In the rest of this paper, we will denote the $i$-th row of a matrix $M$ as $M(i,:)$ and the $j$-th column of $M$ as $M(:,j)$. If $I_{1}$ is a row index set, $M(I_{1},:)$ will be the submatrix formed by the $I_{1}$ rows of $M$. $M(i:end,:)$ will be the submatrix formed by the rows from the $i$-th row to the last. Similarly $M(:,J_{1})$ will be the submatrix formed by the $J_{1}$ columns of $M$. If $I_{1}$ and $I_{2}$ are two index sets, $[M(I_{1},:);M(I_{2},:)]$ denotes the submatrix formed by the $I_{1}$ rows and the $I_{2}$ rows of $M$. In Theorem 3.2, we prove regularity and boundedness of the homotopy paths determined by the homotopy equation (49). ###### Theorem 3.2 For the homotopy $H:\mathbb{R}^{n}\times\mathbb{R}\times[0,1)\rightarrow\mathbb{R}^{n}\times\mathbb{R}$ in (49), for almost all $K\in\mathbb{R}^{n}$, 1. (i) 0 is a regular value of $H$ and therefore the homotopy curves corresponding to different initial points do not intersect each other for $t\in[0,1)$; 2. (ii) Every homotopy path $(\varphi(s),\lambda(s),t(s))\subset H^{-1}(0)$ is bounded. Proof. (i) Define a mapping $\tilde{H}:\mathbb{R}^{n}\times\mathbb{R}\times[0,1]\times\mathbb{R}^{n}\rightarrow\mathbb{R}^{n}\times\mathbb{R}$ related to $H$ as follows $\tilde{H}(\varphi,\lambda,t,K)=\left(\begin{array}[]{c}(1-t)A(K)\varphi+D\varphi+t\beta\varphi^{3}-\lambda\varphi\\\ \frac{1}{2}\left(\frac{1}{h}-\varphi^{\mathrm{T}}\varphi\right)\end{array}\right)=0,$ (51) such that $H(\varphi,\lambda,t)=\tilde{H}(\varphi,\lambda,t,K)$. The Jacobian matrix of $\tilde{H}$, $\frac{\partial\tilde{H}}{\partial(\varphi,\lambda,t,K)}$, is $\left(\begin{array}[]{cccc}(1-t)A(K)+D+3t\beta\mbox{diag}(\varphi^{2})-\lambda I&-\varphi&\beta\varphi^{3}-A(K)\varphi&(1-t)\mbox{diag}(\varphi)\\\ -\varphi^{\mathrm{T}}&0&0&0\end{array}\right).$ Divide $\\{1,\cdots,n\\}$ into two parts $C^{0}$ and $C^{*}$, where $C^{0}$ denotes the indices $i$ such that $\varphi_{i}=0$ and $C^{*}$ denotes the indices $i$ such that $\varphi_{i}\neq 0$. Since $\forall i\in C^{0}$, both the columns $3t\beta\mbox{diag}(\varphi^{2})(:,i)$ and $t\beta\mbox{diag}(\varphi^{2})(:,i)$ are equal to zero, it holds that, $\displaystyle\left((1-t)A(K)+D+3t\beta\mbox{diag}(\varphi^{2})-\lambda I\right)(:,C^{0})$ $\displaystyle=$ $\displaystyle\left((1-t)A(K)+D+t\beta\mbox{diag}(\varphi^{2})-\lambda I\right)(:,C^{0}).$ (52) For the columns in $C^{*}$, the diagonal matrix $\mbox{diag}(\varphi)$ is nonzero. By elementary column transformations, the Jacobian matrix $\frac{\partial\tilde{H}}{\partial(\varphi,\lambda,t,K)}$ is transformed to the following, $\left(\begin{array}[]{cccc}(1-t)A(K)+D+t\beta\mbox{diag}(\varphi^{2})-\lambda I&-\varphi&\beta\varphi^{3}-A(K)\varphi&(1-t)\mbox{diag}(\varphi)\\\ -\varphi^{\mathrm{T}}&0&0&0\end{array}\right).$ Define $\displaystyle F_{1}=\left(\begin{array}[]{cc}(1-t)A(K)+D+t\beta\mbox{diag}(\varphi^{2})-\lambda I&-\varphi\\\ -\varphi^{\mathrm{T}}&0\end{array}\right).$ (55) For $\forall K\in\mathbb{R}^{n}$, $\forall(\varphi,\lambda,t)\in\mathbb{R}^{n}\times R\times[0,1)$ satisfying $(\ref{sec3: mapping-related-to-homotopy-1D})$, since the subdiagonals and supdiagonals of the matrix $(1-t)A(K)+D+t\beta\mbox{diag}(\varphi^{2})$ are nonzero, $\lambda$ is a simple eigenvalue of $(1-t)A(K)+D+t\beta\mbox{diag}(\varphi^{2})$. Therefore $F_{1}$ is nonsingular and $\frac{\partial\tilde{H}}{\partial(\varphi,\lambda,t,K)}$ is row full rank. As a result, 0 is a regular value of the mapping $\tilde{H}$. From Theorem 3.1, for almost all $K\in\mathbb{R}^{n}$, 0 is a regular value of the restricted mapping $\tilde{H}(\cdot,\cdot,\cdot,K)$, i.e., $H$. (ii) From $(\ref{sec 3: homotopy one dimen})$, for fixed $K\in\mathbb{R}^{n}$, $\forall(\varphi,\lambda,t)\in\mathbb{R}^{n}\times\mathbb{R}\times[0,1)$ satisfying $H(\varphi,\lambda,t)$$=$$0$, $\|\varphi\|=1/\sqrt{h}$ and $\lambda=h\left((1-t)\varphi^{\mathrm{T}}A(K)\varphi+\varphi^{\mathrm{T}}D\varphi+t\beta\varphi^{\mathrm{T}}\varphi^{3}\right)$. Then $\displaystyle|\lambda|$ $\displaystyle=$ $\displaystyle h|(1-t)\varphi^{\mathrm{T}}A(K)\varphi+\varphi^{\mathrm{T}}D\varphi+t\beta\varphi^{\mathrm{T}}\varphi^{3}|$ $\displaystyle\leq$ $\displaystyle(1-t)\rho(A(K))+\rho(D)+\frac{t\beta}{h}$ $\displaystyle\leq$ $\displaystyle\rho(A(K))+\rho(D)+\frac{\beta}{h},$ where $\rho(M)$ denotes the spectral radius of $M$. ### 3.2 Two dimensional case #### 3.2.1 The homotopy with random tridiagonal matrix For two dimensional case, the homotopy $H:\mathbb{R}^{mn}\times\mathbb{R}\times[0,1]\rightarrow\mathbb{R}^{mn}\times\mathbb{R}$ is constructed as follows $H(\varphi,\lambda,t)=\left(\begin{array}[]{c}(1-t)A(K)\varphi+D\varphi+t\beta\varphi^{3}-\lambda\varphi\\\ \frac{1}{2}\left(\frac{1}{h_{1}h_{2}}-\varphi^{\mathrm{T}}\varphi\right)\end{array}\right)=0,$ (56) where $A(K)\in\mathbb{R}^{mn\times mn}$ is a random block diagonal matrix with tridiagonal blocks, namely, $A(K)=\left(\begin{array}[]{cccc}A_{1}&&&\\\ &A_{2}&&\\\ &&\ddots&\\\ &&&A_{m}\end{array}\right)\mbox{with}~{}A_{i}=\left(\begin{array}[]{ccccc}a_{11}^{(i)}&a_{12}^{(i)}&&&\\\ a_{12}^{(i)}&a_{22}^{(i)}&a_{23}^{(i)}&&\\\ &a_{23}^{(i)}&a_{33}^{(i)}&\ddots&\\\ &&\ddots&\ddots&a_{n-1,n}^{(i)}\\\ &&&a_{n-1,n}^{(i)}&a_{nn}^{(i)}\end{array}\right),~{}i=1,\ldots,m,$ and $K=\left(K_{1},K_{2},\ldots,K_{m}\right)^{\mathrm{T}}$ with $K_{i}=\left(a_{11}^{(i)},a_{12}^{(i)},a_{22}^{(i)},a_{23}^{(i)},\ldots,a_{n-1,n-1}^{(i)},a_{n-1,n}^{(i)},a_{nn}^{(i)}\right)$. $H(\varphi,\lambda,0)$ corresponds to the linear eigenvalue problem $H(\varphi,\lambda,0)=\left(\begin{array}[]{c}A(K)\varphi+D\varphi-\lambda\varphi\\\ \frac{1}{2}\left(\frac{1}{h_{1}h_{2}}-\varphi^{\mathrm{T}}\varphi\right)\end{array}\right)=0,$ while $H(\varphi,\lambda,1)$ corresponds to the problem $(\ref{sec 2: finite difference - two dimen})$. In order to show the effectiveness of this homotopy $H$ by the Parametrized Sard’s Theorem, define a mapping $\tilde{H}:\mathbb{R}^{mn}\times\mathbb{R}\times[0,1]\times\mathbb{R}^{(2nm-m)}\rightarrow\mathbb{R}^{mn}\times\mathbb{R}$ related to $H$ as follows $\tilde{H}(\varphi,\lambda,t,K)=\left(\begin{array}[]{c}(1-t)A(K)\varphi+D\varphi+t\beta\varphi^{3}-\lambda\varphi\\\ \frac{1}{2}\left(\frac{1}{h_{1}h_{2}}-\varphi^{\mathrm{T}}\varphi\right)\end{array}\right)=0,$ (57) such that $H(\varphi,\lambda,t)=\tilde{H}(\varphi,\lambda,t,K)$. The Jacobian matrix of $\tilde{H}$, $\frac{\partial\tilde{H}}{\partial(\varphi,\lambda,t,K)}$, is $\left(\begin{array}[]{cccc}(1-t)A(K)+D+3t\beta\mbox{diag}(\varphi^{2})-\lambda I&-\varphi&\beta\varphi^{3}-A(K)\varphi&(1-t)B\\\ -\varphi^{\mathrm{T}}&0&0&0\end{array}\right),$ (58) where $B=\frac{\partial(A(K)\varphi)}{\partial K}\in\mathbb{R}^{{(mn)}\times{(2nm-m)}}$. Denote $\varphi_{i}=\left(\varphi_{i1},\varphi_{i2},\ldots,\varphi_{in}\right)^{\mathrm{T}}$. It can be verified that $B=\left(\begin{array}[]{cccc}B_{1}&&&\\\ &B_{2}&&\\\ &&\ddots&\\\ &&&B_{m}\end{array}\right)$ (59) with $B_{i}=\frac{\partial(A_{i}\varphi_{i})}{\partial K_{i}}=\left(\begin{array}[]{ccccccccccc}\varphi_{i1}&\varphi_{i2}&&&&&&&&&\\\ &\varphi_{i1}&\varphi_{i2}&\varphi_{i3}&&&&&&&\\\ &&&\varphi_{i2}&\varphi_{i3}&\varphi_{i4}&&&&&\\\ &&&&&&\ddots&&&&\\\ &&&&&&&\varphi_{i,n-2}&\varphi_{i,n-1}&\varphi_{in}&\\\ &&&&&&&&&\varphi_{i,n-1}&\varphi_{in}\end{array}\right).$ (60) For example, when $m=2,n=3$, we have $B=\begin{pmatrix}\varphi_{11}&\varphi_{12}&{0}&{0}&{0}&{0}&{0}&{0}&{0}&{0}\\\ {0}&\varphi_{11}&\varphi_{12}&\varphi_{13}&{0}&{0}&{0}&{0}&{0}&{0}\\\ {0}&{0}&{0}&\varphi_{12}&\varphi_{13}&{0}&{0}&{0}&{0}&{0}\\\ {0}&{0}&{0}&{0}&{0}&\varphi_{21}&\varphi_{22}&{0}&{0}&{0}\\\ {0}&{0}&{0}&{0}&{0}&{0}&\varphi_{21}&\varphi_{22}&\varphi_{23}&{0}\\\ {0}&{0}&{0}&{0}&{0}&{0}&{0}&{0}&\varphi_{22}&\varphi_{23}\end{pmatrix}.$ Denote by $G_{a}$ all the inner grid points and by $R_{a}$ the ordering of the grid points in $G_{a}$, i.e., $\displaystyle G_{a}=\\{(i,j):i=1,\ldots,m,j=1,\ldots,n\\},$ (61) $\displaystyle R_{a}=\\{j+(i-1)*n:(i,j)\in G_{a}\\},$ (62) and by $G^{0}$ the grid points with function value being zero, $\displaystyle G^{0}=\\{(i,j)\in G_{a}:\varphi_{ij}=0\\}.$ (63) Note that $G_{a}$ and $R_{a}$ have a one to one correspondence. Define such correspondence as a mapping $\Gamma:G_{a}\rightarrow R_{a}$, $\displaystyle\Gamma(i,j)=j+(i-1)*n.$ (64) Denote by $R^{0}$ the indices of rows in which $B$ is zero, by $R^{*}$ the indices of rows, in which $B$ is not zero, and $S_{i}^{0}$ and $S_{i}^{*}$ with similar meanings for $B_{i}$, $\displaystyle R^{0}=\\{r:B(r,:)=0\\},\quad R^{*}=\\{r:B(r,:)\neq 0\\},$ (65) $\displaystyle S^{0}_{i}=\\{r:B_{i}(r,:)=0\\},\quad S_{i}^{*}=\\{r:B_{i}(r,:)\neq 0\\}.$ (66) It can be verified that $\displaystyle R^{0}=\bigcup_{i=1}^{m}S^{0}_{i},\quad R^{*}=\bigcup_{i=1}^{m}S_{i}^{*}.$ (67) ###### Lemma 3.3 Let $B$ be the matrix defined in (59). Then the nonzero rows of $B$ are linearly independent, i.e., the submatrix $B(R^{*},:)$ is row full rank. Proof. Note that $B(R^{*},:)=[B(S_{1}^{*},:);\ldots;B(S_{m}^{*},:)]$. Due to the block structure of $B$, it suffices to prove that for any $i$, $1\leq i\leq m$, $B_{i}(S_{i}^{*},:)$ is row full rank if $S_{i}^{*}$ is not empty. Now suppose that $S_{i}^{*}$ is not empty, that is $\varphi_{i}\neq 0$. Let the nonzero components of $\varphi_{i}$ be $\varphi_{i,i_{1}},\varphi_{i,i_{2}},\ldots,\varphi_{i,i_{r}}$, with $i_{1}<i_{2}<\ldots<i_{r}$. Denote by $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k}})$ the rows of $B_{i}$ containing $\varphi_{i,i_{1}},\varphi_{i,i_{2}},\ldots,\varphi_{i,i_{k}}$, $1\leq k\leq r$. Then $B_{i}(S_{i}^{*},:)=Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{r}})$. The claim will be proved by successively adding rows with nonzero component of $\varphi_{i}$ to $Z$. (1) Prove $Z(\varphi_{i,i_{1}})$ is row full rank. If $i_{1}=1$, $Z(\varphi_{i,i_{1}})$ is the first two rows of $B_{i}$ and it is row full rank. If $1<i_{1}<n$, $Z(\varphi_{i,i_{1}})$ is three successive rows of $B_{i}$ which contains the following submatrix involving $\varphi_{i,i_{1}}$, $\left(\begin{array}[]{ccc}\varphi_{i,i_{1}}&0&0\\\ &\varphi_{i,i_{1}}&*\\\ 0&0&\varphi_{i,i_{1}}\end{array}\right)$ where $*$ stands for an element which may be zero or nonzero. Therefore $Z(\varphi_{i,i_{1}})$ is row full rank. If $i_{1}=n$, $Z(\varphi_{i,i_{1}})$ is the last two rows of $B_{i}$ and is row full rank too. (2) Prove that when $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k}})$ is row full rank, $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k+1}})$ is also row full rank, $1\leq k<r$. If $i_{k}=n-1$, then $i_{k+1}=n$ and $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k+1}})=Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k}})$. Next suppose $i_{k}<n-1$. If $i_{k+1}=i_{k}+1$, i.e., $\varphi_{i,i_{k}}$ and $\varphi_{i,i_{k+1}}$ are successive, then $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k+1}})$ consists of $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k}})$ and a new row with $\varphi_{i,i_{k+1}}$ as its first nonzero element and with $\varphi_{i,i_{k+1}}$ located in a column different from those of $\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k}}$, the connecting submatrix illustrated in the following, $\left(\begin{array}[]{cccccc}*&\varphi_{i,i_{k}}&\varphi_{i,i_{k+1}}&0&0&0\\\ 0&0&\varphi_{i,i_{k}}&\varphi_{i,i_{k+1}}&*&0\\\ 0&0&0&0&\varphi_{i,i_{k+1}}&*\end{array}\right)$ Thus $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k+1}})$ is row full rank. If $i_{k+1}=i_{k}+2<n$, then $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k+1}})$ consists of $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k}})$ and two new rows and similarly these two new rows are linearly independent with $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k}})$. If $i_{k+1}\geq i_{k}+3$ and $i_{k+1}<n$, then $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k+1}})$ consists of $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k}})$ and three new rows. If $i_{k+1}=n$, $i_{k+1}=i_{k}+2$, then $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k+1}})$ consists of $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k}})$ and a new row. If $i_{k+1}=n$, $i_{k+1}=i_{k}+s$ and $s\geq 3$, then $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k+1}})$ consists of $Z(\varphi_{i,i_{1}},\ldots,\varphi_{i,i_{k}})$ and two new rows. All the latter cases can be similarly proved. In the following, first we prove there exists a zero measure set $U_{1}\subset\mathbb{R}^{(2nm-m)}$, such that if $K\in\mathbb{R}^{(2nm-m)}\setminus U_{1}$, the eigenvalues of $A(K)+D$ are simple. Then we prove that 0 is a regular value of $H(\varphi,\lambda,t)$ for almost all $K\in\mathbb{R}^{(2nm-m)}\setminus(U_{1}\cup U_{2})$, where $U_{2}=(\mathbb{R}^{+})^{(2nm-m)}$. The removal of $U_{2}$ is to make the elements in the subdiagonal and supdiagonal of the matrix $(1-t)A(K)+D$ negative for $t\in[0,1)$. ###### Lemma 3.4 The eigenvalues of $A(K)+D$ are simple for $K$ almost everywhere in $\mathbb{R}^{(2nm-m)}$ except on a subset of real codimension 1. Proof. Let $f(\lambda)=det(A(K)+D-\lambda I)$. The polynomial $f(\lambda)$ has no multiple roots if and only if its discriminant $R(K)$ is nonzero [22]. It is obvious that $R(K)$ is not identically zero. Furthermore, since $R(K)$ is a polynomial in the elements of vector $K$, it can vanish only on a hypersurface of real codimension 1. The hypersurface is $U_{1}=\\{K|R(K)=0\\}.$ In the following Lemma 3.5, we prove that for any matrix $F$ consisting of several submatrices by row, if every submatrix of $F$ is row full rank and the index sets of nonzero columns do not intersect for any two submatrices, then $F$ is row full rank. ###### Lemma 3.5 Let $F\in\mathbb{R}^{p\times q}$ be a matrix. Suppose $\\{1,\ldots,p\\}=\bigcup_{i=1}^{s}I_{i}$, with $I_{i}\bigcap I_{j}=\emptyset$, if $i\neq j$. Denote $E_{i}=\\{k\in\\{1,\ldots,q\\}:F(I_{i},k)\neq 0\\}.$ Suppose that for any $1\leq i\leq s$, $F(I_{i},:)$ is row full rank and $E_{i}\bigcap E_{j}=\emptyset$, if $i\neq j$. Then $F$ is row full rank. Proof. Since for any $1\leq i\leq s$, $F(I_{i},:)$ is row full rank, there exist a column index set $J_{i}\subset E_{i}$ such that $F(I_{i},J_{i})$ is a nonsingular submatrix. Correspondingly, the matrix $[F(I_{1},:);F(I_{2},:);\ldots;F(I_{s},:)]$ has a nonsingular submatrix as follows, $\left(\begin{array}[]{cccc}F(I_{1},J_{1})&&&\\\ &F(I_{2},J_{2})&&\\\ &&\ddots&\\\ &&&F(I_{s},J_{s})\end{array}\right).$ As a result, the matrix $[F(I_{1},:);F(I_{2},:);\ldots;F(I_{s},:)]$ is row full rank and so is $F$. To prove that $0$ is a regular value of $\tilde{H}(\varphi,\lambda,t,K):\mathbb{R}^{mn}\times R\times[0,1)\times\mathbb{R}^{(2nm-m)}\setminus(U_{1}\cup U_{2})\to\mathbb{R}^{mn+1}$, we need to prove $\forall(\varphi,\lambda,t,K)\in\mathbb{R}^{mn}\times R\times[0,1)\times\mathbb{R}^{(2nm-m)}\setminus(U_{1}\cup U_{2})$ satisfying $\tilde{H}(\varphi,\lambda,t,K)=0$, the Jacobian matrix of $\tilde{H}(\varphi,\lambda,t,K)$ is row full rank. $\forall(\varphi,\lambda,t,K)$ satisfying $\tilde{H}(\varphi,\lambda,t,K)=0$, for $B$ defined in (59), we have $B(R^{0},:)=0$. Correspondingly for the Jacobian matrix defined in (58), we have $\displaystyle((1-t)A+D+3t\beta\mbox{diag}(\varphi)^{2}-\lambda I)(R^{0},:)=((1-t)A+D-\lambda I)(R^{0},:).$ Through row permutations, the Jacobian matrix of $\tilde{H}(\varphi,\lambda,t,K)$, $\frac{\partial\tilde{H}}{\partial(\varphi,\lambda,t,K)}$, can be rewritten as $\left(\begin{array}[]{cccc}((1-t)A+D+3t\beta\mbox{diag}(\varphi)^{2}-\lambda I)(R^{*},:)&-\varphi(R^{*})&(\beta\varphi^{3}-A\varphi)(R^{*})&(1-t)B(R^{*},:)\\\ ((1-t)A+D-\lambda I)(R^{0},:)&-\varphi(R^{0})&(\beta\varphi^{3}-A\varphi)(R^{0})&0\\\ -\varphi^{\mathrm{T}}&0&0&0\end{array}\right).$ $\forall(\varphi,\lambda,t,K)$ satisfying $\tilde{H}(\varphi,\lambda,t,K)=0$, from Lemma 3.3, we know $B(R^{*},:)$ is row full rank. Therefore, if we can prove $\left(\begin{array}[]{c}((1-t)A+D-\lambda I)(R^{0},:)\\\ -\varphi^{\mathrm{T}}\end{array}\right)$ (68) is row full rank, then $\frac{\partial\tilde{H}}{\partial(\varphi,\lambda,t,K)}$ is row full rank. From $\tilde{H}=0$, we have $((1-t)A+D-\lambda I)(R^{0},:)\varphi=0$, i.e., $\varphi$ is orthogonal to the rows of $((1-t)A+D-\lambda I)(R^{0},:)$. Therefore, if we can prove $((1-t)A+D-\lambda I)(R^{0},:)$ is row full rank, $\frac{\partial\tilde{H}}{\partial(\varphi,\lambda,t,K)}$ is row full rank. Now the problem is turned into proving that $((1-t)A+D-\lambda I)(R^{0},:)$ is row full rank. For easy exposition, some concepts concerning the topology of the grid points with zero function value are introduced. In addition, $(i,j)$ will be considered as grid point in the rest of this section, representing $(x_{i},y_{j})$. ###### Definition 3.6 In the grid $G_{a}$, a zero valued node $(i,j)$ is a grid point with $\varphi_{i,j}=0$. ###### Definition 3.7 Two zero valued nodes $(i_{1},j_{1})$ and $(i_{p},j_{p})$ are said to be zero valued connected, if there exist a sequence of zero valued nodes, $\displaystyle(i_{1},j_{1}),(i_{2},j_{2}),\ldots,(i_{p},j_{p}),$ (69) such that for any two successive nodes $(i_{k},j_{k})$ and $(i_{k+1},j_{k+1})$ of the sequence, the distance of these two nodes is 1 in the sense that $\displaystyle|i_{k}-i_{k+1}|+|j_{k}-j_{k+1}|=1.$ (70) ###### Definition 3.8 A set $S$ consisting of zero valued nodes is called a zero valued connected set if any two nodes of $S$ are zero valued connected. ###### Definition 3.9 A set $S$ consisting of zero valued nodes is called a zero valued connected component if $S$ is connected and S is the largest zero connected set containing $S$. From (56), the discretization of the differential equation in a stencil is written explicitly, $\displaystyle\alpha_{1}^{(ij)}\varphi_{ij}+\alpha_{2}^{(ij)}\varphi_{i-1,j}+\alpha_{3}^{(ij)}\varphi_{i+1,j}+\alpha_{4}^{(ij)}\varphi_{i,j+1}+\alpha_{5}^{(ij)}\varphi_{i,j-1}=0,$ (71) $\displaystyle\varphi_{0j}=\varphi_{m+1,j}=0,$ (72) $\displaystyle\varphi_{i0}=\varphi_{i,n+1}=0,$ (73) $\displaystyle\varphi^{\mathrm{T}}\varphi-\frac{1}{h_{1}h_{2}}=0,$ (74) where $\alpha_{1}^{(ij)}=(1-t)a_{jj}^{(i)}+\frac{1}{h_{1}^{2}}+\frac{1}{h_{2}^{2}}+v_{ij}+t\beta\varphi_{ij}^{2}-\lambda$, $\alpha_{2}^{(ij)}=\alpha_{3}^{(ij)}=(1-t)a_{jj+1}^{(i)}-\frac{1}{2h_{2}^{2}}$, $\alpha_{4}^{(ij)}=\alpha_{5}^{(ij)}=-\frac{1}{2h_{1}^{2}}$, and $i=1,\ldots,m$, $j=1,\ldots,n$. When $K\in R^{(2nm-m)}\setminus(U_{1}\bigcup U_{2})$, the sign relationships among the components of $\varphi$ are given in the following remark. ###### Remark 3.10 Let $(i,j)$ be an inner zero valued node. Assume that all except two of its neighbouring points are known to be zero valued nodes. If one of the rest two points is a zero valued node, then so is the other; If one of the rest two points is not a zero valued node, then neither is the other. For any set $G$ of zero valued nodes, denote by $R_{G}$ the set of indices of rows corresponding to $G$, in which the matrix $B$ is zero, i.e., $\displaystyle R_{G}=\\{r:B(r,:)=0,r=j+(i-1)*n,(i,j)\in G\\}.$ (75) Note that if a zero valued node $(i,j)\in G$ is such that $B(s,:)=0$ with $s=j+(i-1)*n$, then the neighbouring inner grid points in $y$-direction should be zero valued nodes, that is, both $(i,j-1)$ and $(i,j+1)$ are zero valued nodes if $1<j<n$, or $(i,j+1)$ is a zero valued node if $j=1$, or $(i,j-1)$ is a zero valued node if $j=n$. If $(i,j)\in G$ and the upper point $(i,j+1)$ or the lower point $(i,j-1)$ is not a zero valued node, then the corresponding row index of $s=\Gamma(i,j)$ will not be in $R_{G}$. Therefore $\Gamma^{-1}(R_{G})\subset G$. Note that $R_{G}$ may be empty even if $G$ is not empty. Let $M=(1-t)A+D-\lambda I$. Denote by $C_{G}$ the indices of columns corresponding to the zero valued connected set $G$, in which $M(R_{G},:)$ is not zero, i.e., $\displaystyle C_{G}=\\{c\in R_{a}:M(R_{G},c)\neq 0\\}.$ (76) If $R_{G}=\emptyset$, define $C_{G}=\emptyset$. For any $s\in R_{G}$, let $(i,j)=\Gamma^{-1}(s)$. Then the $\Gamma$ images of $(i,j)$ and its neighbouring inner grid points are possibly included in $C_{G}$. Denote by $g_{i}$ the $i$-th column of grid points in $x$-direction, i.e., $\displaystyle g_{i}=\\{(i,j):1\leq j\leq n\\},$ (77) and by $O_{1}$ ($O_{m}$) the ordering of all the inner grid points except the first (last) column in $y$-direction, $\displaystyle O_{1}=\\{r=j+(i-1)*n:(i,j)\in G_{a}\backslash g_{1}\\},$ (78) $\displaystyle O_{m}=\\{r=j+(i-1)*n:(i,j)\in G_{a}\backslash g_{m}\\}.$ (79) ###### Lemma 3.11 Both $M(O_{1},:)$ and $M(O_{m},:)$ are row full rank. Proof. It is obvious that $M(O_{1},:)$ has the following form: $\left(\begin{array}[]{ccccc}*&\ldots&\ldots&\ldots&\ldots\\\ &*&\ldots&\ldots&\ldots\\\ &&\ddots&\ldots&\ldots\\\ &&&*&\ldots\end{array}\right),$ where $*$ represents nonzero element. Therefore $M(O_{1},:)$ is row full rank. $M(O_{m},:)$ has the following form: $\left(\begin{array}[]{ccccc}\ldots&*&&&\\\ \ldots&\ldots&*&&\\\ \ldots&\ldots&\ldots&\ddots&\\\ \ldots&\ldots&\ldots&\ldots&*\end{array}\right).$ Therefore $M(O_{m},:)$ is row full rank. Note that if $G^{0}\bigcap g_{1}=\emptyset$, then $R^{0}\subset O_{1}$, from Lemma 3.11, $M(R^{0},:)$ is row full rank. In the following, we consider the case $G^{0}\bigcap g_{1}\not=\emptyset$. The set $G^{0}\bigcap g_{1}$ can have its own zero valued connected components. ###### Lemma 3.12 For $m\geq n\geq 6$, suppose that there is a zero valued connected component $\gamma$ of $G^{0}\bigcap g_{1}$ with $s$ points, $s\geq 2$. Then 1. (i) $s<n$; 2. (ii) If $\gamma$ contains the point $(1,1)$ or contains the point $(1,n)$, then the zero valued connected component $G$ of $G^{0}$ containing $\gamma$ will be the set of zero valued nodes starting from $\gamma$ and ending on column $s$ with $s+1-i$ zero valued nodes on column $i$, $1\leq i\leq s$; 3. (iii) Suppose that $\gamma$ is located in the inner part of $g_{1}$, i.e., the grid point $(1,j)$ of $\gamma$ is such that $2\leq j\leq n-1$. If $s$ is even, then the zero valued connected component $G$ of $G^{0}$ containing $\gamma$ will be the set of zero valued nodes starting from $\gamma$ and ending on column $s/2$ with $s+2-2i$ zero valued nodes on column $i$, $1\leq i\leq s/2$; 4. (iv) Suppose that $\gamma$ is located in the inner part of $g_{1}$. If $s$ is odd, then there is a zero valued connected set $G$ of $G^{0}$ containing $\gamma$ and arriving at a single zero valued node on column $(s+1)/2$ with $s+2-2i$ zero valued nodes on column $i$, $1\leq i\leq(s+1)/2$; 5. (v) For the cases (2), (3), (4), if $s\geq 3$, the index set of nonzero columns of $M$ corresponding to $G$ satisfies $C_{G}=\Gamma(G)$. Proof. 1. (i) If $s=n$, we have $\varphi_{11}=\varphi_{12}=\ldots=\varphi_{1n}=0$. From the sign relationships among the components of $\varphi$ as stated in Remark 3.10, the grid points in the column 2 are all zero valued nodes, namely, $\varphi_{21}=\varphi_{22}=\ldots=\varphi_{2n}=0$. By induction, all the grid points are zero valued nodes, namely, $\varphi_{11}=\ldots=\varphi_{1n}=\ldots=\varphi_{mn}=0$, a contradiction with the condition that $\varphi^{T}\varphi\neq 0$. Therefore $s<n$. 2. (ii) If $\gamma$ starts from the point $(1,1)$, from the sign relationships among the components of $\varphi$, the zero valued nodes connecting $\gamma$ on the column 2 of grid points are $(2,j)$, $1\leq j\leq s-1$. By induction, it can be seen that the zero valued connected component $G$ of $G^{0}$ containing $\gamma$ will end on column $s$. The zero valued nodes connecting $\gamma$ are illustrated in Fig 1 and form a zero valued component, where black dot represents zero and white dot represents nonzero. Similarly, the case $\gamma$ ends at the point $(1,n)$ can be proved. Figure 1: Illustration of a flag$(1,1)$$(1,2)$$(1,s-1)$$(1,s)$$(1,s+1)$$(1,n)$$(2,1)$$(s-1,1)$$(s,1)$$(s+1,1)$$(m,1)$ 3. (iii) Assume that $\gamma$ starts from the point $(1,j_{1})$ and ends at the point $(1,j_{2})$ with $j_{1}\geq 2$ and $j_{2}\leq n-1$. From the sign relationships among the components of $\varphi$, the zero valued nodes connecting $\gamma$ on the column 2 of grid points are $(2,j)$, $3\leq j\leq s-2$. The zero valued nodes connecting $\gamma$ are illustrated in Fig 2. The zero valued nodes connecting $\gamma$ will end on column $s/2$ with two zero valued nodes and form a zero valued component $G$. Figure 2: Illustration of a flag$(1,1)$$(1,t-1)$$(1,t)$$(1,t+1)$$(1,t+\frac{s}{2}-2)$$(1,t+\frac{s}{2}-1)$$(1,t+\frac{s}{2})$$(1,t+\frac{s}{2}+1)$$(1,t+s-2)$$(1,t+s-1)$$(1,t+s)$$(1,n)$$(2,1)$$(\frac{s}{2}-1,1)$$(\frac{s}{2},1)$$(m,1)$ 4. (iv) Similar to the case (3), the zero valued nodes connecting $\gamma$ are illustrated in Fig 3. However, since $s$ is odd, the zero valued nodes connecting $\gamma$ arrives at one single point on column $(s+1)/2$. All these zero valued nodes connecting $\gamma$ from column 1 to column $(s+1)/2$ form a zero valued connected set $G$. It may be or may not be a zero valued connected component. Figure 3: Illustration of a flag$(1,10)$$(1,t-1)$$(1,t)$$(1,t+1)$$(1,t+\frac{s-3}{2})$$(1,t+\frac{s-1}{2})$$(1,t+\frac{s+1}{2})$$(1,t+s-2)$$(1,t+s-1)$$(1,t+s)$$(1,n)$$(2,1)$$(\frac{s-1}{2},1)$$(\frac{s+1}{2},1)$$(m,1)$ 5. (v) From the figures for cases (2), (3), (4), it can be seen that $\Gamma^{-1}(R_{G})$ is the set of zero valued nodes by shrinking $G$ one layer from its nonzero boundary. $\Gamma^{-1}(C_{G})$ is the set of zero valued nodes by extending $\Gamma^{-1}(R_{G})$ one layer towards the nonzero boundary, which is $G$ itself. ###### Definition 3.13 Let $\gamma$ be a zero valued connected component of $G^{0}\bigcap g_{1}$ with $s$ points, $s\geq 2$. The zero valued connected set $G$ mentioned in (2), (3), (4) in Lemma 3.12 is called a flag of $\gamma$, denoted as $F_{\gamma}$. ###### Lemma 3.14 Let $\gamma$ be a zero valued connected component of $G^{0}\bigcap g_{1}$ of $s$ points with $s>2$ an odd number and $F_{\gamma}$ be its flag. Let $G$ be another set of zero valued nodes. Let $I_{1}=R_{F_{\gamma}}$ and $I_{2}=R_{G}$ be the row index sets as defined in (75) such that $M(I_{2},:)$ is row full rank. Denote $\displaystyle J_{1}=\\{j\in R_{a}:M(I_{1},j)\neq 0\\},\quad J_{2}=\\{j\in R_{a}:M(I_{2},j)\neq 0\\},$ (80) $\displaystyle J_{12}=J_{1}\bigcap J_{2}.$ (81) Suppose that $J_{12}$ has only one element $k$ and that both the column $M(I_{1},k)$ and the column $M(I_{2},k)$ have only one nonzero element, denoted as $M(s_{1},k)\neq 0$ and $M(s_{2},k)\neq 0$ respectively. Let $(i_{1},j_{1})$ and $(i_{2},j_{2})$ be the grid point corresponding to $s_{1}$ and $s_{2}$ respectively. Suppose $i_{2}=i_{1}+2$ and $j_{1}=j_{2}$. Then $[M(I_{1},:);M(I_{2},:)]$ is row full rank. Proof. Note that since $J_{12}$ is not empty, the zero valued connected component $\gamma$ corresponds to the case (4) of Lemma 3.12. Geometrically, for that $k$, the two grid points $(i_{1},j_{1})$ and $(i_{2},j_{2})$ corresponding to $M(s_{1},k)\neq 0$ and $M(s_{2},k)\neq 0$ lie in the same row of grid points with distance 2 in the sense that $|i_{2}-i_{1}|+|j_{2}-j_{1}|=2$, as illustrated locally in Figure 4. Figure 4: Local illustration of connection with a flag$(i_{1},j_{1})$$(i_{2},j_{2})$ Note that by Lemma 3.11, $M(I_{1},:)$ is row full rank. By row permutations and column permutations, $[M(I_{1},:);M(I_{2},:)]$ is transformed to the following form, denoted as $W$, $\begin{array}[]{l@{\hspace{-5pt}}llll}\begin{array}[]{@{\hspace{10pt}}l@{\hspace{4pt}}l@{\hspace{2pt}}l@{\hspace{18pt}}l}\hskip 10.0pt\lx@intercol\overbrace{\hphantom{\begin{array}[]{ccccccc}\ast&\ast&\ast&\ddots&\ddots&\ddots&\ast\end{array}}}^{\displaystyle s}\hfil\hskip 4.0&\overbrace{\hphantom{\begin{array}[]{lcccc}\ast&\ddots&\ddots&\ast&\ast\end{array}}}^{\displaystyle s-2}\hfil\hskip 2.0&\overbrace{\hphantom{\begin{array}[]{lcc}\ast&\ddots&\ast\end{array}}}^{\displaystyle s-4}\hfil\hskip 18.0&\overbrace{\hphantom{\begin{array}[]{lcc}\ast&\ast&\ast\end{array}}}^{\displaystyle 3}\par\end{array}&\\\ \left(\begin{array}[]{ccccccc|ccccc|ccc|c|ccc|c|cc}\ast&\ast&\ast&&&&&\ast&&&&&&&&&&&&&&\\\ &\ast&\ast&\ast&&&&&\ast&&&&&&&&&&&&&\\\ &&\ast&\ast&\ast&&&&&\ast&&&&&&&&&&&&\\\ &&&\ddots&\ddots&\ddots&&&&&\ddots&&&&&&&&&&&\\\ &&&&\ast&\ast&\ast&&&&&\ast&&&&&&&&&\\\ &&\ast&&&&&\ast&\ast&\ast&&&\ast&&&&&&&&&\\\ &&&\ddots&&&&&\ddots&\ddots&\ddots&&&\ddots&&&&&&&&\\\ &&&&\ast&&&&&&\ast&\ast&&&\ast&&&&&&&\\\ &&&&&&&&&\ddots&&&&\ddots&&\ddots&&&&&&\\\ &&&&&&&&&&&&&&&&\ast&&&&&\\\ &&&&&&&&&&&&&&&&&\ast&&&&\\\ &&&&&&&&&&&&&&&&&&\ast&&&\\\ &&&&&&&&&&&&&&&\cdots&\ast&\ast&\ast&\ast&&\\\ \hline\cr&&&&&&&&&&&&&&&&&&&\ast&\ast&\cdots\\\ &&&&&&&&&&&&&&&&&&&&\ast&\\\ &&&&&&&&&&&&&&&&&&&&&\ddots\end{array}\right)&\begin{array}[]{l}\vspace{35pt}\left.\vphantom{\begin{array}[]{c}\ast\\\ \ast\\\ \ast\\\ \ddots\\\ \ast\\\ \ast\\\ \ddots\\\ \ast\\\ \ddots\\\ \ast\\\ \ast\\\ \ast\\\ \ast\end{array}}\right\\}\frac{(s-1)^{2}}{4}\end{array}\end{array}.$ Note that the diagonal entries of the submatrix $W(1:\frac{(s-1)^{2}}{4},s+1:\frac{(s+1)^{2}}{4})$ are all nonzero. Therefore, by elementary matrix transformations on the first $\frac{(s+1)^{2}}{4}$ columns, the above matrix $W$ is transformed to the following, denoted as $\overline{W}$, $\begin{array}[]{l@{\hspace{-5pt}}llll}\begin{array}[]{@{\hspace{1pt}}l@{\hspace{2pt}}l@{\hspace{2pt}}l@{\hspace{15pt}}l}\hskip 1.0pt\lx@intercol\overbrace{\hphantom{\begin{array}[]{ccccccc}\ast&\ast&\ast&\ddots&\ddots&\ddots&\ast\end{array}}}^{\displaystyle s}\hfil\hskip 2.0&\overbrace{\hphantom{\begin{array}[]{lcccc}\ast&\ddots&\ddots&\ast&\ast\end{array}}}^{\displaystyle s-2}\hfil\hskip 2.0&\overbrace{\hphantom{\begin{array}[]{lcc}\ast&\ddots&\ast\end{array}}}^{\displaystyle s-4}\hfil\hskip 15.0&\overbrace{\hphantom{\begin{array}[]{lcc}\ast&\ast&\ast\end{array}}}^{\displaystyle 3}\par\end{array}&\\\ \left(\begin{array}[]{ccccccc|ccccc|ccc|c|ccc|c|cc}&&&&&&&\ast&&&&&&&&&&&&&&\\\ &&&&&&&&\ast&&&&&&&&&&&&&\\\ &&&&&&&&&\ast&&&&&&&&&&&&\\\ &&&&&&&&&&\ddots&&&&&&&&&&&\\\ &&&&&&&&&&&\ast&&&&&&&&&\\\ &&&&&&&&&&&&\ast&&&&&&&&&\\\ &&&&&&&&&&&&&\ddots&&&&&&&&\\\ &&&&&&&&&&&&&&\ast&&&&&&&\\\ &&&&&&&&&&&&&&&\ddots&&&&&&\\\ &&&&&&&&&&&&&&&&\ast&&&&&\\\ &&&&&&&&&&&&&&&&&\ast&&&&\\\ &&&&&&&&&&&&&&&&&&\ast&&&\\\ &&&&&&&&&&&&&&&&&&&\ast&&\\\ \hline\cr\times&\ast&\ast&\cdots&\ast&\ast&\ast&\ast&\ast&\cdots&\ast&\ast&\ast&\cdots&\ast&\cdots&\ast&\ast&\ast&\ast&\ast&\cdots\\\ &&&&&&&&&&&&&&&&&&&&\ast&\\\ &&&&&&&&&&&&&&&&&&&&&\ddots\end{array}\right)&\begin{array}[]{l}\vspace{35pt}\left.\vphantom{\begin{array}[]{c}\ast\\\ \ast\\\ \ast\\\ \ddots\\\ \ast\\\ \ast\\\ \ddots\\\ \ast\\\ \ddots\\\ \ast\\\ \ast\\\ \ast\\\ \ast\end{array}}\right\\}\frac{(s-1)^{2}}{4}\end{array}\end{array}.$ By elementary matrix transformations on the first $\frac{(s-1)^{2}}{4}+1$ rows, the above matrix $\overline{W}$ is further transformed to the following, denoted as $\overline{\overline{W}}$, $\begin{array}[]{l@{\hspace{-5pt}}llll}\begin{array}[]{@{\hspace{0.1pt}}l@{\hspace{1pt}}l@{\hspace{1pt}}l@{\hspace{9pt}}l}\hskip 0.1pt\lx@intercol\overbrace{\hphantom{\begin{array}[]{ccccccc}\ast&\ast&\ast&\ddots&\ddots&\ddots&\ast\end{array}}}^{\displaystyle s}\hfil\hskip 1.0&\overbrace{\hphantom{\begin{array}[]{lcccc}\ast&\ddots&\ddots&\ast&\ast\end{array}}}^{\displaystyle s-2}\hfil\hskip 1.0&\overbrace{\hphantom{\begin{array}[]{lcc}\ast&\ddots&\ast\end{array}}}^{\displaystyle s-4}\hfil\hskip 9.0&\overbrace{\hphantom{\begin{array}[]{lcc}\ast&\ast&\ast\end{array}}}^{\displaystyle 3}\par\end{array}&\\\ \left(\begin{array}[]{ccccccc|ccccc|ccc|c|ccc|c|cc}&&&&&&&\ast&&&&&&&&&&&&&&\\\ &&&&&&&&\ast&&&&&&&&&&&&&\\\ &&&&&&&&&\ast&&&&&&&&&&&&\\\ &&&&&&&&&&\ddots&&&&&&&&&&&\\\ &&&&&&&&&&&\ast&&&&&&&&&\\\ &&&&&&&&&&&&\ast&&&&&&&&&\\\ &&&&&&&&&&&&&\ddots&&&&&&&&\\\ &&&&&&&&&&&&&&\ast&&&&&&&\\\ &&&&&&&&&&&&&&&\ddots&&&&&&\\\ &&&&&&&&&&&&&&&&\ast&&&&&\\\ &&&&&&&&&&&&&&&&&\ast&&&&\\\ &&&&&&&&&&&&&&&&&&\ast&&&\\\ &&&&&&&&&&&&&&&&&&&\ast&&\\\ \hline\cr\times&\ast&\ast&\cdots&\ast&\ast&\ast&&&&&&&&&&&&&&\ast&\cdots\\\ &&&&&&&&&&&&&&&&&&&&\ast&\\\ &&&&&&&&&&&&&&&&&&&&&\ddots\end{array}\right)&\begin{array}[]{l}\vspace{35pt}\left.\vphantom{\begin{array}[]{c}\ast\\\ \ast\\\ \ast\\\ \ddots\\\ \ast\\\ \ast\\\ \ddots\\\ \ast\\\ \ddots\\\ \ast\\\ \ast\\\ \ast\\\ \ast\end{array}}\right\\}\frac{(s-1)^{2}}{4}\end{array}\end{array}.$ Note that in the row $\frac{(s-1)^{2}}{4}+1$, at least one element $\overline{\overline{W}}(\frac{(s-1)^{2}}{4}+1,1)$, denoted as ${}^{\prime}\times^{\prime}$, is not zero. Therefore, the lower submatrix $\overline{\overline{W}}(\frac{(s-1)^{2}}{4}+1:end,:)$ is still row full rank. Now Lemma 3.5 can be applied to conclude that $\overline{\overline{W}}$ is row full rank. Therefore $[M(I_{1},:);M(I_{2},:)]$ is row full rank. ###### Theorem 3.15 For the homotopy $H:\mathbb{R}^{mn}\times\mathbb{R}\times[0,1)\rightarrow\mathbb{R}^{mn}\times\mathbb{R}$ in (56), $\forall n\in\mathbb{N}^{+}$, $m\geq n\geq 6$, for almost all $K\in\mathbb{R}^{(2nm-m)}\setminus(U_{1}\cup U_{2})$, 1. (i) 0 is a regular value of $H$ defined in $(\ref{sec3:eqn-homotopy-3diagonal})$ and therefore the homotopy paths corresponding to different initial points do not intersect each other for $t\in[0,1)$; 2. (ii) Every homotopy path $(\varphi(s),\lambda(s),t(s))\subset H^{-1}(0)$ is bounded. Proof. (i). It suffices to prove that $\forall(\varphi,\lambda,t,K)\in\mathbb{R}^{mn}\times\mathbb{R}\times[0,1)\times\mathbb{R}^{(2nm-m)}\setminus(U_{1}\cup U_{2})$ satisfying $\tilde{H}(\varphi,\lambda,t,K)=0$, $((1-t)A+D-\lambda I)(R^{0},:)$ is row full rank, or in short hand notation $M(R^{0},:)$ is row full rank. Note that $R^{0}$ corresponds to the zero valued nodes $G^{0}$. If $G^{0}\bigcap g_{1}=\emptyset$, then $R^{0}\subset O_{1}$. By Lemma 3.11 $M(R^{0},:)$ is row full rank. Next $G^{0}\bigcap g_{1}\neq\emptyset$ is assumed. If the zero valued connected components of the set $G^{0}\bigcap g_{1}$ are all single point sets, $R^{0}$ is a subset of $O_{1}$, and again by Lemma 3.11, $M(R^{0},:)$ is row full rank. Assume that there are $q$ zero valued connected components of the set $G^{0}\bigcap g_{1}$, each of which has more than one point. These connected components of the set $G^{0}\bigcap g_{1}$ are denoted as $\gamma_{i}$, with corresponding flags $G_{i}=F_{\gamma_{i}}$, $i=1,\cdots,q$. Denote $\displaystyle\overline{G}=G_{a}\setminus(\bigcup_{i=1}^{q}G_{i}),$ (82) $\displaystyle\overline{R}=R_{\overline{G}},\quad\overline{C}=C_{\overline{G}},\quad R_{i}=R_{G_{i}},\quad C_{i}=C_{G_{i}},\quad\quad\forall 1\leq i\leq q.$ (83) By Lemma 3.12, it can be verified that $\displaystyle R^{0}=\overline{R}\bigcup\left(\bigcup_{i=1}^{q}R_{i}\right),$ (84) $\displaystyle R_{i}\bigcap\overline{R}=\emptyset,\quad\forall 1\leq i\leq q,$ (85) $\displaystyle R_{i}\bigcap R_{j}=\emptyset,\quad C_{i}\bigcap C_{j}=\emptyset,\quad\forall i\neq j.$ (86) Note that if a zero valued connected component $\gamma$ of $G^{0}\bigcap g_{1}$ has only two nodes, then the flag of $\gamma$ is the $\gamma$ itself, and $R_{\gamma}=\emptyset$. Assume that all the zero valued connected components $\gamma_{i}$ have more than two nodes. Since $\displaystyle\overline{R}\subset O_{1},\quad R_{i}\subset O_{m},\quad\forall 1\leq i\leq q,$ (87) by Lemma 3.11, all of $M(\overline{R},:)$ and $M(R_{i},:)$, $1\leq i\leq q$, are row full rank. It is possible that for some $\gamma_{i}$, $C_{i}\bigcap\overline{C}\neq\emptyset$. If so, $\gamma_{i}$ is the case described in case $(4)$ in Lemma 3.12, $C_{i}\bigcap\overline{C}$ has only one element and the corresponding grid points are illustrated in Figure 5. Figure 5: Local illustration of a connection$G_{i}$$\overline{G}$ Without loss of generality, suppose that all the $\gamma_{i}$ satisfying such property are the first $p$ $\gamma_{i}$, i.e., $\displaystyle C_{i}\bigcap\overline{C}\neq\emptyset,\quad 1\leq i\leq p,$ (88) $\displaystyle C_{i}\bigcap\overline{C}=\emptyset,\quad p+1\leq i\leq q.$ (89) Denote $\displaystyle G$ $\displaystyle=\overline{G}\bigcup\left(\bigcup_{i=p+1}^{q}G_{i}\right),$ (90) $\displaystyle Z(G_{k},\ldots,G_{1},G)$ $\displaystyle=[M(R_{k},:);\ldots;M(R_{1},:);M(R_{G},:)],\quad\forall 1\leq k\leq p.$ (91) Then $M(R^{0},:)=Z(G_{p},\ldots,G_{1},G)$. The claim that $M(R^{0},:)$ is row full rank will be proved by recursion. Firstly, prove $Z(G_{1},G)$ is row full rank. Take $I_{1}=R_{1}$ and $I_{2}=R_{G}$. The conditions of Lemma 3.14 are satisfied. Thus $[M(R_{1},:);M(R_{G},:)]$ is row full rank. Secondly, prove $Z(G_{k+1},\ldots,G_{1},G)$ is row full rank if $Z(G_{k},\ldots,G_{1},G)$ is, $\forall 1\leq k<p$. Take $I_{1}=R_{k+1}$ and $I_{2}=R_{k}\cup\cdots\cup R_{1}\cup R_{G}$. Since $C_{k+1}\cap C_{i}=\emptyset$, $1\leq i\leq k$, and $C_{k+1}\bigcap C_{G}$ has only one element, $I_{1}$ and $I_{2}$ satisfy the conditions of Lemma 3.14. Thus $[M(R_{k},:);\ldots;M(R_{1},:);M(R_{G},:)]$ is row full rank. (ii). Similar to the one dimensional case, that $H^{-1}(0)$ is bounded can be proved also. #### 3.2.2 The homotopy with random pentadiagonal matrix A homotopy with random pentadiagonal matrix is also possible. Specifically, for $\tilde{H}$ defined in (57), replacing $A(K)$ with $\overline{A}(\overline{K})$, where $\overline{A}(\overline{K})\in\mathbb{R}^{mn\times mn}$ is a random pentadiagonal matrix with the same sparse structure as $D$, namely, $\overline{A}(\overline{K})=\left(\begin{array}[]{cccc}A_{1}&\overline{A}_{1}&&\\\ \overline{A}_{1}&A_{2}&\overline{A}_{2}&\\\ &&\ddots&\overline{A}_{m-1}\\\ &&\overline{A}_{m-1}&A_{m}\end{array}\right),$ where $A_{i}=\left(\begin{array}[]{ccccc}a_{11}^{(i)}&a_{12}^{(i)}&&&\\\ a_{12}^{(i)}&a_{22}^{(i)}&a_{23}^{(i)}&&\\\ &a_{23}^{(i)}&a_{33}^{(i)}&\ddots&\\\ &&\ddots&\ddots&a_{n-1,n}^{(i)}\\\ &&&a_{n-1,n}^{(i)}&a_{nn}^{(i)}\end{array}\right),\quad\overline{A}_{i}=\left(\begin{array}[]{ccccc}b_{1}^{(i)}&&&&\\\ &b_{2}^{(i)}&&&\\\ &&b_{3}^{(i)}&&\\\ &&&\ddots&\\\ &&&&b_{n}^{(i)}\end{array}\right),$ $\displaystyle\overline{K}=\left(K_{1},K_{2},\ldots,K_{m},\overline{K}_{1},\ldots,\overline{K}_{m-1}\right)^{\mathrm{T}},$ $\displaystyle K_{i}=\left(a_{11}^{(i)},a_{12}^{(i)},a_{22}^{(i)},a_{23}^{(i)},\ldots,a_{n-1,n-1}^{(i)},a_{n-1,n}^{(i)},a_{nn}^{(i)}\right),~{}i=1,\ldots,m,$ $\displaystyle\overline{K}_{i}=\left(b_{1}^{(i)},b_{2}^{(i)},\ldots,b_{n}^{(i)}\right),~{}i=1,\ldots,m-1.$ The Jacobian matrix of $\tilde{H}(\varphi,\lambda,t,\overline{K})$, $\frac{\partial\tilde{H}}{\partial(\varphi,\lambda,t,\overline{K})}$, is : $\left(\begin{array}[]{cccc}(1-t)\overline{A}(\overline{K})+D+3t\beta\mbox{diag}(\varphi^{2})-\lambda I&-\varphi&\beta\varphi^{3}-\overline{A}(\overline{K})\varphi&(1-t)\overline{B}\\\ -\varphi^{\mathrm{T}}&0&0&0\end{array}\right),$ where $\overline{B}=\frac{\partial(\overline{A}(\overline{K})\varphi)}{\partial\overline{K}}\in\mathbb{R}^{{(mn)}\times{(3nm- m-n)}}$. Recall $\varphi_{i}=\left(\varphi_{i1},\varphi_{i2},\ldots,\varphi_{in}\right)^{\mathrm{T}}$. It can be verified that $\displaystyle\overline{B}$ $\displaystyle=\left(\begin{array}[]{ccccc|cccc}B_{1}&&&&&\overline{B}_{2}&&&\\\ &B_{2}&&&&\overline{B}_{1}&\overline{B}_{3}&&\\\ &&\ddots&&&&\overline{B}_{2}&\overline{B}_{4}&\\\ &&&\ddots&&&&\ddots&\ddots\\\ &&&&B_{m}&&&&\overline{B}_{m-1}\end{array}\right)$ (97) $\displaystyle=\left(\begin{array}[]{ccccc|cccc}&&&&&\overline{B}_{2}&&&\\\ &&&&&\overline{B}_{1}&\overline{B}_{3}&&\\\ &&B&&&&\overline{B}_{2}&\overline{B}_{4}&\\\ &&&&&&&\ddots&\ddots\\\ &&&&&&&&\overline{B}_{m-1}\end{array}\right),$ (103) with $\displaystyle\overline{B}_{i}=\left(\begin{array}[]{cccc}\varphi_{i1}&&&\\\ &\varphi_{i2}&&\\\ &&\ddots&\\\ &&&\varphi_{in}\end{array}\right).$ (108) Note that the left part $B$ of $\overline{B}$ is nothing but the matrix in (59). ###### Lemma 3.16 The eigenvalues of $\overline{A}(\overline{K})+D$ are simple for $\overline{K}$ almost everywhere except on a subset of real codimension 1. Proof. Similar to the proof of Lemma 3.4. For our discussion, $\overline{U}_{2}=(\mathbb{R}^{+})^{(3nm-m-n)}$ is removed. To prove that $\forall(\varphi,\lambda,t,\overline{K})\in\mathbb{R}^{mn}\times R\times[0,1)\times\mathbb{R}^{(3nm- m-n)}\setminus(\overline{U}_{1}\cup\overline{U}_{2})$ satisfying $\tilde{H}(\varphi,\lambda,t,\overline{K})=0$, the Jacobian matrix of $\tilde{H}(\varphi,\lambda,t,\overline{K})$ in (3.2.2) is row full rank, it suffices to prove that the following submatrix of the Jacobian matrix in (3.2.2) $\left(\begin{array}[]{cccc}(1-t)\overline{A}(\overline{K})+D+3t\beta\mbox{diag}(\varphi^{2})-\lambda I&-\varphi&\beta\varphi^{3}-\overline{A}(\overline{K})\varphi&(1-t)B\\\ -\varphi^{\mathrm{T}}&0&0&0\end{array}\right)$ (109) is row full rank. With the notations $R^{0}$ and $R^{*}$ as in (65) and similar arguments as in Subsection 3.2.1, it can be proved that $((1-t)\overline{A}+D-\lambda I))(R^{0},:)$ is row full rank. Therefore the matrix (109) is row full rank. That is $0$ is a regular value of the homotopy $H$ with random pentadiagonal matrix for almost all $\overline{K}\in\mathbb{R}^{(3nm-m-n)}\setminus(U_{1}\cup U_{2})$ with $m\geq n\geq 6$. ## 4 Algorithm and numerical results ### 4.1 Algorithm Thanks to Theorems 3.2 and 3.15, since 0 is a regular value of the homotopies constructed, the homotopy paths determined by the homotopy equations (49) and (56) have no bifurcation points with probability one. Therefore the usual path following algorithm, i.e., the predictor-corrector method as in [23, 24], can be adapted to trace the homotopy paths of the homotopy equations (49) and (56). The adapted algorithm is stated in Algorithm 1. For notation convenience, in Algorithm 1, we denote the iterate at step $k$ as $x_{k}=(\varphi^{(k)},\lambda_{k})$, $k=1,\ldots$. Initialization: Set $(x_{0},t_{0})=(x_{0},0)$, $k=0$, the minimum step size $ds_{min}$, the initial step size $ds$. Compute the tangent vector $(\dot{x}_{0},\dot{t}_{0})$ such that $\dot{t}_{0}>0$ and record the orientation $ori$ ${H_{x}\dot{x}_{0}+H_{t}\dot{t}_{0}=0},\qquad ori=\mbox{sign}\left(\left|\begin{array}[]{cc}H_{x}&H_{t}\\\ \dot{x}_{0}&\dot{t}_{0}\end{array}\right|\right).$ (110) while _$t_{k} <1$_ do Predictor: $(\bar{x}_{k+1},\bar{t}_{k+1})=(x_{k},t_{k})+ds(\bar{x}_{k},\bar{t}_{k})$; if _$\bar{t}_{k+1} >1$_ then change ds such that $\bar{t}_{k+1}=1$; end if Corrector: if _$\bar{t}_{k+1}=1$_ then $(v,\tau)=(0,1)$; else $(v,\tau)=(\dot{x}_{k},\dot{t}_{k})$; end if Employ Newton method to solve the following nonlinear equations: $\begin{array}[]{ccc}\left(\begin{array}[]{c}H(x,t)\\\ v^{T}(x-\bar{x}_{k+1})+\tau(t-\bar{t}_{k+1})\end{array}\right)&=&0.\end{array}$ (111) Judgement: if _the above iteration converges to $(x_{k+1},t_{k+1})$_ then compute the tangent vector $(\dot{x}_{k+1},\dot{t}_{k+1})$ satisfying ${H_{x}\dot{x}_{k+1}+H_{t}\dot{t}_{k+1}=0},\qquad\mbox{sign}\left(\left|\begin{array}[]{cc}H_{x}&H_{t}\\\ \dot{x}_{k+1}&\dot{t}_{k+1}\end{array}\right|\right)=ori$ (112) else $ds=\frac{1}{2}ds$; go to Predictor; end if Compute angle $\theta$ of $(\bar{x}_{k},\bar{t}_{k})$ and $(\bar{x}_{k+1},\bar{t}_{k+1})$; if _$\theta >18^{0}$_ then $ds=\frac{1}{2}ds,$ go to Predictor; end if Accept iterates: $(x_{k},t_{k})=(x_{k+1},t_{k+1})$,$(\dot{x}_{k},\dot{t}_{k})=(\dot{x}_{k+1},\dot{t}_{k+1})$; if _$\theta <6^{0}$_ then $ds=2ds$; end if if _$ds <ds_{min}$_ then stop the algorithm; end if end while Algorithm 1 Predictor-Corrector method ### 4.2 Numerical results The first numerical example is the 1D discretized problem (9) with $\beta=20$, $V(x)=\frac{1}{2}x^{2}$, $\Omega=[-2,2]$ and $n=999$. Some eigenvectors, i.e., approximate eigenfunctions are plotted in Figures 6(a) to 6(i). The eigenvector in Figure 6(a) is the unique positive solution of (9) as stated in Theorem 2.5, which corresponds to the unique positive ground state, as proved in [4] for the continuous nonlinear eigenvalue problem (4)-(6) when $\beta>0$. The approximate eigenfunction in Figure 6(b) is antisymmetric as described in Theorem 2.6 and is an approximate first excited state. Others are approximate excited states corresponding to higher energy. The order preserving property of the eigenvalue curves as stated in [17] was observed, that is, if $\lambda(0)$ is the $k$th smallest eigenvalue of the initial problem, then $\lambda(t)$ is the $k$th smallest eigenvalue of the intermediate problem for each $t\in[0,1)$. However, we are not able to prove such property for the eigenvector-dependent nonlinear eigen-problem yet. (a) $\lambda=6.76$ (b) $\lambda=8.39$ (c) $\lambda=10.35$ (d) $\lambda=12.73$ (e) $\lambda=15.62$ (f) $\lambda=19.08$ (g) $\lambda=23.13$ (h) $\lambda=27.79$ (i) $\lambda=33.05$ The second numerical example is the 2D discretized problem $(\ref{sec 2: finite difference - two dimen})$ with $\beta=20$, $V(x)=\frac{1}{2}(x_{1}^{2}+x_{2}^{2})$, $\Omega=[0,1]\times[0,1]$ and $m=n=29$. Some approximate eigenfunctions are collected in Figures 6(j) to 6(u). The approximate eigenfunction in Figure 6(j) corresponds to the unique positive ground state. Others are approximate excited states corresponding to higher energy. The order preserving property of the eigenvalue curves is also observed for the 2D case. (j) $\lambda=43.36$ (k) $\lambda=60.89$ (l) $\lambda=65.31$ (m) $\lambda=78.81$ (n) $\lambda=86.28$ (o) $\lambda=94.06$ (p) $\lambda=104.4$ (q) $\lambda=108.82$ (r) $\lambda=120.47$ (s) $\lambda=129.67$ (t) $\lambda=133.81$ (u) $\lambda=138.68$ ## 5 Conclusion Solutions to the discretized problem with the finite difference disretization for the GPE inherit certain properties of the solutions to the continuous problem, such as the existence and uniqueness of positive eigenvector (eigenfunction). The designed homotopy continuation methods are suitable for computing eigenpairs corresponding to excited states of high energy as well as the ground state and the first excited state. In order to make sure that the homotopy paths are regular and that the path following is efficient, artificial homotopy parameter and random matrices with certain structures in the homotopies seem indispensable. ## References * [1] W. Bao and W. Tang, Ground state solution of Bose-Einstein condensate by directly minimizing the energy functional, J. Comput. Phys., 187 (2003), 230–254. * [2] S.-L. Chang, H.-S. Chen, B.-W. Jeng, and C.-S. Chien, A spectral-Galerkin continuation method for numerical solutions of the Gross-Pitaevskii equation, J. Comput. Appl. Math., 254 (2013), 2–16. * [3] W. Bao and Q. Du, Computing the ground state solution of Bose-Einstein condensates by a normalized gradient flow, SIAM J. Sci. Comput., 25(5) (2003), 1674–1697. * [4] W. Bao and Y. Cai, Mathematical theory and numerical methods for Bose-Einstein condensation, J. Kinetic & Related Models., 6 (2013), 1–135. * [5] M. Edwards and K. Burnett, Numerical solution of the nonlinear schrödinger equation for small samples of trapped neutral atoms, Phys. Rev. A., 51 (1995), 1382–1386. * [6] S. K. Adhikari, Numerical solution of the two-dimensional Gross-Pitaevskii equation for trapped interacting atoms, Phys. Lett. A., 265 (2000), 91–96. * [7] S.-L. Chang and C.-S. Chien, Numerical continuation for nonlinear schrödinger equations, Int. J. Bifurcat. Chaos., 17 (2007), 641–656. * [8] S.-L. Chang, C.-S. Chien, and B.-W. Jeng, Computing wave functions of nonlinear Schrödinger equations: a time-independent approach, J. Comput. Phys., 226 (2007), 104–130. * [9] Y. S. Choi, I. Koltracht, P. J. McKenna, and N. Savytska, Global monotone convergence of Newton iteration for a nonlinear eigen-problem, Linear Algebra Appl., 357 (2002), 217–228. * [10] Y. S. Choi, I. Koltracht, and P. J. McKenna, A generalizaion of the Perron-Frobenius theorem for nonlinear perturbations of stiltjes matrices, Contemp. Math. AMS., 281 (2001), 325–330. * [11] Y. Cai, L.-H. Zhang, Z. Bai, and R.-C. Li, On an eigenvector-dependent nonlinear eigenvalue problem, SIAM J. Matrix Anal. Appl., 39 (2018), 1360–1382. * [12] Y. Saad, J. R. Chelikowsky, and S. M. Shontz, Numerical methods for electronic structure calculations of materials, SIAM Rev., 52 (2010), 3–54. * [13] E. Jarlebring, S. Kvaal, and W.Michiels, An inverse iteration method for eigenvalue problems with eigenvector nonlinearities, SIAM J. Sci. Comput., 36 (2014), 1978–2001. * [14] X. D. Yao and J. X. Zhou, Numerical methods for computing nonlinear eigenpairs: part I. Iso-homogeneous cases, SIAM J. Sci. Comput., 29 (2007), 135–1374. * [15] Q. Z. Yang, P. F. Huang and Y. J. Liu, Numerical examples for solving a class of nonlinear eigenvalue problems, Journal on Numerical Methods and Computer Applications, 40 (2019), 130–142. * [16] H. H. Xie, A multigrid method for nonlinear eigenvalue problems (in Chinese), Sci. Sin. Math., 45 (2015), 1193–1204. * [17] T. Y. Li and N. H. Rhee, Homotopy algorithm for symmetric eigenvalue problems, Numer. Math., 55 (1989), 265–280. * [18] T. Y. Li, H. Zhang, and X. H. Sun, Parallel homotopy algorithm for the symmetric tridiagonal eigenvalue problems, SIAM J. Sci. Stat. Comput., 12 (2006), 469–487. * [19] T. Y. Li, Z. Zeng, and L. Cong, Solving eigenvalue problems of real nonsymmetric matrices with real homotopies, SIAM J. Numer. Anal., 29 (1992), 229–248. * [20] S. H. Lui, H. B. Keller, and T. W. Kwok, Homotopy method for the large, sparse, real nonsymmetric eigenvalue problem, SIAM J. Matrix Anal. Appl. 18 (1997), 312–333. * [21] T. Y. Li And T. Sauer, Homotopy method for generalized eigenvalue problems Ax=$\lambda$Bx, Linear Algebra. Appl., 91 (1987), 65–74. * [22] M. T. Chu, T. Y. Li, and T. Sauer, Homotopy method for general $\lambda$-matrix problems, SIAM J. Matrix Anal. Appl., 9 (1988), 528–536. * [23] E. L. Allgower and K. Georg, Introduction to Numerical Continuation Methods, Classics in Applied Mathematics, 45, SIAM, Philadelphia, 2003. * [24] X. D. Huang, Z. G. Zeng and Y. N. Ma, Theories and Methods of Nonlinear Numerical Analysis (in Chinese), Wuhan University Press, Wuhan, 2004.
# Spin dynamics and unconventional Coulomb phase in Nd2Zr2O7 M. Léger Institut Néel, CNRS and Université Grenoble Alpes, 38000 Grenoble, France Laboratoire Léon Brillouin, Université Paris-Saclay, CNRS, CEA, CE- Saclay, F-91191 Gif-sur-Yvette, France E. Lhotel<EMAIL_ADDRESS>Institut Néel, CNRS and Université Grenoble Alpes, 38000 Grenoble, France M. Ciomaga Hatnean Department of Physics, University of Warwick, Coventry, CV4 7AL, United Kingdom J. Ollivier Institut Laue Langevin, F-38042 Grenoble, France A. R. Wildes Institut Laue Langevin, F-38042 Grenoble, France S. Raymond Université Grenoble Alpes, CEA, IRIG, MEM, MDN, 38000 Grenoble, France E. Ressouche Université Grenoble Alpes, CEA, IRIG, MEM, MDN, 38000 Grenoble, France G. Balakrishnan Department of Physics, University of Warwick, Coventry, CV4 7AL, United Kingdom S. Petit<EMAIL_ADDRESS>Laboratoire Léon Brillouin, Université Paris-Saclay, CNRS, CEA, CE-Saclay, F-91191 Gif-sur-Yvette, France ###### Abstract We investigate the temperature dependence of the spin dynamics in the pyrochlore magnet Nd2Zr2O7 by neutron scattering experiments. At low temperature, this material undergoes a transition towards an “all in - all out” antiferromagnetic phase and the spin dynamics encompass a dispersion-less mode, characterized by a dynamical spin ice structure factor. Unexpectedly, this mode is found to survive above $T_{\rm N}\approx 300$ mK. Concomitantly, elastic correlations of the spin ice type develop. These are the signatures of a peculiar correlated paramagnetic phase which can be considered as a new example of Coulomb phase. Our observations near $T_{\rm N}$ do not reproduce the signatures expected for a Higgs transition, but show reminiscent features of the “all in - all out” order superimposed on a Coulomb phase. Geometrical frustration is well known to be one of the key ingredients leading to unconventional states of matter, especially in magnetism [1, 2]. Among them, spin ice and more generally Coulomb phases [3] have attracted significant interest. These can be considered as an original state of matter formed by disordered degenerate configurations where local degrees of freedom remain strongly constrained at the local scale by an organizing principle. In the case of spin ice, these degrees of freedom are Ising spins, sitting on the sites of a pyrochlore lattice formed of corner sharing tetrahedra and aligned along the axes which connect the corners of the tetrahedra to their center. The organizing principle, the “ice rule”, states that each tetrahedron should have two spins pointing in and two out, in close analogy with the rule which controls the hydrogen position in water ice [4]. Importantly, the idea that this local constraint can be considered as the conservation law of an “emergent” magnetic flux (${\bf\nabla}\cdot{\bf B}=0$) was quickly imposed [5, 6, 7]. Quantum fluctuations can cause this flux to change with time, giving rise to an emergent electric field, and eventually to an emergent quantum electromagnetism [8, 9, 10]. This quantum spin ice state hosts spinon (monopole in the spin ice language [11]) and photon like excitations. Despite much work, however, experimental evidence for this enigmatic physics remains elusive, with the possible exception of Pr2Hf2O7 [12]. Indeed, the conditions for the realisation of this so-called quantum spin ice state are drastic: transverse terms have to be sizable in the Hamiltonian to enable fluctuations out of the local Ising axes, but should remain small enough to prevent the stabilization of classical phases, called Higgs phases, characterized by ordered components perpendicular to these axes [13, 11, 14]. The pyrochlore material Nd2Zr2O7 offers the opportunity to approach this issue. Recent studies suggest that below 1 K this compound hosts a correlated state, which could be a remarkable novel example of Coulomb phase [15, 16]. This phase would be described by a “two in – two out” rule as in spin ice, but built on a pseudospin component different from the conventional $\langle 111\rangle$ Ising one. The “all in – all out” (AIAO) ordering previously observed below $T_{\rm N}\approx 300$ mK [17, 18] would then correspond to the pseudospin ordering in directions perpendicular to the components responsible for the “high temperature” Coulomb phase. It was proposed that a Higgs mechanism may account for this transition [16]. Such a process is invoked in $U(1)$ quantum spin liquids when the deconfined spinon excitations undergo a Bose-Einstein condensation, resulting in a Higgs phase along with a gapped photon excitation [19, 13, 20]. In this letter, we show that the paramagnetic phase of Nd2Zr2O7 does carry elastic spin ice-like correlations, and thus confirm the proposed Coulomb phase picture above $T_{\rm N}$. We present a detailed study of the spin dynamics as a function of temperature and explore the nature of this Coulomb phase above and close to the transition. The spin excitations of Nd2Zr2O7 deep in the AIAO phase include a peculiar spectrum with a flat band at the energy $E_{0}\approx$ 70 $\mu$eV characterized by a spin ice-like ${\bf Q}$-dependence [15, 21, 22]. Using neutron scattering experiments, we report the temperature dependence of the gap $E_{0}$, and reveal that this gap persists above $T_{\rm N}$. This result is robust, and withstands a small substitution at the Zr site. The spectra recorded above $T_{\rm N}$ do not show the spinon continuum expected in the Higgs scenario. Instead, we observe dispersive features reminiscent of the AIAO ordered phase superimposed on the Coulomb phase signal. This coexistence suggests that a strong exchange competition is at work in this temperature range, emphasizing the originality of the Coulomb phase above the transition. The single crystal samples used in this work are the same as in our previous studies (labeled #1 [17, 15, 21] and #2 [21]). In addition, results on a single crystal of Nd2(Zr1-xTix)2O7, with $x=2.4$ % (Sample #3) (See supplementary material [23]) are presented, not in order to analyze the role of disorder but to illustrate the robustness of the results. Magnetic properties were measured in very low temperature SQUID magnetometers developed at the Institut Néel [29]. The composition and magnetic structure at low temperature were determined using the D23 (CEA-CRG@ILL) neutron diffractometer [23]. Polarized neutron scattering experiments were carried out at D7 (ILL) on Sample #1. Inelastic neutron scattering (INS) experiments were carried out on the IN5 (ILL) time of flight spectrometer on all samples and on the triple axis spectrometer IN12 (CEA-CRG@ILL) for Sample #1. The INS data have been analyzed using the cefwave software developed at LLB. The XYZ Hamiltonian proposed to describe the properties of Nd based pyrochlores due to the peculiar dipolar-octupolar character of the Nd3+ion [30], writes: ${\cal H}=\sum_{\langle i,j\rangle}\left[{\sf J}_{x}\tau^{x}_{i}\tau^{x}_{j}+{\sf J}_{y}\tau^{y}_{i}\tau^{y}_{j}+{\sf J}_{z}\tau^{z}_{i}\tau^{z}_{j}+{\sf J}_{xz}(\tau^{x}_{i}\tau^{z}_{j}+\tau^{z}_{i}\tau^{x}_{j})\right]$ (1) In this Hamiltonian, $\tau_{i}$ is not the actual spin, but a pseudospin which resides on the rare-earth sites of the pyrochlore lattice. Its $z$ component relates to the usual magnetic moment and is directed along the local $\langle 111\rangle$ directions of the tetrahedra of the pyrochlore lattice. This Hamiltonian can be rewritten by rotating the ${\bf x}$ and ${\bf z}$ axes in the $({\bf x},{\bf z})$ plane by an angle $\theta$. In this $({\bf\tilde{x}},{\bf\tilde{z}})$ rotated frame, the relevant parameters of the Hamiltonian ${\cal H}$ are labeled $\tilde{{\sf J}}_{x,y,z}$, leading to [30, 31]: $\displaystyle{\cal H}_{\rm XYZ}=$ $\displaystyle\sum_{\langle i,j\rangle}\left[{\tilde{\sf J}_{x}}\tilde{\tau}^{\tilde{x}}_{i}\tilde{\tau}^{\tilde{x}}_{j}+{\tilde{\sf J}_{y}}\tilde{\tau}^{\tilde{y}}_{i}\tilde{\tau}^{\tilde{y}}_{j}+{\tilde{\sf J}_{z}}\tilde{\tau}^{\tilde{z}}_{i}\tilde{\tau}^{\tilde{z}}_{j}\right]$ (2) $\displaystyle{\rm with}\quad\tan(2\theta)=\frac{2{\sf J}_{xz}}{{\sf J}_{x}-{\sf J}_{z}}$ With time and maturation of the subject, the estimated parameters for Nd2Zr2O7 have evolved. Determinations of the $\tilde{\sf J}_{i}$ parameters are based on the spin wave spectra measured at very low temperature in zero field [15, 31, 22] or applied field [21], while the angle $\theta$ is deduced from the Curie-Weiss temperature [31] and/or the ordered AIAO magnetic moment [21, 22]. The sets of reported parameters are summarized in Table 1, where we have added the parameters refined here for the Nd2(Zr1-xTix)2O7 sample (Sample #3) [23] and have revisited the ones of Samples #1 and #2. From these values, two interesting features stand out, which remain unexplained to date and should be further explored to ascertain their relevance: (i) the larger the Néel temperature, the larger the ordered moment along z is. (ii) very similar $\tilde{\sf J}_{i}$ parameters are obtained for the various samples, despite differences with regard to the amount of impurities or to the ordering parameters. The ${\sf J}$ parameters lead to an ordered AIAO ground state, where the pseudospins point along the (local) direction ${\bf\tilde{z}}$, turned around the ${\bf z}$-axis towards the ${\bf x}$-axis by the angle $\theta$ [31]. As shown by INS experiments, peculiar excitations are associated with this ground state. They manifest as an inelastic spin ice like flat mode at an energy $E_{0}\approx$ 70 $\mu$eV, above which spin wave branches disperse (See Figure 4a for Sample #1) [15]. This excitation spectrum is understood in the framework of the dynamic fragmentation [31, 32] as the sum of a dynamic divergence-free contribution, giving rise to the flat mode at $E_{0}$ and of a dynamic curl-free contribution, which takes the form of the dispersing branches. These spin waves correspond to the propagation of magnetically charged excitations and have a spectral weight made of half-moons in reciprocal space [15, 33]. Sample / Ref. | $m_{\rm ord}~{}(\mu_{\rm B})$ | $T_{\rm N}$ (mK) | Hamiltonian parameters (K) | $\theta$ (rad) ---|---|---|---|--- | | | $\tilde{{\sf J}}_{x}$ | $\tilde{{\sf J}}_{y}$ | $\tilde{{\sf J}}_{z}$ | $\tilde{{\sf J}}_{x}/|\tilde{{\sf J}}_{z}|$ | #1 | $0.8\pm 0.05$ | 285 | 1.18 | -0.03 | -0.53 | 2.20 | 1.23 #2 | $1.1\pm 0.1$ | 340 | 1.0 | 0.066 | -0.5 | 2.0 | 1.09 #3 | $1.19\pm 0.03$ | 375 | 0.97 | 0.21 | -0.53 | 1.83 | 1.08 [22] | 1.26 | 400 | 1.05 | 0.16 | -0.53 | 1.98 | 0.98 [31] | 1.4 | - | 1.2 | 0.0 | -0.55 | 2.18 | 0.83 Table 1: Ordered moment $m_{\rm ord}$ along ${\bf z}$, transition temperature $T_{\rm N}$ and Hamiltonian parametrization reported in different studies. $\tilde{{\sf J}}_{i}$ parameters for Sample #1 and from Ref. 31 were obtained from fits of the INS data reported in Ref. 15 and, for Sample #2 in Ref. 21. $m_{\rm ord}$ from Ref. 31 is a calculated value. The total Nd3+ magnetic moment is estimated to $\approx 2.4~{}\mu_{\rm B}$ [17, 18]. Figure 1: (a-b) Magnetic instantaneous correlations in Sample #1 as a function of temperature. The 10 K dataset has been subtracted as a background reference. Measurements in (a) were symmetrized. (c) “Spin ice” moment $m_{1}$ and AIAO ordered moment $m_{2}$ along ${\bf z}$ as a function of temperature [23]. Lines are guides to the eye. Instantaneous spin-spin correlations $S({\bf Q})$ were measured in Sample #1 as a function of temperature between 60 mK and 1 K through polarized neutron scattering experiments and are displayed in Figure 1 [23]. These measurements integrate over the neutron energy loss up to 3.5 meV, and thus contain both elastic and inelastic signals. At 1 K, a spin ice pattern can barely be observed, revealing the onset of a Coulomb phase. Upon cooling, the spin ice pattern becomes clearly visible below 600 mK. At 450 mK, the magnetic moment $m_{1}$ responsible for the spin ice-like diffuse scattering is estimated to $2.05\pm 0.3~{}\mu_{\rm B}$ [23], to be compared to the 2.4 $\mu_{\rm B}$ full Nd moment [17, 18]. In addition to this signal, below 800 mK, magnetic diffuse scattering spots appear around $(220)$, $(113)$ and symmetry related positions. Intensity on these positions increases with cooling until they transform into Bragg peaks below $T_{\rm N}$ (285 mK in this sample) characteristic of the AIAO phase. At low temperature, the corresponding ordered magnetic moment is $m_{2}=m_{\rm ord}=0.8\pm 0.05~{}\mu_{\rm B}$ (from diffraction measurements) and the magnetic contribution to the spin ice like diffuse scattering amounts to $m_{1}=2\pm 0.3~{}\mu_{\rm B}$ [23] (see Figure 1c). The moment embedded in the spin ice correlations is thus at maximum around $T_{\rm N}$ and slightly decreases at lower temperature. The diffuse scattering observed in the vicinity of the Bragg peak positions above $T_{\rm N}$ might arise from AIAO diffuse scattering just above the ordering transition, but could also be a signature of deconfined excitations, as proposed in Ref. 16. Figure 2: Spectral function $S(E)$ at different temperatures [23] measured at a wavelength $\lambda=$ 8.5 Å, hence an energy resolution of 20 $\mu$eV: (a) in Sample #1, integrated around ${\bf Q}=(0.8~{}0.8~{}0.8)$. The grey and red lines correspond to the fitted incoherent elastic $I_{c}$ and inelastic $S_{0}$ contributions respectively. (b) and (c): integrated over the measured ${\bf Q}$ range in Samples #2 (b) and #3 (c). To determine the spectral profile contained in those magnetic correlations, and especially the elastic or inelastic nature of the spin ice correlations associated to $m_{1}$, INS measurements have been carried out on the three aforementioned samples (see Table 1) as a function of temperature. To highlight the possible presence of an inelastic flat mode, the ${\bf Q}$-integrated spectral function $S(E)=\int d{\bf Q}S({\bf Q},E)$ was computed. As this quantity is akin to a density of states, it enhances the contribution of the flat modes contained in the spectrum. Figure 2 displays $S(E)$ at different temperatures. As previously shown [15], the inelastic flat band is clearly seen at low temperature. It is still visible at finite energy close to $T_{\rm N}$ (320 mK for Sample #2 and 315 mK for Sample #3) and above $T_{\rm N}$ (340 mK for Sample #1), yet broadens significantly upon warming. At the highest temperatures, the signal looks almost quasielastic. To obtain a quantitative insight into the temperature evolution of the mode, data were fitted for the three samples (as shown in Figure 2a for Sample #1) to the following model [23]: $S(E)=b+I_{c}(E)+F(E,T)\times\left[S_{0}(E)+S_{1}(E)\right]$ (3) $b$ is a flat background, $I_{c}(E)$ is a Gaussian function centered at zero energy to account for the elastic incoherent scattering. $F(E,T)=1+n(E,T)$ is the detailed balance factor ($n$ is the Bose-Einstein distribution). $S_{0}(E)$ and $S_{1}(E)$ are two Lorentzian profiles, centered on the energy $E_{0,1}$ and of intensity $I_{0,1}$, which represent respectively the flat band and the dispersive mode typical of the spin wave spectrum in Nd2Zr2O7. The determined positions $E_{0}$ and intensities $I_{0}$ are shown in Figure 3 as a function of the temperature normalized to $T_{\rm N}$ for the three samples. As anticipated from Figure 2, with increasing temperature, the band at $E_{0}$ softens and broadens while its intensity decreases. Nevertheless, $E_{0}$ is non-zero at $T_{\rm N}$ and a persistent dynamical behaviour is observed in all samples at and above $T_{\rm N}$, up to about $2T_{\rm N}$. Finally, the width of the features above the flat mode makes it hard to extract quantitative information from $S_{1}$. However, close examination of $S({\bf Q},E)$ measured for Sample #1 above $T_{\rm N}$ at 340 mK (see Figure 4) shows that, in all investigated directions, besides a strong quasielastic contribution (the inelastic mode being hardly discernible due to the energy resolution and the color scale), weak features are present close to the position of the low temperature dispersions. These spin wave fingerprints, highlighted by arrows on Figure 4(b) and which manifest as a broad signal in ${\bf Q}$-cuts (Figure 4(c-d)), are not compatible with the excitation spectrum expected in the presence of monopole creation and hopping [16]. Several striking features emerge from these measurements. INS experiments reveal that the intensity $I_{0}$ of the inelastic spin ice mode decreases when increasing temperature. Since D7 polarized experiments show that the full spin ice correlations, elastic and inelastic, are strongest around $T_{\rm N}$, the spin ice pattern observed above $T_{\rm N}$ must contain a new spin ice contribution, likely elastic, and different from the inelastic mode at $E_{0}$. This is confirmed by magnetization measurements, which point to ferromagnetic-like correlations, as expected for spin ice [23]. This elastic signal could not be directly identified in the elastic line of the IN5 data [23] certainly due to background issues, but we should stress that the D7 polarization analysis is definitely the most appropriate way to remove properly nuclear contributions and visualize small magnetic contributions. These results thus point to the coexistence of two spin ice-like contributions, an elastic and an inelastic one with different origins, and different temperature dependences. Figure 3: (a) $E_{0}$ and (b) $I_{0}$ as a function of reduced temperature $T/T_{\rm N}$ obtained from measurements on IN5 (dots - see Figure 2) and IN12 (triangles), together with results from Monte-Carlo (MC) calculations from Ref. 16 (red dots) [23]. Lines are guides to the eye. The large $I_{0}$ experimental value when $E_{0}=0$ is the signature of the persistent quasielastic contribution above $T_{\rm N}$. These two contributions can be understood as the manifestation of the strong competition at play between the pseudospin components of Nd. The negative value of $\tilde{\sf J}_{z}$ (see Table 1) promotes an AIAO phase built on ${\tilde{\tau}}^{\tilde{z}}$ while the positive $\tilde{\sf J}_{x}$ favors a Coulomb phase, similar to a spin ice phase, but built on ${\tilde{\tau}}^{\tilde{x}}$. For $\tilde{\sf J}_{x}/|\tilde{\sf J}_{z}|\approx 2$, the value determined for Nd2Zr2O7, the former is stabilized at low temperature and the latter at finite temperature, due to the large entropy associated to the Coulomb phase. In these two regimes, spin ice contributions are expected, an elastic one in the Coulomb phase at “high” temperature, and an inelastic one in the AIAO ordered phase (accompanied by dispersive excitations). Remarkably, the observable $\tau_{z}$, which corresponds to the magnetic dipolar moment along the local $\langle 111\rangle$ axes, is a combination of the ${\tilde{\tau}}^{\tilde{x}}$ and ${\tilde{\tau}}^{\tilde{z}}$ components of the pseudospin. It thus holds the two competing contributions (AIAO and Coulomb), which contrasts with the conventional spin ice case where the $z$ component carries elastic spin ice correlations only. Figure 4: INS spectra of Sample #1 along several high symmetry directions at 60 mK (a) and 340 mK (b), measured on IN5 with $\lambda$=6 Å. Red arrows highlight the dispersive modes and their fingerprints above $T_{\rm N}$. (c-d) Constant ${\bf Q}$-cuts at these two temperatures, integrated (c) along $(hh0)$ and (d) along $(hh2)$. The present results shed light on the manner in which the system evolves from the “high” temperature Coulomb phase to the low temperature AIAO ordered phase. At high temperature, around 1 K, the elastic spin ice signal characteristic of the ${\tilde{\tau}}^{\tilde{x}}$ Coulomb phase appears first. Upon cooling, the inelastic spin ice contribution along with dispersive spin wave branches emerge above $T_{\rm N}$ and coexist with the elastic one. They can naturally be considered as excitations stemming from the short-range AIAO correlations of the ${\tilde{\tau}}^{\tilde{z}}$ component observed below 800 mK (see Figure 1). The system enters the long-range AIAO ordered state at a temperature $T_{\rm N}\approx 300$ mK. It corresponds to about $|\tilde{\sf J}_{x}|/4$, thus to a temperature scale far above the one obtained theoretically for the stabilization of the quantum regime of spin ice, which is estimated to a few percents of the characteristic exchange interaction [34, 35]. This indicates that the Coulomb phase remains in its thermal regime down to $T_{\rm N}$. Surprisingly, the ordering temperature is larger than semi-classical Monte- Carlo calculations predictions [16]. At $T_{\rm N}$, the excitation spectrum is gapped, with the coexistence between the elastic spin ice component and the inelastic spectrum typical of AIAO ordering. The lack of a spinon continuum which would condense at $T_{\rm N}$ seems to preclude a transition driven by a Higgs mechanism. Deeper in the AIAO phase, the inelastic component - together with the Bragg peaks - develops at the expense of the elastic component. The weak maximum of the spin ice $m_{1}$ moment around $T_{\rm N}$ can thus be interpreted as due to the rise of the inelastic spin ice mode along with the persistence of the elastic contribution of the Coulomb phase. The coexistence of the elastic and inelastic signals is consistent with MC calculations [16], even if, close to $T_{\rm N}$, the two modes are less distinguishable in the experiments than in the calculations due to the strong broadening of the inelastic mode. Although some distribution is observed between the samples, the measured temperature dependence of the inelastic spin ice mode, described by the energy $E_{0}(T)$ and intensity $I_{0}(T)$, is also consistent with calculations [23], despite a slightly stronger inelastic component in experiments above $T_{\rm N}$ (see Figure 3). In summary, we find that with increasing temperature, the now well-established flat spin ice band characteristic of the AIAO ground state in Nd2Zr2O7, softens while its intensity decreases. The energy of this mode remains however finite at and above $T_{\rm N}$ and becomes overdamped with increasing the temperature further. At the same time, a new elastic spin ice component appears. The nature of the correlated phase above $T_{\rm N}$ is thus highly unconventional with the coexistence of an (elastic) Coulomb phase and fragmented excitations, resulting from the competition between the different terms of the Hamiltonian. Our observations support a picture where the AIAO ordering arises from a thermal spin ice phase, a scenario which is well accounted for by semi-classical MC calculations from Ref. 16, and is different from the proposed Higgs transition. When increasing the ratio $\tilde{\sf J}_{x}/|\tilde{\sf J}_{z}|$, reentrant behaviors are predicted [16] while the system approaches a quantum spin liquid ground state [31]. Tuning the parameters of the Hamiltonian (2) with novel materials would thus be of high interest to understand the unusual behavior of Nd2Zr2O7 and explore the frontiers between thermal and quantum regimes. ###### Acknowledgements. The work at the University of Warwick was supported by EPSRC, UK through Grant EP/T005963/1. M. L. and S.P. acknowledge financial support from the French Federation of Neutron Scattering (2FDN). M. L. acknowledges financial support from Université Grenoble-Alpes (UGA). M.L., E.L. and S.P. acknowledge financial support from ANR, France, Grant No. ANR-19-CE30-0040-02. S.P. and E.L. acknowledge F. Damay for helpful remarks and J. Xu for providing the data of his calculations. E.L. acknowledges C. Paulsen for the use of his magnetometers. ## References * Lacroix _et al._ [2011] C. Lacroix, P. Mendels, and F. Mila, eds., _Introduction to Frustrated Magnetism_ (Springer-Verlag, Berlin, 2011). * Gardner _et al._ [2010] J. S. Gardner, M. J. P. Gingras, and J. E. Greedan, Magnetic pyrochlore oxides, Rev. Mod. Phys. 82, 53 (2010). * Henley [2010] C. L. Henley, The “Coulomb phase” in frustrated systems, Annu. Rev. Condens. Matter Phys. 1, 179 (2010). * Harris _et al._ [1997] M. J. Harris, S. T. Bramwell, D. F. McMorrow, T. Zeiske, and K. W. Godfrey, Geometrical frustration in the ferromagnetic pyrochlore Ho2Ti2O7, Phys. Rev. Lett. 79, 2554 (1997). * Isakov _et al._ [2004] S. V. Isakov, K. Gregor, R. Moessner, and S. L. Sondhi, Dipolar spin correlations in classical pyrochlore magnets, Phys. Rev. Lett. 93, 167204 (2004). * Henley [2005] C. L. Henley, Power-law spin correlations in pyrochlore antiferromagnets, Phys. Rev. B 71, 014424 (2005). * Castelnovo _et al._ [2008] C. Castelnovo, R. Moessner, and S. L. Sondhi, Magnetic monopoles in spin ice, Nature 451, 42 (2008). * Hermele _et al._ [2004] M. Hermele, M. P. A. Fisher, and L. Balents, Pyrochlore photons: The $U(1)$ spin liquid in a $S=1/2$ three-dimensional frustrated magnet, Phys. Rev. B 69, 064404 (2004). * Shannon _et al._ [2012] N. Shannon, O. Sikora, F. Pollmann, K. Penc, and P. Fulde, Quantum ice: A quantum Monte Carlo study, Phys. Rev. Lett. 108, 067204 (2012). * Benton _et al._ [2012] O. Benton, O. Sikora, and N. Shannon, Seeing the light: Experimental signatures of emergent electromagnetism in a quantum spin ice, Phys. Rev. B 86, 075154 (2012). * Gingras and McClarty [2014] M. J. P. Gingras and P. A. McClarty, Quantum spin ice: a search for gapless quantum spin liquids in pyrochlore magnets, Rep. Prog. Phys. 77, 056501 (2014). * Sibille _et al._ [2018] R. Sibille, N. Gauthier, H. Yan, M. Ciomaga Hatnean, J. Ollivier, B. Winn, G. Balakrishnan, M. Kenzelmann, N. Shannon, and T. Fennell, Experimental signatures of emergent quantum electrodynamics in a quantum spin ice, Nature Phys. 14, 711 (2018). * Savary and Balents [2012] L. Savary and L. Balents, Coulombic quantum liquids in spin-1/2 pyrochlores, Phys. Rev. Lett. 108, 037202 (2012). * Hao _et al._ [2014] Z. Hao, A. G. R. Day, and M. J. P. Gingras, Bosonic many-body theory of quantum spin ice, Phys. Rev. B 90, 214430 (2014). * Petit _et al._ [2016] S. Petit, E. Lhotel, B. Canals, M. Ciomaga Hatnean, J. Ollivier, H. Mutka, E. Ressouche, A. R. Wildes, M. R. Lees, and G. Balakrishnan, Observation of magnetic fragmentation in spin ice, Nature Phys. 12, 746 (2016). * Xu _et al._ [2020] J. Xu, O. Benton, A. T. M. N. Islam, T. Guidi, G. Ehlers, and B. Lake, Order out of a Coulomb phase and Higgs transition: Frustrated transverse interactions in Nd2Zr2O7, Phys. Rev. Lett. 124, 097203 (2020). * Lhotel _et al._ [2015] E. Lhotel, S. Petit, S. Guitteny, O. Florea, M. Ciomaga Hatnean, C. Colin, E. Ressouche, M. R. Lees, and G. Balakrishnan, Fluctuations and all-in–all-out ordering in dipole-octupole Nd2Zr2O7, Phys. Rev. Lett. 115, 197202 (2015). * Xu _et al._ [2015] J. Xu, V. K. Anand, A. K. Bera, M. Frontzek, D. L. Abernathy, N. Casati, K. Siemensmeyer, and B. Lake, Magnetic structure and crystal-field states of the pyrochlore antiferromagnet Nd2Zr2O7, Phys. Rev. B 92, 224430 (2015). * Pekker and Varma [2015] D. Pekker and C. M. Varma, Amplitude / Higgs modes in condensed matter physics, Annu. Rev. Condens. Matter Phys. 6, 269 (2015). * Chang _et al._ [2012] L.-J. Chang, S. Onoda, Y. Su, Y.-J. Kao, K.-D. Tsuei, Y. Yasui, K. Kakurai, and M. R. Lees, Higgs transition from a magnetic Coulomb liquid to a ferromagnet in Yb2Ti2O7, Nature Commun. 3, 992 (2012). * Lhotel _et al._ [2018] E. Lhotel, S. Petit, M. Ciomaga Hatnean, J. Ollivier, H. Mutka, E. Ressouche, M. R. Lees, and G. Balakrishnan, Evidence for dynamic kagome ice, Nature Commun. 9, 3786 (2018). * Xu _et al._ [2019] J. Xu, O. Benton, V. K. Anand, A. T. M. N. Islam, T. Guidi, G. Ehlers, E. Feng, Y. Su, A. Sakai, P. Gegenwart, and B. Lake, Anisotropic exchange hamiltonian, magnetic phase diagram, and domain inversion of Nd2Zr2O7, Phys. Rev. B 99, 144420 (2019). * [23] See Supplemental Material for details on single crystal growth, neutron diffraction, polarized neutron experiments, inelastic neutron experiments, spin dynamics in Ti doped sample, analysis of classical dynamics results, magnetization, which includes Refs. [24, 25, 26, 27, 28]. * Ciomaga Hatnean _et al._ [2015] M. Ciomaga Hatnean, M. R. Lees, and G. Balakrishnan, Growth of single-crystals of rare-earth zirconate pyrochlores, ${Ln}_{2}$Zr2O7 (with ${Ln}=$La, Nd, Sm, and Gd) by the floating zone technique, J. Cryst. Growth 418, 1 (2015). * Ciomaga Hatnean _et al._ [2016] M. Ciomaga Hatnean, C. Decorse, M. R. Lees, O. A. Petrenko, and G. Balakrishnan, Zirconate pyrochlore frustrated magnets: crystal growth by the floating zone technique, Crystals 6, 79 (2016). * Rodríguez-Carvajal [1993] J. Rodríguez-Carvajal, Recent advances in magnetic structure determination by neutron powder diffraction, Physica B 192, 55 (1993). * Arnold and et al. [2014] O. Arnold and et al., Mantid - data analysis and visualization package for neutron scattering and $\mu$SR experiments, Nucl. Instrum. Methods Phys. Res. Sect. A 764, 156 (2014). * Ewings _et al._ [2016] R. A. Ewings, A. Buts, M. D. Lee, J. van Duijn, I. Bustinduy, and T. G. Perring, Horace: Software for the analysis of data from single crystal spectroscopy experiments at time-of-flight neutron instruments, Nucl. Instrum. Methods Phys. Res. Sect. A 834, 132 (2016). * Paulsen [2001] C. Paulsen, Dc magnetic measurements, in _Introduction to Physical Techniques in Molecular Magnetism: Structural and Macroscopic Techniques - Yesa 1999_ , edited by F. Palacio, E. Ressouche, and J. Schweizer (Servicio de Publicaciones de la Universidad de Zaragoza, 2001) p. 1. * Huang _et al._ [2014] Y.-P. Huang, G. Chen, and M. Hermele, Quantum spin ices and topological phases from dipolar-octupolar doublets on the pyrochlore lattice, Phys. Rev. Lett. 112, 167203 (2014). * Benton [2016] O. Benton, Quantum origins of moment fragmentation in Nd2Zr2O7, Phys. Rev. B 94, 104430 (2016). * Brooks-Bartlett _et al._ [2014] M. E. Brooks-Bartlett, S. T. Banks, L. D. C. Jaubert, A. Harman-Clarke, and P. C. W. Holdsworth, Magnetic-moment fragmentation and monopole crystallization, Phys. Rev. X 4, 011007 (2014). * Yan _et al._ [2018] H. Yan, R. Pohle, and N. Shannon, Half moons are pinch points with dispersion, Phys. Rev. B 98, 140402(R) (2018). * Savary and Balents [2013] L. Savary and L. Balents, Spin liquid regimes at nonzero temperature in quantum spin ice, Phys. Rev. B 87, 205130 (2013). * Huang _et al._ [2018] C.-J. Huang, Y. Deng, Y. Wan, and Z. Y. Meng, Dynamics of topological excitations in a model quantum spin ice, Phys. Rev. Lett. 120, 167202 (2018). Spin dynamics and unconventional Coulomb phase in Nd2Zr2O7 Supplementary Material ## I Single crystal growth Single crystals of Nd2(Zr1-xTix)2O7 ($x=0$ and 0.025) were grown by the floating zone method, using a four-mirror xenon arc lamp optical image furnace [24, 25]. A summary of the conditions used for each crystal growth is given in Table S1. Crystal | Sample label | Lattice parameter | Growth rate | Growth atmosphere, | Feed / seed ---|---|---|---|---|--- | Å | (mm/h) | pressure | rotation rate (rpm) Nd2Zr2O7 | Sample #1 | $10.66\pm 0.02$ | 12.5 | Air, ambient | 15 / 30 Nd2Zr2O7 | Sample #2 | $10.66\pm 0.04$ | 15 | Air, ambient | 20 / 25 Nd2(Zr1-xTix)2O7 | Sample #3 | $10.65\pm 0.02$ | 10 | Air, ambient | 15 / 5 Table S1: Summary of the samples with their crystal growth conditions. The lattice parameters were obtained at 6 K on neutron diffractometers (Sample #1 and #3) and triple axis spectrometers (all samples). Two different pure Nd2Zr2O7 samples had to be used in inelastic neutron scattering experiments, because the first one broke when warming up the dilution fridge after an experiment. ## II Characterization of the Ti substituted sample (Sample #3) Figure S1: (a) Refinement of the crystal neutron structure factors at 6 K, giving a refined Ti content equal to 2.4 %. (b) Measured intensity on the magnetic peaks (220), (113), (351) and (260), and symmetry related peaks at 60 mK, obtained from the difference with the 6 K data, and compared to the refined intensity. We have studied a substituted sample, in which a small content of Zr is replaced by Ti, slightly shrinking the structure. The nominal composition of the studied sample is 2.5 % of Ti atoms. As shown below, this substitution only slightly affects the magnetic properties of the magnetic Nd3+ sublattice and the low temperature properties are qualitatively the same. The value of the Ti content was refined by neutron diffraction, thanks to the significant contrast between Zr and Ti. A series of Bragg peak intensities was collected at 6 K on the single crystal neutron diffractometer D23 (CEA CRG- ILL). The data are in agreement with the pyrochlore structure ($Fd{\bar{3}}m$ space group), with a lattice parameter of 10.65 Å and the 48f oxygen atoms at the position $x_{\rm 48f}=0.336$. The Ti content is found to be 2.4 %. The Fullprof refinement [26] of the structure factor is shown on Fig. S1(a). The Néel temperature was determined from very low temperature magnetization measurements, and found to be $T_{\rm N}=375$ mK. The magnetic contribution raises below $T_{\rm N}$ in neutron diffraction measurements on top of the crystalline peaks. The Fullprof refinement below $T_{\rm N}$ confirms the same “all in - all out” (AIAO) magnetic structure as in the pure sample (Fig. S1(b)). At 60 mK, the refined ordered Nd3+ magnetic moment is $1.19\pm 0.03~{}\mu_{\rm B}$. ## III Measurements and analysis of polarized neutron scattering experiments Figure S2: D7 data and analysis. Panels (a,b,c) display respectively the raw data at 60 mK, the symmetrized and noise filtered data (10 K data have been subtracted). Panel (d) shows the spin ice magnetic scattering function calculated for a lattice containing 5488 spins. Panel (e) shows the same calculation assuming random $\pm 1$ Ising spins. It serves as a background reference, and is subtracted from (d), just as the 10 K data are subtracted from the low temperature data. In polarized neutron experiments carried out at D7 (ILL, France), we used the $P_{z}$ polarization mode, ${\bf z_{Q}}$ being the axis normal to the scattering plane and parallel to $[1\bar{1}0]$. We measured $N+I^{(z)}$ in the NSF channel, and $I^{(y)}$ in the SF channel. Here we use conventional notations: $N$ is the crystalline structure factor, while $I^{(y)}$ and $I^{(z)}$ denote the spin-spin correlation functions between spin components parallel to ${\bf y_{Q}}$ and ${\bf z_{Q}}$ respectively. The ${\bf y_{Q}}$ axis lies within the scattering plane, perpendicular both to ${\bf Q}$ and ${\bf z_{Q}}$. We used a wavelength $\lambda=4.85$ Å. The sample was rotated by steps of 1 degree, and 2 positions of the detector bank have been combined. Standard corrections (vanadium and quartz) have been processed. Finally, in order to eliminate any background contribution, the data recorded at 10 K have been subtracted from the data taken at lower temperatures. In Figure 1 of the main article, the data have been symmetrized (for the three lowest temperatures) while the noise was reduced by a mean filtering. This image processing based treatment tends to reduce the variation between one pixel and the next. The idea of mean filtering is to replace each pixel value with the average value of its neighbors, including itself. This has the effect of eliminating pixel values which are unrepresentative of their surroundings. For the sake of illustration, Figure S2(a-c) shows the different steps of this processing for the 60 mK data. Unfortunately, it was not possible to determine intensities in absolute units from the D7 measurements. To determine the magnetic moment responsible for the spin ice-like diffuse scattering, we had to proceed in an alternative manner. To this end, we carried out a series of calculations, assuming a “theoretical sample crystal” consisting of Ising spins (of length unity) located at the rare earth sites of a pyrochlore lattice of size $L$. We have considered $n$ spin ice configurations generated on this pyrochlore lattice (via a Monte- Carlo algorithm) and computed the average structure factor from the obtained magnetic moment ${\bf m}_{i,a=x,y,z}$ at each site $i$. The total magnetic neutron intensity, proportional to the spin-spin correlation, is calculated as: $I^{(y)}+I^{(z)}=\sum_{i,j}\sum_{a,b=x,y,z}m_{i,a}\left(\delta_{ab}-\frac{{\bf Q}_{a}{\bf Q}_{b}}{{\bf Q}^{2}}\right)m_{j,b}~{}e^{i{\bf Q}.({\bf R}_{i}-{\bf R}_{j})}$ and the intensity in the SF $P_{z}$ mode, corresponding to the data, is given by $\displaystyle I^{(y)}=\sum_{i,j}{\bf m}_{i}.{\bf y}_{\bf Q}\ {\bf m}_{j}.{\bf y}_{\bf Q}\ e^{i{\bf Q}.({\bf R}_{i}-{\bf R}_{j})}$. Noteworthy, the Monte-Carlo sampling was checked using the analytical method proposed by C.L. Henley [6]. The spin-spin correlation function (per unit cell) is written as: $I^{(y)}=4~{}t_{0}^{2}~{}\sum_{a=x,y,z}M^{T}_{a}\left[I-E(E^{+}E)^{-1}E^{+}\right]M_{a}$ $E$ is a 2-column matrix and $M_{a=x,y,z}$ is a collection of 1-column vectors containing the coordinates $a=x,y,z$ of the four magnetic moments ${\bf m}_{i}$ belonging to a given tetrahedron: $\displaystyle E$ $\displaystyle=\begin{pmatrix}e^{-i\pi{\bf Q.u_{1}}}&e^{i\pi{\bf Q.u_{1}}}\\\ e^{-i\pi{\bf Q.u_{2}}}&e^{i\pi{\bf Q.u_{2}}}\\\ e^{-i\pi{\bf Q.u_{3}}}&e^{i\pi{\bf Q.u_{3}}}\\\ e^{-i\pi{\bf Q.u_{4}}}&e^{i\pi{\bf Q.u_{4}}}\\\ \end{pmatrix}$ $\displaystyle M_{a}$ $\displaystyle=\begin{pmatrix}M_{1,a}\\\ M_{2,a}\\\ M_{3,a}\\\ M_{4,a}\end{pmatrix}\quad{\rm where}\quad M_{i,a}=m_{i,a}-\frac{{\bf m}_{i}.{\bf Q}}{Q^{2}}Q_{a}-\left(\sum_{b}\left(m_{i,b}-\frac{{\bf m}_{i}.{\bf Q}}{Q^{2}}Q_{b}\right).z_{{\bf Q},b}\right)z_{{\bf Q},a}\quad{\rm for}\quad{i=1,2,3,4}$ The four moments are defined as ${\bf m}_{1}=(c,c,c)$, ${\bf m}_{2}=(-c,-c,c)$, ${\bf m}_{3}=(-c,c,-c)$, ${\bf m}_{4}=(c,-c,-c)$ with $c=1/\sqrt{3}$ and attached to a tetrahedron. The ${\bf u_{i}}$ are vectors pointing towards the four corners of a tetrahedron: ${\bf u_{1}}=(d,d,d)$, ${\bf u_{2}}=(-d,-d,d)$, ${\bf u_{3}}=(-d,d,-d)$, ${\bf u_{4}}=(d,-d,-d)$ and $d=1/4$. Proper normalization condition imposes $t_{0}^{2}=2$. In the same way, we computed the structure factor $I^{(y)}_{\rm rd}$ assuming that the Ising spins have purely random $\pm 1$ values. This quantity was used as a background reference, just as the 10 K measurement was used as described above. We eventually considered the case where the spins are arranged in an “all in – all out” (AIAO) ordering, leading to magnetic Bragg peaks with a structure factor denoted hereafter $I^{(y)}_{\rm AIAO}$. First we calculated the integrated intensity around $(11\bar{3})$ from the experimental data at different temperatures, yielding $I^{\rm exp}_{\rm AIAO}(T)$. On the other hand, the same quantity was determined from $I^{(y)}_{\rm AIAO}$, yielding $I^{\rm c}_{\rm AIAO}$. Since the actual value of the ordered AIAO moment at low temperature $m_{\rm AIAO}$ (called $m_{2}$ in the main text) is precisely known from diffraction measurements (D23), we introduced the normalization factor $c$: $c=\frac{I^{\rm exp}_{\rm AIAO}}{m_{\rm AIAO}^{2}~{}I^{\rm c}_{\rm AIAO}}$ In a second step, we have computed the experimental integrated intensity within a box delineating the arm along $(hh{\bar{h}})$, yielding $\Delta I^{\rm exp}_{\rm arm}$. To obtain an estimate of the moment $m_{\rm SI}$ (called $m_{1}$ in the main text) involved in the spin ice component, which reflects the evolution seen on the maps presented in Figure 1 of the main text, we proposed to compare $\Delta I^{\rm exp}_{\rm arm}$ to $\Delta I^{\rm c}_{\rm arm}=m_{\rm SI}^{2}(I^{(y)}-I^{(y)}_{\rm rd})$. The estimation is then made quantitative by looking for $m_{\rm SI}$ such that : $\Delta I^{\rm exp}_{\rm arm}(T)=c\times\Delta I^{\rm c}_{\rm arm}$ The obtained values are listed in Table S2. Other calculation methods have been tested and lead to similar results in terms of absolute values and evolution with temperature. Furthermore, using $m_{\rm AIAO}=0.8\pm 0.05~{}\mu_{\rm B}$, the accuracy on $m_{\rm SI}$ is estimated to $\pm 0.3~{}\mu_{\rm B}$. This analysis confirms that the spin ice pattern has a weak maximum close to $T_{\rm N}$ and persists up to 600 mK, i.e. far above $T_{\rm N}$. $T$ (mK) | $I^{\rm exp}_{\rm arm}(T)$ | $m_{\rm SI}$ ($\mu_{\rm B}$) ---|---|--- 60 | 0.027 | 1.97 235 | 0.035 | 2.25 450 | 0.030 | 2.05 600 | 0.020 | 1.70 800 | 0.015 | 1.46 1000 | 0.0095 | 1.16 Table S2: Values of the spin moment $m_{\rm SI}$ vs temperature, determined from the procedure described in the text. Note that the temperature of 235 mK was estimated from the amplitude of the magnetic Bragg peaks (the thermometer indicated 300 mK). For 450 mK, no precise determination of the sample temperature could be done, but, in the absence of magnetic Bragg peaks, the temperature was definitely above $T_{\rm N}$. ## IV Time of flight inelastic scattering measurements Inelastic neutron scattering experiments were carried out on the IN5 disk chopper time of flight spectrometer (ILL, France). A good compromise between flux, energy resolution and accessible ${\bf Q}$ space was obtained with a wave length $\lambda=6$ or 6.5 Å. However, to ensure a better energy resolution $\Delta E=20~{}\mu$eV, necessary to fully resolve the dynamic spin ice mode at $E_{0}$, experiments were also conducted with $\lambda=8.5$ Å. The data were processed with the Mantid [27] and horace [28] softwares, transforming the recorded time of flight, sample rotation and scattering angle into energy transfer and ${\bf Q}$-wave vectors. The offset of the sample rotation was determined based on the Bragg peak positions. In all the experiments, the sample was rotated in steps of 1 degree and the counting time was about 10 minutes per sample position. It should be noticed that a very long thermalization time was systematically necessary to cool down the sample to the lowest temperature. In addition, we realized that when warming up from the lowest temperature, the sample temperature was not necessarily the same as the temperature indicated by the thermometer. For this reason, when possible (depending on the ratio between the resolution and the temperature), we have refined the “true” temperature by fitting the negative energy part of the spectra. It leads to the temperatures indicated on Figures 2, 3 and 4 of the main text, which are quite different from the thermometer temperatures. These temperatures are summarized in Table S3. Sample | Thermometer | Estimated ---|---|--- | temperature | temperature Sample #1 | 450 mK | $341\pm 100$ mK Sample #2 | 60 mK | $323\pm 78$ mK | 300 mK | $313\pm 68$ mK | 450 mK | $444\pm 111$ mK Sample #3 | 275 mK | $242\pm 35$ mK | 350 mK | $317\pm 38$ mK Table S3: Estimated effective temperatures in the different experiments performed on IN5. Figure S3: Constant ${\bf Q}$-cuts at two temperatures, the base temperature of 60 mK (blue) and above $T_{\rm N}$ (red) and which clearly show the vestiges of spin waves. Constant ${\bf Q}$-cuts from the data have been performed at the base temperature (typically 60 mK) and above $T_{\rm N}$ (340 mK), to clearly show the persistence of the spin wave signal above $T_{\rm N}$. These cuts, displayed in Figure S3 are along $(1,1,\ell)$, $(2,2,\ell)$, $(h,h,0)$ and $(h,h,2)$. Two of them are reproduced in the main text. This residual spin wave signal above $T_{\rm N}$ is not expected in conventional three dimensional paramagnets, in the absence of magnetic frustration. It thus would not be observed in a standard “all in – all out” antiferromagnet, which is predicted to behave classically close to the antiferromagnetic transition. The persistence of the spin wave signal (and of “all in – all out” diffuse scattering) quite far above $T_{\rm N}$ in Nd2Zr2O7 thus points out the unconventional nature of the magnetism in this compound and is likely related to the strong competition at play with the Coulomb phase observed above $T_{\rm N}$. ## V Inelastic scattering measurements on a triple axis spectrometer The temperature dependence of the spin dynamics in Sample #1 was also investigated on the cold TAS spectrometer IN12 (ILL, France). Scans at specific ${\bf Q}$ positions (0.5 0.5 2), (0 0 2.5) and (1.8 1.8 0) have been performed at different temperatures ranging from 50 up to 800 mK. Those positions were chosen since they probe different regions with respect to the dispersion. (0.5 0.5 2) essentially probes the flat spin ice band, (1.8 1.8 0) is sensitive to the zone boundary dispersive spin wave mode and (0.5 0.5 2) is somehow intermediate. A final wave vector $k_{f}=1.05$ Å-1 was used (in combination with nitrogen cooled Be filter) to ensure the best energy resolution, $\Delta E=50~{}\mu$eV. A magnetic field was also applied along $[1\bar{1}0]$. After correction from the detailed balance factor, we computed the difference between data taken a given temperature $T$ and the 800 mK data. Where applicable, we subtracted the data obtained at the same temperature but under a 1 T magnetic field. We could then extract the energy and intensity of the inelastic mode in the same way as for TOF measurements. The temperatures below $T_{\rm N}$ were estimated from the intensity of the (220) magnetic Bragg peak. ## VI Spin dynamics in the Ti substituted sample (Sample #3) ### VI.1 Determination of the parameters Inelastic neutron scattering data carried out at IN5 (ILL) on a single crystal sample show little evolution compared to the pure sample. The inelastic flat spin ice mode is observed at $E_{0}\approx 70~{}\mu$eV, while the dispersing mode stemming from the pinch point positions unfolds towards the zone centers, for instance $(220)$ or $(113)$. This is illustrated in Figure S4, which shows the dispersion along several reciprocal directions at 45 mK. Figure S4: Top: INS data taken at IN5 at 45 mK on the Sample #3 along high symmetry directions. Black and white dots are the energies $E_{0}$ and $E_{1}$ respectively, fitted according to the procedure described in the main text (see also equation (S1). Bottom: Spin wave calculations performed with the parameters given in Table 1 (main text). To determine the parameters of the XYZ Hamiltonian ${\cal H}_{\rm XYZ}=\sum_{\langle i,j\rangle}\left[{\tilde{\sf J}_{x}}\tilde{\tau}^{\tilde{x}}_{i}\tilde{\tau}^{\tilde{x}}_{j}+{\tilde{\sf J}_{y}}\tilde{\tau}^{\tilde{y}}_{i}\tilde{\tau}^{\tilde{y}}_{j}+{\tilde{\sf J}_{z}}\tilde{\tau}^{\tilde{z}}_{i}\tilde{\tau}^{\tilde{z}}_{j}\right]$ (see also equation (2) of the main text), we use analytic calculations giving the energy of the spin ice band [31]: $E_{0}=\sqrt{(3|\tilde{{\sf J}}_{z}|-\tilde{{\sf J}}_{x})(3|\tilde{{\sf J}}_{z}|-\tilde{{\sf J}}_{y})}$ as well as the energy of the dispersive modes at some high symmetry ${\bf Q}$ vectors [22]: $\displaystyle{\bf Q}=(110),(112)$ $\displaystyle,~{}$ $\displaystyle\Delta_{2}=\sqrt{(3|\tilde{{\sf J}}_{z}|+\tilde{{\sf J}}_{x})(3|\tilde{{\sf J}}_{z}|+\tilde{{\sf J}}_{y})}$ $\displaystyle{\bf Q}=(220),(113)$ $\displaystyle,~{}$ $\displaystyle\Delta_{3}=3\sqrt{(|\tilde{{\sf J}}_{z}|+\tilde{{\sf J}}_{x})(|\tilde{{\sf J}}_{z}|+\tilde{{\sf J}}_{y})}$ Simulations have then been performed to reproduce the data with the cefwave software developed at LLB using the Hamiltonian (1): ${\cal H}=\sum_{\langle i,j\rangle}\left[{\sf J}_{x}\tau^{x}_{i}\tau^{x}_{j}+{\sf J}_{y}\tau^{y}_{i}\tau^{y}_{j}+{\sf J}_{z}\tau^{z}_{i}\tau^{z}_{j}+{\sf J}_{xz}(\tau^{x}_{i}\tau^{z}_{j}+\tau^{z}_{i}\tau^{x}_{j})\right]$ The ground state configuration is first determined by solving this Hamiltonian at the mean field level, where the expectation values $\langle\tau^{x,y,z}_{j}\rangle$ are determined in a self-consistent manner. Spin wave calculations are performed using a generalized susceptibility approach out of the obtained configurations. Finally, the neutron cross section is calculated from $\tau^{z}_{i}\tau^{z}_{j}$ correlations. Notably, the simulations performed with the parameters of Table 1 (main text) reproduce quite well the data, as shown in Figure S4. ### VI.2 Temperature dependence of the spin dynamics Inelastic data in Sample #3 were fitted in the whole measured ${\bf Q}$ space using the model described in the main text: $S({\bf Q},E)=b+I_{c}(E)+F(E,T)\times\left[S_{0}(E)+S_{1}(E)\right]$ (S1) $b$ is a flat background (wavelength dependent), $I_{c}$ is a Gaussian function centered at zero energy to represent the elastic incoherent scattering. $F(E,T)=(1+n(E))$ is the detailed balance factor, and $S_{0}$ and $S_{1}$ are two Lorentzian profiles which represent respectively the flat band and the dispersive mode typical of the spin wave spectrum in Nd2Zr2O7. Figure S5 shows the energy $E_{0}$ of the flat band at different temperatures in the form of a map over the sector probed by TOF measurements. Figure S6 displays the intensities $I_{0}$ (panel a, upper row) and $I_{1}$ (panel b, lower row). The map on the right of the same figure shows the energy $E_{1}$ of the dispersive spin wave mode. To check the overall consistency of the fitting procedure, dashed lines visualize the directions of the scans reported in Figure S4. Note that for Sample #1, the fit was carried out at selected ${\bf Q}$ values (${\bf Q}=(0.8\ 0.8\ 0.8)$, $(1.1\ 1.1\ 1.1)$, $(1/2\ 1/2\ 1/2)$, $(1/2\ 1/2\ 3/2)$ and $(3/4\ 3/4\ 3/2)$). Figure S5: Temperature dependence of the flat spin ice band at $E_{0}$ deduced from the fit in Sample #3, as described in the main text. The portion of $({\bf Q},E)$ space corresponds to the sector probed by TOF measurements with $\lambda$=8.5 Å. Figure S6: (a) Temperature dependence of the intensity $I_{0}$ of the flat spin ice band from the fit in Sample #3, as described in the main text. (b) shows the intensity $I_{1}$ of the dispersive mode, and (c) shows its energy $E_{1}$. Dashed lines correspond to the directions of the cuts shown in Figure S4. The portion of $({\bf Q},E)$ space corresponds to the sector probed by TOF measurements with $\lambda=8.5$ Å. Figure S7: Temperature dependence of the intensity of the incoherent scattering $I_{c}(E=0)$ in Sample #3. The portion of $({\bf Q},E)$ space corresponds to the sector probed by TOF measurements with $\lambda=8.5$ Å. Finally, aiming at identifying a possible elastic contribution with the spin ice structure factor, Figure S7 shows the temperature evolution of the intensity of the incoherent elastic contribution, i.e. the dominant contribution $I_{c}(E=0)$ in the spectrum, to which the 45 mK map was subtracted. Within experimental uncertainties, these maps are featureless and no spin ice pattern can be clearly distinguished. ## VII Analysis of classical dynamics results Figure S8: (a) From Ref. 16 (courtesy of J. Xu): Evolution of the gapped flat mode for several temperatures (0.05, 0.1, 0.125, 0.15, 0.165, 0.175, 0.18, 0.185, 0.2 K) simulated using semi-classical molecular dynamics averaging over Q from (0.1 0.1 0) to (0.9 0.9 0). (b-c) Temperature dependences fitted from (a) (see equation S2): (b) Temperature dependence of the intensities $I_{0}$ and $I^{\prime}_{0}$ of the flat spin ice band and of the elastic contributions respectively. (c) Temperature dependence of the energy $E_{0}$ of the flat spin ice band. The main text of the present work compares the measured temperature dependence of $E_{0}$ and $I_{0}$ in our three samples with Monte Carlo calculations reported in Ref. 16. These calculations use effective exchange parameters from Ref. 22, which are detailed in Table 1 of the main text. They give a Néel temperature of 0.18 K. From these calculations, as illustrated in the Figure 3 of Ref. 16, two contributions are obtained: a spin ice elastic contribution which projects onto the ${\bf z}$ axis with a factor $\sin^{2}\theta$, as well as spin waves features characteristic of the AIAO phase, with especially a flat spin ice band. To compare quantitatively these results with our data, those theoretical curves have been fitted to two modes, following: $I(E)=I_{0}~{}e^{-4\log 2(\frac{E-E_{0})}{\delta_{0}})^{2}}+\frac{I^{\prime}_{0}}{1+(\frac{E}{\delta^{\prime}_{0}})^{2}}$ (S2) The result of this fit is illustrated in Figures S8(b) and S8(c), which display respectively the intensity of the modes ($I_{0}$ and $I^{\prime}_{0}$) and the position $E_{0}$ of the flat spin ice band. Interestingly, this energy remains finite even above the calculated critical temperature $T_{\rm N}=0.18$ K. ## VIII Correlations in magnetization measurements Figure S9: Magnetization $M$ vs $H/T$ measured for Sample #1 with the field applied along: (a) [100], (b) [110] and (c) [111] at 1, 1.8 and 4.2 K. In a paramagnet, isothermal magnetization curves scale as a function of the variable $H/T$. The deviations to this scaling give insight into the nature of the correlations that develop in the system. Upon cooling, if the magnetization curve increases faster (slower) than the higher temperature curve, it is the signature of the development of ferromagnetic-like (antiferromagnetic-like) correlations. We have plotted the magnetization as a function of $H/T$ for Nd2Zr2O7, measured in Sample #1. As shown in Figure S9, the $M(H/T)$ curves rise above the 4.2 K curves upon cooling down to 1 K, which indicates the development of ferromagnetic correlations, consistent with the elastic spin ice picture inferred from our neutron scattering measurements. At 500 mK, the curves lie between the 1 and 4.2 K curves (not shown on the figure for clarity), showing the development of antiferromagnetic correlations compared to 1 K, but the persistence of global ferromagnetic correlations. These antiferromagnetic correlations will end in the “all in – all out” ordering at about 300 mK.
# Growth of Sobolev norms in linear Schrödinger equations as a dispersive phenomenon A. Maspero111 International School for Advanced Studies (SISSA), Via Bonomea 265, 34136, Trieste, Italy Email<EMAIL_ADDRESS> ###### Abstract In this paper we consider linear, time dependent Schrödinger equations of the form ${\rm i}\partial_{t}\psi=K_{0}\psi+V(t)\psi$, where $K_{0}$ is a strictly positive selfadjoint operator with discrete spectrum and constant spectral gaps, and $V(t)$ a time periodic potential. We give sufficient conditions on $V(t)$ ensuring that $K_{0}+V(t)$ generates unbounded orbits. The main condition is that the resonant average of $V(t)$, namely the average with respect to the flow of $K_{0}$, has a nonempty absolutely continuous spectrum and fulfills a Mourre estimate. These conditions are stable under perturbations. The proof combines pseudodifferential normal form with dispersive estimates in the form of local energy decay. We apply our abstract construction to the Harmonic oscillator on ${\mathbb{R}}$ and to the half-wave equation on ${\mathbb{T}}$; in each case, we provide large classes of potentials which are transporters. ## 1 Introduction We consider the abstract linear Schrödinger equation ${\rm i}\partial_{t}\psi=K_{0}\psi+V(t)\psi\ $ (1.1) on a scale of Hilbert spaces ${\mathcal{H}}^{r}$; here $V(t)$ is a time $2\pi$-periodic potential and $K_{0}$ a selfadjoint, strictly positive operator with compact resolvent, pure point spectrum and constant spectral gaps. We prove some abstract results ensuring, $\forall r>0$, the existence of solutions $\psi(t)$ whose ${\mathcal{H}}^{r}$-norms grow polynomially fast, $\|\psi(t)\|_{r}\geq C_{r}\,\left\langle t\right\rangle^{r},\quad\forall t\gg 1\ ,$ whereas their ${\mathcal{H}}^{0}$-norms are constant for all times, $\|\psi(t)\|_{0}=\|\psi(0)\|_{0}$ $\forall t$. Here $\left\langle t\right\rangle:=\sqrt{1+t^{2}}$. These solutions therefore exhibit weak turbulent behavior in the form of energy cascade towards high frequencies. We apply our abstract result to two models: the Harmonic oscillator on ${\mathbb{R}}$ and the half-wave equation on ${\mathbb{T}}$. In both cases we exhibit large classes of potentials $V(t)$, bounded, smooth and periodic in time, so that the Hamiltonian $K_{0}+V(t)$ generates unbounded orbits. The phenomenon is purely perturbative: for $V=0$ each norm of each solution is constant for all times. So the central question is the existence of potentials able to transport energy to high-frequencies; we formalize this notion in the following definition: ###### Definition 1.1. We shall say that $V(t)$ is a transporter if $\forall r>0$ there exists a solution $\psi(t)\in{\mathcal{H}}^{r}$ of (1.1) with unbounded growth of norm, i.e. $\limsup_{t\to\infty}\|\psi(t)\|_{r}=\infty.$ (1.2) If this happens for every nonzero solution, we shall say that $V(t)$ is a universal transporter. Starting with the pioneering work of Bourgain [9], in the last few years there has been some efforts to construct both transporters [17, 22, 44] and universal transporters [6, 34] for different types of Schrödinger equations. All these papers provide explicit examples of potentials, constructed ad hoc for the problem at hand. The novelty of our result is that we identify sufficient, explicit and robust (i.e. stable under perturbations) conditions ensuring $V(t)$ to be a transporter. Precisely, its resonant average $\displaystyle\left\langle V\right\rangle:=\frac{1}{2\pi}\int_{0}^{2\pi}e^{{\rm i}sK_{0}}\,V(s)\,e^{-{\rm i}sK_{0}}\,{\rm d}s\ $ (1.3) must have nontrivial absolutely continuous spectrum in an interval, and over this interval it has to fulfill a Mourre estimate – see (2.7) below (actually we also require that both $K_{0}$ and $V(t)$ belong to some abstract graded algebra of pseudodifferential operators, as in [5]). The crucial point is that these conditions imply dispersive estimates for $\left\langle V\right\rangle$ of the form $\|K_{0}^{-k}e^{-{\rm i}t\left\langle V\right\rangle}P_{c}\phi\|_{0}\lesssim\left\langle t\right\rangle^{-k}\|K_{0}^{k}\phi\|_{0}\ ,\quad\forall t\in{\mathbb{R}},$ (1.4) where $P_{c}$ is a projection on a subset of the absolutely continuous spectral space of $\left\langle V\right\rangle$. A consequence of (1.4) is that we obtain solutions of ${\rm i}\partial_{t}\phi=\left\langle V\right\rangle\phi$ with decaying negative Sobolev norms and so, by duality, growing positive Sobolev norms. The fact that Mourre estimates imply dispersive estimates as above has origin from the work of Sigal-Soffer in quantum scattering theory [41] and it has been extended by many authors (see e.g. [42, 23, 31, 30, 24, 2]). See also the recent results [13, 12, 20] where similar dispersive properties are studied for pseudodifferential operators of order 0 on compact manifolds of dimension greater equal $2$. To explain the connection between the dynamics of (1.1) and the dispersive properties of the flow of $\left\langle V\right\rangle$, let us briefly describe the main ideas of the proof. The first step is to put system (1.1) into its resonant pseudodifferential normal form. This is the resonant variant of the normal form developed in [5] for non-resonant systems (and essentially an abstract version of the normal form of Delort [17]); it allows, $\forall N\in{\mathbb{N}}$, to conjugate equation (1.1) to ${\rm i}\partial_{t}{\varphi}=\big{(}K_{0}+Z^{(N)}(t)+R^{(N)}(t)\big{)}{\varphi}$ (1.5) where $Z^{(N)}(t)$ is a time dependent operator fulfilling ${\rm i}\partial_{t}Z^{(N)}(t)=[K_{0},Z^{(N)}(t)],\qquad Z^{(N)}(0)=\left\langle V\right\rangle+\mbox{lower order terms}$ (1.6) whereas $R^{(N)}(t)$ is an $N$-smoothing operator (it maps ${\mathcal{H}}^{r}\to{\mathcal{H}}^{r+N}$ continuously $\forall r$). The difference with the non-resonant case of [5] is that, in that paper, $Z^{(N)}(t)$ commutes with $K_{0}$. This is not anymore true in the resonant case we deal with; however (1.6) implies that $e^{{\rm i}tK_{0}}Z^{(N)}(t)e^{-{\rm i}tK_{0}}$ is time independent and thus coincides with $Z^{(N)}(0)$. Thus, conjugating (1.5) with $e^{-{\rm i}tK_{0}}$, we arrive at the equation ${\rm i}\partial_{t}\phi=\big{(}\left\langle V\right\rangle+T+R(t)\big{)}\phi$ (1.7) where $T:=Z^{(N)}(0)-\left\langle V\right\rangle$ is a time independent selfadjoint compact operator and $R(t)$ is $N$-smoothing. Then we analyze the dynamics of the truncated equation ${\rm i}\partial_{t}\phi=\big{(}\left\langle V\right\rangle+T\big{)}\phi$ (1.8) and prove that it has solutions with decaying negative Sobolev norms and so, by duality, growing positive Sobolev norms. This is the core of the proof; after this step, it is not difficult to construct a solution of the complete equation (1.7) exhibiting energy cascade, exploiting that $R(t)$ is regularizing. So let us concentrate on (1.8). The goal is to prove a dispersive estimate of the form (1.4) with $\left\langle V\right\rangle$ replaced by $\left\langle V\right\rangle+T$. However the point is delicate because the absolutely continuous spectrum of $\left\langle V\right\rangle$ (which exists by assumption) could be completely destroyed by adding $T$; a celebrated theorem by Weyl-von Neumann ensures that any selfadjoint operator (in a separable Hilbert space) can be perturbed by a (arbitrary small) compact selfadjoint operator so that its spectrum becomes pure point (see e.g. [32, pag. 525]). This is exactly the situation we want to avoid, as pure point spectrum prevents dispersive estimates. To get around this, we exploit that Mourre estimates are stable under pseudodifferential perturbations. This allows us to prove that $\left\langle V\right\rangle+T$ fulfills Mourre estimates and thus a dispersive estimate as (1.4). We also stress that fulfilling a Mourre estimate seems to be a quite general condition, and in the applications we exhibit large classes of operators which are transporters. For example, for the half wave equation we prove that any operator of the form $\cos(mt)v(x)$ with $v\in C^{\infty}({\mathbb{T}},{\mathbb{R}})$ and $m\in{\mathbb{Z}}$ is a transporter provided the $m$-th Fourier coefficient of $v(x)$ is not zero. Finally, the conditions we identity to be transporters are robust: if a potential $V(t)$ fulfills them, so does $V(t)+W(t)$ for any sufficiently small pseudodifferential operator $W(t)$. This shows that weakly turbulent phenomena induced by certain transporters are stable under perturbations. Up to our knowledge, this “stability of instability” is new in the literature and we consider it one of the main novelty of the paper. We conclude the introduction by reviewing the known results about existence of transporters for linear time dependent Schrödinger equations. As we already mentioned, the first result is due to Bourgain [9], who constructed a transporter for the Schrödinger equation on the torus; in this case $V(t,x)$ is a bounded real analytic function. Delort [17] constructs a transporter for the harmonic oscillator on ${\mathbb{R}}$, which is a time $2\pi$-periodic pseudodifferential operator of order zero. In [6] we proved that $ax\sin(t)$, $a>0$, is a universal transporter for the harmonic oscillator on ${\mathbb{R}}$; in this case the potential is an unbounded operator. In [34] we constructed universal transporters for the abstract equation (1.1), and applied the result to the harmonic oscillator on ${\mathbb{R}}$, the half-wave equation on ${\mathbb{T}}$ and on a Zoll manifold; in all cases the universal transporters are time periodic pseudodifferential operators of order 0. Finally recently Faou-Raphael [22] constructed a transporter for the harmonic oscillator on ${\mathbb{R}}$ which is a time dependent function (and not a pseudodifferential operator), and Thomann [44] has constructed a transporter for the harmonic oscillator on the Bargman-Fock space. Finally we recall the long-time growth result [25] for the semiclassical anharmonic oscillator on ${\mathbb{R}}^{d}$. Acknowledgments: We thank Matteo Gallone for helpful discussions on spectral theory and Dario Bambusi and Didier Robert for useful suggestions during the preparation of this work. ## 2 The abstract result We start with a Hilbert space ${\mathcal{H}}$, endowed with the scalar product $\left\langle\cdot,\cdot\right\rangle$, and a reference operator $K_{0}$, which we assume to be selfadjoint, positive, namely such that $\langle\psi;K_{0}\psi\rangle\geq c_{K_{0}}\|\psi\|^{2}\ ,\quad\forall\psi\in D(K_{0}^{1/2})\ ,\quad c_{K_{0}}>0\ ,$ and with compact resolvent. We define as usual a scale of Hilbert spaces by ${\mathcal{H}}^{r}:=D(K_{0}^{r})$ (the domain of the operator $K_{0}^{r}$) if $r\geq 0$, and ${\mathcal{H}}^{r}=({\mathcal{H}}^{-r})^{\prime}$ (the dual space) if $r<0$. Finally we denote by ${\mathcal{H}}^{-\infty}=\bigcup_{r\in{\mathbb{R}}}{\mathcal{H}}^{r}$ and ${\mathcal{H}}^{+\infty}=\bigcap_{r\in{\mathbb{R}}}{\mathcal{H}}^{r}$. We endow ${\mathcal{H}}^{r}$ with the natural norm $\|\psi\|_{r}:=\|(K_{0})^{r}\psi\|_{0}$, where $\|\cdot\|_{0}$ is the norm of ${\mathcal{H}}^{0}\equiv{\mathcal{H}}$. Notice that for any $m\in{\mathbb{R}}$, ${\mathcal{H}}^{+\infty}$ is a dense linear subspace of ${\mathcal{H}}^{m}$ (this is a consequence of the spectral decomposition of $K_{0}$). ###### Remark 2.1. By the very definition of ${\mathcal{H}}^{r}$, the unperturbed flow $e^{-{\rm i}tK_{0}}$ preserves each norm, $\|e^{-{\rm i}tK_{0}}\psi\|_{r}=\|\psi\|_{r}$ $\,\forall t\in{\mathbb{R}}$. Consequently, every orbit of equation (1.1) with $V(t)=0$ is bounded. Following [5], we introduce now a graded algebra ${\mathcal{A}}$ of operators which mimic some fundamental properties of different classes of pseudodifferential operators. For $m\in{\mathbb{R}}$ let ${\mathcal{A}}_{m}$ be a linear subspace of $\bigcap_{s\in{\mathbb{R}}}{\mathcal{L}}({\mathcal{H}}^{s},{\mathcal{H}}^{s-m})$ and define ${\mathcal{A}}:=\bigcup_{m\in{\mathbb{R}}}{\mathcal{A}}_{m}$. We notice that the space $\bigcap_{s\in{\mathbb{R}}}{\mathcal{L}}({\mathcal{H}}^{s},{\mathcal{H}}^{s-m})$ is a Fréchet space equipped with the semi-norms: $\|A\|_{m,s}:=\|A\|_{{\mathcal{L}}({\mathcal{H}}^{s},{\mathcal{H}}^{s-m})}$. We shall need to control the smoothing properties of the operators in the scale $\\{{\mathcal{H}}^{r}\\}_{r\in{\mathbb{R}}}$. If $A\in{\mathcal{A}}_{m}$ then $A$ is more and more smoothing if $m\rightarrow-\infty$ and the opposite as $m\rightarrow+\infty$. We will say that $A$ is of order $m$ if $A\in{\mathcal{A}}_{m}$. ###### Definition 2.2. We say that $S\in{\mathcal{L}}({\mathcal{H}}^{+\infty},{\mathcal{H}}^{-\infty})$ is $N$-smoothing if $\forall\kappa\in{\mathbb{R}}$, it can be extended to an operator in ${\mathcal{L}}({\mathcal{H}}^{\kappa},{\mathcal{H}}^{\kappa+N})$. When this is true for every $N\geq 0$, we say that $S$ is a smoothing operator. The first set of assumptions concerns the properties of ${\mathcal{A}}_{m}$: Assumption I: Pseudodifferential algebra * (i) For each $m\in{\mathbb{R}}$, $K_{0}^{m}\in{\mathcal{A}}_{m}$; in particular $K_{0}$ is an operator of order one. * (ii) For each $m\in{\mathbb{R}}$, ${\mathcal{A}}_{m}$ is a Fréchet space for a family of filtering semi-norms $\\{\wp^{m}_{j}\\}_{j\geq 0}$ such that the embedding ${\mathcal{A}}_{m}\hookrightarrow\bigcap_{s\in{\mathbb{R}}}{\mathcal{L}}({\mathcal{H}}^{s},{\mathcal{H}}^{s-m})$ is continuous222A family of seminorms $\\{\wp^{m}_{j}\\}_{j\geq 0}$ is called filtering if for any $j_{1},j_{2}\geq 0$ there exist $k\geq 0$ and $c_{1},c_{2}>0$ such that the two inequalities $\wp^{m}_{j_{1}}(A)\leq c_{1}\wp^{m}_{k}(A)$ and $\wp^{m}_{j_{2}}(A)\leq c_{2}\wp^{m}_{k}(A)$ hold for any $A\in{\mathcal{A}}_{m}$.. If $m^{\prime}\leq m$ then ${\mathcal{A}}_{m^{\prime}}\subseteq{\mathcal{A}}_{m}$ with a continuous embedding. * (iii) ${\mathcal{A}}$ is a graded algebra, i.e. $\forall m,n\in{\mathbb{R}}$: if $A\in{\mathcal{A}}_{m}$ and $B\in{\mathcal{A}}_{n}$ then $AB\in{\mathcal{A}}_{m+n}$ and the map $(A,B)\mapsto AB$ is continuous from ${\mathcal{A}}_{m}\times{\mathcal{A}}_{n}$ into ${\mathcal{A}}_{m+n}$. * (iv) ${\mathcal{A}}$ is a graded Lie-algebra333This property will impose the choice of the semi-norms $\\{\wp^{m}_{j}\\}_{j\geq 1}$. We will see in the examples that the natural choice $(\|\cdot\|_{m,s})_{s\geq 0}$ has to be refined. : if $A\in{\mathcal{A}}_{m}$ and $B\in{\mathcal{A}}_{n}$ then the commutator $[A,B]\in{\mathcal{A}}_{m+n-1}$ and the map $(A,B)\mapsto[A,B]$ is continuous from ${\mathcal{A}}_{m}\times{\mathcal{A}}_{n}$ into ${\mathcal{A}}_{m+n-1}$. * (v) ${\mathcal{A}}$ is closed under perturbation by smoothing operators in the following sense: let $A$ be a linear map: ${\mathcal{H}}^{+\infty}\rightarrow{\mathcal{H}}^{-\infty}$. If there exists $m\in{\mathbb{R}}$ such that for every $N>0$ we have a decomposition $A=A^{(N)}+S^{(N)}$, with $A^{(N)}\in{\mathcal{A}}_{m}$ and $S^{(N)}$ is $N$-smoothing, then $A\in{\mathcal{A}}_{m}$. * (vi) If $A\in{\mathcal{A}}_{m}$ then also the adjoint operator $A^{*}\in{\mathcal{A}}_{m}$. The duality here is defined by the scalar product $\langle\cdot,\cdot\rangle$ of ${\mathcal{H}}={\mathcal{H}}^{0}$. The adjoint $A^{*}$ is defined by $\langle u,Av\rangle=\langle A^{*}u,v\rangle$ for $u,v\in{\mathcal{H}}^{\infty}$ and extended by continuity. It is well known that classes of pseudodifferential operators satisfy these properties, provided one chooses for $K_{0}$ a suitable operator of the right order (see e.g. [28]). ###### Remark 2.3. One has that $\forall A\in{\mathcal{A}}_{m}$, $\forall B\in{\mathcal{A}}_{n}$ $\displaystyle\forall m,s\quad\exists N\ s.t.\ $ $\displaystyle\|A\|_{m,s}\leq C_{1}\,\wp^{m}_{N}(A)\ ,$ (2.1) $\displaystyle\forall m,n,j\quad\exists N\ s.t.\ $ $\displaystyle\wp^{m+n}_{j}(AB)\leq C_{2}\,\wp^{m}_{N}(A)\,\wp^{n}_{N}(B)\ ,$ (2.2) $\displaystyle\forall m,n,j\quad\exists N\ s.t.\ $ $\displaystyle\wp^{m+n-1}_{j}([A,B])\leq C_{3}\,\wp^{m}_{N}(A)\,\wp^{n}_{N}(B)\ ,$ (2.3) for some positive constants $C_{1}(s,m)$, $C_{2}(m,n,j)$, $C_{3}(m,n,j)$. ###### Remark 2.4. Any $A\in{\mathcal{A}}_{m}$ with $m<0$ is a compact operator on ${\mathcal{H}}$. Indeed write $A=AK_{0}^{-m}\,K_{0}^{m}$. Then $AK_{0}^{-m}\in{\mathcal{A}}_{0}$ is a bounded operator on ${\mathcal{H}}$ (Assumption I (i)–(iii)), whereas $K_{0}^{m}\equiv(K_{0}^{-1})^{-m}$ is compact on ${\mathcal{H}}$, as $K_{0}^{-1}$ is a compact operator by assumption. For $\Omega\subseteq{\mathbb{R}}^{d}$ and ${\mathcal{F}}$ a Fréchet space, we will denote by $C_{b}^{m}(\Omega,{\mathcal{F}})$ the space of $C^{m}$ maps $f:\Omega\ni x\mapsto f(x)\in{\mathcal{F}}$ such that, for every seminorm $\|\cdot\|_{j}$ of ${\mathcal{F}}$, one has $\sup_{x\in\Omega}\|\partial_{x}^{\alpha}f(x)\|_{j}<+\infty\ ,\quad\forall\alpha\in{\mathbb{N}}^{d}\ :\ \left|\alpha\right|\leq m\ .$ (2.4) If (2.4) is true $\forall m$, we say $f\in C^{\infty}_{b}(\Omega,{\mathcal{F}})$. Similarly we denote by $C^{\infty}({\mathbb{T}},{\mathcal{F}})$ the space of smooth maps from the torus ${\mathbb{T}}={\mathbb{R}}/(2\pi{\mathbb{Z}})$ to the Fréchet space ${\mathcal{F}}$. The second set of assumptions concerns the operator $K_{0}$, its spectral structure and an Egorov-like property, also well known for pseudo-differential operators. Assumption II: Properties of $K_{0}$ * (i) The operator $K_{0}$ has purely discrete spectrum fulfilling ${\rm spec}(K_{0})\subseteq{\mathbb{N}}+\lambda$ (2.5) for some $\lambda\geq 0$. * (ii) For any $m\in{\mathbb{R}}$ and $A\in{\mathcal{A}}_{m}$, the map defined on ${\mathbb{R}}$ by $\tau\mapsto A(\tau):={\rm e}^{{\rm i}\tau K_{0}}\,A\,{\rm e}^{-{\rm i}\tau K_{0}}$ belongs to $C^{\infty}_{b}({\mathbb{R}},{\mathcal{A}}_{m})$ and one has $\forall j\quad\exists N\ s.t.\ \ \sup_{\tau\in{\mathbb{R}}}\wp^{m}_{j}(A(\tau))\leq C_{4}\,\wp^{m}_{N}(A)$ (2.6) for some positive constant $C_{4}(m,j)$. ###### Remark 2.5. Assumption II (i) guarantees that $e^{{\rm i}2\pi K_{0}}=e^{{\rm i}2\pi\lambda}$. As a consequence, for any operator $V$, the map $\tau\mapsto e^{{\rm i}\tau K_{0}}Ve^{-{\rm i}\tau K_{0}}$ is $2\pi$-periodic. We denote by $C^{\infty}_{c}({\mathbb{R}}^{d},{\mathbb{R}}_{\geq 0})$ the set of smooth functions with compact support from ${\mathbb{R}}^{d}$ to ${\mathbb{R}}_{\geq 0}$ (hence non-negative). Furthermore from now on, given two operators $\mathsf{A},\mathsf{B}\in{\mathcal{L}}({\mathcal{H}})$, we write $\mathsf{A}\leq\mathsf{B}$ with the meaning $\left\langle\mathsf{A}{\varphi},{\varphi}\right\rangle\leq\left\langle\mathsf{B}{\varphi},{\varphi}\right\rangle$ $\,\forall{\varphi}\in{\mathcal{H}}$. The last set of assumptions concerns the resonant average $\left\langle V\right\rangle$ of the potential $V(t)$ (see (1.3)) and its spectrum $\sigma(\left\langle V\right\rangle)$. Note that if $V(t)$ is selfadjoint $\forall t$, so is $\left\langle V\right\rangle$. Assumption III: Properties of the potential $V(t)$ The operator $V\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$, $V(t)$ selfadjoint $\forall t$, and its resonant average $\left\langle V\right\rangle$ fulfills: * (i) There exists an interval $I_{0}\subset{\mathbb{R}}$ such that $\left|\sigma(\left\langle V\right\rangle)\cap I_{0}\right|>0$; here $\left|\cdot\right|$ denotes the Lebesgue measure. * (ii) Mourre estimate over $I_{0}$: there exist a selfadjoint operator $A\in{\mathcal{A}}_{1}$ and a function $g_{I_{0}}\in C_{c}^{\infty}({\mathbb{R}},{\mathbb{R}}_{\geq 0})$ with $g_{I_{0}}\equiv 1$ on $I_{0}$ such that $g_{I_{0}}(\left\langle V\right\rangle)\,{\rm i}[\left\langle V\right\rangle,A]\,g_{I_{0}}(\left\langle V\right\rangle)\geq\theta\,g_{I_{0}}(\left\langle V\right\rangle)^{2}+K$ (2.7) for some $\theta>0$ and $K$ a selfadjoint compact operator. The operator $g_{I_{0}}(\left\langle V\right\rangle)$ above is defined via functional calculus, see Appendix B. Following the literature, we shall say that $\left\langle V\right\rangle$ is conjugated to $A$ over $I_{0}$. ###### Remark 2.6. By Mourre theory [36] $\left\langle V\right\rangle$ has, in the interval $I_{0}$, a nontrivial absolutely continuous spectrum with finitely many eigenvalues of finite multiplicity and no singular continuous spectrum. In general one cannot exclude the existence of embedded eigenvalues in the absolutely continuous spectrum.444 For example consider $H\in{\mathcal{L}}(L^{2}({\mathbb{T}}))$ given by $(Hu)(x):=\cos(x)u(x)+\delta(1-\delta^{-1}\cos(x))\frac{1}{2\pi}\int_{\mathbb{T}}u(x)\big{(}1-\delta^{-1}\cos(x)\big{)}\,{\rm d}x\ ,\quad\delta\in\left(-\frac{1}{2},\frac{1}{2}\right)\setminus\\{0\\}\ .$ $H$ is selfadjoint, a 1-rank perturbation of the multiplication operator by $\cos(x)$, it has absolutely continuous spectrum in the interval $(-1,1)$, and $\delta$ is an embedded eigenvalue with eigenvector $u(x)\equiv 1$. Moreover $H$ is conjugated to $\sin(x)\frac{\partial_{x}}{{\rm i}}+\frac{\partial_{x}}{{\rm i}}\sin(x)$ over $[-\frac{1}{2},\frac{1}{2}]$. We are ready to state our main results. The first says that, under the set of assumptions above, $V(t)$ is a transporter in the sense of Definition 1.1: ###### Theorem 2.7. Assume that ${\mathcal{A}}$ is a graded algebra as in Assumption I, and that $K_{0}$ and $V(t)\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$ satisfy Assumptions II and III. Then $V(t)$ is a transporter for the equation ${\rm i}\partial_{t}\psi=(K_{0}+V(t))\psi\ .$ (2.8) More precisely, for any $r>0$ there exist a solution $\psi(t)$ of (2.8) in ${\mathcal{H}}^{r}$ and constants $C,T>0$ such that $\|\psi(t)\|_{r}\geq C\left\langle t\right\rangle^{r},\quad\forall t\geq T\ .$ (2.9) We also prove a stronger result: namely not only $V(t)$ is a transporter, but also any operator sufficiently close to it (in the ${\mathcal{A}}_{0}$-topology). Here the precise statement: ###### Theorem 2.8. With the same assumptions of Theorem 2.7, there exist $\epsilon_{0}>0$ and ${\mathtt{M}}\in{\mathbb{N}}$ such that for any $W\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$, $W(t)$ selfadjoint $\forall t$, fulfilling $\sup_{t\in{\mathbb{T}}}\,\wp^{0}_{\mathtt{M}}(W(t))\leq\epsilon_{0},$ (2.10) then $V(t)+W(t)$ is a transporter for the equation ${\rm i}\partial_{t}\psi=\big{(}K_{0}+V(t)+W(t)\big{)}\psi\ .$ (2.11) More precisely, for any $r>0$ there exist a solution $\psi(t)$ in ${\mathcal{H}}^{r}$ of (2.11) and constants $C,T>0$ such that $\|\psi(t)\|_{r}\geq C\left\langle t\right\rangle^{r},\quad\forall t\geq T\ .$ (2.12) Let us comment the above results. 1. 1. The growth of Sobolev norms of Theorem 2.7 is truly an energy cascade phenomenon; indeed the ${\mathcal{H}}^{0}$-norm of any solution of (2.8) is preserved for all times, $\|\psi(t)\|_{0}=\|\psi(0)\|_{0}$, $\,\forall t\in{\mathbb{R}}$. This is due to the selfadjointness of $K_{0}+V(t)$ (the same happens to solutions of (2.11)). 2. 2. Estimates (2.9), (2.12) provide optimal lower bounds for the speed of growth of the Sobolev norms. Indeed we proved [33] that, under the assumptions above555 in particular the fact that $[K_{0},V(t)]$ and $[K_{0},V(t)+W(t)]$ are uniformly (in $t$) bounded operators on the scale ${\mathcal{H}}^{r}$, any solution of (2.8) or (2.11) fulfills the upper bounds $\forall r>0\ \ \ \exists\,{\widetilde{C}}_{r}>0\colon\quad\|\psi(t)\|_{r}\leq{\widetilde{C}}_{r}\left\langle t\right\rangle^{r}\,\|\psi(0)\|_{r}.$ Thus, Theorems 2.7, 2.8 construct unbounded solutions with optimal growth. 3. 3. Theorem 2.8 proves robustness of certain type of transporters under small pseudodifferential perturbations. This shows a sort of “stability of instability”, which, up to our knowledge, is new in this context. 4. 4. Actually there are infinitely many distinct solutions undergoing growth of Sobolev norms. Their initial data are constructed in a unique way starting from functions belonging to the absolutely continuous spectral subspace of the operator $\left\langle V\right\rangle$. We describe such initial data in Corollary 3.16. 5. 5. Energy cascade is a resonant phenomenon; here it happens because $V(t)$ oscillates at frequency $\omega=1$ which resonates with the spectral gaps of $K_{0}$. In [5] we proved that if $V(t)\equiv\mathsf{V}(\omega t)$ is quasiperiodic in time with a frequency vector $\omega\in{\mathbb{R}}^{n}$ fulfilling the non-resonant condition $\exists\gamma,\tau>0\colon\quad\left|\ell+\omega\cdot k\right|\geq\frac{\gamma}{\left\langle k\right\rangle^{\tau}}\quad\forall\ell,k\in{\mathbb{Z}}\times{\mathbb{Z}}^{n}\setminus\\{0\\}$ (which is violated if $V(t)$ is $2\pi$-periodic) then the Sobolev norms grow at most as $\left\langle t\right\rangle^{\epsilon}$ $\forall\epsilon>0$. The $\left\langle t\right\rangle^{\epsilon}$-speed of growth is also known for systems with increasing [37, 33, 5] or shrinking [21, 35] spectral gaps and for Schrödinger equation on ${\mathbb{T}}^{d}$ with bounded [10, 16, 8] and even unbounded [7] potentials. 6. 6. In concrete models one can typically prove that if $V(t)$ is sufficiently small in size and oscillates in time with a strongly non resonant frequency $\omega$ (typically belonging to some Cantor set of large measure), then all solutions have uniformly in time bounded Sobolev norms. Therefore the stability/instability of the system depends only on the resonance property of the frequency $\omega$. We mention just the recent results [4, 6] which deal with the harmonic oscillator (as we consider it in the applications) and refer to those papers for a complete bibliography. 7. 7. The most delicate assumption to verify is (2.7). In the applications, one can try to construct an escape function for the principal symbol $\left\langle v\right\rangle$ of $\left\langle V\right\rangle$. This means to find a symbol $a(x,\xi)$ of order 1 such that the Poisson bracket $\\{\left\langle v\right\rangle,a\\}$ is strictly positive in some energy levels: $\exists c>0\colon\quad\\{\left\langle v\right\rangle,a\\}\geq c\qquad\mbox{ in }\\{(x,\xi):\ \ \left|\left\langle v\right\rangle(x,\xi)-\lambda\right|\leq\delta\\}\ .$ Then symbolic calculus and sharp Gårding inequality imply that (2.7) holds in the interval $I=(\lambda-\delta,\lambda+\delta)$; see [13] Section 6.2 for details. We finally note that the second theorem is stronger than the first one and implies it in the special case $W(t)\equiv 0$. However we think that the statement of Theorem 2.7 is clear and useful in the applications (see e.g. Section 4), so we decided to state it on its own. Having said so, in the sequel we shall only prove Theorem 2.8. ## 3 Proof of the abstract result As already mentioned, we shall only prove Theorem 2.8. The proof is divided in three steps; in the first one we put system (2.11) in its resonant pseudodifferential normal form. In the second one we analyze the dynamics of the effective Hamiltonian and prove the existence of solutions with decaying negative Sobolev norms. The final step is to construct a solution of the complete equation exhibiting growth of Sobolev norms. ### 3.1 Resonant pseudodifferential normal form The goal of this section is to put system (2.11) into its resonant pseudodifferential normal form up to an arbitrary $N$-smoothing operator. In this first step we shall only require Assumptions I and II. It is slightly more convenient to deal with the equation ${\rm i}\partial_{t}\psi=\big{(}K_{0}+\mathsf{V}(t)\big{)}\psi$ (3.1) and then to specify the result for $\mathsf{V}(t)=V(t)+W(t)$ as in (2.11). Given $\mathsf{V}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{m})$, $m\in{\mathbb{R}}$, we define the averaged operator $\displaystyle{\widehat{\mathsf{V}}}(t):=\frac{1}{2\pi}\int_{0}^{2\pi}e^{{\rm i}sK_{0}}\,\mathsf{V}(t+s)\,e^{-{\rm i}sK_{0}}\,{\rm d}s\ .$ (3.2) We shall prove below that ${\widehat{\mathsf{V}}}(t)\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{m})$ (see Lemma 3.2). ###### Proposition 3.1 (Resonant pseudodifferential normal form). Consider equation (3.1) with $\mathsf{V}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$, $\mathsf{V}(t)$ selfadjoint $\forall t$. There exists a sequence $\\{X_{j}(t)\\}_{j\geq 1}$ of selfadjoint (time-dependent) operators in ${\mathcal{H}}$ with $X_{j}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{1-j})$ and fulfilling $\forall r\in{\mathbb{R}},\ \exists c_{r,j},C_{r,j}>0\colon\qquad c_{r,j}\|{\varphi}\|_{r}\leq\|e^{\pm{\rm i}X_{j}(t)}{\varphi}\|_{r}\leq C_{r,j}\|{\varphi}\|_{r},\qquad\forall t\in{\mathbb{R}},$ (3.3) such that the following holds true. For any $N\geq 1$, the change of variables $\psi=e^{-{\rm i}X_{1}(t)}\cdots e^{-{\rm i}X_{N}(t)}{\varphi}$ (3.4) transforms (3.1) into the equation ${\rm i}\partial_{t}{\varphi}=\big{(}K_{0}+Z^{(N)}(t)+\mathsf{V}^{(N)}(t)\big{)}{\varphi}\ ;$ (3.5) here $\mathsf{V}^{(N)}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{-N})$ whereas $Z^{(N)}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$, it is selfadjoint $\forall t$, it fulfills ${\rm i}\partial_{t}Z^{(N)}(t)=[K_{0},Z^{(N)}(t)]$ (3.6) and it has the expansion $Z^{(N)}(t)={\widehat{\mathsf{V}}}(t)+T^{(N)}(t),\qquad T^{(N)}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{-1})\ .$ (3.7) Here ${\widehat{\mathsf{V}}}(t)$ is the averaged operator defined in (3.2). In order to prove the proposition we start with some preliminary results. The first regards the properties of ${\widehat{\mathsf{V}}}(t)$. ###### Lemma 3.2. Let $\mathsf{V}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{m})$, $m\in{\mathbb{R}}$, $\mathsf{V}(t)$ selfadjoint $\forall t$. Then the following holds true. * (i) ${\widehat{\mathsf{V}}}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{m})$, it is selfadjoint $\forall t$, it commutes with ${\rm i}\partial_{t}-K_{0}$, i.e. ${\rm i}\partial_{t}{\widehat{\mathsf{V}}}(t)=[K_{0},{\widehat{\mathsf{V}}}(t)]$ and $\forall j,\ell\geq 0\quad\exists\,M\in{\mathbb{N}},\ C>0\ \ {\rm s.t.}\ \ \sup_{t\in{\mathbb{T}}}\wp^{m}_{j}(\partial_{t}^{\ell}{\widehat{\mathsf{V}}}(t))\leq C\,\sup_{t\in{\mathbb{T}}}\ \wp^{m}_{M}(\mathsf{V}(t))\ .$ (3.8) * (ii) The resonant averaged operator $\left\langle\mathsf{V}\right\rangle$, defined in (1.3), belongs to ${\mathcal{A}}_{m}$, it is selfadjoint and $\forall j\geq 0\quad\exists\,M\in{\mathbb{N}},\ C>0\ \ {\rm s.t.}\ \ \wp^{m}_{j}(\left\langle\mathsf{V}\right\rangle)\leq C\,\sup_{t\in{\mathbb{T}}}\ \wp^{m}_{M}(\mathsf{V}(t))\ .$ (3.9) * (iii) One has the chain of identities ${\widehat{\mathsf{V}}}(0)=\left\langle\mathsf{V}\right\rangle=e^{{\rm i}tK_{0}}\,{\widehat{\mathsf{V}}}(t)\,e^{-{\rm i}tK_{0}}=\langle\,{\widehat{\mathsf{V}}}\,\rangle,\qquad\forall t\in{\mathbb{R}}\ .$ (3.10) ###### Proof. $(i)$ The properties ${\widehat{\mathsf{V}}}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{m})$ and ${\widehat{\mathsf{V}}}(t)$ selfadjoint $\forall t$ follow from Assumption II and the fact that $\mathsf{V}(t)$ is $2\pi$-periodic in $t$ and selfadjoint $\forall t$. Let us prove it commutes with ${\rm i}\partial_{t}-K_{0}$. Using $\partial_{s}\left(e^{{\rm i}sK_{0}}\,\mathsf{V}(t+s)\,e^{-{\rm i}sK_{0}}\right)=e^{{\rm i}sK_{0}}\big{(}{\rm i}[K_{0},\mathsf{V}(t+s)]+\partial_{s}\mathsf{V}(t+s)\big{)}e^{-{\rm i}sK_{0}}$ we get $\displaystyle\partial_{t}{\widehat{\mathsf{V}}}(t)$ $\displaystyle=\frac{1}{2\pi}\int_{0}^{2\pi}e^{{\rm i}sK_{0}}\,\partial_{t}\mathsf{V}(t+s)\,e^{-{\rm i}sK_{0}}\,{\rm d}s=\frac{1}{2\pi}\int_{0}^{2\pi}e^{{\rm i}sK_{0}}\,\partial_{s}\mathsf{V}(t+s)\,e^{-{\rm i}sK_{0}}\,{\rm d}s$ $\displaystyle=\frac{1}{2\pi{\rm i}}\int_{0}^{2\pi}e^{{\rm i}sK_{0}}\,[K_{0},\mathsf{V}(t+s)]\,e^{-{\rm i}sK_{0}}\,{\rm d}s={\rm i}^{-1}\,[K_{0},{\widehat{\mathsf{V}}}(t)]$ where in the second line we used the periodicity of $s\mapsto e^{{\rm i}sK_{0}}\,\mathsf{V}(t+s)\,e^{-{\rm i}sK_{0}}$ (see Remark 2.5) to remove the boundary terms. Estimate (3.8) for $\ell=0$ follows from Assumption II. For $\ell\geq 1$ we use induction: assume (3.8) is true up to a certain $\ell$; using $\partial_{t}^{\ell+1}{\widehat{\mathsf{V}}}(t)=-{\rm i}\partial_{t}^{\ell}[K_{0},{\widehat{\mathsf{V}}}(t)]=-{\rm i}[K_{0},\partial_{t}^{\ell}{\widehat{\mathsf{V}}}(t)]$, we get $\forall j\in{\mathbb{N}}$ $\wp_{j}^{m}(\partial_{t}^{\ell+1}{\widehat{\mathsf{V}}}(t))\leq\ \wp_{j}^{m}([K_{0},\partial_{t}^{\ell}{\widehat{\mathsf{V}}}(t)])\leq C\wp_{j_{1}}^{m}(\partial_{t}^{\ell}{\widehat{\mathsf{V}}}(t))\leq C\wp_{j_{2}}^{m}(\mathsf{V}(t))$ using also the inductive assumption. This proves (3.8). $(ii)$ It is clear that $\left\langle\mathsf{V}\right\rangle$ is time independent, selfadjoint and in ${\mathcal{A}}_{m}$ by Assumption II. Estimate (3.9) follows from Assumption II. $(iii)$ Clearly ${\widehat{\mathsf{V}}}(0)=\left\langle\mathsf{V}\right\rangle$. Then, as the map $\tau\mapsto e^{{\rm i}\tau K_{0}}\,\mathsf{V}(\tau)\,e^{-{\rm i}\tau K_{0}}$ is $2\pi$-periodic, one has $\forall t\in{\mathbb{R}}$ $e^{{\rm i}tK_{0}}\,{\widehat{\mathsf{V}}}(t)\,e^{-{\rm i}tK_{0}}=\frac{1}{2\pi}\int_{0}^{2\pi}e^{{\rm i}(t+s)K_{0}}\,\mathsf{V}(t+s)\,e^{-{\rm i}(s+t)K_{0}}\,{\rm d}s=\left\langle\mathsf{V}\right\rangle\ .$ Finally, exploiting this last identity, one has $\langle\,{\widehat{\mathsf{V}}}\,\rangle=\frac{1}{2\pi}\int_{0}^{2\pi}e^{{\rm i}tK_{0}}\,{\widehat{\mathsf{V}}}(t)\,e^{-{\rm i}tK_{0}}{\rm d}t=\frac{1}{2\pi}\int_{0}^{2\pi}\left\langle\mathsf{V}\right\rangle{\rm d}t=\left\langle\mathsf{V}\right\rangle$ which completes the proof of (3.10). ∎ The second preliminary result regards how to solve the homological equations which appear during the normal form procedure. More precisely we look for a time periodic operator $X(t)$ solving the homological equation $\partial_{t}X(t)+{\rm i}[K_{0},X(t)]=\mathsf{V}(t)-{\widehat{\mathsf{V}}}(t),$ (3.11) where ${\widehat{\mathsf{V}}}(t)$ is the averaged operator defined in (3.2). This is done in the next lemma. ###### Lemma 3.3. Let $\mathsf{V}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{m})$, $m\in{\mathbb{R}}$, $\mathsf{V}(t)$ selfadjoint $\forall t$. The homological equation (3.11) has a solution $X\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{m})$ and $X(t)$ is selfadjoint $\forall t$. ###### Proof. We look for a solution of (3.11) using the method of variation of constants. In particular we take $X(t)=e^{-{\rm i}tK_{0}}\,Y(t)\,e^{{\rm i}tK_{0}}$ for some $Y\in C^{\infty}({\mathbb{R}},{\mathcal{A}}_{m})$ with $Y(0)=0$ to be determined. Then $X$ solves (3.11) provided $\partial_{t}Y(t)=e^{{\rm i}tK_{0}}\,(\mathsf{V}(t)-{\widehat{\mathsf{V}}}(t))\,e^{-{\rm i}tK_{0}}$, giving $Y(t)=\int_{0}^{t}e^{{\rm i}sK_{0}}\big{(}\mathsf{V}(s)-{\widehat{\mathsf{V}}}(s)\big{)}\,e^{-{\rm i}sK_{0}}\,{\rm d}s.$ By Lemma 3.2 and Assumption II, $Y\in C^{\infty}({\mathbb{R}},{\mathcal{A}}_{m})$ and it is selfadjoint $\forall t$. Therefore one gets $X(t)=\int_{0}^{t}e^{{\rm i}(s-t)K_{0}}\big{(}\mathsf{V}(s)-{\widehat{\mathsf{V}}}(s)\big{)}\,e^{-{\rm i}(s-t)K_{0}}\,{\rm d}s.$ Again $X\in C^{\infty}({\mathbb{R}},{\mathcal{A}}_{m})$ and it is selfadjoint $\forall t$. Finally (recall Remark 2.5) $\displaystyle X(t+2\pi)-X(t)$ $\displaystyle=\int_{t}^{t+2\pi}e^{{\rm i}(s-t)K_{0}}\,\big{(}\mathsf{V}(s)-{\widehat{\mathsf{V}}}(s)\big{)}\,e^{-{\rm i}(s-t)K_{0}}\,{\rm d}s$ $\displaystyle=e^{-{\rm i}tK_{0}}\int_{0}^{2\pi}e^{{\rm i}sK_{0}}\,\big{(}\mathsf{V}(s)-{\widehat{\mathsf{V}}}(s)\big{)}\,e^{-{\rm i}sK_{0}}\,{\rm d}s\ e^{{\rm i}tK_{0}}$ $\displaystyle=2\pi e^{-{\rm i}tK_{0}}\big{(}\left\langle\mathsf{V}\right\rangle-\langle\,{\widehat{\mathsf{V}}}\,\rangle\big{)}e^{{\rm i}tK_{0}}\stackrel{{\scriptstyle\eqref{av.W}}}{{=}}0$ which proves the periodicity of $t\mapsto X(t)$. ∎ We are ready to prove Proposition 3.1. During the proof we shall use some results proved in [5] about the flow generated by pseudodifferential operators; we collect them, for the reader’s convenience, in Appendix A. ###### Proof of Proposition 3.1. The proof is inductive on $N$. Let us start with $N=1$. We look for a change of variables of the form $\psi=e^{-{\rm i}X_{1}(t)}{\varphi}$ where $X_{1}(t)\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$ is selfadjoint $\forall t$, to be determined. By Lemma A.1, ${\varphi}$ fulfills the Schrödinger equation ${\rm i}\partial_{t}{\varphi}=H^{+}(t){\varphi}$ with $\displaystyle H^{+}(t)$ $\displaystyle:=e^{{\rm i}X_{1}(t)}\,\big{(}K_{0}+\mathsf{V}(t)\big{)}\,e^{-{\rm i}X_{1}(t)}-\int_{0}^{1}e^{{\rm i}sX_{1}(t)}\,(\partial_{t}X_{1}(t))\,e^{-{\rm i}sX_{1}(t)}\ {\rm d}s\ .$ Then a commutator expansion, see Lemma A.2, gives $\displaystyle H^{+}(t)$ $\displaystyle=K_{0}+{\rm i}[X_{1}(t),K_{0}]+\mathsf{V}(t)-\partial_{t}X_{1}+\mathsf{V}^{(1)}(t)$ with $\mathsf{V}^{(1)}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{-1})$, selfadjoint $\forall t$. By Lemma 3.3, we choose $X_{1}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$, selfadjoint $\forall t$, s.t. ${\rm i}[K_{0},X_{1}(t)]+\partial_{t}X_{1}(t)=\mathsf{V}(t)-{\widehat{\mathsf{V}}}(t)\ ,$ (3.12) where ${\widehat{\mathsf{V}}}(t)$ is the averaged operator (see (3.2)). With this choice we have $\displaystyle H^{+}(t)$ $\displaystyle=K_{0}+Z^{(1)}(t)+\mathsf{V}^{(1)}(t)\ ,\quad Z^{(1)}(t):={\widehat{\mathsf{V}}}(t)\ .$ (3.13) By Lemma 3.2, $Z^{(1)}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$, it is selfadjoint $\forall t$, it commutes with ${\rm i}\partial_{t}-K_{0}$. The map $e^{-{\rm i}X_{1}(t)}$ fulfills (3.3) thanks to Lemma A.3. This concludes the first step. The iterative step $N\to N+1$ is proved following the same lines, just adding the remark that $e^{{\rm i}X_{N+1}}Z^{(N)}e^{-{\rm i}X_{N+1}}-Z^{(N)}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{-N-1})$, and solving the homological equation ${\rm i}[K_{0},X_{N+1}(t)]+\partial_{t}X_{N+1}(t)=\mathsf{V}^{(N)}(t)-{\widehat{\mathsf{V}^{(N)}}}(t)\ .$ (3.14) So one puts $Z^{(N+1)}:=Z^{(N)}+{\widehat{\mathsf{V}^{(N)}}}$. Note that ${\widehat{\mathsf{V}^{(N)}}}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{-N})$, so $Z^{(N)}$ has an expansion in operators of decreasing order. ∎ It turns out that property (3.6) implies that $e^{{\rm i}tK_{0}}\,Z^{(N)}(t)\,e^{-{\rm i}tK_{0}}$ is time independent. A consequence of this fact is the following corollary. ###### Corollary 3.4. Consider equation (3.1) with $\mathsf{V}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$, $\mathsf{V}(t)$ selfadjoint $\forall t$. Fix $N\in{\mathbb{N}}$ arbitrary. There exists a change of coordinates ${\mathcal{U}}_{N}(t)$ unitary in ${\mathcal{H}}$ and fulfilling $\forall r\geq 0\quad\exists c_{r},C_{r}>0\colon\qquad c_{r}\|{\varphi}\|_{r}\leq\|{\mathcal{U}}_{N}(t)^{\pm}{\varphi}\|_{r}\leq C_{r}\|{\varphi}\|_{r},\qquad\forall t\in{\mathbb{R}},$ (3.15) such that $\psi(t)$ is a solution of (3.1) if and only if $\phi(t):={\mathcal{U}}_{N}(t)\psi(t)$ solves ${\rm i}\partial_{t}\phi=\big{(}\left\langle\mathsf{V}\right\rangle+T_{N}+R_{N}(t)\big{)}\phi\ ;$ (3.16) here $\left\langle\mathsf{V}\right\rangle$ is the resonant average of $\mathsf{V}$ (see (1.3)), $T_{N}\in{\mathcal{A}}_{-1}$ is time independent and selfadjoint and $R_{N}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{-N})$. ###### Proof. Fix $N\in{\mathbb{N}}$ and apply Proposition 3.1 to conjugate equation (3.1) to the form (3.5) via the change of variables (3.4). Then we gauge away $K_{0}$ by the change of coordinates ${\varphi}=e^{-{\rm i}tK_{0}}\phi$, getting ${\rm i}\partial_{t}\phi=e^{{\rm i}tK_{0}}\,\big{(}Z^{(N)}(t)+\mathsf{V}^{(N)}(t)\big{)}\,e^{-{\rm i}tK_{0}}\,\phi.$ Define $\mathsf{H}_{N}:=e^{{\rm i}tK_{0}}\,Z^{(N)}(t)\,e^{-{\rm i}tK_{0}},\qquad R_{N}(t):=e^{{\rm i}tK_{0}}\,\mathsf{V}^{(N)}(t)\,e^{-{\rm i}tK_{0}}.$ The operator $R_{N}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{-N})$ by Assumption II since $\mathsf{V}^{(N)}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{-N})$. Let us now prove that $\mathsf{H}_{N}$ is time independent. We know by Lemma 3.1 that $Z^{(N)}(t)$ commutes with ${\rm i}\partial_{t}-K_{0}$; therefore $\partial_{t}\big{(}e^{{\rm i}tK_{0}}\,Z^{(N)}(t)\,e^{-{\rm i}tK_{0}}\big{)}=e^{{\rm i}tK_{0}}\,\big{(}{\rm i}[K_{0},Z^{(N)}(t)]+\partial_{t}Z^{(N)}(t)\big{)}\,e^{-{\rm i}tK_{0}}=0$ and we get $\mathsf{H}_{N}=e^{{\rm i}tK_{0}}\,Z^{(N)}(t)\,e^{-{\rm i}tK_{0}}|_{t=0}=Z^{(N)}(0)\stackrel{{\scriptstyle\eqref{Z.exp}}}{{=}}{\widehat{\mathsf{V}}}(0)+T^{(N)}(0)\stackrel{{\scriptstyle\eqref{av.W}}}{{=}}\left\langle\mathsf{V}\right\rangle+T^{(N)}(0).$ So we put $T_{N}:=T^{(N)}(0)$; clearly it belongs to ${\mathcal{A}}_{-1}$, it is selfadjoint and time independent. Finally we put ${\mathcal{U}}_{N}(t):=e^{{\rm i}tK_{0}}\,e^{{\rm i}tX_{N}(t)}\,\cdots e^{{\rm i}tX_{1}(t)}$; estimate (3.15) follows from (3.3) and Remark 2.1. ∎ Coming back to the original equation (2.11), we apply Corollary 3.4 with $\mathsf{V}=V+W\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$, getting the following result: ###### Corollary 3.5. With the same assumptions of Theorem 2.8, the following holds true. Fix $N\in{\mathbb{N}}$ arbitrary. There exists a change of coordinates ${\mathcal{U}}_{N}(t)$, unitary in ${\mathcal{H}}$ and fulfilling (3.15) such that $\psi(t)$ is a solution of (2.11) if and only if $\phi(t):={\mathcal{U}}_{N}(t)\psi(t)$ solves ${\rm i}\partial_{t}\phi=\big{(}\left\langle V\right\rangle+\left\langle W\right\rangle+T_{N}+R_{N}(t)\big{)}\phi$ (3.17) where $T_{N}\in{\mathcal{A}}_{-1}$ is selfadjoint and time independent whereas $R_{N}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{-N})$. ### 3.2 Local energy decay estimates From now on we are going to assume also Assumption III. In the previous section we have conjugated the original equation (2.11) to the resonant equation (3.17). In this section we consider the effective equation obtained removing $R_{N}(t)$ from (3.17), namely ${\rm i}\partial_{t}{\varphi}=H_{N}{\varphi},\qquad H_{N}:=\left\langle V\right\rangle+\left\langle W\right\rangle+T_{N},$ (3.18) with $T_{N}\in{\mathcal{A}}_{-1}$ of Corollary 3.5. Note that $H_{N}$ is selfadjoint by Lemma 3.2 and Corollary 3.5. The goal is to construct a solution of (3.18) with polynomially in time growing Sobolev norms. Actually we will prove the following slightly stronger result, namely the existence of a solution with decaying negative Sobolev norms: ###### Proposition 3.6 (Decay of negative Sobolev norms). With the same assumptions of Theorem 2.8, consider the operator $H_{N}$ in (3.18). For any $k\in{\mathbb{N}}$, there exist a nontrivial solution ${\varphi}(t)\in{\mathcal{H}}^{k}$ of (3.18) and $\forall r\in[0,k]$ a constant $C_{r}>0$ such that $\|{\varphi}(t)\|_{{-r}}\leq C_{r}\left\langle t\right\rangle^{-r}\,\|{\varphi}(0)\|_{r}\ ,\qquad\forall t\in{\mathbb{R}}\ .$ (3.19) ###### Remark 3.7. As $H_{N}$ is selfadjoint, the conservation of the ${\mathcal{H}}^{0}$-norm and Cauchy-Schwartz inequality give $\|{\varphi}(0)\|_{0}^{2}=\|{\varphi}(t)\|_{0}^{2}\leq\|{\varphi}(t)\|_{r}\ \|{\varphi}(t)\|_{{-r}}\ ,\qquad\forall t\in{\mathbb{R}}\ ,$ so that (3.19) implies the growth of positive Sobolev norms: $\|{\varphi}(t)\|_{r}\geq\frac{1}{C_{r}}\frac{\|{\varphi}(0)\|_{0}^{2}}{\|{\varphi}(0)\|_{r}}\,\left\langle t\right\rangle^{r}\ ,\quad\forall t\in{\mathbb{R}}\ .$ The rest of the section is devoted to the proof of Proposition 3.6. As we shall see, it follows from a local energy decay estimate for the operator $H_{N}$, namely a dispersive estimate of the form $\|\left\langle A\right\rangle^{-k}\,e^{-{\rm i}H_{N}t}\,g_{J}(H_{N})\,{\varphi}\|_{0}\leq C_{k}\left\langle t\right\rangle^{-k}\|\left\langle A\right\rangle^{k}g_{J}(H_{N}){\varphi}\|_{0}\ ,\qquad\forall t\in{\mathbb{R}}$ (3.20) where $A\in{\mathcal{A}}_{1}$, $J\subset I_{0}$ is an interval and $g_{J}\in C^{\infty}_{c}({\mathbb{R}},{\mathbb{R}}_{\geq 0})$ with $g_{J}\equiv 1$ on $J$. ###### Remark 3.8. Actually estimate (3.20) show the existence of infinitely many solutions of (3.18) with decaying negative Sobolev norms. In particular this happens to any solution whose initial datum ${\varphi}(0)$ belongs to the (infinite dimensional) set ${\rm Ran}\,E_{J}(H_{N})$, where $E_{J}(H_{N})$ is the spectral projection of $H_{N}$ corresponding to the interval $J$. A possible approach (which we will follow here) to obtain such estimate is via Sigal-Soffer minimal velocity estimates [42, 23, 31, 30, 24, 2]. These estimates are based on Mourre theory, let us recall this last one. #### Mourre theory. Let $\mathsf{H}$ be a selfadjoint operator on the Hilbert space ${\mathcal{H}}$, and denote by $\sigma(\mathsf{H})$ its spectrum. We further denote by $\sigma_{d}(\mathsf{H})$ its discrete spectrum, $\sigma_{ess}(\mathsf{H})$ its essential spectrum, $\sigma_{pp}(\mathsf{H})$ its pure point spectrum, $\sigma_{ac}(\mathsf{H})$ its absolutely continuous spectrum and $\sigma_{sc}(\mathsf{H})$ its singular spectrum; see e.g. [38] pag. 236 and 231 for their definitions. Furthermore we denote by $E_{\Omega}(\mathsf{H})$ the spectral projection of $\mathsf{H}$ corresponding to the Borel set $\Omega$ and by $m_{\varphi}(\Omega):=\left\langle E_{\Omega}(\mathsf{H}){\varphi},{\varphi}\right\rangle$ the spectral measure associated to ${\varphi}\in{\mathcal{H}}$. Assume a selfadjoint operator $\mathsf{A}$ can be found such that $D(\mathsf{A})\cap{\mathcal{H}}$ is dense in ${\mathcal{H}}$. We put ${\rm ad}^{0}_{\mathsf{A}}(\mathsf{H}):=\mathsf{H},\qquad{\rm ad}_{\mathsf{A}}(\mathsf{H}):=[\mathsf{H},\mathsf{A}],\qquad{\rm ad}^{n}_{\mathsf{A}}(\mathsf{H}):=[{\rm ad}^{n-1}_{\mathsf{A}}(\mathsf{H}),\mathsf{A}],\quad\forall n\geq 2\ .$ (3.21) Consider the following properties: * (M1) For some ${\mathtt{N}}\geq 1$, the operators ${\rm ad}^{n}_{\mathsf{A}}(\mathsf{H})$ with $n=1,\ldots,{\mathtt{N}}$, can all be extended to bounded operators on ${\mathcal{H}}$. * (M2) Mourre estimate: there exists an open interval $I\subset{\mathbb{R}}$ with compact closure and a function $g_{I}\in C^{\infty}_{c}({\mathbb{R}},{\mathbb{R}}_{\geq 0})$ with $g_{I}\equiv 1$ on $I$ such that $g_{I}(\mathsf{H})\,{\rm i}[\mathsf{H},\mathsf{A}]\,g_{I}(\mathsf{H})\geq\theta g_{I}(\mathsf{H})^{2}+\mathsf{K}$ (3.22) for some $\theta>0$ and $\mathsf{K}$ a selfadjoint compact operator on ${\mathcal{H}}$. If the estimate (3.22) holds true with $\mathsf{K}=0$, we shall say that $\mathsf{H}$ fulfills a strict Mourre estimate. Mourre theorem [36] says the following: ###### Theorem 3.9 (Mourre). Assume conditions (M1) – (M2) with ${\mathtt{N}}=2$. In the interval $I$, the operator $\mathsf{H}$ can have only absolutely continuous spectrum and finitely many eigenvalues of finite multiplicity. If $\mathsf{K}=0$, there are no eigenvalues in the interval $I$, i.e. $\sigma(\mathsf{H})\cap I=\sigma_{ac}(\mathsf{H})\cap I$. ###### Remark 3.10. The version stated here of Mourre theorem is taken from [3, Lemma 5.6] and [14, Theorem 4.7 – 4.9], and it has slightly weaker assumptions compared to [36]. ###### Remark 3.11. Mourre theorem guarantees that $\sigma_{sc}(\mathsf{H})\cap I=\emptyset$ and, in case $\mathsf{K}=0$, $\sigma_{pp}(\mathsf{H})\cap I=\emptyset$. However it does not guarantee that $\sigma(\mathsf{H})\cap I\neq\emptyset$; in our case we shall verify this property explicitly. The key point is that if $H_{N}$ fulfills a strict Mourre estimate (namely with $\mathsf{K}=0$) then one can prove a local energy decay estimate like (3.20) for the Schrödinger flow of $H_{N}$. This is a quite general fact which follows exploiting minimal velocity estimates [30] and we prove it for completeness in Appendix C. So the next goal is to prove that $H_{N}$ satisfies a strict Mourre estimate over a certain interval $J\subset I_{0}$. During the proof we will use some standard results from functional calculus; we recall them in Appendix B. We shall also use the following lemma: ###### Lemma 3.12. Let $\mathsf{H}\in{\mathcal{L}}({\mathcal{H}})$ be selfadjoint. If $\lambda\in\sigma_{ac}(\mathsf{H})$, then $\forall\delta>0$ one has $\left|[\lambda-\delta,\lambda+\delta]\cap\sigma(\mathsf{H})\right|>0\ .$ ###### Proof. By contradiction, assume that $\exists\delta_{0}>0$ such that $\left|[\lambda-\delta_{0},\lambda+\delta_{0}]\cap\sigma(\mathsf{H})\right|=0$. As $\lambda\in\sigma_{ac}(\mathsf{H})$, there exists $f\in{\mathcal{H}}$ such that $E_{[\lambda-\delta_{0},\lambda+\delta_{0}]}(\mathsf{H})f\neq 0$ and the spectral measure $m_{f}=\left\langle E(\mathsf{H})f,f\right\rangle$ is absolutely continuous. Then $0=m_{f}([\lambda-\delta_{0},\lambda+\delta_{0}])=\left\langle E_{[\lambda-\delta_{0},\lambda+\delta_{0}]}(\mathsf{H})f,f\right\rangle=\|E_{[\lambda-\delta_{0},\lambda+\delta_{0}]}(\mathsf{H})f\|_{0}^{2}>0$ giving a contradiction. ∎ ###### Lemma 3.13. There exist $\epsilon_{0},{\mathtt{M}}>0$ such that, provided $W$ fulfills (2.10), the following holds true: * (i) There exists an interval $I\subset I_{0}$ such that $\left|I\cap\sigma(H_{N})\right|>0$. * (ii) $H_{N}$ fulfills a strict Mourre estimate over $I$: there exists a function $g_{I}\in C^{\infty}_{c}({\mathbb{R}},{\mathbb{R}}_{\geq 0})$ with ${\rm supp}\,g_{I}\subset I_{0}$, $g_{I}\equiv 1$ on $I$, and $\theta^{\prime}>0$ such that $g_{I}(H_{N})\,{\rm i}[H_{N},A]\,g_{I}(H_{N})\geq\theta^{\prime}g_{I}(H_{N})^{2}\ .$ (3.23) Here $I_{0}$ is the interval and $A$ is the operator of Assumption III. ###### Proof. During the proof we shall often use that for $\mathsf{A},\mathsf{B},\mathsf{C}\in{\mathcal{L}}({\mathcal{H}})$ and selfadjoints $\displaystyle\mathsf{A}\leq\mathsf{B}\ \ \Rightarrow\ \ \mathsf{C}\mathsf{A}\mathsf{C}\leq\mathsf{C}\mathsf{B}\mathsf{C},\qquad\|\mathsf{A}\|_{{\mathcal{L}}({\mathcal{H}})}\leq a\ \ \Rightarrow\ \ -a\leq\mathsf{A}\leq a\ .$ (3.24) To shorten notation, throughout the proof we shall put $H_{0}:=\left\langle V\right\rangle\ .$ We split the proof in several steps. Step 1: By Assumption III, $H_{0}$ fulfills a Mourre estimate over the interval $I_{0}$. The first step of the proof is to exhibit a subinterval $I_{1}\subset I_{0}$ containing only absolutely continuous spectrum of $H_{0}$, namely $\sigma(H_{0})\cap I_{1}=\sigma_{ac}(H_{0})\cap I_{1}\ ,\qquad\left|\sigma(H_{0})\cap I_{1}\right|>0\ ,$ (3.25) and over which $H_{0}$ fulfills a strict Mourre estimate: $\exists g_{I_{1}}\in C^{\infty}_{c}({\mathbb{R}},{\mathbb{R}}_{\geq 0})$, $g_{I_{1}}\equiv 1$ on $I_{1}$, ${\rm supp}\,g_{I_{1}}\subset I_{0}$, such that $g_{I_{1}}(H_{0})\,{\rm i}[H_{0},A]\,g_{I_{1}}(H_{0})\geq\frac{\theta}{2}g_{I_{1}}(H_{0})^{2}\ .$ (3.26) To prove this claim, first apply Mourre theorem to $H_{0}$ (note that (M1) and (M2) are verified $\forall{\mathtt{N}}\in{\mathbb{N}}$ by symbolic calculus and Assumption III), getting that $\sigma(H_{0})\cap I_{0}$ contains only finitely many eigenvalues with finite multiplicity and absolutely continuous spectrum. In particular $|\overline{\sigma_{pp}(H_{0})}\cap I_{0}|=0$ and by Assumption III $(i)$ it follows that $|\sigma_{ac}(H_{0})\cap I_{0}|=|\sigma(H_{0})\cap I_{0}|>0$. So we take $\lambda_{0}\in I_{0}\cap(\sigma_{ac}(H_{0})\setminus\sigma_{pp}(H_{0}))$ and a sufficiently small interval $I_{1}(\overline{\delta}):=(\lambda_{0}-\overline{\delta},\lambda_{0}+\overline{\delta})\subset I_{0}$, $\overline{\delta}>0$, which does not contain eigenvalues of $H_{0}$; this is possible as the eigenvalues of $H_{0}$ in $I_{0}$ are finite. Moreover by Lemma 3.12, $\left|\sigma(H_{0})\cap I_{1}(\delta)\right|>0$ for any $\delta>0$. Now take $\delta\in(0,\overline{\delta})$ and a function $g_{\delta}\in C^{\infty}_{c}({\mathbb{R}},{\mathbb{R}}_{\geq 0})$ with ${\rm supp}\,g_{\delta}\subset I_{1}(\delta)$ and $g_{\delta}=1$ on $I_{1}(\frac{\delta}{2})$. We claim that provided $\delta\in(0,\overline{\delta})$ is sufficiently small $\|g_{\delta}(H_{0})Kg_{\delta}(H_{0})\|_{{\mathcal{L}}({\mathcal{H}})}\leq\frac{\theta}{2}\ ,$ (3.27) where $\theta>0$ is the one of Assumption III. Indeed in $I_{1}(\overline{\delta})$ the spectrum of $H_{0}$ is absolutely continuous; this means that $\forall{\varphi}\in{\mathcal{H}}$, the vector ${\varphi}^{\prime}:=E_{I_{1}(\overline{\delta})}(H_{0}){\varphi}$ belongs to the absolutely continuous subspace of $H_{0}$, namely its spectral measure $m_{{\varphi}^{\prime}}$ is absolutely continuous w.r.t. the Lebesgue measure. Now, since for any ${\varphi}\in{\mathcal{H}}$ one has by functional calculus $g_{\delta}(H_{0})=g_{\delta}(H_{0})E_{I_{1}(\overline{\delta})}(H_{0})$, one has that $\|g_{\delta}(H_{0}){\varphi}\|_{0}^{2}=\|g_{\delta}(H_{0})E_{I_{1}(\overline{\delta})}(H_{0}){\varphi}\|_{0}^{2}=\int_{\mathbb{R}}g_{\delta}(\lambda)^{2}\,{\rm d}m_{{\varphi}^{\prime}}(\lambda)\to 0\ \ \ \mbox{ as }\delta\to 0$ by Lebesgue dominated convergence theorem. In particular $g_{\delta}(H_{0})\to 0$ strongly as $\delta\to 0$ and then, being $K$ compact, $g_{\delta}(H_{0})K\to 0$ uniformly as $\delta\to 0$ (see e.g. [1]). Therefore for $\delta\in(0,\overline{\delta})$ sufficiently small (3.27) holds true. Using the assumption (2.7), (3.27) and (3.24) we deduce that $\displaystyle g_{\delta}(H_{0})\,g_{I_{0}}(H_{0})\,{\rm i}[H_{0},A]\,g_{I_{0}}(H_{0})\,g_{\delta}(H_{0})$ $\displaystyle\geq\theta g_{\delta}(H_{0})\,g_{I_{0}}(H_{0})^{2}\,g_{\delta}(H_{0})-\frac{\theta}{2}\ ;$ next apply $g_{\frac{\delta}{2}}(H_{0})$ to the right and left of the previous inequality, use again (3.24) and the identity $g_{I_{0}}(H_{0})\,g_{\delta}(H_{0})\,g_{\frac{\delta}{2}}(H_{0})=g_{\frac{\delta}{2}}(H_{0})$ (which follows from $g_{I_{0}}\,g_{\delta}\,g_{\frac{\delta}{2}}=g_{\frac{\delta}{2}}$), to get the strict Mourre estimate $g_{I_{1}}(H_{0})\,{\rm i}[H_{0},A]\,g_{I_{1}}(H_{0})\geq\frac{\theta}{2}g_{I_{1}}(H_{0})^{2}$ (3.28) where $I_{1}:=I_{1}(\frac{\delta}{4})$ and $g_{I_{1}}:=g_{\frac{\delta}{2}}$ fulfills $g_{I_{1}}\equiv 1$ on $I_{1}$, ${\rm supp}\,g_{I_{1}}\subset I_{1}(\frac{\delta}{2})$. Clearly $I_{1}$ fulfills (3.25). Step 2: We shall prove that the selfadjoint operator $H_{\left\langle W\right\rangle}:=H_{0}+\left\langle W\right\rangle$ has a nontrivial spectrum in a subinterval $I_{2}\subseteq I_{1}$, and over this interval it fulfills the strict Mourre estimate $g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}\,{\rm i}[H_{\left\langle W\right\rangle},A]\,g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}\geq\frac{\theta}{4}g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}^{2}\ $ (3.29) for any $g_{I_{2}}\in C^{\infty}_{c}({\mathbb{R}},{\mathbb{R}}_{\geq 0})$ with ${\rm supp}\,g_{I_{2}}\subset I_{1}$, $g_{I_{2}}\equiv 1$ on $I_{2}$. To prove this, we exploit that $\left\langle W\right\rangle\in{\mathcal{A}}_{0}$ is a small bounded perturbation of $H_{0}$, fulfilling, by (2.1), (3.9) $\exists M_{0}\in{\mathbb{N}},C_{0}>0\colon\qquad\|\left\langle W\right\rangle\|_{{\mathcal{L}}({\mathcal{H}})}{\leq}\,C_{0}[W]_{M_{0}},$ (3.30) where we denoted $[W]_{M}:=\sup_{t\in{\mathbb{T}}}\,\wp^{0}_{M}(W(t))\ .$ First let us prove that $\sigma(H_{\left\langle W\right\rangle})\cap I_{1}\neq\emptyset$. Take again the same $\lambda_{0}\in\sigma(H_{0})\cap I_{1}$ as in the previous step. We claim that ${\rm dist}\big{(}\lambda_{0},\sigma(H_{\left\langle W\right\rangle})\big{)}\leq C_{0}\,[W]_{M_{0}}.$ (3.31) If $\lambda_{0}\in\sigma(H_{\left\langle W\right\rangle})$ this is trivial. So assume that $\lambda_{0}$ belongs to the resolvent set of $H_{\left\langle W\right\rangle}$. As $\lambda_{0}\in\sigma(H_{0})$, by Weyl criterion $\exists(f_{n})_{n\geq 1}\in{\mathcal{H}}$ with $\|f_{n}\|_{0}=1$ such that $\|(H_{0}-\lambda_{0})f_{n}\|_{0}\to 0$ as $n\to\infty$. Then $\forall n\geq 1$ $\displaystyle 1$ $\displaystyle=\|f_{n}\|_{0}=\|(H_{\left\langle W\right\rangle}-\lambda_{0})^{-1}\,(H_{\left\langle W\right\rangle}-\lambda_{0})f_{n}\|_{0}\leq\frac{1}{{\rm dist}\big{(}\lambda_{0},\sigma(H_{\left\langle W\right\rangle})\big{)}}\|(H_{\left\langle W\right\rangle}-\lambda_{0})f_{n}\|_{0}$ $\displaystyle\stackrel{{\scriptstyle\eqref{pm.s20}}}{{\leq}}\frac{1}{{\rm dist}\big{(}\lambda_{0},\sigma(H_{\left\langle W\right\rangle})\big{)}}\Big{(}\|(H_{0}-\lambda_{0})f_{n}\|_{0}+C_{0}[W]_{M_{0}}\Big{)}$ which proves (3.31) passing to the limit $n\to\infty$. Then, provided $[W]_{M_{0}}$ is sufficiently small, (3.31) implies that ${\rm dist}\big{(}\lambda_{0},\sigma(H_{\left\langle W\right\rangle})\big{)}<\delta/8$. From this we learn that (recall $I_{1}=(\lambda_{0}-\frac{\delta}{4},\lambda_{0}+\frac{\delta}{4})$) $\sigma(H_{\left\langle W\right\rangle})\cap I_{1}\neq\emptyset\ .$ (3.32) Next we prove the Mourre estimate (3.29); we shall work perturbatively from (3.26). First $\displaystyle g_{I_{1}}(H_{0})\,{\rm i}[H_{\left\langle W\right\rangle},A]\,g_{I_{1}}(H_{0})=g_{I_{1}}(H_{0})\,{\rm i}[H_{0},A]\,g_{I_{1}}(H_{0})+g_{I_{1}}(H_{0})\,{\rm i}[\left\langle W\right\rangle,A]\,g_{I_{1}}(H_{0});$ we bound the first term in the right hand side above from below using (3.26). Concerning the second term, we use $\displaystyle\exists\,M_{1}\in{\mathbb{N}},\,C_{1}>0\colon\quad\|{\rm i}[\left\langle W\right\rangle,A]\|_{{\mathcal{L}}({\mathcal{H}})}\leq C_{1}[W]_{M_{1}}\ $ (3.33) (by (2.1), (2.3), (3.9)) and the inequalities (3.24) to bound it from above getting $g_{I_{1}}(H_{0})\,{\rm i}[\left\langle W\right\rangle,A]\,g_{I_{1}}(H_{0})\geq- C_{1}\,[W]_{M_{1}}\,g_{I_{1}}(H_{0})^{2}\ .$ Therefore we find $\displaystyle g_{I_{1}}(H_{0})\,{\rm i}[H_{\left\langle W\right\rangle},A]\,g_{I_{1}}(H_{0})\geq\left(\frac{\theta}{2}-C_{1}[W]_{M_{1}}\right)\,g_{I_{1}}(H_{0})^{2}\ .$ (3.34) Take now an open interval $I_{2}\subset I_{1}$ such that $\sigma(H_{\left\langle W\right\rangle})\cap{I}_{2}\neq\emptyset$ (it is possible by (3.32)); take also $g_{I_{2}}\in C^{\infty}_{c}({\mathbb{R}},{\mathbb{R}}_{\geq 0})$ with ${\rm supp}\,g_{I_{2}}\subseteq I_{1}$ and $g_{I_{2}}\equiv 1$ on $I_{2}$; remark that $g_{I_{1}}g_{I_{2}}=g_{I_{2}}$. Now we wish to replace $g_{I_{1}}(H_{0})$ by $g_{I_{2}}(H_{\left\langle W\right\rangle})$ in (3.34), thus getting the claimed estimate (3.29). So write $\displaystyle g_{I_{2}}(H_{\left\langle W\right\rangle})\,{\rm i}[H_{\left\langle W\right\rangle},A]$ $\displaystyle\,g_{I_{2}}(H_{\left\langle W\right\rangle})=g_{I_{2}}(H_{\left\langle W\right\rangle})\,g_{I_{1}}(H_{\left\langle W\right\rangle})\,{\rm i}[H_{\left\langle W\right\rangle},A]\,g_{I_{1}}(H_{\left\langle W\right\rangle})\,g_{I_{2}}(H_{\left\langle W\right\rangle})$ $\displaystyle=g_{I_{2}}(H_{\left\langle W\right\rangle})\,g_{I_{1}}(H_{0})\,{\rm i}[H_{\left\langle W\right\rangle},A]\,g_{I_{1}}(H_{0})\,g_{I_{2}}(H_{\left\langle W\right\rangle})$ (3.35) $\displaystyle\ +g_{I_{2}}(H_{\left\langle W\right\rangle})\,\Big{(}\big{(}g_{I_{1}}(H_{\left\langle W\right\rangle})-g_{I_{1}}(H_{0})\big{)}\,{\rm i}[H_{\left\langle W\right\rangle},A]\,g_{I_{1}}(H_{0})$ (3.36) $\displaystyle\ +g_{I_{1}}(H_{\left\langle W\right\rangle})\,{\rm i}[H_{\left\langle W\right\rangle},A]\,\big{(}g_{I_{1}}(H_{\left\langle W\right\rangle})-g_{I_{1}}(H_{0})\big{)}\Big{)}\,g_{I_{2}}(H_{\left\langle W\right\rangle})$ (3.37) Again we estimate (3.35) from below and the other lines from above. First $\displaystyle\eqref{pm6}$ $\displaystyle\stackrel{{\scriptstyle\eqref{pm5}}}{{\geq}}\left(\frac{\theta}{2}-C_{1}[W]_{M_{1}}\right)\,g_{I_{2}}(H_{\left\langle W\right\rangle})\,g_{I_{1}}(H_{0})^{2}\,g_{I_{2}}(H_{\left\langle W\right\rangle})\ .$ (3.38) We still have to bound from below $g_{I_{2}}(H_{\left\langle W\right\rangle})\,g_{I_{1}}(H_{0})^{2}\,g_{I_{2}}(H_{\left\langle W\right\rangle})$. To proceed we use that $g_{I_{1}}(H_{\left\langle W\right\rangle})-g_{I_{1}}(H_{0})$ is small in size, being bounded, via Lemma B.6 and (3.30), by $\displaystyle\|g_{I_{1}}(H_{\left\langle W\right\rangle})-g_{I_{1}}(H_{0})\|_{{\mathcal{L}}({\mathcal{H}})}\leq C\,[W]_{M_{0}}\,.$ (3.39) So write $\displaystyle g_{I_{2}}(H_{\left\langle W\right\rangle})\,g_{I_{1}}(H_{0})^{2}\,g_{I_{2}}(H_{\left\langle W\right\rangle})=g_{I_{2}}(H_{\left\langle W\right\rangle})\,g_{I_{1}}(H_{\left\langle W\right\rangle})^{2}\,g_{I_{2}}(H_{\left\langle W\right\rangle})$ (3.40) $\displaystyle\ \ +g_{I_{2}}(H_{\left\langle W\right\rangle})\,\Big{(}g_{I_{1}}(H_{\left\langle W\right\rangle})\,(g_{I_{1}}(H_{0})-g_{I_{1}}(H_{\left\langle W\right\rangle}))\,+(g_{I_{1}}(H_{0})-g_{I_{1}}(H_{\left\langle W\right\rangle}))\,g_{I_{1}}(H_{0})\Big{)}\,g_{I_{2}}(H_{\left\langle W\right\rangle}).$ Therefore, using $g_{I_{1}}g_{I_{2}}=g_{I_{2}}$, estimates (3.39) and (3.24), we deduce $g_{I_{2}}(H_{\left\langle W\right\rangle})\,g_{I_{1}}(H_{0})^{2}\,g_{I_{2}}(H_{\left\langle W\right\rangle})\geq\big{(}1-C[W]_{M_{0}}\big{)}\,g_{I_{2}}(H_{\left\langle W\right\rangle})^{2}\ .$ Thus we can finally estimate line (3.35) from below using (3.38) and the previous estimate, concluding $\eqref{pm6}\geq\left(\frac{\theta}{2}-C_{1}[W]_{M_{1}}\right)\big{(}1-C[W]_{M_{0}}\big{)}\,g_{I_{2}}(H_{\left\langle W\right\rangle})^{2}.$ (3.41) Next consider lines (3.36), (3.37). We use the bound (see (3.33)) $\|[H_{\left\langle W\right\rangle},A]\|_{{\mathcal{L}}({\mathcal{H}}^{0})}\leq C\big{(}1+[W]_{M_{1}}\big{)}\ ,$ and (3.39) to get $\eqref{pm7}+\eqref{pm8}\geq-C\,[W]_{M_{0}}\,(1+[W]_{M_{1}})\,g_{I_{2}}(H_{\left\langle W\right\rangle})^{2}.$ (3.42) Putting together (3.41) and (3.42) we finally find $g_{I_{2}}(H_{\left\langle W\right\rangle})\,{\rm i}[H_{\left\langle W\right\rangle},A]\,g_{I_{2}}(H_{\left\langle W\right\rangle})\geq\left(\frac{\theta}{2}-C([W]_{M_{1}}+[W]_{M_{0}}+[W]_{M_{0}}\,[W]_{M_{1}})\right)\,g_{I_{2}}(H_{\left\langle W\right\rangle})^{2}.$ Thus, provided (2.10) holds true for ${\mathtt{M}}$ sufficiently large and $\epsilon_{0}$ sufficiently small, the strict Mourre estimate (3.29) follows. Mourre theorem implies that the spectrum of $H_{\left\langle W\right\rangle}$ in $I_{2}$ is absolutely continuous and by (3.32) it is also nonempty; summarizing (use also Lemma 3.12) $\sigma(H_{\left\langle W\right\rangle})\cap I_{2}=\sigma_{ac}(H_{\left\langle W\right\rangle})\cap I_{2}\qquad\mbox{and}\qquad\left|\sigma(H_{\left\langle W\right\rangle})\cap I_{2}\right|>0\ .$ (3.43) Step 3: The last step is to consider the operator $H_{N}=H_{0}+\left\langle W\right\rangle+T_{N}=H_{\left\langle W\right\rangle}+T_{N}$, which, for the remaining part of the proof, we shall denote just by $H$. We shall constantly use that any pseudodifferential operator of strictly negative order is a compact operator on ${\mathcal{H}}$ (see Remark 2.4); in particular $T_{N}\in{\mathcal{A}}_{-1}$ is compact. We begin by proving that $\left|\sigma(H)\cap I_{2}\right|>0\ .$ (3.44) Indeed by Weyl theorem $\sigma_{ess}(H)=\sigma_{ess}(H_{\left\langle W\right\rangle})$ and therefore $\displaystyle\sigma(H)\cap I_{2}$ $\displaystyle\supset\sigma_{ess}(H)\cap I_{2}=\sigma_{ess}(H_{\left\langle W\right\rangle})\cap I_{2}=\sigma(H_{\left\langle W\right\rangle})\cap I_{2}\ ,$ since $\sigma_{d}(H_{\left\langle W\right\rangle})\cap I_{2}=\emptyset$ having $H_{\left\langle W\right\rangle}$ no eigenvalues in $I_{2}$. Then (3.44) follows by (3.43). Next we prove that $H$ fulfills a Mourre estimate over $I_{2}$, i.e. $g_{I_{2}}\big{(}H\big{)}\,{\rm i}[H,A]\,g_{I_{2}}\big{(}H\big{)}\geq\frac{\theta}{4}g_{I_{2}}\big{(}H\big{)}^{2}+K$ (3.45) with $K$ a compact operator. We work perturbatively from (3.29). Again first we compute $g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}\,{\rm i}[H,A]\,g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}=g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}\,{\rm i}[H_{\left\langle W\right\rangle},A]\,g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}+g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}\,{\rm i}[T_{N},A]\,g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}\ ;$ we estimate the first term in the r.h.s. above by (3.29), whereas the second term is a compact operator since $[T_{N},A]\in{\mathcal{A}}_{-1}$. We obtain $g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}\,{\rm i}[H,A]\,g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}\geq\frac{\theta}{4}g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}^{2}+K_{1}$ (3.46) with $K_{1}$ a compact operator. Now we must replace $g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}$ with $g_{I_{2}}(H)$. We write $\displaystyle g_{I_{2}}(H)\,{\rm i}[H,A]\,g_{I_{2}}(H)=g_{I_{2}}(H_{\left\langle W\right\rangle})\,{\rm i}[H,A]\,g_{I_{2}}(H_{\left\langle W\right\rangle})$ (3.47) $\displaystyle\quad+\big{(}g_{I_{2}}(H)-g_{I_{2}}(H_{\left\langle W\right\rangle})\big{)}\,{\rm i}[H,A]\,g_{I_{2}}(H_{\left\langle W\right\rangle})+g_{I_{2}}(H)\,{\rm i}[H,A]\,\big{(}g_{I_{2}}(H)-g_{I_{2}}(H_{\left\langle W\right\rangle})\big{)}$ (3.48) This time we use that $g_{I_{2}}(H)-g_{I_{2}}(H_{\left\langle W\right\rangle})$ is a compact operator, see Lemma B.6. Thus $\displaystyle\eqref{pm600}$ $\displaystyle\stackrel{{\scriptstyle\eqref{pm.12}}}{{\geq}}\frac{\theta}{4}\,g_{I_{2}}\big{(}H_{\left\langle W\right\rangle}\big{)}^{2}+K_{1}=\frac{\theta}{4}g_{I_{2}}(H)^{2}+K_{2}$ where $K_{1}$, $K_{2}$ are compact operators. Similarly, using that ${\rm i}[H,A]\in{\mathcal{A}}_{0}$ is a bounded operator, we deduce that (3.48) is a compact operator. Estimate (3.45) follows. In particular $H$ is conjugated to $A$ over the interval $I_{2}$ fulfilling (3.43). Proceeding as in Step 1, we produce a subinterval $I\subset I_{2}$ such that $\left|I\cap\sigma(H)\right|>0\ ,\qquad I\cap\sigma(H)=I\cap\sigma_{ac}(H)$ and over which $H$ fulfills the strict Mourre estimate (3.23). ∎ The previous result has proved the existence of an interval $I$ over which $H_{N}$ fulfills a strict Mourre estimate. This implies that $H_{N}$ fulfills dispersive estimates in the form of local energy decay. In the literature there are various variants of this result, thus in Appendix C we state and prove the one we apply here. ###### Corollary 3.14. Fix $k\in{\mathbb{N}}$. For any interval $J\subset I$, any function $g_{J}\in C^{\infty}_{c}({\mathbb{R}},{\mathbb{R}}_{\geq 0})$ with ${\rm supp}\,g_{J}\subset I$, $g_{J}\equiv 1$ on $J$, there exists a constant $C_{k}>0$ such that $\|\left\langle A\right\rangle^{-k}\,e^{-{\rm i}H_{N}t}\,g_{J}(H_{N})\,{\varphi}\|_{0}\leq C_{k}\left\langle t\right\rangle^{-k}\|\left\langle A\right\rangle^{k}g_{J}(H_{N}){\varphi}\|_{0}\ ,\quad\forall t\in{\mathbb{R}}\ ,\ \ \ \forall{\varphi}\in{\mathcal{H}}^{k}\ .$ (3.49) Moreover $J$ can be chosen so that $\left|J\cap\sigma(H_{N})\right|>0$ and $\sigma(H_{N})\cap J=\sigma_{ac}(H_{N})\cap J$. ###### Proof. Apply Theorem C.1, noting that condition (M1) at page (M1) is trivially satisfied $\forall n\in{\mathbb{N}}$ as ${\rm ad}^{n}_{A}(H_{N})\in{\mathcal{A}}_{0}\subset{\mathcal{L}}({\mathcal{H}})$, whereas the whole point of Lemma 3.13 was to verify (M2). This gives estimate (3.49). The right hand side is finite for ${\varphi}\in{\mathcal{H}}^{k}$ by Lemma 3.15 below, which ensures that $g_{J}(H_{N}){\varphi}\in{\mathcal{H}}^{k}$. Finally note that, since $\left|I\cap\sigma(H_{N})\right|>0$, it is certainly possible to choose $J\subset I$ so that $\left|J\cap\sigma(H_{N})\right|>0$; as $H_{N}$ fulfills a strict Mourre estimate over $I$, its spectrum in this interval is absolutely continuous, so the same is true in $J$. ∎ ###### Lemma 3.15. For any $k\in{\mathbb{N}}$, $g_{J}(H_{N})$ extends to a bounded operator ${\mathcal{H}}^{k}\to{\mathcal{H}}^{k}$. ###### Proof. As $K_{0}^{k}\,g_{J}(H_{N})\,K_{0}^{-k}=g_{J}(H_{N})-[g_{J}(H_{N}),K_{0}^{k}]K_{0}^{-k},$ it is clearly sufficient to show that $[g_{J}(H_{N}),K_{0}^{k}]K_{0}^{-k}$ is bounded on ${\mathcal{H}}$. The adjoint formula (B.6) gives $[g_{J}(H_{N}),K_{0}^{k}]K_{0}^{-k}=\sum_{j=1}^{k}c_{k,j}\,{\rm ad}^{j}_{K_{0}}(g_{J}(H_{N}))\,K_{0}^{-j}\ ;$ then it is enough to show that ${\rm ad}^{j}_{K_{0}}(g_{J}(H_{N}))\in{\mathcal{L}}({\mathcal{H}})$. As ${\rm ad}^{j}_{K_{0}}(H_{N})$ is a bounded operator $\forall j$ (symbolic calculus), the result is an immediate application of Lemma B.5. ∎ We finally prove Proposition 3.6. ###### Proof of Proposition 3.6. First we show that for any $k\in{\mathbb{N}}$, there exists $C_{2k}>0$ such that $\|e^{-{\rm i}tH_{N}}g_{J}(H_{N}){\varphi}\|_{-2k}\leq C_{2k}\,\left\langle t\right\rangle^{-2k}\,\|g_{J}(H_{N}){\varphi}\|_{2k},\qquad\forall t\in{\mathbb{R}},\ \ \ \forall{\varphi}\in{\mathcal{H}}^{2k}\ .$ (3.50) This follows from Corollary 3.14 with $k\leadsto 2k$. Indeed, as $A\in{\mathcal{A}}_{1}$, the operator $\left\langle A\right\rangle^{2k}=(1+A^{2})^{k}\in{\mathcal{A}}_{2k}$ and therefore, by symbolic calculus, $K_{0}^{-2k}\left\langle A\right\rangle^{2k}$ and $\left\langle A\right\rangle^{2k}K_{0}^{-2k}$ belong to ${\mathcal{A}}_{0}\subset{\mathcal{L}}({\mathcal{H}})$. Then $\displaystyle\|e^{-{\rm i}tH_{N}}g_{J}(H_{N}){\varphi}\|_{-2k}$ $\displaystyle\leq\|K_{0}^{-2k}\,\left\langle A\right\rangle^{2k}\|_{{\mathcal{L}}({\mathcal{H}})}\,\|\left\langle A\right\rangle^{-2k}e^{-{\rm i}tH_{N}}g_{J}(H_{N}){\varphi}\|_{0}$ $\displaystyle\leq C_{2k}\left\langle t\right\rangle^{-2k}\|\left\langle A\right\rangle^{2k}g_{J}(H_{N}){\varphi}\|_{0}$ $\displaystyle\leq C_{2k}\left\langle t\right\rangle^{-2k}\|\left\langle A\right\rangle^{2k}K_{0}^{-2k}\|_{{\mathcal{L}}({\mathcal{H}})}\|g_{J}(H_{N}){\varphi}\|_{2k}$ proving (3.50). Then linear interpolation with the equality $\|e^{-{\rm i}tH_{N}}{\varphi}_{0}\|_{0}=\|{\varphi}_{0}\|_{0}$ $\,\forall t$ gives $\forall r\in[0,2k]$ $\|e^{-{\rm i}tH_{N}}g_{J}(H_{N}){\varphi}\|_{-r}\leq C_{r}\left\langle t\right\rangle^{-r}\|g_{J}(H_{N}){\varphi}\|_{r}\ ,\quad\forall t\in{\mathbb{R}}\ ,\ \ \ \forall{\varphi}\in{\mathcal{H}}^{r}\ .$ Finally we must show that this estimate is not trivial, namely that $\exists{\varphi}\in{\mathcal{H}}^{k}$ so that $g_{J}(H_{N}){\varphi}\neq 0$. So take $J\subset I$ with $\left|J\cap\sigma(H_{N})\right|>0$ and $\sigma(H_{N})\cap J=\sigma_{ac}(H_{N})\cap J$, which is possible by Corollary 3.14. As $g_{J}(H_{N}){\mathcal{H}}\neq\\{0\\}$ and ${\mathcal{H}}^{k}$ is dense in ${\mathcal{H}}$, we have that $g_{J}(H_{N}){\mathcal{H}}^{k}\neq\\{0\\}$. Then it is enough to take $f\in{\mathcal{H}}^{k}$ so that $g_{J}(H_{N})f\neq 0$, and put ${\varphi}_{0}:=g_{J}(H_{N})f$ which, by Lemma 3.15, belongs to ${\mathcal{H}}^{k}$. Such initial datum fulfills the claim of Proposition 3.6. ∎ ### 3.3 Proof of Theorem 2.8 We are finally in position of proving Theorem 2.8. Recall that in Corollary 3.5 we have conjugated equation (2.11) to (3.17) with a change of variables bounded ${\mathcal{H}}^{r}\to{\mathcal{H}}^{r}$ uniformly in time, whereas in Proposition 3.6 we have constructed a solution of the effective equation ${\rm i}\partial_{t}\psi=H_{N}\psi$ with decaying negative Sobolev norms, therefore with growing positive Sobolev norms. The last step is to construct a solution of the full equation (3.17) with growing Sobolev norms. To achieve this, we exploit that the perturbation $R_{N}(t)$ is $N$-smoothing (Definition 2.2). So to proceed we fix the parameters. First fix $r>0$, then choose $N,k\in{\mathbb{N}}$ such that $N\geq 2r+2,\quad k\geq N-r.$ (3.51) Apply Corollary 3.5 with such $N$, producing the operators $T_{N}$, $R_{N}(t)$ and conjugating (2.11) to (3.17). By Proposition 3.6, $\exists\,{\varphi}_{0}\in{\mathcal{H}}^{k}$ such that ${\varphi}(t):=e^{-{\rm i}tH_{N}}{\varphi}_{0}$ fulfills $\forall{\tt r}\in[0,k]$: $\|{\varphi}(t)\|_{-{\tt r}}\leq C_{{\tt r},N}\left\langle t\right\rangle^{-{\tt r}}\,\|{\varphi}_{0}\|_{{\tt r}},\qquad\forall t\in{\mathbb{R}}\ .$ (3.52) We look for an exact solution $\phi(t)$ of (3.17) of the form $\phi(t)={\varphi}(t)+u(t)$, i.e. $u(t)$ has to satisfy ${\rm i}\partial_{t}u=\big{(}H_{N}+R_{N}(t)\big{)}u+R_{N}(t){\varphi}(t).$ (3.53) Denoting by $U_{N}(t,s)$ the linear propagator of $H_{N}+R_{N}(t)$, we choose $u(t):={\rm i}\int\limits_{t}^{+\infty}U_{N}(t,s)\,R_{N}(s)\,{\varphi}(s)\,{\rm d}s.$ (3.54) We estimate the ${\mathcal{H}}^{r}$ norm of $u(t)$. As $\sup_{t}\|[H_{N}+R_{N}(t),\,K_{0}]\|_{{\mathcal{L}}({\mathcal{H}}^{m})}<C_{m}<\infty\ ,\qquad\forall m\in{\mathbb{R}},$ Theorem 1.5 of [33] guarantees that the propagator $U_{N}(t,s)$ extends to a bounded operator ${\mathcal{H}}^{r}\to{\mathcal{H}}^{r}$ fulfilling666apply the theorem with $\tau=0$ and note that in that paper we defined $\|\psi\|_{r}\equiv\|K_{0}^{r/2}\psi\|_{0}$, therefore the estimate in that paper reads explicitly $\|K_{0}^{r/2}U_{N}(t,s)\psi\|_{0}\leq C_{r}\,\left\langle t-s\right\rangle^{r/2}\|K_{0}^{r/2}\psi\|_{0}$ $\forall r>0\ \ \ \exists\,C_{r}>0\colon\qquad\|U_{N}(t,s)\|_{{\mathcal{L}}({\mathcal{H}}^{r})}\leq C_{r}\,\left\langle t-s\right\rangle^{r},\quad\forall t,s\in{\mathbb{R}}\ .$ This estimate, the smoothing property $R_{N}(t)\colon{\mathcal{H}}^{r-N}\to{\mathcal{H}}^{r}$ and (3.52) with ${\tt r}:=N-r\in[0,k]$ give $\displaystyle\|u(t)\|_{r}$ $\displaystyle\leq C_{r}\int\limits_{t}^{+\infty}\left\langle t-s\right\rangle^{r}\|R_{N}(s)\,{\varphi}(s)\|_{r}\,{\rm d}s\leq C_{r}\int\limits_{t}^{+\infty}\left\langle t-s\right\rangle^{r}\,\|{\varphi}(s)\|_{-(N-r)}\,{\rm d}s$ $\displaystyle\leq C_{r,N}\,\|{\varphi}_{0}\|_{N-r}\int\limits_{t}^{+\infty}\left\langle t-s\right\rangle^{r}\,\frac{1}{\left\langle s\right\rangle^{N-r}}\,\,{\rm d}s\leq C_{r,N}\,\|{\varphi}_{0}\|_{k}\left\langle t\right\rangle^{-1}\ .$ In particular the ${\mathcal{H}}^{r}$ norm of $u(t)$ decreases to 0 as $t\to\infty$. Then $\phi(t)={\varphi}(t)+u(t)$ fulfills $\|\phi(t)\|_{r}\geq\|{\varphi}(t)\|_{r}-\|u(t)\|_{r}\geq c_{r}\frac{\|{\varphi}_{0}\|_{0}^{2}}{\|{\varphi}_{0}\|_{r}}\left\langle t\right\rangle^{r}-C_{r,N}\|{\varphi}_{0}\|_{k}\left\langle t\right\rangle^{-1}\geq C\left\langle t\right\rangle^{r}\ ,\quad\forall|t|\geq T\,,$ (3.55) where we used (3.52) with ${\tt r}=r$ and Remark 3.7. Finally we get a solution of the original equation (2.11) putting $\psi(t)={\mathcal{U}}_{N}(t)^{-1}\phi(t)$, recall Proposition 3.4. The operator ${\mathcal{U}}_{N}(t)$ fulfills (3.15), thus $\psi(t)$ has polynomially growing Sobolev norms as (2.12), concluding the proof of Theorem 2.8. We can also prove the existence of infinitely many solutions undergoing growth of Sobolev norms. ###### Corollary 3.16. There are infinitely many distinct solutions of equation (2.11) with growing Sobolev norms. ###### Proof. We fix $r>0$ and choose $N,k$ as in (3.51). From the previous proof, it follows that any initial data of the form $\psi(0):=({\rm Id}+{\mathcal{K}}_{0}){\varphi}\ ,\qquad{\mathcal{K}}_{t}{\varphi}:={\rm i}\int_{t}^{+\infty}U_{N}(t,s)R_{N}(s)e^{-{\rm i}sH_{N}}{\varphi}\,{\rm d}s,\,\qquad t\geq 0,$ with ${\varphi}\in{\rm Ran}\,g_{J}(H_{N})\cap{\mathcal{H}}^{k}$, gives rise to a solution with growing Sobolev norms (see also Remark 3.8). Here $J$ is the interval of Corollary 3.14. In particular, as $\left|J\cap\sigma(H_{N})\right|>0$ and $\sigma(H_{N})\cap J=\sigma_{ac}(H_{N})\cap J$, the set ${\rm Ran}\,g_{J}(H_{N})$ has infinite dimension. Let us prove that ${\rm Id}+{\mathcal{K}}_{0}$ is injective. Assume there are ${\varphi}_{1}\neq{\varphi}_{2}\in{\rm Ran}\,g_{J}(H_{N})\cap{\mathcal{H}}^{k}$ with $({\rm Id}+{\mathcal{K}}_{0}){\varphi}_{1}=({\rm Id}+{\mathcal{K}}_{0}){\varphi}_{2}$. Put $u_{j}(t):={\mathcal{K}}_{t}{\varphi}_{j}$, $j=1,2$; arguing as in the previous proof one has $\|u_{j}(t)\|_{r}\to 0$ as $t\to\infty$. Then ${\mathcal{U}}_{N}(t)^{-1}(e^{-{\rm i}tH_{N}}{\varphi}_{j}+u_{j}(t))$, $j=1,2$, both solve (2.11) and have the same initial datum, so they are the same solution $\psi(t)$ of equation (2.11). Then $\displaystyle\|{\varphi}_{1}-{\varphi}_{2}\|_{0}$ $\displaystyle=\|e^{-{\rm i}tH_{N}}({\varphi}_{1}-{\varphi}_{2})\|_{0}$ $\displaystyle\leq C_{r}\|{\mathcal{U}}_{N}^{-1}(t)e^{-{\rm i}tH_{N}}({\varphi}_{1}-{\varphi}_{2})\|_{r}\leq C_{r}\big{(}\|u_{1}(t)\|_{r}+\|u_{2}(t)\|_{r}\big{)}\to 0$ as $t\to\infty$. Hence ${\varphi}_{1}={\varphi}_{2}$. ∎ ## 4 Applications In the following section we apply Theorem 2.8 to the harmonic oscillator on ${\mathbb{R}}$ and the half-wave equation on ${\mathbb{T}}$. In both cases we construct transporters which are stable under small, time periodic, pseudodifferential perturbations. ### 4.1 Harmonic oscillator on ${\mathbb{R}}$ Consider the quantum harmonic oscillator $\displaystyle{\rm i}\partial_{t}\psi=\frac{1}{2}(-\partial_{x}^{2}+x^{2})\psi+V(t,x,D)\psi,\quad x\in{\mathbb{R}}.$ (4.1) Here $K_{0}:=\frac{1}{2}\left(-\partial_{x}^{2}+x^{2}\right)$ is the quantum Harmonic oscillator, the scale of Hilbert spaces is defined as usual by ${\mathcal{H}}^{r}={\rm Dom}\left(K_{0}^{r}\right)$, and the base space $({\mathcal{H}}^{0},\left\langle\cdot,\cdot\right\rangle)$ is $L^{2}({\mathbb{R}},{\mathbb{C}})$ with its standard scalar product. The perturbation $V$ is chosen as the Weyl quantization of a symbol belonging to the following class: ###### Definition 4.1. A function $f$ is a symbol of order $\rho\in{\mathbb{R}}$ if $f\in C^{\infty}({\mathbb{R}}_{x}\times{\mathbb{R}}_{\xi},{\mathbb{C}})$ and $\forall\alpha,\beta\in{\mathbb{N}}_{0}$, there exists $C_{\alpha,\beta}>0$ such that $|\partial_{x}^{\alpha}\,\partial_{\xi}^{\beta}f(x,\xi)|\leq C_{\alpha,\beta}\ (1+|x|^{2}+|\xi|^{2})^{\rho-\frac{\beta+\alpha}{2}}\ .$ We will write $f\in S^{\rho}_{{\rm har}}$. We endow $S^{\rho}_{\rm har}$ with the family of seminorms $\wp^{\rho}_{j}(f):=\sum_{|\alpha|+|\beta|\leq j}\ \ \sup_{(x,\xi)\in{\mathbb{R}}^{2}}\frac{\left|\partial_{x}^{\alpha}\,\partial_{\xi}^{\beta}f(x,\xi)\right|}{\left(1+|x|^{2}+|\xi|^{2}\right)^{\rho-\frac{\beta+\alpha}{2}}}\ ,\qquad j\in{\mathbb{N}}\cup\\{0\\}\ .$ Such seminorms turn $S^{\rho}_{\rm har}$ into a Fréchet space. If a symbol $f$ depends on additional parameters (e.g. it is time dependent), we ask that all the seminorms are uniform w.r.t. such parameters. To a symbol $f\in S^{\rho}_{\rm har}$ we associate the operator $f(x,D)$ by standard Weyl quantization $\Big{(}f(x,D)\psi\Big{)}(x):=\frac{1}{2\pi}\iint_{y,\xi\in{\mathbb{R}}}{\rm e}^{{\rm i}(x-y)\xi}\,f\left(\frac{x+y}{2},\xi\right)\,\psi(y)\,{\rm d}y{\rm d}\xi\ .$ ###### Definition 4.2. We say that $F\in{\mathcal{A}}_{\rho}$ if it is a pseudodifferential operator with symbol of class $S^{\rho}_{{\rm har}}$, i.e., if there exists $f\in S^{\rho}_{{\rm har}}$ and $S$ smoothing (in the sense of Definition 2.2) such that $F=f(x,D_{x})+S$. ###### Remark 4.3. With our numerology, the symbol of the harmonic oscillator $K_{0}$ is of order 1, $\frac{1}{2}({x^{2}+\xi^{2}})\in S^{1}_{{\rm har}}$, and not of order 2 as typically in the literature. As an application of the abstract theorems, we describe a class of operators which are transporters. This class, which we call smooth Töplitz operators, is easily described in terms of their matrix elements, which we now introduce. We denote by $\\{{\bf e}_{n}\\}_{n\in{\mathbb{N}}}$ the Hermite basis, formed by the (orthonormal) eigenvectors of the Harmonic oscillator $K_{0}$: $K_{0}{\bf e}_{n}=\left(n-\frac{1}{2}\right){\bf e}_{n},\quad\|{\bf e}_{n}\|_{0}=1,\quad n\in{\mathbb{N}}\ .$ (4.2) To each operator $\mathsf{H}\in{\mathcal{L}}({\mathcal{H}})$ we associate its matrix $(\mathsf{H}_{mn})_{m,n\in{\mathbb{N}}}$ with respect to the Hermite basis, whose elements are given by $\mathsf{H}_{mn}:=\left\langle\mathsf{H}\,{\bf e}_{n},{\bf e}_{m}\right\rangle\ ,\qquad\forall m,n\in{\mathbb{N}}\ .$ (4.3) ###### Remark 4.4. If $\mathsf{H}$ is selfadjoint, so is its matrix $(\mathsf{H}_{mn})_{m,n\in{\mathbb{N}}}$, in particular $\mathsf{H}_{mn}=\overline{\mathsf{H}_{nm}}$. ###### Definition 4.5 (Smooth Töplitz operators). A linear operator $\mathsf{H}\in{\mathcal{L}}({\mathcal{H}})$ is said a Töplitz operator if the entries of its matrix are constant along each diagonal, i.e. $\mathsf{H}_{m_{1}n_{1}}=\mathsf{H}_{m_{2}n_{2}},\quad\forall m_{1},n_{1},m_{2},n_{2}\in{\mathbb{N}}\colon\ \ \ m_{1}-n_{1}=m_{2}-n_{2}\ .$ (4.4) A Töplitz operator is said smooth if its matrix elements decay fast off diagonal, i.e. $\forall N>0$, $\exists C_{N}>0$ such that $\left|\mathsf{H}_{mn}\right|\leq\frac{C_{N}}{\left\langle m-n\right\rangle^{N}}\ ,\qquad\forall m,n\in{\mathbb{N}}\ .$ (4.5) ###### Example 4.6. The shift operators $S$ and its adjoint $S^{*}$ are defined on the Hermite functions $\\{{\bf e}_{n}\\}_{n\geq 1}$ by $S{\bf e}_{n}={\bf e}_{n+1}\ ,\quad\forall n\in{\mathbb{N}}\ ,\qquad S^{*}{\bf e}_{n}=\begin{cases}0&\mbox{ if }n=1\\\ {\bf e}_{n-1}&\mbox{ if }n\geq 2\end{cases}\ .$ (4.6) The action of $S$ (and of $S^{*}$) is extended on all ${\mathcal{H}}$ by linearity, giving $S\psi=\sum_{n\geq 1}\psi_{n}{\bf e}_{n+1}$, where we defined $\psi_{n}:=\left\langle\psi,{\bf e}_{n}\right\rangle$ for $n\geq 1$. Their matrices are given by $(S_{mn})_{m,n\in{\mathbb{N}}}=\begin{pmatrix}0&&&\\\ 1&0&&\\\ &1&0&\\\ &&\ddots&\ddots\end{pmatrix},\qquad(S^{*}_{mn})_{m,n\in{\mathbb{N}}}=\begin{pmatrix}0&1&&\\\ &0&1&\\\ &&0&1\\\ &&&\ddots\end{pmatrix},$ from which it is clear that both $S$ and $S^{*}$ are smooth Töplitz operators. We prove in the following that any smooth Töplitz operator is actually a pseudodifferential operator in ${\mathcal{A}}_{0}$, see Lemma 4.10. As an application of the abstract theorems, we show that any smooth Töplitz operator becomes a transporter for the Harmonic oscillator once it is multiplied by an appropriate scalar time periodic function. ###### Theorem 4.7. Let $\mathsf{V}(x,D)$ be a selfadjoint and smooth Töplitz operator (see Definition 4.5). Take $m,n\in{\mathbb{N}}$, $m>n$, such that the matrix element $\mathsf{V}_{m-n}:=\left\langle\mathsf{V}(x,D)\,{\bf e}_{n},{\bf e}_{m}\right\rangle\neq 0\ .$ Then $V(t,x,D):=\cos((m-n)t)\,\mathsf{V}(x,D)$ (4.7) is a transporter for (4.1). More precisely, $\forall r\geq 0$ there exist a solution $\psi(t)\in{\mathcal{H}}^{r}$ of (4.1) and constants $C,T>0$ such that $\|\psi(t)\|_{r}\geq C\left\langle t\right\rangle^{r},\quad\forall t>T.$ The theorem follows applying Theorem 2.7. So we check that Assumptions I-III are fulfilled. Regarding Assumption I, it is the usual Weyl calculus for symbols in $S^{\rho}_{{\rm har}}$, see e.g. [40]. Concerning Assumption II, one has $\sigma(K_{0})=\\{n-\frac{1}{2}\\}_{n\in{\mathbb{N}}}$. Furthermore Egorov theorem for the Harmonic oscillator [27] states that the map $t\mapsto e^{-{\rm i}tK_{0}}\mathsf{A}e^{{\rm i}tK_{0}}\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{\rho})$ for any $\mathsf{A}\in{\mathcal{A}}_{\rho}$ (use also the periodicity of the flow of $K_{0}$). This can be seen e.g. by remarking that the symbol of $e^{-{\rm i}tK_{0}}\mathsf{A}e^{{\rm i}tK_{0}}$ is $a\circ\phi_{{\rm har}}^{t}$, where $a\in S^{\rho}_{\rm har}$ is the symbol of $\mathsf{A}$ and $\phi^{t}_{{\rm har}}$ is the time $t$ flow of the harmonic oscillator; explicitly $\displaystyle\left(a\circ\phi^{t}_{{\rm har}}\right)(x,\xi)=a(x\cos t+\xi\sin t,-x\sin t+\xi\cos t)\ .$ (4.8) Verification of Assumption III. First we show that smooth Töplitz operators belong to ${\mathcal{A}}_{0}$. We exploit Chodosh’s characterization [11], which we now recall. Define the discrete difference operator $\triangle$ on a function $M\colon{\mathbb{N}}\times{\mathbb{N}}\to{\mathbb{C}}$ by $(\triangle M)(m,n):=M(m+1,n+1)-M(m,n)\ ,$ and its powers $\triangle^{\gamma}$, $\gamma\in{\mathbb{N}}$, by $\triangle$ applied $\gamma$-times. ###### Definition 4.8 (Symbol matrix). A function $M\colon{\mathbb{N}}\times{\mathbb{N}}\to{\mathbb{C}}$ will be said to be a symbol matrix of order $\rho$ if for any $\gamma\in{\mathbb{N}}_{0}$, $N\in{\mathbb{N}}$, there exists $C_{\gamma,N}>0$ such that $\left|(\triangle^{\gamma}M)(m,n)\right|\leq C_{\gamma,N}\frac{(1+m+n)^{\rho-|\gamma|}}{\left\langle m-n\right\rangle^{N}}\ ,\quad\forall m,n\in{\mathbb{N}}\ .$ (4.9) The connection between pseudodifferential operators and symbol matrices is given by Chodosh’s characterization: ###### Theorem 4.9 ([11]). An operator $\mathsf{H}$ belongs to ${\mathcal{A}}_{\rho}$ if and only if its matrix $M^{(\mathsf{H})}(m,n):=\mathsf{H}_{mn}$ (as defined in (4.3)) is a symbol matrix of order $\rho$. As a direct consequence we have the following result: ###### Lemma 4.10. Any smooth Töplitz operator is a pseudodifferential operator in ${\mathcal{A}}_{0}$. ###### Proof. We use Theorem 4.9. Let $\mathsf{H}$ be smooth Töplitz and put $M^{(\mathsf{H})}(m,n):=\mathsf{H}_{mn}$. Then (4.9) holds with $\rho=\gamma=0$ by (4.5). By (4.4) one has $\triangle M^{(\mathsf{H})}=0$; so (4.9) holds also $\forall\gamma\geq 1$. ∎ In particular $V(t,x,D)=\cos((m-n)t)\mathsf{V}(x,D)$ belongs to $C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$, which is the first required property of Assumption III. ###### Remark 4.11. The shift operators $S,S^{*}$, defined in (4.6), belong to ${\mathcal{A}}_{0}$ being smooth Töplitz. Also their (integer) powers $S^{k}$, $S^{*k}$, given for $k\in{\mathbb{N}}$ by $S^{k}{\bf e}_{n}={\bf e}_{n+k}\ ,\quad\forall n\in{\mathbb{N}}\ ,\qquad S^{*k}{\bf e}_{n}=\begin{cases}0&\mbox{ if }n\leq k\\\ {\bf e}_{n-k}&\mbox{ if }n\geq k+1\end{cases}\ $ (4.10) are smooth Töplitz, so in ${\mathcal{A}}_{0}$. Next we compute the resonant average of $V(t,x,D)$. ###### Lemma 4.12. Let $V(t,x,D)$ as in (4.7). Its resonant average $\left\langle V\right\rangle$ (see (1.3)) is $\left\langle V\right\rangle=\frac{1}{2}\left(\mathsf{V}_{k}\,S^{k}+\overline{\mathsf{V}_{k}}\,S^{*k}\right)\ ,\qquad k:=m-n\in{\mathbb{N}}\ ,$ (4.11) where $S\in{\mathcal{A}}_{0}$ is defined in (4.6) and $\mathsf{V}_{k}:=\mathsf{V}_{m-n}:=\left\langle\mathsf{V}\,{\bf e}_{n},{\bf e}_{m}\right\rangle\in{\mathbb{C}}$. ###### Proof. For $\ell\in{\mathbb{N}}$, denote by $\Pi_{\ell}{\varphi}:=\left\langle{\varphi},{\bf e}_{\ell}\right\rangle\,{\bf e}_{\ell}$ the projector on the Hermite function ${\bf e}_{\ell}$. Clearly $e^{{\rm i}sK_{0}}\,\Pi_{\ell}=\Pi_{\ell}\,e^{{\rm i}sK_{0}}=e^{{\rm i}s(\ell-\frac{1}{2})}\,\Pi_{\ell},\quad\forall\ell\in{\mathbb{N}}\ .$ From now on we simply write $\mathsf{V}\equiv\mathsf{V}(x,D)$. Using this identity and writing ${\rm Id}=\sum_{\ell\geq 1}\Pi_{\ell}$ we get $\displaystyle e^{{\rm i}sK_{0}}\,\mathsf{V}\,e^{-{\rm i}sK_{0}}$ $\displaystyle=\sum_{j,\ell\geq 1}e^{{\rm i}s(j-\ell)}\Pi_{j}\,\mathsf{V}\,\Pi_{\ell}=\sum_{j,\ell\geq 1}e^{{\rm i}s(j-\ell)}\,\left\langle\cdot,{\bf e}_{\ell}\right\rangle\,\left\langle\mathsf{V}{\bf e}_{\ell},{\bf e}_{j}\right\rangle\,{\bf e}_{j}\ .$ Now we compute, with $k:=m-n\in{\mathbb{N}}$, $\displaystyle\left\langle V\right\rangle$ $\displaystyle=\frac{1}{2\pi}\int_{0}^{2\pi}\cos(ks)\,e^{{\rm i}sK_{0}}\,\mathsf{V}\,e^{-{\rm i}sK_{0}}{\rm d}s=\sum_{j,\ell\geq 1}\left\langle\mathsf{V}{\bf e}_{\ell},{\bf e}_{j}\right\rangle\,\left\langle\cdot,{\bf e}_{\ell}\right\rangle\,{\bf e}_{j}\,\frac{1}{2\pi}\int_{0}^{2\pi}\cos(ks)\,e^{{\rm i}s(j-\ell)}\,{\rm d}s$ $\displaystyle=\frac{1}{2}\sum_{\ell\geq 1}\left\langle\mathsf{V}{\bf e}_{\ell},{\bf e}_{\ell+k}\right\rangle\left\langle\cdot,{\bf e}_{\ell}\right\rangle\,{\bf e}_{\ell+k}+\frac{1}{2}\sum_{\ell\geq k+1}\left\langle\mathsf{V}{\bf e}_{\ell},{\bf e}_{\ell-k}\right\rangle\left\langle\cdot,{\bf e}_{\ell}\right\rangle\,{\bf e}_{\ell-k}=\frac{1}{2}\mathsf{V}_{k}\,S^{k}+\frac{1}{2}\overline{\mathsf{V}_{k}}\,S^{*k}$ where in the last line we used $\mathsf{V}_{-k}=\left\langle\mathsf{V}{\bf e}_{\ell},{\bf e}_{\ell-k}\right\rangle=\overline{\left\langle\mathsf{V}{\bf e}_{\ell-k},{\bf e}_{\ell}\right\rangle}=\overline{\mathsf{V}_{k}}$ being $\mathsf{V}$ selfadjoint and smooth Töplitz (see Remark 4.4). ∎ Now define the selfadjoint operator $A:=\frac{\mathsf{V}_{k}}{{\rm i}}\,(K_{0}+\frac{1}{2})\,S^{k}-\frac{\overline{\mathsf{V}_{k}}}{{\rm i}}\,S^{*k}\,(K_{0}+\frac{1}{2})-\frac{\overline{\mathsf{V}_{k}}}{{\rm i}}\,(K_{0}+\frac{1}{2})\,S^{*k}+\frac{{\mathsf{V}_{k}}}{{\rm i}}\,S^{k}\,(K_{0}+\frac{1}{2})\ ,$ (4.12) which belongs to ${\mathcal{A}}_{1}$ by symbolic calculus as $K_{0}\in{\mathcal{A}}_{1}$ and $S,S^{*}\in{\mathcal{A}}_{0}$ (see Remark 4.11). The next lemma verifies Assumption III. ###### Lemma 4.13. Assume that $\mathsf{V}_{k}\neq 0$. The following holds true: * (i) The spectrum of the operator $H_{0}:=\left\langle V\right\rangle$ fulfills $\sigma(H_{0})\supseteq\left[-|\mathsf{V}_{k}|,\,|\mathsf{V}_{k}|\right]$. * (ii) For any interval $I_{0}\subset\left[-|\mathsf{V}_{k}|,\,|\mathsf{V}_{k}|\right]$, any $g_{I_{0}}\in C^{\infty}_{c}({\mathbb{R}},{\mathbb{R}}_{\geq 0})$ with $g_{I_{0}}\equiv 1$ over $I_{0}$ and ${\rm supp}\,g_{I_{0}}\subset\left[-|\mathsf{V}_{k}|,\,|\mathsf{V}_{k}|\right]$, there exist $\theta>0$ and $\mathsf{K}$ compact operator such that $g_{I_{0}}(H_{0})\,{\rm i}[H_{0},A]\,g_{I_{0}}(H_{0})\geq\theta\,g_{I_{0}}(H_{0})^{2}+\mathsf{K}\ .$ Here $A$ is defined in (4.12). ###### Proof. $(i)$ Let ${\mathtt{f}}(\rho):={\rm Re}(\mathsf{V}_{k}\,e^{-{\rm i}\rho k})$. We shall prove that ${\mathtt{f}}(\rho)\in\sigma(H_{0})$ $\forall\rho\in{\mathbb{R}}$, from which the claim follows. As $H_{0}$ is selfadjoint, it is enough to construct a Weyl sequence for ${\mathtt{f}}(\rho)$, i.e. a sequence $(\psi^{(n)})_{n\geq 1}$ with $\|\psi^{(n)}\|_{0}=1$ $\,\forall n$ and $\|(H_{0}-{\mathtt{f}}(\rho))\psi^{(n)}\|_{0}\to 0$ as $n\to\infty$. We put $\psi^{(n)}:=\frac{1}{\sqrt{n}}\sum_{\ell=1}^{n}e^{{\rm i}\rho\ell}{\bf e}_{\ell}\ .$ Then $\|\psi^{(n)}\|_{0}=1$ $\,\forall n$ and a direct computation shows that for $n>k$ $H_{0}\psi^{(n)}=\frac{1}{\sqrt{n}}\frac{\overline{\mathsf{V}}_{k}}{2}e^{{\rm i}\rho k}\sum_{m=1}^{k}e^{{\rm i}\rho m}\,{\bf e}_{m}+\frac{1}{\sqrt{n}}{\mathtt{f}}(\rho)\sum_{m=k+1}^{n-k}e^{{\rm i}\rho m}\,{\bf e}_{m}+\frac{1}{\sqrt{n}}\frac{\mathsf{V}_{k}}{2}e^{-{\rm i}\rho k}\sum_{m=n-k+1}^{n+k}e^{{\rm i}\rho m}\,{\bf e}_{m}\ .$ Thus one finds a constant $C_{k}>0$ such that $\|(H_{0}-{\mathtt{f}}(\rho))\psi^{(n)}\|_{0}\leq\frac{C_{k}}{\sqrt{n}}\to 0\quad\mbox{ as }n\to\infty\ ,$ proving that $\psi^{(n)}$ is a Weyl sequence; by Weyl criterium ${\mathtt{f}}(\rho)\in\sigma(H_{0})$. $(ii)$ First note that, by (4.2) and (4.6), one has $\forall k\in{\mathbb{N}}$ $\displaystyle[S^{k},K_{0}]=-kS^{k},\qquad[S^{*k},K_{0}]=kS^{*k},\qquad[S^{*k},S^{k}]=\Pi_{\leq k}$ (4.13) $\displaystyle S^{k}S^{*k}={\rm Id}-\Pi_{\leq k}\,\qquad S^{*k}S^{k}={\rm Id}$ (4.14) where $\Pi_{\leq k}:=\sum_{\ell=1}^{k}\Pi_{\ell}$ is the projector on the Hermite modes with index $\leq k$. Using (4.13) a direct computation gives $\displaystyle{\rm i}[H_{0},A]$ $\displaystyle=k\big{(}2|\mathsf{V}_{k}|^{2}-\mathsf{V}_{k}^{2}S^{2k}-\overline{\mathsf{V}}^{2}_{k}S^{*2k}-|\mathsf{V}_{k}|^{2}\Pi_{\leq k}\big{)}+2|\mathsf{V}_{k}|^{2}(K_{0}+\frac{1}{2})\Pi_{\leq k}$ $\displaystyle=4k\big{(}|\mathsf{V}_{k}|^{2}-H_{0}^{2}\big{)}+2|\mathsf{V}_{k}|^{2}(K_{0}+\frac{1}{2}-k)\Pi_{\leq k}\ .$ Clearly $\mathsf{K}:=2|\mathsf{V}_{k}|^{2}(K_{0}+\frac{1}{2}-k)\Pi_{\leq k}$ is compact, being finite rank. Next put $\tilde{f}(\lambda)=4k(|\mathsf{V}_{k}|^{2}-\lambda^{2})$ getting $\forall{\varphi}\in{\mathcal{H}}$ $\left\langle g_{I_{0}}(H_{0})\,{\rm i}[H_{0},A]\,g_{I_{0}}(H_{0}){\varphi},{\varphi}\right\rangle=\left\langle g_{I_{0}}(H_{0})\,\tilde{f}(H_{0})\,g_{I_{0}}(H_{0}){\varphi},{\varphi}\right\rangle+\left\langle g_{I_{0}}(H_{0})\,\mathsf{K}\,g_{I_{0}}(H_{0}){\varphi},{\varphi}\right\rangle\ .$ (4.15) Note that $\tilde{f}$ is strictly positive in the interior of $\left[-|\mathsf{V}_{k}|,\,|\mathsf{V}_{k}|\right]$; we put $\theta:=\inf\\{\tilde{f}(\lambda)\colon\lambda\in{\rm supp}\,g_{I_{0}}\\}>0\ .$ With this information we apply the spectral theorem and get $\displaystyle\left\langle g_{I_{0}}(H_{0})\,\tilde{f}(H_{0})\,g_{I_{0}}(H_{0}){\varphi},{\varphi}\right\rangle$ $\displaystyle=\int\limits_{\lambda\in\sigma(H_{0})}g_{I_{0}}(\lambda)^{2}\,\tilde{f}(\lambda)\,{\rm d}m_{\varphi}(\lambda)$ $\displaystyle\geq\theta\int\limits_{\lambda\in\sigma(H_{0})}g_{I_{0}}(\lambda)^{2}\,\,{\rm d}m_{\varphi}(\lambda)=\theta\|g_{I_{0}}(H_{0}){\varphi}\|_{0}^{2}\ .$ This estimate and (4.15) proves that $H_{0}$ fulfills a Mourre estimate over $I_{0}$. ∎ To conclude this section, we recall that in [34] it is proved that the pseudodifferential operator $\displaystyle V(t):=e^{-{\rm i}tK_{0}}\,(S+S^{*})\,e^{{\rm i}tK_{0}}$ (4.16) is a universal transporter (see Definition 1.1). Using the abstract Theorem 2.8 we prove its stability under perturbations of class $C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$: ###### Theorem 4.14. Consider equation (4.1) with $V(t)$ defined in (4.16). There exist $\epsilon_{0},{\mathtt{M}}>0$ such that $\forall W\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$ with $\sup_{t}\wp^{0}_{\mathtt{M}}(W(t))\leq\epsilon_{0}$, the operator $V+\epsilon W$ is a transporter. More precisely $\forall r>0$ there exist a solution $\psi(t)\in{\mathcal{H}}^{r}$ of ${\rm i}\partial_{t}\psi=\big{(}\frac{-\partial_{x}^{2}+x^{2}}{2}+V(t)+W(t)\big{)}\psi$ and constants $C,T>0$ such that $\|\psi(t)\|_{r}\geq C\left\langle t\right\rangle^{r},\quad\forall t\geq T\ .$ ###### Proof. Again we verify Assumption III. It is clear that $V(t)\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{0})$ and that $\left\langle V\right\rangle=S+S^{*}$, so it has the form (4.11) with $k=1$ and $\mathsf{V}_{1}=2$. Then Lemma 4.13 implies that $\left\langle V\right\rangle$ fulfills a Mourre estimate. ∎ ### 4.2 Half-wave equation on ${\mathbb{T}}$ The half-wave equation on ${\mathbb{T}}$ is given by ${\rm i}\partial_{t}\psi=|D|\psi+V(t,x,D)\psi\ ,\qquad x\in{\mathbb{T}}\ .$ (4.17) Here $|D|$ is the Fourier multiplier defined by $|D|\psi:=\sum_{j\in{\mathbb{Z}}}|j|\,\psi_{j}\,e^{{\rm i}jx}\ ,\qquad\psi_{j}:=\frac{1}{2\pi}\int_{\mathbb{T}}\psi(x)e^{-{\rm i}jx}{\rm d}x\ ,$ whereas $V(t,x,D)$ is a pseudodifferential operator of order 0. In this case $K_{0}:=|D|+1$, the scale of Hilbert spaces defined as ${\mathcal{H}}^{r}={\rm Dom}\left(K_{0}^{r}\right)$ coincides with standard Sobolev spaces on the torus $H^{r}({\mathbb{T}})$, and the base space $({\mathcal{H}}^{0},\left\langle\cdot,\cdot\right\rangle)$ is $L^{2}({\mathbb{T}},{\mathbb{C}})$ with its standard scalar product. In this setting we shall use pseudodifferential operators with periodic symbols, belonging to the following class: ###### Definition 4.15. A function $a(x,\xi)$ is a periodic symbol of order $\rho\in{\mathbb{R}}$ if $a\in C^{\infty}({\mathbb{T}}_{x}\times{\mathbb{R}}_{\xi},{\mathbb{C}})$ and for any $\alpha,\beta\in{\mathbb{N}}_{0}$, there exists a constant $C_{\alpha\beta}>0$ such that $\left|\partial_{x}^{\alpha}\,\partial_{\xi}^{\beta}\,a(x,\xi)\right|\leq C_{\alpha\beta}\,\left\langle\xi\right\rangle^{\rho-\beta},\quad\forall x\in{\mathbb{T}},\,\forall\xi\in{\mathbb{R}}\,.$ (4.18) We will write $a\in S^{\rho}_{{\rm per}}$. We also put $S^{-\infty}_{\rm per}:=\bigcap_{\rho\in{\mathbb{R}}}S^{\rho}_{\rm per}$ the class of smoothing symbols. We endow $S^{\rho}_{\rm per}$ with the family of seminorms $\wp^{\rho}_{j}(a):=\sum_{|\alpha|+|\beta|\leq j}\ \ \sup_{(x,\xi)\in{\mathbb{T}}\times{\mathbb{R}}}{\left|\partial_{x}^{\alpha}\,\partial_{\xi}^{\beta}a(x,\xi)\right|\,\left\langle\xi\right\rangle^{-\rho+\beta}}\ ,\qquad j\in{\mathbb{N}}_{0}\ .$ (4.19) Such seminorms turn $S^{\rho}_{\rm per}$ into a Fréchet space. If a symbol $a$ depends on additional parameters (e.g. it is time dependent), we ask that all the seminorms are uniform w.r.t. such parameters. To a symbol $a\in S^{\rho}_{\rm per}$ we associate its quantization $a(x,D)$ acting on a $2\pi$-periodic function $u(x)=\sum_{j\in{\mathbb{Z}}}u_{j}e^{{\rm i}jx}$ as $a(x,D)u:={\rm Op}{(a)}[u]:=\sum_{j\in{\mathbb{Z}}}\,a(x,j)\,u_{j}\,e^{{\rm i}jx}\,.$ (4.20) ###### Remark 4.16. Given a symbol $a(\xi)$ independent of $x$, then ${\rm Op}{(a)}$ is the Fourier multiplier operator $a(D)u=\sum_{j\in{\mathbb{Z}}}\,a(j)\,u_{j}\,e^{{\rm i}jx}$. If instead the symbol $a(x)$ is independent of $\xi$, then ${\rm Op}{(a)}$ is the multiplication operator ${\rm Op}{(a)}u=a(x)u$. ###### Definition 4.17. We say that $A\in{\mathcal{A}}_{\rho}$ if $A={\rm Op}(a)$ with $a\in S^{\rho}_{{\rm per}}$. ###### Example 4.18. The operator $|D|\in{\mathcal{A}}_{1}$ with symbol given by ${\mathtt{d}}(\xi):=|\xi|\chi(\xi)$ where $\chi$ is an even, positive smooth cut-off function satisfying $\chi(\xi)=0$ for $|\xi|\leq\frac{1}{5}$, $\chi(\xi)=1$ for $|\xi|\geq\frac{2}{5}$ and $\partial_{\xi}\chi(\xi)>0$ $\,\forall\xi\in(\frac{1}{5},\frac{2}{5})$. Also the Fourier projectors $\Pi_{\pm}$ and $\Pi_{0}$ defined by $\Pi_{+}u:=\sum_{j\geq 1}u_{j}\,e^{{\rm i}jx},\qquad\Pi_{-}u:=\sum_{j\leq-1}u_{j}\,e^{{\rm i}jx},\qquad\Pi_{0}u:=u_{0}$ (4.21) are pseudodifferential operators. In particular $\Pi_{\pm}={\rm Op}\left(\pi_{\pm}\right)\in{\mathcal{A}}_{0}$ and $\Pi_{0}={\rm Op}\left(\pi_{0}\right)\in{\mathcal{A}}_{-\infty}$, where $\pi_{\pm},\pi_{0}$ are a smooth partition of unity, $\pi_{+}(\xi)+\pi_{-}(\xi)+\pi_{0}(\xi)=1$ $\,\forall\xi$, fulfilling $\pi_{+}(\xi)=\begin{cases}1&\mbox{ if }\xi\geq\frac{4}{5}\\\ 0&\mbox{ if }\xi\leq\frac{3}{5}\end{cases}\ ,\quad\pi_{-}(\xi)=\begin{cases}1&\mbox{ if }\xi\leq-\frac{4}{5}\\\ 0&\mbox{ if }\xi\geq-\frac{3}{5}\end{cases}\ ,\quad\pi_{0}(\xi)=\begin{cases}1&\mbox{ if }|\xi|\leq\frac{3}{5}\\\ 0&\mbox{ if }|\xi|\geq\frac{1}{5}\end{cases}\ .$ (4.22) In this setting we prove that any multiplication operator, multiplied by an appropriate time periodic function, becomes a transporter. Here the result. ###### Theorem 4.19. Let $v\in C^{\infty}({\mathbb{T}},{\mathbb{R}})$. Choose $j\in{\mathbb{Z}}\setminus\\{0\\}$ such that the Fourier coefficient $v_{j}\neq 0$. Then the selfadjoint operator $V(t,x):=\cos(jt)\,v(x)$ (4.23) is a transporter. More precisely, $\forall r>0$ there exist a solution $\psi(t)\in{\mathcal{H}}^{r}$ of ${\rm i}\partial_{t}\psi=\big{(}|D|+V(t,x)\big{)}\psi$ and constants $C,T>0$ such that $\|\psi(t)\|_{r}\geq C\left\langle t\right\rangle^{r},\quad\forall t>T.$ The theorem follows from Theorem 2.7. So first we put ourselves in the setting of the abstract theorem and rewrite (4.17) as ${\rm i}\partial_{t}\psi=K_{0}\psi+{\widetilde{V}}(t,x)\psi,\qquad{\widetilde{V}}(t,x):=\cos(jt)v(x)-1\in{\mathcal{A}}_{0}\ .$ (4.24) Again we check Assumptions I-III. Regarding Assumption I, it is the usual pseudodifferential calculus for periodic symbols, see e.g. [39]. Verification of Assumption II. One has $\sigma(K_{0})=\\{n\\}_{n\in{\mathbb{N}}}$. To prove Assumption II $(ii)$ we use the identity $e^{-{\rm i}tK_{0}}\mathsf{A}\,e^{{\rm i}tK_{0}}=e^{-{\rm i}t|D|}\mathsf{A}\,e^{{\rm i}t|D|}$ and Egorov theorem for $|D|$, see e.g. [43, Theorem 4.3.6]. Actually we need also the following version of Egorov theorem. ###### Lemma 4.20. Let $a\in S^{\rho}_{\rm per}$, $\rho\in{\mathbb{R}}$. Then $e^{{\rm i}t|D|}\,{\rm Op}\left(a\right)\,e^{-{\rm i}t|D|}={\rm Op}\left(a(x+t,\xi)\right)\Pi_{+}+{\rm Op}\left(a(x-t,\xi)\right)\Pi_{-}+R(t)$ (4.25) where $\Pi_{\pm}$ are defined in (4.21) and $R(t)\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{\rho-1})$. If ${\rm Op}\left(a\right)$ is selfadjoint, so is $e^{{\rm i}t|D|}{\rm Op}\left(a\right)\,e^{-{\rm i}t|D|},$ $\,\forall t$. ###### Proof. The classical Egorov theorem for the half-Laplacian $|D|$ says that $e^{{\rm i}t|D|}\,{\rm Op}\left(a\right)\,e^{-{\rm i}t|D|}={\rm Op}\left(a\circ\phi_{\mathtt{d}}^{t}(x,\xi)\right)+{\widetilde{R}}(t)$ where $\phi_{\mathtt{d}}^{t}(x,\xi)$ is the time $t$ flow of the classical Hamiltonian ${\mathtt{d}}(\xi)=|\xi|\chi(\xi)$ (the symbol of $|D|$) and ${\widetilde{R}}(t)\in C^{\infty}({\mathbb{R}},{\mathcal{A}}_{\rho-1})$, see e.g. [43, Theorem 4.3.6]. We compute more explicitly $a\circ\phi_{\mathtt{d}}^{t}(x,\xi)$. The Hamiltonian equations of ${\mathtt{d}}(\xi)$ and its flow $\phi^{t}_{\mathtt{d}}$ are given by $\begin{cases}\dot{x}=\partial_{\xi}{\mathtt{d}}(\xi)={\mathtt{d}}^{\prime}(\xi)\\\ \dot{\xi}=-\partial_{x}{\mathtt{d}}(\xi)=0\end{cases}\ ,\qquad\phi^{t}_{\mathtt{d}}(x,\xi)=(x+t{\mathtt{d}}^{\prime}(\xi),\,\xi)\ .$ As ${\mathtt{d}}^{\prime}(\xi)=1$ for $\xi\geq\frac{2}{5}$ and ${\mathtt{d}}^{\prime}(\xi)=-1$ for $\xi\leq-\frac{2}{5}$, we write $(a\circ\phi_{\mathtt{d}}^{t})(x,\xi)=a(x+t,\xi)\,\pi_{+}(\xi)+a(x-t,\xi)\,\pi_{-}(\xi)+a(x+t{\mathtt{d}}^{\prime}(\xi),\xi)\,\pi_{0}(\xi)\ .$ As $\pi_{0}\in S^{-\infty}_{\rm per}$, the operator ${\rm Op}\left(a(x+t{\mathtt{d}}^{\prime}(\xi),\xi)\,\pi_{0}(\xi)\right)\in C^{\infty}({\mathbb{R}},{\mathcal{A}}_{-\infty})$. Moreover by symbolic calculus ${\rm Op}\left(a(x\pm t,\xi)\,\pi_{\pm}(\xi)\right)={\rm Op}\left(a(x\pm t,\xi)\right)\Pi_{\pm}+R_{\pm}(t),\quad R_{\pm}(t)\in C^{\infty}({\mathbb{R}},{\mathcal{A}}_{\rho-1})\ .$ Formula (4.25) follows with $R(t):={\widetilde{R}}(t)+R_{+}(t)+R_{-}(t)+{\rm Op}\left(a(x+t{\mathtt{d}}^{\prime}(\xi),\xi)\,\pi_{0}(\xi)\right)$. We claim that $R(t)$ is periodic in time. This follows by difference since both $e^{{\rm i}t|D|}\,{\rm Op}\left(a\right)\,e^{-{\rm i}t|D|}$ and ${\rm Op}\left(a(x\pm t,\xi)\right)\Pi_{\pm}$ are periodic in $t$ (recall that the symbol $a(x,\xi)$ is periodic in $x$). Finally as $e^{\pm{\rm i}t|D|}$ are unitary, the claim on the selfadjointness of $e^{{\rm i}t|D|}\,{\rm Op}\left(a\right)\,e^{-{\rm i}t|D|}$ follows. ∎ Verification of Assumption III. First we compute $\langle{\widetilde{V}}\rangle$. ###### Lemma 4.21. The resonant average $\langle{\widetilde{V}}\rangle\in{\mathcal{A}}_{0}$ of ${\widetilde{V}}$ (defined in (4.24)) is given by $\langle{\widetilde{V}}\rangle={\mathtt{v}}(x)-1+R,\qquad{\mathtt{v}}(x):={\rm Re}(v_{j}e^{{\rm i}jx})$ (4.26) and $R\in{\mathcal{A}}_{-1}$ is selfadjoint. ###### Proof. First remark that, as $e^{{\rm i}tK_{0}}=e^{{\rm i}t|D|}e^{{\rm i}t}$, $\displaystyle\langle{\widetilde{V}}\rangle$ $\displaystyle=\frac{1}{2\pi}\int_{0}^{2\pi}e^{{\rm i}s|D|}\,{\widetilde{V}}(s)\,e^{-{\rm i}s|D|}{\rm d}s=\frac{1}{2\pi}\int_{0}^{2\pi}\cos(js)\,e^{{\rm i}s|D|}\,v(x)\,e^{-{\rm i}s|D|}\,{\rm d}s-1\ .$ (4.27) We compute $e^{{\rm i}s|D|}\,v(x)\,e^{-{\rm i}s|D|}$ with the aid of Lemma 4.20, getting $e^{{\rm i}s|D|}\,v(x)\,e^{-{\rm i}s|D|}=v(x+s)\,\Pi_{+}+v(x-s)\,\Pi_{-}+{\widetilde{R}}(s),$ (4.28) where ${\widetilde{R}}(s)\in C^{\infty}({\mathbb{T}},{\mathcal{A}}_{-1})$. Then, recalling that $v_{j}=\overline{v_{-j}}$ being $v(x)$ real valued, $\displaystyle\frac{1}{2\pi}\int_{0}^{2\pi}\cos(js)\,e^{{\rm i}s|D|}\,v(x)\,e^{-{\rm i}s|D|}\,{\rm d}s$ $\displaystyle\stackrel{{\scriptstyle\eqref{eg.per}}}{{=}}\sum_{\sigma=\pm}\frac{1}{2\pi}\int_{0}^{2\pi}\cos(js)\,v(x\sigma s)\,{\rm d}s\,\Pi_{\sigma}+\frac{1}{2\pi}\int_{0}^{2\pi}\cos(js){\widetilde{R}}(s){\rm d}s$ $\displaystyle={\rm Re}\left(v_{j}e^{{\rm i}jx}\right)\,(\Pi_{+}+\Pi_{-})+\frac{1}{2\pi}\int_{0}^{2\pi}\cos(js){\widetilde{R}}(s){\rm d}s$ $\displaystyle={\rm Re}\left(v_{j}e^{{\rm i}jx}\right)+R$ where $R:=\frac{1}{2\pi}\int_{0}^{2\pi}\cos(js){\widetilde{R}}(s){\rm d}s-{\rm Re}\left(v_{j}e^{{\rm i}jx}\right)\Pi_{0}\in{\mathcal{A}}_{-1}$. Together with (4.27), this proves (4.26). Finally $R$ is selfadjoint by difference, since both $\langle{\widetilde{V}}\rangle$ and ${\mathtt{v}}(x)-1$ are selfadjoint operators. ∎ Define the selfadjoint operator $\displaystyle A:={\mathtt{w}}(x)\,\frac{\partial_{x}}{{\rm i}}+\frac{\partial_{x}}{{\rm i}}\,{\mathtt{w}}(x)\ ,\qquad{\mathtt{w}}(x):={\rm Im}(v_{j}e^{{\rm i}jx})\ $ (4.29) belonging to ${\mathcal{A}}_{1}$. The next lemma verifies Assumption III. ###### Lemma 4.22. Assume that $v_{j}\neq 0$. The following holds true: * (i) The operator $H_{0}:=\langle{\widetilde{V}}\rangle$ has spectrum $\sigma(H_{0})\supseteq[-|v_{j}|-1,\,|v_{j}|-1]=:I$. * (ii) For any interval $I_{0}\subset I$, any $g_{I_{0}}\in C^{\infty}_{c}({\mathbb{R}},{\mathbb{R}}_{\geq 0})$ with $g_{I_{0}}\equiv 1$ over $I_{0}$ and ${\rm supp}\,g_{I_{0}}\subset I$, there exist $\theta>0$ and a compact operator $\mathsf{K}$ such that $g_{I_{0}}(H_{0})\,{\rm i}[H_{0},A]\,g_{I_{0}}(H_{0})\geq\theta\,g_{I_{0}}(H_{0})^{2}+\mathsf{K}\ .$ Here $A$ is defined in (4.29). ###### Proof. During the proof we shall use that any operator in ${\mathcal{A}}_{-1}$ is compact. Moreover we shall simply denote any compact operator by $\mathsf{K}$, which can change from line to line. $(i)$ By Lemma 4.21, $H_{0}$ is a compact perturbation of the multiplication operator by ${\mathtt{v}}(x)-1$, whose spectrum coincides with $I$. Then by Weyl’s theorem $\sigma(H_{0})\supseteq\sigma_{ess}(H_{0})=\sigma_{ess}({\mathtt{v}}(x)-1)=I\ .$ $(ii)$ First notice that, as ${\mathtt{v}}(x)={\rm Re}(v_{j}e^{{\rm i}jx})$ and ${\mathtt{w}}(x)={\rm Im}(v_{j}e^{{\rm i}jx})$, one has the identities ${\mathtt{v}}(x)^{2}+{\mathtt{w}}(x)^{2}=|v_{j}|^{2},\qquad{\mathtt{v}}^{\prime}(x)=-j\,{\mathtt{w}}(x)\ .$ (4.30) Next we compute $\displaystyle{\rm i}[H_{0},A]$ $\displaystyle={\rm i}[{\mathtt{v}}(x)-1+R,A]=-2{\mathtt{w}}(x)\,{\mathtt{v}}^{\prime}(x)+{\rm i}[R,A]$ $\displaystyle\stackrel{{\scriptstyle\eqref{vw}}}{{=}}2j\big{(}|v_{j}|^{2}-{\mathtt{v}}(x)^{2}\big{)}+\mathsf{K}=2j\big{(}|v_{j}|^{2}-(H_{0}+1-R)^{2}\big{)}+\mathsf{K}$ $\displaystyle=2j\big{(}|v_{j}|^{2}-(H_{0}+1)^{2}\big{)}+\mathsf{K}\ .$ Putting $f(\lambda):=2j\big{(}|v_{j}|^{2}-(\lambda+1)^{2}\big{)}$, we get $\forall{\varphi}\in{\mathcal{H}}$ $\displaystyle\left\langle g_{I_{0}}(H_{0})\,{\rm i}[H_{0},A]\,g_{I_{0}}(H_{0}){\varphi},{\varphi}\right\rangle=\left\langle g_{I_{0}}(H_{0})\,f(H_{0})\,g_{I_{0}}(H_{0}){\varphi},{\varphi}\right\rangle+\left\langle\mathsf{K}{\varphi},{\varphi}\right\rangle.$ (4.31) Now we notice that $f(\lambda)$ is positive in the interior of $I$; so we put $\theta:=\inf\left\\{f(\lambda)\colon\ \ \lambda\in{\rm supp}\,g_{I_{0}}\right\\}>0\ .$ With this information we apply the spectral theorem, getting, as in the previous section, $\displaystyle\left\langle g_{I_{0}}(H_{0})\,f(H_{0})\,g_{I_{0}}(H_{0}){\varphi},{\varphi}\right\rangle$ $\displaystyle\geq\theta\int\limits_{\lambda\in I}g_{I_{0}}(\lambda)^{2}\,\,{\rm d}m_{\varphi}(\lambda)=\theta\|g_{I_{0}}(H_{0}){\varphi}\|_{0}^{2}\ .$ (4.32) This together with (4.31) establishes the Mourre estimate over $I_{0}$. ∎ ## Appendix A Flows of pseudodifferential operators In this appendix we collect some known results about the flow generated by pseudodifferential operators belonging to the algebra ${\mathcal{A}}$. The setting is the same as [5] and we refer to that paper for the proofs. The first result describes how a Schrödinger equation is changed under a change of variables induced by the flow of a pseudodifferential operator, see Lemma 3.1 of [5]: ###### Lemma A.1. Let $H(t)$ be a time dependent selfadjoint operator, and $X(t)$ be a selfadjoint family of operators. Assume that $\psi(t)=e^{-{\rm i}X(t)}{\varphi}(t)$ then ${\rm i}\partial_{t}\psi=H(t)\psi\ \quad\iff\quad{\rm i}\partial_{t}{\varphi}=H^{+}(t){\varphi}$ (A.1) where $\displaystyle H^{+}(t):=e^{{\rm i}X(t)}\,H(t)\,e^{-{\rm i}X(t)}-\int_{0}^{1}e^{{\rm i}sX(t)}\,(\partial_{t}X(t))\,e^{-{\rm i}sX(t)}\ {\rm d}s\ .$ (A.2) The next property we shall need is the Lie expansion of $e^{{\rm i}X}\,A\,e^{-{\rm i}X}$ in operators of decreasing order, see Lemma 3.2 of [5]: ###### Lemma A.2. Let $X\in{\mathcal{A}}_{\rho}$ with $\rho<1$ be a symmetric operator. Let $A\in{\mathcal{A}}_{m}$ with $m\in{\mathbb{R}}$. Then $e^{{\rm i}\tau X}\,A\,e^{-{\rm i}\tau X}$ is selfadjoint and for any $M\geq 1$ we have777in [5] we have defined ${\rm ad}_{X}(A)={\rm i}[X,A]$ rather than (3.21); so we formulate the next result with the current notation $e^{{\rm i}\tau X}\,A\,e^{-{\rm i}\tau X}=\sum_{\ell=0}^{M}\frac{\tau^{\ell}}{{\rm i}^{\ell}\,\ell!}{\rm ad}_{X}^{\ell}(A)+R_{M}(\tau,X,A)\ ,\qquad\forall\tau\in{\mathbb{R}}\ ,$ (A.3) where $R_{M}(\tau,X,A)\in{\mathcal{A}}_{m-(M+1)(1-\rho)}$. In particular ${\rm ad}_{X}^{\ell}(A)\in{\mathcal{A}}_{m-\ell(1-\rho)}$ and $e^{{\rm i}\tau X}\,A\,e^{-{\rm i}\tau X}\in{\mathcal{A}}_{m}$, $\forall\tau\in{\mathbb{R}}$. The last result concerns boundedness properties of the operator $e^{-{\rm i}\tau X}$, see Lemma 3.3 of [5]: ###### Lemma A.3. Assume that $X(t)$ is a family of selfadjoint operators in ${\mathcal{A}}_{1}$ s.t. $\sup_{t\in{\mathbb{R}}}\wp^{1}_{j}(X(t))<\infty\ ,\quad\forall j\geq 1\ .$ (A.4) Then $e^{-{\rm i}\tau X(t)}$ extends to an operator in ${\mathcal{L}}({\mathcal{H}}^{r})$ $\,\forall r\in{\mathbb{R}}$, and moreover there exist $c_{r},C_{r}>0$ s.t. $c_{r}\|\psi\|_{r}\leq\|e^{-{\rm i}\tau X(t)}\psi\|_{r}\leq C_{r}\|\psi\|_{r}\ ,\qquad\forall t\in{\mathbb{R}}\ ,\quad\forall\tau\in[0,1]\ .$ (A.5) ## Appendix B Functional calculus In this section we collect some known results about functional calculus of selfadjoint operators which are used thorough the paper. We begin recalling Helffer-Sjöstrand formula [26], following the presentation of [15]. ###### Definition B.1. A function $f\in C^{\infty}({\mathbb{R}},{\mathbb{C}})$ will be said to belong to the class $S^{\rho}$, $\rho\in{\mathbb{R}}$, if $\forall m\in{\mathbb{N}}_{0}$, $\exists C_{m}>0$ such that $\left|\frac{{\rm d}^{m}}{{\rm d}x^{m}}f(x)\right|\leq C_{m}\left\langle x\right\rangle^{\rho-m},\quad\forall x\in{\mathbb{R}}\ .$ As usual we set the seminorms $\wp^{\rho}_{m}(f):=\sum_{0\leq j\leq m}\ \sup_{x\in{\mathbb{R}}}\left|\frac{{\rm d}^{m}f(x)}{{\rm d}x^{m}}\right|\,\left\langle x\right\rangle^{-\rho+m}\ ,\qquad m\in{\mathbb{N}}_{0}\ .$ Given $f\in S^{\rho}$, we define its almost analytic extension as follows: for any $N\in{\mathbb{N}}$, put ${\widetilde{f}}_{N}\colon{\mathbb{R}}^{2}\to{\mathbb{C}},\qquad{\widetilde{f}}_{N}(x,y):=\left(\sum_{\ell=0}^{N}f^{(\ell)}(x)\frac{({\rm i}y)^{\ell}}{\ell!}\right)\,\tau\left(\frac{y}{\left\langle x\right\rangle}\right)$ (B.1) where $\tau\in C^{\infty}({\mathbb{R}},{\mathbb{R}}_{\geq 0})$ is a cut-off function fulfilling $\tau(s)=1$ for $|s|\leq 1$ and $\tau(s)=0$ for $|s|\geq 2$. It is well known [15] that the choice of $N$ and of the cut-off function $\tau$ are by no means critical, and even other choices of $\widetilde{f}_{N}$ are possible (see e.g. [18]). The following properties are true [15]: let $f\in S^{\rho}$ with $\rho<0$, then $\displaystyle{\widetilde{f}}_{N}|_{\mathbb{R}}=f,\qquad{\rm supp}\,{\widetilde{f}}_{N}\subset\left\\{x+{\rm i}y\ \colon\ \ \ x\in{\rm supp}\,f,\quad|y|\leq 2\left\langle x\right\rangle\right\\}\ ,$ (B.2) $\displaystyle\left|\frac{\partial{\widetilde{f}}_{N}(x,y)}{\partial{\overline{z}}}\right|\leq C_{N}\left\langle x\right\rangle^{\rho-N-1}\,|y|^{N},\qquad\frac{\partial{\widetilde{f}}_{N}}{\partial{\overline{z}}}:=\left(\frac{\partial{\widetilde{f}}_{N}}{\partial x}+{\rm i}\frac{\partial{\widetilde{f}}_{N}}{\partial y}\right)$ (B.3) $\displaystyle\int_{{\mathbb{R}}^{2}}\left|\frac{\partial{\widetilde{f}}_{N}(z)}{\partial{\overline{z}}}\right|\,\left|{\rm Im}\,(z)\right|^{-p-1}{\rm d}\overline{z}\wedge{\rm d}z\leq C_{N}\,\wp^{\rho}_{N+2}(f),\qquad\forall p=0,\ldots,N,$ (B.4) where $z=x+{\rm i}y$ and ${\rm d}\overline{z}\wedge{\rm d}z$ is the Lebesgue measure on ${\mathbb{C}}$. Given $\mathsf{H}$ a selfadjoint operator and $f\in S^{\rho}$, $\rho<0$, the Helffer-Sjöstrand formula defines $f(\mathsf{H})$ as $f(\mathsf{H}):=\frac{{\rm i}}{2\pi}\int_{{\mathbb{R}}^{2}}\frac{\partial{\widetilde{f}}_{N}(z)}{\partial{\overline{z}}}\,(z-\mathsf{H})^{-1}{\rm d}\overline{z}\wedge{\rm d}z=-\frac{1}{\pi}\int_{{\mathbb{R}}^{2}}\frac{\partial{\widetilde{f}}_{N}(z)}{\partial{\overline{z}}}\,(z-\mathsf{H})^{-1}{\rm d}x\,{\rm d}y\ .$ (B.5) ###### Theorem B.2 ([15]). Let $f\in S^{\rho}$, $g\in S^{\mu}$ with $\rho,\mu<0$ and $\mathsf{H}$ a selfadjoint operator. Then * (i) The operator $f(\mathsf{H})$ is independent of $N$ and of the cut-off function $\tau$. * (ii) The integral in (B.5) is norm convergent and $\|f(\mathsf{H})\|_{{\mathcal{L}}({\mathcal{H}})}\leq\|f\|_{L^{\infty}}$. * (iii) $f(\mathsf{H})\,g(\mathsf{H})=(fg)(\mathsf{H})$. * (iv) $\overline{f}(\mathsf{H})=f(\mathsf{H})^{*}$. * (v) If $f\in C^{\infty}_{c}$ has support disjoint from $\sigma(\mathsf{H})$, then $f(\mathsf{H})=0$. * (vi) If $z\notin{\mathbb{R}}$ and $f_{z}(x):=(z-x)^{-1}$ for all $x\in{\mathbb{R}}$, then $f_{z}\in S^{-1}$ and $f_{z}(\mathsf{H})=(z-\mathsf{H})^{-1}$. ###### Remark B.3. Given $f\in S^{\rho}$, $\rho<0$ and $\mathsf{H}$ selfadjoint, the operator $f(\mathsf{H})$ defined via Helffer-Sjöstrand formula coincides with the classical definition given by the spectral theorem, namely $f(\mathsf{H})=\int_{{\mathbb{R}}}f(\lambda)\,{\rm d}E(\lambda)$ where ${\rm d}E(\lambda)$ is the spectral resolution of $\mathsf{H}$. For a proof, see e.g. [19], Theorem 8.1. Next we recall expansion formulas for commutators. We start from the basic identities $\displaystyle{\rm ad}^{n}_{\mathsf{A}}(\mathsf{P}\mathsf{Q})$ $\displaystyle=\sum_{k=0}^{n}\binom{n}{k}\,{\rm ad}^{n-k}_{\mathsf{A}}(\mathsf{P})\,{\rm ad}^{k}_{\mathsf{A}}(\mathsf{Q}),\qquad[\mathsf{P},\mathsf{A}^{n}]=\sum_{j=1}^{n}c_{n,j}\,{\rm ad}^{j}_{\mathsf{A}}(\mathsf{P})\,\mathsf{A}^{n-j}\ .$ (B.6) For the next lemma see e.g. [18, Lemma C.3.1] or [29, Appendix B]. ###### Lemma B.4 (Commutator expansion formula). Let $k\in{\mathbb{N}}$ and $\mathsf{A},\mathsf{B}$ selfadjoint operators with $\|{\rm ad}_{\mathsf{A}}^{j}(\mathsf{B})\|_{{\mathcal{L}}({\mathcal{H}})}<\infty,\qquad\forall\,1\leq j\leq k\ .$ Let $f\in S^{\rho}$ with $\rho<0$, then one has the right and left commutator expansions $\displaystyle[\mathsf{B},f(\mathsf{A})]$ $\displaystyle=\sum_{j=1}^{k-1}\frac{1}{j!}\,f^{(j)}(\mathsf{A})\,{\rm ad}_{\mathsf{A}}^{j}(\mathsf{B})+R_{k}(f,\mathsf{A},\mathsf{B})$ (B.7) $\displaystyle=\sum_{j=1}^{k-1}\frac{(-1)^{j-1}}{j!}\,{\rm ad}_{\mathsf{A}}^{j}(\mathsf{B})\,f^{(j)}(\mathsf{A})\,+{\widetilde{R}}_{k}(f,\mathsf{A},\mathsf{B})$ (B.8) where the operators $R_{k},{\widetilde{R}}_{k}$ fulfill $\|R_{k}(f,\mathsf{A},\mathsf{B})\|_{{\mathcal{L}}({\mathcal{H}})},\quad\|{\widetilde{R}}_{k}(f,\mathsf{A},\mathsf{B})\|_{{\mathcal{L}}({\mathcal{H}})}\leq\,C_{N}\,\wp^{\rho}_{k+2}(f)\,\|{\rm ad}_{\mathsf{A}}^{k}(\mathsf{B})\|_{{\mathcal{L}}({\mathcal{H}})}\ .$ (B.9) ###### Lemma B.5. Let $k\in{\mathbb{N}}$ and $\mathsf{A},\mathsf{H}$ selfadjoint operators such that $\|{\rm ad}_{\mathsf{A}}^{j}(\mathsf{H})\|_{{\mathcal{L}}({\mathcal{H}})}<\infty,\qquad\forall\,1\leq j\leq k\ .$ (B.10) Let $g\in S^{\rho}$ with $\rho<0$. Then $\|{\rm ad}_{\mathsf{A}}^{j}(g(\mathsf{H}))\|_{{\mathcal{L}}({\mathcal{H}})}<\infty\,\quad\forall 1\leq j\leq k\ .$ ###### Proof. Take $N\geq k$ and use Helffer-Sjöstrand formula to write ${\rm ad}^{j}_{\mathsf{A}}(g(\mathsf{H}))=\frac{{\rm i}}{2\pi}\int_{{\mathbb{R}}^{2}}\frac{\partial{\widetilde{g}}_{N}(z)}{\partial\overline{z}}\,{\rm ad}^{j}_{\mathsf{A}}\big{(}(z-\mathsf{H})^{-1}\big{)}\,{\rm d}\overline{z}\wedge{\rm d}z.$ (B.11) As ${\rm ad}_{\mathsf{A}}\big{(}(z-\mathsf{H})^{-1}\big{)}=(z-\mathsf{H})^{-1}\,{\rm ad}_{\mathsf{A}}(\mathsf{H})\,(z-\mathsf{H})^{-1}$, by induction one gets for $j=1,\ldots,k$ ${\rm ad}_{\mathsf{A}}^{j}\big{(}(z-\mathsf{H})^{-1}\big{)}=\sum_{\ell=1}^{j}\sum_{k_{1}+\cdots+k_{\ell}=j\atop k_{1},\ldots,k_{\ell}\geq 1}c_{k_{1}\cdots k_{\ell}}^{\ell,j}\,(z-\mathsf{H})^{-1}\,{\rm ad}_{\mathsf{A}}^{k_{1}}(\mathsf{H})\,(z-\mathsf{H})^{-1}{\rm ad}_{\mathsf{A}}^{k_{2}}(\mathsf{H})\cdots(z-\mathsf{H})^{-1}\,{\rm ad}_{\mathsf{A}}^{k_{\ell}}(\mathsf{H})\,(z-\mathsf{H})^{-1}$ Using (B.10) and the estimate $\|(z-\mathsf{H})^{-1}\|_{{\mathcal{L}}({\mathcal{H}})}\leq{\left|{\rm Im}\,(z)\right|^{-1}}$, $\forall z\in{\mathbb{C}}\setminus{\mathbb{R}}$, one has for $j=1,\ldots,k$ $\|{\rm ad}_{\mathsf{A}}^{j}\big{(}(z-\mathsf{H})^{-1}\big{)}\|_{{\mathcal{L}}({\mathcal{H}})}\leq\sum_{\ell=1}^{j}C_{\ell}\,{\left|{\rm Im}\,(z)\right|}^{-\ell-1},\quad\forall z\in{\mathbb{C}}\setminus{\mathbb{R}}.$ Inserting this estimate into (B.11) and using (B.4) we bound for any $j=1,\ldots,k$ $\|{\rm ad}^{j}_{\mathsf{A}}(g_{J}(\mathsf{H}))\|_{{\mathcal{L}}({\mathcal{H}})}\lesssim\sum_{\ell=1}^{j}\int_{{\mathbb{R}}^{2}}\left|\frac{\partial{\widetilde{g}}_{N}(z)}{\partial\overline{z}}\right|\,\,{\left|{\rm Im}\,(z)\right|}^{-\ell-1}{\rm d}\overline{z}\wedge{\rm d}z\lesssim\wp^{\rho}_{N+2}(g)<\infty\ .$ ∎ ###### Lemma B.6. Let $g\in C^{\infty}_{c}({\mathbb{R}},{\mathbb{R}})$. Let $\mathsf{H},\mathsf{B}\in{\mathcal{L}}({\mathcal{H}})$ be selfadjoint. Then $\exists\,C>0$ such that $\|g(\mathsf{H}+\mathsf{B})-g(\mathsf{H})\|_{{\mathcal{L}}({\mathcal{H}})}\leq C\|\mathsf{B}\|_{{\mathcal{L}}({\mathcal{H}})}\ .$ (B.12) If $\mathsf{B}$ is compact on ${\mathcal{H}}$, so is $g(\mathsf{H}+\mathsf{B})-g(\mathsf{H})$. ###### Proof. Take $N\geq 1$. Using Helffer-Sjöstrand formula and the resolvent identity we obtain $g(\mathsf{H}+\mathsf{B})-g(\mathsf{H})=\frac{{\rm i}}{2\pi}\int_{{\mathbb{R}}^{2}}\frac{\partial{\widetilde{g}}_{N}(z)}{\partial\overline{z}}\,\big{(}z-(\mathsf{H}+\mathsf{B})\big{)}^{-1}\,\mathsf{B}\,(z-\mathsf{H})^{-1}\,{\rm d}\overline{z}\wedge{\rm d}z\ .$ Then use $\|\big{(}z-(\mathsf{H}+\mathsf{B})\big{)}^{-1}\|_{{\mathcal{L}}({\mathcal{H}})}$, $\|(z-\mathsf{H})^{-1}\|_{{\mathcal{L}}({\mathcal{H}})}\leq\left|{\rm Im}(z)\right|^{-1}$ for $z\in{\mathbb{C}}\setminus{\mathbb{R}}$ and (B.4). If $\mathsf{B}$ is compact then $\big{(}z-(\mathsf{H}+\mathsf{B})\big{)}^{-1}\,\mathsf{B}\,(z-\mathsf{H})^{-1}$ is a compact operator for any $z\in{\mathbb{C}}\setminus{\mathbb{R}}$. ∎ ## Appendix C Local energy decay estimates In this section we prove a local energy decay estimate starting from Mourre estimate. The result is essentially known but we could not find in the literature a statement exactly as the one we use in the paper, so we include here a proof, which follows closely the one of Lemma 4.1 of [24]. In this part we do not require pseudodifferential properties of the operators. We shall assume conditions (M1) and (M2) at page (M1). ###### Theorem C.1 (Local energy decay estimate). Fix $k\in{\mathbb{N}}$ and assume (M1)–(M2) with ${\mathtt{N}}\geq 4k+2$ and $\mathsf{K}=0$. Then for any interval $J\subset I$, any function $g_{J}\in C^{\infty}_{c}({\mathbb{R}},{\mathbb{R}}_{\geq 0})$ with ${\rm supp}\,g_{J}\subset I$, $g_{J}=1$ on $J$, there exists $C>0$ such that $\|\left\langle\mathsf{A}\right\rangle^{-k}\,e^{-{\rm i}\mathsf{H}t}\,g_{J}(\mathsf{H})\,\psi\|_{0}\leq C\left\langle t\right\rangle^{-k}\|\left\langle\mathsf{A}\right\rangle^{k}\,g_{J}(\mathsf{H})\psi\|_{0},\quad\forall t\in{\mathbb{R}},$ (C.1) for any $\psi$ such that the r.h.s. is finite. ###### Proof. Take $\chi(\xi):=\frac{1}{2}(1-\tanh\xi)$. Put $\eta(\xi):=\frac{1}{\sqrt{2}\cosh\xi}$ and note that $\chi^{\prime}=-\eta^{2},\qquad\left|\eta^{(m)}(\xi)\right|\leq C_{m}\,\eta(\xi),\quad\forall\xi\in{\mathbb{R}},\ \ \forall m\in{\mathbb{N}}.$ (C.2) Next we set for $a\in{\mathbb{R}}$, $s\geq 1$ and $\vartheta:=\frac{\theta}{2}$ (with $\theta$ of (M2) ) $\mathsf{A}_{t,s}:=\frac{1}{s}\big{(}\mathsf{A}-a-\vartheta t\big{)}$ and define via functional calculus the operators $\chi(\mathsf{A}_{t,s})$ and $\eta(\mathsf{A}_{t,s})$; both are bounded and selfadjoint on ${\mathcal{H}}$. To shorten the notation, from now on we write $\chi_{t,s}\equiv\chi(\mathsf{A}_{t,s})$, $\eta_{t,s}\equiv\eta(\mathsf{A}_{t,s})$, $g_{J}\equiv g_{J}(\mathsf{H})$ and $\psi_{t}:=e^{-{\rm i}t\mathsf{H}}\psi$. Note that $e^{-{\rm i}\mathsf{H}t}g_{J}(\mathsf{H})\psi=g_{J}(\mathsf{H})e^{-{\rm i}\mathsf{H}t}\psi\equiv g_{J}\psi_{t}$. The starting point of the proof is an energy estimate for the quantity $\|(\chi_{t,s})^{\frac{1}{2}}\,g_{J}\psi_{t}\|_{0}$. We have $\frac{d}{dt}\|(\chi_{t,s})^{\frac{1}{2}}\,g_{J}\psi_{t}\|_{0}^{2}=\frac{d}{dt}\left\langle\chi_{t,s}g_{J}\psi_{t},\,g_{J}\psi_{t}\right\rangle=\frac{\vartheta}{s}\|\eta_{t,s}\,g_{J}\,\psi_{t}\|_{0}^{2}+\left\langle{\rm i}[\mathsf{H},\chi_{t,s}]\,g_{J}\psi_{t},\,g_{J}\psi_{t}\right\rangle.$ (C.3) To evaluate the right hand side we shall use the commutator formulas in Lemma B.4, the identity ${\rm ad}^{j}_{\mathsf{A}_{t,s}}(\mathsf{H})=\frac{1}{s^{j}}{\rm ad}^{j}_{\mathsf{A}}(\mathsf{H})\ ,\quad\forall s\geq 1,\ \ \forall 1\leq j\leq{\mathtt{N}}$ (C.4) and the fact that all the operators ${\rm ad}^{j}_{\mathsf{A}}(\mathsf{H})$ are bounded $\forall 1\leq j\leq{\mathtt{N}}$ by (M1). The goal now is to estimate the second term in the right hand side of (C.3). For an arbitrary $f\in{\mathcal{H}}$ we write $\left\langle{\rm i}[\mathsf{H},\chi_{t,s}]f,f\right\rangle\stackrel{{\scriptstyle\eqref{r.exp},\eqref{ad.Ats}}}{{=}}-\frac{1}{s}\left\langle\eta_{t,s}^{2}{\rm i}[\mathsf{H},\mathsf{A}]f,f\right\rangle+\sum_{j=2}^{{\mathtt{N}}-1}\frac{1}{j!}\frac{1}{s^{j}}\left\langle\chi_{t,s}^{(j)}\,{\rm i}\,{\rm ad}_{\mathsf{A}}^{j}(\mathsf{H})f,f\right\rangle+\frac{1}{s^{\mathtt{N}}}\left\langle R_{\mathtt{N}}f,f\right\rangle$ (C.5) where $\chi_{t,s}^{(j)}:=\chi^{(j)}(\mathsf{A}_{t,s})$ and the remainder $R_{\mathtt{N}}$ fulfills the estimate (see (B.9)) $\|R_{\mathtt{N}}\|_{{\mathcal{L}}({\mathcal{H}})}\leq C_{\mathtt{N}}\,\|{\rm ad}_{\mathsf{A}}^{\mathtt{N}}(\mathsf{H})\|\leq C_{{\mathtt{N}}}.$ (C.6) Note that the constant $C_{\mathtt{N}}>0$ in the previous estimate is uniform in $a\in{\mathbb{R}}$. In the following we shall simply denote by $R_{\mathtt{N}}$ any bounded operator fulfilling an estimate like (C.6). Consider now the first term in the expansion (C.5) above. This time we use the left expansion (B.8) and write $\displaystyle\frac{1}{s}\left\langle\eta_{t,s}^{2}\,{\rm i}[\mathsf{H},\mathsf{A}]f,f\right\rangle$ $\displaystyle=\frac{1}{s}\left\langle\,{\rm i}[\mathsf{H},\mathsf{A}]\,\eta_{t,s}f,\,\eta_{t,s}f\right\rangle$ (C.7) $\displaystyle\quad+\sum_{j=2}^{{\mathtt{N}}-1}\frac{(-1)^{j-1}}{(j-1)!}\frac{1}{s^{j}}\langle\,{\rm i}\,{\rm ad}_{\mathsf{A}}^{j}(\mathsf{H})\,\eta_{t,s}^{(j-1)}f,\eta_{t,s}f\rangle+\frac{1}{s^{\mathtt{N}}}\left\langle R_{\mathtt{N}}f,f\right\rangle$ (C.8) where $R_{\mathtt{N}}$ is estimated as in (C.6). Consider now the second term in (C.5). From $\chi^{\prime}=-\eta^{2}$, we have by functional calculus $\chi^{(j)}(\mathsf{A}_{t,s})=\sum_{\ell=1}^{j}c_{\ell j}\,\eta^{(j-\ell)}(\mathsf{A}_{t,s})\,\eta^{(\ell)}(\mathsf{A}_{t,s})\,.$ Thus we get that $\displaystyle\frac{1}{s^{j}}\langle\chi_{t,s}^{(j)}\,{\rm i}\,{\rm ad}_{\mathsf{A}}^{j}(\mathsf{H})f,f\rangle$ $\displaystyle\stackrel{{\scriptstyle\eqref{l.exp}}}{{=}}\frac{1}{s^{j}}\sum_{\ell=1}^{j}c_{\ell j}\langle\,{\rm i}\,{\rm ad}_{\mathsf{A}}^{j}(\mathsf{H})\,\eta_{t,s}^{(\ell)}f,\eta_{t,s}^{(j-\ell)}f\rangle$ (C.9) $\displaystyle\quad+\sum_{\ell=1}^{j}\sum_{n=1}^{{\mathtt{N}}-j-1}\frac{c_{\ell jn}}{s^{j+n}}\langle\,{\rm i}\,{\rm ad}_{\mathsf{A}}^{j+n}(\mathsf{H})\,\eta_{t,s}^{(\ell+n)}f,\,\eta_{t,s}^{(j-\ell)}f\rangle+\frac{1}{s^{{\mathtt{N}}}}\left\langle R_{\mathtt{N}}f,f\right\rangle$ (C.10) By (C.5), (C.7), (C.9) we have found that $\left\langle{\rm i}[\mathsf{H},\chi_{t,s}]f,f\right\rangle$ is a sum of terms of the form $\displaystyle\left\langle{\rm i}[\mathsf{H},\chi_{t,s}]f,f\right\rangle=-\frac{1}{s}\left\langle\,{\rm i}[\mathsf{H},\mathsf{A}]\,\eta_{t,s}f,\,\eta_{t,s}f\right\rangle+\sum_{j=2}^{{\mathtt{N}}-1}\frac{1}{s^{j}}\sum_{n,\ell,m}\langle\,R_{n}\,\eta_{t,s}^{(\ell)}f,\eta_{t,s}^{(m)}f\rangle+\frac{1}{s^{\mathtt{N}}}\left\langle R_{\mathtt{N}}f,f\right\rangle$ where $R_{n},R_{\mathtt{N}}$ are bounded operators. Furthermore, from the second of (C.2) and the spectral theorem, we bound $\left|\langle\,R_{n}\,\eta_{t,s}^{(\ell)}f,\eta_{t,s}^{(m)}f\rangle\right|\leq C\,\|\eta_{t,s}f\|_{0}^{2}\ .$ (C.11) We thus obtain, for any $f\in{\mathcal{H}}$ and $s\geq 1$, the estimate $\left\langle{\rm i}[\mathsf{H},\chi_{t,s}]f,f\right\rangle\leq-\frac{1}{s}\left\langle\,{\rm i}[\mathsf{H},\mathsf{A}]\,\eta_{t,s}f,\,\eta_{t,s}f\right\rangle+\frac{C_{\mathtt{N}}}{s^{2}}\|\eta_{t,s}f\|_{0}^{2}+\frac{C_{\mathtt{N}}}{s^{\mathtt{N}}}\|f\|_{0}^{2}\ .$ (C.12) Now we evaluate such inequality at $f=g_{J}\psi_{t}$, getting $\left\langle{\rm i}[\mathsf{H},\chi_{t,s}]g_{J}\psi_{t},\,g_{J}\psi_{t}\right\rangle\leq-\frac{1}{s}\left\langle\,{\rm i}[\mathsf{H},\mathsf{A}]\,\eta_{t,s}g_{J}\psi_{t},\,\eta_{t,s}g_{J}\psi_{t}\right\rangle+\frac{C_{\mathtt{N}}}{s^{2}}\|\eta_{t,s}g_{J}\psi_{t}\|_{0}^{2}+\frac{C_{\mathtt{N}}}{s^{\mathtt{N}}}\|g_{J}\psi_{t}\|_{0}^{2}\ .$ (C.13) The next step is to prove that the first term in the right hand side above has a sign, up to higher order terms in $s^{-j}$. This is the point where the Mourre estimate (M2) comes into play. To see this, we analyze $\displaystyle\left\langle\,{\rm i}[\mathsf{H},\mathsf{A}]\,\eta_{t,s}\,g_{J}\psi_{t},\,\eta_{t,s}\,g_{J}\psi_{t}\right\rangle\equiv\left\langle\,{\rm i}[\mathsf{H},\mathsf{A}]\,\eta_{t,s}\,g_{I}\,g_{J}\psi_{t},\,\eta_{t,s}g_{I}\,g_{J}\psi_{t}\right\rangle$ (C.14) where we used that $g_{J}g_{I}=g_{J}$. Next we commute and expand in commutators $\eta_{t,s}g_{I}$: $\displaystyle\eta_{t,s}g_{I}=g_{I}\eta_{t,s}+[\eta_{t,s},g_{I}]\stackrel{{\scriptstyle\eqref{l.exp}}}{{=}}g_{I}\,\eta_{t,s}+\sum_{j=1}^{{\mathtt{N}}-2}\frac{c_{j}}{s^{j}}\,{\rm ad}_{\mathsf{A}}^{j}(g_{I}(\mathsf{H}))\,\eta_{t,s}^{(j)}+\frac{1}{s^{{\mathtt{N}}-1}}{\widetilde{R}}_{{\mathtt{N}}-1}\ ;$ (C.15) note that Lemma B.5 assures that the operators ${\rm ad}_{\mathsf{A}}^{j}(g_{I}(\mathsf{H}))$ are bounded $\forall j=1,\ldots,{\mathtt{N}}$, so is the operator ${\widetilde{R}}_{{\mathtt{N}}-1}$ which fulfills $\|{\widetilde{R}}_{{\mathtt{N}}-1}\|_{{\mathcal{L}}({\mathcal{H}})}\leq C_{\mathtt{N}}\,{\rm ad}_{\mathsf{A}}^{{\mathtt{N}}-1}(g_{I}(\mathsf{H}))<\infty\ .$ (C.16) Again in the following we shall denote by ${\widetilde{R}}_{{\mathtt{N}}-1}$ any operator fulfilling an estimate like (C.16). Inserting the expansion (C.15) into (C.14) one gets, with $w:=g_{J}\psi_{t}$, $\displaystyle\eqref{HAeta}=\left\langle{\rm i}[\mathsf{H},\mathsf{A}]\,g_{I}\,\eta_{t,s}w,\ g_{I}\,\eta_{t,s}w\right\rangle+\sum_{j=1}^{{\mathtt{N}}-2}\frac{c_{j}}{s^{j}}\sum_{n,\ell,m}\langle R_{n}\,\eta^{(\ell)}_{t,s}w,\ \eta^{(m)}_{t,s}w\rangle+\frac{1}{s^{{\mathtt{N}}-1}}\langle{\widetilde{R}}_{{\mathtt{N}}-1}w,w\rangle$ where each term of the form $\langle R_{n}\,\eta^{(\ell)}_{t,s}w,\ \eta^{(m)}_{t,s}w\rangle$ fulfills an estimate like (C.11). It is finally time to use the strict Mourre estimate: by assumption (M2) we have for $s\geq 1$ $\displaystyle\left\langle{\rm i}[\mathsf{H},\mathsf{A}]\,g_{I}\,\eta_{t,s}w,\ g_{I}\,\eta_{t,s}w\right\rangle$ $\displaystyle\geq\theta\|g_{I}\,\eta_{t,s}w\|_{0}^{2}\ .$ (C.17) Using again the expansion (C.15) and estimates (C.11), (C.16) we get therefore $\displaystyle\left\langle{\rm i}[\mathsf{H},\mathsf{A}]\,g_{I}\,\eta_{t,s}w,\ g_{I}\,\eta_{t,s}w\right\rangle$ $\displaystyle\geq\theta\|\,\eta_{t,s}g_{I}w\|_{0}^{2}-\frac{C_{\mathtt{N}}}{s}\|\eta_{t,s}w\|_{0}^{2}-\frac{C_{\mathtt{N}}}{s^{{\mathtt{N}}-1}}\|w\|_{0}^{2}.$ (C.18) This proves that the first term in the right hand side of (C.13) has a sign; we proceed from (C.13) and using inequality (C.18) (recall $w=g_{J}\psi_{t})$ we get $\left\langle{\rm i}[\mathsf{H},\chi_{t,s}]g_{J}\psi_{t},\,g_{J}\psi_{t}\right\rangle\leq-\frac{\theta}{s}\|\eta_{t,s}\,g_{J}\psi_{t}\|_{0}^{2}+\frac{C_{\mathtt{N}}}{s^{2}}\|\eta_{t,s}\,g_{J}\psi_{t}\|_{0}^{2}+\frac{C_{\mathtt{N}}}{s^{\mathtt{N}}}\|\,g_{J}\psi_{t}\|_{0}^{2}\ .$ (C.19) We come back to the estimate (C.3) of $\|(\chi_{t,s})^{\frac{1}{2}}\,g_{J}\psi_{t}\|_{0}$. We finally obtain, with $\vartheta=\frac{\theta}{2}$ and $s\geq 1$ sufficiently large, $\displaystyle\frac{d}{dt}\|(\chi_{t,s})^{\frac{1}{2}}\,g_{J}\psi_{t}\|_{0}^{2}$ $\displaystyle\stackrel{{\scriptstyle\eqref{min53}}}{{\leq}}\frac{1}{s}\left(\vartheta-\theta\right)\|\eta_{t,s}\,g_{J}\psi_{t}\|_{0}^{2}+\frac{C_{\mathtt{N}}}{s^{2}}\|\eta_{t,s}\,g_{J}\psi_{t}\|_{0}^{2}+\frac{C_{\mathtt{N}}}{s^{\mathtt{N}}}\|\,g_{J}\psi_{t}\|_{0}^{2}$ $\displaystyle\leq\frac{1}{s}\left(-\frac{\theta}{2}+\frac{C_{\mathtt{N}}}{s}\right)\|\eta_{t,s}\,g_{J}\psi_{t}\|_{0}^{2}+\frac{C_{\mathtt{N}}}{s^{\mathtt{N}}}\|\,g_{J}\psi_{t}\|_{0}^{2}$ So, for $s\geq 1$ sufficiently large, the first term in the right hand side above is negative and, using also that $e^{-{\rm i}t\mathsf{H}}$ is unitary and commutes with $g_{J}\equiv g_{J}(\mathsf{H})$, we get $\frac{d}{dt}\|(\chi_{t,s})^{\frac{1}{2}}\,g_{J}\psi_{t}\|_{0}^{2}\leq\frac{C_{\mathtt{N}}}{s^{\mathtt{N}}}\|\,g_{J}\psi_{0}\|_{0}^{2}\ .$ Integrating this inequality between $0$ and $t$ we find $\forall t>0$ $\|\chi^{\frac{1}{2}}\Big{(}\frac{\mathsf{A}-a-\vartheta t}{s}\Big{)}\,g_{J}\psi_{t}\|_{0}^{2}\leq\|\chi^{\frac{1}{2}}\Big{(}\frac{\mathsf{A}-a}{s}\Big{)}\,g_{J}\psi\|_{0}^{2}+\frac{C_{\mathtt{N}}\,t}{s^{\mathtt{N}}}\|\,g_{J}\psi\|_{0}^{2},$ uniformly for $a\in{\mathbb{R}}$ and $s\geq 1$ sufficiently large. We evaluate this inequality at $a=-\frac{\vartheta}{2}t$ and $s=\sqrt{t}$, obtaining for $t\geq 1$ sufficiently large, the minimal velocity estimate $\|\chi^{\frac{1}{2}}\Big{(}\frac{\mathsf{A}-\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\,g_{J}\psi_{t}\|_{0}\leq\|\chi^{\frac{1}{2}}\Big{(}\frac{\mathsf{A}+\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\,g_{J}\psi\|_{0}+C_{\mathtt{N}}\,t^{-\frac{{\mathtt{N}}}{4}+\frac{1}{2}}\|\,g_{J}\psi\|_{0}\ .$ (C.20) To conclude, take $k\in{\mathbb{N}}$ and consider $\|\left\langle\mathsf{A}\right\rangle^{-k}\,g_{J}\,\psi_{t}\|_{0}$. Clearly $\displaystyle\|\left\langle\mathsf{A}\right\rangle^{-k}\,g_{J}\,\psi_{t}\|_{0}$ $\displaystyle\leq\|\chi^{\frac{1}{2}}\Big{(}\frac{\mathsf{A}-\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\,\left\langle\mathsf{A}\right\rangle^{-k}\,g_{J}\,\psi_{t}\|_{0}$ (C.21) $\displaystyle\quad+\|\Big{(}1-\chi^{\frac{1}{2}}\Big{(}\frac{\mathsf{A}-\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\,\Big{)}\,\left\langle\mathsf{A}\right\rangle^{-k}\,g_{J}\,\psi_{t}\|_{0}$ (C.22) We estimate first (C.22). By Theorem B.2 (ii) we have $\|\Big{(}1-\chi^{\frac{1}{2}}\Big{(}\frac{\mathsf{A}-\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\,\Big{)}\,\left\langle\mathsf{A}\right\rangle^{-k}\|_{{\mathcal{L}}({\mathcal{H}})}\leq\sup_{\lambda\in{\mathbb{R}}}\left|\Big{(}1-\chi^{\frac{1}{2}}\Big{(}\frac{\lambda-\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\Big{)}\,\left\langle\lambda\right\rangle^{-k}\right|\leq C_{k}\left\langle t\right\rangle^{-k}.$ (C.23) To prove the last inequality, use that for $\lambda\geq\frac{\vartheta}{4}t$ then $\left\langle\lambda\right\rangle^{-k}\leq\left\langle t\right\rangle^{-k}$, whereas when $\lambda<\frac{\vartheta}{4}t$ then, being $\lambda\mapsto 1-\chi^{\frac{1}{2}}(\lambda)$ monotone increasing and exponentially decaying at $-\infty$, $1-\chi^{\frac{1}{2}}\Big{(}\frac{\lambda-\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\leq 1-\chi^{\frac{1}{2}}\Big{(}-\frac{\vartheta}{4}\sqrt{t}\Big{)}\leq C\left(e^{-\frac{\vartheta}{4}\sqrt{t}}\right)^{\frac{1}{2}}\leq C_{k}\left\langle t\right\rangle^{-k}.$ Next we estimate (C.21) using the minimal velocity estimate. As $\left\langle\mathsf{A}\right\rangle^{-k}$ is a bounded operator, $\displaystyle\|\chi^{\frac{1}{2}}\Big{(}\frac{\mathsf{A}-\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\,\left\langle\mathsf{A}\right\rangle^{-k}\,g_{J}\,\psi_{t}\|_{0}$ $\displaystyle\leq\|\chi^{\frac{1}{2}}\Big{(}\frac{\mathsf{A}-\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\,g_{J}\,\psi_{t}\|_{0}$ $\displaystyle\stackrel{{\scriptstyle\eqref{min61}}}{{\leq}}\|\chi^{\frac{1}{2}}\Big{(}\frac{\mathsf{A}+\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\,g_{J}\psi\|_{0}+C_{\mathtt{N}}\,t^{-\frac{{\mathtt{N}}}{4}+\frac{1}{2}}\|\,g_{J}\psi\|_{0}$ $\displaystyle\leq\|\chi^{\frac{1}{2}}\Big{(}\frac{\mathsf{A}+\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\,\left\langle\mathsf{A}\right\rangle^{-k}\|_{{\mathcal{L}}({\mathcal{H}})}\,\|\left\langle\mathsf{A}\right\rangle^{k}\,g_{J}\psi\|_{0}+C_{\mathtt{N}}\,t^{-\frac{{\mathtt{N}}}{4}+\frac{1}{2}}\|\,g_{J}\psi\|_{0}$ Again we have $\|\chi^{\frac{1}{2}}\Big{(}\frac{\mathsf{A}+\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\,\left\langle\mathsf{A}\right\rangle^{-k}\|_{{\mathcal{L}}({\mathcal{H}})}\leq\sup_{\lambda\in{\mathbb{R}}}\left|\chi^{\frac{1}{2}}\Big{(}\frac{\lambda+\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\,\left\langle\lambda\right\rangle^{-k}\right|\leq C_{k}\left\langle t\right\rangle^{-k},$ (C.24) since for $\lambda\leq-\frac{\vartheta}{4}t$ one has $\left\langle\lambda\right\rangle^{-k}\leq C\left\langle t\right\rangle^{-k}$, whereas in case $\lambda>-\frac{\vartheta}{4}t$, as $\lambda\mapsto\chi^{\frac{1}{2}}(\lambda)$ is monotone decreasing exponentially fast at $+\infty$, one has $\chi^{\frac{1}{2}}\Big{(}\frac{\lambda+\frac{\vartheta}{2}t}{\sqrt{t}}\Big{)}\leq\chi^{\frac{1}{2}}\Big{(}\frac{\vartheta}{4}\sqrt{t}\Big{)}\leq C\left(e^{-\frac{\vartheta}{4}\sqrt{t}}\right)^{\frac{1}{2}}\leq C_{k}\left\langle t\right\rangle^{-k}.$ Altogether, from (C.21), (C.22) we have proved that for $t\geq 1$ sufficiently large, $\displaystyle\|\left\langle\mathsf{A}\right\rangle^{-k}\,g_{J}\,\psi_{t}\|_{0}$ $\displaystyle\leq C_{k}\left\langle t\right\rangle^{-k}\|g_{J}\psi_{t}\|_{0}+C_{k}\left\langle t\right\rangle^{-k}\|\left\langle\mathsf{A}\right\rangle^{k}g_{J}\psi\|_{0}+C_{\mathtt{N}}\,t^{-\frac{{\mathtt{N}}}{4}+\frac{1}{2}}\|\,g_{J}\psi\|_{0}$ $\displaystyle\leq C_{k}\left\langle t\right\rangle^{-k}\,\|\left\langle\mathsf{A}\right\rangle^{k}g_{J}\psi\|_{0}$ provided ${\mathtt{N}}=4k+2$. This proves the estimate (C.1) for $t\geq 1$ sufficiently large, and it is also clearly true for $t$ in any bounded interval. ∎ ## References * [1] P. Anselone. Collectively compact operator approximation theory and applications to integral equations. Prentice-Hall, Inc., Englewood Cliffs, N. J., 1971. * [2] J. Arbunich, F. Pusateri, I.M. Sigal, A. Soffer. Growth of Sobolev norms for linear Schrödinger operators. ArXiv e-print, arXiv:2011.04570, 2020\. * [3] V. Bach, J. Fröhlich, I.M. Sigal, and A. Soffer. Positive commutators and the spectrum of Pauli-Fierz Hamiltonian of atoms and molecules. Comm. Math. Phys., 207(3):557–587, 1999. * [4] D. Bambusi. Reducibility of 1-d Schrödinger equation with time quasiperiodic unbounded perturbations, I. Trans. Amer. Math. Soc., 370(3):1823–1865, 2018. * [5] D. Bambusi, B. Grébert, A. Maspero, and D. Robert. Growth of Sobolev norms for abstract linear Schrödinger equations. J. Eur. Math. Soc. (JEMS), 2020. doi: 10.4171/JEMS/1017 * [6] D. Bambusi, B. Grébert, A. Maspero, and D. Robert. Reducibility of the quantum harmonic oscillator in d-dimensions with polynomial time-dependent perturbation. Anal. PDE, 11(3):775–799, 2018. * [7] D. Bambusi, B. Langella, and R. Montalto. Growth of Sobolev norms for unbounded perturbations of the Laplacian on flat tori. ArXiv e-print, arXiv:2012.02654, 2020. * [8] M. Berti and A. Maspero. Long time dynamics of Schrödinger and wave equations on flat tori. J. Diff. Eq., 267(2):1167 – 1200, 2019. * [9] J. Bourgain. Growth of Sobolev norms in linear Schrödinger equations with quasi-periodic potential. Comm. Math. Phys., 204(1):207–247, 1999. * [10] J. Bourgain. On growth of Sobolev norms in linear Schrödinger equations with smooth time dependent potential. J. Anal. Math., 77:315–348, 1999. * [11] O. Chodosh. Infinite matrix representations of isotropic pseudodifferential operators. Methods Appl. Anal., 18(4):351–371, 2011. * [12] Y. Colin de Verdière. Spectral theory of pseudodifferential operators of degree 0 and an application to forced linear waves. Anal. PDE, 13(5):1521–1537, 2020. * [13] Y. Colin de Verdière and Laure Saint-Raymond. Attractors for two-dimensional waves with homogeneous Hamiltonians of degree 0. Comm. Pure Appl. Math., 73(2):421–462, 2020. * [14] H. Cycon, R. Froese, W. Kirsch, and B. Simon. Schrödinger operators with application to quantum mechanics and global geometry. Texts and Monographs in Physics. Springer-Verlag, Berlin, study edition, 1987. * [15] E. Davies. The functional calculus. J. London Math. Soc. (2), 52(1):166–176, 1995. * [16] J.-M. Delort. Growth of Sobolev norms of solutions of linear Schrödinger equations on some compact manifolds. Int. Math. Res. Not. IMRN, (12):2305–2328, 2010. * [17] J.-M. Delort. Growth of Sobolev norms for solutions of time dependent Schrödinger operators with harmonic oscillator potential. Comm. Partial Differential Equations, 39(1):1–33, 2014. * [18] J. Dereziński and C. Gérard. Scattering theory of classical and quantum $N$-particle systems. Texts and Monographs in Physics. Springer-Verlag, Berlin, 1997. * [19] M. Dimassi and J. Sjostrand. Spectral Asymptotics in the Semi-Classical Limit. London Mathematical Society Lecture Note Series. Cambridge University Press, 1999. * [20] S. Dyatlov and M. Zworski. Microlocal analysis of forced waves. Pure Appl. Anal. , 1(3): 359–384, 2019. * [21] P. Duclos, O. Lev, and P. Sťovíček. On the energy growth of some periodically driven quantum systems with shrinking gaps in the spectrum. J. Stat. Phys., 130(1):169–193, 2008. * [22] E. Faou and P. Raphael. On weakly turbulent solutions to the perturbed linear harmonic oscillator. ArXiv e-print, arXiv:2006.08206, 2020\. * [23] C. Gérard and I. M. Sigal. Space-time picture of semiclassical resonances. Comm. Math. Phys., 145(2):281–328, 1992. * [24] E. Grenier, T. Nguyen, F. Rousset, and A. Soffer. Linear inviscid damping and enhanced viscous dissipation of shear flows by using the conjugate operator method. J. Funct. Anal., 278(3):108339, 27, 2020. * [25] E. Haus and A. Maspero. Growth of Sobolev norms in time dependent semiclassical anharmonic oscillators. J. Funct. Anal. , 278(2), 108316, 2020. * [26] B. Helffer and J. Sjöstrand. Équation de Schrödinger avec champ magnétique et équation de Harper. In Schrödinger operators, volume 345 of Lecture Notes in Phys., 118–197. Springer, Berlin, 1989. * [27] L. Hörmander. The Weyl calculus of pseudodifferential operators. Comm. Pure Appl. Math., 32(3):360–444, 1979. * [28] L. Hörmander. The analysis of linear partial differential operators I-IV. Grundlehren der mathematischen Wissenschaften 256. Springer-Verlag, 1985\. * [29] W. Hunziker and I. M. Sigal. Time-dependent scattering theory of n-body quantum systems. Reviews in Mathematical Physics, 12(08):1033–1084, 2000. * [30] W. Hunziker, I. M. Sigal, and A. Soffer. Minimal escape velocities. Comm. Partial Differential Equations, 24(11-12):2279–2295, 1999\. * [31] A. Jensen, É. Mourre, and P. Perry. Multiple commutator estimates and resolvent smoothness in quantum scattering theory. Ann. Inst. H. Poincaré Phys. Théor., 41(2):207–225, 1984\. * [32] T. Kato. Perturbation theory for linear operators. Classics in Mathematics. Springer-Verlag, Berlin, 1995. * [33] A. Maspero and D. Robert. On time dependent Schrödinger equations: Global well-posedness and growth of Sobolev norms. J. Fun. Anal., 273(2):721 – 781, 2017. * [34] A. Maspero. Lower bounds on the growth of Sobolev norms in some linear time dependent Schrödinger equations. Math. Res. Lett., 26(4):1197–1215, 2019. * [35] R. Montalto. Growth of Sobolev norms for time dependent periodic Schrödinger equations with sublinear dispersion. J. Diff. Eq., 266(8):4953 – 4996, 2019. * [36] E. Mourre. Absence of singular continuous spectrum for certain selfadjoint operators. Comm. Math. Phys., 78(3):391–408, 1980/81. * [37] G. Nenciu. Adiabatic theory: stability of systems with increasing gaps. Annales de l’I. H. P, 67-4:411–424, 1997. * [38] M. Reed and B. Simon. Methods of modern mathematical physics. I. Academic Press, Inc., New York, second edition, 1980. Functional analysis. * [39] J. Saranen and G. Vainikko. Periodic integral and pseudodifferential equations with numerical approximation. Springer Monographs in Mathematics. Springer-Verlag, Berlin, 2002. * [40] M. Shubin. Pseudodifferential operators and spectral theory. Springer-Verlag, Berlin, second edition, 2001. * [41] I.M. Sigal and A. Soffer. Local decay and velocity bounds for quantum propagation. preprint (Princeton), 1988. http://www.math.toronto.edu/sigal/publications/SigSofVelBnd.pdf * [42] E. Skibsted. Propagation estimates for $N$-body Schroedinger operators. Comm. Math. Phys., 142(1):67–98, 1991. * [43] C. Sogge. Hangzhou Lectures on Eigenfunctions of the Laplacian. Princeton University Press, 2014. * [44] L. Thomann. Growth of Sobolev norms for linear Schrödinger operators. ArXiv e-print, arXiv:2006.02674, 2020\.
# An ensemble solver for segregated cardiovascular FSI Xue Li Department of Applied and Computational Mathematics and Statistics University of Notre Dame Notre Dame, IN 46556 <EMAIL_ADDRESS> &Daniele E. Schiavazzi Department of Applied and Computational Mathematics and Statistics University of Notre Dame Notre Dame, IN 46556 <EMAIL_ADDRESS> ###### Abstract Computational models are increasingly used for diagnosis and treatment of cardiovascular disease. To provide a quantitative hemodynamic understanding that can be effectively used in the clinic, it is crucial to quantify the variability in the outputs from these models due to multiple sources of uncertainty. To quantify this variability, the analyst invariably needs to generate a large collection of high-fidelity model solutions, typically requiring a substantial computational effort. In this paper, we show how an explicit-in-time _ensemble_ cardiovascular solver offers superior performance with respect to the embarrassingly parallel solution with implicit-in-time algorithms, typical of an inner-outer loop paradigm for non-intrusive uncertainty propagation. We discuss in detail the numerics and efficient distributed implementation of a segregated FSI cardiovascular solver on both CPU and GPU systems, and demonstrate its applicability to idealized and patient-specific cardiovascular models, analyzed under steady and pulsatile flow conditions. _K_ eywords Ensemble solvers $\cdot$ Uncertainty Quantification $\cdot$ Computational Hemodynamics $\cdot$ Explicit Time Integration $\cdot$ Biomechanics ## 1 Introduction In this study we focus on the development of efficient solvers for complex fluid-structure interaction (FSI) phenomena arising in cardiovascular (CV) hemodynamics. For this and many other applications, output variability is induced by uncertainty or ignorance in the input processes, e.g., material property distribution, physiologically sound boundary conditions or model anatomy, resulting from operator-dependent image volume segmentation. In this context, the new paradigm of Uncertainty Quantification (UQ) is rapidly becoming an integral part of the modeling exercise, and an indispensable tool to rigorously quantify confidence in the simulation outputs, enabling robust predictions of greater clinical impact. However, running a complete UQ study on a large scale cardiovascular model is typically associated with a substantial computational cost. Non-intrusive approaches for the solution of the forward problem in uncertainty quantification (also known as _uncertainty propagation_) typically consider the underlying deterministic solver as a black box (see Figure 1), requiring model solutions at various parameter realizations to be _independently_ (and possibly simultaneously) computed. The scalability of this paradigm is limited due to two main reasons. First, in the embarrassingly parallel solution of multiple instances of the same problem, a large number of operations is repeated. Second, the solution of large linear systems of equations from numerical integration in time with implicit schemes presents, in general, a less than ideal scalability when computed on large multi-core architectures. Figure 1: Schematic representation of the forward and inverse problems in UQ. To tackle these challenges, we propose an efficient computational approach for solving multiple instances of the same model (so-called model _ensemble_) at the same time, in a highly scalable fashion, enabled by CPU/GPU implementations of numerical integration schemes that rely heavily on distributed sparse matrix-vector products. We validate our approach by characterizing the effect of variability in vessel wall thickness and elastic modulus on the mechanical response of ideal and patient-specific cardiovascular models analyzed under steady and pulsatile flow conditions. A stochastic model for the spatial distribution of thickness and elastic modulus is provided, in this study, by approximating a Gaussian random field with Matérn covariance through the solution of a stochastic partial differential equation on a triangular finite element mesh of the vessel lumen [1]. Additionally, we follow a one-way coupled, segregated approach to fluid- structure interaction, where the wall stress is computed by an implicit Variational Multiscale fluid solver and passed to a structural, three-d.o.f.s shell model of the vascular walls [2]. Unique contributions of our approach are: 1. 1. We propose, for the first time, an ensemble solver in the context of cardiovascular hemodynamics for UQ, with the aim of drastically reduce the computational effort to perform campaigns of high-fidelity model solutions. 2. 2. Our approach paves the way to a fully explicit treatment of fluid structure interaction in cardiovascular modeling, whose potential has not yet been fully explored in the literature. In our opinion, this new paradigm will provide simpler approaches to simulating complex physiological dysfunctions (e.g. aortic dissection, involving large vessel deformations and contact/auto- contact phenomena, see [3]). 3. 3. The combination of explicit time integration schemes with the solution of model ensembles leads to an efficient distribution of computing load and memory usage in the GPU, enabling scalability in a way that is currently not possible with embarrassingly parallel model solutions and implicit solvers. We motivate the computational advantage of explicit-in-time ensemble solvers through a back-of-the-envelope argument. It is well known how explicit numerical integration schemes are only conditionally stable. For structural problems, the larger stable time step is determined through a CFL condition as $0.9\cdot\Delta t$ where $\Delta t$ is the amount of time required by an elastic wave to cross the smallest element in the mesh. A reasonable value of $\Delta t$ for CV modeling is $\Delta t=l_{e,\text{min}}/c$, where $l_{e,\text{min}}$ is the diameter of the circle/sphere inscribed in the smallest element in the mesh, $c=\sqrt{E/\rho}$ is the elastic wave speed, $E$ and $\rho$ the elastic modulus and density of the vascular tissue (assumed homogeneous and isotropic in the current argument). Typical values of $l_{e,\text{min}}=1.0\times 10^{-3}$ m, $E=0.7$ MPa and $\rho=1.06$ kg/m3 lead to a stable time step equal to approximately $0.9\cdot 1.2\times$10${}^{-6}\approx 1.0\times$10-6. In contrast, the typical time step adopted by implicit CV FSI solvers is one millisecond (an upper bound, typically much smaller). At every time step, these solvers require multiple linear systems to be solved with iterative methods, each consisting in several matrix-vector products per iteration. Thus, if we assume an average of 10 non- linear iterations consisting of 10 linear solver iterations each (with two matrix-vector multiplications per iteration), an explicit solver is only a factor of five more expensive. However, the cost of explicit methods can be further reduced through several means, for example, increasing the critical time step via mass scaling (see, e.g., [4]), selectively updating the stiffness, damping and mass matrices between successive time steps to avoid the repeated assembly of element matrices (in the linear regime), and by solving multiple realizations of the boundary conditions, material properties and geometry at the same time. Additionally, large matrix-vector operations needed by the explicit solution of model ensembles are particularly well suited for GPU computing. In summary, explicit time integration schemes have many advantages over implicit approaches in the simulation of phenomena occurring over small time intervals and, combined with the solution of model ensembles, bear substantial potential to boost the efficiency of solving CV models on modern GPU-based systems. Even though our work focuses on CV hemodynamics, the proposed solver paradigm is applicable to study fluid- structure interaction phenomena in other fields, but its efficiency is affected by the stiffness and mass properties in the selected application. Use of explicit structural solvers in cardiovascular flow problems is mainly related to their flexibility in modeling complex contact configurations. Studies involving coronary stent deployment following endoscopic balloon inflation are proposed, e.g., in [5, 6], and coupling with an implicit fluid solver is discussed in [7]. Implementation of structural explicit solvers on GPU are discussed in various studies in the literature. In [8] the authors describe in detail an application involving thin shells, while an overview on applications in biomechanics is discussed in [9]. Additionally, Ensemble methods for fluid problems have been recently proposed by [10, 11, 12, 13, 14] in the context of the Navier-Stokes equations with distinct initial conditions and forcing terms. This is based on the observation that solution of linear systems is responsible for a significant fraction of the overall running time for linearly implicit methods, and that is far more efficient to solve multiple times a system of equation with the same coefficient matrix and different right-hand-side than different systems altogether. Extensions have also been proposed to magnetohydrodynamics [15], natural convection problems [16] and parametrized flow problems [17, 18]. We note how, in our case, there is no approximation introduced in the formulation of the ensemble numerical scheme. In addition, no ensemble method appears to be available from the literature in the context of fluid-structure interaction problems. The basic methodology behind the proposed solver is discussed in Section 2, including the generation of random field material properties, the structural finite element formulation, the variational multiscale fluid solver, thier weak coupling and CPU/GPU implementation. Validation of the proposed approach is discussed in Section 3 with reference to an idealized model of the thoracic aorta and a patient-specific coronary model. Performance and scalability of the approach is discussed in Section 3.3 followed by conclusions and future work in Section 4. ## 2 Methodology ### 2.1 Random field material properties A homogeneous and isotropic vascular tissue with uncertain elastic modulus and thickness is assumed in this study, modeled through a Gaussian random field with Matérn covariance. A Gaussian marginal distribution appears to be the simplest idealized distribution compatible with the scarce experimental observations, while the choice of a Matérn covariance relates to its finite differentiability, which make this model more desirable than other kernels [19]. Let $\left\|\cdot\right\|$ denote the Euclidean distance in $\mathbb{R}^{d}$. The Matérn covariance between two points at distance $\|\bm{h}\|$ is $r(\|\bm{h}\|)=\frac{\sigma^{2}}{2^{\nu-1}\Gamma(\nu)}(\kappa\|\bm{h}\|)^{\nu}K_{\nu}(\kappa\|\bm{h}\|),\,\,\bm{h}\in\mathbb{R}^{d}$ (1) where $\Gamma(\cdot)$ is the gamma function, $K_{\nu}$ is the modified Bessel function of the second kind, $\sigma^{2}$ is the marginal variance, $\nu$ is a scaling parameter which determines the mean square differentiability of the underling process, and $\kappa$ is related to the correlation length $\rho=\sqrt{8\nu}/\kappa$, i.e., the distance which corresponds to a correlation of approximately $0.1$, for all $\nu$. It is known from the literature [20, 21] how Gaussian random fields with Matérn covariance can be obtained as solutions of a linear fractional stochastic partial differential equation (SPDE) of the form $(\kappa^{2}-\Delta)^{\alpha/2}\,x(\bm{s})=\mathcal{W}(\bm{s}),\,\,\bm{s}\in\mathbb{R}^{d},$ (2) where $\alpha=\nu+d/2$, $\kappa>0$, $\nu>0$, $\mathcal{W}(\bm{s})$ is a white noise spatial process, $\Delta=\Sigma_{i}\,\partial^{2}/\partial\,s_{i}^{2}$, and the marginal variance is $\sigma^{2}=\frac{\Gamma(\nu)}{\Gamma(\nu+d/2)(4\pi)^{d/2}\kappa^{2\nu}}.$ Since the lumen wall is modelled with a surface of triangular elements, we are interested in generating realizations from discretely indexed Gaussian random fields. This is achieved through an approximate stochastic weak solution of the SPDE (2), as discussed in [1]. We construct a discrete approximation of the solution $x(\bm{s})$ using a linear combination of basis functions, $\\{\psi_{k}\\},k=1,\dots,n$, and appropriate weights, $\\{w_{k}\\},k=1,\dots,n$, i.e., $x(\bm{s})=\sum_{k=1}^{n}\,\psi_{k}(\bm{s})\,w_{k}$. We then introduce an appropriate Sobolev space with inner product $\langle\cdot,\cdot\rangle$, a family of _test functions_ $\\{\varphi_{k}\\},k=1,\dots,n$, and derive a Galerkin functional for (2) of the form $\langle\varphi_{i},(\kappa^{2}-\Delta)^{\alpha/2}\,\psi_{j}\rangle\,w_{j}=\langle\varphi_{i},\mathcal{W}\rangle$ (3) We then choose $\varphi_{k}=(\kappa^{2}-\Delta)^{1/2}\,\psi_{k}$ for $\alpha=1$ and $\varphi_{k}=\psi_{k}$ for $\alpha=2$, leading to precision matrices $\bm{Q}_{\alpha}$ expressed as $\displaystyle\bm{Q}_{\alpha}=\varkappa^{2}\bm{C+G}$ $\displaystyle\text{for}\,\,\alpha=1$ (4) $\displaystyle\bm{Q}_{\alpha}=(\varkappa^{2}\bm{C+G})^{T}\bm{C}^{-1}(\varkappa^{2}\bm{C+G})$ $\displaystyle\text{for}\,\,\alpha=2$ $\displaystyle\bm{Q}_{\alpha}=(\varkappa^{2}\bm{C+G})^{T}\bm{C}^{-1}\bm{Q}_{\alpha-2}\bm{C}^{-1}(\varkappa^{2}\bm{C+G})$ $\displaystyle\text{for}\,\,\alpha>2,$ where, for $\alpha\geq 3$ a recursive Galerkin formulation is used, letting $\alpha=2$ on the left-hand side of equation (2) and replacing the right-hand side with a field generated by $\alpha-2$, assigning $\varphi_{k}=\psi_{k},\,k=1,\dots,n$. Note how the use of piecewise linear basis functions $\\{\psi_{k}\\},k=1,\dots,n$, lead to matrices $\bm{G}_{ij}=\langle\nabla\psi_{i},\nabla\psi_{j}\rangle\,\,\text{and}\,\,\bm{C}_{ij}=\langle\psi_{i},\psi_{j}\rangle,$ (5) that are _sparse_ , and often found in the finite element discretization of second order elliptic PDEs. However, the precision matrices $\bm{Q}_{\alpha}$ are, in general, not sparse as they contain the inverse $\bm{C}^{-1}$. Thus, by replacing the matrix $\bm{C}$ with the _lumped_ diagonal matrix $\widetilde{\bm{C}}$, sparsity is restored and $\bm{Q}_{\alpha}$ can be efficiently manipulated and decomposed through fast routines for sparse linear algebra available on a wide range of architectures. In addition, the introduction of $\widetilde{\bm{C}}$ leads to non zero terms on each row only for the _immediate neighbors_ $\mathcal{I}(s_{k})$ of a given node $k$ on the triangular surface mesh, since the basis function $\\{\psi_{k}\\}$ is supported only on the elements connected to node $k$. This reduces the Gaussian random field to a Gaussian Markov random field, for which $\begin{split}&\rho\left(x(s_{i}),x(s_{k})\,|\,x(s_{j}),\,s_{j}\in\mathcal{I}(s_{i})\right)=\\\ &=\rho\left(x(s_{i})\,|\,x(s_{j}),\,s_{j}\in\mathcal{I}(s_{i})\right),\end{split}$ (6) or, in other words, _given the value of $x$ on its neighbors_, at any node $k$ the random field $x(s_{k})$ is statistically independent from any other location. The approximation error introduced in (6) is, however, small and often negligible in applications [22]. The interested reader is referred to [1] for additional detail on the derivation of $\bm{Q}_{\alpha}$. Numerically generated realizations for various correlation lengths are shown in Figure 3 on an ideal cylindrical representation of the descending thoracic aorta, while Figure 3 shows the agreement of the generated field with the Matérn model. (a) $\rho=.95$ cm (b) $\rho=3.7$ cm (c) $\rho=7.2$ cm Figure 2: Random field generated on a cylindrical mesh for various correlation lengths. Figure 3: Comparison between spatial correlations from a numerically generated field $x(\bm{s})$ with precision matrix $\bm{Q}_{\alpha}$ and the exact Matérn model. ### 2.2 A segregated solver for fluid-structure interaction phenomena #### 2.2.1 Finite element model for the vessel wall We use a small strain, linear, 3 d.o.f. elastic thin shell which allows for a full compatibility between the fluid mesh discretized with tetrahedral elements and the solid walls. The in-plane stiffness of the shell is complemented with a transverse shear stiffness which provides stability under transverse loading [2]. Using a superscript $l$ and lower case $x,y,z$ to indicate quantities expressed in the local shell reference frame, we introduce a constitutive relation in Voigt notation expressed as $\bm{\sigma}^{l}=\bm{C}\cdot\bm{\varepsilon}^{l},\,\,\text{with}\,\,\bm{\sigma}^{l}=\begin{bmatrix}\sigma_{xx}\\\ \sigma_{yy}\\\ \tau_{xy}\\\ \tau_{xz}\\\ \tau_{yz}\end{bmatrix},\bm{\varepsilon}^{l}=\begin{bmatrix}\partial u_{x}/\partial x\\\ \partial u_{y}/\partial y\\\ \left(\partial u_{x}/\partial y+\partial u_{y}/\partial x\right)\\\ \partial u_{z}/\partial x\\\ \partial u_{z}/\partial y\end{bmatrix}.$ (7) We assume $\varepsilon^{l}_{zz}=0$, i.e., zero deformation through the thickness, disregarding the effect of both the normal pressure acting at the lumen surface and the Poisson effect due to the membrane deformations. Strains $\bm{\varepsilon}^{l}$ and nodal displacements $\bm{u}$ are related through the matrix $\bm{B}$ of shape function derivatives for a linear triangular element, i.e. $\bm{\varepsilon}^{l}=\bm{B}\,\bm{u}=\frac{1}{2\,A_{e}}\begin{bmatrix}y_{23}&0&0&y_{31}&0&0&y_{12}&0&0\\\ 0&x_{32}&0&0&x_{13}&0&0&x_{21}&0\\\ x_{32}&y_{23}&0&x_{13}&y_{31}&0&x_{21}&y_{12}&0\\\ 0&0&y_{23}&0&0&y_{31}&0&0&y_{12}\\\ 0&0&x_{32}&0&0&x_{13}&0&0&x_{21}\end{bmatrix}\begin{bmatrix}u_{x,1}\\\ u_{y,1}\\\ u_{z,1}\\\ u_{x,2}\\\ u_{y,2}\\\ u_{z,2}\\\ u_{x,3}\\\ u_{y,3}\\\ u_{z,3}\end{bmatrix}$ (8) where $x_{j},y_{j},\,j\in\\{1,2,3\\}$ are the local coordinates of the $j$-th element node, $x_{ij}=x_{i}-x_{j}$ (similarly for $y_{ij}$), $u_{x,j},u_{y,j},u_{z,j}$ are the local nodal displacements and $A_{e}$ is the triangular element area. The constitutive matrix is expressed as $\bm{C}=\frac{E}{(1-\nu^{2})}\begin{bmatrix}1&\nu&0&0&0\\\ \nu&1&0&0&0\\\ 0&0&0.5\,(1-\nu)&0&0\\\ 0&0&0&0.5\,k\,(1-\nu)&0\\\ 0&0&0&0&0.5\,k\,(1-\nu)\end{bmatrix}$ (9) where $E$ and $\nu$ are the Young’s modulus and Poisson’s ratio coefficient, respectively, and the shear factor $k$ accounts for a parabolic variation of transverse shear stress through the shell thickness (assumed as $5/6$ for a shell with rectangular cross section). Finally, the local element stiffness matrix $\bm{k}_{e}\in\mathbb{R}^{9\times 9}$ can be expressed as $\bm{k}_{e}=\int_{\Omega^{s}_{e}}\,\bm{B}^{T}\bm{C}\bm{B}\,\,\mathrm{d}\Omega^{s}_{e}=\sum_{i=1}^{n_{\text{gp}}}\,\bm{B}^{T}\bm{C}\bm{B}\,A_{e}\,\zeta_{i}\,w_{i},$ (10) where $n_{\text{gp}}$ is the total number of integration points and $\zeta_{i},w_{i},i\in\\{1,2,\dots,n_{\text{gp}}\\}$, are the element thickness and integration rule weights, respectively. In this study, we adopt a three- point Gauss integration rule to capture a linear variation for $E$ and $\zeta$ through each element, generated from a Gauss Markov random field with Matérn covariance $\bm{Q}_{\alpha}$, i.e., $E=E(x,y,\omega)$ and $\zeta=\zeta(x,y,\omega)$. Nodal vectors $\bm{E}\sim\mathcal{N}(\overline{\bm{E}},\bm{Q}^{-1}_{\alpha})$ and $\bm{\zeta}\sim\mathcal{N}(\overline{\bm{\zeta}},\bm{Q}^{-1}_{\alpha})$ are generated as $\bm{E}=\overline{\bm{E}}+(\bm{L}^{T})^{-1}\bm{z},\,\,\text{and}\,\,\bm{\zeta}=\overline{\bm{\zeta}}+(\bm{L}^{T})^{-1}\bm{z},$ (11) where $\bm{z}\sim\mathcal{N}(\bm{0},\bm{I}_{n})$ is a vector with standard Gaussian components and $\bm{L}$ is the sparse Cholesky factor of $\bm{Q}_{\alpha}$, i.e., $\bm{Q}_{\alpha}=\bm{L}\,\bm{L}^{T}$. The covariance matrix $\bm{Q}_{\alpha}$ is assembled in compressed sparse column (CSC) format and the Cholesky decomposition is computed using the _cholmod_ routine provided by the _scikit-sparse_ Python library. Finally, the product $(\bm{L}^{T})^{-1}\bm{z}$ is performed using the _solve_Lt_ routine, as the solution of a triangular system. #### 2.2.2 Variational multiscale finite element fluid solver The evolution of blood flow and pressure in the human cardiovascular system can be modeled using the Navier-Stokes equations. Even though many simplifying assumptions can be made to the equations to reduce the computational complexity, here we focus on high-fidelity models, i.e., models associated with large discretizations of a three-dimensional ($n_{\text{sd}}=3$) fluid domain $\Omega^{f}\subseteq\mathbb{R}^{n_{\text{sd}}}$. The boundary $\Gamma^{f}$ of $\Omega^{f}$ coincides with the mid-plane of the solid domain $\Omega^{s}$, and is partitioned into $\Gamma^{f}=\Gamma_{g}^{f}\cup\Gamma_{h}^{f}\cup\Gamma_{s}^{f}$ which correspond to the application of Dirichlet, Neumann boundary conditions and interaction with the solid, respectively. Consider also the vector fields $\bm{h}:\Gamma_{h}\times(0,T)\to\mathbb{R}^{n_{sd}}$, $\bm{g}:\Gamma_{g}\times(0,T)\to\mathbb{R}^{n_{sd}}$, $\bm{f}:\Omega\times(0,T)\to\mathbb{R}^{n_{sd}}$ and $\bm{v}^{0}:\Omega\to\mathbb{R}^{n_{sd}}$. We would like to solve the problem of finding $\bm{v}(\bm{X},t)$ and $p(\bm{X},t)$, $\forall\,\bm{X}\in\Omega$, $\forall\,t\in[0,T]$ such that $\begin{cases}\rho\,\dot{\bm{v}}+\rho\,\bm{v}\cdot\nabla\bm{v}=-\nabla p+\nabla\cdot\bm{\tau}+\bm{f}&(\bm{X},t)\in\Omega\times(0,T)\\\ \nabla\cdot\bm{v}=0&(\bm{X},t)\in\Omega\times(0,T),\\\ \end{cases}$ (12) subject to the boundary conditions $\begin{cases}\bm{v}=\bm{g}&(\bm{X},t)\in\Gamma_{g}\times(0,T)\\\ \bm{t}_{\bm{n}}=\bm{\sigma}\cdot\bm{n}=[-p\,\bm{I}+\bm{\tau}]\cdot\bm{n}=\bm{h}&(\bm{X},t)\in\Gamma_{h}\times(0,T)\\\ \bm{t}_{\bm{n}}=\bm{t}^{f}&(\bm{X},t)\in\Gamma_{s}\times(0,T)\\\ \bm{v}(\bm{X},0)=\bm{v}^{0}(\bm{X})&\bm{X}\in\Omega,\end{cases}$ (13) where $\bm{\tau}=\mu(\nabla\bm{v}+\nabla\bm{v}^{\,T})$ is the viscous stress tensor resulting from considering blood as a Newtonian fluid. Solution of (12) in weak form requires to define four approximation spaces, i.e., two trial spaces for the velocity $\bm{v}$ and pressure $p$ $\begin{split}\mathscr{S}_{k}^{h}&=\Big{\\{}\bm{v}\,|\,\bm{v}(\cdot,t)\in\bm{H}^{1}(\Omega),\mathnormal{t}\in[0,\mathnormal{T}],\\\ &\bm{v}\,|_{\bm{x}\in\Omega_{e}}\in\mathnormal{P}_{k}(\Omega_{e}),\,\bm{v}(\cdot,t)=\bm{g}\text{ on }\Gamma_{g}\Big{\\}},\\\ \mathscr{P}_{k}^{h}&=\left\\{p\,|\,p(\cdot,t)\in L^{2}(\Omega),t\in[0,\mathnormal{T}],\,p\,|_{\bm{x}\in\bar{\Omega}_{e}}\in\mathnormal{P}_{k}(\Omega_{e})\right\\},\end{split}$ and two test spaces for $\vec{w}$ and $q$ $\begin{split}\mathscr{W}_{k}^{h}=\Big{\\{}&\bm{w}\,|\,\bm{w}(\cdot,t)\in\bm{H}^{1}(\Omega),\,\mathnormal{t}\in[0,\mathnormal{T}],\\\ &\bm{w}\,|_{\bm{x}\in\Omega_{e}}\in\mathnormal{P}_{k}(\Omega_{e}),\,\bm{w}(\cdot,t)=\bm{0}\text{ on }\Gamma_{g}\Big{\\}},\,\,\mathscr{Q}_{k}^{h}=\mathscr{P}_{k}^{h}\end{split}$ where $\bm{H}^{1}(\Omega)$ is the Sobolev space of function triplets in $L^{2}(\Omega)$ with derivatives in $L^{2}(\Omega)$ and $P_{k}(\Omega)$ is the space of polynomials of order $k$ in $\Omega$. The above spaces are separated into large and a small scale contributions $\mathscr{S}_{k}^{h}=\overline{\mathscr{S}_{k}^{h}}\oplus\widetilde{\mathscr{S}_{k}^{h}}$, $\mathscr{W}_{k}^{h}=\overline{\mathscr{W}_{k}^{h}}\oplus\widetilde{\mathscr{W}_{k}^{h}}$ and $\mathscr{P}_{k}^{h}=\overline{\mathscr{P}_{k}^{h}}\oplus\widetilde{\mathscr{P}_{k}^{h}}$ with a corresponding decomposition of velocity and pressures as $\bm{v}=\overline{\bm{v}}+\widetilde{\bm{v}}$ and $p=\overline{p}+\widetilde{p}$, respectively. This decomposition is introduced in a weak form of the Navier-Stokes equations (12) and a _closure_ obtained by expressing the small scale variables $\widetilde{\bm{v}}$ and $\widetilde{p}$ in terms of their large scale counterparts $\overline{\bm{v}}$ and $\overline{p}$ using [23] $\begin{bmatrix}\widetilde{\bm{v}}\\\ \widetilde{\bm{p}}\end{bmatrix}=\begin{bmatrix}\tau_{M}\,\bm{R}_{M}(\overline{\bm{v}},\overline{p})\\\ \tau_{C}\,R_{C}(\overline{\bm{v}},\overline{p})\end{bmatrix},$ (14) where $\bm{R}_{M}$ and $R_{C}$ represent the momentum and continuity residual expressed as $\begin{cases}\bm{R}_{M}(\overline{\bm{v}},\overline{p})=\bm{R}_{M}=\rho\,\dot{\overline{\bm{v}}}+\rho\,\overline{\bm{v}}\cdot\nabla\overline{\bm{v}}+\nabla\overline{p}-\nabla\cdot\bm{\tau}-\bm{f},\\\ R_{C}(\overline{\bm{v}},\overline{p})=R_{C}=\nabla\cdot\overline{\bm{v}},\end{cases}$ (15) and the _stabilization_ coefficients are expressed as $\begin{split}\tau_{M}&=\left(\frac{4}{\Delta t^{2}}+\bm{u}\cdot\bm{G}\,\bm{u}+C_{I}\,\nu^{2}\,\bm{G}:\bm{G}\right)^{-1/2}\\\ \tau_{C}&=(\tau_{M}\,\bm{g}\cdot\bm{g})^{-1},\,\,G_{i,j}=\sum_{k=1}^{3}\,\dfrac{\partial\xi_{k}}{\partial x_{i}}\,\dfrac{\partial\xi_{k}}{\partial x_{j}},\,\,\,g_{i}=\sum_{k=1}^{3}\,\dfrac{\partial\xi_{k}}{\partial x_{i}},\end{split}$ (16) where $\partial\bm{\xi}/\partial\bm{x}$ is the inverse Jacobian of the element mapping between the parametric and physical domains. To simplify the notation, in what follows the large scale variables $\overline{\bm{v}}$ and $\overline{p}$ will be denoted simply by $\bm{v}$ and $p$ and the spaces $\overline{\mathscr{S}_{k}^{h}}$, $\overline{\mathscr{W}_{k}^{h}}$ and $\overline{\mathscr{P}_{k}^{h}}$ by $\mathscr{S}_{k}^{h}$, $\mathscr{W}_{k}^{h}$ and $\mathscr{P}_{k}^{h}$. A weak solution of the Navier-Stokes equations can now be determined by finding $\bm{v}\in\mathscr{S}_{h}^{k}$ and $p\in\mathscr{P}_{h}^{k}$ such that $\begin{split}&B(\bm{w},q\,;\,\bm{v},p)=B_{G}(\bm{w},q\,;\,\bm{v},p)+\\\ &+\sum_{e=1}^{n_{\text{el}}}\,\int_{\Omega_{e}}\left\\{(\bm{v}\cdot\nabla)\,\bm{w}\cdot(\tau_{M}\,\bm{R}_{M})+\nabla\cdot\bm{w}\,\tau_{C}\,R_{C}\right\\}\,\mathrm{d}\Omega_{e}+\\\ &+\sum_{e=1}^{n_{\text{el}}}\,\int_{\Omega_{e}}\left\\{\bm{w}\cdot\left[-\tau_{M}\,\bm{R}_{M}\cdot\nabla\bm{v}\right]+\left[\bm{R}_{M}\cdot\nabla\bm{w}\right]\cdot\left[\overline{\tau}\,\bm{R}_{M}\cdot\bm{v}\right]\right\\}\,\mathrm{d}\Omega_{e}+\\\ &+\sum_{e=1}^{n_{\text{el}}}\,\int_{\Omega_{e}}\nabla q\cdot\frac{\tau_{M}}{\rho}\,\bm{R}_{M}\,\mathrm{d}\Omega_{e}=0,\end{split}$ (17) for all $\vec{w}\in\vec{\mathscr{W}}_{h}^{k}$ and $q\in\mathscr{P}_{h}^{k}$, where the Galerkin functional $B_{G}$ is expressed as $\begin{split}&B_{G}(\bm{w},q\,;\,\bm{v},p)=\int_{\Omega}\bm{w}\cdot(\rho\,\dot{\bm{v}}+\rho\,\bm{v}\cdot\nabla\bm{v}-\bm{f})\,\,\mathrm{d}\Omega+\\\ &+\int_{\Omega}\left\\{\nabla\bm{w}\,:\,(-p\bm{I}+\bm{\tau})-\nabla q\cdot\bm{v}\right\\}\,\mathrm{d}\Omega+\\\ &+\int_{\Gamma_{h}}\,\left\\{-\bm{w}\cdot\bm{h}+q\,v_{n}\right\\}\,\mathrm{d}\Gamma+\int_{\Gamma_{s}}\,\left\\{-\bm{w}\cdot\bm{t}^{f}+q\,v_{n}\right\\}\,\mathrm{d}\Gamma+\\\ &+\int_{\Gamma_{g}}\,q\,v_{n}\,\mathrm{d}\Gamma.\end{split}$ (18) The discrete variables $\bm{w}^{h}$, $\bm{v}^{h}$, $q^{h}$ and $p^{h}$ are introduced in (17), leading to a non linear system of equations of the form $\begin{cases}\bm{N}_{M}(\dot{\bm{v}}^{h},\bm{v}^{h},p^{h})=0\\\ N_{C}(\dot{\bm{v}}^{h},\bm{v}^{h},p^{h})=0,\end{cases}$ (19) A predictor-multicorrector scheme [24] is used for time integration and, at each time step, the resulting non linear system (19) is solved through successive Newton iterations. Additional details can be found in [23, 25]. #### 2.2.3 Fluid-structure coupling Since this study focuses more on the development of an ensemble solver, we provide a one-way coupled, simplified treatment of the interaction between fluid and structure, leaving a more rigorous treatment to future work. Here we assume a fluid model with rigid walls and a structural model of the lumen governed by the three d.o.f. shell discussed in Section 2.2.1. The lumen deformation does not, in turn, affect the geometry of the fluid region. The elastic forces in the lumen wall are in equilibrium with the shear and pressure exerted by the fluid, or, in other words $\bm{t}^{s}=-\bm{t}^{f}$, where the wall stress $\bm{t}^{f}$ computed by the fluid solver is $\bm{t}^{f}=\bm{\sigma}^{f}\cdot\bm{n},\,\,\bm{\sigma}=2\mu\bm{\varepsilon}-p\bm{I},$ (20) $p$ is the main _nodal_ pressure unknown in the VMS fluid solver, and $\bm{\varepsilon}=\nabla^{s}\bm{u}$ is the symmetric part of the velocity gradient, which is constant on each P1 element in the fluid mesh. An example of shear forces computed by the VMS fluid solver for a coronary model is illustrated in Figure 4. At every node on the wall, the shear forces and normal vectors from adjacent elements are averaged and the nodal pressure added, leading to the three components of the nodal force that are passed to the structural solver. Finally, stress components in the circumferential, radial and axial directions are obtained by transforming the solid stress tensor to a local cylindrical coordinate system (see Figure 4). For an arbitrary Gauss point or lumen shell node $\bm{s}$, we identify the closest location on the vessel centerline. The tangent vector to the centerline at $\bm{s}$ is defined as the local _axial_ direction $z$, the normal at the Gauss point is the local _radial_ direction $r$, and the local _circumferential_ direction $\theta$ is obtained as the cross product between $r$ and $z$. Figure 4: Visualization of interface shear forces exchanged between the fluid and the structural solver. The local axis system used for stress post- processing is also shown in the close up. ### 2.3 Distributed explicit FSI solver on multiple CPUs The solution in time for the dynamics of the undamped non-linear structural system $\bm{M}\,\bm{\ddot{u}}+\bm{C}\,\bm{\dot{u}}+\bm{K}(\bm{u})\,\bm{u}=\bm{f}(t),$ (21) is computed using central differences in time, resulting in an update formula in the $[n\,\Delta t,(n+1)\,\Delta t]$ intervals expressed as $\left(\bm{\widetilde{M}}+\frac{\Delta t}{2}\bm{\widetilde{C}}\right)\,\bm{u}_{n+1}=\Delta t^{2}\,\bm{f}_{n}-\left(\Delta t^{2}\bm{K}-2\,\bm{M}\right)\,\bm{u}_{n}-\left(\bm{M}-\frac{\Delta t}{2}\bm{C}\right)\,\bm{u}_{n-1}$ (22) which is performed independently by an arbitrary number of mesh partitions. To do so efficiently, $\bm{\widetilde{M}},\bm{\widetilde{C}}$ in (22) are lumped mass matrix pre-assembled before the beginning of the time loop, containing the elemental contributions from all finite elements, even those belonging to separate mesh partitions. Given the limited amount of deformation typically observed in cardiovascular applications over a single heart cycle, the assumption of a fixed nodal mass over time is considered realistic. This assumption removes the need of communicating nodal masses during finite element assembly, improving scalability. In some cases, a viscous force $\bm{f}_{v}$ is added to the right-hand-side in order to damp the high- frequency oscillations as $\bm{f}_{v}=-c_{d}\,\dot{\bm{u}}_{n}$, through an appropriate damping coefficient $c_{d}$. Even though the geometry of the vessel lumen is updated at every step by adding $\bm{u}_{n+1}$, the displacement magnitudes over a typical heart cycle remain of the same order as the thickness of the vessel wall, hence in the linear regime. In this context, the stiffness do not significantly change and it is possible to save computational time by avoiding to assemble it at every iteration. In our code, we therefore provide the option to selectively update the stiffness matrix after a prescribed number of iterations. Synchronization of displacements for the nodes shared by multiple partitions is implemented using _Send_ , _Recv_ to the root CPU and broadcast back. The performance of the proposed ensemble solver was first tested on multiple CPUs. We use distributed sparse matrices in the Yale compressed sparse row (CSR) format [26] with dense coefficient entries of size $9\cdot n_{s}$, where $n_{s}$ is the number of material property realizations and a local element matrix of size 3$\times$3 results from selecting a three-d.o.f.s shell finite element. The code for the finite element assembly and sparse matrix-vector multiplication was developed in Cython+MPI+openMP and compared both with a C implementation and with the mkl_cspblas_dcsrgemv routine provided through the Intel MKL library, using single and multiple threads. We verified the satisfactory performance of our implementation under a wide range of mesh sizes, number of cores, and with/without multithreading. Encouraging speedups were obtained on multiple CPUs, as shown in Figure 5 and Figure 6. (a) (b) Figure 5: Performance of matrix-vector product kernel on CPU by our Cython+openMP implementation, GPU implementation and MKL library on multiple threads (a). Optimization of GPU matrix-vector kernel and CPU-GPU communication performance (b). Tests were performed using 1 CPU and 1 GPU on a cylindrical model associated with a sparse matrix having nnz = 160,587 and 1,000 material property realizations. (a) (b) Figure 6: Performance of explicit ensemble solver on multiple CPUs for a mesh with 5,074 elements, 2,565 nodes (a) and for a mesh with 15,136 elements and 7,628 nodes (b). (a) (b) Figure 7: Performance of explicit ensemble solver on multiple CPUs/GPUs for a mesh with 5,074 elements, 2,565 nodes (a) and for a mesh with 15,136 elements and 7,628 nodes (b). ### 2.4 Distributed explicit FSI solver on multiple GPUs We developed an hardware-independent openCL implementation of the solver running on multiple GPUs. We started with a naïve CSR-based parallel sparse matrix-vector product (SpMV), known as _CSR-Scalar_ , where each row of the sparse matrix is assigned to a separate thread [27]. This works well on CPUs, but causes uncoalesced, slow memory accesses on GPUs, since elements in each row occupy consecutive addresses in memory, but consecutive threads access elements on different rows. In addition, long rows lead to an unequal amount of work among the threads and some of them need to wait for others to finish. We then transitioned to a _CSR-Vector_ scheme [28], assigning a wavefront (or so-called _warp_ on NVIDIA architectures) to work on a single row of the matrix. This allows for access to consecutive memory locations in parallel, resulting in fast coalesced loads. However, CSR-Vector can lead to poor GPU occupancy for short rows due to unused execution resources. Improved performance can be achieved using _CSR-Stream_ [29] which statically fixes the number of nonzeros that will be processed by one wavefront and streams all of these values into the local scratchpad memory, effectively utilizing the GPU’s DRAM bandwidth and improving over CSR-Scalar. CSR-Stream also dynamically determines the number of rows on which each wavefront will operate, thus improving over CSR-Vector. While CSR-Stream substantially improves the performance of the spMV product kernel, the CPU-to-GPU data transfer still dominates the time step update, as shown in Figure 5(b). This problem can be mitigated by sending data to the GPU in smaller chunks, thus overlapping data transfer and kernel execution. In addition, data transfer can be minimized by an assembly-free approach. Even though this is a standard practice in explicit finite element codes, it is particularly effective on a GPU for three reasons. First, storage of a sparse global matrix would occupy a significant portion of the GPU memory, posing restrictions on the model size. Second, indexing operations to access entries in the global matrix would cause severe uncoalesced memory access, reducing significantly the degree of parallelism in the GPU. Third, for the most common sparse matrix storage schemes, indexing always include searching, which would be particularly slow on GPU. We compute local matrices directly in the GPU and take their product with a partition-based displacement vector, where only the displacements of shared nodes need to be synchronized through the root CPU. To compute element matrices, we buffer data to the GPU before the beginning of the time integration loop, in order to avoid any CPU to GPU data transfer due to element assembly. Buffered quantities include mesh geometry and material properties, particularly the product between the Young’s modulus and thickness at each Gauss point for all random field realizations. Note how the rest of the local stiffness matrix is constant for linear triangular elements. In addition, the left-hand-side lumped mass matrix resulting from the central difference scheme does not change throughout the time loop. We also leverage a mesh coloring algorithm, allowing working units to process different elements at the same time. During each time step, each computing group in the GPU works on one element, while the working units in the same group work on different realizations and therefore have access to coalesced GPU memory. Communication is only triggered by displacement synchronization. We use pinned host memory to speed up the data transfer between CPU and GPU. All displacements for each realization and the local matrices are stored in private memory since they do not need to be shared with other working units. Final GPU speed ups are illustrated in Figure 7. ### 2.5 A Python code-base The CVFES solver is developed in Python 3 with optimization in Cython [30] and element assembly and matrix product implementation on openCL [31] (though the Python pyOpenCL library [32]). The code leverages the VTK library [33] to read the solid, fluid mesh and boundary conditions. In this context the solver is fully compatible with the input files generated by the SimVascular software platform [34] and can be easily integrated with the SimVascular modeling workflow. Partitioning on multiple CPUs and GPUs is obtained for both solvers using parMETIS [35]. The code used to generate the results discussed in Section 3 is available through a public GitHub repository at https://github.com/desResLab/CVFES. ## 3 Results ### 3.1 Ideal cylindrical benchmark The first benchmark represents an ideal cylindrical lumen subject to aortic flow. The cylinder has a diameter equal to 4 cm and length of 30 cm, while two Matérn random fields for thickness and elastic modulus have been assigned as discussed in Section 2.1. Specifically, a mean $\mu=7.0\times 10^{6}$ Barye and a standard deviation $\sigma=7.0\times 10^{5}$ Barye have been assumed for the elastic modulus, whereas the thickness random field is characterized through a mean $\mu=0.4$ cm and a standard deviation $\sigma=0.04$ cm. Three values of the correlation length were considered, equal to 0.95 cm, 3.7 cm and 7.2 cm, respectively (see discussion in [36]). A uniform pressure of 13 mmHg was added to the pressure computed by the VMS fluid solver. We considered diastole as a natural (unstressed) state and applied the difference between a diastolic pressure of 80 mmHg and the mean brachial pressure in a healthy subject (i.e., with systolic pressure equal to 120 mmHg). This configuration is analyzed both under steady state and pulsatile flow conditions, under fully fixed structural boundary conditions at the two cylinder ends. #### 3.1.1 Steady state analysis The fluid solution is computed with the VMS fluid solver discussed in Section 2.2.2 using a parabolic velocity profile at the inflow corresponding to a $-66.59$ mL/s volumetric flow rate, zero-traction boundary condition at the outlet and a no-slip condition at the lumen wall. As expected, the fluid solution show a perfectly linear relative pressure profile along the cylinder center path and a uniform parabolic velocity profile from inlet to outlet, typical of viscous-dominated Poiseuille flow, as shown in Figure 8. The steady state pressure resulting from the fluid solver is applied as discussed in Section 2.2.3 and 100 thickness and elastic modulus realizations are solved simultaneously. The explicit structural simulation is run for $0.5$ seconds, until a steady state was observed. Three wall mesh densities (coarse, medium and fine) are finally considered, consisting of 5,074, 15,136 and 32,994 triangular shell elements, with explicit time steps set to $\Delta t=4.0\times 10^{-5}$ s, $\Delta t=4.0\times 10^{-5}$ s and $\Delta t=1.0\times 10^{-5}$ s, respectively and no viscous damping. Displacement magnitudes along the longitudinal $z$ axis (cylinder generator) are shown in the top row of Figure 9 for the finer mesh and various correlation lengths. The mean displacement one diameter away from the fully fixed ends (thick black line) is consistent with an homogeneous solution for a thick cylinder with average elastic modulus and thickness (blue dashed line). Displacements associated with single random field realizations are also shown, in Figure 9 (top row), using colors. As expected, the displacement wave length increases with the correlation length, and so does the displacement uncertainty quantified through the 5%-95% confidence interval (gray shaded area). Finally, the second row of Figure 9 shows how increasing the mesh density produces a limited difference in the 5%-95% confidence interval. Circumferential stress was found to be the most significant component, as expected, showing uncertainty increasing with the correlation length, similarly to what observed for the displacement magnitude. The in-plane and out-of-plane shear components ($\sigma_{\theta z}$ and $\sigma_{rz}$), though much smaller than circumferential and axial stresses, are essentially related to the material property non-homogeneity and reduce for an increasing correlation length. These stress components are zero both on average and by solving the model with average material properties. Thus, they can only be captured by explicitly modeling the spatial variability of material properties as in the proposed approach. (a) (b) Figure 8: Steady state pressure (a) and velocity (b) distributions from variational multi-scale fluid solver. Figure 9: Displacement magnitude along cylinder generator (longitudinal $z$ axis) for various correlation lengths (first row). 5%-95% confidence intervals for displacement magnitudes for three increasing mesh densities (i.e., coarse, medium and fine, second row). Figure 10: Ensemble means, 5%-95% confidence intervals and single realizations of longitudinal stress profiles for various mesh densities and random field correlation lengths. #### 3.1.2 Validation under pulsatile flow conditions A parabolic pulsatile inflow (Figure 11(a)) was applied to the same cylindrical geometry discussed in the previous section, while all the other boundary conditions (walls and outlets) were kept the same as in the steady case. The time step is set to $1.0\times 10^{-5}$ and the simulation run for two heart cycles and 100 material property realizations. To avoid the application of impulsive loads which can excite a broad range of frequencies, significantly affecting the undamped dynamics, a ramp was applied to the wall loads resulting from the fluid solver in the last heart cycle. The ramp follows a sine wave and is kept active for the first $0.2$ seconds of the simulation. No damping was applied to the simulation, i.e., $\bm{f}_{v}=0$. As expected, the resulting displacements follow a time profile similar to the inflow, and the 5%-95%confidence intervals increase with the correlation length of the underlying random field. Circumferential and axial stress components are the dominant stress components, which also increase with the correlation length, as observed for the steady state results. (a) (b) (c) Figure 11: Results for pulsatile flow on ideal cylindrical vessel. Inflow time history and spatial location for acquisition of displacement and stress outputs (a). Displacement time history at selected spatial location with thick lines representing ensemble averages, and thin lines used for 5%/95% percentiles (b). Circumferential and axial stress time history at selected spatial location (c). ### 3.2 Benchmark on patient-specific coronary model We also demonstrate the results of the proposed ensemble solver on a patient- specific model of the left anterior descending coronary branch. An ensemble of 100 model solutions were obtained using random field parameters for the elastic modulus equal to $\mu=1.15\times 10^{7}$ Barye and $\sigma=1.7\times 10^{6}$ Barye. For the thickness, we considered a mean equal to $\mu=0.075$ cm and a standard deviation of $\sigma=0.017$ cm. A slip-free boundary condition was applied at the outlets of the fluid domain, whereas fully fixed mechanical restraints (i.e., all three nodal translations) where applied along the edges at both the inlets and outlets. A uniform pressure has been superimposed to the lumen stress obtained from the fluid solver as discussed for the ideal aortic model in Section 3.1. #### 3.2.1 Steady state analysis A constant flow rate equal to $-0.28$ mL/s is applied at the model inlet with a parabolic profile, while a zero-traction boundary condition is applied at the outlets and a no-slip condition at the walls. The explicit time step is set to $2.5\times 10^{-6}$ and the model was run for $0.13$ seconds to reach the steady state. No additional viscous force $\bm{f}_{v}$ was considered. Figure 12 shows the displacements and stress for all three analyzed correlation lengths, averaged through a cross sectional slice of the branch of interest, and plotted for successive slices along the longitudinal $z$ axis. Displacement results confirm the absence of torsion, and a prevalent radial deformation mode, with rigid body motion evident from the axial displacements $d_{z}$, affected by the centerline path geometry. The stress results confirm the importance of the circumferential followed by the axial component. The circumferential stress reduces with the coronary branch radius as intuitively suggested by the Barlow (or Mariotte) formula for thin-walled cylinders. Similar cylindrical displacements, circumferential and axial stress are observed for all three correlation lengths. For the selected coronary branch, the smaller correlation length (0.95 cm) is approximately equal to twice the largest diameter, inducing minimal changes in local deformability. Figure 12: Displacement and stress profiles in left coronary artery LAD branch under steady flow, averaged over all material property realizations and cross- sectional slice. #### 3.2.2 Validation under pulsatile flow conditions The same geometry analyzed in the previous section is subject to a parabolic pulsatile inflow shown in Figure 13. A time step equal to $2.0\times 10^{-6}$ is selected, and the model is run for two complete heart cycles (1.6 seconds) and 100 material property realizations. Similar to the pulsatile cylindrical test case, a time ramp is applied during the first 0.2 s, to avoid impulsive loads produced by a non-zero wall stress at $t=0$, and a pressure of 13 mmHg is superimposed to the fluid wall stress. A viscous force $\bm{f}_{v}$ was applied, using a damping coefficient equal to $c_{d}=0.005$, which was found from various tests to remove the high-frequency oscillations without affecting the system dynamics. Results for the average displacements and stress are shown in Figure 13 at four successive locations over the center path of the LAD branch. Even for this case the circumferential and axial stress are the most relevant components and their time history is similar to the inflow and exhibit a maximum at diastole. Figure 13: Displacement and stress profiles in left coronary artery LAD branch under pulsatile flow, averaged over all material property realizations and cross-sectional slice. ### 3.3 Performance assessment In this section, we compare the performance obtained by running the proposed ensemble solver on three cylindrical models with an increasing number of elements, as shown in Table 1, Table 2 and Table 3. The explicit time step was set to $1.0\times 10^{-5}$ s for all models, and each run consisted of 1,000 time steps. The GPUs used for these tests are four _GeForce RTX 2080 Ti_ with 11GB of RAM equipped with 4352 NVIDIA CUDA Cores and connected to the server main board through PCI express ports. Two types of speedup are investigated, the first relates to solving the same number of material property realizations on an increasing number of processors, either CPUs or GPUs, and provides an idea of the effectiveness of the various optimizations presented in Section 2.3 and Section 2.4. This speedup (first number in parenthesis) is observed to increase with the model size and the number of random field realizations. The computation/communication tradeoff and the small mesh sizes selected for these tests also justify the negligible benefits of using 24 CPU cores instead of 12. The speedup achieved by our GPU implementation is instead very relevant, i.e., approximately three orders of magnitude with respect to a single CPU implementation. The second type of speedup quantifies the efficiency in the proposed ensemble solver, i.e., how much faster one can obtain the solution of multiple realizations by solving them at the same time, with respect to the solution of a single realization on the same hardware, multiplied by the total number of realizations. The computational savings of an ensemble solution are quantified between one and two orders of magnitude, confirming our initial claim. Table 1: Speedup - Model with 5074 elements, 2565 nodes - 1000 time steps with $\Delta t=1.0\times 10^{-5}$ Hardware | 1 Smp (Spd) | 10 Smp (Spd) | 50 Smp (Spd) | 100 Smp (Spd) | 200 Smp (Spd) | 500 Smp (Spd) | ---|---|---|---|---|---|---|--- 1 CPU | 0:00:16 (1.0,1.0) | 0:01:54 (1.43,1.0) | 0:09:46 (1.39,1.0) | 0:18:53 (1.44,1.0) | 0:37:20 (1.46,1.0) | 1:40:13 (1.36,1.0) | 12 CPU | 0:00:10 (1.0,1.50) | 0:00:18 (5.92,6.20) | 0:01:00 (8.92,9.62) | 0:01:58 (9.21,9.59) | 0:03:45 (9.65,9.94) | 0:09:35 (9.46,10.45) | 24 CPU | 0:00:06 (1.0,2.35) | 0:00:17 (3.93,6.45) | 0:01:21 (4.27,7.22) | 0:02:18 (5.03,8.21) | 0:05:25 (4.26,6.87) | 0:11:38 (4.98,8.61) | 1 GPU | 0:00:01 (1.0,9.42) | 0:00:01 (10.20,67.19) | 0:00:02 (29.82,201.96) | 0:00:05 (32.75,214.30) | 0:00:09 (37.84,244.67) | 0:00:20 (42.31,293.77) | 2 GPU | 0:00:01 (1.0,10.56) | 0:00:01 (9.17,67.69) | 0:00:02 (36.27,275.28) | 0:00:03 (44.01,322.76) | 0:00:05 (55.86,404.83) | 0:00:11 (66.74,519.27) | 3 GPU | 0:00:01 (1.0,9.49) | 0:00:01 (9.53,63.25) | 0:00:01 (44.86,305.92) | 0:00:02 (59.66,393.12) | 0:00:04 (75.29,490.21) | 0:00:09 (91.75,641.43) | 4 GPU | 0:00:01 (1.0,9.34) | 0:00:01 (9.70,63.35) | 0:00:01 (46.12,309.52) | 0:00:02 (58.67,380.35) | 0:00:04 (75.09,481.08) | 0:00:09 (90.21,620.49) | (*) All time entries are in a _days:hours:minutes:seconds_ format. The speedup is indicated as $(x,y)$, where $x$ is the speedup with respect to the number of samples and $y$ the speedup by distributing the computation on multiple CPUs or GPUs. Table 2: Speedup - Model with 15136 elements, 7628 nodes - 1000 time steps with $\Delta t=1.0\times 10^{-5}$ Hardware | 1 Smp (Spd) | 10 Smp (Spd) | 50 Smp (Spd) | 100 Smp (Spd) | 200 Smp (Spd) | 500 Smp (Spd) | ---|---|---|---|---|---|---|--- 1 CPU | 0:00:37 (1.0,1.0) | 0:05:24 (1.16,1.0) | 0:25:33 (1.23,1.0) | 0:52:29 (1.20,1.0) | 2:03:22 (1.02,1.0) | 4:48:25 (1.09,1.0) | 12 CPU | 0:00:10 (1.0,3.59) | 0:00:21 (4.83,14.89) | 0:01:50 (4.74,13.82) | 0:04:05 (4.29,12.83) | 0:07:23 (4.74,16.69) | 0:19:19 (4.54,14.92) | 24 CPU | 0:00:11 (1.0,3.32) | 0:00:28 (4.02,11.48) | 0:01:57 (4.83,13.01) | 0:03:58 (4.77,13.21) | 0:07:20 (5.16,16.79) | 0:18:38 (5.08,15.47) | 1 GPU | 0:00:03 (1.0,9.72) | 0:00:03 (10.16,84.84) | 0:00:07 (26.56,209.58) | 0:00:13 (27.89,225.99) | 0:00:24 (31.22,297.26) | 0:00:56 (34.28,305.28) | 1 GPU | 0:00:02 (1.0,15.01) | 0:00:02 (10.11,130.45) | 0:00:04 (29.17,355.60) | 0:00:08 (30.94,387.40) | 0:00:13 (37.20,547.27) | 0:00:30 (40.80,561.29) | 1 GPU | 0:00:02 (1.0,17.02) | 0:00:02 (9.47,138.46) | 0:00:03 (29.37,405.87) | 0:00:06 (33.57,476.46) | 0:00:10 (40.38,673.50) | 0:00:24 (44.55,694.70) | 1 GPU | 0:00:01 (1.0,20.01) | 0:00:02 (8.21,141.15) | 0:00:03 (28.94,470.08) | 0:00:05 (35.15,586.47) | 0:00:08 (42.02,823.93) | 0:00:19 (48.60,891.06) | (*) All time entries are in a _days:hours:minutes:seconds_ format. The speedup is indicated as $(x,y)$, where $x$ is the speedup with respect to the number of samples and $y$ the speedup by distributing the computation on multiple CPUs or GPUs. Table 3: Speedup - Model with 131552 elements, 65896 nodes - 1000 time steps with $\Delta t=1.0\times 10^{-5}$ Hardware | 1 Smp (Spd) | 10 Smp (Spd) | 50 Smp (Spd) | 100 Smp (Spd) | 200 Smp (Spd) | 500 Smp (Spd) | ---|---|---|---|---|---|---|--- 1 CPU | 0:06:47 (1.0,1.0) | 0:56:58 (1.19,1.0) | 4:06:55 (1.38,1.0) | 7:55:06 (1.43,1.0) | - | - | 12 CPU | 0:00:22 (1.0,18.13) | 0:02:03 (1.83,27.74) | 0:10:59 (1.71,22.48) | 0:21:32 (1.74,22.06) | 0:50:50 (1.46,1.0) | - | 24 CPU | 0:00:23 (1.0,17.50) | 0:02:57 (1.32,19.29) | 0:13:34 (1.43,18.20) | 0:26:51 (1.45,17.69) | 0:54:27 (1.43,0.93) | 2:47:47 (1.16,1.0) | 1 GPU | 0:00:29 (1.0,13.93) | 0:00:30 (9.74,113.79) | 0:01:00 (24.17,244.68) | 0:02:00 (24.33,236.96) | 0:03:33 (27.48,14.32) | - | 2 GPU | 0:00:16 (1.0,25.45) | 0:00:16 (9.95,212.32) | 0:00:31 (25.24,466.69) | 0:01:02 (25.54,454.35) | 0:01:53 (28.31,26.95) | 0:04:18 (30.99,38.94) | 3 GPU | 0:00:11 (1.0,37.06) | 0:00:11 (9.68,300.82) | 0:00:21 (25.24,679.48) | 0:00:42 (25.69,665.52) | 0:01:17 (28.57,39.60) | 0:02:54 (31.48,57.60) | 4 GPU | 0:00:08 (1.0,46.39) | 0:00:09 (9.52,370.13) | 0:00:17 (25.33,853.57) | 0:00:33 (26.47,858.07) | 0:00:59 (29.47,51.13) | 0:02:16 (32.24,73.82) | (*) All time entries are in a _days:hours:minutes:seconds_ format. The speedup is indicated as $(x,y)$, where $x$ is the speedup with respect to the number of samples and $y$ the speedup by distributing the computation on multiple CPUs or GPUs. ## 4 Conclusions and future work Our study investigates the efficiency achievable by a ensemble cardiovascular solver on modern GPU architectures. The basic methodological approaches are discussed together with details on our efforts to optimize execution on such architectures, both algorithmically and in terms of host-to-device communication. In particular, computation of local element matrices is performed directly on the GPU, and working units are used to determine displacement increments for independent material property realizations. The main result from our analysis is the observation that explicit-in-time ensemble solvers based on matrix-vector product naturally achieve high scalability, due to their ability to generate large amount of computations and re-usable data patterns provided by independent realizations. This also suggests how ensemble cardiovascular solvers are _ideal_ to efficiently generate campaigns of high-fidelity model solutions that form a pre-requisite for uncertainty quantification studies, the state-of-the-art paradigm for the analysis of cardiovascular systems, due to their ability to quantify the effects of multiple sources of uncertainty on the simulation outputs. We have also validated our solver for the steady state and pulsatile solution of an idealized aortic flow and a patient-specific model of the left coronary artery, under vessel wall mechanical property uncertainty, modeled through a Gaussian random field approximation. Future work will be devoted to further improve the computational efficiency of the proposed approach and to transition from a segregated to a fully coupled Arbitrary Lagrangian-Eulerian fluid-structure interaction paradigm. In addition, the proposed approach only considered uncertainties due to vessel wall thickness and elastic modulus. We plan to support uncertainty also in the boundary conditions and model geometry. ## Acknowledgements This work was supported by a National Science Foundation award #1942662 _CAREER: Bayesian Inference Networks for Model Ensembles_ (PI Daniele E. Schiavazzi). This research used computational resources provided through the Center for Research Computing at the University of Notre Dame. We also acknowledge support from the open source SimVascular project at www.simvascular.org. ## Conflict of interest The authors declare that they have no conflict of interest. ## References * [1] F. Lindgren, H. Rue, and J. Lindström. An explicit link between gaussian fields and Gaussian Markov random fields: the stochastic partial differential equation approach. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 73(4):423–498, 2011. * [2] C.A. Figueroa, I.E. Vignon-Clementel, K.E. Jansen, T.J.R. Hughes, and C.A. Taylor. A coupled momentum method for modeling blood flow in three-dimensional deformable arteries. Computer methods in applied mechanics and engineering, 195(41-43):5685–5706, 2006. * [3] K. Bäumler, V. Vedula, A.M. Sailer, J. Seo, P. Chiu, G. Mistelbauer, F.P. Chan, M.P. Fischbein, A.L. Marsden, and D. Fleischmann. Fluid–structure interaction simulations of patient-specific aortic dissection. Biomechanics and Modeling in Mechanobiology, pages 1–22, 2020. * [4] H. Askes, D.C.D. Nguyen, and A. Tyas. Increasing the critical time step: micro-inertia, inertia penalties and mass scaling. Computational Mechanics, 47(6):657–667, 2011. * [5] S. Morlacchi, C. Chiastra, D. Gastaldi, G. Pennati, G. Dubini, and F. Migliavacca. Sequential structural and fluid dynamic numerical simulations of a stented bifurcated coronary artery. Journal of biomechanical engineering, 133(12), 2011. * [6] C. Chiastra, G. Dubini, and F. Migliavacca. Modeling the stent deployment in coronary arteries and coronary bifurcations. In Biomechanics of Coronary Atherosclerotic Plaque, pages 579–597. Elsevier, 2020. * [7] C. Chiastra, S. Morlacchi, D. Gallo, U. Morbiducci, R. Cárdenes, I. Larrabide, and F. Migliavacca. Computational fluid dynamic simulations of image-based stented coronary bifurcation models. Journal of The Royal Society Interface, 10(84):20130193, 2013. * [8] A. Bartezzaghi, M. Cremonesi, N. Parolini, and U. Perego. An explicit dynamics gpu structural solver for thin shell finite elements. Computers & Structures, 154:29–40, 2015. * [9] V. Strbac, D.M. Pierce, J. Vander Sloten, and N. Famaey. GPGPU-based explicit finite element computations for applications in biomechanics: the performance of material models, element technologies, and hardware generations. Computer Methods in Biomechanics and Biomedical Engineering, 20(16):1643–1657, 2017. * [10] N. Jiang and W. Layton. An algorithm for fast calculation of flow ensembles. International Journal of Uncertainty Quantification, 4(4):273–301, 2014. * [11] N. Jiang. A higher order ensemble simulation algorithm for fluid flows. Journal of Scientific Computing, 64(1):264–288, 2015. * [12] N. Jiang and W. Layton. Numerical analysis of two ensemble eddy viscosity numerical regularizations of fluid motion. Numerical Methods for Partial Differential Equations, 31(3):630–651, 2015. * [13] A. Takhirov, M. Neda, and J. Waters. Time relaxation algorithm for flow ensembles. Numerical Methods for Partial Differential Equations, 32(3):757–777, 2016. * [14] N. Jiang. A second-order ensemble method based on a blended backward differentiation formula timestepping scheme for time-dependent Navier–Stokes equations. Numerical Methods for Partial Differential Equations, 33(1):34–61, 2017. * [15] M. Mohebujjaman and L.G. Rebholz. An efficient algorithm for computation of MHD flow ensembles. Computational Methods in Applied Mathematics, 17(1):121–137, 2017\. * [16] J.A. Fiordilino. Ensemble time-stepping algorithms for the heat equation with uncertain conductivity. Numerical Methods for Partial Differential Equations, 34(6):1901–1916, 2018. * [17] Y. Luo and Z. Wang. An ensemble algorithm for numerical solutions to deterministic and random parabolic PDEs. SIAM Journal on Numerical Analysis, 56(2):859–876, 2018. * [18] M. Gunzburger, N. Jiang, and Z. Wang. An efficient algorithm for simulating ensembles of parameterized flow problems. IMA Journal of Numerical Analysis, 39(3):1180–1205, 2019. * [19] M.L. Stein. Interpolation of spatial data: some theory for kriging. Springer Science & Business Media, 2012. * [20] P. Whittle. On stationary processes in the plane. Biometrika, pages 434–449, 1954. * [21] P. Whittle. Stochastic-processes in several dimensions. Bulletin of the International Statistical Institute, 40(2):974–994, 1963. * [22] D. Bolin and F. Lindgren. A comparison between Markov approximations and other methods for large spatial data sets. Computational Statistics & Data Analysis, 61:7–21, 2013. * [23] Y. Bazilevs, V.M. Calo, J.A. Cottrell, T.J.R. Hughes, A. Reali, and G. Scovazzi. Variational multiscale residual-based turbulence modeling for large eddy simulation of incompressible flows. Computer Methods in Applied Mechanics and Engineering, 197(1):173–201, 2007. * [24] K.E. Jansen, C.H. Whiting, and G.M. Hulbert. A generalized-$\alpha$ method for integrating the filtered Navier–Stokes equations with a stabilized finite element method. Computer methods in applied mechanics and engineering, 190(3-4):305–319, 2000. * [25] J. Seo, D.E. Schiavazzi, and A.L. Marsden. Performance of preconditioned iterative linear solvers for cardiovascular simulations in rigid and deformable vessels. Computational Mechanics, pages 1–23, 2019. * [26] A. Buluç, J.T. Fineman, M. Frigo, J.R. Gilbert, and C.E. Leiserson. Parallel sparse matrix-vector and matrix-transpose-vector multiplication using compressed sparse blocks. In Proceedings of the twenty-first annual symposium on Parallelism in algorithms and architectures, pages 233–244, 2009. * [27] M. Garland. Sparse matrix computations on manycore GPUs. In Proceedings of the 45th annual Design Automation Conference, pages 2–6, 2008. * [28] N. Bell and M. Garland. Implementing sparse matrix-vector multiplication on throughput-oriented processors. In Proceedings of the conference on high performance computing networking, storage and analysis, pages 1–11, 2009. * [29] J.L. Greathouse and M. Daga. Efficient sparse matrix-vector multiplication on GPUs using the CSR storage format. In SC’14: Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis, pages 769–780. IEEE, 2014. * [30] S. Behnel, R. Bradshaw, C. Citro, L. Dalcin, D.S. Seljebotn, and K. Smith. Cython: The best of both worlds. Computing in Science Engineering, 13(2):31–39, march-april 2011\. * [31] J.E. Stone, D. Gohara, and G. Shi. Opencl: A parallel programming standard for heterogeneous computing systems. Computing in science & engineering, 12(3):66–73, 2010. * [32] A. Klöckner, N. Pinto, Y. Lee, B. Catanzaro, P. Ivanov, and A. Fasih. Pycuda and pyopencl: A scripting-based approach to gpu run-time code generation. Parallel Computing, 38(3):157–174, 2012. * [33] Will Schroeder, Ken Martin, and Bill Lorensen. The Visualization Toolkit (4th ed.). Kitware, 2006. * [34] H. Lan, A. Updegrove, N.M. Wilson, G.D. Maher, S.C. Shadden, and A.L. Marsden. A re-engineered software interface and workflow for the open-source simvascular cardiovascular modeling package. Journal of biomechanical engineering, 140(2), 2018. * [35] George Karypis and Vipin Kumar. MeTis: Unstructured Graph Partitioning and Sparse Matrix Ordering System, Version 4.0. http://www.cs.umn.edu/~metis, 2009. * [36] J.S. Tran, D.E. Schiavazzi, A.M. Kahn, and A.L. Marsden. Uncertainty quantification of simulated biomechanical stimuli in coronary artery bypass grafts. Computer Methods in Applied Mechanics and Engineering, 345:402–428, 2019.
Present address: ]Department of Physics, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA Present address: ]Westlake University, 310024 Hangzhou, China Present address: ]Department of Quantum Matter Physics, University of Geneva, 24 Quai Ernest-Ansermet, 1211 Geneva 4, Switzerland # Quasi-isotropic orbital magnetoresistance in lightly doped SrTiO3 Clément Collignon [ JEIP, USR 3573 CNRS, Collège de France, PSL Research University, 11, place Marcelin Berthelot, 75231 Paris Cedex 05, France Laboratoire de Physique et d’Étude des Matériaux (ESPCI Paris - CNRS - Sorbonne Université), PSL Research University, 75005 Paris, France Yudai Awashima Department of Engineering Science, University of Electro- Communications, Chofu, Tokyo 182-8585, Japan Ravi Laboratoire de Physique et d’Étude des Matériaux (ESPCI Paris - CNRS - Sorbonne Université), PSL Research University, 75005 Paris, France Xiao Lin [ Laboratoire de Physique et d’Étude des Matériaux (ESPCI Paris - CNRS - Sorbonne Université), PSL Research University, 75005 Paris, France Carl Willem Rischau [ Laboratoire de Physique et d’Étude des Matériaux (ESPCI Paris - CNRS - Sorbonne Université), PSL Research University, 75005 Paris, France Anissa Acheche JEIP, USR 3573 CNRS, Collège de France, PSL Research University, 11, place Marcelin Berthelot, 75231 Paris Cedex 05, France Baptiste Vignolle Laboratoire National des Champs Magnétiques Intenses (LNCMI-EMFL), CNRS ,UGA, UPS, INSA, Grenoble/Toulouse, France Institut de Chimie de la Matière Condensée, Bordeaux, France Cyril Proust Laboratoire National des Champs Magnétiques Intenses (LNCMI-EMFL), CNRS ,UGA, UPS, INSA, Grenoble/Toulouse, France Yuki Fuseya Department of Engineering Science, University of Electro- Communications, Chofu, Tokyo 182-8585, Japan Institute for Advanced Science, University of Electro-Communications, Chofu, Tokyo 182-8585, Japan Kamran Behnia Laboratoire de Physique et d’Étude des Matériaux (ESPCI Paris - CNRS - Sorbonne Université), PSL Research University, 75005 Paris, France Benoit Fauqué<EMAIL_ADDRESS>JEIP, USR 3573 CNRS, Collège de France, PSL Research University, 11, place Marcelin Berthelot, 75231 Paris Cedex 05, France ###### Abstract A magnetic field parallel to an electrical current does not produce a Lorentz force on the charge carriers. Therefore, orbital longitudinal magnetoresistance is unexpected. Here we report on the observation of a large and non saturating magnetoresistance in lightly doped SrTiO3-x independent of the relative orientation of current and magnetic field. We show that this quasi-isotropic magnetoresistance can be explained if the carrier mobility along all orientations smoothly decreases with magnetic field. This anomalous regime is restricted to low concentrations when the dipolar correlation length is longer than the distance between carriers. We identify cyclotron motion of electrons in a potential landscape tailored by polar domains as the cradle of quasi-isotropic orbital magnetoresistance. The result emerges as a challenge to theory and may be a generic feature of lightly-doped quantum paralectric materials. ††preprint: APS/123-QED Magnetoresistance (MR), the change in electrical resistivity under the application of a magnetic field is an old topic in condensed matter physics Pippard (1989). It can be simply understood as a consequence of the Lorentz force exerted on mobile electrons by the magnetic field. This orbital magnetoresistance (which neglects the spin of electrons) is largest when the magnetic field is perpendicular to the electrical current. The transverse magnetoresistance (labelled TMR) is expected to increase quadratically with magnetic field at low fields and then saturate at high fields. The boundary between the two regimes is set by $\mu_{H}B\approx 1$ (where $\mu_{H}$ is the Hall mobility). When the field and the current are parallel, we are in presence of longitudinal magnetoresistance (labelled LMR), expected to be negligibly small due to the cancellation of the Lorentz force. However, this simple picture is known to fail in numerous cases. Non- saturating linear TMR has been observed in electronic systems ranging from potassium Pippard (1989), to doped silicon Delmo _et al._ (2009), 2D electron gas system Khouri _et al._ (2016), 3D doped semi-conductors and Dirac materials Xu _et al._ (1997); Hu _et al._ (2005); Kozlova _et al._ (2012); Schneider _et al._ (2014); Fauqué _et al._ (2013); Novak _et al._ (2015); Narayanan _et al._ (2015); Xiong _et al._ (2016), density wave materials Feng _et al._ (2019), or correlated materials Hayes _et al._ (2016). LMR, one order of magnitude smaller (than and with an opposite sign to) its transverse counterpart has been observed in silver chalcogenides Hu _et al._ (2005) and topological materials Wiedmann _et al._ (2016); dos Reis _et al._ (2016). The exact conditions for the emergence of a sizeable LMR is the subject of ongoing debate Pal and Maslov (2010); Burkov (2015); Goswami _et al._ (2015). Figure 1: Electrical transport properties of a lightly doped SrTiO3-x at low temperature : a) Fermi surface of the lower band of SrTiO3 according to Allen _et al._ (2013) b) Temperature dependence of the resistivity ($\rho$) of sample S7. Insert: $\rho$ vs $T^{2}$. c) Transverse ($j\perp B$) and longitudinal ($j//B$) magnetoresistance ($\frac{\Delta\rho}{\rho_{0}}$=$\frac{\rho(B)-\rho(B=0)}{\rho(B=0)}$) as function of the magnetic field up to 54 T, at $T=1.5$ K, for S6 ($n_{H}$(S${}_{6})=3.2\times 10^{17}$ cm-3). d) Normalised angular magneto- resistance of S1 ($n_{H}($S${}_{1})=6.5\times 10^{16}$ cm-3) in polar plot at $B=10$ T for two temperatures : $T=2$ K (in blue) and $T=10$ K (in red). $\theta=0^{\circ}$ and $90^{\circ}$ correspond respectively to ${\bf{j}}\perp{\bf{B}}$ and ${\bf{j}}//{\bf{B}}$) In this paper, we report on the case of lightly doped SrTiO3. Undoped strontium titanate is an incipient ferroelectric, dubbed quantum paraelectric Müller and Burkard (1979), which can be turned into a metal by non-covalent substitution or by removing oxygen Spinelli _et al._ (2010). This dilute metal Collignon _et al._ (2019) has attracted recent attention Zhou and Bernardi (2019); Kumar _et al._ (2020) due to the persistence of $T$-square resistivity in absence of Umklapp and interband scattering among electrons Lin _et al._ (2015) and the unexpected survival of metallicity at high temperatures Collignon _et al._ (2020). We will see below that its magnetoresistance is also remarkably non-trivial. In contrast with any other documented material, it shows a large and quasi-linear TMR, accompanied by a positive LMR of comparable amplitude. Intriguingly, the amplitude of magnetoresistance depends only on the amplitude of the magnetic field, independent of the mutual orientation of the current and the magnetic field. We will show that this unusual quasi-isotropic magnetoresistance is restricted to a range of doping where the inter-electron distance exceeds the typical size of a polar domain. The observation implies that this phenomenon is driven by the interplay between cyclotron orbits and the potential landscape shaped by polar domains and suggests that it may be generic to lightly doped quantum paraelectrics. Fig.1 presents our main result. When the carrier density in SrTiO3-δ is $n_{H}=3\times 10^{17}$ cm-3, there is a single Fermi pocket at the center of the Brillouin zone shown Fig.1a). This is what is expected by band calculations van der Marel _et al._ (2011) and found by quantum oscillations Uwe _et al._ (1985); Allen _et al._ (2013); Lin _et al._ (2014). Nevertheless, not only this dilute metal displays a $T^{2}$behavior (see Fig.1b)), but it also responds to magnetic field in a striking manner. Upon the application of a magnetic field of 54 T, there is a forty (twenty)-fold enhancement of resistance for the transverse (longitudinal) configuration (see Fig.1c)). In both cases, the evolution with field is quasi-linear and there is no sign of saturation even if the high field regime ($\mu_{H}B>>1$) is clearly attained. . Figure 2: Temperature dependence of the transverse and longitudinal magneto- resistance : a)-c) Field dependence of the resistivity for ${\bf{j}}\perp{\bf{B}}$ ($\rho_{\perp}$), $\frac{\rho_{xy}}{B}$ and the resistivity for ${\bf{j}}//{\bf{B}}$ ($\rho_{//}$) from $T=2$ to 14 K for S7 ($n_{H}($S${}_{7})=3.3\times 10^{17}$ cm-3) d)-f) same as a)-c) from $T=16$ to 60 K. g) Field dependence of the Hall mobility $\mu_{H}=\frac{\rho_{xy}}{B\rho_{\perp}}$ for $T=2$, 18 and 60 K. h) Longitudinal magnetoconductivity ($\frac{1}{\rho_{//}}$) compared with $\sigma_{//}=ne\mu_{H}(B)$ with the deduced $\mu_{H}(B)$ shown on g). A polar plot of the normalised angular magnetoresistance (AMR) at a fixed magnetic field for another sample (S1) with a slightly lower carrier density ($n_{H}($S${}_{1})=6.5\times 10^{16}$ cm-3) is shown on Fig.1d). The magnetic field rotates from the transverse ($\theta=0^{\circ}$) to the longitudinal ($\theta=90^{\circ}$) configuration at two different temperatures. While at $T=10$ K, the longitudinal magnetoresistance shrinks towards zero, at $T=2$ K the magnetoresistance is quasi-isotropic and the relative direction of the magnetic field and the current injection barely affects its amplitude. When a magnetic field is aligned along the $z$-axis, in presence of a single- component Fermi surface, the three components of the conductivity tensor have remarkably simple expressions: $\sigma_{zz}=\sigma_{//}=ne\mu_{H}$ (1) $\sigma_{xx}=\sigma_{\perp}=\frac{ne\mu_{H}}{1+\mu^{2}_{H}B^{2}}$ (2) $\sigma_{xy}=\mu_{H}B\frac{ne\mu_{H}}{1+\mu^{2}_{H}B^{2}}$ (3) Now, if $\mu_{H}$ remains constant as a function of magnetic field, one does not expect the longitudinal magnetoresistance, since $\rho_{//}=\sigma_{//}^{-1}$ would not depend on magnetic field. One would not even expect a transverse magnetoresistance, because the same is true for $\rho_{\perp}=\frac{\sigma_{\perp}}{\sigma_{\perp}^{2}+\sigma_{xy}^{2}}=\frac{1}{ne\mu_{H}}$. These equations hold in presence of a quadratic dispersion when the effective mass $m^{*}$ and the Hall mobility, $\mu_{H}=e\tau/m^{*}$ are well defined. The Fermi pocket associated with the lowest band in dilute metallic strontium titanate is not an ellipsoid. This can lead to a finite TMR and LMR Pippard (1989); Pal and Maslov (2010). However, as discussed in the supplement SM , the results computed using the specific geometry of the Fermi surface are well below the experimentally observed magnitudes at low temperature. As we will see below, to explain our result, one needs to assume a field-dependent $\mu_{H}$. Figure 3: Doping dependence of the transverse and the longitudinal magnetoresistance at $T=2$ K : a) and b) TMR for nH ranging from $6.5\times 10^{16}$ to $3.6\times 10^{19}$ cm-3. c) and d) same as a) and b) for the LMR. e) Field dependence of Hall mobility ($\mu_{H}$) deduced from the TMR and the Hall effect measurements in four low doped samples. f) Comparison of the longitudinal magnetoconductance ($\frac{1}{\rho_{//}}$) with $\sigma_{//}=ne\mu_{H}(B)$ with the deduced $\mu_{H}(B)$ shown on e). Fig.2 shows the evolution of the quasi-isotropic magnetoresistance with temperature. The amplitude of the TMR decreases with warming (see Fig.2a) and d)). The same is true for the LMR (see Fig.2c) and f)). On the other hand, the Hall coefficient is barely temperature-dependent (see Fig.2b) and e)). Upon warming, the longitudinal magnetoresistance decreases faster than its transverse counterpart and above 14 K it almost vanishes (Fig.2c) and d)). Above this temperature, a small TMR persists with an amplitude comparable with what the semi-classical theory expects (see supplement SM ). Fig.3 shows the evolution with doping. Increasing carrier concentration diminishes both TMR and LMR (see Fig.3a)-d)). As in the case of thermal evolution, the LMR decreases faster than the TMR. At low doping, the two configurations yield a similar amplitude. With increasing carrier concentration, the LMR becomes smaller than the TMR (see Fig.3 b) and d)). Therefore, the unusual regime of the magnetoresistance detected by the present study emerges only at low temperature (when resistivity is dominated by its elastic component) and at low carrier concentration. Remarkably, even in this unusual context, the three components of the conductivity tensor keep the links expected by Eq.(1-3). This is demonstrated in the final panels of Fig.2 and Fig.3. The Hall mobility at a given magnetic field can be extracted using $\mu_{H}=\frac{1}{B}\frac{\sigma_{xy}}{\sigma_{\perp}}=\frac{1}{B}\frac{\rho_{xy}}{\rho_{\perp}}$ (see Fig.2g)). The deduced $\mu_{H}(B)$ can then be compared with the field dependence of the longitudinal conductance $\sigma_{//}$ (see Fig.2h)). As seen in the figure, there is a satisfactory agreement. This is the case of all samples at low doping levels, as shown in Fig.3e) and f). Thus, assuming that mobility smoothly evolves with magnetic field, would explain both the quasi-linear non saturating TMR and the large finite LMR, which emerge at low doping. Fig.4a) shows the doping dependence of the LMR to TMR ratio ($\frac{\Delta\rho_{//}}{\Delta\rho_{\perp}}$ at $B=10$ T and $T=2.5$ K). Clearly, the finite LMR kicks in below a cut-off concentration and grows steadily with decreasing carrier density. The unusual magnetoresistance of lightly doped SrTiO3-δ is therefore restricted to carrier densities below a threshold of $3\times 10^{18}$ cm-3. As we will see below, a clue to the origin of this phenomenon is provided by this boundary. Figure 4: Size of polar domains vs. inter-electron distance in lightly doped SrTiO3 : a) Doping dependence of the ratio of the LMR and TMR at B=10T for T=2.5K (in blue closed circles). Insert : sketch of the cubic unit cell of SrTiO3 lattices in presence of an oxygen vacancy. b) Doping dependence of $\ell_{ee}$=$(n_{H})^{\frac{-1}{3}}$ (the inter-electron distance), of the screening length scale from charged impurities $r_{TF}=\sqrt{\frac{\pi a_{B}}{4k_{F}}}$ compares with the polar domain diameter, 2Rc=5.4nm at low temperature. c) Sketch of the SrTiO3 lattice in presence of two oxygen vacancies separated by a distance $\ell_{ee}$ and of the polar domains (in gray light) which form around each oxygen vacancies with a radius Rc. Below a critical doping where $\ell_{ee}$ becomes shorter than 2Rc non zero LMR appears. How can the mobility decrease with magnetic field along both orientations? Why does this decrease happens in a restricted window of doping? We will see below that a length scale specific to quantum paraelectrics plays a key role in finding answers to both of these questions. A field-dependent mobility has been previously invoked in other contexts Song _et al._ (2015); Fauqué _et al._ (2018). The time between collision events can become shorter in presence of magnetic field, because disorder is scanned differently at zero and finite magnetic fields. Compared to zero-field counterparts, charge carriers following a cyclotron orbit are more vulnerable to shallow scattering centers. Such a picture has been invoked to explain the linear TMR in 3D high mobility dilute semiconductors Song _et al._ (2015) and the sub-quadratic TMR in semi-metals Fauqué _et al._ (2018). There are three already identified relevant length scales to the problem. These are, i) the Thomas-Fermi screening radius $r_{TF}=\sqrt{\frac{\pi a_{B}}{4k_{F}}}$; ii) the magnetic length, $\ell_{B}=\sqrt{\frac{\hbar}{eB}}$; and iii) the Fermi wavelength, $\lambda_{F}=2\pi k_{F}^{-1}$. When disorder is smooth and $r_{TF}$ is longer than the cyclotron radius ($r_{c}=\ell_{B}^{2}k_{F}$), the magnetic field, by quenching the kinetic energy of electrons in the plane of cyclotron motion, would guide them along the minimum of the electrostatic potential fluctuations Song _et al._ (2015). This would lead to a decrease in mobility in the plane perpendicular to the magnetic field. The doping dependence of the Thomas-Fermi screening radius is shown in Fig.4b). Thanks to a Bohr radius as long as 600 nm in strontium titanate, $r_{TF}$ is remarkably long Behnia (2015) and easily exceeds the cyclotron radius in a field of the order of Tesla. Therefore, shallow extended disorder, screened at zero-field will become visible as the cyclotron radius shrinks. One can invoke this picture to explain the quasi-linear TMR. However, the finite LMR and the low-field TMR remain both unexplained, because only the plane perpendicular of the orientation to the magnetic field is concerned. In a polar crystal, defects, by distorting the lattice, generate electric dipoles. The typical length for correlation between such dipoles is set by $R_{c}$=$\frac{v_{s}}{\omega_{O}}$ (where $v_{s}$ and $\omega_{O}$ are the sound velocity and the frequency of the soft optical mode, respectively). In highly polarizable crystals, $\omega_{O}$ is small and $R_{c}$ can become remarkably long Vugmeister and Glinchuk (1990); Samara (2003). In the specific case of strontium titanate $v_{s}\simeq 7500$ m.s-1 Rehwald (1970), $\omega_{0}(300K)\simeq 11$ meV and $\omega_{0}(2$K$)\simeq 1.8$ meV Yamada and Shirane (1969), therefore, $R_{c}$ varies from 0.5 nm at 300 K to 2.7 nm at 2 K. As a consequence, defects can cooperate with other defects over long distances to generate mesoscopic dipoles. In the case of a co-valent substitution, such as Sr1-xCaxTiO3, a Ca atom can break the local inversion symmetry. It can cooperate with other Ca sites within a range of $R_{c}$ to choose the same orientation for dipole alignment. When the Ca density exceeds a threshold, these domains percolate and generate a ferroelectric ground states. Remarkably, this critical density ($x>0.002$ Bednorz and Müller (1984)) corresponds to replacement of 1 out of 500 Sr atoms by Ca, that is when their average distance falls below $\frac{R_{c}}{a}$ (here, $a=0.39$ nm is the lattice parameter and $\frac{R_{c}}{a}\approx 500^{-1/3}$). In the case of an oxygen vacancy, the donor, in addition to a local potential well, brings also a local dipole capable of cooperation with neighboring donors over long distances. Recent studies Rischau _et al._ (2017); Wang _et al._ (2019) have confirmed the survival of dipolar physics in presence of dilute metallicity and the generation of ripples by electric dipoles inside the shallow Fermi sea. Specifically, it was found that in Sr1-xCaxTiO3-δ, the ferroelectric-like alignment of dipoles is destroyed when there is more than one mobile electron per $7.9\pm 0.6$ Ca atoms, the Fermi sea is dense enough to impede the percolation between polar domains. This threshold corresponds to an inter- electron distance approximately twice ($7.9^{-1/3}\approx 2$) the inter-dipole distance. An oxygen vacancy (in addition to being a donor and an ionized point defect) generates an extended distortion of the size of $R_{c}$. This leads us to identify the origin of the doping window for the unusually isotropic magnetoresistance. If the inter-electron distance, ($\ell_{ee}$), which increases with decreasing carrier concentration, becomes significantly longer than $R_{c}$, then mobile carriers cannot adequately screen polar domains. Our data indicates that this is where the large quasi-isotropic magnetoresistance emerges. Fig.4b) shows the evolution of $\ell_{ee}=(n_{H})^{-1/3}$ with doping. One can see that the threshold of $3\times 10^{18}$ cm-3 corresponds to $\ell_{ee}=6.7$ nm. In other words, when the inter-electron distance becomes shorter than $2R_{c}$, the unusual magnetoresistance disappears, presumably because the Fermi sea is dense enough to impede the inhomogeneity generated by polar domains. This unusual MR is the largest at low temperature when resistivity is dominated by elastic scattering events. It vanishes with warming, when the inelastic $T^{2}$-term dominates over residual resistivity. This cross-over typically occurs around 15 K (see Fig.1b)). A possible solution to the mystery of the isotropic reduction of the mobility with magnetic field is offered by this length scale, which does not depend on the orientation of magnetic field. In presence of randomly oriented mesoscopic dipoles, the charge current does not align locally parallel to its macroscopic orientation. Instead, it will meander along a trajectory set by dipoles’ electric field. The disorder affecting the whirling electrons will reduce mobility along different orientations. Remarkably, the inhomogeneity brought by these polar domains do not impede the existence of a percolated Fermi sea and the observation of quantum oscillations in this range of carrier concentration. A solid explanation of this apparent paradox remains a a task for future theoretical investigations. In summary, we found a large and quasi-isotropic magnetoresistance in lightly doped strontium titanate. We found that the longitudinal and the transverse magnetoresistance can be both explained in a picture where mobility changes with magnetic field and this arises as long the inter-electron distance is twice larger than the typical size of a polar domains. Other doped quantum paraelectrics, such as PbTe, KTaO3 appear as potential candidates for displaying the same phenomenon. We thank M. Feigelman and B. Skinner for useful discussions. We acknowledge the support of the LNCMI-CNRS, member of the European Magnetic Field Laboratory (EMFL). This work was supported by JEIP-Collège de France, by the Agence Nationale de la Recherche (ANR-18-CE92-0020-01; ANR-19-CE30-0014-04) and by a grant attributed by the Ile de France regional council. ## References * Pippard (1989) A. B. Pippard, _Magnetoresistance in Metals_ (Cambridge University Press, 1989). * Delmo _et al._ (2009) M. P. Delmo, S. Yamamoto, S. Kasai, T. Ono, and K. Kobayashi, Nature 457, 1112 (2009). * Khouri _et al._ (2016) T. Khouri, U. Zeitler, C. Reichl, W. Wegscheider, N. E. Hussey, S. Wiedmann, and J. C. Maan, Phys. Rev. Lett. 117, 256601 (2016). * Xu _et al._ (1997) R. Xu, A. Husmann, T. F. Rosenbaum, M.-L. Saboungi, J. E. Enderby, and P. B. Littlewood, Nature 390, 57 (1997). * Hu _et al._ (2005) J. Hu, T. F. Rosenbaum, and J. B. Betts, Phys. Rev. Lett. 95, 186603 (2005). * Kozlova _et al._ (2012) N. V. Kozlova, N. Mori, O. Makarovsky, L. Eaves, Q. D. Zhuang, A. Krier, and A. Patanè, Nature Communications 3, 1097 EP (2012), article. * Schneider _et al._ (2014) J. M. Schneider, M. L. Peres, S. Wiedmann, U. Zeitler, V. A. Chitta, E. Abramof, P. H. O. Rappl, S. de Castro, D. A. W. Soares, U. A. Mengui, and N. F. Oliveira, Applied Physics Letters 105, 162108 (2014). * Fauqué _et al._ (2013) B. Fauqué, N. P. Butch, P. Syers, J. Paglione, S. Wiedmann, A. Collaudin, B. Grena, U. Zeitler, and K. Behnia, Phys. Rev. B 87, 035133 (2013). * Novak _et al._ (2015) M. Novak, S. Sasaki, K. Segawa, and Y. Ando, Phys. Rev. B 91, 041203 (2015). * Narayanan _et al._ (2015) A. Narayanan, M. D. Watson, S. F. Blake, N. Bruyant, L. Drigo, Y. L. Chen, D. Prabhakaran, B. Yan, C. Felser, T. Kong, P. C. Canfield, and A. I. Coldea, Phys. Rev. Lett. 114, 117201 (2015). * Xiong _et al._ (2016) J. Xiong, S. Kushwaha, J. Krizan, T. Liang, R. J. Cava, and N. P. Ong, EPL (Europhysics Letters) 114, 27002 (2016). * Feng _et al._ (2019) Y. Feng, Y. Wang, D. M. Silevitch, J.-Q. Yan, R. Kobayashi, M. Hedo, T. Nakama, Y. Onuki, A. V. Suslov, B. Mihaila, P. B. Littlewood, and T. F. Rosenbaum, Proceedings of the National Academy of Sciences 116, 11201 (2019). * Hayes _et al._ (2016) I. M. Hayes, R. D. McDonald, N. P. Breznay, T. Helm, P. J. W. Moll, M. Wartenbe, A. Shekhter, and J. G. Analytis, Nature Physics 12, 916 (2016). * Wiedmann _et al._ (2016) S. Wiedmann, A. Jost, B. Fauqué, J. van Dijk, M. J. Meijer, T. Khouri, S. Pezzini, S. Grauer, S. Schreyeck, C. Brüne, H. Buhmann, L. W. Molenkamp, and N. E. Hussey, Phys. Rev. B 94, 081302 (2016). * dos Reis _et al._ (2016) R. D. dos Reis, M. O. Ajeesh, N. Kumar, F. Arnold, C. Shekhar, M. Naumann, M. Schmidt, M. Nicklas, and E. Hassinger, New Journal of Physics 18, 085006 (2016). * Pal and Maslov (2010) H. K. Pal and D. L. Maslov, Phys. Rev. B 81, 214438 (2010). * Burkov (2015) A. A. Burkov, Phys. Rev. B 91, 245157 (2015). * Goswami _et al._ (2015) P. Goswami, J. H. Pixley, and S. Das Sarma, Phys. Rev. B 92, 075205 (2015). * Allen _et al._ (2013) S. J. Allen, B. Jalan, S. Lee, D. G. Ouellette, G. Khalsa, J. Jaroszynski, S. Stemmer, and A. H. MacDonald, Phys. Rev. B 88, 045114 (2013). * Müller and Burkard (1979) K. A. Müller and H. Burkard, Phys. Rev. B 19, 3593 (1979). * Spinelli _et al._ (2010) A. Spinelli, M. A. Torija, C. Liu, C. Jan, and C. Leighton, Phys. Rev. B 81, 155110 (2010). * Collignon _et al._ (2019) C. Collignon, X. Lin, C. W. Rischau, B. Fauqué, and K. Behnia, Annual Review of Condensed Matter Physics 10, 25 (2019). * Zhou and Bernardi (2019) J.-J. Zhou and M. Bernardi, Phys. Rev. Research 1, 033138 (2019). * Kumar _et al._ (2020) A. Kumar, V. I. Yudson, and D. L. Maslov, “Quasiparticle and non-quasiparticle transport in doped quantum paraelectrics,” (2020), arXiv:2007.14947 [cond-mat.str-el] . * Lin _et al._ (2015) X. Lin, B. Fauqué, and K. Behnia, Science 349, 945 (2015). * Collignon _et al._ (2020) C. Collignon, P. Bourges, B. Fauqué, and K. Behnia, Phys. Rev. X 10, 031025 (2020). * van der Marel _et al._ (2011) D. van der Marel, J. L. M. van Mechelen, and I. I. Mazin, Phys. Rev. B 84, 205111 (2011). * Uwe _et al._ (1985) H. Uwe, R. Yoshizaki, T. Sakudo, A. Izumi, and T. Uzumaki, Japanese Journal of Applied Physics 24, 335 (1985). * Lin _et al._ (2014) X. Lin, G. Bridoux, A. Gourgout, G. Seyfarth, S. Krämer, M. Nardone, B. Fauqué, and K. Behnia, Phys. Rev. Lett. 112, 207002 (2014). * (30) See Supplemental Material for more details on the sample properties, additional temperature and angular measurements and the theoritical model used to compute the semi-classical MR. * Song _et al._ (2015) J. C. W. Song, G. Refael, and P. A. Lee, Physical Review B - Condensed Matter and Materials Physics 92, 1 (2015), arXiv:1507.04730 . * Fauqué _et al._ (2018) B. Fauqué, X. Yang, W. Tabis, M. Shen, Z. Zhu, C. Proust, Y. Fuseya, and K. Behnia, Phys. Rev. Materials 2, 114201 (2018). * Behnia (2015) K. Behnia, Journal of Physics: Condensed Matter 27, 375501 (2015). * Vugmeister and Glinchuk (1990) B. E. Vugmeister and M. D. Glinchuk, Rev. Mod. Phys. 62, 993 (1990). * Samara (2003) G. A. Samara, Journal of Physics: Condensed Matter 15, R367 (2003). * Rehwald (1970) W. Rehwald, Solid State Communications 8, 607 (1970). * Yamada and Shirane (1969) Y. Yamada and G. Shirane, Journal of the Physical Society of Japan 26, 396 (1969). * Bednorz and Müller (1984) J. G. Bednorz and K. A. Müller, Phys. Rev. Lett. 52, 2289 (1984). * Rischau _et al._ (2017) C. W. Rischau, X. Lin, C. P. Grams, D. Finck, S. Harms, J. Engelmayer, T. Lorenz, Y. Gallais, B. Fauqué, J. Hemberger, and K. Behnia, Nature Physics 13, 643 EP (2017). * Wang _et al._ (2019) J. Wang, L. Yang, C. W. Rischau, Z. Xu, Z. Ren, T. Lorenz, J. Hemberger, X. Lin, and K. Behnia, npj Quantum Materials 4, 61 (2019).
# Convergence Results of Forward-Backward Algorithms for Sum of Monotone Operators in Banach Spaces Yekini Shehu111Department of Mathematics, Zhejiang Normal University, Jinhua, 321004, People’s Republic of China; Institute of Science and Technology (IST), Am Campus 1, 3400, Klosterneuburg, Vienna, Austria; e-mail: <EMAIL_ADDRESS> (January 22, 2021) ###### Abstract It is well known that many problems in image recovery, signal processing, and machine learning can be modeled as finding zeros of the sum of maximal monotone and Lipschitz continuous monotone operators. Many papers have studied forward-backward splitting methods for finding zeros of the sum of two monotone operators in Hilbert spaces. Most of the proposed splitting methods in the literature have been proposed for the sum of maximal monotone and inverse-strongly monotone operators in Hilbert spaces. In this paper, we consider splitting methods for finding zeros of the sum of maximal monotone operators and Lipschitz continuous monotone operators in Banach spaces. We obtain weak and strong convergence results for the zeros of the sum of maximal monotone and Lipschitz continuous monotone operators in Banach spaces. Many already studied problems in the literature can be considered as special cases of this paper. Keywords: inclusion problem; 2-uniformly convex Banach space; forward-backward algorithm; weak convergence; strong convergence. 2010 MSC classification: 47H05, 47J20, 47J25, 65K15, 90C25. ## 1 Introduction Let $E$ be a real Banach space with norm $\|.\|,$ we denote by $E^{*}$ the dual of $E$ and $\langle f,x\rangle$ the value of $f\in E^{*}$ at $x\in E.$ Let $B:E\rightarrow 2^{E^{*}}$ be a maximal monotone operator and $A:E\rightarrow E^{*}$ be a Lipschitz continuous monotone operator. We consider the following inclusion problem: find $x\in E$ such that $\displaystyle 0\in(A+B)x.$ (1) Throughout this paper, we denote the solution set of the inclusion problem (1) by $(A+B)^{-1}(0)$. The inclusion problem (1) contains, as special cases, convexly constrained linear inverse problem, split feasibility problem, convexly constrained minimization problem, fixed point problems, variational inequalities, Nash equilibrium problem in noncooperative games, and many more. See, for instance, [11, 15, 28, 33, 35, 36] and the references therein. A popular method for solving problem (1) in real Hilbert spaces, is the well- known forward–backward splitting method introduced by Passty [35] and Lions and Mercier [28]. The method is formulated as $\displaystyle x_{n+1}=(I+\lambda_{n}B)^{-1}(I-\lambda_{n}A)x_{n},\leavevmode\nobreak\ \leavevmode\nobreak\ \lambda_{n}>0,$ (2) under the condition that $Dom(B)\subset Dom(A)$. It was shown, see for example [11], that weak convergence of (2) requires quite restrictive assumptions on $A$ and $B$, such that the inverse of $A$ is strongly monotone or $B$ is Lipschitz continuous and monotone and the operator $A+B$ is strongly monotone on $Dom(B)$. Tseng in [48], weakened these assumptions and included an extra step per each step of (2) (called Tseng’s splitting algorithm) and obtained weak convergence result in real Hilbert spaces. Quite recently, Gibali and Thong [18] have obtained strong convergence result by modifying Tseng’s splitting algorithm in real Hilbert spaces. In this paper, we extend Tseng’s result [48] to a Banach space. We first prove the weak convergence of the sequence generated by our proposed method, assuming that the duality mapping is weakly sequentially continuous. This weak convergence is a generalization of Theorem 3.4 given in [48]. We next prove the strong convergence result for problem (1) under some mild assumptions and this extends Theorems 1 and 2 in [18] to Banach spaces. Finally, we apply our convergence results to the composite convex minimization problem in Banach spaces. ## 2 Preliminaries In this section, we define some concepts and state few basic results that we will use in our subsequent analysis. Let $S_{E}$ be the unit sphere of $E$, and $B_{E}$ the closed unit ball of $E$. Let $\rho_{E}:[0,\infty)\rightarrow[0,\infty)$ be the modulus of smoothness of $E$ defined by $\rho_{E}(t):=\sup\Big{\\{}\frac{1}{2}(\|x+y\|+\|x-y\|)-1:\,x\in S_{E},\,\|y\|\leq t\Big{\\}}.$ A Banach space $E$ is said to be $2$-uniformly smooth, if there exists a fixed constant $c>0$ such that $\rho_{E}(t)\leq ct^{2}$. The space $E$ is said to be smooth if $\displaystyle\lim_{t\rightarrow 0}\frac{\|x+ty\|-\|x\|}{t}$ (3) exists for all $x,y\in S_{E}$. The space $E$ is also said to be uniformly smooth if (3) converges uniformly in $x,y\in S_{E}$. It is well known that if $E$ is $2$-uniformly smooth, then $E$ is uniformly smooth. It is said to be strictly convex if $\|(x+y)/2\|<1$ whenever $x,y\in S_{E}$ and $x\neq y$. It is said to be uniformly convex if $\delta_{E}(\epsilon)>0$ for all $\epsilon\in(0,2]$, where $\delta_{E}$ is the modulus of convexity of $E$ defined by $\displaystyle\delta_{E}(\epsilon):=\inf\Big{\\{}1-\Big{|}\Big{|}\frac{x+y}{2}\Big{|}\Big{|}\mid x,y\in B_{E},\|x-y\|\geq\epsilon\Big{\\}}$ (4) for all $\epsilon\in[0,2]$. The space $E$ is said to be 2-uniformly convex if there exists $c>0$ such that $\delta_{E}(\epsilon)\geq c\epsilon^{2}$ for all $\epsilon\in[0,2]$. It is obvious that every 2-uniformly convex Banach space is uniformly convex. It is known that all Hilbert spaces are uniformly smooth and 2-uniformly convex. It is also known that all the Lebesgue spaces $L_{p}$ are uniformly smooth and 2-uniformly convex whenever $1<p\leq 2$ (see [7]). The normalized duality mapping of $E$ into $E^{*}$ is defined by $Jx:=\\{x^{*}\in E^{*}\mid\langle x^{*},x\rangle=\|x^{*}\|^{2}=\|x\|^{2}\\}$ for all $x\in E$. The normalized duality mapping $J$ has the following properties (see, e.g., [47]): * • if $E$ is reflexive and strictly convex with the strictly convex dual space $E^{*}$, then $J$ is single-valued, one-to-one and onto mapping. In this case, we can define the single-valued mapping $J^{-1}:E^{*}\rightarrow E$ and we have $J^{-1}=J_{*}$, where $J_{*}$ is the normalized duality mapping on $E^{*}$; * • if $E$ is uniformly smooth, then $J$ is uniformly norm-to-norm continuous on each bounded subset of $E.$ Let us recall from [1, 13] some examples for the normalized duality mapping $J$ in the uniformly convex and uniformly smooth Banach spaces $\ell_{p}$ and $L_{p},1<p<\infty$. * • For $\ell_{p}:Jx=\|x\|_{\ell_{p}}^{2-p}y\in\ell_{q}$, where $x=(x_{j})_{j\geq 1}$ and $y=(x_{j}|x_{j}|^{p-2})_{j\geq 1}$, $\frac{1}{p}+\frac{1}{q}=1$. * • For $L_{p}:Jx=\|x\|_{L_{p}}^{2-p}|x|^{p-2}x\in L_{q}$, $\frac{1}{p}+\frac{1}{q}=1$. Now, we recall some fundamental and useful results. ###### Lemma 2.1. The space $E$ is 2-uniformly convex if and only if there exists $\mu_{E}\geq 1$ such that $\displaystyle\frac{\|x+y\|^{2}+\|x-y\|^{2}}{2}\geq\|x\|^{2}+\|\mu^{-1}_{E}y\|^{2}$ (5) for all $x,y\in E$. The minimum value of the set of all $\mu_{E}\geq 1$ satisfying (5) for all $x,y\in E$ is denoted by $\mu$ and is called the 2-uniform convexity constant of $E$; see [5]. It is obvious that $\mu=1$ whenever $E$ is a Hilbert space. ###### Lemma 2.2 ([4]). Let $\displaystyle\frac{1}{p}+\frac{1}{q}=1,\leavevmode\nobreak\ \leavevmode\nobreak\ p,q>1$. The space $E$ is $q-$uniformly smooth if and only if its dual $E^{*}$ is $p-$uniformly convex. ###### Lemma 2.3 ([51]). Let $E$ be a real Banach space. The following are equivalent: * (1) $E$ is 2-uniformly smooth * (2) There exists a constant $\kappa>0$ such that $\forall\ x,y\in E$, $\|x+y\|^{2}\leq\|x\|^{2}+2\langle y,J(x)\rangle+2\kappa^{2}\|y\|^{2},$ where $\kappa$ is the 2-uniform smoothness constant. In Hilbert spaces, $\kappa=\frac{1}{\sqrt{2}}$. ###### Definition 2.4. Let $X\subseteq E$ be a nonempty subset. Then a mapping $A:X\to E^{*}$ is called * (a) strongly monotone with modulus $\gamma>0$ on $X$ if $\langle Ax-Ay,x-y\rangle\geq\gamma\|x-y\|^{2},\forall x,y\in X.$ In this case, we say that $A$ is $\gamma$-strongly monotone; * (b) monotone on $X$ if $\langle Ax-Ay,x-y\rangle\geq 0,\forall x,y\in X;$ * (c) Lipschitz continuous on $X$ if there exists a constant $L>0$ such that $\|Ax-Ay\|\leq L\|x-y\|$ for all $x,y\in X$. We give some examples of monotone operator in Banach spaces as given in [2]. ###### Example 2.5. Let $G\subset\mathbb{R}^{n}$ be a bounded measurable domain. Define the operator $A:L^{p}(G)\rightarrow L^{q}(G),\leavevmode\nobreak\ \leavevmode\nobreak\ \frac{1}{p}+\frac{1}{q}=1,\leavevmode\nobreak\ \leavevmode\nobreak\ p>1$, by the formula $Ay(x):=\varphi(x,|y(x)|^{p-1})|y(x)|^{p-2}y(x),\leavevmode\nobreak\ \leavevmode\nobreak\ x\in G,$ where the function $\varphi(x,s)$ is measurable as a function of $x$ for every $s\in[0,\infty)$ and continuous for almost all $x\in G$ as a function on $s,|\varphi(x,s)|\leq M$ for all $s\in[0,\infty)$ and for almost all $x\in G$. Observe that the operator $A$ really maps $L^{p}(G)$ to $L^{q}(G)$ because of the inequality $|Ay|\leq M|y|^{p-1}$. Then it can be shown that $A$ is a monotone map on $L^{p}(G)$. Let us consider another example from quantum mechanics. ###### Example 2.6. Define the operator $Au:=-a^{2}\triangle u+(g(x)+b)u(x)+u(x)\int_{\mathbb{R}^{3}}\frac{u^{2}(y)}{|x-y|}dy,$ where $\triangle:=\sum_{i=1}^{3}\frac{\partial^{2}}{\partial x_{i}^{2}}$ is the Laplacian in $\mathbb{R}^{3}$, $a$ and $b$ are constants, $g(x)=g_{0}(x)+g_{1}(x),\leavevmode\nobreak\ \leavevmode\nobreak\ g_{0}(x)\in L^{\infty}(\mathbb{R}^{3}),g_{1}(x)\in L^{2}(\mathbb{R}^{3})$. Let $A:=L+B$, where the operator $L$ is the linear part of $A$ (it is the Schrödinger operator) and $B$ is defined by the last term. It is known that $B$ is a monotone operator on $L^{2}(\mathbb{R}^{3})$ (see page 23 of [2]) and this implies that $A:L^{2}(\mathbb{R}^{3})\rightarrow L^{2}(\mathbb{R}^{3})$ is also a monotone operator. ###### Example 2.7. This example gives one of the perhaps most famous example of monotone operators, viz. the $p$-Laplacian $-{\rm div}(|\nabla u|^{p-2}\nabla u):W^{1}_{0}(L_{p}(\Omega))\rightarrow\Big{(}W^{1}_{0}(L_{p}(\Omega))\Big{)}^{*}$, where $u:\Omega\rightarrow\mathbb{R}$ is a real function defined on a domain $\Omega\subset\mathbb{R}^{n}$. The $p$-Laplacian operator is a monotone operator for $1<p<\infty$ (in fact, it is strongly monotone for $p\geq 2$, and strictly monotone for $1<p<2$). The $p$-Laplacian operator is an extremely important model in many topical applications and certainly played an important role in the development of the theory of monotone operators. ###### Definition 2.8. A multi-valued operator $B:E\rightarrow 2^{E^{*}}$ with graph $G(T)=\\{(x,x^{*}):x^{*}\in Tx\\}$ is said to be monotone if for any $x,y\in D(T),x^{*}\in Tx$ and $y^{*}\in Ty$ $\langle x-y,x^{*}-y^{*}\rangle\geq 0.$ A monotone operator $B$ is said to be maximal if $B=S$ whenever $S:E\rightarrow 2^{E^{*}}$ is monotone and $G(B)\subset G(S)$. Let $E$ be a reflexive, strictly convex and smooth Banach space and let $B:E\rightarrow 2^{E^{*}}$ be a maximal monotone operator. Then for each $r>0$ and $x\in E$, there corresponds a unique element $x_{r}\in E$ such that $Jx\in Jx_{r}+rBx_{r}.$ We define this unique element $x_{r}$, the resolvent of $B$, denoted by $J^{B}_{r}x$. In other words, $J_{r}^{B}=(J+rB)^{-1}J$ for all $r>0$. It is easy to show that $B^{-1}0=F(J^{B}_{r})$ for all $r>0$, where $F(J^{B}_{r})$ denotes the set of all fixed points of $J^{B}_{r}$. We can also define, for each $r>0$, the Yosida approximation of $B$ by $A_{r}=\frac{J-JJ^{B}_{r}}{r}$. For more details, see, for instance [6]. Suppose $E$ is a smooth Banach space. We introduce the functional studied in [1, 25, 38]: $\phi:E\times E\rightarrow\mathbb{R}$ defined by: $\displaystyle\phi(x,y):=\|x\|^{2}-2\langle x,Jy\rangle+\|y\|^{2}.$ (6) Clearly, $\phi(x,y)\geq(\|x\|-\|y\|)^{2}\geq 0.$ The following lemma gives some identities of functional $\phi$ defined in (6). ###### Lemma 2.9. (see [3] and [1]) Let $E$ be a real uniformly convex, smooth Banach space. Then, the following identities hold: (i) $\displaystyle\phi(x,y)=\phi(x,z)+\phi(z,y)+2\langle x-z,Jz-Jy\rangle,\ \forall x,y,z\in E.$ (ii) $\displaystyle\phi(x,y)+\phi(y,x)=2\langle x-y,Jx-Jy\rangle,\ \forall x,y\in E.$ Let $C\subseteq E$ be a nonempty, closed and convex subset of a real, uniformly convex Banach space $E$. Let us introduce the functional $V(x,y):E\times E^{*}\rightarrow\mathbb{R}$ by the formula: $V(x,y):=\|x\|^{2}_{E}-2\langle x,y\rangle+\|y\|^{2}_{E^{*}}.$ (7) Then, it is easy to see that $V(y,x)=\phi(y,J^{-1}x),\leavevmode\nobreak\ \leavevmode\nobreak\ \forall x\in E^{*},y\in E.$ In the next lemma, we describe the property of the operator $V(.,.)$ defined in (7). ###### Lemma 2.10. ([1]) $V(x,x^{*})+2\langle J^{-1}x^{*}-x,y^{*}\rangle\leq V(x,x^{*}+y^{*}),\leavevmode\nobreak\ \leavevmode\nobreak\ \forall x\in E,\leavevmode\nobreak\ \leavevmode\nobreak\ x^{*},y^{*}\in E^{*}.$ The lemma that follows is stated and proven in [3, Lem. 2.2]. ###### Lemma 2.11. Suppose that $E$ is 2-uniformly convex Banach space. Then, there exists $\mu\geq 1$ such that $\frac{1}{\mu}\|x-y\|^{2}\leq\phi(x,y)\leavevmode\nobreak\ \leavevmode\nobreak\ \forall x,y\in E.$ The following lemma was given in [21]. ###### Lemma 2.12. Let $S$ be a nonempty, closed convex subset of a uniformly convex, smooth Banach space $E$. Let $\\{x_{n}\\}$ be a sequence in $E$. Suppose that, for all $u\in S$, $\phi(u,x_{n+1})\leq\phi(u,x_{n}),\leavevmode\nobreak\ \leavevmode\nobreak\ \forall n\geq 1.$ Then $\\{\Pi_{S}(x_{n})\\}$ is a Cauchy sequence. The following property of $\phi(.,.)$ was given in [1, Thm. 7.5] (see also [16, 17]). ###### Lemma 2.13. Let $E$ be a uniformly smooth Banach space which is also uniformly convex. If $\|x\|\leq c,\|y\|\leq c$, then $2L_{1}^{-1}c^{2}\delta_{E}\Big{(}\frac{\|x-y\|}{4c}\Big{)}\leq\phi(y,x)\leq 4L_{1}^{-1}c^{2}\rho_{E}\Big{(}\frac{4\|x-y\|}{c}\Big{)},$ where $L_{1}(1<L_{1}<3.18)$ is the Figiel’s constant. We next recall some existing results from the literature to facilitate our proof of strong convergence. The first is taken from [31]. ###### Lemma 2.14. Let $\\{a_{n}\\}$ be sequence of real numbers such that there exists a subsequence $\\{n_{i}\\}$ of $\\{n\\}$ such that $a_{n_{i}}<a_{{n_{i}}+1}$, for all $i\in\mathbb{N}$. Then there exists a nondecreasing sequence $\\{m_{k}\\}\subset\mathbb{N}$ such that $m_{k}\rightarrow\infty$ and the following properties are satisfied by all (sufficiently large) numbers $k\in\mathbb{N}$ $a_{m_{k}}\leq a_{{m_{k}}+1}\leavevmode\nobreak\ \leavevmode\nobreak\ {\rm and}\leavevmode\nobreak\ \leavevmode\nobreak\ a_{k}\leq a_{{m_{k}}+1}.$ In fact, $m_{k}=\max\\{j\leq k:a_{j}<a_{j+1}\\}$. ###### Lemma 2.15. ([52]) Let $\\{a_{n}\\}$ be a sequence of nonnegative real numbers satisfying the following relation: $a_{n+1}\leq(1-\alpha_{n})a_{n}+\alpha_{n}\sigma_{n}+\gamma_{n},\leavevmode\nobreak\ \leavevmode\nobreak\ n\geq 1,$ where * (a) $\\{\alpha_{n}\\}\subset[0,1],$ $\sum_{n=1}^{\infty}\alpha_{n}=\infty;$ * (b) $\limsup\sigma_{n}\leq 0$; * (c) $\gamma_{n}\geq 0\ (n\geq 1),$ $\sum_{n=1}^{\infty}\gamma_{n}<\infty.$ Then, $a_{n}\rightarrow 0$ as $n\rightarrow\infty$. The following lemma is needed in our proof to show that the weak limit point is a solution to the inclusion problem (1). ###### Lemma 2.16. ([6]) Let $B:E\to 2^{E^{*}}$ be a maximal monotone mapping and $A:E\to E^{*}$ be a Lipschitz continuous and monotone mapping. Then the mapping $A+B$ is a maximal monotone mapping. The following result gives an equivalence of fixed point problem and problem (1). ###### Lemma 2.17. Let $B:E\to 2^{E^{*}}$ be a maximal monotone mapping and $A:E\to E^{*}$ be a mapping. Define a mapping $T_{\lambda}x:=J_{\lambda}^{B}oJ^{-1}(J-\lambda A)(x),\leavevmode\nobreak\ \leavevmode\nobreak\ x\in E,\lambda>0.$ Then $F(T_{\lambda})=(A+B)^{-1}(0),$ where $F(T_{\lambda})$ denotes the set of all fixed points of $T_{\lambda}$. ###### Proof. Let $x\in F(T_{\lambda})$. Then $\displaystyle x\in F(T_{\lambda})$ $\displaystyle\Leftrightarrow$ $\displaystyle x=T_{\lambda}x=J_{\lambda}^{B}oJ^{-1}(J-\lambda A)(x)$ $\displaystyle\Leftrightarrow$ $\displaystyle x=(J+\lambda B)^{-1}JoJ^{-1}(Jx-\lambda Ax)$ $\displaystyle\Leftrightarrow$ $\displaystyle Jx-\lambda Ax\in Jx+\lambda Bx$ $\displaystyle\Leftrightarrow$ $\displaystyle 0\in\lambda(Ax+Bx)$ $\displaystyle\Leftrightarrow$ $\displaystyle 0\in Ax+Bx$ $\displaystyle\Leftrightarrow$ $\displaystyle x\in(A+B)^{-1}(0).$ ∎ We shall adopt the following notation in this paper: . $x_{n}\rightarrow x$ means that $x_{n}\rightarrow x$ strongly. . $x_{n}\rightharpoonup x$ means that $x_{n}\rightarrow x$ weakly. ## 3 Approximation Method In this section, we propose our method and state certain conditions under which we obtain the desired convergence for our proposed methods. First, we give the conditions governing the cost function and the sequence of parameters below. ###### Assumption 3.1. * (a) Let $E$ be a real 2-uniformly convex Banach space which is also uniformly smooth. * (b) Let $B:E\to 2^{E^{*}}$ be a maximal monotone operator; $A:E\to E^{*}$ a monotone and $L$-Lipschitz continuous. * (c) The solution set $(A+B)^{-1}(0)$ of the inclusion problem (1) is nonempty. Throughout this paper, we assume that the duality mapping $J$ and the resolvent $J_{\lambda_{n}}^{B}:=(J+\lambda_{n}B)^{-1}J$ of maximal monotone operator $B$ are easy to compute. ###### Assumption 3.2. Suppose the sequence $\\{\lambda_{n}\\}_{n=1}^{\infty}$ of step-sizes satisfies the following condition: $0<a\leq\lambda_{n}\leq b<\displaystyle\frac{1}{\sqrt{2\mu}\kappa L}$ where * $\mu$ is the 2-uniform convexity constant of $E$; * $\kappa$ is the 2-uniform smoothness constant of $E^{*}$; * $L$ is the Lipschitz constant of $A$. Assumption 3.2 is satisfied, e.g., for $\lambda_{n}=a+\frac{n}{n+1}\Big{(}\frac{1}{\sqrt{2\mu}\kappa L}-a\Big{)}$ for all $n\geq 1$. We now give our proposed method below. ###### Algorithm 3.3. Step 0: Let Assumptions 3.1 and 3.2 hold. Let $x_{1}\in E$ be a given starting point. Set $n:=1$. Step 1: Compute $y_{n}:=J_{\lambda_{n}}^{B}oJ^{-1}(Jx_{n}-\lambda_{n}Ax_{n})$. If $x_{n}-y_{n}=0$: STOP. Step 2: Compute $\displaystyle x_{n+1}=J^{-1}[Jy_{n}-\lambda_{n}(Ay_{n}-Ax_{n})].$ (8) Step 3: Set $n\leftarrow n+1$, and go to Step 1. We observe that in real Hilbert spaces, the duality mapping $J$ becomes the identity mapping and our Algorithm 3.3 reduces to the algorithm proposed by Tseng in [48]. Note that both sequences $\\{y_{n}\\}$ and $\\{x_{n}\\}$ are in $E$. Furthermore, by Lemma 2.17, we have that if $x_{n}=y_{n}$, then $x_{n}$ is a solution of problem (1). To the best of our knowledge, the proposed Algorithm 3.3 is the only known algorithm which can solve monotone inclusion problem (1) without the inverse- strongly monotonicity of $A$. We consider some various cases of Algorithm 3.3. * • When $A=0$ in Algorithm 3.3, then Algorithm 3.3 reduces to the methods proposed in [6, 20, 24, 26, 27, 28, 32, 35, 38, 39, 43]. In this case, the assumption that $E$ is 2-uniformly convex Banach space and uniformly smooth is not needed. In fact, the convergence can be obtained in reflexive Banach spaces in this case. However, we do not know if the convergence of Algorithm 3.3 can be obtained in a more general reflexive Banach space for problem (1). * • When $B=N_{C}$, the normal cone for closed and convex subset $C$ of $E$ ($N_{C}(x):=\\{x^{*}\in E^{*}:\langle y-x,x^{*}\rangle\leq 0,\forall y\in C\\}$), then the inclusion problem (1) reduces to a variational inequality problem (i.e., find $x\in C:\langle Ax,y-x\rangle\geq 0,\leavevmode\nobreak\ \forall y\in C$). It is well known that $N_{C}=\partial\delta_{C}$, where $\delta_{C}$ is the indicator function of $C$ at $x$, defined by $\delta_{C}(x)=0$ if $x\in C$ and $\delta_{C}(x)=+\infty$ if $x\notin C$ and $\partial(.)$ is the subdifferential, defined by $\partial f(x):=\\{x^{*}\in E^{*}:f(y)\geq f(x)+\langle x^{*},y-x\rangle,\leavevmode\nobreak\ \leavevmode\nobreak\ \forall y\in E\\}$ for a proper, lower semicontinuous convex functional $f$ on $E$. Using the theorem of Rockafellar in [40, 41], $N_{C}=\partial\delta_{C}$ is maximal monotone. Hence, $Jz\in J(J_{\lambda_{n}}^{B})+\lambda_{n}\partial\delta_{C}(J_{\lambda_{n}}^{B}),\leavevmode\nobreak\ \leavevmode\nobreak\ \forall z\in E.$ This implies that $\displaystyle 0\in\partial\delta_{C}(J_{\lambda_{n}}^{B})+\frac{1}{\lambda_{n}}J(J_{\lambda_{n}}^{B})-\frac{1}{\lambda_{n}}Jz=\partial\Big{(}\delta_{C}+\frac{1}{2\lambda_{n}}\|.\|^{2}-\frac{1}{\lambda_{n}}Jz\Big{)}J_{\lambda_{n}}^{B}.$ Therefore, $J_{\lambda_{n}}^{B}(z)={\rm argmin}_{y\in E}\Big{\\{}\delta_{C}(y)+\frac{1}{2\lambda_{n}}\|y\|^{2}-\frac{1}{\lambda_{n}}\langle y,Jz\rangle\Big{\\}}$ and $y_{n}$ in Algorithm 3.3 reduces to $y_{n}={\rm argmin}_{y\in E}\Big{\\{}\delta_{C}(y)+\frac{1}{2\lambda_{n}}\|y\|^{2}-\frac{1}{\lambda_{n}}\langle y,Jx_{n}-\lambda_{n}Ax_{n}\rangle\Big{\\}}.$ However, in implementing our proposed Algorithm 3.3, we assume that the resolvent $(J+\lambda_{n}B)^{-1}J$ is easy to compute and the duality mapping $J$ is easily computable as well. On the other hand, one has to obtain the Lipschitz constant, $L$, of the monotone mapping $A$ (or an estimate of it). In a case when the Lipschitz constant cannot be accurately estimated or overestimated, this might result in too small step-sizes $\lambda_{n}$. This is a drawback of our proposed Algorithm 3.3. One way to overcome this obstacle is to introduce linesearch in our Algorithm 3.3. This case will be considered in Algorithm 3.8. ### 3.1 Convergence Analysis In this Section, we give the convergence analysis of the proposed Algorithm 3.3. First, we establish the boundedness of the sequence of iterates generated by Algorithm 3.3. ###### Lemma 3.4. Let Assumptions 3.1 and 3.2 hold. Assume that $x^{*}\in(A+B)^{-1}(0)$ and let the sequence $\\{x_{n}\\}_{n=1}^{\infty}$ be generated by Algorithm 3.3. Then $\\{x_{n}\\}$ is bounded. ###### Proof. By the Lyaponuv functional $\phi$, we have $\displaystyle\phi(x^{*},x_{n+1})=$ $\displaystyle\phi(x^{*},J^{-1}(Jy_{n}-\lambda_{n}(Ay_{n}-Ax_{n})))$ $\displaystyle=$ $\displaystyle\|x^{*}\|^{2}-2\langle x^{*},JJ^{-1}(Jy_{n}-\lambda_{n}(Ay_{n}-Ax_{n}))\rangle$ $\displaystyle+\|J^{-1}(Jy_{n}-\lambda_{n}(Ay_{n}-Ax_{n}))\|^{2}$ $\displaystyle=$ $\displaystyle\|x^{*}\|^{2}-2\langle x^{*},Jy_{n}-\lambda_{n}(Ay_{n}-Ax_{n})\rangle+\|(Jy_{n}-\lambda_{n}(Ay_{n}-Ax_{n}))\|^{2}$ $\displaystyle=$ $\displaystyle\|x^{*}\|^{2}-2\langle x^{*},Jy_{n}\rangle+2\lambda_{n}\langle x^{*},Ay_{n}-Ax_{n}\rangle$ $\displaystyle+\|Jy_{n}-\lambda_{n}(Ay_{n}-Ax_{n})\|^{2}.$ (9) Using Lemma 2.2, we get that $E^{*}$ is 2-uniformly smooth and so by Lemma 2.3, we get $\displaystyle\|Jy_{n}-\lambda_{n}(Ay_{n}-Ax_{n})\|^{2}\leq$ $\displaystyle\|Jy_{n}\|^{2}-2\lambda_{n}\langle Ay_{n}-Ax_{n},y_{n}\rangle$ $\displaystyle+2\kappa^{2}\|\lambda_{n}(Ay_{n}-Ax_{n})\|^{2}.$ (10) Substituting (10) into (9), we get $\displaystyle\phi(x^{*},x_{n+1})\leq$ $\displaystyle\|Jy_{n}\|^{2}-2\lambda_{n}\langle Ay_{n}-Ax_{n},y_{n}\rangle+2\kappa^{2}\|\lambda_{n}(Ay_{n}-Ax_{n})\|^{2}$ $\displaystyle+\|x^{*}\|^{2}-2\langle x^{*},Jy_{n}\rangle+2\lambda_{n}\langle x^{*},Ay_{n}-Ax_{n}\rangle$ $\displaystyle=$ $\displaystyle\|x^{*}\|^{2}-2\langle x^{*},Jy_{n}\rangle+\|y_{n}\|^{2}-2\lambda_{n}\langle Ay_{n}-Ax_{n},y_{n}-x^{*}\rangle$ $\displaystyle+2\kappa^{2}\|\lambda_{n}(Ay_{n}-Ax_{n})\|^{2}$ $\displaystyle=$ $\displaystyle\phi(x^{*},y_{n})-2\lambda_{n}\langle Ay_{n}-Ax_{n},y_{n}-x^{*}\rangle+2\kappa^{2}\|\lambda_{n}(Ay_{n}-Ax_{n})\|^{2}.$ (11) Using Lemma 2.9 (i), we get $\displaystyle\phi(x^{*},y_{n})=$ $\displaystyle\phi(x^{*},x_{n})+\phi(x_{n},y_{n})+2\langle x^{*}-x_{n},Jx_{n}-Jy_{n}\rangle$ $\displaystyle=$ $\displaystyle\phi(x^{*},x_{n})+\phi(x_{n},y_{n})+2\langle x_{n}-x^{*},Jy_{n}-Jx_{n}\rangle.$ (12) Putting (12) into (11), we get $\displaystyle\phi(x^{*},x_{n+1})=$ $\displaystyle\phi(x^{*},x_{n})+\phi(x_{n},y_{n})+2\langle x_{n}-x^{*},Jy_{n}-Jx_{n}\rangle$ $\displaystyle-2\lambda_{n}\langle Ay_{n}-Ax_{n},y_{n}-x^{*}\rangle+2\kappa^{2}\|\lambda_{n}(Ay_{n}-Ax_{n})\|^{2}$ $\displaystyle=$ $\displaystyle\phi(x^{*},x_{n})+\phi(x_{n},y_{n})-2\langle y_{n}-x_{n},Jy_{n}-Jx_{n}\rangle+2\langle y_{n}-x^{*},Jy_{n}-Jx_{n}\rangle$ $\displaystyle-2\lambda_{n}\langle Ay_{n}-Ax_{n},y_{n}-x^{*}\rangle+2\kappa^{2}\|\lambda_{n}(Ay_{n}-Ax_{n})\|^{2}.$ (13) Using Lemma 2.9 (ii), we get $-\phi(y_{n},x_{n})+2\langle y_{n}-x_{n},Jy_{n}-Jx_{n}\rangle=\phi(x_{n},y_{n}).$ (14) Substituting (14) into (13), we have $\displaystyle\phi(x^{*},x_{n+1})\leq$ $\displaystyle\phi(x^{*},x_{n})-\phi(y_{n},x_{n})+2\langle y_{n}-x^{*},Jy_{n}-Jx_{n}\rangle-2\lambda_{n}\langle Ay_{n}-Ax_{n},y_{n}-x^{*}\rangle$ $\displaystyle+2\kappa^{2}\|\lambda_{n}(Ay_{n}-Ax_{n})\|^{2}$ $\displaystyle=$ $\displaystyle\phi(x^{*},x_{n})-\phi(y_{n},x_{n})+2\kappa^{2}\|\lambda_{n}(Ay_{n}-Ax_{n})\|^{2}$ $\displaystyle-2\langle Jx_{n}-Jy_{n}-\lambda_{n}(Ax_{n}-Ay_{n}),y_{n}-x^{*}\rangle.$ (15) Since $y_{n}=(J+\lambda_{n}B)^{-1}JoJ^{-1}(Jx_{n}-\lambda_{n}Ax_{n})$, we have $Jx_{n}-\lambda_{n}Ax_{n}\in(J+\lambda_{n}B)y_{n}$. Using the fact that $B$ is maximal monotone, then there exists $v_{n}\in By_{n}$ such that $Jx_{n}-\lambda_{n}Ax_{n}=Jy_{n}+\lambda_{n}v_{n}$. Therefore $v_{n}=\frac{1}{\lambda_{n}}(Jx_{n}-Jy_{n}-\lambda_{n}Ax_{n}).$ (16) On the other hand, we know that $0\in(Ax^{*}+Bx^{*})$ and $Ay_{n}+v_{n}\in(A+B)y_{n}$. Since $A+B$ is maximal monotone, we obtain $\displaystyle\langle Ay_{n}+v_{n},y_{n}-x^{*}\rangle\geq 0.$ (17) Putting (16) into (17), we get $\displaystyle\langle Jx_{n}-Jy_{n}-\lambda_{n}(Ax_{n}-Ay_{n}),y_{n}-x^{*}\rangle\geq 0.$ (18) Now, using (18) in (15), we get $\displaystyle\phi(x^{*},x_{n+1})\leq$ $\displaystyle\phi(x^{*},x_{n})-\phi(y_{n},x_{n})+2\kappa^{2}\|\lambda_{n}(Ay_{n}-Ax_{n})\|^{2}$ $\displaystyle\leq$ $\displaystyle\phi(x^{*},x_{n})-\phi(y_{n},x_{n})+2\kappa^{2}\lambda_{n}^{2}L^{2}\mu\phi(y_{n},x_{n})$ $\displaystyle=$ $\displaystyle\phi(x^{*},x_{n})-(1-2\kappa^{2}\lambda_{n}^{2}L^{2}\mu)\phi(y_{n},x_{n}).$ (19) Using Assumption 3.2, we get $\displaystyle\phi(x^{*},x_{n+1})\leq\phi(x^{*},x_{n}),$ (20) which shows that $\lim\phi(x^{*},x_{n})$ exists and hence, $\\{\phi(x^{*},x_{n})\\}$ is bounded. Therefore $\\{x_{n}\\}$ is bounded. ∎ ###### Definition 3.5. The duality mapping $J$ is weakly sequentially continuous if, for any sequence $\\{x_{n}\\}\subset E$ such that $x_{n}\rightharpoonup x$ as $n\rightarrow\infty$, then $Jx_{n}\rightharpoonup^{*}Jx$ as $n\rightarrow\infty$. It is known that the normalized duality map on $\ell_{p}$ spaces, $1<p<\infty$, is weakly sequentially continuous. We now obtain the weak convergence result of Algorithm 3.3 in the next theorem. ###### Theorem 3.6. Let Assumptions 3.1 and 3.2 hold. Assume that $J$ is weakly sequentially continuous on $E$ and let the sequence $\\{x_{n}\\}_{n=1}^{\infty}$ be generated by Algorithm 3.3. Then $\\{x_{n}\\}$ converges weakly to $z\in(A+B)^{-1}(0)$. Moreover, $z:=\underset{n\rightarrow\infty}{\lim}\Pi_{(A+B)^{-1}(0)}(x_{n})$. ###### Proof. Let $x^{*}\in(A+B)^{-1}(0)$. From (19), we have $\displaystyle 0$ $\displaystyle<$ $\displaystyle[1-2\kappa^{2}b^{2}L^{2}\mu]\phi(y_{n},x_{n})\leq[1-2\kappa^{2}\lambda_{n}^{2}L^{2}\mu]\phi(y_{n},x_{n})$ (21) $\displaystyle\leq$ $\displaystyle\phi(x^{*},x_{n})-\phi(x^{*},x_{n+1}).$ Since $\lim_{n\rightarrow\infty}\phi(x^{*},x_{n})$ exists, we obtain from (21) that $\underset{n\rightarrow\infty}{\lim}\phi(y_{n},x_{n})=0.$ Applying Lemma 2.11, we get $\underset{n\rightarrow\infty}{\lim}\|x_{n}-y_{n}\|=0.$ Since $E$ is uniformly smooth, the duality mapping $J$ is uniformly norm-to- norm continuous on each bounded subset of $E$. Hence, we have $\underset{n\rightarrow\infty}{\lim}\|Jx_{n}-Jy_{n}\|=0.$ Since $\\{x_{n}\\}$ is bounded by Lemma 3.4, there exists a subsequence $\\{x_{n_{i}}\\}$ of $\\{x_{n}\\}$ and $z\in C$ such that $x_{n_{i}}\rightharpoonup z$. Since $\underset{n\rightarrow\infty}{\lim}\|x_{n}-y_{n}\|=0$, it follows that $x_{{n_{i}}+1}\rightharpoonup z$. We now show that $z\in(A+B)^{-1}(0)$. Suppose $(v,u)\in\textrm{Graph}(A+B)$. This implies that $Ju-Av\in Bv$. Furthermore, we obtain from $y_{n_{i}}=(J+\lambda_{n_{i}}B)^{-1}JoJ^{-1}(Jx_{n_{i}}-\lambda_{n_{i}}Ax_{n_{i}})$ that $(J-\lambda_{n_{i}}A)x_{n_{i}}\in(J+\lambda_{n_{i}}B)y_{n_{i}},$ and thus $\frac{1}{\lambda_{n_{i}}}(Jx_{n_{i}}-Jy_{n_{i}}-\lambda_{n_{i}}Ax_{n_{i}})\in By_{n_{i}}.$ Using the fact that $B$ is maximal monotone, we obtain $\langle v-y_{n_{i}},Ju- Av-\frac{1}{\lambda_{n_{i}}}(Jx_{n_{i}}-Jy_{n_{i}}-\lambda_{n_{i}}Ax_{n_{i}})\rangle\geq 0.$ Therefore, $\displaystyle\langle v-y_{n_{i}},Ju\rangle$ $\displaystyle\geq$ $\displaystyle\langle v-y_{n_{i}},Av+\frac{1}{\lambda_{n_{i}}}(Jx_{n_{i}}-Jy_{n_{i}}-\lambda_{n_{i}}Ax_{n_{i}})\rangle$ $\displaystyle=$ $\displaystyle\langle v-y_{n_{i}},Av- Ax_{n_{i}}\rangle+\langle v-y_{n_{i}},\frac{1}{\lambda_{n_{i}}}(Jx_{n_{i}}-Jy_{n_{i}})\rangle$ $\displaystyle=$ $\displaystyle\langle v-y_{n_{i}},Av- Ay_{n_{i}}\rangle+\langle v-y_{n_{i}},Ay_{n_{i}}-Ax_{n_{i}}\rangle$ $\displaystyle+\langle v-y_{n_{i}},\frac{1}{\lambda_{n_{i}}}(Jx_{n_{i}}-Jy_{n_{i}})\rangle$ $\displaystyle\geq$ $\displaystyle\langle v-y_{n_{i}},Ay_{n_{i}}-Ax_{n_{i}}\rangle+\langle v-y_{n_{i}},\frac{1}{\lambda_{n_{i}}}(Jx_{n_{i}}-Jy_{n_{i}})\rangle.$ By the fact that $\underset{n\rightarrow\infty}{\lim}\|x_{n}-y_{n}\|=0$ and $A$ is Lipschitz continuous, we obtain $\underset{n\rightarrow\infty}{\lim}\|Ax_{n}-Ay_{n}\|=0$. Consequently, we obtain that $\langle v-z,Ju\rangle\geq 0.$ By the maximal monotonicity of $A+B$, we have $0\in(A+B)z$. Hence, $z\in(A+B)^{-1}(0)$. Let $u_{n}:=\Pi_{(A+B)^{-1}(0)}(x_{n})$. By (20) and Lemma 2.12, we have that $\\{u_{n}\\}$ is a Cauchy sequence. Since $(A+B)^{-1}(0)$ is closed, we have that $\\{u_{n}\\}$ converges strongly to $w\in(A+B)^{-1}(0)$. By the uniform smoothness of $E$, we also have $\underset{n\rightarrow\infty}{\lim}\|Ju_{n}-Jw\|=0$. We then show that $z=w$. Using Lemma 2.10 (i), $u_{n}=\Pi_{(A+B)^{-1}(0)}(x_{n})$ and $z\in(A+B)^{-1}(0)$, we have $\langle z-u_{n},Ju_{n}-Jx_{n}\rangle\geq 0,\leavevmode\nobreak\ \leavevmode\nobreak\ \forall n\geq 1.$ Therefore, $\displaystyle\langle z-w,Jx_{n}-Ju_{n}\rangle$ $\displaystyle=$ $\displaystyle\langle z-u_{n},Jx_{n}-Ju_{n}\rangle+\langle u_{n}-w,Jx_{n}-Ju_{n}\rangle$ $\displaystyle\leq$ $\displaystyle\|u_{n}-w\|\|Jx_{n}-Ju_{n}\|\leq M\|u_{n}-w\|,\leavevmode\nobreak\ \leavevmode\nobreak\ \forall n\geq 1,$ where $M:=\underset{n\geq 1}{\sup}\|Jx_{n}-Ju_{n}\|$. Using $n=n_{i}$ in $\underset{n\rightarrow\infty}{\lim}\|u_{n}-w\|=0,\underset{n\rightarrow\infty}{\lim}\|Ju_{n}-Jw\|=0$ and the weakly sequential continuity of $J$, we obtain $\langle z-w,Jz-Jw\rangle\leq 0$ as $i\rightarrow\infty$. Therefore, $\langle z-w,Jz-Jw\rangle=0$. Since $E$ is strictly convex, we have $z=w$. Therefore, the sequence $\\{x_{n}\\}$ converges weakly to $z=\underset{n\rightarrow\infty}{\lim}\Pi_{(}A+B)^{-1}(0)(x_{n})$. This completes the proof. ∎ It is easy to see from Algorithm 3.3 above and Lemma 2.17 that $x_{n}=y_{n}$ if and only if $x_{n}\in(A+B)^{-1}(0)$. Also, we have already established that $\|x_{n}-y_{n}\|\rightarrow 0$ holds when $(A+B)^{-1}(0)\neq\emptyset$. Therefore, using the $\|x_{n}-y_{n}\|$ as a measure of convergence rate, we obtain the following non asymptotic rate of convergence of our proposed Algorithm 3.3. ###### Theorem 3.7. Let Assumptions 3.1 and 3.2 hold. Let the sequence $\\{x_{n}\\}_{n=1}^{\infty}$ be generated by Algorithm 3.3. Then $\min_{1\leq k\leq n}\|x_{k}-y_{k}\|=O(1/\sqrt{n})$. ###### Proof. We obtain from (19) that $\displaystyle\phi(x^{*},x_{n+1})\leq\phi(x^{*},x_{n})-(1-2\kappa^{2}\lambda_{n}^{2}L^{2}\mu)\phi(y_{n},x_{n}).$ Hence, we have from Lemma 2.11 that $\displaystyle\frac{1}{\mu}(1-2\kappa^{2}\lambda_{n}^{2}L^{2}\mu)\|x_{n}-y_{n}\|^{2}$ $\displaystyle\leq$ $\displaystyle(1-2\kappa^{2}\lambda_{n}^{2}L^{2}\mu)\phi(y_{n},x_{n})$ $\displaystyle\leq$ $\displaystyle\phi(x^{*},x_{n})-\phi(x^{*},x_{n+1}).$ By Assumption 3.2, we get $\displaystyle\sum_{k=1}^{n}\|x_{k}-y_{k}\|^{2}\leq\frac{\mu}{(1-2\kappa^{2}\lambda_{n}^{2}L^{2}\mu)}\phi(x^{*},x_{1}).$ Therefore, $\min_{1\leq k\leq n}\|x_{k}-y_{k}\|^{2}\leq\frac{\mu}{n(1-2\kappa^{2}\lambda_{n}^{2}L^{2}\mu)}\phi(x^{*},x_{1}).$ This implies that $\min_{1\leq k\leq n}\|x_{k}-y_{k}\|=O(1/\sqrt{n}).$ ∎ Next, we propose another iterative method such that the sequence of step-sizes does not depend on the Lipschitz constant of monotone operator $A$ in problem (1). ###### Algorithm 3.8. Step 0: Let Assumption 3.1 hold. Given $\gamma>0,l\in(0,1)$ and $\theta\in(0,\frac{1}{\sqrt{2\mu}\kappa})$. Let $x_{1}\in E$ be a given starting point. Set $n:=1$. Step 1: Compute $y_{n}:=J_{\lambda_{n}}^{B}J^{-1}(Jx_{n}-\lambda_{n}Ax_{n})$, where $\lambda_{n}$ is chosen to be the largest $\lambda\in\\{\gamma,\gamma l,\gamma l^{2},\ldots\\}$ satisfying $\displaystyle\lambda\|Ax_{n}-Ay_{n}\|\leq\theta\|x_{n}-y_{n}\|.$ (22) If $x_{n}-y_{n}=0$: STOP. Step 2: Compute $\displaystyle x_{n+1}=J^{-1}[Jy_{n}-\lambda_{n}(Ay_{n}-Ax_{n})].$ (23) Step 3: Set $n\leftarrow n+1$, and go to Step 1. Before we establish the weak convergence analysis of Algorithm 3.8, we first show that the line search rule given in (22) is well-defined in this lemma. ###### Lemma 3.9. The line search rule (22) in Algorithm 3.8 is well-defined and $\min\Big{\\{}\gamma,\frac{\theta l}{L}\Big{\\}}\leq\lambda_{n}\leq\gamma.$ ###### Proof. Using the Lipschitz continuity of $A$ on $E$, we obtain $\|Ax_{n}-A(J_{\lambda_{n}}^{B}J^{-1}(Jx_{n}-\lambda_{n}Ax_{n}))\|\leq L\|x_{n}-J_{\lambda_{n}}^{B}J^{-1}(Jx_{n}-\lambda_{n}Ax_{n})\|.$ This implies that $\frac{\theta}{L}\|Ax_{n}-A(J_{\lambda_{n}}^{B}J^{-1}(Jx_{n}-\lambda_{n}Ax_{n}))\|\leq\theta\|x_{n}-J_{\lambda_{n}}^{B}J^{-1}(Jx_{n}-\lambda_{n}Ax_{n})\|.$ Therefore, (22) holds whenever $\lambda_{n}\leq\frac{\theta}{L}$. Hence, $\lambda_{n}$ is well-defined. From the way $\lambda_{n}$ is chosen, we can clearly see that $\lambda_{n}\leq\gamma$. Now, suppose $\lambda_{n}=\gamma$, then (22) is satisfied and the lemma is proved. Suppose $\lambda_{n}<\gamma$. Then $\frac{\lambda_{n}}{l}$ violates (22) and we get $\displaystyle L\|x_{n}-J_{\lambda_{n}}^{B}J^{-1}(Jx_{n}-\lambda_{n}Ax_{n})\|$ $\displaystyle\geq$ $\displaystyle\|Ax_{n}-A(J_{\lambda_{n}}^{B}J^{-1}(Jx_{n}-\lambda_{n}Ax_{n}))\|$ $\displaystyle>$ $\displaystyle\frac{\theta}{\frac{\lambda_{n}}{l}}\|x_{n}-J_{\lambda_{n}}^{B}J^{-1}(Jx_{n}-\lambda_{n}Ax_{n})\|.$ This implies that $\lambda_{n}>\frac{\theta l}{L}$. This completes the proof. ∎ We now give a weak convergence result using Algorithm 3.8 in the next theorem. ###### Theorem 3.10. Let Assumptions 3.1. Assume that $J$ is weakly sequentially continuous on $E$ and let the sequence $\\{x_{n}\\}_{n=1}^{\infty}$ be generated by Algorithm 3.8. Then $\\{x_{n}\\}$ converges weakly to $z\in(A+B)^{-1}(0)$. Moreover, $z:=\underset{n\rightarrow\infty}{\lim}\Pi_{(A+B)^{-1}(0)}(x_{n})$. ###### Proof. Using the same line of arguments as in the proof of Lemma 3.4, we can obtain from (19) that $\displaystyle\phi(x^{*},x_{n+1})$ $\displaystyle\leq$ $\displaystyle\phi(x^{*},x_{n})-\phi(y_{n},x_{n})+2\kappa^{2}\|\lambda_{n}(Ay_{n}-Ax_{n})\|^{2}$ (24) $\displaystyle\leq$ $\displaystyle\phi(x^{*},x_{n})-\phi(y_{n},x_{n})+2\kappa^{2}\theta^{2}\|y_{n}-x_{n}\|^{2}$ $\displaystyle\leq$ $\displaystyle\phi(x^{*},x_{n})-\phi(y_{n},x_{n})+2\kappa^{2}\theta^{2}\mu\phi(y_{n},x_{n})$ $\displaystyle=$ $\displaystyle\phi(x^{*},x_{n})-(1-2\kappa^{2}\theta^{2}\mu)\phi(y_{n},x_{n}).$ Since $\theta^{2}<\frac{1}{2\kappa^{2}\mu}$, we get $\displaystyle\phi(x^{*},x_{n+1})\leq\phi(x^{*},x_{n}),$ (25) which shows that $\lim\phi(x^{*},x_{n})$ exists and hence, $\\{\phi(x^{*},x_{n})\\}$ is bounded. Therefore $\\{x_{n}\\}$ is bounded. The rest of the proof follows by using the same arguments as in the proof of Theorem 3.6. The completes the proof. ∎ Finally, we give a modification of Algorithm 3.3 and consequently obtain the strong convergence analysis below. ###### Algorithm 3.11. Step 0: Let Assumptions 3.1 and 3.2 hold. Suppose that $\\{\alpha_{n}\\}$ is a real sequence in (0,1) and let $x_{1}\in E$ be a given starting point. Set $n:=1$. Step 1: Compute $y_{n}:=J_{\lambda_{n}}^{B}J^{-1}(Jx_{n}-\lambda_{n}Ax_{n})$. If $x_{n}-y_{n}=0$: STOP. Step 2: Compute $\displaystyle w_{n}=J^{-1}[Jy_{n}-\lambda_{n}(Ay_{n}-Ax_{n})]$ (26) and $\displaystyle x_{n+1}=J^{-1}[\alpha_{n}Jx_{1}+(1-\alpha_{n})Jw_{n}].$ (27) Step 3: Set $n\leftarrow n+1$, and go to Step 1. ###### Theorem 3.12. Let Assumptions 3.1 and 3.2 hold. Suppose that $\lim_{n\to\infty}\alpha_{n}=0$ and $\sum_{n=1}^{\infty}\alpha_{n}=\infty$. Let the sequence $\\{x_{n}\\}_{n=1}^{\infty}$ be generated by Algorithm 3.11. Then $\\{x_{n}\\}$ converges strongly to $z=\Pi_{(A+B)^{-1}(0)}(x_{1})$. ###### Proof. By Lemma 3.4, we have that $\\{x_{n}\\}$ is bounded. Furthermore, using Lemma 2.10 with (26) and (27), we have $\displaystyle\phi(z,x_{n+1})$ $\displaystyle=$ $\displaystyle\phi(z,J^{-1}(\alpha_{n}Jx_{1}+(1-\alpha_{n})Jw_{n}))$ (28) $\displaystyle=$ $\displaystyle V(z,\alpha_{n}Jx_{1}+(1-\alpha_{n})Jw_{n}))$ $\displaystyle\leq$ $\displaystyle V(z,\alpha_{n}Jx_{1}+(1-\alpha_{n})Jw_{n}-\alpha_{n}(Jx_{1}-Jz))$ $\displaystyle+2\alpha_{n}\langle Jx_{1}-Jz,x_{n+1}-z\rangle$ $\displaystyle=$ $\displaystyle V(z,\alpha_{n}Jz+(1-\alpha_{n})Jw_{n})+2\alpha_{n}\langle Jx_{1}-Jz,x_{n+1}-z\rangle$ $\displaystyle\leq$ $\displaystyle\alpha_{n}V(z,Jz)+(1-\alpha_{n})V(z,Jw_{n})+2\alpha_{n}\langle Jx_{1}-Jz,x_{n+1}-z\rangle$ $\displaystyle=$ $\displaystyle(1-\alpha_{n})V(z,Jw_{n})+2\alpha_{n}\langle Jx_{1}-Jz,x_{n+1}-z\rangle$ $\displaystyle\leq$ $\displaystyle(1-\alpha_{n})V(z,Jx_{n})+2\alpha_{n}\langle Jx_{1}-Jz,x_{n+1}-z\rangle$ $\displaystyle=$ $\displaystyle(1-\alpha_{n})\phi(z,x_{n})+2\alpha_{n}\langle Jx_{1}-Jz,x_{n+1}-z\rangle.$ Set $a_{n}:=\phi(x_{n},z)$ and divide the rest of the proof into two parts as follows. Case 1: Suppose that there exists $n_{0}\in\mathbb{N}$ such that $\\{\phi(z,x_{n})\\}_{n=n_{0}}^{\infty}$ is non-increasing. Then $\\{\phi(z,x_{n})\\}_{n=1}^{\infty}$ converges, and we therefore obtain $a_{n}-a_{n+1}\rightarrow 0,\leavevmode\nobreak\ \leavevmode\nobreak\ n\rightarrow\infty.$ (29) Using (20) in (27), we have $\displaystyle V(z,Jx_{n+1})$ $\displaystyle\leq$ $\displaystyle\alpha_{n}V(z,Jx_{1})+(1-\alpha_{n})V(z,Jw_{n})$ (30) $\displaystyle\leq$ $\displaystyle\alpha_{n}V(Jx_{1},z)+(1-\alpha_{n})V(Jx_{n},z)$ $\displaystyle-(1-\alpha_{n})[1-2\kappa^{2}\theta^{2}\mu]V(y_{n},Jx_{n}).$ This implies from (30) that $(1-\alpha_{n})[1-2\kappa^{2}\theta^{2}\mu]V(y_{n},Jx_{n})\leq V(Jx_{n},z)-V(Jx_{n+1},z)+\alpha_{n}M_{1},$ for some $M_{1}>0$. Thus, $(1-\alpha_{n})[1-2\kappa^{2}\theta^{2}\mu]\phi(y_{n},x_{n})\rightarrow 0,\leavevmode\nobreak\ \leavevmode\nobreak\ n\rightarrow\infty.$ Hence, $\phi(y_{n},x_{n})\rightarrow 0,\leavevmode\nobreak\ \leavevmode\nobreak\ n\rightarrow\infty.$ Consequently, $\|x_{n}-y_{n}\|\rightarrow 0,\leavevmode\nobreak\ \leavevmode\nobreak\ n\rightarrow\infty.$ By (26), we get $\displaystyle\|Jw_{n}-Jy_{n}\|$ $\displaystyle=$ $\displaystyle\lambda_{n}\|Ay_{n}-Ax_{n}\|$ $\displaystyle\leq$ $\displaystyle b\|Ay_{n}-Ax_{n}\|\rightarrow 0,\leavevmode\nobreak\ \leavevmode\nobreak\ n\rightarrow\infty.$ Therefore, $\|w_{n}-y_{n}\|\rightarrow 0,\leavevmode\nobreak\ \leavevmode\nobreak\ n\rightarrow\infty.$ Moreover, we obtain from (27) that $\displaystyle\|Jx_{n+1}-Jw_{n}\|$ $\displaystyle=$ $\displaystyle\alpha_{n}\|Jx_{1}-Jw_{n}\|\leq\alpha_{n}M_{2}\rightarrow 0,\leavevmode\nobreak\ \leavevmode\nobreak\ n\rightarrow\infty,$ (31) for some $M_{2}>0$. Since $J^{-1}$ is norm-to-norm uniformly continuous on bounded subsets of $E^{*}$, we have that $\|x_{n+1}-w_{n}\|\rightarrow 0,\leavevmode\nobreak\ \leavevmode\nobreak\ n\rightarrow\infty.$ Now, $\|x_{n+1}-x_{n}\|\leq\|x_{n+1}-w_{n}\|+\|w_{n}-y_{n}\|+\|y_{n}-x_{n}\|\rightarrow 0,\leavevmode\nobreak\ \leavevmode\nobreak\ n\rightarrow\infty.$ Since $\\{x_{n}\\}$ is a bounded sunset of $E$, we can choose a subsequence $\\{x_{n_{k}}\\}$ of $\\{x_{n}\\}$ such that $x_{n_{k}}\rightharpoonup p\in E$ and $\displaystyle\limsup_{n\rightarrow\infty}\langle Jx_{1}-Jz,x_{n}-z\rangle\leq 2\lim_{k\rightarrow\infty}\langle Jx_{1}-Jz,x_{n_{k}}-z\rangle.$ Since $z=\Pi_{C}x_{1}$, we get $\displaystyle\limsup_{n\rightarrow\infty}\langle Jx_{1}-Jz,x_{n}-z\rangle$ $\displaystyle\leq$ $\displaystyle 2\lim_{k\rightarrow\infty}\langle Jx_{1}-Jz,x_{n_{k}}-z\rangle$ (32) $\displaystyle=$ $\displaystyle 2\langle Jx_{1}-Jz,p-z\rangle\leq 0.$ This implies that $\limsup_{n\rightarrow\infty}\langle Jx_{1}-Jz,x_{n}-z\rangle\leq 0.$ Using Lemma 2.15 and (32) in (28), we obtain $\lim_{n\rightarrow\infty}\phi(z,x_{n})=0.$ Thus, $x_{n}\rightarrow z$, $n\rightarrow\infty$. Case 2: Suppose that there exists a subsequence $\\{x_{n_{j}}\\}$ of $\\{x_{n}\\}$ such that $\phi(z,x_{m_{j}})<\phi(z,x_{{m_{j}}+1}),\leavevmode\nobreak\ \leavevmode\nobreak\ \forall j\in\mathbb{N}.$ From Lemma 2.14, there exists a nondecreasing sequence $\\{n_{k}\\}$ of $\mathbb{N}$ such that $\lim_{k\rightarrow\infty}\lim n_{k}=\infty$ and the following inequalities hold for all $k\in\mathbb{N}$: $\displaystyle\phi(z,x_{n_{k}})\leq\phi(z,x_{{n_{k}}+1})\leavevmode\nobreak\ \leavevmode\nobreak\ {\rm and}\leavevmode\nobreak\ \leavevmode\nobreak\ \phi(z,x_{k})\leq\phi(z,x_{{n_{k}}+1}).$ (33) Observe that $\displaystyle\phi(z,x_{n_{k}})$ $\displaystyle\leq$ $\displaystyle\phi(z,x_{{n_{k}}+1})\leq\alpha_{n_{k}}\phi(z,x_{1})+(1-\alpha_{n_{k}})\phi(z,w_{n_{k}})$ $\displaystyle\leq$ $\displaystyle\alpha_{n_{k}}\phi(z,x_{1})+(1-\alpha_{n_{k}})\phi(z,x_{n_{k}}).$ Since $\lim_{n\rightarrow\infty}\alpha_{n}=0$, we get $\phi(z,x_{{n_{k}}+1})-\phi(z,x_{n_{k}})\rightarrow 0,\leavevmode\nobreak\ \leavevmode\nobreak\ k\rightarrow\infty.$ Since $\\{x_{n_{k}}\\}$ is bounded, there exists a subsequence of $\\{x_{n_{k}}\\}$ still denoted by $\\{x_{n_{k}}\\}$ which converges weakly to $p\in E$. Repeating the same arguments as in Case 1 above, we can show that $\|x_{n_{k}}-y_{n_{k}}\|\rightarrow 0,\leavevmode\nobreak\ \leavevmode\nobreak\ k\rightarrow\infty,\|y_{n_{k}}-w_{n_{k}}\|\rightarrow 0,\leavevmode\nobreak\ \leavevmode\nobreak\ k\rightarrow\infty\leavevmode\nobreak\ \leavevmode\nobreak\ {\rm and}\leavevmode\nobreak\ \leavevmode\nobreak\ \|x_{{n_{k}}+1}-x_{n_{k}}\|\rightarrow 0,\leavevmode\nobreak\ \leavevmode\nobreak\ k\rightarrow\infty.$ Similarly, we can conclude that $\displaystyle\limsup_{k\rightarrow\infty}\langle x_{{n_{k}}+1}-z,Jx_{1}-Jz\rangle$ $\displaystyle=$ $\displaystyle\limsup_{k\rightarrow\infty}\langle x_{n_{k}}-z,Jx_{1}-Jz\rangle\leq 0.$ (34) It then follows from (28) and (33) that $\displaystyle\phi(z,x_{{n_{k}}+1})$ $\displaystyle\leq$ $\displaystyle(1-\alpha_{n_{k}})\phi(z,x_{n_{k}})+\alpha_{n_{k}}\langle x_{{n_{k}}+1}-z,Jx_{1}-Jz\rangle$ $\displaystyle\leq$ $\displaystyle(1-\alpha_{n_{k}})\phi(z,x_{{n_{k}}+1})+\alpha_{n_{k}}\langle x_{{n_{k}}+1}-z,Jx_{1}-Jz\rangle.$ Since $\alpha_{n_{k}}>0$, we get $\phi(z,x_{n_{k}})\leq\phi(z,x_{{n_{k}}+1})\leq\langle x_{{n_{k}}+1}-z,Jx_{1}-Jz\rangle.$ By (34), we have that $\limsup_{k\rightarrow\infty}\phi(z,x_{n_{k}})\leq\limsup_{k\rightarrow\infty}\langle x_{{n_{k}}+1}-z,Jx_{1}-Jz\rangle.$ Therefore, $x_{k}\rightarrow z,\leavevmode\nobreak\ \leavevmode\nobreak\ k\rightarrow\infty.$ This concludes the proof. ∎ ###### Remark 3.13. Our proposed Algorithms 3.3 and 3.11 are more applicable than the proposed methods in [10, 12, 23, 29, 30, 44, 45, 46, 42, 49] even in Hilbert spaces. The methods proposed in [12, 23, 29, 30, 44, 45, 46, 42, 49] are only applicable for solving problem (1) in the case when $B$ is maximal monotone and $A$ is inverse-strongly monotone (co-coercive) operator in real Hilbert spaces. Our Algorithms 3.3 and 3.11 are applicable for the case when $B$ is maximal monotone and $A$ is monotone operator even in 2-uniformy convex and uniformly smooth Banach spaces (e.g., $L_{p},1<p\leq 2$). Our results in this paper also complement the results of [14, 22]. ## 4 Application In this section, we apply our results to the minimization of composite objective function of the type $\displaystyle\min_{x\in E}f(x)+g(x),$ (35) where $f:E\rightarrow\mathbb{R}\cup\\{+\infty\\}$ is proper, convex and lower semi-continuous functional and $g:E\rightarrow\mathbb{R}$ is convex functional. Many optimization problems from image processing [9], statistical regression, machine learning (see, e.g., [50] and the references contained therein), etc can be adapted into the form of (35). In this setting, we assume that $g$ represents the ”smooth part” of the functional where $f$ is assumed to be non- smooth. Specifically, we assume that $g$ is G$\hat{a}$teaux-differentiable with derivative $\nabla g$ which is Lipschitz-continuous with constant $L$. Then by [37, thm. 3.13], we have $\langle\nabla g(x)-\nabla g(y),x-y\rangle\geq\frac{1}{L}\|\nabla g(x)-\nabla g(y)\|^{2},\leavevmode\nobreak\ \leavevmode\nobreak\ \forall x,y\in E.$ Therefore, $\nabla g$ is monotone and Lipschitz continuous with Lipschitz constant $L$. Observe that problem (35) is equivalent to find $\in E$ such that $\displaystyle 0\in\partial f(x)+\nabla g(x).$ (36) Then problem (36) is a special case of inclusion problem (1) with $A:=\nabla g$ and $B:=\partial f$. Next, we obtain the resolvent of $\partial f$. Let us fix $r>0$ and $z\in E$. Suppose $J_{r}^{\partial f}$ is the resolvent of $\partial f$. Then $Jz\in J(J_{r}^{\partial f})+r\partial f(J_{r}^{\partial f}).$ Hence we obtain $\displaystyle 0\in\partial f(J_{r}^{\partial f})+\frac{1}{r}J(J_{r}^{\partial f})-\frac{1}{r}Jz=\partial\Big{(}f+\frac{1}{2r}\|.\|^{2}-\frac{1}{r}Jz\Big{)}J_{r}^{\partial f}.$ Therefore, $J_{r}^{\partial f}(z)={\rm argmin}_{y\in E}\Big{\\{}f(y)+\frac{1}{2r}\|y\|^{2}-\frac{1}{r}\langle y,Jz\rangle\Big{\\}}.$ We can then write $y_{n}$ in Algorithm 3.3 as $y_{n}={\rm argmin}_{y\in E}\Big{\\{}f(y)+\frac{1}{2\lambda_{n}}\|y\|^{2}-\frac{1}{\lambda_{n}}\langle y,Jx_{n}-\lambda_{n}\nabla g(x_{n})\rangle\Big{\\}}.$ We obtain the following weak and strong convergence results for problem (35). ###### Theorem 4.1. Let $E$ be a real 2-uniformly convex Banach space which is also uniformly smooth and the solution set $S$ of problem (35) be nonempty. Suppose $\\{\lambda_{n}\\}_{n=1}^{\infty}$ satisfies the condition $0<a\leq\lambda_{n}\leq b<\displaystyle\frac{1}{\sqrt{2\mu}\kappa L}$. Assume that $J$ is weakly sequentially continuous on $E$ and let the sequence $\\{x_{n}\\}_{n=1}^{\infty}$ be generated by $\displaystyle\left\\{\begin{array}[]{llll}&x_{1}\in E,\\\ &y_{n}={\rm argmin}_{y\in E}\Big{\\{}f(y)+\frac{1}{2\lambda_{n}}\|y\|^{2}-\frac{1}{\lambda_{n}}\langle y,Jx_{n}-\lambda_{n}\nabla g(x_{n})\rangle\Big{\\}}\\\ &x_{n+1}=J^{-1}[Jy_{n}-\lambda_{n}(\nabla g(y_{n})-\nabla g(x_{n}))],\leavevmode\nobreak\ \leavevmode\nobreak\ n\geq 1.\end{array}\right.$ (40) Then $\\{x_{n}\\}$ converges weakly to $z\in S$. Moreover, $z:=\underset{n\rightarrow\infty}{\lim}\Pi_{S}(x_{n})$. ###### Theorem 4.2. Let $E$ be a real 2-uniformly convex Banach space which is also uniformly smooth and the solution set $S$ of problem (35) be nonempty. Suppose $\\{\lambda_{n}\\}_{n=1}^{\infty}$ satisfies the condition $0<a\leq\lambda_{n}\leq b<\displaystyle\frac{1}{\sqrt{2\mu}\kappa L}$. Suppose that $\\{\alpha_{n}\\}$ is a real sequence in (0,1) with $\lim_{n\to\infty}\alpha_{n}=0$ and $\sum_{n=1}^{\infty}\alpha_{n}=\infty$. Let the sequence $\\{x_{n}\\}_{n=1}^{\infty}$ be generated by $\displaystyle\left\\{\begin{array}[]{llll}&x_{1}\in E,\\\ &y_{n}={\rm argmin}_{y\in E}\Big{\\{}f(y)+\frac{1}{2\lambda_{n}}\|y\|^{2}-\frac{1}{\lambda_{n}}\langle y,Jx_{n}-\lambda_{n}\nabla g(x_{n})\rangle\Big{\\}}\\\ &w_{n}=J^{-1}[Jy_{n}-\lambda_{n}(\nabla g(y_{n})-\nabla g(x_{n}))],\\\ &x_{n+1}=J^{-1}[\alpha_{n}Jx_{1}+(1-\alpha_{n})Jw_{n}],\leavevmode\nobreak\ \leavevmode\nobreak\ n\geq 1.\end{array}\right.$ (45) Then $\\{x_{n}\\}$ converges strongly to $z=\Pi_{S}(x_{1})$. ###### Remark 4.3. * • Our result in Theorems 4.1 and 4.2 complement the results of Bredies [9, 19]. Consequently, our results in Section 3.1 extend the results of Bredies [9, 19] to inclusion problem (1). In particular, we do not assume boundedness of $\\{x_{n}\\}$ (which was imposed on the results of [9, 19]) in our results. Therefore, our result improves on the results of [9, 19]. * • The minimization problem (35) in this section extends the problem studied in [8, 15, 34, 50] and other related papers from Hilbert spaces to Banach spaces. ## 5 Conclusion We study the Tseng-type algorithm for finding a solution to monotone inclusion problem involving a sum of maximal monotone and a Lipschitz continuous monotone mapping in 2-uniformly convex Banach space which is also uniformly smooth. We prove both weak and strong convergence of sequences of iterates to the solution of the inclusion problem under some appropriate conditions. Many results on monotone inclusion problems with single maximal monotone operator can be considered as special cases of the problem studied in this paper. As far as we know, this is the first time an inclusion problem involving sum of maximal monotone and Lipschitz continuous monotone operators will be studied in Banach spaces. Therefore, the results of this paper open up many forthcoming results regarding the inclusion problem studied in this paper. Our next project involves the following. * • The results in this paper exclude $L_{p}$ spaces with $p>2$. Therefore, extension of the results in this paper to a more general reflexive Banach space will be desired. * • How to effectively compute the duality mapping $J$ and the resolvent of maximal monotone mapping $B$ during implementations of our proposed algorithms will be considered further. * • The numerical implementations of problem (1) arising from signal processing, image reconstruction, etc will be studied; * • Other ways of implementation of the step-sizes $\lambda_{n}$ to give faster convergence of the proposed methods in this paper will be given. Acknowledgements The project of the author has received funding from the European Research Council (ERC) under the European Union’s Seventh Framework Program (FP7 - 2007-2013) (Grant agreement No. 616160) ## References * [1] Alber, Y. I. Metric and generalized projection operators in Banach spaces: properties and applications. Theory and applications of nonlinear operators of accretive and monotone type, 15-50, Lecture Notes in Pure and Appl. Math., 178, Dekker, New York, 1996. * [2] Alber, Y.; Ryazantseva, I. Nonlinear ill-posed problems of monotone type. Springer, Dordrecht, 2006. xiv+410 pp. ISBN: 978-1-4020-4395-6; 1-4020-4395-3. * [3] Aoyama, K.; Kohsaka, F. Strongly relatively nonexpansive sequences generated by firmly nonexpansive-like mappings. Fixed Point Theory Appl. 2014, 2014:95, 13 pp. * [4] Avetisyan, K.; Djordjević, O.; Pavlović, M. Littlewood-Paley inequalities in uniformly convex and uniformly smooth Banach spaces. J. Math. Anal. Appl. 336 (2007), no. 1, 31–43. * [5] Ball, K.; Carlen, E. A.; Lieb, E. H. Sharp uniform convexity and smoothness inequalities for trace norms. Invent. Math. 115 (1994), no. 3, 463–482. * [6] Barbu, V. Nonlinear Semigroups and Differential Equations in Banach Spaces, Editura Academiei R.S.R., Bucharest, 1976. * [7] Beauzamy, B. Introduction to Banach spaces and their geometry. Second edition. North-Holland Mathematics Studies, 68. Notas de Matemática [Mathematical Notes], 86. North-Holland Publishing Co., Amsterdam, 1985. xv+338 pp. ISBN: 0-444-87878-5. * [8] Beck, A.; Teboulle, M. A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2 (2009), no. 1, 183-202. * [9] Bredies, K. A forward-backward splitting algorithm for the minimization of non-smooth convex functionals in Banach space. Inverse Problems 25 (2009), no. 1, 015005, 20 pp. * [10] Briceño-Arias, L. M. Forward-partial inverse-forward splitting for solving monotone inclusions. J. Optim. Theory Appl. 166 (2015), no. 2, 391-413. * [11] Chen, G. H.-G.; Rockafellar, R. T. Convergence rates in forward-backward splitting. SIAM J. Optim. 7 (1997), no. 2, 421–444. * [12] Cho, S. Y.; Qin, X.; Wang, L. Strong convergence of a splitting algorithm for treating monotone operators. Fixed Point Theory Appl. 2014, 2014:94, 15 pp. * [13] Cioranescu, I. Geometry of Banach spaces, duality mappings and nonlinear problems. Mathematics and its Applications, 62. Kluwer Academic Publishers Group, Dordrecht, 1990. xiv+260 pp. ISBN: 0-7923-0910-3. * [14] Combettes, P. L.; Nguyen, Q. V. Solving composite monotone inclusions in reflexive Banach spaces by constructing best Bregman approximations from their Kuhn-Tucker set. J. Convex Anal. 23 (2016), no. 2, 481-510. * [15] Combettes, P.; Wajs, V. R. Signal recovery by proximal forward-backward splitting. Multiscale Model. Simul. 4 (2005), no. 4, 1168-1200. * [16] Diestel, J. Geometry of Banach spaces—selected topics. Lecture Notes in Mathematics, Vol. 485. Springer-Verlag, Berlin-New York, 1975, xi+282 pp. * [17] Figiel, T. On the moduli of convexity and smoothness. Studia Math. 56 (1976), no. 2, 121-155. * [18] Gibali, A.; Thong, D. V. Tseng type methods for solving inclusion problems and its applications. Calcolo 55 (2018), no. 4, 55:49. * [19] Guan, W.-B., Song, W. The generalized forward-backward splitting method for the minimization of the sum of two functions in Banach spaces. Numer. Funct. Anal. Optim. 36 (2015), no. 7, 867-886. * [20] Güler, O. On the convergence of the proximal point algorithm for convex minimization. SIAM J. Control Optim. 29 (1991), 403-419. * [21] Iiduka, H.; Takahashi, W. Weak convergence of a projection algorithm for variational inequalities in a Banach space. J. Math. Anal. Appl. 339 (2008), no. 1, 668–679. * [22] Iusem, A. N.; Svaiter, B. F. Splitting methods for finding zeroes of sums of maximal monotone operators in Banach spaces. J. Nonlinear Convex Anal. 15 (2014), no. 2, 379-397. * [23] Jiao, H.; Wang, F. On an iterative method for finding a zero to the sum of two maximal monotone operators. J. Appl. Math. 2014, Art. ID 414031, 5 pp. * [24] Kamimura, S.; Kohsaka, F.; Takahashi, W. Weak and Strong Convergence Theorems for Maximal Monotone Operators in a Banach Space, Set-Valued Anal. 12 (2004), 417-429. * [25] Kamimura, S.; Takahashi, W. Strong convergence of a proximal-type algorithm in a Banach space. SIAM J. Optim. 13 (2002), no. 3, 938-945 (2003). * [26] Kohsaka, F.; Takahashi, W. Strong convergence of an iterative sequence for maximal monotone operators in a Banach space. Abstr. Appl. Anal. 2004, no. 3, 239-249. * [27] Lions, P. L. Une méthode itérative de résolution d’une inéquation variationnelle, Israel J. Math. 31 (1978), 204-208. * [28] Lions, P. L.; Mercier, B. Splitting algorithms for the sum of two nonlinear operators, SIAM J. Numer. Anal., 16 (1979), 964-979. * [29] Lin, L.-J.; Takahashi, W. A general iterative method for hierarchical variational inequality problems in Hilbert spaces and applications. Positivity 16 (2012), no. 3, 429-453. * [30] López, G.; Martín-Márquez, V.; Wang, F.; Xu, H.-K. Forward-backward splitting methods for accretive operators in Banach spaces. Abstr. Appl. Anal. 2012, Art. ID 109236, 25 pp. * [31] Maingé, P.-E. Strong convergence of projected subgradient methods for nonsmooth and nonstrictly convex minimization. Set-Valued Anal. 16 (2008), no. 7-8, 899–912. * [32] Martinet, B. Régularisation d’inéquations variationnelles par approximations successives. (French) Rev. Française Informat. Recherche Opérationnelle 4 (1970), Sér. R-3, 154–158. * [33] Moudafi, A.; Thera, M. Finding a zero of the sum of two maximal monotone operators, J. Optim. Theory Appl., 94 (1997), 425–448. * [34] Nguyen, T. P.; Pauwels, E.; Richard, E.; Suter, B. W. Extragradient method in optimization: convergence and complexity. J. Optim. Theory Appl. 176 (2018), no. 1, 137-162. * [35] Passty, G. B. Ergodic convergence to a zero of the sum of monotone operators in Hilbert spaces, J. Math. Anal. Appl., 72 (1979), 383-390. * [36] Peaceman, D. H.; Rachford, H. H. The numerical solutions of parabolic and elliptic differential equations, J. Soc. Indust. Appl. Math., 3 (1955), 28-41. * [37] Peypouquet, J. Convex optimization in normed spaces. Theory, methods and examples. With a foreword by Hedy Attouch. Springer Briefs in Optimization. Springer, Cham, 2015. xiv+124 pp. ISBN: 978-3-319-13709-4; 978-3-319-13710-0. * [38] Reich, S. A weak convergence theorem for the alternating method with Bregman distances, in: A.G. Kartsatos (Ed.), Theory and Applications of Nonlinear Operators of Accretive and Monotone Type, in: Lecture Notes Pure Appl. Math., vol. 178, Dekker, New York, 1996, pp. 313-318. * [39] Rockafellar, R. T. Monotone operators and the proximal point algorithm. SIAM J. Control. Optim. 14 (1976), 877-898. * [40] Rockafellar, R. T. Characterization of the subdifferentials of convex functions. Pacific J. Math. 17 (1966), 497-510. * [41] Rockafellar, R. T. On the maximal monotonicity of subdifferential mappings. Pacific J. Math. 33 (1970), 209-216. * [42] Shehu, Y.; Cai, G. Strong convergence result of forward-backward splitting methods for accretive operators in Banach spaces with applications. Rev. R. Acad. Cienc. Exactas Fís. Nat. Ser. A Math. RACSAM 112 (2018), no. 1, 71-87. * [43] Solodov, M. V.; Svaiter, B. F. Forcing strong convergence of proximal point iterations in a Hilbert space. Math. Programing 87 (2000), 189-202. * [44] Takahashi, S.; Takahashi, W.; Toyoda, M. Strong convergence theorems for maximal monotone operators with nonlinear mappings in Hilbert spaces. J. Optim. Theory Appl. 147 (2010), no. 1, 27-41. * [45] Takahashi, W.; Wong, N.-C.; Yao, J.-C. Two generalized strong convergence theorems of Halpern’s type in Hilbert spaces and applications. Taiwanese J. Math. 16 (2012), no. 3, 1151-1172. * [46] Takahashi, W. Strong convergence theorems for maximal and inverse-strongly monotone mappings in Hilbert spaces and applications. J. Optim. Theory Appl. 157 (2013), no. 3, 781-802. * [47] Takahashi, W. Nonlinear Functional Analysis, Yokohama Publishers, Yokohama 2000. * [48] Tseng, P. A modified forward-backward splitting method for maximal monotone mappings. SIAM J. Control Optim. 38 (2000), no. 2, 431–446. * [49] Wang, Y.; Wang, F. Strong convergence of the forward-backward splitting method with multiple parameters in Hilbert spaces. Optimization 67 (2018), no. 4, 493–505. * [50] Wang, Y.; Xu, H.-K. Strong convergence for the proximal-gradient method. J. Nonlinear Convex Anal. 15 (2014), no. 3, 581–593. * [51] Xu, H. K. Inequalities in Banach spaces with applications. Nonlinear Anal. 16 (1991), no. 12, 1127–1138. * [52] Xu, H. K. Iterative algorithms for nonlinear operators. J. London Math. Soc. (2) 66 (2002), no. 1, 240–256.
# A sparse grid discrete ordinate discontinuous Galerkin method for the radiative transfer equation Jianguo Huang******Corresponding author<EMAIL_ADDRESS>Yue Yu <EMAIL_ADDRESS>School of Mathematical Sciences, and MOE-LSC, Shanghai Jiao Tong University Shanghai 200240, China (Version 0.0, June 12, 2019) ###### Abstract The radiative transfer equation is a fundamental equation in transport theory and applications, which is a 5-dimensional PDE in the stationary one-velocity case, leading to great difficulties in numerical simulation. To tackle this bottleneck, we first use the discrete ordinate technique to discretize the scattering term, an integral with respect to the angular variables, resulting in a semi-discrete hyperbolic system. Then, we make the spatial discretization by means of the discontinuous Galerkin (DG) method combined with the sparse grid method. The final linear system is solved by the block Gauss-Seidal iteration method. The computational complexity and error analysis are developed in detail, which show the new method is more efficient than the original discrete ordinate DG method. A series of numerical results are performed to validate the convergence behavior and effectiveness of the proposed method. ###### keywords: Radiative transfer equation, Sparse grid method , Discrete ordinate method , Discontinuous Galerkin method ## 1 Introduction Radiation transport is a physical process of energy transfer in the form of electromagnetic radiation which is affected by absorption, emission and scattering as it passes through the background materials. The radiative transfer equation (RTE) is an important mathematical model used to describe these interactions, finds applications in a wide variety of subjects, including neutron transport, heat transfer, optics, astrophysics, inertial confinement fusion, and high temperature flow systems, see for examples [27, 12, 16, 2, 20, 40]. The RTE can be viewed as a hyperbolic-type integro-differential equation. Even for the stationary monochromatic RTE, it is five-dimensional in the phase space, and hence cannot have a closed-form solution in general. Thus, the numerical solution of the equation is unavoidable and critical in applications. In history, the Monte-Carlo method is a typical approach for numerical simulation (cf. [11] and the references therein). The advantage is its simplicity and dimension-free convergence, and the weakness is its heavy computational cost and slow convergence. Until now, there have developed many other numerical methods as well. For the angular discretization, the typical methods include discrete ordinate methods (or $S_{N}$ methods) and spherical harmonic methods (or $P_{N}$ method); for the spatial discretization, the typical methods include finite difference methods, finite element methods and spectral methods. We refer to [31, 7, 6, 27, 12, 18, 30, 20] for details. Due to the flexibility and easy implementation, the discrete ordinate method is frequently used for angular discretization in practice. If the spatial domain is regular, this semi-discrete method is further discretized by the Chebyshev spectral method in [28, 17, 5] and the meshless discretization in [34, 29, 42]. In recent years, the positivity-preserving schemes are also developed very technically in [45, 15, 47]. For numerical solvers such as source iteration and multigrid algorithms, one can refer to [13, 1, 38, 36]. On the other hand, except the Monte-Carlo method, all the methods mentioned above solve the problems with reduced dimensions. In this paper, we intend to attack the problem in its original form with 3-spatial variables and 2-angular variables. In this case, most usual methods suffer from the so-called “the curse of dimension”, which indicates the low rate of convergence in terms of number of degrees of freedom due to the high dimensionality of the underlying problem. To the best of our knowledge, the sparse grid method, also called the sparse tensor product method, is an effective way to overcome the bottleneck. Historically, the idea of sparse grids can be traced back to Smoljak’s construction of multivariate quadrature formulas using combinations of tensor products of suitable one-dimensional formulas (cf. [39, 19]). More recently, the systematic and thorough studies on the method can be found in [46, 22, 23, 19]. In addition, several sparse grid methods are devised in [44, 21] for solving the RTE through conforming spatial discretization. However, according to the computational experience, it is preferable to use the discontinuous Galerkin (DG) method for spatial discretization for hyperbolic problems (cf. [9, 14, 10]), in order to capture non-smooth physical solutions. In [43], the sparse grid technique combined with the DG method has been developed for elliptic equations. This method is also applied to transport equations in [24, 25], but the scattering effect is not considered. The adaptive analogues of their methods are also given in [25, 41]. In this paper, we are intended to propose and analyze a sparse grid DG method to solve the RTE, following the ideas in [27] and [24]. Unlike the studies in [44, 21], the DG method will be used to carry out the spatial discretization. And different from [24], we will discuss in detail the efficient solution of the 5-dimensional RTE with scattering effect. Concretely speaking, the discrete ordinate technique is first applied to discretize the scattering term, an integral with respect to the angular variables, by simply picking several directions spanning the solid angle, resulting in a semi-discrete coupled hyperbolic system. In view of the hyperbolic nature of the semi- discrete system, the DG method is further employed for spatial discretization, yielding a fully discrete method. To overcome the curse of dimension, the sparse DG space is constructed by using the techniques in wavelet analysis to replace the original piecewise polynomial approximation space. We achieve the complexity analysis and error analysis of the method using some arguments in [27] and [24], which show the new approach can greatly reduce the spatial degrees of freedom while keeping almost the same accuracy up to multiplication of an $\log$ factor. For the resulting linear system, considering its block structure, we solve it using the block Gauss-Seidal iteration method. A series of numerical examples are reported to validate the accuracy and performance of the proposed method. Furthermore, we also extend the method to solve the RTE efficiently for some non-tensor product spatial domains in two dimensions. We end this section by introducing some notations and symbols frequently used in this paper. For a bounded Lipschitz domain $D$, the symbol $(\cdot,\cdot)_{D}$ denotes the $L^{2}$-inner product on $D$, $\|\cdot\|_{0,D}$ denotes the $L^{2}$-norm, and $|\cdot|_{s,D}$ is the $H^{s}(D)$-seminorm. For all integer $k\geq 0$, $\mathbb{P}_{k}(D)$ is the set of polynomials of degree $\leq k$ on $D$. The jumps and averages for scalar and vector-valued functions ($v,\boldsymbol{\tau}$, respectively) on an edge $e$ common to two elements $K_{1},K_{2}$ are defined by $[\\![v]\\!]=v_{1}\boldsymbol{n}_{1}+v_{2}\boldsymbol{n}_{2},\quad\\{\\!\\!\\{v\\}\\!\\!\\}=\frac{v_{1}+v_{2}}{2},$ $[\\![\boldsymbol{\tau}]\\!]=\boldsymbol{\tau}_{1}\cdot\boldsymbol{n}_{1}+\boldsymbol{\tau}_{2}\cdot\boldsymbol{n}_{2},\quad\\{\\!\\!\\{\boldsymbol{\tau}\\}\\!\\!\\}=\frac{\boldsymbol{\tau}_{1}+\boldsymbol{\tau}_{2}}{2},$ where $\boldsymbol{n}_{1},\boldsymbol{n}_{2}$ are the unit outward normals to $K_{1},K_{2}$, respectively. On a boundary edge or face, $[\\![v]\\!]=v\boldsymbol{n}$ and $\\{\\!\\!\\{\boldsymbol{\tau}\\}\\!\\!\\}=\boldsymbol{\tau}$. Moreover, for any two quantities $a$ and $b$, “$a\lesssim b$” indicates “$a\leq Cb$” with the hidden constant $C$ independent of the mesh size $h_{K}$, and “$a\eqsim b$” abbreviates “$a\lesssim b\lesssim a$”. ## 2 Radiative transfer equation The steady-state monoenergetic version of the radiative transfer equation is expressed as (cf. [27, 6]) $\boldsymbol{\omega}\cdot\nabla u(\boldsymbol{x},\boldsymbol{\omega})+\sigma_{t}(\boldsymbol{x})u(\boldsymbol{x},\boldsymbol{\omega})=\sigma_{s}(\boldsymbol{x})(Su)(\boldsymbol{x},\boldsymbol{\omega})+f(\boldsymbol{x},\boldsymbol{\omega}),\quad\boldsymbol{x}\in D,\boldsymbol{\omega}\in S^{2}.$ (2.1) Here, $D$ is a domain in $\mathbb{R}^{3}$ and $S^{2}$ denotes the unit sphere in $\mathbb{R}^{3}$, $u(\boldsymbol{x},\boldsymbol{\omega})$ is a function of three space variables $\boldsymbol{x}$ and two angular variables $\boldsymbol{\omega}$, $\sigma_{t}=\sigma_{a}+\sigma_{s}$ with $\sigma_{a}$ being the macroscopic absorption cross section, and $\sigma_{s}$ the macroscopic scattering cross section, and $f$ is a source function in $D$. We impose an inflow boundary value condition $u(\boldsymbol{x},\boldsymbol{\omega})=\alpha(\boldsymbol{x},\boldsymbol{\omega}),\quad(\boldsymbol{x},\boldsymbol{\omega})\in{\Gamma_{-}},$ (2.2) where $\Gamma_{-}$ is defined by $\Gamma_{-}=\\{(\boldsymbol{x},\boldsymbol{\omega}):\boldsymbol{x}\in\partial D,~{}\boldsymbol{\omega}\in S^{2},\quad\boldsymbol{n}(\boldsymbol{x})\cdot\boldsymbol{\omega}<0\\}.$ (2.3) The symbol $S$ on the right-hand side of (2.1) is an integral operator defined by $(Su)(\boldsymbol{x},\boldsymbol{\omega})=\int_{S^{2}}g(\boldsymbol{x},\boldsymbol{\omega}\cdot\hat{\boldsymbol{\omega}})u(\boldsymbol{x},\hat{\boldsymbol{\omega}}){\rm d}\sigma(\hat{\boldsymbol{\omega}})$ (2.4) with $g$ being a nonnegative normalized phase function $\int_{S^{2}}g(\boldsymbol{x},\boldsymbol{\omega}\cdot\hat{\boldsymbol{\omega}}){\rm d}\sigma(\hat{\boldsymbol{\omega}})=1,\quad\boldsymbol{x}\in D,\boldsymbol{\omega}\in S^{2}.$ In most applications, the function $g$ is assumed to be independent of $\boldsymbol{x}$. One well-known example considered in this paper is the Henyey-Greenstein phase function $g(t)=\frac{1-\eta^{2}}{4\pi(1+\eta^{2}-2\eta t)^{3/2}},~{}~{}t\in[-1,1],$ (2.5) where the parameter $\eta\in(-1,1)$ is the anisotropy factor for the scattering medium which measures the strength of forward peakedness of the phase function. Note that $\eta=0$ for isotropic scattering, $\eta>0$ for forward scattering, and $\eta<0$ for backward scattering. We assume that * 1. $\sigma_{t},\sigma_{s}\in L^{\infty}(D)$, $\sigma_{s}\geq 0$ a.e. in $D$, $\sigma_{a}=\sigma_{t}-\sigma_{s}\geq c_{0}$ in $D$ for a constant $c_{0}>0$. * 2. $f(\boldsymbol{x},\boldsymbol{\omega})\in L^{2}(D\times S^{2})$ and is a continuous function with respect to $\boldsymbol{\omega}\in S^{2}$. Under these assumptions, the problem (2.1)-(2.2) has a unique solution $u\in H_{2}^{1}(D\times S^{2})$ (cf. [27]), where $H_{2}^{1}(D\times S^{2}):=\\{v\in L^{2}(D\times S^{2}):\boldsymbol{\omega}\cdot\nabla v\in L^{2}(D\times S^{2})\\}.$ ## 3 The sparse grid discrete-ordinate DG method for the RTE In this section, we first recall the construction of sparse discontinuous finite element spaces; One can refer to [43, 3, 4] and the references therein for details. Then, we will present in detail the sparse grid discrete-ordinate DG method for the RTE. ### 3.1 Construction of sparse DG spaces Let $\Omega=[0,1]$ and partition it into $2^{n}$ cells with uniform cell size $h=2^{-n}$. The resulting $n$-th level grid is denoted by ${\Omega_{n}}$ and the $j$-th cell is given by $I_{n}^{j}=(2^{-n}j,2^{-n}(j+1)],\quad j=0,1,\cdots,2^{n}-1.$ We define $V_{n}^{k}=\\{v:v|_{I_{n}^{j}}\in\mathbb{P}_{k}(I_{n}^{j}),\quad j=0,1,\cdots,2^{n}-1\\}$ to be the piecewise polynomial space on ${\Omega_{n}}$. One can check that there exists the nested structure for different values of $n$: $V_{0}^{k}\subset V_{1}^{k}\subset\cdots\subset V_{n}^{k}\subset\cdots$. Denote $W_{n}^{k}$ to be the orthogonal complement of $V_{n-1}^{k}$ in $V_{n}^{k}$ with respect to the $L^{2}(\Omega)$ inner product, i.e., $V_{n-1}^{k}\oplus W_{n}^{k}=V_{n}^{k},\quad W_{n}^{k}\bot V_{n-1}^{k},\quad n\geq 1,$ where for simplicity set $W_{0}^{k}=V_{0}^{k}$. We then obtain an orthogonal decomposition of the DG space $V_{N}^{k}=\mathop{\bigoplus}\limits_{0\leq n\leq N}W_{n}^{k}.$ We proceed to review the construction in multi-dimensions. For $\Omega=[0,1]^{d}$, let $h_{m}=2^{-n_{m}}$ be the step size along $x_{m}$-direction. For simplicity, we use the notations of multi-indices in the following. Let $\boldsymbol{n}=(n_{1},n_{2},\cdots,n_{d})$. Then the cell size can be denoted by $h_{\boldsymbol{n}}=(2^{-n_{1}},2^{-n_{2}},\cdots,2^{-n_{d}})=2^{-\boldsymbol{n}}$ and the associated grid is written by $\Omega_{\boldsymbol{n}}$ whose $\boldsymbol{j}$-th cell is given by $I_{\boldsymbol{n}}^{\boldsymbol{j}}=I_{n_{1}}^{j_{1}}\times I_{n_{2}}^{j_{2}}\times\cdots\times I_{n_{d}}^{j_{d}},\quad\boldsymbol{j}=(j_{1},j_{2},\cdots,j_{d}),$ where $I_{n_{m}}^{j_{m}}=(2^{-n_{m}}j_{m},2^{-n_{m}}(j_{m}+1)],\quad j_{m}=0,1,\cdots,2^{n_{m}}-1$ is the element along $x_{m}$-axis. With multi-indices notation we have $\boldsymbol{0}\leq\boldsymbol{j}\leq 2^{\boldsymbol{n}}-\boldsymbol{1}$. Introduce a tensor-product piecewise polynomial space as $\boldsymbol{V}_{\boldsymbol{n}}^{k}=\\{\boldsymbol{v}:\boldsymbol{v}(\boldsymbol{x})\in\mathbb{Q}_{k}(I_{\boldsymbol{n}}^{\boldsymbol{j}}),\quad\boldsymbol{0}\leq\boldsymbol{j}\leq 2^{\boldsymbol{n}}-\boldsymbol{1}\\},$ where $\mathbb{Q}_{k}(I_{\boldsymbol{n}}^{\boldsymbol{j}})$ consists of polynomials of degree up to $k$ in each dimension on cell $I_{\boldsymbol{n}}^{\boldsymbol{j}}$. If we use an equal refinement of size $h:=h_{N}=2^{-N}$ in each coordinate direction, the grid and space will be denoted by $\Omega_{N}$ and $\boldsymbol{V}_{N}^{k}$, respectively. With the usual convention, we also use $\mathcal{T}_{h}$ and $V_{h}^{k}$ instead. It is obvious that $\boldsymbol{V}_{\boldsymbol{n}}^{k}=V_{n_{1}}^{k}\times V_{n_{2}}^{k}\times\cdots\times V_{n_{d}}^{k}.$ We similarly define the tensor-product multiwavelet space as $\boldsymbol{W}_{\boldsymbol{n}}^{k}=W_{n_{1}}^{k}\times W_{n_{2}}^{k}\times\cdots\times W_{n_{d}}^{k}.$ Observing the fact that $V_{n_{m}}^{k}=\mathop{\bigoplus}\limits_{0\leq j_{m}\leq n_{m}}W_{j_{m}}^{k},$ we have the following expansion $\boldsymbol{V}_{\boldsymbol{n}}^{k}=\mathop{\bigoplus}\limits_{\boldsymbol{0}\leq\boldsymbol{j}\leq\boldsymbol{n}}\boldsymbol{W}_{\boldsymbol{j}}^{k},\quad~{}\boldsymbol{V}_{N}^{k}=\mathop{\bigoplus}\limits_{|\boldsymbol{j}|_{\infty}\leq N}\boldsymbol{W}_{\boldsymbol{j}}^{k}.$ The sparse finite element approximation space on $\Omega_{N}$ is defined by the following truncated space $\widehat{\boldsymbol{V}}_{N}^{k}:=\mathop{\bigoplus}\limits_{|\boldsymbol{n}|_{1}\leq N}\boldsymbol{W}_{\boldsymbol{n}}^{k},\quad|\boldsymbol{n}|_{1}=n_{1}+n_{2}+\cdots+n_{d}.$ The number of degrees of freedom of sparse DG space is $\mathcal{O}(h^{-1}|\log_{2}h|^{d-1})$ with $h=2^{-N}$, which is significantly less than that of DG space with exponential dependence on $d$. ### 3.2 The sparse grid discrete-ordinate DG method For any continuous function $F(\boldsymbol{\omega})$ defined on the unit sphere $S^{2}$, we write the numerical quadrature to be used in the form $\int_{S^{2}}F(\boldsymbol{\omega}){\rm d}\sigma(\boldsymbol{\omega})\approx\sum\limits_{l=1}^{L}{w_{l}F(\boldsymbol{\omega}^{l})},\quad\boldsymbol{\omega}^{l}\in S^{2},~{}1\leq l\leq L.$ (3.1) The integral operator $S$ is then approximated by $(Su)(\boldsymbol{x},\boldsymbol{\omega})\approx(S_{d}u)(\boldsymbol{x},\boldsymbol{\omega}):=\sum\limits_{l=1}^{L}w_{l}g(\boldsymbol{x},\boldsymbol{\omega}\cdot\boldsymbol{\omega}^{l})u(\boldsymbol{x},\boldsymbol{\omega}^{l}).$ (3.2) Regarding the accuracy of the quadrature (3.1), we will write $n$ for the algebraic precision, i.e., the quadrature integrates exactly all spherical polynomials of total degree no more than $n$ and does not integrate exactly some spherical polynomial of total degree $n+1$. Then we have the following estimate (cf. [27]) $\Big{|}\int_{S^{2}}F(\boldsymbol{\omega}){\rm d}\sigma(\boldsymbol{\omega})-\sum\limits_{l=1}^{L}w_{l}F(\boldsymbol{\omega}^{l})\Big{|}\leq c_{s}n^{-s}\|F\|_{s,S^{2}},\quad F\in H^{s}({S^{2}}),\quad s>1,$ (3.3) where $c_{s}$ is a universal constant depending only on $s$. Associated with the numerical quadrature, we further define $m(\boldsymbol{x})=\mathop{\max}\limits_{1\leq i\leq L}\sum\limits_{l=1}^{L}w_{l}g(\boldsymbol{x},\boldsymbol{\omega}^{l}\cdot\boldsymbol{\omega}^{i})$ (3.4) and make the following assumption (cf. [27]): $\sigma_{t}-m\sigma_{s}\geq c_{0}^{\prime}~{}\mbox{in $D$ for some constant $c_{0}^{\prime}>0$}.$ (3.5) Using the quadrature (3.2), we can discretize (2.1) in angular direction to get $\boldsymbol{\omega}^{l}\cdot\nabla u^{l}+\sigma_{t}u^{l}=\sigma_{s}(\boldsymbol{x})\sum\limits_{i=1}^{L}w_{i}g(\boldsymbol{x},\boldsymbol{\omega}^{l}\cdot\boldsymbol{\omega}^{i})u^{i}+f^{l},\quad 1\leq l\leq L$ (3.6) with boundary value condition $u^{l}(\boldsymbol{x})=\alpha^{l}(\boldsymbol{x}),\quad(\boldsymbol{x},\boldsymbol{\omega}^{l})\in\Gamma_{-},~{}~{}1\leq l\leq L,$ (3.7) where $u^{l}=u^{l}(\boldsymbol{x})$ is the approximation to $u(\boldsymbol{x},\boldsymbol{\omega}^{l})$. The system (3.6) is a first-order hyperbolic problem in space, which will be further discretized by DG method. Let $\\{\mathcal{T}_{h}\\}_{h>0}$ be a regular family of triangulations of $D$. Assume that $\mathcal{E}_{h}$ consists of the set of all edges ($d=2$) or faces ($d=3$) in $\mathcal{T}_{h}$ and $\mathcal{E}_{h}^{0}$ the set of all interior edges or faces. By a direct manipulation, we obtain the following identity (cf. [10]): ###### Lemma 3.1. For $(\varphi,\boldsymbol{\tau})\in H^{s}(\mathcal{T}_{h})\times[H^{s}(\mathcal{T}_{h})]^{d}$, $s>1/2$, there holds $\sum\limits_{K\in\mathcal{T}_{h}}\int_{\partial K}\varphi\boldsymbol{\tau}\cdot\boldsymbol{n}{\rm d}s=\sum\limits_{e\in\mathcal{E}_{h}}\int_{e}\\{\\!\\!\\{\boldsymbol{\tau}\\}\\!\\!\\}\cdot[\\![\varphi]\\!]{\rm d}s+\sum\limits_{e\in\mathcal{E}_{h}^{0}}\int_{e}\\{\\!\\!\\{\varphi\\}\\!\\!\\}\cdot[\\![\boldsymbol{\tau}]\\!]{\rm d}s.$ (3.8) Further, if $u\in H^{s}(\omega_{e})$ and $s>1/2$, then we have the following weak continuity $\int_{e}[\\![u]\\!]v{\rm d}s=0,\quad v\in L^{2}(e),\quad e\in\mathcal{E}_{h}^{0},$ where $\omega_{e}$ is the set of elements sharing $e$ as an edge ($d=2$) or faces ($d=3$). We define a discontinuous finite element space by $V_{h}=\Big{\\{}v\in L^{2}(D):v|_{K}\in\mathbb{P}_{k}(K),~{}~{}K\in\mathcal{T}_{h}\Big{\\}},$ (3.9) where $\mathbb{P}_{k}(K)$ denotes the set of all polynomials on $K$ with degree $\leq k$. Multiplying (3.6) by any $v_{h}\in V_{h}$, we obtain from the integration by parts that $\displaystyle\sum\limits_{K\in\mathcal{T}_{h}}\Big{[}\int_{K}(-u^{l}(\boldsymbol{\omega}^{l}\cdot\nabla v_{h})+\sigma_{t}u^{l}v_{h}){\rm d}x+\int_{\partial K}(\boldsymbol{\omega}^{l}\cdot\boldsymbol{n})u^{l}v_{h}{\rm d}s\Big{]}$ $\displaystyle\quad\quad=\int_{D}\sigma_{s}\sum\limits_{i=1}^{L}w_{i}g(\cdot,\boldsymbol{\omega}^{l}\cdot\boldsymbol{\omega}^{i})u^{i}v_{h}{\rm d}x+\int_{D}f^{l}v_{h}{\rm d}x,~{}~{}1\leq l\leq L.$ Taking $\boldsymbol{\tau}=\boldsymbol{\omega}^{l}u^{l}$ and $\varphi=v_{h}$ in (3.8), we immediately obtain the following system $a_{h}^{(l)}(u^{l},v_{h})+b_{h}^{(l)}(u^{l},v_{h})=(f^{l},v_{h})+\langle\alpha^{l},v_{h}\rangle^{(l)},\quad v_{h}\in V_{h},$ where $\displaystyle a_{h}^{(l)}(u^{l},v_{h})$ $\displaystyle=\sum\limits_{K\in\mathcal{T}_{h}}\int_{K}(-u^{l}(\boldsymbol{\omega}^{l}\cdot\nabla v_{h})+\sigma_{t}u^{l}v_{h}){\rm d}x$ $\displaystyle\quad-\int_{D}\sigma_{s}\sum\limits_{i=1}^{L}w_{i}g(\cdot,\boldsymbol{\omega}^{l}\cdot\boldsymbol{\omega}^{i})u^{i}v_{h}{\rm d}x,$ (3.10) $b_{h}^{(l)}(u^{l},v_{h})=\sum\limits_{e\not\subset\Gamma_{-}}\int_{e}\\{\\!\\!\\{\boldsymbol{\omega}^{l}u^{l}\\}\\!\\!\\}\cdot[\\![v_{h}]\\!]{\rm d}s,$ (3.11) $(f^{l},v_{h})=\int_{D}f^{l}v_{h}{\rm d}x,\quad\langle\alpha^{l},v_{h}\rangle^{(l)}=-\sum\limits_{e\subset\Gamma_{-}}\int_{e}\boldsymbol{\omega}^{l}\cdot\boldsymbol{n}\alpha^{l}v_{h}{\rm d}s.$ (3.12) Define $\boldsymbol{V}_{h}=(V_{h})^{L}$ and write a generic element as $\boldsymbol{v}_{h}:=\\{v_{h}^{l}\\}_{l=1}^{L}$. The global formulation can be expressed as $\sum\limits_{l=1}^{L}w_{l}(a_{h}^{(l)}(u^{l},v_{h}^{l})+b_{h}^{(l)}(u^{l},v_{h}^{l}))=\sum\limits_{l=1}^{L}w_{l}((f^{l},v_{h}^{l})+\langle\alpha^{l},v_{h}^{l}\rangle^{(l)}).$ Then the discrete-ordinate DG method is: Find $\boldsymbol{u}_{h}:=\\{u_{h}^{l}\\}\in\boldsymbol{V}_{h}$ such that $a_{h}(\boldsymbol{u}_{h},\boldsymbol{v}_{h})=F(\boldsymbol{v}_{h}),\quad\boldsymbol{v}_{h}\in\boldsymbol{V}_{h},$ (3.13) where $a_{h}(\boldsymbol{u}_{h},\boldsymbol{v}_{h})=\sum\limits_{l=1}^{L}w_{l}(a_{h}^{(l)}(u_{h}^{l},v_{h}^{l})+b_{h}^{(l)}(u_{h}^{l},v_{h}^{l})),$ $F(\boldsymbol{v}_{h})=\sum\limits_{l=1}^{L}w_{l}((f^{l},v_{h}^{l})+\langle\alpha^{l},v_{h}^{l}\rangle^{(l)}).$ It is preferable to add some stabilization terms in the DG scheme to penalize the jump of the solution across interior edges or faces of the triangulation. One approach introduced in [10] is to replace the average $\\{\\!\\!\\{\boldsymbol{\omega}^{l}u^{l}\\}\\!\\!\\}$ in (3.11) by $\\{\\!\\!\\{\boldsymbol{\omega}^{l}u_{h}^{l}\\}\\!\\!\\}+c_{e}^{l}[\\![u_{h}^{l}]\\!]$, where $c_{e}^{l}$ is a nonnegative function over $e$ satisfying $c_{e}^{l}=\theta_{0}|\boldsymbol{\omega}^{l}\cdot\boldsymbol{n}|$ with $\theta_{0}$ a constant independent of $e$ and $h$. The stabilized discrete- ordinate DG method is to find $\boldsymbol{u}_{h}:=\\{u_{h}^{l}\\}\in\boldsymbol{V}_{h}$ such that $a_{h}^{s}(\boldsymbol{u}_{h},\boldsymbol{v}_{h})=F(\boldsymbol{v}_{h}),\quad\boldsymbol{v}_{h}\in\boldsymbol{V}_{h},$ (3.14) where $a_{h}^{s}(\boldsymbol{u}_{h},\boldsymbol{v}_{h})=\sum\limits_{l=1}^{L}w_{l}(a_{h}^{(l)}(u_{h}^{l},v_{h}^{l})+b_{hs}^{(l)}(u_{h}^{l},v_{h}^{l}))$ (3.15) and $\displaystyle b_{hs}^{(l)}(u_{h}^{l},v_{h}^{l})$ $\displaystyle=b_{h}^{(l)}(u_{h}^{l},v_{h}^{l})+\sum\limits_{e\in\mathcal{E}_{h}^{0}}\int_{e}c_{e}^{l}[\\![u_{h}^{l}]\\!]\cdot[\\![v_{h}^{l}]\\!]{\rm d}s$ $\displaystyle=\sum\limits_{e\not\subset\Gamma_{-}}\int_{e}\\{\\!\\!\\{\boldsymbol{\omega}^{l}u_{h}^{l}\\}\\!\\!\\}\cdot[\\![v_{h}^{l}]\\!]{\rm d}s+\sum\limits_{e\in\mathcal{E}_{h}^{0}}\int_{e}c_{e}^{l}[\\![u_{h}^{l}]\\!]\cdot[\\![v_{h}^{l}]\\!]{\rm d}s.$ (3.16) ###### Remark 3.1. The sparse grid discrete-ordinate DG method is obtained by replacing the DG space $V_{h}$ in (3.9) with the sparse DG space $\widehat{V}_{h}^{k}:=\widehat{\boldsymbol{V}}_{N}^{k}\subset V_{h}$. ## 4 Error analysis ### 4.1 Error estimate of the sparse projection operator We define the broken $H^{s}$ Sobolev norm on $\Omega_{N}$ by $\|v\|_{H^{s}(\Omega_{N})}^{2}=\sum\limits_{\boldsymbol{0}\leq\boldsymbol{j}\leq 2^{\boldsymbol{N}-\boldsymbol{1}}-\boldsymbol{1}}\|v\|_{H^{s}(I_{\boldsymbol{N}}^{\boldsymbol{j}})}^{2}.$ For any nonnegative integer $m$ and the multi-index $\alpha=\\{i_{1},i_{2},\cdots,i_{r}\\}\subset\\{1,2,\cdots,d\\}$, define $|v|_{H^{m,\alpha}(\Omega)}=\Big{\|}\Big{(}\frac{\partial^{m}}{\partial x_{i_{1}}^{m}}\cdots\frac{\partial^{m}}{\partial x_{i_{r}}^{m}}\Big{)}v\Big{\|}_{L^{2}(\Omega)}$ and $|v|_{\mathcal{H}^{q+1}(\Omega)}=\mathop{\max}\limits_{1\leq r\leq d}\Big{(}\mathop{\max}\limits_{\alpha\in\\{1,2,\cdots,d\\},|\alpha|=r}|v|_{H^{q+1,\alpha}(\Omega)}\Big{)},$ which is the norm for the mixed derivative of $v$ of at most degree $q+1$ in each direction. In the following, we denote by $\boldsymbol{P}$ the sparse projection operator to be the $L^{2}$ projection onto $\widehat{\boldsymbol{V}}_{N}^{k}$. ###### Lemma 4.1. Let $\boldsymbol{P}$ be the sparse projector, $k\geq 1$, $N\geq 1$ and $d\geq 2$. Then for $v\in\mathcal{H}^{p+1}(\Omega)$ there hold $|\boldsymbol{P}v-v|_{L^{2}(\Omega_{N})}\lesssim|\log_{2}h|^{d}h^{k+1}|v|_{\mathcal{H}^{k+1}(\Omega)},$ $|\boldsymbol{P}v-v|_{H^{1}(\Omega_{N})}\lesssim h^{k}|v|_{\mathcal{H}^{k+1}(\Omega)},$ and $\Big{(}\sum\limits_{K\in\mathcal{T}_{h}}\|\boldsymbol{P}v-v\|_{0,\partial K}^{2}\Big{)}^{1/2}\lesssim|\log_{2}h|^{d}h^{k+1/2}|v|_{\mathcal{H}^{k+1}(\Omega)}.$ ###### Proof. It follows from [35, 43, 24] that for any $v\in\mathcal{H}^{p+1}(\Omega)$ and $1\leq q\leq\min\\{p,k\\}$, there holds $|\boldsymbol{P}v-v|_{H^{s}(\Omega_{N})}\lesssim\begin{cases}N^{d}2^{-N(q+1)}|v|_{\mathcal{H}^{q+1}(\Omega)},\quad&s=0,\\\ 2^{-Nq}|v|_{\mathcal{H}^{q+1}(\Omega)},\quad&s=1.\end{cases}$ Noting that $h=h_{N}=2^{-N}$, we have $|\boldsymbol{P}v-v|_{L^{2}(\Omega_{N})}\lesssim|\log_{2}h|^{d}h^{k+1}|v|_{\mathcal{H}^{k+1}(\Omega)}$ and $|\boldsymbol{P}v-v|_{H^{1}(\Omega_{N})}\lesssim h^{k}|v|_{\mathcal{H}^{k+1}(\Omega)}.$ Recalling the trace inequality (cf. [8]) $\|\phi\|_{0,\partial K}^{2}\lesssim h_{K}^{-1}\|\phi\|_{0,K}^{2}+h_{K}|\phi|_{1,K}^{2},$ (4.1) where $K\in\mathcal{T}_{h}$ with diameter $h_{K}$, we then have $\displaystyle\Big{(}\sum\limits_{K\in\mathcal{T}_{h}}\|\boldsymbol{P}v-v\|_{0,\partial K}^{2}\Big{)}^{1/2}$ $\displaystyle\lesssim\Big{(}\sum\limits_{K\in\mathcal{T}_{h}}h_{K}^{-1}\|\boldsymbol{P}v-v\|_{0,K}^{2}+h_{K}|\boldsymbol{P}v-v|_{1,K}^{2}\Big{)}^{1/2}$ $\displaystyle\lesssim(|\log_{2}h|^{2d}+1)^{1/2}h^{k+1/2}|v|_{\mathcal{H}^{k+1}(\Omega)}$ $\displaystyle\lesssim|\log_{2}h|^{d}h^{k+1/2}|v|_{\mathcal{H}^{k+1}(\Omega)}.$ This completes the proof. ∎ ### 4.2 Error analysis of the sparse grid discrete-ordinate DG method Using the similar arguments in [10, 27], one can deduce the following stability result whose proof is omitted for simplicity. ###### Lemma 4.2. Let $|\\!|\\!|\boldsymbol{v}_{h}|\\!|\\!|=\Big{[}\sum\limits_{l=1}^{L}w_{l}\Big{(}\sum\limits_{e\in\mathcal{E}_{h}}\int_{e}c_{e}^{l}|[\\![v_{h}^{l}]\\!]|^{2}{\rm d}s+\int_{D}(v_{h}^{l})^{2}{\rm d}x\Big{)}\Big{]}^{1/2}.$ Under the assumption (3.5), there holds $a_{h}^{s}(\boldsymbol{v}_{h},\boldsymbol{v}_{h})\gtrsim|\\!|\\!|\boldsymbol{v}_{h}|\\!|\\!|^{2},\quad\boldsymbol{v}_{h}\in\boldsymbol{V}_{h}.$ We denote the solutions of the original problem (2.1), the semi-discrete problem (3.6) and the stabilized discrete-ordinate DG method (3.14) by $\\{u(\boldsymbol{x},\boldsymbol{\omega}^{l})\\}$, $\boldsymbol{u}=\\{u^{l}(\boldsymbol{x})\\}$ and $\boldsymbol{u}_{h}=\\{u_{h}^{l}\\}$, respectively. The error is decomposed as $\\{u(\boldsymbol{x},\boldsymbol{\omega}^{l})\\}-\boldsymbol{u}_{h}=(\\{u(\boldsymbol{x},\boldsymbol{\omega}^{l})\\}-\\{u^{l}(\boldsymbol{x})\\})+(\\{u^{l}(\boldsymbol{x})\\}-\\{u_{h}^{l}(\boldsymbol{x})\\}),$ (4.2) and measured by $\|u-u_{h}\|_{h}=\Big{(}\sum\limits_{l=1}^{L}w_{l}\|u(\cdot,\boldsymbol{\omega}^{l})-u_{h}^{l}\|_{0,D}^{2}\Big{)}^{1/2}.$ (4.3) ###### Theorem 4.1. Let $n$ be the degree of precision of the numerical quadrature and $d\geq 2$. Then under the assumption (3.5), for the sparse grid discrete-ordinate DG method, we have $\displaystyle\|u-u_{h}\|_{h}$ $\displaystyle\lesssim c(\theta_{0})|\log_{2}h|^{d}h^{k+1/2}\Big{(}\sum\limits_{l=1}^{L}w_{l}|u^{l}|_{\mathcal{H}^{k+1}(\Omega)}^{2}\Big{)}^{1/2}$ $\displaystyle\quad+c(r^{\prime},g)n^{-r^{\prime}}\Big{(}\int_{D}\|u(\boldsymbol{x},\cdot)\|_{r^{\prime},S^{2}}^{2}{\rm d}x\Big{)}^{1/2},$ where, $h=2^{-N}$, $u^{l}$ is the solution of (3.6), $c(\theta_{0})=\theta_{0}^{-1/2}+\theta_{0}^{1/2}$ and $c(r^{\prime},g)$ is defined in (4.4). ###### Proof. For the first part in (4.2), let $\varepsilon^{l}(\boldsymbol{x}):=u(\boldsymbol{x},\boldsymbol{\omega}^{l})-u^{l}(\boldsymbol{x}),\quad 1\leq l\leq L.$ In view of the equation (4.19) in [27], one has $\sum\limits_{l=1}^{L}w_{l}\int_{D}(\varepsilon^{l})^{2}{\rm d}x\lesssim c(r^{\prime},g)^{2}n^{-2r^{\prime}}\int_{D}\|u(\boldsymbol{x},\cdot)\|_{r^{\prime},S^{2}}^{2}{\rm d}x,$ where $c(r^{\prime},g):=c(r^{\prime})\mathop{\sup}\limits_{\boldsymbol{x}\in D,\boldsymbol{\omega}\in S^{2}}\|g(\boldsymbol{x},\boldsymbol{\omega}\cdot)\|_{r^{\prime},S^{2}}$ (4.4) and $c(r^{\prime})$ is a positive constant depending only on $r^{\prime}$. For the second part, let $u^{l}-u_{h}^{l}=(u^{l}-P_{h}^{k}u^{l})-(u_{h}^{l}-P_{h}^{k}u^{l})=:\eta^{l}+\delta^{l},$ where $P_{h}^{k}$ is the $L^{2}$-projection onto the sparse DG space $\widehat{V}_{h}^{k}$ (cf. Remark 3.1). Similarly, we denote $\boldsymbol{\eta}=\\{\eta^{l}\\}$ and $\boldsymbol{\delta}=\\{\delta^{l}\\}$. When $\boldsymbol{u}_{h}$ is replaced by the exact solution $\boldsymbol{u}$ of the semi-discrete problem, the weak continuity in Lemma 3.1 yields $\int_{e}c_{e}^{l}[\\![u^{l}]\\!][\\![v_{h}^{l}]\\!]{\rm d}s=0,\quad e\in\mathcal{E}_{h}^{0}.$ From (3.2) we have $b_{hs}^{(l)}(u^{l},v_{h}^{l})=b_{h}^{(l)}(u^{l},v_{h}^{l})$, and hence $a_{h}^{s}(\boldsymbol{u},\boldsymbol{v}_{h})=a_{h}(\boldsymbol{u},\boldsymbol{v}_{h})$, which yields the following Galerkin orthogonality $a_{h}^{s}(\boldsymbol{u}-\boldsymbol{u}_{h},\boldsymbol{v}_{h})=a_{h}(\boldsymbol{u}-\boldsymbol{u}_{h},\boldsymbol{v}_{h})=0,\quad\boldsymbol{v}_{h}\in\boldsymbol{V}_{h}.$ (4.5) According to the stability estimate in Lemma 4.2, we have $|\\!|\\!|\boldsymbol{\delta}|\\!|\\!|^{2}\lesssim a_{h}^{s}(\boldsymbol{\delta},\boldsymbol{\delta})=a_{h}^{s}(\boldsymbol{u}_{h}-P_{h}^{k}\boldsymbol{u},\boldsymbol{\delta})=a_{h}^{s}(\boldsymbol{u}-P_{h}^{k}\boldsymbol{u},\boldsymbol{\delta})=a_{h}^{s}(\boldsymbol{\eta},\boldsymbol{\delta}).$ (4.6) We now estimate the right-hand side of (4.6). Let $a_{h}^{(l)}(u^{l},v_{h})={\rm I}_{1}^{l}(u^{l},v_{h})-{\rm I}_{2}^{l}(u^{l},v_{h}),$ where $\displaystyle{\rm I}_{1}^{l}(u^{l},v_{h})$ $\displaystyle=\sum\limits_{K\in\mathcal{T}_{h}}\int_{K}\left(-u^{l}(\boldsymbol{\omega}^{l}\cdot\nabla v_{h})+\sigma_{t}u^{l}v_{h}\right){\rm d}x,$ $\displaystyle{\rm I}_{2}^{l}(u^{l},v_{h})$ $\displaystyle=\int_{D}\sigma_{s}\sum\limits_{i=1}^{L}w_{i}g(\cdot,\boldsymbol{\omega}^{l}\cdot\boldsymbol{\omega}^{i})u^{i}v_{h}{\rm d}x.$ For the first term ${\rm I}_{1}^{l}$, noting that $\boldsymbol{\omega}^{l}\cdot\nabla\delta^{l}|_{K}\in\mathbb{Q}_{k}(K)$, by the definition of the projector $P_{h}^{k}$, $\int_{K}\eta^{l}(\boldsymbol{\omega}^{l}\cdot\nabla\delta^{l}){\rm d}x=\int_{K}(u^{l}-P_{h}^{k}u^{l})(\boldsymbol{\omega}^{l}\cdot\nabla\delta^{l}){\rm d}x=0,$ which gives $|{\rm I}_{1}^{l}(\eta^{l},\delta^{l})|=\Big{|}\sum\limits_{K\in\mathcal{T}_{h}}\int_{K}\left(-\eta^{l}(\boldsymbol{\omega}^{l}\cdot\nabla\delta^{l})+\sigma_{t}\eta^{l}\delta^{l}\right){\rm d}x\Big{|}\lesssim\sum\limits_{K\in\mathcal{T}_{h}}\|\eta^{l}\|_{0,K}\|\delta^{l}\|_{0,K}.$ The Cauchy-Schwarz inequality yields $\Big{|}\sum\limits_{l=1}^{L}w_{l}{\rm I}_{1}^{l}(\eta^{l},\delta^{l})\Big{|}=\Big{(}\sum\limits_{l=1}^{L}w_{l}\|\eta^{l}\|_{0,D}^{2}\Big{)}^{1/2}\Big{(}\sum\limits_{l=1}^{L}w_{l}\|\delta^{l}\|_{0,D}^{2}\Big{)}^{1/2}.$ For the second one, using Lemma 4.3 in [27], we obtain $\displaystyle\Big{|}\sum\limits_{l=1}^{L}{w_{l}{\rm I}_{2}^{l}(\eta^{l},\delta^{l})}\Big{|}$ $\displaystyle\leq\Big{(}\sum\limits_{l=1}^{L}w_{l}\int_{D}m\sigma_{s}(\eta^{l})^{2}{\rm d}x\Big{)}^{1/2}\Big{(}\sum\limits_{l=1}^{L}w_{l}\int_{D}m\sigma_{s}(\delta^{l})^{2}{\rm d}x\Big{)}^{1/2}$ $\displaystyle\lesssim\Big{(}\sum\limits_{l=1}^{L}w_{l}\|\eta^{l}\|_{0,D}^{2}\Big{)}^{1/2}\Big{(}\sum\limits_{l=1}^{L}w_{l}\|\delta^{l}\|_{0,D}^{2}\Big{)}^{1/2},$ where $m=m(\boldsymbol{x})$ is given in (3.4). It remains to consider $b_{hs}^{(l)}(\eta^{l},\delta^{l})$. From (49) in [10] we have $|b_{hs}^{(l)}(\eta^{l},\delta^{l})|\leq\sum\limits_{e\in\mathcal{E}_{h}}\Big{(}\frac{1}{\theta_{0}}\|c_{e}^{1/2}\\{\\!\\!\\{\eta^{l}\\}\\!\\!\\}\|_{0,e}+\|c_{e}^{1/2}[\\![\eta^{l}]\\!]\|_{0,e}\Big{)}\|c_{e}^{1/2}[\\![\delta^{l}]\\!]\|_{0,e}.$ According to the choice of $c_{e}^{l}$, the Cauchy-Schwarz inequality gives $\Big{|}\sum\limits_{l=1}^{L}{w_{l}b_{hs}^{(l)}(\eta^{l},\delta^{l})}\Big{|}\lesssim c(\theta_{0})\Big{(}\sum\limits_{l=1}^{L}w_{l}\sum\limits_{K\in\mathcal{T}_{h}}\|\eta^{l}\|_{0,\partial K}^{2}\Big{)}^{1/2}\Big{(}\sum\limits_{l=1}^{L}w_{l}\sum\limits_{e\in\mathcal{E}_{h}}\|c_{e}^{1/2}[\\![\delta^{l}]\\!]\|_{0,e}^{2}\Big{)}^{1/2},$ where $c(\theta_{0})=\theta_{0}^{-1/2}+\theta_{0}^{1/2}$ is a constant, and hence $a_{h}^{s}(\boldsymbol{\eta},\boldsymbol{\delta})\lesssim c(\theta_{0})\Big{[}\sum\limits_{l=1}^{L}w_{l}\Big{(}\|\eta^{l}\|_{0,D}^{2}+\sum\limits_{K\in\mathcal{T}_{h}}\|\eta^{l}\|_{0,\partial K}^{2}\Big{)}\Big{]}^{1/2}|\\!|\\!|\boldsymbol{\delta}|\\!|\\!|.$ This combined with (4.6) yields $|\\!|\\!|\boldsymbol{\delta}|\\!|\\!|\lesssim c(\theta_{0})\Big{[}\sum\limits_{l=1}^{L}w_{l}\Big{(}\|\eta^{l}\|_{0,D}^{2}+\sum\limits_{K\in\mathcal{T}_{h}}\|\eta^{l}\|_{0,\partial K}^{2}\Big{)}\Big{]}^{1/2},$ and with the error estimates of the sparse projection in Lemma 4.1 leads to the desired result. ∎ ## 5 Numerical results In this section, we shall provide a series of numerical examples for solving the RTE (2.1)-(2.2) to illustrate the performance of the proposed sparse grid discrete coordinate DG method. ### 5.1 The linear system from the discrete problem The $S_{n}$ method has $n(n+2)$ directions with $n$ an even natural number. The discrete-ordinate sets satisfying the required moment equations to fourteen digits of accuracy have been given in [7]. Note that only the ordinates in the first octant are given there. The remaining ordinates can be obtained by using symmetry arguments. For example, $S_{2}$ data is given in Tab. 1. Tab. 1: Discrete-ordinate sets for $S_{2}$ ($\boldsymbol{\omega}=(s_{1},s_{2},s_{3})$) $s_{1}$ | $s_{2}$ | $s_{3}$ | $w$ ---|---|---|--- $\pm 0.5773502691896257$ | $\pm 0.5773502691896257$ | $\pm 0.5773502691896257$ | 1.5707963267948966 Relabel the sparse bases by a single index $i=1,2,\cdots,M$ and denote them by $\varphi_{i}$, where $M=\dim\widehat{\boldsymbol{V}}_{N}^{k}$. The variational problem (3.14) can be written in matrix form $\boldsymbol{A}^{(l)}\hat{\boldsymbol{U}}^{l}-\sum\limits_{i=1}^{L}\boldsymbol{B}_{i}^{(l)}\hat{\boldsymbol{U}}^{i}+\boldsymbol{C}^{(l)}\hat{\boldsymbol{U}}^{l}=\boldsymbol{F}^{(l)},\quad 1\leq l\leq L,$ (5.1) where $\boldsymbol{A}^{(l)}=(a_{nm}^{(l)}),\quad\boldsymbol{B}_{i}^{(l)}=(b_{nmi}^{(l)}),\quad\boldsymbol{C}^{(l)}=(c_{nm}^{(l)}),$ $\hat{\boldsymbol{U}}^{l}=[\hat{u}_{1}^{l},\hat{u}_{2}^{l},\cdots,\hat{u}_{M}^{l}]^{T},\quad\boldsymbol{F}^{(l)}=[F_{1}^{(l)},F_{2}^{(l)},\cdots,F_{M}^{(l)}]^{T},$ and $a_{nm}^{(l)}=\sum\limits_{K\in\mathcal{T}_{h}}\int_{K}(-\varphi_{m}(\boldsymbol{\omega}^{l}\cdot\nabla\varphi_{n})+\sigma_{t}\varphi_{m}\varphi_{n}){\rm d}x,$ $b_{nmi}^{(l)}=w_{i}\int_{D}\sigma_{s}g(\cdot,\boldsymbol{\omega}^{l}\cdot{\boldsymbol{\omega}}^{i}){\varphi_{m}}\varphi_{n}{\rm d}x,$ $c_{nm}^{(l)}=\sum\limits_{e\not\subset\Gamma_{-}}\int_{e}\\{\\!\\!\\{\boldsymbol{\omega}^{l}\varphi_{m}\\}\\!\\!\\}\cdot[\\![\varphi_{n}]\\!]{\rm d}s+\sum\limits_{e\in\mathcal{E}_{h}^{0}}\int_{e}c_{e}^{l}[\\![\varphi_{m}]\\!]\cdot[\\![\varphi_{n}]\\!]{\rm d}s,$ $F_{n}^{(l)}=\int_{D}f^{l}\varphi_{n}{\rm d}x-\sum\limits_{e\subset\Gamma_{-}}\int_{e}\boldsymbol{\omega}^{l}\cdot\boldsymbol{n}\alpha^{l}\varphi_{n}{\rm d}s.$ The system (5.1) can be further rewritten in block matrix form $\boldsymbol{D}^{(l)}\hat{\boldsymbol{U}}=\boldsymbol{F}^{(l)},\quad 1\leq l\leq L,$ where $\boldsymbol{D}^{(l)}:=[-\boldsymbol{B}_{1}^{(l)},\cdots,-\boldsymbol{B}_{l-1}^{(l)},{\boldsymbol{A}^{(l)}}-\boldsymbol{B}_{l}^{(l)}+\boldsymbol{C}^{(l)},\cdots,-\boldsymbol{B}_{L}^{(l)}].$ The final linear system is $\boldsymbol{D}\hat{\boldsymbol{U}}=\boldsymbol{F},$ (5.2) where $\boldsymbol{D}=\left[\begin{array}[]{*{20}{c}}\boldsymbol{D}^{(1)}\\\ \vdots\\\ \boldsymbol{D}^{(L)}\end{array}\right],\quad\hat{\boldsymbol{U}}=\left[\begin{array}[]{*{20}{c}}\hat{\boldsymbol{U}}^{1}\\\ \vdots\\\ {\hat{\boldsymbol{U}}}^{L}\end{array}\right],\quad\boldsymbol{F}=\left[{\begin{array}[]{*{20}{c}}\boldsymbol{F}^{(1)}\\\ \vdots\\\ \boldsymbol{F}^{(L)}\end{array}}\right].$ We solve (5.2) by using the block Gauss-Seidal iteration method. The accuracy is measured by the weighted relative error defined by $\|u-u_{h}\|_{rel}=\frac{\left(\sum\limits_{l=1}^{L}\omega_{l}\|u(\cdot,\boldsymbol{\omega}^{l})-u_{h}^{l}\|_{0,D}^{2}\right)^{1/2}}{\left(\sum\limits_{l=1}^{L}\omega_{l}\|u(\cdot,\boldsymbol{\omega}^{l})\|_{0,D}^{2}\right)^{1/2}},$ where $u_{h}^{l}=\sum\limits_{m=1}^{M}\hat{u}_{m}^{l}\varphi_{m}$ is the numerical solution. ### 5.2 Examples in three dimensions ###### Example 5.1. We take $\sigma_{t}=2$, $\sigma_{s}=1$ and $\eta=0$. The domain $D$ is a unit cube. With the right-hand side function $\displaystyle f(\boldsymbol{x},\boldsymbol{\omega})$ $\displaystyle=\pi s_{1}\cos(\pi x_{1})\sin(\pi x_{2})\sin(\pi x_{3})+\pi s_{2}\sin(\pi x_{1})\cos(\pi x_{2})\sin(\pi x_{3})$ $\displaystyle\quad+\pi s_{3}\sin(\pi x_{1})\sin(\pi x_{2})\cos(\pi x_{3})+\sin(\pi x_{1})\sin(\pi x_{2})\sin(\pi x_{3}),$ where $\boldsymbol{\omega}=(s_{1},s_{2},s_{3})$, the exact solution is $u(\boldsymbol{x},\boldsymbol{\omega})=\sin(\pi x_{1})\sin(\pi x_{2})\sin(\pi x_{3}).$ Fig. 1: Sparse pattern of the coefficient matrix for Example 5.1 ($N=3,k=2,n=2$) The sparse pattern for the coefficient matrix is shown in Fig. 1. The total number of the entries is $8200\times 8200={\text{67371264}}$ and the number of nonzero elements is ${\text{nz}}=239760$. Thus the sparsity ratio is 99.64%. (a) Exact (b) $\theta_{0}=10$ (c) $\theta_{0}=100$ (d) $\theta_{0}=500$ Fig. 2: The expansion coefficients for Example 5.1 with different stabilization parameters ($S_{2},k=2$) Tab. 2: Relative errors for Example 5.1: $k$ v.s. $S_{n}$ ($N=2$) $n$ | 2 | 4 | 6 | 8 | 10 ---|---|---|---|---|--- $k=1$ | 1.7133e-01 | 1.7329e-01 | 1.7480e-01 | 1.7491e-01 | 1.7500e-01 $k=2$ | 8.9453e-03 | 8.3749e-03 | 8.0365e-03 | 8.0532e-03 | 8.0755e-03 Tab. 3: Relative errors for Example 5.1: $N$ v.s. $S_{n}$ ($k=2$) $n$ | 2 | 4 | 6 | 8 | 10 ---|---|---|---|---|--- $N=2$ | 8.9453e-03 | 8.3749e-03 | 8.0365e-03 | 8.0532e-03 | 8.0755e-03 $N=3$ | 2.2512e-03 | 2.1269e-03 | 2.0150e-03 | 2.0138e-03 | 2.0136e-03 Fig. 2 displays the expansion coefficients which coincide in each angular direction since the true solution is independent of the angular variable $\boldsymbol{\omega}$. We observe a better result for bigger stabilization parameter $\theta_{0}$. In the following we always choose $\theta_{0}=10^{N+k}$ due to its good performance in different cases. For the isotropic case $\eta=0$, $S_{2}$ method is enough to resolve the solution accurately in angle as indicated by the numerical results in Tabs. 2 and 3. For the given example, the error is then dominated by the spatial problems. According to Theorem 4.1, the error bound is $\mathcal{O}(c(\theta_{0})|\log_{2}h|^{d}h^{k+1/2})$. With the choice for $\theta_{0}$ in this case, we have $c(\theta_{0})|\log_{2}h|^{d}h^{k+1/2}\approx\mathcal{O}(|\log_{2}h|^{d}h^{k})$ and the logarithmic factor implies a slightly lower order than $k$. From Tab. 4, we see that the convergence rates for $k=1,3$ are better than $\mathcal{O}(h^{k+1/2})$ and even the $(k+1)$-th order can be obtained for $k=3$. For $k=2,4$ the order is about $k$. Tab. 4: $L^{2}$ errors of $S_{2}$ method for Example 5.1 $N$ | $k=1$ | | $k=2$ | | $k=3$ | | $k=4$ ---|---|---|---|---|---|---|--- Err | rate | | Err | rate | | Err | rate | | Err | rate 1 | 4.8695e-01 | - | | 3.7626e-02 | - | | 3.8603e-03 | - | | 2.9324e-04 | - 2 | 1.7133e-01 | 1.5070 | | 8.9453e-03 | 2.0725 | | 2.1133e-04 | 4.1911 | | 1.6406e-05 | 4.1598 3 | 5.6436e-02 | 1.6021 | | 2.2512e-03 | 1.9904 | | 1.2971e-05 | 4.0261 | | 7.8260e-07 | 4.3898 4 | 1.6990e-02 | 1.7319 | | 5.6295e-04 | 1.9996 | | 8.2285e-07 | 3.9785 | | - | - ###### Example 5.2. We take $\sigma_{t}=3$ and $\sigma_{s}=1$. The domain $D$ is a unit cube. The true solution is taken as $u(\boldsymbol{x},\boldsymbol{\omega})=10\omega_{3}\sin(\pi x_{1})\sin(\pi x_{2})\sin(\pi x_{3}),$ from which we know after a direct manipulation that the right-hand side function is $\displaystyle f(\boldsymbol{x},\boldsymbol{s})=$ $\displaystyle 10(\sigma_{t}-\eta\sigma_{s})s_{3}\sin(\pi x_{1})\sin(\pi x_{2})\sin(\pi x_{3})$ $\displaystyle+10\pi s_{3}^{2}\sin(\pi x_{1})\sin(\pi x_{2})\cos(\pi x_{3})+10\pi s_{2}s_{3}\sin(\pi x_{1})\cos(\pi x_{2})\sin(\pi x_{3})$ $\displaystyle+10\pi s_{1}s_{3}\cos(\pi x_{1})\sin(\pi x_{2})\sin(\pi x_{3})$ where $\boldsymbol{\omega}=(s_{1},s_{2},s_{3})$. Tab. 5: Relative errors for Example 5.2 ($S_{2},\eta=0.1$) $N$ | $k=1$ | $k=2$ | $k=3$ | $k=4$ ---|---|---|---|--- 1 | 2.2797e-01 | 1.6584e-02 | 1.7683e-03 | 2.1850e-04 2 | 8.2048e-02 | 3.7848e-03 | 2.0045e-04 | 1.7282e-04 Tab. 6: Relative errors for Example 5.2 ($N=1,\eta=0.9$) $n$ | $k=1$ | $k=2$ | $k=3$ | $k=4$ ---|---|---|---|--- 2 | 6.6685e-01 | 6.6872e-01 | 6.6969e-01 | 2.0653e+01 4 | 6.4693e-01 | 1.4665e+01 | 1.5884e+01 | 1.6226e+01 6 | 1.1411e+00 | 1.2068e+00 | 1.2539e+00 | 1.2606e+00 8 | 1.1511e-01 | 1.3098e-01 | 1.3206e-01 | 1.3212e-01 10 | 7.3980e-02 | 7.8657e-02 | 7.9128e-02 | 7.9144e-02 12 | 4.3524e-02 | 3.2606e-02 | 3.2684e-02 | 3.2690e-02 We observe from Tab. 5 that $S_{2}$ method is accurate enough for the anisotropy factor close to isotropic cases. However, for strong forward scattering of $\eta=0.9$, it does not give a satisfactory result. We have to choose a larger $n$ to get an improved result, which, however, is not expected in real applications since $S_{n}$ method has $n(n+2)$ angular directions and hence $n(n+2)$ coupled spatial problems. In this case, some models have been developed to approximate the integral operator (cf. [26, 37, 48]). Another approach is to combine the sparse grid technique with the spherical harmonic method. Fig. 3: Sparse pattern of the coefficient matrix for Example 5.3 ($N=1,k=2,n=2$) ###### Example 5.3. This example is taken from the reference [32], where the Henyey-Greenstein function is replaced by the simplified approximate Mie (SAM): $g(t)=K_{S}(1+t)^{n_{p}},~{}~{}t\in[-1,1],$ where $n_{p}=\frac{2\eta}{1-\eta}$ is the anisotropic index and $K_{S}=\frac{1}{2\pi}\frac{n_{p}+1}{2^{n_{p}+1}}$ is the normalization factor. The geometric parameters and the true solution are the same as Example 5.2. For $S_{2}$ method with $N=1$ and $k=2$, the sparse pattern for the coefficient matrix is shown in Fig. 3. We also display the numerical and exact coefficients and $L^{2}$ projections at $z=0$ associated with the first angular direction in Fig. 4. We repeat the test for highly forward-peaked scattering with $\eta=0.9$. From Tab. 7 we observe a relatively smaller errors than that from Tab. 6, but the convergence behaviours are the same since the errors do not decrease significantly with the increase of $k$ and $n$. Fig. 4: Numerical and exact coefficients and $L^{2}$ projections for Example 5.3 ($N=1,k=2,n=2$) Tab. 7: Relative errors for Example 5.3 ($N=1,\eta=0.9$) $n$ | $k=1$ | $k=2$ | $k=3$ | $k=4$ ---|---|---|---|--- 2 | 6.0818e-01 | 7.2904e-01 | 7.3901e-01 | 7.3967e-01 4 | 3.7295e-01 | 4.0517e-02 | 3.1238e-02 | 3.1143e-02 6 | 3.7622e-01 | 3.7490e-02 | 2.7350e-02 | 2.7243e-02 8 | 3.7823e-01 | 2.8034e-02 | 1.0114e-02 | 9.7213e-03 10 | 3.7872e-01 | 2.6477e-02 | 3.5129e-03 | 2.0719e-03 12 | 3.7875e-01 | 2.6452e-02 | 3.2775e-03 | 1.6395e-03 ### 5.3 Flux distributions in two and three dimensions We now investigate the impact of the source term on the flux distributions. The isotropic photon flux is defined by $q(\boldsymbol{x})=\frac{1}{4\pi}\int_{S^{2}}u(\boldsymbol{x},\hat{\boldsymbol{\omega}}){\rm d}\sigma(\hat{\boldsymbol{\omega}}).$ For simplicity, vacuum boundary conditions are applied on all the boundaries. We always consider the isotropic scattering, and take $N=k=2$ for the spatial discretization. The examples in this subsection are taken from the reference [33]. ###### Example 5.4. This problem is defined on a unit cube with vacuum boundaries. The first 0.2 by 0.2 by 0.2 region $R$ contains a uniform isotropic source. For simplicity, we consider the following right-hand side function: $f(\boldsymbol{x},\boldsymbol{\omega})=f(\boldsymbol{x})=\begin{cases}1,\quad\boldsymbol{x}\in R=[0,0.2]^{3},\\\ 0,\quad\boldsymbol{x}\in D\backslash R.\end{cases}$ The entire box is of uniform composition with the following data: $\sigma_{t}=1$ and $\sigma_{s}=0.4$. For $z=0.1$ fixed, the contour plot of the flux distributions with varying orders of the discrete ordinates is displayed in Fig. 5. We can see clearly that the contour map shows rays emanating from the source. (a) $S_{2}$ (b) $S_{4}$ (c) $S_{6}$ Fig. 5: Contour map of the photon flux distributions for Example 5.4 We now perform the test on the problem in $(x,y)$-geometry. In this case, all coefficients, the boundary data and the solution of (2.1)-(2.2) are independent of the space variable $x_{3}=z$. ###### Example 5.5. This problem is defined on a unit square with vacuum boundaries. The first 0.2 by 0.2 region localized in the lower left corner contains a uniform isotropic source. The entire box is of uniform composition with the following data: $\sigma_{t}=1$ and $\sigma_{s}=0.4$. For this example, we only consider the $S_{4}$ method. The contour plot of the flux distributions is displayed in Fig. 6 (a). Numerical results of other cases are listed in Fig. 6 (b)-(f) by changing the positions or increasing the numbers of the isotropic sources. In all cases, we again observe the rays emanating from the sources. Fig. 6: Contour map of the photon flux distributions for Example 5.5 ### 5.4 Examples with complex spatial domains in two dimensions In the following, we extend the method to solve the RTE for some non-tensor product spatial domains in two dimensions. We always consider the isotropic scattering. Fig. 7: Initial subdivision of a $L$-shaped region for Example 5.6 ###### Example 5.6. The spatial domain $D$ is an $L$-shaped region in 2-D displayed in Fig. 7, consisting of three rectangles $R_{1}$, $R_{2}$ and $R_{2}$, where $R_{1}=\Big{\\{}(x_{1},x_{2}):0\leq x_{1}\leq 1,~{}~{}1\leq x_{2}\leq 2\Big{\\}},$ $R_{2}=\Big{\\{}(x_{1},x_{2}):0\leq x_{1}\leq 1,~{}~{}0\leq x_{2}\leq 1\Big{\\}},$ $R_{3}=\Big{\\{}(x_{1},x_{2}):1\leq x_{1}\leq 2,~{}~{}0\leq x_{2}\leq 1\Big{\\}}.$ The parameters are the same as Example 5.1 and the true solution is $u(\boldsymbol{x},\boldsymbol{\omega})=\sin(\pi x_{1})\sin(\pi x_{2}).$ (a) Exact solution (b) $N=2,k=1$ (c) $N=2,k=2$ Fig. 8: Exact and numerical solutions for Example 5.6 Let $R=R_{1}\cup R_{2}\cup R_{3}$ be the initial subdivision of the $L$-shaped region and $\boldsymbol{V}_{0}^{k}(R)$ denote the piecewise polynomial space on $R$. We have the following orthogonal decomposition $\boldsymbol{V}_{0}^{k}(R)=\boldsymbol{V}_{0}^{k}(R_{1})\oplus\boldsymbol{V}_{0}^{k}(R_{2})\oplus\boldsymbol{V}_{0}^{k}(R_{3}),$ where functions in $\boldsymbol{V}_{0}^{k}({R_{j}})~{}(j=1,2,3)$ are extended by zero to $\mathbb{R}^{2}$. For each $R_{j}$, one can regard it as $[0,1]^{2}$ and give the sparse representation by using an affine transformation. In Fig. 8, we display the numerical solutions for different $N$ and $k$ and the relative errors are given in Tab. 8. Tab. 8: Relative errors for Example 5.6 $N$ | $k=1$ | $k=2$ | $k=3$ | $k=4$ ---|---|---|---|--- 1 | 2.2059e-01 | 1.6769e-02 | 1.7691e-03 | 1.3197e-04 2 | 6.1359e-02 | 2.1758e-03 | 1.1360e-04 | 4.1841e-06 3 | 1.7434e-02 | 3.0707e-04 | 7.2792e-06 | 2.1799e-07 4 | 4.8163e-03 | 4.2519e-05 | 4.6425e-07 | - (a) Initial division (b) The final division $\mathcal{T}_{h}$ Fig. 9: Subdivision of the circular domain ###### Example 5.7. The spatial domain $D$ is a circular region displayed in Fig. 9. The parameters and the true solution are the same as the last example. To use the sparse grid method, we first plot a sufficiently large rectangle in the domain and approximate the boundary curve by a polygon as depicted in Fig. 9 (a). For simplicity, the boundary data corresponding to the polygon is obtained from the exact solution. For the general case, some approximation should be implemented, for example, the technique from the isoparametric finite elements. To avoid hanging nodes, the polygon approximation and the corresponding triangulation can be made consistent with the final partition of the rectangle, see Fig. 9 (b). We should note that the hanging nodes are allowed in our procedure since no interelement continuity is required. Denote the rectangle by $R$ and the other triangles by $T_{1},\cdots,T_{8}$, respectively. Let $\Omega=R\cup{T_{1}}\cup\cdots\cup{T_{8}}$. We then consider the initial DG space given by $\boldsymbol{V}_{0}^{k}(\Omega)=\boldsymbol{V}_{0}^{k}({R})\oplus\boldsymbol{V}_{0}^{k}({T_{1}})\oplus\cdots\oplus\boldsymbol{V}_{0}^{k}({T_{8}}).$ Tab. 9: Orthonormal bases on the reference triangle $\tau$ for $k\leq 2$ ($r:=\lambda_{1},s:=\lambda_{2}$) $k=0$ | ---|--- | $\varphi_{1}=\sqrt{2}$ $k=1$ | | $\begin{gathered}\varphi_{2}=6r-2\hfill\\\ \varphi_{3}=2\sqrt{3}(2r+s-1)\hfill\\\ \end{gathered}$ $k=2$ | | $\begin{gathered}\varphi_{4}=\sqrt{6}(10r^{2}-8r+1)\hfill\\\ \varphi_{5}=3\sqrt{2}(5r-1)(r+2s-1)\hfill\\\ \varphi_{6}=\sqrt{30}(r^{2}+6rs-2r+6s^{2}-6s+1)\hfill\\\ \end{gathered}$ The orthonormal bases corresponding to $R$ has been given in the previous section, while the orthonormal bases on each $T_{i}$ can be obtained by using the Gram-Schmidt procedure. For any triangle $T$ with vertices $z_{i}={({x_{i}},{y_{i}})}$, $i=1,2,3$, any point $z={(x,y)}$ can be represented by the barycentric coordinates as $\begin{cases}x=x_{1}\lambda_{1}+x_{2}\lambda_{2}+x_{3}\lambda_{3},\\\ y=y_{1}\lambda_{1}+y_{2}\lambda_{2}+y_{3}\lambda_{3},\\\ 1=\lambda_{1}+\lambda_{2}+\lambda_{3}.\end{cases}$ Note that $\iint_{T}f(x,y)g(x,y){\rm d}x{\rm d}y=2|T|\int_{0}^{1}\int_{0}^{1-\lambda_{1}}\tilde{f}(\lambda_{1},\lambda_{2})\tilde{g}(\lambda_{1},\lambda_{2}){\rm d}\lambda_{2}{\rm d}\lambda_{1},$ where $\tilde{f}(\lambda_{1},\lambda_{2}):=f(x(\lambda_{1},\lambda_{2}),y(\lambda_{1},\lambda_{2})).$ We then define an inner product on the reference triangle $\tau$ by $(\tilde{f},\tilde{g})_{\tau}=\int_{0}^{1}\int_{0}^{1-\lambda_{1}}\tilde{f}(\lambda_{1},\lambda_{2})\tilde{g}(\lambda_{1},\lambda_{2}){\rm d}\lambda_{2}{\rm d}\lambda_{1}.$ Given the orthonormal bases on $\tau$ by $\\{\tilde{\varphi}_{i}\\}$, we then obtain the bases on $T$ given by $\psi_{i}(x,y)=\sqrt{\frac{1}{2|T|}}\varphi_{i}(x,y).$ For any function $f(x,y)$ defined on $T$, the projection coefficients are computed as $\displaystyle c_{i}$ $\displaystyle=\iint_{T}f(x,y)\psi_{i}(x,y){\rm d}x{\rm d}y=2|T|\int_{0}^{1}\int_{0}^{1-\lambda_{1}}\tilde{f}(\lambda_{1},\lambda_{2}){{\tilde{\psi}}_{i}}(\lambda_{1},\lambda_{2}){\rm d}\lambda_{2}{\rm d}\lambda_{1}$ $\displaystyle=\sqrt{2|T|}\int_{0}^{1}\int_{0}^{1-\lambda_{1}}\tilde{f}(\lambda_{1},\lambda_{2}){\tilde{\varphi}}_{i}(\lambda_{1},\lambda_{2}){\rm d}\lambda_{2}{\rm d}\lambda_{1}=:\sqrt{2|T|}{\tilde{c}}_{i}.$ The orthogonal bases on $\tau$ are obtained by using Gram-Schmidt procedure to the polynomial set $\\{1,\lambda_{1},\lambda_{2},\lambda_{1}^{2},\lambda_{1}\lambda_{2},\lambda_{2}^{2},\cdots\\}$, some of which are listed in Tab. 9. The relative error is defined by ${\rm Err}=\|f-f_{h}\|_{L^{2}(\mathcal{T}_{h})}/\|f\|_{L^{2}(\mathcal{T}_{h})}$ and given in Tab. 10. Tab. 10: Relative errors for Example 5.7 $(N=2)$ $k$ | 0 | 1 | 2 | 3 ---|---|---|---|--- Err | 5.8510e-01 | 6.1678e-02 | 9.3273e-03 | 5.5965e-04 Summarizing our main observations from the numerical results reported in all previous examples, we may conclude that * 1. The sparse discrete ordinate DG method can greatly reduce the spatial degrees of freedom while keeping almost the same accuracy up to multiplication of an log factor. * 2. The proposed method is highly effective for problems away from strong forward scattering. To get an improved result, large discrete-ordinate sets are needed for highly forward-peaked case. * 3. The method can be extended to solve the RTE efficiently for some non-tensor product spatial domains in two dimensions. ## 6 Conclusions and remarks In this paper, we combine the sparse grid technique with the discrete ordinate DG method to solve the RTE with inflow boundary conditions, which can be adapted to other types of boundary conditions. Under suitable regularity assumptions, we derive error estimates for the numerical solutions. Results from many numerical examples show the good convergence behavior of the method. For highly forward-peaked scattering, there have been substantial efforts made to develop simpler approximations to integral scattering operator $S$. One well-established example is the so-called Fokker-Planck equation (cf. [37]), to which the sparse grid techniques can also be applied. ## Acknowledgments The work was partially supported by NSFC (Grant No. 12071289) and the Strategic Priority Research Program of Chinese Academy of Sciences (Grant No. XDA25010402). ## References * [1] M. L. Adams and E. W. Larsen. Fast iteration methods for discrete-ordinates partical transport calculations. Prog. Nucl. Energy, 40(1):3–159, 2002. * [2] V. Agoshkov. Boundary Value Problems for Transport Equations. Birkhauser, Boston, 1998. * [3] B. Alpert. A class of bases in $L^{2}$ for the sparse representation of integral operators. SIAM J. Math. Anal., 24(1):246–262, 1993. * [4] B. Alpert, G. Beylkin, D. Gines, and L. Vozovoi. Adaptive solution of partial differential equations in multiwavelet bases. J. Comput. Phys., 182:149–190, 2002. * [5] M. Asadzadeh and A. Kadem. Chebyshev spectral-$S_{N}$ method for the neutron transport equation. Comput. Math. Appl., 52(3-4):509–524, 2006. * [6] K. Atkinson and W. Han. Spherical Harmonics and Approximations on the Unit Sphere: An Introduction. Springer, Heidelberg, 2012. * [7] D. Balsara. Fast and accurate discrete ordinates methods for multidimensional radiative transfer. Part I, basic methods. J. Quant. Spectrosc. Radiat. Transf., 69(6):671–707, 2001. * [8] S. C. Brenner and L. R. Scott. The Mathematical Theory of Finite Element Methods. Springer-Verlag, New York, 2008. * [9] F. Brezzi, B. Cockburn, L. D. Marini, and E. Süli. Stabilization mechanisms in discontinuous galerkin finite element methods. Comput. Methods Appl. Mech. Engrg., 195(25-28):3293–3310, 2006\. * [10] F. Brezzi, L. D. Marini, and E. Süli. Discontinuous Galerkin methods for first-order hyperbolic problems. Math. Models Meth. Appl. Sci., 14(12):1893–1903, 2004. * [11] R. E. Caflisch. Monte carlo and quasi-monte carlo methods. Acta Numer., 7:1–49, 1998. * [12] K. M. Case and P. F. Zweifel. Linear Transport Theory. Addison-Wesley, Reading, MA, 1967. * [13] B. Chang, T. Manteuffel, S. McCormick, J. Ruge, and B. Sheehan. Spatial multigrid for isotropic neutron transport. SIAM J. Sci. Comput., 29:1900–1917, 2007. * [14] B. Cockburn. Discontinuous Galerkin methods. ZAMM Z. Angew. Math. Mech., 83(11):731–754, 2003. * [15] L. Dan, J. Cheng, and C. Shu. Conservative high order positivity-preserving discontinuous Galerkin methods for linear hyperbolic and radiative transfer equations. J. Sci. Comput., 77(3):1801–1831, 2018. * [16] J. J. Duderstadt and W. R. Martin. Transport Theory. John Wiley, New York, 1978. * [17] P. Edström. A fast and stable solution method for the radiative transfer problem. SIAM Rev., 47(3):447–468, 2005. * [18] M. Frank, A. Klar, E. W. Larsen, and S. Yasuda. Time-dependent simplified $P_{N}$ approximation to the equations of radiative transfer. J. Comput. Phys., 226:2289–2305, 2007. * [19] T. Gerstner and M. Griebel. Numerical integration using sparse grids. Numer. Algorithms, 18:209–232, 1998. * [20] F. Golse, S. Jin, and C. Levermore. The convergence of numerical transfer schemes in diffusive regimes. I. Discrete-ordinate method. SIAM J. Numer. Anal., 36(5):1333–1369, 1999. * [21] K. Grella and C. Schwab. Sparse discrete ordinates method in radiative transfer. Comput. Methods Appl. Math., 11(3):305–326, 2011. * [22] M. Griebel. A parallelizable and vectorizable multi-level algorithm on sparse grids. Parallel algorithms for partial differential equations (kiel, 1990). Notes Numer. Fluid Mech., 31:94–100, 1991. * [23] M. Griebel. Adaptive sparse grid multilevel methods for elliptic PDEs based on finite differences. Computing, 61(2):151–179, 1998. * [24] W. Guo and Y. Cheng. A sparse grid discontinuous Galerkin method for high-dimensional transport equations and its application to kinetic simulations. SIAM J. Sci. Comput., 38(6):A3381–A3409, 2016. * [25] W. Guo and Y. Cheng. An adaptive multiresolution discontinuous Galerkin method for time-dependent transport equations in multidimensions. SIAM J. Sci. Comput., 39(6):A2962–A2992, 2017. * [26] W. Han, J. Eichholz, and G. Wang. On a family of differential approximations of the radiative transfer equation. J. Math. Chem., 50(4):689–702, 2012. * [27] W. Han, J. Huang, and J. A. Eichholz. Discrete-ordinate discontinuous Galerkin methods for solving the radiative transfer equation. SIAM J. Sci. Comput., 32(2):477–497, 2010. * [28] A. D. Kim and M. Moscoso. Chebyshev spectral methods for radiative transfer. SIAM J. Sci. Comput., 23:2074–2094, 2002. * [29] M. Kindelan, F. Bernal, P. González-Rodríguez, and M. Moscoso. Application of the RBF meshless method to the solution of the radiative transport equation. J. Comput. Phys., 229:1897–1908, 2010. * [30] E. W. Larsen and J. E. Morel. Advances in Discrete-ordinates Methodology. Springer, New York, 2010. * [31] E. E. Lewis and W. F. Miller. Computational Methods of Neutron Transport. John Wiley & Sons, New York, 1984. * [32] P. Liu. A new phase function approximating to Mie scattering for radiative transport equations. Phys. Med. Biol., 39:1025–1036, 1994. * [33] J. A. Roberts. Direct solution of the discrete ordinates equations. 2010\. * [34] H. Sadat. On the use of a meshless method for solving radiative transfer with the discrete ordinates formulations. J. Quant. Spectrosc. Radiat. Transf., 101:263–268, 2006. * [35] C. Schwab, E. Süli, and R. A. Todor. Sparse finite element approximation of high-dimensional transport-dominated diffusion problems. M2AN Math. Model. Numer. Anal., 42(5):777–819, 2008. * [36] W. Shao, Q. Sheng, and C. Wang. A cascadic multigrid asymptotic-preserving discrete ordinate discontinuous streamline diffusion method for radiative transfer equations with diffusive scalings. Comput. Math. Appl., 80(6):1650–1667, 2020. * [37] Q. Sheng and W. Han. Well-posedness of the Fokker-Planck equation in a scattering process. J. Math. Anal. Appl., 406(2):531–536, 2013. * [38] Q. Sheng, C. Wang, and W. Han. An optimal cascadic multigrid method for the radiative transfer equation. J. Comput. Appl. Math., 303:189–205, 2016. * [39] S. A. Smoljak. Quadrature and interpolation formulae on tensor products of certain function classes. Dokl. Akad. Nauk SSSR, 148:1042–1045, 1963. * [40] M. Tang. A uniform first-order method for the discrete ordinate transport equation with interfaces in X,Y-geometry. J. Comput. Math., 27(6):764–786, 2009. * [41] Z. Tao, Y. Jiang, and Y. Cheng. An adaptive high-order piecewise polynomial based sparse grid collocation method with applications. arXiv:1912.03982v1, pages 1–33, 2019. * [42] C. Wang, H. Sadat, and J. Tan. First-order and second-order meshless formulations of the radiative transfer equation: a comparative study. Numer. Heat Transfer B, 66:21–42, 2014. * [43] Z. Wang, Q. Tang, and W. Guo. Sparse grid discontinuous Galerkin methods for high-dimensional elliptic equations. J. Comput. Phys., 314:244–263, 2016. * [44] G. Widmer, R. Hiptmair, and C. Schwab. Sparse adaptive finite elements for radiative transfer. J. Comput. Phys., 227(12):6071–6105, 2008. * [45] D. Yuan, J. Cheng, and C. Shu. High order positivity-preserving discontinuous Galerkin methods for radiative transfer equations. SIAM J. Sci. Comput., 38(5):A2987–A3019, 2016. * [46] C. Zenger. Sparse grids. In W. Hackbusch, editor, Parallel algorithms for partial differential equations, Proceedings of the Sixth GAMM-Seminar, Kiel, 1990\. Notes on Num. Fluid Mech. Vieweg-Verlag, 31:241–251, 1990. * [47] M. Zhang, J. Cheng, and J. Qiu. High order positivity-preserving discontinuous galerkin schemes for radiative transfer equations on triangular meshes. J. Comput. Phys., 397:108811, 2019. * [48] H. Zheng and W. Han. On simplified spherical harmonics equations for the radiative transfer equation. J. Math. Chem., 49(8):1785–1797, 2011.
# Lower bounds for the warping degree of a knot projection Atsushi Ohya Department of Computer Science and Engineering, University of Yamanashi, 4-4-37, Takeda, Kofu-shi, Yamanashi, 400-8510, Japan. Ayaka Shimizu Department of Mathematics, National Institute of Technology (KOSEN), Gunma College, 580 Toriba, Maebashi-shi, Gunma, 371-8530, Japan. Email: <EMAIL_ADDRESS><EMAIL_ADDRESS> ###### Abstract The warping degree of an oriented knot diagram is the minimal number of crossings which we meet as an under-crossing first when we travel along the diagram from a fixed point. The warping degree of a knot projection is the minimal value of the warping degree for all oriented alternating diagrams obtained from the knot projection. In this paper, we consider the maximal number of regions which share no crossings for a knot projection with a fixed crossing, and give lower bounds for the warping degree. ## 1 Introduction In this paper we assume that every knot diagram and knot projection has at least one crossing. A based knot diagram is a knot diagram which is given a base point on the diagram avoiding crossings. We denote by $D_{b}$ a based diagram $D$ with the base point $b$. The warping degree, $d(D_{b})$, of an oriented based knot diagram $D_{b}$ is the number of crossings such that we encounter the crossing as an under-crossing first when we travel along $D$ with the orientation starting at $b$. We call such a crossing a warping crossing point of $D_{b}$ (see Figure 1). Figure 1: The oriented based knot diagram $D_{b}$ has warping degree one. The crossing $p$ is the warping crossing point of $D_{b}$. The warping degree, $d(D)$, of an oriented knot diagram $D$ is the minimal value of $d(D_{b})$ for all base points $b$ of $D$ ([4]). A knot diagram is said to be monotone, or descending, if the warping degree is zero. Conversely, we can assume that the warping degree represents a complexity of a diagram in terms of how distant a knot diagram is from a monotone diagram. Note that a monotone knot diagram is a diagram of the trivial knot. A knot diagram is said to be alternating if we encounter an over-crossing and an under-crossing alternatively when traveling the diagram starting at any point on the diagram. Let $P$ be an unoriented knot projection. The warping degree of $P$ is defined to be the minimal value of the warping degree for all the oriented alternating diagrams obtained from $P$ by giving the orientation and crossing information. In Figure 2, all the reduced knot projections with warping degree one and two are shown ([6]). Further examples for warping degree three or four are listed in the table in Section 5 in this paper. As we may see, the warping degree of a knot projection shows somewhat complexity of a knot projection, like how “curly” a knot projection is, or how “quick” to back to a crossing when traveling the projection. Figure 2: All the reduced knot projections of warping degree one or two. The first two knot projections have warping degree one, and the others have two. The knot projections of warping degree two are determined in [6] by considering all possibilities of connections of the unavoidable parts. Further explorations in the same way for warping degree three or more would be difficult since there are too many kinds of unavoidable parts and too many possibilities of their connections. In this paper, we introduce the maximal independent region number, $\mathrm{IR}(P)$, of a knot projection in Section 2, and show the following inequality which is useful to estimate the warping degree. x ###### Theorem 1.1. The inequality $\mathrm{IR}(P)\leq d(P)\leq c(P)-\mathrm{IR}(P)-1$ holds for every reduced knot projection $P$, where $c(P)$ denotes the crossing number of $P$. x As mentioned in Sections 2 and 3, the value of $\mathrm{IR}(P)$ can be obtained without traveling along the knot projection, and also calculated just by solving simultaneous equations. Figure 3: Knot projections of warping degree three with 10, 11, 12 crossings. Since all the reduced knot projections with warping degree one and two are determined and we can find some knot projections with warping degree three (see Figure 3 and Section 5), we obtain the following table about the minimal value of the warping degree of reduced knot projections for each crossing number. $c$ | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 ---|---|---|---|---|---|---|---|---|---|--- $d^{\text{min}}(c)$ | 1 | 1 | 2 | 2 | 2 | 2 | 3 | 3 | 3 | 3 Table 1: The crossing number $c$ and the minimal value of warping degree $d^{\text{min}}(c)$ for all reduced knot projections with $c$ crossings. Regions of a knot or link projection are independent if they share no crossings. We also give the following lower bound for the warping degree which would be helpful to extend the above table. x ###### Theorem 1.2. If all the connected link projections with $n$, $n+1$ or $n+2$ crossings have $m$ or more independent regions, then $d(P)\geq m-1$ holds for all reduced knot projections $P$ with $n$ or more crossings. x The rest of the paper is organized as follows: In Section 2, we define the maximal independent region number $\mathrm{IR}(P)$ and prove Theorem 1.1. In Section 3, we introduce the calculation for $\mathrm{IR}(P)$ by simultaneous equations. In Section 4, we estimate $\mathrm{IR}(P)$ and the warping degree $d(P)$ and prove Theorem 1.2. In Section 5, we list and compare the values of $\mathrm{IR}(P)$ and $d(P)$. ## 2 Independent region sets In this section, we define the maximal independent region number, and using it we estimate the warping degree of a knot projection. Throughout this section, we assume that every knot diagram and knot projection is reduced. We have the following lemma (cf. [6]). x ###### Lemma 2.1. Let $D$ be an oriented alternating knot diagram. Let $c$ be a crossing of $D$. Take a base point $b$ just before an over-crossing of $c$. If $D$ has a region $R$ which does not incident to $c$, then one of the crossings on the boundary of $R$ is a warping crossing point of $D_{b}$, and one of that is a non- warping crossing point of $D_{b}$. x ###### Proof. Let $e$ be the edge on the boundary of $R$ such that we meet it first from $b$. Then $e$ has an under-crossing and an over-crossing, that is, a warping crossing point and a non-warping crossing point. Figure 4: The edge $e$ on the boundary of $R$ which we meet first from the base point $b$ has one under-crossing and one over-crossing, and they are a warping crossing point and non-warping crossing point of $D_{b}$, respectively, regardless of the orientation. ∎ x Similarly, we have the following. x ###### Corollary 2.2. Let $D$ be an oriented alternating knot diagram. Let $c$ be a crossing of $D$. Take a base point $b$ just before an over-crossing of $c$. If $D$ has independent $n$ regions which are not incident to $c$, then the inequality $n\leq d(D)\leq c(D)-n-1$ holds. x ###### Proof. By Lemma 2.1, $D$ has at least $n$ warping crossing points of $D_{b}$. Also, $D$ has at least $n+1$ non-warping crossing points since the crossing $c$ is a non-warping crossing point, too. Therefore we have $n\leq d(D_{b})\leq c(D)-n-1$. By the location of the base point $b$, we have $d(D_{b})=d(D)$ ([8]). ∎ x By definition, we have the following corollary for knot projections. x ###### Corollary 2.3. Let $P$ be a knot projection, and $c$ a crossing of $P$. If $P$ has independent $n$ regions which are not incident to $c$, then the inequality $n\leq d(P)\leq c(P)-n-1$ holds. x The strong point is that we can estimate the warping degree without traveling the projection (see Figure 5). Figure 5: The warping degree can be determined for some knot projections from the inequality of Corollary 2.3, without traveling along the knot projection. The knot projection $P$ has $1\leq d(P)\leq 3-1-1$, and $d(P)=1$. We also obtain $d(Q)=2,d(R)=3$ from the inequalities. This would enable us to estimate the warping degree more combinatorically. We call the set of regions of a knot projection $P$ which are independent and are not incident to a crossing $c$ an independent region set for $P^{c}$. We call the crossing $c$ a base crossing. We define the maximal independent region number of $P^{c}$, $\mathrm{IR}(P^{c})$, to be the maximal cardinality of an independent region set for $P^{c}$. We define the maximal independent region number of $P$, $\mathrm{IR}(P)$, to be the maximal value of $\mathrm{IR}(P^{c})$ for all base crossings $c$. We prove Theorem 1.1. Proof of Theorem 1.1. It follows from Corollary 2.3. $\Box$ x ## 3 Independent region sets and region choice matrix In this section we explore how to find the independent region sets. A region choice matrix $M$, defined in [2], of a knot projection $P$ of $n$ crossings is the following $n\times(n+2)$ matrix. (The transposition is known as an incidence matrix defined in [3].) If a crossing $c_{i}$ is on the boundary of a region $R_{j}$, the $(i,j)$ component of $M$ is 1, and otherwise 0 (see Figure 6). Figure 6: A region choice matrix $M=(m_{ij})$, where $m_{ij}=1$ if $R_{j}$ is incident to $c_{i}$, and otherwise $m_{ij}=0$. Now we find out all of the independent region sets for $P^{c_{3}}$ for the knot projection $P$ and the crossing $c_{3}$ in Figure 6 by looking at its region choice matrix. Since the crossing $c_{3}$ is involved with the four regions $R_{1},R_{4},R_{5}$ and $R_{7}$, namely, the third row has 1 at the first, fourth, fifth and seventh column, we can not choose them as independent regions for $P^{c_{3}}$. Hence we choose the regions from the rest regions $R_{2},R_{3}$ and $R_{6}$. Namely, we choose columns from the second, third and sixth so that there are no components with the value two or more in the sum of the columns. Thus we obtain all the independent region sets for $P^{c_{3}}$, as $\\{R_{2}\\}$, $\\{R_{3}\\}$, $\\{R_{6}\\}$ and $\\{R_{2},R_{6}\\}$. More generally, we can find out all the independent region sets for a knot projection $P^{c}$ for all base crossings $c$ from the region choice matrix by solving the following simultaneous equations111 If it works on $\mathbb{Z}_{2}$, it is known that the simultaneous equations have solutions for any $b_{i}$’s and any region choice matrix of a knot projection ([7], [3]). Besides, if $x_{i}$’s are permitted to have the value for any integer, it is also known that the simultaneous equations have solutions for any region choice matrix of a knot projection even if $b_{i}$’s have the value for any integers ([2]). In this case, however, the equation has no solutions for some $b_{i}$’s. In Lemma 4.2, we will see that it definitely has solutions for some $b_{i}$’s. for $x_{i}\in\\{1,0\\}\ (i=1,2,\dots,7)$ $\displaystyle\left(\begin{array}[]{ccccccc}1&1&1&0&0&0&1\\\ 1&1&1&1&0&0&0\\\ 1&0&0&1&1&0&1\\\ 0&0&0&1&1&1&1\\\ 0&0&1&1&0&1&1\end{array}\right)\left(\begin{array}[]{c}x_{1}\\\ x_{2}\\\ x_{3}\\\ x_{4}\\\ x_{5}\\\ x_{6}\\\ x_{7}\end{array}\right)=\left(\begin{array}[]{c}b_{1}\\\ b_{2}\\\ b_{3}\\\ b_{4}\\\ b_{5}\end{array}\right),$ for all $b_{i}\in\\{1,0\\}\ (i=1,2,\dots,7)$, where $b_{k}\neq b_{l}$ for some $k$ and $l$; If $b_{i}=0$ for all $i$, it implies that no regions are chosen. If $b_{i}=1$ for all $i$, it means all the crossings are on the chosen regions, and we can not have a base crossing. ## 4 Estimation for the maximal independent region number As Theorem 1.1 implies, the warping degree is estimated by the maximal independent region numbers. In this section, we estimate the maximal independent region number itself, and prove Theorem 1.2. First, we have the following: x ###### Lemma 4.1. The inequality $\displaystyle\mathrm{IR}(P)\leq\frac{c(P)-1}{2}$ holds for every reduced knot projection $P$. x ###### Proof. By Theorem 1.1, we have $\mathrm{IR}(P)\leq c(P)-\mathrm{IR}(P)-1$, and then have $2\mathrm{IR}(P)\leq c(P)-1$. ∎ x Next, we give a routine lower bound for $\mathrm{IR}(P)$. x ###### Lemma 4.2. For any knot projection $P$ with $c(P)\geq 2$, we have $\mathrm{IR}(P)\geq 1$. x ###### Proof. For the case that $c(P)=2$, $P$ has two independent bigons. By taking a base crossing at a crossing which belongs to one of the two bigons, we can take an independent region at the other bigon. For the case that $c(P)\geq 3$, then the number of regions is $3+2=5$ or more by the Euler characteristic (see, for example, [2]). Take a base crossing $c$. Then either three or four regions are incident to $c$. This means there exists a region which is not incident to $c$. Hence $\mathrm{IR}(P)\geq 1$ holds. ∎ x To give further lower bounds for $\mathrm{IR}(P)$, we show the following lemma for link projections. x ###### Lemma 4.3. If all the connected link projections with $n$, $n+1$ or $n+2$ crossings have $m$ or more independent regions, then all the connected link projections with $n$ or more crossings have $m$ or more independent regions. x ###### Proof. Let $P$ be a link projection with $n+3$ crossings. If $P$ is reducible, splice it at a reducible crossing as shown in Figure 7. Figure 7: Splice $P$ at a reducible crossing. Then we obtain a link projection, $P^{\prime}$, with $n+2$ crossings. By assumption, $P^{\prime}$ has $m$ independent regions. Take the corresponding regions of $P$; If the region $R$ of $P^{\prime}$ created by the splice has been chosen, take one of the parts $R^{1}$ and $R^{2}$ as a corresponding region (see Figure 7). Thus, we obtain $m$ independent regions of $P$. If $P$ is reduced, it is shown in [1] that $P$ has a bigon or trigon. Splice it at a bigon or a trigon as shown in Figure 8. Figure 8: Splice $P$ at a bigon or a trigon and obtain a link projection $P^{\prime}$ or $P^{\prime\prime}$, respectively. Note that $P^{\prime}$ has $n+2$ and $P^{\prime\prime}$ has $n$ crossings. Then $P^{\prime}$ and $P^{\prime\prime}$ have $m$ independent regions. Similarly, take $m$ regions of $P$ properly from the corresponding regions, which are independent. ∎ x We have the following corollary. x ###### Corollary 4.4. If all the connected link projections with $n$, $n+1$ or $n+2$ crossings have $m$ or more independent regions, then $\mathrm{IR}(P)\geq m-1$ holds for all reduced knot projections $P$ with $n$ or more crossings. x ###### Proof. All reduced knot projections with $n$ or more crossings have $m$ independent regions by Lemma 4.3. Take a base crossing $c$ at the boundary of one of the $m$ regions. Then, the rest $m-1$ regions are independent regions for $P^{c}$. ∎ x We prove Theorem 1.2. Proof of Theorem 1.2. From Corollary 4.4 we have $\mathrm{IR}(P)\geq m-1$, and from Theorem 1.1 we have $d(P)\geq\mathrm{IR}(P)$. $\Box$ ## 5 Table of $\mathrm{IR}(P)$ and $d(P)$ For the knot projections $P$ of prime alternating knots with up to nine crossings which are obtained from the knot diagrams in Rolfsen’s knot table ([5]), the values of $\mathrm{IR}(P)$ and $d(P)$ ([8]) are listed below. The values of $\mathrm{IR}(P)$ were obtained by the calculation using the SAT solver. ## Acknowledgment The authors thank Yoshiro Yaguchi for helpful comments. ## References * [1] C. C. Adams, R. Shinjo and K. Tanaka, Complementary regions of knot and link diagrams, Ann. Comb. 15 (2011), 549–563. * [2] K. Ahara and M. Suzuki, An integral region choice problem on knot projection, J. Knot Theory Ramifications 21 (2012), 1250119 [20 pages]. * [3] Z. Cheng and H. Gao, On region crossing change and incidence matrix, Sci. China Math. 55 (2012), 1487–1495. * [4] A. Kawauchi, Lectures on knot theory (in Japanese), Kyoritsu shuppan Co. Ltd, 2007. * [5] D. Rolfsen, Knots and links, Publish or Perish, Inc. (1976). * [6] A. Shimizu, Prime alternating knots of minimal warping degree two, J. Knot Theory Ramifications 29 (2020), 2050060. * [7] A. Shimizu, Region crossing change is an unknotting operation, J. Math. Soc. Japan 66 (2014), 693–708. * [8] A. Shimizu, The warping degree of a knot diagram, J. Knot Theory Ramifications 19 (2010), 849–857.
Distance Estimation for BLE-based Contact Tracing – A Measurement Study Bernhard Etzlinger, Barbara Nußbaummüller, Philipp Peterseil, and Karin Anna Hummel Johannes Kepler University, Linz, Austria<EMAIL_ADDRESS> This work has been supported by the LCM K2 Center within the framework of the Austrian COMET-K2 program. Mobile contact tracing apps are – in principle – a perfect aid to condemn the human-to-human spread of an infectious disease such as COVID-19 due to the wide use of smartphones worldwide. Yet, the unknown accuracy of contact estimation by wireless technologies hinders the broader use. We address this challenge by conducting a measurement study with a custom testbed to show the capabilities and limitations of Bluetooth Low Energy (BLE) in different scenarios. Distance estimation is based on interpreting the signal pathloss with a basic linear and a logarithmic model. Further, we compare our results with accurate ultra-wideband (UWB) distance measurements. While the results indicate that distance estimation by BLE is not accurate enough, a contact detector can detect contacts below 2.5$\,$m with a true positive rate of 0.65 for the logarithmic and of 0.54 for the linear model. Further, the measurements reveal that multi-path signal propagation reduces the effect of body shielding and thus increases detection accuracy in indoor scenarios. Contact Tracing, Wireless Communications, BLE § INTRODUCTION Contact tracing aims at fighting human-to-human infection spreading by identifying – and isolating – persons who were in close contact to an infected person. Quarantine is one of the most effective measures to break infection chains [1]. To support time-intense manual contact tracing, mobile contact tracing apps have been recently introduced that estimate and capture contacts as a measure against COVID-19. Bluetooth Low Energy (BLE) is seen as the most promising wireless technology for contact tracing [2]. The spatial distance between two smartphones is estimated based on the received signal strength indicator (RSSI), which is known to be challenged by non line-of-sight (LOS) conditions such as shielding by human bodies and multi-path propagation in particular in indoor environments. It has been shown that BLE RSSI based distance estimation requires additional concepts to increase accuracy [3]. Among alternative technologies, ultra-wideband (UWB) is a well-known technology often used for location estimation. Ultra-wideband (UWB) based distance calculation uses time-of-flight (ToF) measures and may be leveraged for contact tracing. Yet, UWB is currently only available in a few flagship smartphones such as Samsung Galaxy Note 20 or Apple iPhone 11. Further, UWB distance estimation relies on the cooperative exchange of time information [4], which is a potential privacy threat and will consume additional energy. In our work, we will use UWB for comparison and consider it as ground-truth measurement system. The focus of this paper rests on characterizing BLE RSSI based distance estimation for contact tracing. We make the following contributions: * We introduce our measurement testbed consisting of a mobile app that allows to capture BLE and UWB-based distance estimates, the latter for comparison. BLE-based distance estimation makes use of a linear and a logarithmic pathloss model that interpret BLE RSSI values measured onboard of the smartphone. UWB distance readings are retrieved from connected external UWB modules. * We summarize the achieved accuracy of distance and exposure estimation in our measurement campaign in different scenarios consisting of 20'535 BLE logs. Our results confirm that using BLE RSSI for distance estimation is challenging and that the linear distance estimation model provides better distance estimation accuracy than the logarithmic estimation model. We further show that in order to enhance exposure detection, awareness of the smartphones carrying position is more beneficial than knowledge about the environment. * We study the effect of body shielding and multi-path propagation in an anechoic chamber (comparable to outdoor scenarios), LOS indoors, and in a corner scenario indoors with varying phone carrying positions. We find that BLE multi-path propagation can reduce body shielding effects. Finally, we compare BLE-based distance estimation with the more robust and accurate UWB ToF based estimation. § RELATED WORK BLE is widely used in mobile contact tracing apps currently promoted by health authorities worldwide. To unify and support contact tracing apps, the Google/Apple API for exposure notification based on BLE has been proposed for iOS[https://developer.apple.com/documentation/exposurenotification, accessed November 17th, 2020] and Android[https://developers.google.com/android/exposure-notifications/exposure-notifications-api, accessed November 17th, 2020] platforms. These APIs notify users about exposures to infected people according to commonly agreed thresholds (too close, too long) while preserving privacy to a high degree. The Google/Apple notification API has been widely adopted in contact tracing apps [2]. Yet, BLE-based distance estimation is error-prone. The study presented in [5] shows that staged contacts in a tram (less than 2 m, longer than 15 minutes) do not lead to expected notifications of the API when using the contact tracing app of Switzerland or Germany, and only 50% of exposures are detected by the Italian app. Thus, it is crucial to study and improve the effectiveness of BLE for contact tracing in real environments. The use of BLE for distance estimation and indoor localization has been thoroughly studied in the past, see, e.g. [6, 7, 8, 9]. These studies focus on communication of devices with no or little human interaction. As the contact tracing use case requires to capture realistic use scenarios of smartphones, those studies need to be extended. Major contact tracing data-sets and data repositories are provided by the MIT PACT project [10] and the DP-3T initiative [11]. Closest to the research presented in this paper is the study of BLE RSSI based distance estimation presented in [3], where several isolated effects such as body shielding or complex real-world situations are investigated. In our study, we will add a new dataset (joint effect of body shielding and multi-path), study model options to derive distance estimates based on BLE RSSI values, and extend the investigation by providing UWB-based measurements in addition to BLE RSSI logs. § TESTBED IMPLEMENTATION The measurement testbed comprises off-the-shelf smartphones running a mobile measurement app that has been developed for the purpose of evaluating contact tracing technologies. The app captures BLE RSSI values provided by the onboard BLE module and UWB distance measurements retrieved from an external module. Fig. <ref> visualizes the testbed and its use. §.§ Testbed Hardware and Software The testbed smartphones are of type Samsung Galaxy S7 and Samsung XCover 4s (Android version 8 and 9, Android BLE API). We do not make use of the Google/Apple API as the aim of the study is to investigate lower-level information provided by the BLE module. Via USB port, the smartphones connect to the UWB sensors [12]. The core component of the sensor is the Decawave DW1000 UWB transceiver chip, which is widely used for indoor localization. (a) (b) Measurement equipment: (a) smartphones connected to UWB sensors, (b) UWB sensor mounted on the left arm of a person, smartphone held in front of trunk. The measurement app is a native Android app that provides a user interface to control an experiment and to log BLE RSSI and UWB distance values. The app logs last recent scan values at configurable time intervals; 1 s is the default setting. Upon start, the app sends BLE advertisements based on the Generic Attribute Profile (GATT). When enabled by the user, a BLE scan is started resulting in an asynchronous return of newly scanned devices in BLE range. To restrict the devices reported in the scan, filtering by universally unique identifier (UUID) is employed. §.§ BLE RSSI Transformation BLE RSSI values are recorded on the smartphones as obtained through the getRSSI() Android function, which returns the received signal strength $P_{RX,i}$ in dBm, limited to the range of $[-127, 126]$ dBm. Note that in the practical implementation, the lowest recorded RSSI value was -105$\,$dBm. From the RSSI values, the pathloss (PL) is calculated on the receiving device using calibration values from the GAEN database[https://developers.google.com/android/exposure-notifications/files/en-calibration-2020-08-12.csv] as \begin{equation} \text{PL} = P_{\text{TX},j} - (P_{\text{RX},i} + \Delta_{\text{RX},i})\, , \label{eq:BLEcalibration} \end{equation} where $\Delta_{\text{RX},i}$ denotes the RX calibration value for scanning device $i$ and $P_{\text{TX},j}$ the calibrated TX power level of advertising device $j$. For the used Samsung XCover these are $P_{\text{TX}} = -24\,$dBm and $\Delta_{\text{RX}} = 5$ dB, and for the Samsung S7 $P_{\text{TX}} = -33\,$dBm and $\Delta_{\text{RX}} = 10$ dB. §.§ BLE Distance Estimation and Exposure Detection To detect exposures, we apply a distance based threshold detector. Thereby, we first obtain a model-based distance estimate $\hat{d}$ from a pathloss measurement. If the estimated distance is below the threshold distance, the data point is marked as positive, and otherwise negative. The model-based exposure estimation makes use of a linear pathloss model and a logarithmic pathloss model corresponding to known free space signal propagation properties. The linear pathloss model is chosen to describe a basic correlation of RSSI values with ground truth distance and has no physical interpretation. It is given by \begin{equation} \text{PL}_\text{lin} = \text{PL}_\text{0,lin} + k\, d\, , \label{eq:model_lin} \end{equation} where $\text{PL}_\text{0,lin}$ in dB is the pathloss at distance $d=0$ and $k$ is the slope in dB/m. The distance is then estimated by \begin{equation} \hat{d}_\text{lin} = \text{max}\Big( \frac{\text{PL} - \text{PL}_\text{0,lin}}{k}, 0 \Big)\, ,\label{eq:estimation_lin} \end{equation} where the maximum function $\text{max}(\cdot,0)$ avoids the estimation of negative distances. The log-distance pathloss model [6] is given by \begin{equation} \text{PL}_\text{ld} = \text{PL}_\text{0,ld} + 10\, \gamma \, \text{log}10\Big(\frac{d}{d_0}\Big) + X_g \, , \label{eq:model_ld} \end{equation} where $\text{PL}_\text{0,ld}$ in dB is the pathloss at reference distance $d_0$, $\gamma$ is the pathloss exponent and $X_g$ is the zero-mean Gaussian noise in dB. In this work, we have chosen $d_0 = 2\,$m as it reflects the COVID-19 distance of interest. The estimated distance is then given by \begin{equation} \hat{d}_\text{ld} = d_0 \, 10^{\frac{\text{PL} - \text{PL}_\text{0,ld}}{10\,\gamma}} \,. \label{eq:estimation_ld} \end{equation} §.§ UWB Distance Estimation The UWB modules retrieve a distance estimate by time-of-flight (ToF) calculation, obtained from timestamps that are recorded upon packet transmission and reception. As the clocks of the receiver and the transmitter are not synchronized, the ToF can only be estimated. The most common estimation approach double-sided two-way ranging (DS-TWR) [4], which requires TX and RX time stamps of three packet exchanges. We implement an extension to DS-TWR, known as cooperative synchronization and ranging [13]. The distance estimates are updated every 250$\,$ms. § MEASUREMENT STUDY In our measurement study, we aim at quantifying the accuracy of our proposed distance estimators (linear model, logarithmic model) based on BLE RSSI values. As BLE signal propagation is known to be effected by body shielding and multi-path propagation [14], we will in particular study these effects. §.§ Experiment Setup Two test persons are each equipped with a smartphone (Samsung xCover, Samsung Galaxy S7) and connected UWB sensor that is mounted on the left upper arm of the person. The experiments are recorded by the measurement app. The following properties are varied in our experiments: Carrying position: The two test persons are always carrying the smartphone at the same position, which is either (i) head – the smartphone is held at the left ear, (ii) trunk – the smartphone is held in front of the trunk, or (iii) pelvis – the smartphone is carried in the left front trouser's pocket. Distance $d$: The distance between the two persons is varied from 1$\,$m to 6$\,$m in steps of Environment: An anechoic chamber, a corridor, and a corner are selected, as depicted in Fig. <ref> (a)-(c). Orientation: The relative orientation between the persons can be 0$^{\circ}$, 90$^{\circ}$, 180$^{\circ}$, or 270$^{\circ}$ for head and pelvis carrying positions, and 0$^{\circ}$or 180$^{\circ}$ for trunk; see Fig. <ref> (d). Overall, the experiment consists of 180 combinations, each setting is measured with a duration of 3 min. The BLE pathloss of the whole experiment is visualized in Fig. <ref> and ranges from 22 dB to 70 dB. In addition, the fitted linear model in (<ref>), here referred to as 'lin', and the fitted logarithmic model in (<ref>), referred to as 'l-d', are depicted. The correlation coefficient of the linear approximation is $r = 0.51$, which is a weak correlation between distance and pathloss caused by the high variability of pathloss at each distance. The noise $X_g$ standard deviation is high with $\sigma = 8.48$ dB. Schematics of environments: (a) anechoic chamber, (b) corridor, (c) corner. Locations of Person 1 and Person 2 are depicted by blue and red circles, respectively, at six timesteps (at each location, each person is $d/2$ away from 'x', the start location), and (d) relative orientations. Heat map of BLE pathloss at six distances, grouped by three environments (corner, corridor, anechoic chamber). The model parameters of the linear model are: $\text{PL}_{0,\text{lin}}=\arrayij{mydatalin}{1}{1}\,$dB, $k=\arrayij{mydatalin}{2}{1}\,$dB/m with a correlation coefficient of $r = \arrayij{mydatalin}{3}{1}$. The parameters of the log-distance model are: $\text{PL}_{0,\text{ld}}=\arrayij{mydatald}{1}{1}\,$dB for reference distance $d_0=2.5\,$m, $\gamma=\arrayij{mydatald}{2}{1}$, $X_g$ noise standard deviation $\sigma = \arrayij{mydatald}{3}{1}\,$dB. BLE mean pathloss over actual distance (mean std. deviation is 3.90$\,$dB). The black dashed and dotted lines depict the linear and, respectively, the log-distance fit, per environment and carrying position. Distance estimation error for (a) BLE with known carrying position and unknown environment, (b) BLE with unknown carrying position and known environment, and (c) UWB. The boxplots depict the median, 0.25 and 0.75 quantile and the corresponding whiskers. §.§ Distance and Exposure Estimation Accuracy The distance estimation accuracy is described by the root-mean-square error (RMSE) of the distance estimate and the ground truth distance. The exposure detector is configured with a threshold of 2.5 m, meaning that an exposure is detected when the estimated distance is below this threshold. The detector is evaluated by the true positive rate $r_p$, the true negative rate $r_n$, the $F_1$ score and the Matthews correlation coefficient (MCC).[The MCC is in the interval $[-1,1]$, where values >0 indicate a performance better than a random guess.] Four scenarios are studied to assess how awareness of the environment and/or the carrying position influences estimation accuracy. For evaluation, the dataset is first divided along the known context settings (e.g., anechoic, corridor, corner), then each class is split equally into a training set and a test set. The training data is used to derive individual pathloss models for both model types (lin, l-d) considering censored pathloss measurements [15]. Comparing the case without context awareness (unknown/unknown) with the case of full awareness (known/known) in Tab. <ref>, the detection rates $r_p$ and $r_n$ increase; hereby the impact of knowing the carrying position is larger than the impact of knowing the environment. This effect is also described by the MCC. Note that the MCC is always above zero, which indicates that the estimate is always better than a random guess. The highest observed accuracy is MCC$=0.47$ and $F_1=0.7$. Accuracy of Exposure Estimation §.§ Effect of Body Shielding To isolate the effect of body shielding, we now discuss the BLE attenuation in the anechoic chamber as visualized in Fig. <ref>, left column. The phone's carrying position and orientation have a major influence on the pathloss. In LOS scenarios (head, 180$^{\circ}$ and 270$^{\circ}$; trunk, 0$^{\circ}$; pelvis, 0$^{\circ}$) the mean pathloss is always lower than in the other non-LOS scenarios, at the respective same distance. The spread of pathloss is high in all carrying positions, i.e., up to 7.11 dB (head, $d=1$ m), 11.69 dB (trunk, $d=1$ m), and 6.53 dB (pelvis, $d=1$ m). The spread due to body shielding is often higher than the distance-dependent increase of attenuation. This is confirmed by the weak linear correlation coefficients between pathloss and distance (lin model), which are $r=0.4$ (head), $r=0.5$ (trunk), and $r=0.36$ (pelvis). (The l-d model shows a similar behavior.) These results show that in particular in short ranges important for contact tracing (up to 2m), accurate distance estimation cannot be expected. The largest spread is found for head scenarios. For example, at a distance of 1 m, the pathloss is 42$\,$dB and 44$\,$dB under LOS conditions (180$^{\circ}$ and 270$^{\circ}$) and when fully blocked by the head, it is 57$\,$dB (90$^{\circ}$). It is worth noting that with increasing distance this difference is decreasing due to a mean pathloss saturation at approximately 64$\,$dB (BLE packets with higher pathloss are lost). §.§ Effect of Multi-path Propagation Multi-path propagation is usually thought to hinder precise distance measurements. However, comparing the pathloss in a multi-path propagation environment (corridor, corner) to the pathloss in the anechoic chamber, the effect of body shielding is less severe due to reflections from the walls, as expressed by the smaller spread of pathloss at a given distance (see Fig. <ref>). The linear correlation coefficients between distance and pathloss (lin model) reflect this effect as well: $0.35 \leq r \leq 0.56$ (anechoic), $0.46 \leq r \leq 0.6$ (corridor), and $0.77 \leq r \leq 0.83$ (corner). These results indicate that in environments with multi-path signal propagation, distance estimation based on BLE pathloss may be possible without knowing the carrying position and orientation, which is not the case for non multi-path environments such as outdoor environments. The highest correlation between distance and pathloss is observed in the corner scenario. At lower distances $d < 2$ m, a low pathloss below 42 dB is measured, which increases strongly with larger distances. The corner itself is a major cause of this behavior as at smaller distances LOS conditions are given. The corner obstructs the LOS path of the signal at larger distances. §.§ Distance Estimation Error Fig. <ref>(a) and (b) visualize the error statistics of BLE (lin and l-d models, individually fit to the scenario (head, trunk, pelvis and anechoic, corridor, corner)), and (c) the error statistics of UWB. For UWB, an overall distance RMSE of $0.9\,$m is observed, while BLE shows a RMSE of $3\,$m. This makes UWB a model technology w.r.t. its accuracy. Further, UWB may be used to collect ground truth measurements in realistic scenarios. UWB measurements yield precise distance estimates for all experiment settings that allow LOS propagation. The signal of the LOS path is correctly recognized by the UWB transceiver amongst other reflected paths. Notably, in the anechoic chamber either the signal is strong enough to propagate through particular body parts or no signal is received at all. Wall penetration of the UWB signal may lead to small error-prone distance estimates (corridor, corner). For example, see the negative distance errors in the corner environment. (Note that due to mounting the UWB sensors on the arm, position and orientation is not an issue.) Concerning BLE distance estimation errors, Fig. <ref>(a) and (b) show that the spread of the distance errors increases with distance, irrespective of the context. The l-d model also shows a larger spread of values than the lin model in most § CONCLUSIONS To assess mobile contact tracing technology, we introduced a flexible measurement smartphone app capable to capture BLE RSSI and UWB time-of-flight distance measurements. Our experimental results reveal that BLE-based estimation of distance is sensitive to carrying positions and distance estimation. Distance estimation is not accurate but exposure detection is feasible (distances below 2.5 m). Remarkably, multi-path propagation can reduce the effect of body shielding which may be leveraged in indoor environments where reflections from walls occur. We further showed that exposure detection can be best improved, if knowledge about the carrying position can be inferred. In future work, we collect extended measurement sets. We plan to enhance the capabilities of BLE-based distance estimation by further analysis of RSSI patterns over time to come closer to the estimation accuracy of an accurate measurement technology such as UWB. [1] N. Haug, L. Geyrhofer, A. Londei, A. Desvars-Larrive, V. Loreto, B. Pinior, S. Thurnher, and P. Klimek, “Ranking the effectiveness of worldwide covid-19 government interventions,” Nature Human Behaviour, 2020. [2] T. Martin, G. Karopoulos, J. L. Hernández-Ramos, G. Kambourakis, and I. N. Fovino, “Demystifying covid-19 digital contact tracing: A survey on frameworks and mobile apps,” Wireless Commun. Mobile Computing, pp. 1–29, Oct. 2020. [3] D. J. Leith and S. Farrell, “Coronavirus contact tracing: Evaluating the potential of using bluetooth received signal strength for proximity detection,” SIGCOMM Comput. Commun. Rev., vol. 50, pp. 66–74, Oct. [4] D. Neirynck, E. Luk, and M. McLaughlin, “An alternative double-sided two-way ranging method,” in 13th Workshop Pos., Navig. Commun. (WPNC), pp. 1–4, IEEE, 2016. [5] D. J. Leith and S. Farrell, “Measurement-based evaluation of google/apple exposure notification api for proximity detection in a light-rail tram,” PLoS ONE, no. 9, 2020. [6] K. Benkic, M. Malajner, P. Planinsic, and Z. Cucej, “Using rssi value for distance estimation in wireless sensor networks based on zigbee,” in 2008 15th International Conference on Systems, Signals and Image Processing, pp. 303–306, IEEE, 2008. [7] X. Zhao, Z. Xiao, A. Markham, N. Trigoni, and Y. Ren, “Does BTLE measure up against WiFi? a comparison of indoor location performance,” in European Wireless 2014; 20th European Wireless Conference, pp. 1–6, VDE, [8] F. Touvat, J. Poujaud, and N. Noury, “Indoor localization with wearable rf devices in 868mhz and 2.4 ghz bands,” in 2014 IEEE 16th International Conference on e-Health Networking, Applications and Services (Healthcom), pp. 136–139, IEEE, 2014. [9] S. Sadowski and P. Spachos, “RSSI-based indoor localization with the internet of things,” IEEE Access, vol. 6, pp. 30149–30161, 2018. [10] C. Corey, “PACT datasets and evaluation website.” <https://github.com/mitll/PACT>, 2020. [11] DP-3T, “Bluetooth measurements.” <https://github.com/DP-3T/bt-measurements>, 2020. [12] B. Etzlinger, A. Ganhr̈, J. Karoliny, R. Hüttner, and A. Springer, “WSN implementation of cooperative localization,” in Proc. IEEE MTT-S Int. Conf. Microwaves and Intelligent Mobility, pp. 1–4, IEEE, 2020. to appear. [13] B. Etzlinger, H. Wymeersch, et al., “Synchronization and localization in wireless networks,” Foundations and Trends® in Signal Processing, vol. 12, no. 1, pp. 1–106, 2018. [14] S. Gezici, Z. Tian, G. B. Giannakis, H. Kobayashi, A. F. Molisch, H. V. Poor, and Z. Sahinoglu, “Localization via ultra-wideband radios: a look at positioning aspects for future sensor networks,” IEEE signal processing magazine, vol. 22, no. 4, pp. 70–84, 2005. [15] C. Gustafson, T. Abbas, D. Bolin, and F. Tufvesson, “Statistical modeling and estimation of censored pathloss data,” IEEE Wireless Communications Letters, vol. 4, no. 5, pp. 569–572, 2015.
# Evaluation Discrepancy Discovery: A Sentence Compression Case-study Yevgeniy Puzikov Ubiquitous Knowledge Processing Lab (UKP Lab), Department of Computer Science, Technical University of Darmstadt https://www.ukp.tu-darmstadt.de Research done during an internship at Bloomberg L.P., London, United Kingdom. ###### Abstract Reliable evaluation protocols are of utmost importance for reproducible NLP research. In this work, we show that sometimes neither metric nor conventional human evaluation is sufficient to draw conclusions about system performance. Using sentence compression as an example task, we demonstrate how a system can game a well-established dataset to achieve state-of-the-art results. In contrast with the results reported in previous work that _showed correlation_ between human judgements and metric scores, our manual analysis of state-of- the-art system outputs demonstrates that high metric scores _may only indicate a better fit to the data_ , but not better outputs, as perceived by humans. The prediction and error analysis files are publicly released. 111https://github.com/UKPLab/arxiv2021-evaluation-discrepancy-nsc ## 1 Introduction ### 1.1 Task description Sentence compression is a Natural Language Processing (NLP) task in which a system produces a concise summary of a given sentence, while preserving the grammaticality and the important content of the original input. Both abstractive [Cohn and Lapata, 2008, Rush et al., 2015] and extractive [Filippova and Altun, 2013, Filippova et al., 2015, Wang et al., 2017, Zhao et al., 2018] approaches have been proposed to tackle this problem. Most researchers have focused on the extractive methods, which treat this as a deletion-based task where each compression is a subsequence of tokens from its original sentence (Figure 1). Dickinson, who competed in triple jump at the 1936 Berlin Games, was also a bronze medalist in both the long jump and triple jump at the 1938 Empire Games. (a) Input sentence Dickinson competed in triple jump at the 1936 Berlin Games. (b) Reference compression Dickinson was a bronze medalist in the long jump and triple jump at the 1938 Empire Games. (c) Another possible compression Figure 1: Sentence compression example from the Google Dataset: an input sentence and a reference compression. The compression candidate at the bottom is also a valid one, but would score low, because the n-gram overlap with the reference is small. In the past few years several novel methods have been proposed to tackle the task of sentence compression. Most of these methods have been evaluated using the Google Dataset [Filippova and Altun, 2013] or its derivatives. Most authors present approaches that show better metric scores; a few of them also describe human evaluation experiments and show that the proposed methods outperform previous work. However, there has been a serious lack of analysis done on actual model predictions. In this work we show that metric scores obtained on the Google Dataset might be misleading; a closer look at model predictions reveals a considerable amount of noise, which renders the trained model predictions ungrammatical. Another problem is that valid system outputs which do not match the references are severely penalized. For example, a plausible compression for the introductory example we used above would be Dickinson was a bronze medalist in the long jump and triple jump at the 1938 Empire Games. However, with the established evaluation protocol, this compression would score very low because of the insignificant token overlap with the reference. We showed that evaluating a system on the Google Dataset is tricky, and even human evaluation done in the previous years could not detect the issues described in this chapter. To summarize, our contributions in this study are: * • We introduce a simple method of sentence compression that established new state-of-the-art results, as measured by common metrics. * • We design an experiment with a contrived system which achieved even higher scores, but produced less grammatical and less informative outputs. * • We show that this discrepancy may be attributed to the noise in the dataset. ## 2 Data Analysis In our experiments we use the Google Dataset introduced by ?) 222https://bit.ly/2ZvTK9z This dataset was constructed automatically by collecting English news articles from the Internet, treating the first sentence of each article as an uncompressed sentence and creating its extractive compression using a set of heuristics and the headline of the article. The dataset contains 200 000 training and 10 000 evaluation instances; the first 1 000 data points from the latter are commonly used as a test set and the remaining 9 000 as a development set. Exploratory data analysis showed that the distribution of the training data is highly skewed, which is not surprising though, given the nature of the data. (a) Sentence length (b) Reference length (c) Token length (d) Compression ratio (CR) values Figure 2: Data analysis of the Google Dataset: length distributions of sentences, ground-truth compressions and tokens, and the distribution of compression ratio values. The numbers to the right of each box denote median values. In order to remove outliers and fit the computation budget, we removed instances which contained sentences longer than 50 tokens and compressions longer than 17 tokens. We also removed examples with tokens longer than 15 characters, since those in most cases denoted website links. Finally, we excluded cases with a compression ratio of more than 0.85 — those rare cases in most cases were too long to qualify as compressions. Evaluation on the development and test sets was done without any data filtering. ## 3 BERT-based Sentence Compression Most modern deletion-based compression systems adopt either a tree-pruning, or a sequence labeling approach. The former uses syntactic information to navigate over a syntactic tree of a sentence and decide which parts of it to remove [Knight and Marcu, 2000, McDonald, 2006, Filippova and Altun, 2013]. With the advent of sequence-to-sequence models it became possible to skip the syntactic parsing step and solve the task directly, by processing a sentence one token at a time and making binary decisions as to whether to keep a token or delete it [Filippova et al., 2015, Wang et al., 2017, Zhao et al., 2018, Kamigaito and Okumura, 2020]. The advantages of such approaches include a lesser chance of introducing error propagation from incorrect parsing decisions, as well as higher training and inference speed. For a long time the space of sequence-to-sequence models has been dominated by different variants of Recurrent Neural Networks (RNN) [Rumelhart et al., 1986]. However, a more recent Transformer architecture [Vaswani et al., 2017] has shown very promising results in many NLP tasks. Given the success of Bidirectional Encoder Representations from Transformers (BERT) [Devlin et al., 2019], and the fact that there has been no empirical evaluation of its performance in sentence compression, we decided to fill this gap and find out how well BERT-based models would cope with the task. We used pretrained BERT-base-uncased model weights 333https://huggingface.co/bert-base-uncased provided by the HuggingFace library [Wolf et al., 2020], and implemented a simple BertUni model which encodes the source sentence $S=\\{w_{1},w_{2},\dots w_{n}\\}$ and produces a sequence of vectors $V=\\{v_{1},v_{2},\dots v_{n}\\},v_{i}\in\mathbb{R}^{h}$. Each vector is fed into a dense layer with a logistic function as a non-linear function to produce a score $s_{i}\in[0,1]$ (Figure 3). If $s_{i}\geq 0.5$, the model output is 1 (and 0, otherwise). Figure 3: Schematic of the BertUni architecture. A simple decision rule was used to make a binary prediction: $decision=\begin{cases}\text{1},&\text{if }\mathit{s}\geq 0.5,\\\ \text{0},&\text{otherwise}.\end{cases}$ ## 4 Experiments ### 4.1 Automatic Metric Evaluation For automatic evaluation of sentence compression systems, most researchers follow ?; ?) and use the following two metrics: * • F1-score: harmonic mean of the recall and precision in terms of tokens kept in the target and the generated compressions. * • Compression ratio (CR): the length of the compression divided over the original sentence length. The former metric shows how close the model outputs are to the references. The latter one is supposed to measure the compression effectiveness. To make our results comparable to previous work, in our experiments we followed the same convention. However, we would like to note that _measuring CR in sentence compression might be redundant_ for several reasons. The first reason comes from the fact that data-driven sentence compressors are likely to produce outputs with a compression ratio most commonly seen in the training data references. In other words, CR is less a property of a system and more a characteristic of the dataset. This is supported by the fact that most models reported in the literature have the same compression ratio (in the range of 0.38–0.43, see Table 1). Secondly, it is not even clear how to treat compression ratio values: is a CR of 0.4 better than a CR of 0.5? Intuitively, yes, because it means a more concise compression. However, the compression is really better only if it retained more valuable information from the source. On the other side, defining the notion of informativeness/importance in sentence compression (and document summarization, in general) is an open problem and currently is not measured automatically. This means that the CR metric is a very one-sided proxy, too crude to be used for real automatic evaluation without balancing it with some recall-oriented metric. To put the evaluation of our approach into better context, we compare it with the following systems. All systems predict a sequence of binary labels which decide which tokens to keep or remove from the input sentence. ##### LSTM. ?) use a three-layer uni-directional Long Short-term Memory (LSTM) network [Hochreiter and Schmidhuber, 1997] and pretrained word2vec [Mikolov et al., 2013] embeddings as input representations. For comparison, we use the results for the best configuration reported in the paper (LSTM-PAR-PRES). This system parses the input sentence into a dependency tree, encodes the tree structure and passes the aggregated feature representations to the decoder LSTM. Unlike our approach, this system relies on beam search at inference time. ##### BiLSTM. ?) build upon LSTM approach, but introduces several modifications. It employs a bi-directional LSTM encoder and enriches the feature representation with syntactic context. In addition, it uses Integer Linear Programming (ILP) methods to enforce explicit constraints on the syntactic structure and sentence length of the output. ##### Evaluator-LM. ?) uses a bi-directional RNN to encode the input sentence and predict a binary label for each input token. In addition to token embeddings, the network uses vector representations of part-of-speech (POS) tags and dependency relations. The system is trained using the REINFORCE algorithm [Williams, 1992], the reward signal comes from a pretrained syntax-based language model (LM). ##### SLAHAN. ?) propose a modular sequence-to-sequence model that consists of several components. The system encodes a sequence of tokens using a combination of pretrained embeddings (Glove [Pennington et al., 2014], ELMO [Peters et al., 2018], BERT) and parses the input into a dependency graph. Three attention modules are employed to encode the relations in the graph, their weighted sum is passed to a selective gate. The output of the latter forms an input to a LSTM decoder. Despite its simplicity, the proposed BERT-based approach achieved very competitive scores (Table 1). Model | F1$\uparrow$ | CR$\downarrow$ ---|---|--- Evaluator-LM | 0.851 | 0.39 BiLSTM | 0.800 | 0.43 LSTM | 0.820 | 0.38 SLAHAN | 0.855 | 0.407 BertUni | 0.857 $\pm 0.002$ | 0.413 $\pm 0.004$ BertUni (dev) | 0.860 $\pm 0.001$ | 0.418 $\pm 0.004$ Table 1: The performance of the BertUni model on the test portion of the Google Dataset, compared to recent approaches. The last row shows BertUni’s performance on the development set. Comparing single performance scores (and not score distributions) of neural approaches is meaningless, because training neural models is non-deterministic in many aspects and depends on random weight initialization, random shuffling of the training data for each epoch, applying random dropout masks [Reimers and Gurevych, 2017]. This makes it hard to compare the scores reported in previous works and our approach. To facilitate a fair comparison with future systems, we report the mean and standard deviation of the BertUni scores averaged across ten runs with different random seeds. In order to understand where BertUni fails and what we could potentially improve upon, we conducted manual error analysis of its predictions. ### 4.2 Error analysis The purpose of error analysis is to find weak spots of a system, from the point of view of human evaluation. In sentence compression, previous work typically analyzed system predictions of the first 200 sentences of the test set, using a 5-point Likert scale to assess annotators’ opinions of the compressions’ _readability_ and _informativeness_ [Filippova et al., 2015]. Since error analysis is used for further system improvement and test sets should be used only for final evaluation, we perform error analysis on the development set. In order to do that, we retrieved BertUni’s predictions on the 200 dev set sentences which received the lowest F1 scores and manually examined them. Note that those are not random samples; the reason why we chose worst predictions is because we know that the system performed poorly on them. As for the quality criteria, we had to make certain adjustments. ?) mention that readability _covers the grammatical correctness, comprehensibility and fluency of the output_ , while informativeness measures _the amount of important content preserved in the compression_. In our opinion, merging several criteria into one synthetic index is a bad idea, because annotators can’t easily decide on the exact facet of evaluation. Given that there already exists a problem of distinguishing fluency and grammaticality, adding both of them to assess readability seems to be a bad design decision. The problem is aggravated by the fact that readability as a text quality criterion is already used by NLP researchers for estimating the _text complexity_ from a reader’s point of view [Vajjala and Meurers, 2012, Štajner and Saggion, 2013, Venturi et al., 2015, De Clercq and Hoste, 2016]. This made us conclude that readability is another overloaded criterion. Instead, we chose _grammaticality_ as the first quality criterion. We manually analyzed BertUni predictions on the 200 aforementioned samples, trying to identify common error patterns. The results are presented below. ##### Grammaticality. Out of 200 compressions, 146 (73 %) were deemed to be grammatical. The errors in the remaining instances have been classified into several groups (marked with _G_ in Figure 4a). (a) BertUni errors (b) BertBi-TF errors (c) Ground-truth errors Figure 4: Number of errors made by the evaluated approaches on the 200 development set instances where BertUni achieved the lowest F1 scores, as well as errors found in ground-truth compressions. Error types marked with _G_ are _grammaticality_ flaws; the remaining ones are errors of _informativeness_. Most of them were cases where grammatical clauses miss linking words, are _stitched_ together, making the output ungrammatical, as in the following compressions: * • I ’m said It ’s not Kitty Pryde superhero is the leader of the X-Men . * • He first Postal Vote result can be announced before 10PM . Another large error category was _finish_ : the compression was grammatical until the last retained token, where the sentence ended abruptly, rendering the compression incomplete: * • Activision Blizzard has confirmed some new statistics for its games including . * • The South Sydney star had no case to . A few system outputs incorrectly started with a relative or demonstrative pronoun. This happened when the system failed to retain parts of main clause of the sentence (_rd-pron_): * • That shows young people rapping while flashing cash and a handgun in a public park . Finally, one output missed a verb which was essential for ensuring grammaticality (_verb-miss_): * • People giant waves crash against the railway line and buildings at Dawlish . ##### Informativeness. Out of 200 compressions, 105 (52.5 %) were deemed to be informative, the errors in the remaining instances have been classified into several groups (marked with _I_ in Figure 4a). Most of these erroneous cases were compressions which missed certain information that was needed for understanding the context (_info-miss_). For example: * • Dolly Bindra filed a case . * • Mount Hope became the third largest city . A smaller, but still a large group of compressions started with unresolved personal pronouns, which made it hard to understand the subject (_p-pron_): * • She hopes her album Britney Jean will inspire people . * • He should be allowed to work freely till proven guilty . In some cases, omitting the context caused a change in the meaning of the sentence (_mean-change_). For example: * • Reference: […] Aleksandar Vucic […] voiced hope that Germany will give even stronger support to Serbia […] * • System: Aleksandar spoke Germany will give stronger support to Serbia . A large number of both grammatical and informative compressions did not match references (_I2_). Interestingly enough, in some cases the system outputs were better then the references: * • Reference: We saw their two and raised to three. * • System: Newport beat Hartlepool 2 0 . * • Reference: Who joins for the remainder of the season subject . * • System: Watford have announced the signing of Lucas Neill . More examples of compression errors are provided in Section A.1. ## 5 Evaluation Discrepancy When assessing the sentence compressions, we needed to compare system outputs with references. Manual examination revealed that many references themselves were flawed. This, in turn, meant that noise is inherent to the Google Dataset, and metric-based improvements on this data are misleading. To corroborate this claim, we conducted two experiments: the first tested the capacity of a more accurate system to ignore the noise and output compressions of better quality. In the second, we verified whether the noise came from the ground-truth data and attempted to quantify it. At first, we decided to implement more complex models that could potentially achieve better scores. We attempted to improve the grammatical quality of BertUni compressions by using the history of model predictions for making more informed decisions. We impelented and tested models that use BERT-encoded lastly-retained tokens at each prediction step as an additional input to the model (prediction history), similar to n-gram language models. As a history, BertBi and BertTri used one and two previously predicted tokens, respectively. BertBiSS and BertTriSS were the same as BertBi and BertTri, but used scheduled sampling training scheme to mitigate the exposure bias issue [Bengio et al., 2015]. According to the metric evaluation results, none of the more complex models outperformed BertUni (Table 2). Model | F1$\uparrow$ | CR$\downarrow$ ---|---|--- BertUni | 0.860 $\pm 0.001$ | 0.418 $\pm 0.004$ BertBi | 0.849 $\pm 0.001$ | 0.423 $\pm 0.005$ BertBiSS | 0.840 $\pm 0.003$ | 0.370 $\pm 0.005$ BertTri | 0.847 $\pm 0.002$ | 0.423 $\pm 0.007$ BertTriSS | 0.843 $\pm 0.003$ | 0.382 $\pm 0.006$ BertBi-TF | 0.901 | 0.423 Table 2: BERT-based model variants’ performance on the development set (mean and standard deviation across ten random seed values). BertBi-TF was run only once, since it is a “cheating” model that is not meant to be used in production. We used an unrealistic scenario and artificially made it easier for the model to make correct predictions. We trained a BertBi-TF model which builds upon BertBi, but at prediction time for history instead of model predictions uses ground-truth labels 444We call this model BertBi-TF, since it builds upon BertBi, but uses teacher forcing (TF) both at training and prediction time.. The development set result of BertBi-TF was an F1 score of 0.901, a 4-point improvement over BertUni. We retrieved this model’s predictions for the same 200 dev set sentences used for the error analysis of BertUni outputs, and manually examined them. The usual evaluation practice is to draw samples randomly, in order to not give an advantage to any system and not to bias the evaluation. However, in this work we approached the problem from a system- development perspective and attempted to assess the comparative performance of the approaches in the _worst-case_ scenario. If such a comparison is biased, then only in favor of BertBi-TF, because the drawn samples were the worst ones for BertUni, not BertBi-TF. We view this as sanity step, a regression test to ensure that the newer version of the system performs at least as well as the baseline on the challenging cases. We assessed BertBi-TF outputs from the same aspects of _grammaticality_ and _informativeness_ , as described in Section 4.2. ##### Grammaticality. Out of 200 compressions, only 44 (22 %) were found to be grammatical; we classified the errors in the remaining instances into groups (marked with _G_ in Figure 4b). The first and most prevalent is the already mentioned _stitch_ group which comprises around 80 % of all grammatical errors: * • The program has received FBS college game 2014 season . * • Tskhinvali region with Russia . The remaining errors are faulty compression endings (_finish_): * • The fine has been described as a slap on the . * • P Chidambaram sought . ##### Informativeness. A similar situation was observed when assessing the compressions’ informativeness — only 41 (20.5 %) instances were considered as correct. The distribution of errors (marked with _I_ in Figure 4b) indicates that more than 80 % of cases miss information by omitting important words: * • Dickinson was a . * • Wynalda is mixing . A smaller fraction of errors was comprised by the cases with unresolved personal pronouns: * • He is an education . * • It would win 45 to 55 seats in Odisha . The remaining errors were the cases where the system compressions changed the semantics of the input: * • Sentence: 612 ABC Mornings intern Saskia Edwards hit the streets of Brisbane to find out what frustrates you about other people. * • System: Saskia frustrates people . More examples of BertBi-TF errors are provided in Section A.2. We counted the cases in which predictions of BertBi-TF had better or worse quality, compared to BertUni. In terms of informativeness, BertBi-TF improved 15 and worsened 78 instances; in terms of grammaticality, 115 instances were perceived as less grammatical, versus only 13 improved cases, which makes it clear that BertBi-TF makes many more mistakes than BertUni, despite the higher metric scores. In order to verify our findings, we examined the ground-truth compressions in more detail. Only 63 (31.5 %) of these compressions were both grammatical and informative. Figure 4c shows a visualization of the error type distribution. We provide examples of noisy ground-truth compressions in Section A.3. The abundant errors related to the use of pronouns in the compressions were predominantly caused by the fact that many instances contained ground-truth compressions with unresolved pronouns; cleaning the data would likely result in better outputs. The _stitch_ , _finish_ and _info_miss_ errors can be attributed to the fact that many references have missing information or artifacts remaining from the automatic procedure that was used to create these compressions [Filippova and Altun, 2013]. Resolving these issues may require more elaborate strategies, beyond simple text substitution. ## 6 Discussion In this study we advanced the state-of-the-art for the task of sentence compression, and achieved that by designing a simple, but effective sequence labeling system based on the Transformer neural network architecture. While the proposed approach achieved the highest scores reported in the research literature, the main message of the study is not a higher score — it is the idea that NLP system evaluation might need to go beyond simple comparison of metric scores with human judgements. We found that a higher-scoring system can produce worse-quality outputs. We further provided some empirical evidence that this issue is caused by the noise in the training data. We call this finding a _discrepancy discovery_ , because existent sentence compression work does not explain our results, based on the established evaluation practices. The research papers we analyzed present automatic and human evaluation statistics that seem to overlook the data quality issue. Of course, the approaches proposed so far could still produce high-quality sentence compressions, but the absence of error analysis plants a seed of doubt into the reader. In this work, we question not the reported results, but the principles of the conventional evaluation workflow. None of the examined research papers drew attention to the quality of the data, even though it is known that the dataset was constructed automatically, and therefore should contain noisy examples, which should affect the output quality of any data-driven system. Previous work also overlooked the use of the compression ratio which seems to be too simplistic to call it a metric that measures the compression effectiveness. Finally, the employed sentence compression evaluation protocols do not assume having multiple references. We did not go into much detail about this issue, but provided an illustration at the beginning of the paper (the Dickson example). The space of possible compressions in deletion-based sentence compression is bound by sentence length. But because the definition of importance is left out, the candidate space is very large. The existence of only one reference brings additional requirements for evaluation metrics to work, and commonly used n-gram overlap metrics clearly do not satisfy these requirements. ## 7 Conclusion The presented results show that system output analysis is indispensable when assessing the quality of NLP systems. It is a well-established fact that metric scores used for system development do not always reflect the actual quality of a system; usually this is revealed via human evaluation experiments. However, in our case study of automatic sentence compression we have discovered that they might not be sufficient. Further investigation is needed to make stronger claims; the study’s findings are yet to be confirmed for other datasets and, perhaps, tasks. ## Acknowledgments We thank Andy Almonte, Joshua Bambrick, Minh Duc Nguyen, Vittorio Selo, Umut Topkara and Minjie Xu for their support and insightful comments. ## References * [Bengio et al., 2015] Samy Bengio, Oriol Vinyals, Navdeep Jaitly, and Noam Shazeer. 2015\. Scheduled sampling for sequence prediction with recurrent neural networks. In Proceedings of the 28th International Conference on Neural Information Processing Systems - Volume 1, NIPS’15, page 1171–1179, Cambridge, MA, USA. MIT Press. * [Clarke and Lapata, 2006] James Clarke and Mirella Lapata. 2006\. Models for sentence compression: a comparison across domains, training requirements and evaluation measures. In Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics, pages 377–384, Sydney, Australia, July. Association for Computational Linguistics. * [Cohn and Lapata, 2008] Trevor Cohn and Mirella Lapata. 2008\. Sentence compression beyond word deletion. In Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008), pages 137–144, Manchester, UK, August. Coling 2008 Organizing Committee. * [De Clercq and Hoste, 2016] Orphée De Clercq and Véronique Hoste. 2016\. All mixed up? Finding the optimal feature set for general readability prediction and its application to English and Dutch. Computational Linguistics, 42(3):457–490, September. * [Devlin et al., 2019] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019\. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota, June. Association for Computational Linguistics. * [Filippova and Altun, 2013] Katja Filippova and Yasemin Altun. 2013\. Overcoming the lack of parallel data in sentence compression. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, pages 1481–1491, Seattle, Washington, USA, October. Association for Computational Linguistics. * [Filippova et al., 2015] Katja Filippova, Enrique Alfonseca, Carlos A. Colmenares, Lukasz Kaiser, and Oriol Vinyals. 2015\. Sentence compression by deletion with LSTMs. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 360–368, Lisbon, Portugal, September. Association for Computational Linguistics. * [Hochreiter and Schmidhuber, 1997] Sepp Hochreiter and Jürgen Schmidhuber. 1997\. Long Short-term Memory. Neural Computation, 9(8):1735–1780. * [Kamigaito and Okumura, 2020] Hidetaka Kamigaito and Manabu Okumura. 2020\. Syntactically look-ahead attention network for sentence compression. ArXiv, abs/2002.01145. * [Knight and Marcu, 2000] Kevin Knight and Daniel Marcu. 2000\. Statistics-based summarization - step one: Sentence compression. In Proceedings of the Seventeenth National Conference on Artificial Intelligence and Twelfth Conference on Innovative Applications of Artificial Intelligence, page 703–710. AAAI Press. * [McDonald, 2006] Ryan McDonald. 2006\. Discriminative sentence compression with soft syntactic evidence. In Proceedings of the 11th Conference of the European Chapter of the Association for Computational Linguistics, Trento, Italy, April. Association for Computational Linguistics. * [Mikolov et al., 2013] Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013\. Distributed representations of words and phrases and their compositionality. In C. J. C. Burges, L. Bottou, M. Welling, Z. Ghahramani, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems 26, pages 3111–3119. Curran Associates, Inc. * [Pennington et al., 2014] Jeffrey Pennington, Richard Socher, and Christopher Manning. 2014\. GloVe: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1532–1543, Doha, Qatar, October. Association for Computational Linguistics. * [Peters et al., 2018] Matthew E. Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, and Luke Zettlemoyer. 2018\. Deep contextualized word representations. CoRR, abs/1802.05365. * [Reimers and Gurevych, 2017] Nils Reimers and Iryna Gurevych. 2017\. Reporting Score Distributions Makes a Difference: Performance Study of LSTM-networks for Sequence Tagging. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 338–348, Copenhagen, Denmark. Association for Computational Linguistics. * [Rumelhart et al., 1986] David E. Rumelhart, Geoffrey E. Hinton, and Ronald J. Williams. 1986\. Learning representations by back-propagating errors. Nature, 323(6088):533–536, October. * [Rush et al., 2015] Alexander M. Rush, Sumit Chopra, and Jason Weston. 2015\. A neural attention model for abstractive sentence summarization. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 379–389, Lisbon, Portugal, September. Association for Computational Linguistics. * [Štajner and Saggion, 2013] Sanja Štajner and Horacio Saggion. 2013\. Readability indices for automatic evaluation of text simplification systems: a feasibility study for Spanish. In Proceedings of the Sixth International Joint Conference on Natural Language Processing, pages 374–382, Nagoya, Japan, October. Asian Federation of Natural Language Processing. * [Vajjala and Meurers, 2012] Sowmya Vajjala and Detmar Meurers. 2012\. On improving the accuracy of readability classification using insights from second language acquisition. In Proceedings of the Seventh Workshop on Building Educational Applications Using NLP, NAACL HLT ’12, page 163–173, USA. Association for Computational Linguistics. * [Vaswani et al., 2017] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017\. Attention is all you need. CoRR, abs/1706.03762. * [Venturi et al., 2015] Giulia Venturi, Tommaso Bellandi, Felice Dell’Orletta, and Simonetta Montemagni. 2015\. NLP–based readability assessment of health–related texts: a case study on Italian informed consent forms. In Proceedings of the Sixth International Workshop on Health Text Mining and Information Analysis, pages 131–141, Lisbon, Portugal, September. Association for Computational Linguistics. * [Wang et al., 2017] Liangguo Wang, Jing Jiang, Hai Leong Chieu, Chen Hui Ong, Dandan Song, and Lejian Liao. 2017\. Can syntax help? Improving an LSTM-based sentence compression model for new domains. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1385–1393, Vancouver, Canada, July. Association for Computational Linguistics. * [Williams, 1992] Ronald J. Williams. 1992\. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning, 8(3–4):229–256, May. * [Wolf et al., 2020] Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander M. Rush. 2020\. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online, October. Association for Computational Linguistics. * [Zhao et al., 2018] Yang Zhao, Zhiyuan Luo, and Akiko Aizawa. 2018\. A language model based evaluator for sentence compression. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 170–175, Melbourne, Australia, July. Association for Computational Linguistics. ## Appendix A Error Examples This section contains error examples of BertUni and BertBi-TF models, as well as errors in the gold standard. We publicly share the prediction and error analysis files at https://github.com/UKPLab/arxiv2021-evaluation-discrepancy- nsc. ##### Grammaticality. Error types: * • _finish_ : incomplete sentence with an abrupt ending, caused by omitting the last token(s); * • _stitch_ : grammatical clauses with missing linking words, as if they are stitched together, which renders the sentence ungrammatical; * • _rd-pron_ : incorrect sentence start with a relative or demonstrative pronoun; * • _verb-miss_ : missing a verb which is essential for ensuring the grammaticality of a sentence. ##### Informativeness. Error types: * • _info-miss_ : missing certain information that is needed for understanding the context; * • _p-pron_ : starting with an unresolved personal pronoun, which makes it hard to understand the subject; * • _mean-change_ : omitting the context causing a change of the meaning of the sentence. We also provide examples of alternative compressions which focus on parts of the input sentence, which are different from those present in the reference compressions. They are marked with an _I2_ label and are listed here, together with erroneous cases, because due to mismatches with references they lower the metric scores of the evaluated approaches. ### A.1 BertUni #### Grammatical Errors Table 3: Manual error analysis results: examples of _ungrammatical_ outputs of BertUni. Error | Sentence | Compression ---|---|--- finish | A young cow trying to grab a cooling drink from a river in Hampshire in the hot weather had to be rescued by firefighters when it got stuck in the mud. | A cow trying to grab a drink had rescued . | Japan and the US reaffirmed Monday during a meeting in Tokyo with visiting Under Secretary for Terrorism and Financial Intelligence David S Cohen their ongoing cooperation on sanctions against Iran, a Japanese government source said Tuesday. | Japan and the US reaffirmed . | 612 ABC Mornings intern Saskia Edwards hit the streets of Brisbane to find out what frustrates you about other people. | Saskia Edwards hit . | Google Helpouts, which was launched this week, is a service that allows you to pay for brief one-on-one webcam master classes with a range of experts in various fields. | Google Helpouts is . rd-pron | Some parents and others in Bessemer City are complaining about a YouTube video that shows young people rapping while flashing cash and a handgun in a public park. | That shows young people rapping while flashing cash and a handgun in a public park . | “In my district, already reeling from the shutdown of our largest private employer, the highest energy costs in the country, and reduced government revenues, this shutdown, if it continues any longer, can be the final nail in our economic coffin”, said Rep. Christensen. | This shutdown can be the final nail in our economic coffin . stitch | Our obsession with Sachin Tendulkar and records has made us lose perspective to such an extent that what should have been widely condemned is being conveniently ignored. | Should have been condemned is . | Direct Link Pisegna JR. Without will amoxicillin work for a uti this sweet little boy told his best friend that he loved him. | boy told his he loved . | Visitors will find a mixture of old and new at Silver Springs State Park, which opened Tuesday near Ocala. | will find and at Silver Springs State Park opened . | Twitter took the first step toward its IPO by filing with US regulators on Thursday, September 12, 2013. | Twitter took by filing with US regulators . | Arsenio Hall lost control of his brand new Porsche Cayenne S and crashed his car on Monday night in El Lay! | Arsenio Hall lost his and crashed his car in El Lay verb-miss | People watch giant waves crash against the already damaged railway line and buildings at Dawlish during storms in south west England February 8, 2014. | People giant waves crash against the railway line and buildings at Dawlish . #### Informativeness Errors Table 4: Manual error analysis results: examples of _uninformative_ outputs of BertUni. We also show examples of alternative compressions (_I2_) which deviate from ground-truth compressions, but cannot be considered as errors. Error | Sentence | Compression ---|---|--- info-miss | Buyers should beware, even though there was a safety recall on some GM cars those vehicles are still being sold on Craigslist. | Buyers should beware . | New results show some improvement in test scores and graduation rates for local students, but experts say there’s still more work to be done. | There ’s still more work to be done . | Counting of postal votes have already commenced while Elections Commissioner Mahinda Deshapriya stated that he first Postal Vote result can be announced before 10PM. | He first Postal Vote result can be announced before 10PM . | When President Obama was elected in 2008 for his first term, he made a presidential decision that he would not give up his blackberry. | He would not give up his blackberry . mean-change | On this week’s “Hostages” season 1, episode 13: “Fight or Flight,” Ellen reveals to Duncan that she will not kill the President but will help him get what he needs, as long as he gives her something in return. | Ellen reveals she will not kill the President . | Serbia’s First Deputy Prime Minister Aleksandar Vucic spoke in Germany with former German chancellor Helmut Kohl about Serbia’s path towards the EU and its economic recovery/ During the talks, Vucic highlighted the important role of German investors in Serbia and voiced hope that Germany will give even stronger support to Serbia in the realisation of its European goals. | Aleksandar spoke Germany will give stronger support to Serbia . p-pron | “Being a five-time champion, he knows how to handle pressure. Anand generally puts a lid on his emotions.” | He knows how to handle pressure | Chelsea manager Jose Mourinho has made light of the managerial instability under Roman Abramovich by admitting he’s trying to break his own record. | He ’s trying to break his own record . | Katy Perry wasn’t lying when she said she had some “beautiful news to share,” because she is now the new face of COVERGIRL. | She had some beautiful news to share . I2 | Provisur Technologies has entered into an agreement with Scanico in which Scanico has become Provisur’s global partner in commercial freezing technology. | Reference: Scanico has become Provisur ’s global partner in commercial freezing technology . | | System: Provisur Technologies has entered into an agreement with Scanico . | Davina McCall has undergone medical tests after fears she may be suffering from hypothermia after battling severe weather in a Sport Relief challenge. | Reference: she may be suffering from hypothermia after battling in a Sport Relief challenge . | | System: Davina McCall has undergone medical tests . ### A.2 BertBi-TF #### Grammatical Errors Table 5: Manual error analysis results: examples of _ungrammatical_ outputs of BertBi-TF Error | Sentence | Compression ---|---|--- finish | A UKIP candidate who is standing for election in Enfield Town has defended a tweet in which he said a black comedian “should emigrate to a black country.” | UKIP candidate has defended a tweet he said a black comedian should | Wynalda, who was introduced as the Silverbacks’ new manager on Tuesday, is mixing a bit of Europe with a bit of Mexico with a bit of Silicon Valley in an approach that will eliminate the head-coaching position. | Wynalda is mixing . | Cell C has announced new pre-paid and contract packages that offer unlimited calls to any network. | Cell C has announced . | Radnor police received a report Sept. 3 from a cadet at Valley Forge Military Academy that another cadet struck him in the face. | Radnor police received a report another . | Diego Forlan scored directly from a corner this weekend to help Internacional to a 3-2 win over Fluminense. | Diego Forlan scored to help Internacional to . | This 1930s-built four bedroom detached seafront home in Worthing is immaculately presented and has been expertly modernised. | This seafront home is . stitch | Buyers should beware, even though there was a safety recall on some GM cars those vehicles are still being sold on Craigslist. | Buyers should beware are being sold on Craigslist . | Davina McCall has undergone medical tests after fears she may be suffering from hypothermia after battling severe weather in a Sport Relief challenge. | Davina McCall has undergone tests be suffering from hypothermia . | New results show some improvement in test scores and graduation rates for local students, but experts say there’s still more work to be done. | Show some improvement in test scores graduation rates students but there . | POLICE are looking for witnesses after a car was hit by a van which failed to stop on Friday, January 7. | POLICE a . | Our obsession with Sachin Tendulkar and records has made us lose perspective to such an extent that what should have been widely condemned is being conveniently ignored. | With Sachin Tendulkar has made lose perspective what . | Watford have this evening announced the signing of experienced defender Lucas Neill, who joins for the remainder of the season subject to international clearance. | Watford announced the signing Lucas season . . | | #### Informativeness Errors Table 6: Manual error analysis results: examples of _uninformative_ outputs of BertBi-TF. We also show examples of alternative compressions (_I2_) which deviate from ground-truth compressions, but cannot be considered as errors. Error | Sentence | Compression ---|---|--- info-miss | “Former President Mandela is still in a critical condition in hospital but shows sustained improvement,” President Jacob Zuma said in a statement. | Mandela is still in a critical condition shows improvement Jacob . | This 1930s-built four bedroom detached seafront home in Worthing is immaculately presented and has been expertly modernised. | This seafront home is . | ASI’s additional director general BR Mani said he was hopeful of Nalanda making it to the list, claiming that Nalanda was an important centre of art and culture even before the university came into being. | He Nalanda was . | Tata Martino explained that after two months in Barcelona, he is starting to understand how things work at the club, assuring that there seems to be a need to create a weekly crisis regarding the team’s style of play. | He crisis . mean-change | Provisur Technologies has entered into an agreement with Scanico in which Scanico has become Provisur’s global partner in commercial freezing technology. | Provisur has become Provisur ’s partner in commercial freezing technology . | TV presenter Katie Piper shows off her baby daughter for the first time and said motherhood has given her “true fulfillment”. | Piper shows off her baby daughter motherhood . | AFL powerbrokers yesterday met Collingwood officials over convicted criminal Marley Williams, who faces a possible club ban or even a league suspension. | Powerbrokers met ban suspension . | British rapper Professor Green was arrested on suspicion of drink driving on Sunday after he called police to report a robbery. | Professor Green was arrested on suspicion of report a robbery . p-pron | “Being a five-time champion, he knows how to handle pressure. Anand generally puts a lid on his emotions.” | He puts a lid on his emotions | PersonalWeb® today announced that it has reached a confidential settlement in its lawsuit with NEXSAN®, to which Nexsan has licensed the Truename patents from PersonalWeb. | It has reached patents PersonalWeb . | Britney Spears has said she hopes her latest album Britney Jean will inspire people and she wants to ’project positive energy out into the world. | She hopes her album Britney will inspire . I2 | On this week’s “Hostages” season 1, episode 13: “Fight or Flight,” Ellen reveals to Duncan that she will not kill the President but will help him get what he needs, as long as he gives her something in return. | Reference: On this week ’s Hostages season 1 episode 13 Fight or Flight Ellen reveals . | | System: Episode Fight or Flight Ellen reveals she will not kill the President . | Blustery winds arrived in Gwinnett on Wednesday and brought with them lower temperatures that caused the National Weather Service to issue a freeze warning for the area. | Reference: that caused the National Weather Service to issue a freeze warning . | | System: Blustery winds arrived issue a freeze warning . | BJP and JD today welcomed the five-year jail term handed down to RJD chief Lalu Prasad in the fodder scam case, saying it would send out a message that the law will catch up with the corrupt, however influential they might be. | Reference: The law will catch up with the corrupt influential . | | System: BJP law will catch up with the corrupt . ### A.3 Ground Truth #### Grammatical Errors Table 7: Manual error analysis results: examples of _grammatical errors_ in ground-truth compressions, sampled from 200 development set instances with lowest BertUni F1 scores). Error | Sentence | Compression ---|---|--- finish | Police investigating the unexplained death of a man in Taupo say his van appears to have broken down. | Police investigating the unexplained death say . | Akkineni Nageswara Rao was one of the Indian cinema’s stalwarts, who will be remembered for his rich contribution. | Akkineni Nageswara Rao was one . | Mortgage fees are going up so where does Pa. | Where does Pa . | Coffee chain Starbucks has said guns are no longer “welcome” in its US cafes, although it has stopped short of an outright ban. | Starbucks has said guns are . rd-pron | Way back in May 2011, Google filed a patent application for eye tracking technology, which would allow it to charge advertisers on a ’pay per gaze’ basis. | Which would allow it to charge advertisers on a pay per gaze basis . | POLICE are looking for witnesses after a car was hit by a van which failed to stop on Friday, January 7. | Which failed to stop . | Tomorrow South Africa will celebrate the centenary of the Union Buildings in Pretoria which have recently been declared a national heritage site by the South African Heritage Resources Agency. | Which have been declared a national heritage site . | In a press release, Patrick said Goldstein will be replaced by Rachel Kaprielian, who is currently the state’s registrar of motor vehicles. | Who is the state ’s registrar . stitch | Iran wants to end the stand-off with global powers over its nuclear programme swiftly, but will not sacrifice its rights or interests for the sake of a solution, President Hassan Rouhani said on Friday. | Iran wants but will not sacrifice its rights Hassan Rouhani said . | Maggie Rose sheds her innocence in her brand new music video for “Looking Back Now.” | Maggie Rose sheds for Looking Back Now | The Muskingum University chapter of Omicron Delta Kappa has made a donation of more than $600 to the New Concord Food Pantry in an effort to give back to the community. | The Muskingum University chapter of Omicron Delta Kappa has made in an effort to give back to the community . | Dolly Bindra filed a case on an unknown person for having threatened her at gun point today in Oshiwara, Mumbai. | Dolly Bindra filed for having threatened her at gun point in Oshiwara Mumbai . | North Korean leader Kim Jong-un has met with the top military leaders and warned them of a grave situation and threatened a new nuclear test. | Kim Jong-un has met and warned of a grave situation and threatened a nuclear test . #### Informativeness Errors Table 8: Manual error analysis results: examples of _informativeness_ errors in ground-truth compressions, sampled from 200 development set instances with lowest BertUni F1 scores). Error | Sentence | Compression ---|---|--- info-miss | Some parents and others in Bessemer City are complaining about a YouTube video that shows young people rapping while flashing cash and a handgun in a public park. | Some parents in Bessemer City are complaining about a video . | Tata Martino explained that after two months in Barcelona, he is starting to understand how things work at the club, assuring that there seems to be a need to create a weekly crisis regarding the team’s style of play. | There seems to be a need to create a weekly crisis . | Nothing is ever left behind in a BREACHED performance as the loud rocking, heavy amp cranky band announce Toronto show dates since performing last October at Indie Week. | Band announce Toronto show dates | Prime Minister Kevin Rudd has missed the deadline for an August 24 election, with his deputy saying “people should just chill out” about the election date. | People should chill out about the election date . p-pron | Davina McCall has undergone medical tests after fears she may be suffering from hypothermia after battling severe weather in a Sport Relief challenge. | She may be suffering from hypothermia after battling in a Sport Relief challenge . | TV presenter Nick Knowles has been the recipient of some unexpected abuse as a result of an announcement that he will not be present at the birth of his child. | He will not be present at the birth . | England fast bowler James Anderson does not feel sorry for Australia and has said his team wants to win the Ashes 5-0. | His team wants to win the Ashes 5 0 . | If he decides to run for president, New Jersey Gov. Chris Christie will need to push back against the inevitable pressure that he will encounter to move to the right. | He will encounter to move to the right . | Armaan will be taken for a medical examination and post that he will be presented in the court today. | He will be presented in the court .
# Repeated randomized algorithm for the Multicovering Problem Abbass Gorgi<EMAIL_ADDRESS>Mourad El Ouali<EMAIL_ADDRESS>Anand Srivastav<EMAIL_ADDRESS>Mohamed Hachimi <EMAIL_ADDRESS>Engineering Science Laboratory, University Ibn Zohr, Agadir, Morocco Department of Computer Science, Christian Albrechts University, Kiel, Germany ###### Abstract Let $\mathcal{H}=(V,\mathcal{E})$ be a hypergraph with maximum edge size $\ell$ and maximum degree $\Delta$. For given numbers $b_{v}\in\mathbb{N}_{\geq 2}$, $v\in V$, a set multicover in $\mathcal{H}$ is a set of edges $C\subseteq\mathcal{E}$ such that every vertex $v$ in $V$ belongs to at least $b_{v}$ edges in $C$. set multicover is the problem of finding a minimum-cardinality set multicover. Peleg, Schechtman and Wool conjectured that unless $\mathcal{P}=\mathcal{NP}$, for any fixed $\Delta$ and $b:=\min_{v\in V}b_{v}$, no polynomial-time approximation algorithm for the set multicover problem has an approximation ratio less than $\delta:=\Delta-b+1$. Hence, it’s a challenge to know whether the problem of set multicover is not approximable within a ratio of $\beta\delta$ with a constant $\beta<1$. This paper proposes a repeated randomized algorithm for the set multicover problem combined with an initial deterministic threshold step. Boosting success by repeated trials, our algorithm yields an approximation ratio of $\max\left\\{\frac{15}{16}\delta,\left(1-\frac{(b-1)\exp\left(\frac{3\delta+1}{8}\right)}{72\ell}\right)\delta\right\\}$. The crucial fact is not only that our result improves over the approximation ratio presented by Srivastav et al (Algorithmica 2016) for any $\delta\geq 13$, but it’s more general since we set no restriction on the parameter $\ell$. Furthermore, we prove that it is NP-hard to approximate the set multicover problem on $\Delta$-regular hypergraphs within a factor of $(\delta-1-\epsilon)$. Moreover we show that the integrality gap for the set multicover problem is at least $\frac{\ln_{2}(n+1)}{2b}$, which for constant $b$ is $\Omega(\ln n)$. ###### keywords: Integer linear programs, hypergraphs, approximation algorithms, randomized rounding, set cover and set multicover. ## 1 Introduction This work was intended as an attempt to solve approximately the set multicover problem. A nice formulation of this problem may be given by the notion of hypergraphs. A hypergraph is a pair $\mathcal{H}=(V,\mathcal{E})$, where $V$ is a finite set and $\mathcal{E}\subseteq 2^{V}$ is a family of some subsets of $V$. We call the elements of $V$ vertices and the elements of $\mathcal{E}$ (hyper-)edges. Further, let $n:=|V|$, $m:=|{\cal E}|$. W.l.o.g. let the vertices be enumerated as $v_{1},v_{2},\dots,v_{n}$ and the edges as $E_{1},E_{2},\dots,E_{m}$. As usually the degree of a vertex $v$ (notation $d(v)$) is the number of hyperedges it appears in. Let $\Delta:=\max_{v\in V}d(v)$ be the maximum degree. Furthermore, if the degree of every vertex is exactly $\Delta$, then ${\cal H}$ is called $\Delta$-regular. We define the number of vertices of a hyperedge as its size. If the size of all hyperedges is exactly $\ell$, i.e., $\forall E\in\mathcal{E},\,|E|=\ell$, then $\mathcal{H}$ is $\ell$-uniform. Let $\mathbf{b}:=(b_{1},b_{2},\dots,b_{n})\in\mathbb{N}_{\geq 2}^{n}$ be given. If a vertex $v_{i}$, $i\in[n]$, is contained in at least $b_{i}$ edges of some subset $C\subseteq\mathcal{E}$, we say that the vertex $v_{i}$ is fully covered by $b_{i}$ edges in $C$. A set multicover in $\mathcal{H}$ is a set of edges $C\subseteq\mathcal{E}$ such that every vertex $v_{i}$ in $V$ is fully covered by $b_{i}$ edges in $C$. The set multicover problem is the task of finding a set multicover of minimum cardinality. Related Work. The set cover problem $(b=1)$ is known to be NP-hard [14] and has been intensively explored for decades. Several deterministic approximation algorithms are exhibited for this problem [1, 10, 12, 16], all with approximation ratios $\Delta$. Furthermore, Johnson [13] and Lovász [17] gave a greedy algorithm with performance ratio $H(\ell)$, where $H(\ell)=\sum_{i=1}^{\ell}\frac{1}{i}$ is the harmonic number. Notice that $H(\ell)\leq 1+\ln(\ell)$. For hypergraphs with bounded $\ell$, Duh and Fürer [4] used the technique called semi-local optimization, improving $H(\ell)$ to $H(\ell)-\frac{1}{2}$. Unlike the set cover problem, the case $b\geq 2$ of the set multicover problem is less known. Let us give a summary of the known approximability results. In paper [21], Vazirani using primal-dual schema extended the result of Lovász [17] for $b\geq 1$. Later Fujito et al. [9] improved the algorithm of Vazirani and achieved an approximation ratio of $H(\ell)-\frac{1}{6}$ for $\ell$ bounded. Hall and Hochbaum [11] achieved by a greedy algorithm based on LP duality an approximation ratio of $\Delta$. By a deterministic threshold algorithm Peleg, Schechtman, and Wool in 1997 [19, 20] improved this result and gave an approximation ratio of $\delta$. They were also the first to propose an approximation algorithm for the set multicover problem with approximation ratio below $\delta$, namely a randomized rounding algorithm with performance ratio $(1-(\frac{c}{n})^{\frac{1}{\delta}})\cdot\delta$ for a small constant $c>0$. However, their ratio is depending on $n$, and asymptotically tends to $\delta$. Furthermore Peleg, Schechtman and Wool conjectured that for any fixed $\Delta$ and $b:=\min_{i\in[n]}b_{i}$ the problem cannot be approximated by a ratio smaller than $\delta:=\Delta-b+1$ unless $\mathcal{P}=\mathcal{NP}$. Hence it remained an open problem whether an approximation ratio of $\beta\delta$ with $\beta<1$ constant can be proved. A randomized algorithm of hybrid type was later given by Srivastav et al [7]. Their algorithm achieves for hypergraphs with $l\in\mathcal{O}\left(\max\\{(nb)^{\frac{1}{5}},n^{\frac{1}{4}}\\}\right)$ an approximation ratio of $\left(1-\frac{11(\Delta-b)}{72l}\right)\cdot\delta$ with constant probability. Concerning the algorithmic complexity, the set multicover problem has still not been investigated. In contrast to the set cover problem, it is known that the problem is hard to approximate to within $\Delta-1-\epsilon$, unless $\mathcal{P}=\mathcal{NP}$ [2], and to within $\Delta-\epsilon$ under the UGC [15] for any fixed $\epsilon>0$. Unless $\mathcal{P}=\mathcal{NP}$ there is no $(1-\epsilon)\ln n$ approximation [8]. This motivated us to study this aspect of the problem. Our Results. The main contribution of our paper is the combination of a deterministic threshold-based algorithm with repeated randomized rounding steps. The idea is to algorithmically discard instances that can be handled deterministically in favor of instances for which we obtain a constant-factor approximation less than $\delta$ using a repeated randomized strategy. Our hybrid randomized algorithm is designed as a cascade of a deterministic and a repeated randomized rounding step followed by greedy repair if the randomized solution is not feasible. First, the relaxed problem of the set multicover problem is solved. The successive actions depend on the cardinality of a set of hyperedges that will be defined according to the relaxed problem output. Our algorithm is an extension of an example given in [5, 6, 7, 10, 11, 20] for the vertex cover, partial vertex cover and set multicover problem in graphs and hypergraphs. The methods used in this paper rely on an application of an extension of the Chernoff-Hoeffding bound theorem for sums of independent random variables and are based on estimating the variance of the summed random variables for invoking the Chebychev-Cantelli inequality. Our algorithm yields a performance ratio of $\max\left\\{\frac{15}{16}\delta,\left(1-\frac{(b-1)\exp\left(\frac{3\delta+1}{8}\right)}{72\ell}\right)\delta\right\\}$. This ratio means a constant factor of less than $\delta$ for many settings of the parameters $\delta$, $b$, and $\ell$. It is asymptotically better than the former approximation ratios due to Peleg et al. and Srivastav et al. Furthermore, using a reduction of the set cover problem on $\Delta$-regular hypergraphs to the set multicover problem on $\Delta+b-1$-regular hypergraphs, we show that it is NP-hard to approximate the set multicover problem on $\Delta$-regular hypergraphs within a factor of $(\delta-1-\epsilon)$. Moreover, we show that the integrality gap for the natural LP formulation of the set multicover problem is at least $\frac{\ln_{2}(n+1)}{2b}$, which for constant $b$ is $\Omega(\ln n)$. Fundamental results and approximations for set multicover problem Hypergraph | Approximation ratio ---|--- - | $H(\ell)$[21] bounded $\ell$ | $H(\ell)-\frac{1}{6}$ [9] - | $\delta$ [11, 20] - | $(1-(\frac{c}{n})^{\frac{1}{\delta}})\cdot\delta$ where $c>0$ is a constant. [19] $l\in\mathcal{O}\left(\max\\{(nb)^{\frac{1}{5}},n^{\frac{1}{4}}\\}\right)$ | $\left(1-\frac{11(\Delta-b)}{72\ell}\right)\cdot\delta$ [7] - | $\max\left\\{\frac{15}{16}\delta,\left(1-\frac{(b-1)\exp\left(\frac{3\delta+1}{8}\right)}{72\ell}\right)\delta\right\\}$ (this paper) Outline of the paper. In Section 2, we give all the definitions and the tools needed for our analysis. In Section 3, we present a randomized algorithm of hybrid type and its analysis. In Section 4, we give a lower bound for the problem. In Section 5, we discuss the integrality gap of the LP formulation of the problem. ## 2 Definitions and preliminaries For the later analysis we will use the following extension of Chernoff- Hoeffding Bound inequality for a sum of independent random variables. It is often used if one only has a bound on the expectation: ###### Theorem 1 (see [3]) Let $X_{1},\ldots,X_{n}$ be independent $\\{0,1\\}$-random variables. Let $X=\sum_{i=1}^{n}X_{i}$ and suppose $\mathbb{E}(X)<\mu$. For every $0<\beta\leq 1$ we have $\Pr[X\geq(1+\beta)\mu]\leq\exp{\left(-\frac{\beta^{2}\mu}{3}\right)}.$ A further useful concentration theorem we will use is the Chebychev-Cantelli inequality: ###### Theorem 2 (see [18], page 64) Let $X$ be a non-negative random variable with finite mean $\mathbb{E}(X)$ and variance Var$(X)$. Then for any $a>0$ it holds that $\displaystyle\Pr(X\leq\mathbb{E}(X)-a)$ $\displaystyle\leq$ $\displaystyle\frac{{\rm Var}(X)}{{\rm Var}(X)+a^{2}}\cdot$ Our lower bound proof for the problem relies on extending the following theorem from the case of $b=1$ to the case of $b\geq 2$. ###### Theorem 3 (I. Dinur et al, 2005 [2]) For every integer $l\geq 3$ and every $\epsilon>0$, it is NP-hard to approximate the minimum vertex cover problem on $\ell$-uniform hypergraphs within a factor of $(\ell-1-\epsilon)$. A key notion of linear programming relaxations is the concept of Integrality Gap. ###### Definition 1 Let $\cal{I}$ be a set of instances, the Integrality Gap for minimization problems is defined as $\sup_{i\in\cal{I}}{\frac{{\rm Opt}(I)}{{\rm Opt}^{*}(I)}}.$ ## 3 The multi-randomized rounding algorithm Let ${\cal H}=(V,{\cal E})$ be a hypergraph with maximum vertex degree $\Delta$ and maximum edge size $\ell$. An integer linear programming formulation of the set multicover problem is the following: $\displaystyle\min\sum_{j=1}^{m}x_{j},$ $\displaystyle\mbox{ILP}(\Delta,{\bf b}):\qquad$ $\displaystyle\sum_{j=1}^{m}a_{ij}x_{j}\geq b_{i}\quad\mbox{ for all }i\in[n],$ $\displaystyle x_{j}\in\\{0,1\\}\quad\mbox{ for all }j\in[m],$ where $A=(a_{ij})_{i\in[n],\,j\in[m]}\in\\{0,1\\}^{n\times m}$ is the vertex- edge incidence matrix of ${\cal H}$ and ${\bf b}=(b_{1},b_{2},\dots,b_{n})\in\mathbb{N}_{\geq 2}^{n}$ is the given integer vector. For every vertex $v$, we define $\Gamma(v):=\\{E\in\mathcal{E}\mathrel{|}v\in E\\}$ the set of edges incident to $v$. The linear programming relaxation LP($\Delta,\,{\bf b}$) of ILP($\Delta,\,{\bf b}$) is given by relaxing the integrality constraints to $x_{j}\in[0,1]$ for all $j\in[m]$. Let $\mathrm{Opt}$ resp. ${\rm Opt}^{*}$ be the value of an optimal solution to ILP($\Delta,\,{\bf b}$) resp. LP($\Delta,\,{\bf b}$). Let $(x^{\ast}_{1},\ldots,x^{\ast}_{m})$ be the optimal solution of the LP($\Delta,\,{\bf b}$). So ${\rm Opt}^{*}=\sum_{j=1}^{m}x^{*}_{j}$ and ${\rm Opt}^{*}\leq\mathrm{Opt}$. The next lemma shows that the $b_{i}$ greatest values of the LP variables corresponding to the incident edges for any vertex $v_{i}$ are all greater than or equal to $\frac{1}{\delta}$. ###### Lemma 1 (see [20]) Let $b_{i},d,\Delta,n\in\mathbb{N}$ with $2\leqslant b_{i}\leqslant d-1\leqslant\Delta-1,i\in[n]$ . Let $x_{j}\in[0,1],j\in[d]$, such that $\displaystyle\sum_{j=1}^{d}x_{j}\geqslant b_{i}$. Then at least $b_{i}$ of the $x_{j}$ fulfill the inequality $x_{j}\geqslant\frac{1}{\delta}$. Our second lemma shows that the $b_{i}-1$ greatest values of the LP variables corresponding to the incident edges for any vertex $v_{i}$ are all greater than or equal to $\frac{2}{\delta+1}$ and with Lemma 1 we take the sum over the $b_{i}$ greatest values of the LP variables corresponding to the incident edges for any vertex $v_{i}$. ###### Lemma 2 Let $b_{i},d,\Delta,n\in\mathbb{N}$ with $2\leqslant b_{i}\leqslant d-1\leqslant\Delta-1,i\in[n]$ . Let $x_{j}\in[0,1],j\in[d]$, such that $\displaystyle\sum_{j=1}^{d}x_{j}\geqslant b_{i}$. Then at least $b_{i}-1$ of the $x_{j}$ fulfill the inequality $x_{j}\geqslant\frac{2}{\delta+1}$ and there exists an element $x_{j}$, distinct to all of them, that fulfills the inequality $x_{j}\geqslant\frac{1}{\delta}$ . ###### Proof 1 W.l.o.g. we suppose $x_{1}\geq x_{2}\geq\cdots\geq x_{b_{i}}\geq\cdots\geq x_{d}$. Hence $b_{i}-2\geq\displaystyle\sum_{j=1}^{b_{i}-2}x_{j}$ and $(d-b_{i}+2)x_{b_{i-1}}\geq\displaystyle\sum_{j=b_{i}-1}^{d}x_{j}$. Then $\displaystyle b_{i}-2+(\Delta-b+2)x_{b_{i}-1}$ $\displaystyle\geq$ $\displaystyle b_{i}-2+(\Delta-b_{i}+2)x_{b_{i}-1}$ $\displaystyle\geq$ $\displaystyle b_{i}-2+(d-b_{i}+2)x_{b_{i}-1}$ $\displaystyle\geq$ $\displaystyle\sum_{j=1}^{b_{i}-2}x_{j}+\sum_{j=b_{i}-1}^{d}x_{j}=\displaystyle\sum_{j=1}^{d}x_{j}$ $\displaystyle\geq$ $\displaystyle b_{i}$ So we have $x_{b_{i}-1}\geq\frac{2}{\delta+1}$. Since for all $j\in[b_{i}-1]\;,\ x_{j}\geq x_{b_{i}-1}$ then for all $j\in[b_{i}-1]\;,\;x_{j}\geq\frac{2}{\delta+1}$. Furthermore, by Lemma $1$ and the assumption on the orders of the variables $x_{j}$, for all $j\in[b_{i}]\;$ we have $x_{j}\geq\frac{1}{\delta}$ and particularly $x_{b_{i}}\geq\frac{1}{\delta}$. ### 3.1 The algorithm In this section we present an algorithm with conditioned randomized rounding based on the properties satisfied by two generated sets, $C_{1}$ and $C_{2}$. Input : A hypergraph $\mathcal{H}=(V,\,\mathcal{E})$ with maximum degree $\Delta$ and maximum hyperedge size $\ell$, numbers $b_{i}\in\mathbb{N}_{\geq 2}\text{ for }i\in[n]$, $b:=\min_{i\in[n]}b_{i}$, $\epsilon\in(0,1)$, a constant $k\in\mathbb{N}_{\geq 2}$ and $\delta=\Delta-b+1$. Output : A set multicover $C$ 1. 1. Initialize $C:=\emptyset$. Set $\lambda=\frac{\delta+1}{2}\;$, $\alpha=\frac{(b-1)\delta\epsilon^{k}}{6\ell}\times\exp\left(a_{k,\epsilon}\right)$ with $a_{k,\epsilon}=\frac{k(1-\epsilon)+(\delta-1)(1-\epsilon^{a})}{2}$ and $\lambda_{0}=(1-\epsilon)\delta$. 2. 2. Obtain an optimal solution $x^{*}\in[0,1]^{m}$ by solving the LP($\Delta,\,{\bf b}$) relaxation. 3. 13. Set $C_{1}:=\\{E_{j}\in\mathcal{E}\mathrel{|}x_{j}^{\ast}\geq\frac{1}{\lambda}\\}$, $\ C_{2}:=\\{E_{j}\in\mathcal{E}\mathrel{|}\frac{1}{\lambda}>x_{j}^{\ast}\geq\frac{1}{\delta}\\}$ and $C_{3}:=\\{E_{j}\in\mathcal{E}\mathrel{|}0<x_{j}^{\ast}<\frac{1}{\delta}\\}$. * 4. Take all edges of the set $C_{1}$ in the cover $C$. 25. if $|C_{1}|\geq\alpha\cdot\mathrm{Opt}^{*}$ then return $C=C_{1}\cup C_{2}$. Else (Multi-randomized Rounding) 1. 3(a) For all edges $E_{j}\in C_{2}$ include the edge $E_{j}$ in the cover $C$, independently for all such $E_{j}$, with probability $\lambda_{0}x_{j}^{*}$, $k$ times. 4 $\left(\text{ If, in any of these $k$ biased coin flips shows head, include the edge }E_{j}\text{ in the cover.}\right)$ * (b) For all edges $E_{j}\in C_{3}$ include the edge $E_{j}$ in the cover $C$, independently for all such $E_{j}$, with probability $(1-\epsilon^{k})\delta x_{j}^{*}$. (c) (Repairing) Repair the cover $C$ (if necessary) as follows: Include arbitrary edges from $C_{2}$, incident to the vertices $v_{i}$ not fully covered, to $C$ until all vertices are fully covered. (d) Return the cover $C$. Algorithm 1 SET MULTICOVER In step $2$ we solve the linear programming relaxation LP($\Delta,\,{\bf b}$) in polynomial time, using some known polynomial-time procedure, e.g. the interior point method. Next we take into the cover all edges of the sets $C_{1}$ resp. $C_{2}$. Since the LP variable value $x^{*}_{j}$ that corresponds to an edge $E_{j}$ from the set $C_{1}$ is greater than or equal to $\frac{2}{\delta+1}$ and the value $x^{*}_{j}$ that corresponds to an edge $E_{j}$ from the set $C_{2}$ is less than $\frac{2}{\delta+1}$, we have $|C_{1}|+|C_{2}|=|C|\text{\quad and\quad}C_{1}\cap C_{2}=\emptyset$ (1) ### 3.2 Analysis of the algorithm Case $\mathbf{|C_{1}|\geq\alpha\cdot\mathrm{Opt}^{*}}$. ###### Theorem 4 Let $\mathcal{H}$ be a hypergraph with maximum vertex degree $\Delta$ and maximum edge size $\ell$. Let $\alpha=\frac{(b-1)\delta\epsilon^{k}}{6\ell}\times\exp\left(a_{k,\epsilon}\right)$ with $a_{k,\epsilon}=\frac{k(1-\epsilon)+(\delta-1)(1-\epsilon^{a})}{2}$ as defined in Algorithm 1. If $|C_{1}|\geq\alpha\cdot\mathrm{Opt}^{*}$ then Algorithm 1 returns a set multicover $C$ such that $|C|<\left(1-\frac{(b-1)\epsilon^{k}}{18\ell}\times\exp\left(a_{k,\epsilon}\right)\right)\delta\cdot\mathrm{Opt}^{*}$ ###### Proof 2 The proof is straightforward, using the definitions of the sets $C_{1}$ and $C_{2}$. $\displaystyle\delta\mathrm{Opt}^{*}=\sum_{j=1}^{m}\delta x^{*}_{j}$ $\displaystyle\geq$ $\displaystyle\displaystyle\sum_{E_{j}\in C_{1}}\delta x^{*}_{j}+\sum_{E_{j}\in C_{2}}\delta x^{*}_{j}$ $\displaystyle\geq$ $\displaystyle\frac{2\delta}{\delta+1}|C_{1}|+|C_{2}|$ $\displaystyle\geq$ $\displaystyle\frac{2\delta}{\delta+1}|C_{1}|+\left(|C|-|C_{1}|\right)$ $\displaystyle\geq$ $\displaystyle\frac{\delta-1}{\delta+1}|C_{1}|+|C|$ $\displaystyle\overset{\delta\geq 2}{\geq}$ $\displaystyle\frac{1}{3}|C_{1}|+|C|$ $\displaystyle\geq$ $\displaystyle\frac{1}{3}\alpha\cdot\mathrm{Opt}^{*}+|C|.$ Hence $|C|\leq\left(1-\frac{(b-1)\epsilon^{k}}{18\ell}\times\exp\left(a_{k,\epsilon}\right)\right)\delta\cdot\mathrm{Opt}^{*}$ Case $\mathbf{|C_{1}|<\alpha\cdot\mathrm{Opt}^{*}}$. Let $X_{1},\ldots,X_{m}$ be $\\{0,1\\}$-random variables defined as follows: $\displaystyle X_{j}=\begin{cases}1&\text{if the edge}\,E_{j}\,\text{was picked into the cover before repairing}\\\ 0&\text{otherwise}.\end{cases}$ Note that the $X_{1},\ldots,X_{m}$ are independent for a given $x^{*}\in[0,1]^{m}$. For all $i\in[n]$ we define the $\\{0,1\\}$\- random variables $Y_{i}$ as follows: $\displaystyle Y_{i}=\begin{cases}1&\text{if the vertex}~{}v_{i}~{}\text{is fully covered before repairing}\\\ 0&\text{otherwise}.\end{cases}$ We denote by $X:=\sum_{j=1}^{m}X_{j}$ and $Y:=\sum_{i=1}^{n}Y_{i}$ the cardinality of the cover and the cardinality of the set of fully covered vertices before the step of repairing, respectively. At this step by Lemma 2, one more edge for each vertex is at most needed to be fully covered. The cover $C$ obtained by Algorithm 1 is bounded by $\left\lvert C\right\rvert\leq X+n-Y.$ (2) Our next lemma provides upper bounds on the expectation of the random variable $X$ and the expectation and variance of the random variable $Y$, which we will use to proof Theorem 5. This is a restriction of Lemma $4$ in [7] to the last case in Algorithm 1. ###### Lemma 3 Let $l$ and $\Delta$ be the maximum size of an edge and the maximum vertex degree, respectively. Let $\epsilon\in\left[\frac{\delta-1}{2\delta},\left(\frac{\delta-1}{2\delta}\right)^{\frac{1}{k}}\right]$, $a_{k,\epsilon}=\frac{k(1-\epsilon)+(\delta-1)(1-\epsilon^{a})}{2}$, $\lambda_{0}=(1-\epsilon)\delta$ and $\lambda=\frac{\delta+1}{2}$ as in Algorithm 1. We have $\mathrm{(i)}$ $\mathbb{E}(Y)\geq(1-\exp\left(-2a_{k,\epsilon}\right))n$. $\mathrm{(ii)}$ ${\rm Var}(Y)\leq 2n^{2}\exp\left(-2a_{k,\epsilon}\right)$. $\mathrm{(iii)}$ $\mathbb{E}(X)\leq(1-\epsilon^{k})\delta\mathrm{Opt}^{*}$. $\mathrm{(iv)}$ $\dfrac{(b-1)n}{\alpha\ell}<\mathrm{Opt}^{*}$. ###### Proof 3 (i) Let $i\in[n]$, $r=d(i)-b_{i}+1$. If $\left\lvert C_{1}\cap\Gamma(v_{i})\right\rvert\geq b_{i}$, then the vertex $v_{i}$ is fully covered and $\Pr(Y_{i}=0)=0$. Otherwise we get by Lemma 2 that $\left\lvert C_{1}\cap\Gamma(v_{i})\right\rvert=b_{i}-1$ and there exists at least one more edge from $C_{2}$ with $x_{j}\geq\frac{1}{\delta}$, so we have $\sum_{E_{j}\in\Gamma(v_{i})\cap C_{2}}x_{j}^{*}\geq\frac{1}{\delta}$ and by the inequality constraints it holds that $\sum_{E_{j}\in\Gamma(v_{i})\cap\left(C_{2}\cup C_{3}\right)}x_{j}^{*}\geq 1$. Therefore $\displaystyle\Pr(Y_{i}=0)$ $\displaystyle=$ $\displaystyle\left(\prod_{E_{j}\in\Gamma(v_{i})\cap C_{2}}(1-\lambda_{0}x_{j}^{*})\right)^{k}\prod_{E_{j}\in\Gamma(v_{i})\cap C_{3}}\left(1-(1-\epsilon^{k})\delta x_{j}^{*}\right)$ $\displaystyle=$ $\displaystyle\prod_{E_{j}\in\Gamma(v_{i})\cap C_{2}}\left(1-\lambda_{0}x_{j}^{*}\right)^{k}\prod_{E_{j}\in\Gamma(v_{i})\cap C_{3}}\left(1-(1-\epsilon^{k})\delta x_{j}^{*}\right)$ $\displaystyle\leq$ $\displaystyle\prod_{E_{j}\in\Gamma(v_{i})\cap C_{2}}\exp(-k\lambda_{0}x_{j}^{*})\prod_{E_{j}\in\Gamma(v_{i})\cap C_{3}}\exp(-(1-\epsilon^{k})\delta x_{j}^{*})$ $\displaystyle=$ $\displaystyle\exp\left(-k(1-\epsilon)\delta\sum_{E_{j}\in\Gamma(v_{i})\cap C_{2}}x_{j}^{*}\right)\cdot\exp\left(-(1-\epsilon^{k})\delta\sum_{E_{j}\in\Gamma(v_{i})\cap C_{3}}x_{j}^{*}\right)$ $\displaystyle=$ $\displaystyle\exp\left(\left(-k(1-\epsilon)+(1-\epsilon^{k})\right)\delta\sum_{E_{j}\in\Gamma(v_{i})\cap C_{2}}x_{j}^{*}\right)\cdot\exp\left(-(1-\epsilon^{k})\delta\sum_{E_{j}\in\Gamma(v_{i})\cap\left(C_{2}\cup C_{3}\right)}x_{j}^{*}\right).$ Since $1-\epsilon^{k}=(1-\epsilon)\sum_{i=0}^{k-1}\epsilon^{i}\leq k(1-\epsilon)$, we have $-k(1-\epsilon)+1-\epsilon^{k}\leq 0$. It follows that $\displaystyle\Pr(Y_{i}=0)$ $\displaystyle\leq$ $\displaystyle\exp\left(-k(1-\epsilon)+(1-\epsilon^{k})\right)\cdot\exp\left(-(1-\epsilon^{k})\delta\right)$ $\displaystyle=$ $\displaystyle\exp\left(-2a_{k,\epsilon}\right).$ Therefore $\displaystyle\mathbb{E}(Y)$ $\displaystyle=\sum_{i=1}^{n}\Pr(Y_{i}=1)=\sum_{i=1}^{n}(1-\Pr(Y_{i}=0))$ $\displaystyle\geq\sum_{i=1}^{n}(1-\exp\left(-2a_{k,\epsilon}\right))$ $\displaystyle\geq(1-\exp\left(-2a_{k,\epsilon}\right))n.$ (ii) Since $Y=\sum_{i=1}^{n}Y_{i}\leq n,$ we have $\mathbb{E}(Y^{2})\leq n^{2}.$ Thus, $\displaystyle{\rm Var}(Y)$ $\displaystyle=\mathbb{E}(Y^{2})-\mathbb{E}(Y)^{2}\leq n^{2}-(1-\exp\left(-2a_{k,\epsilon}\right))^{2}n^{2}$ $\displaystyle\leq n^{2}\left(1-(1-\exp\left(-2a_{k,\epsilon}\right))^{2}\right)$ $\displaystyle\leq 2n^{2}\exp\left(-2a_{k,\epsilon}\right).$ (iii) Let $E_{j}$ be an edge from $C_{2}$. By Lemma 2 we have $\frac{1}{\delta}\leq x^{*}_{j}<\frac{2}{\delta+1}$. Recall that we include independently the edge $E_{j}$ in the cover $C$, with probability $\lambda_{0}x_{j}^{*}$, $k$ times. Since $\frac{\delta-1}{2\delta}\leq\epsilon$, we have $1-\epsilon\leq\lambda_{0}x^{*}_{j}<\frac{2}{\delta+1}(1-\epsilon)\delta\leq\frac{2}{\delta+1}(1-\frac{\delta-1}{2\delta})\delta=1$. Furthermore with $\epsilon\leq\left(\frac{\delta-1}{2\delta}\right)^{\frac{1}{k}}$ we have $\left(1-\epsilon^{k}\right)\delta\geq\left(1-\frac{\delta-1}{2\delta}\right)\delta=\frac{\delta+1}{2}=\lambda$. Then $\lambda\leq\left(1-\epsilon^{k}\right)\delta.$ (3) Clearly $\Pr\left(X_{j}=1\right)=1-\left(1-\lambda_{0}x^{*}_{j}\right)^{k}$. Define the function $f$ by $f(x)=\frac{1-(1-x)^{k}}{x}$. $f$ is strictly decreasing on $(0,1]$. Therefore, $\ \frac{1-\left(1-\lambda_{0}x^{*}_{j}\right)^{k}}{\lambda_{0}x^{*}_{j}}\leq\frac{1-\left(1-(1-\epsilon)\right)^{k}}{1-\epsilon}=\frac{1-\epsilon^{k}}{1-\epsilon}$. It follows that $\Pr\left(X_{j}=1\right)\leq\frac{1-\epsilon^{k}}{1-\epsilon}\cdot\lambda_{0}x^{*}_{j}$. Then $\Pr\left(X_{j}=1\right)\leq\left(1-\epsilon^{k}\right)\delta x^{*}_{j}.$ (4) By using the LP relaxation and the definition of the sets $C_{1}$ and $C_{2}$, and since $\lambda x^{*}_{j}\geq 1$ for all ${E_{j}\in C_{1}}$, we get $\displaystyle\mathbb{E}(X)$ $\displaystyle=$ $\displaystyle|C_{1}|+\sum_{E_{j}\in C_{2}}\Pr\left(X_{j}=1\right)+\sum_{E_{j}\in C_{3}}\Pr\left(X_{j}=1\right)$ $\displaystyle\overset{(\ref{probability5})}{\leq}$ $\displaystyle\sum_{E_{j}\in C_{1}}\lambda x^{*}_{j}+\sum_{E_{j}\in C_{2}}(1-\epsilon^{k})\delta x^{*}_{j}+\sum_{E_{j}\in C_{3}}(1-\epsilon^{k})\delta x^{*}_{j}$ $\displaystyle\overset{(\ref{probability4})}{\leq}$ $\displaystyle(1-\epsilon^{k})\delta\sum_{E_{j}\in\mathcal{E}}x^{*}_{j}$ $\displaystyle\leq$ $\displaystyle(1-\epsilon^{k})\delta\mathrm{Opt}^{*}.$ (iv) Let us consider $\tilde{\mathcal{H}}$ the subhypergraph induced by $C_{1}$ in which degree equality gives $\sum_{i\in V}d(i)=\sum_{E_{j}\in C_{1}}|E_{j}|.$ As the minimum vertex degree in the subhypergraph $\tilde{\mathcal{H}}$ is $b-1$ with $b:=\min_{i\in[n]}b_{i}$, we have $(b-1)n\leq\sum_{i\in V}d(i)=\sum_{E\in C_{1}}|E_{j}|\leq\ell|C_{1}|.$ Therefore $\frac{(b-1)n}{\ell}\leq|C_{1}|.$ Since $|C_{1}|<\alpha\cdot\mathrm{Opt}^{*}$ we obtain $\displaystyle\frac{(b-1)n}{\alpha\ell}<\mathrm{Opt}^{*}.$ ###### Theorem 5 Let $\mathcal{H}$ be a hypergraph with fixed maximum vertex degree $\Delta$ and maximum edge size $\ell$. Let $\alpha=\frac{(b-1)\delta\epsilon^{k}}{6\ell}\times\exp\left(a_{k,\epsilon}\right)$ with $a_{k,\epsilon}=\frac{k(1-\epsilon)+(\delta-1)(1-\epsilon^{k})}{2}$ and $\epsilon\in\left[\frac{\delta-1}{2\delta},\left(\frac{\delta-1}{2\delta}\right)^{\frac{1}{k}}\right]$ as in Algorithm 1. The Algorithm 1 returns a set multicover $C$ such that $|C|<\max\left\\{\left(1-\frac{1}{2}\left(1-\epsilon\right)\epsilon^{k}\right)\delta,\left(1-\frac{(b-1)\epsilon^{k}}{18\ell}\times\exp\left(a_{k,\epsilon}\right)\right)\delta\right\\}\cdot\mathrm{Opt}^{*}$ with probability greater than $0.65$. ###### Proof 4 Let $\mathcal{C}$ be the event that the inequality $|C|<\left(1-\frac{1}{2}\left(1-\epsilon\right)\epsilon^{k}\right)\delta\cdot\mathrm{Opt}^{*}$ is satisfied. It suffices to prove that event $\mathcal{C}$ holds with the given probability in the case $|C_{1}|<\alpha\cdot\mathrm{Opt}^{*}$ since the opposite case is discussed in Theorem 4. For this purpose, we estimate both the concentration of $X$ and $Y$ around their expectation. Choose $t=2n\exp\left(-a_{k,\epsilon}\right)$ and consider $\mathcal{A}$ the event $Y\leq n(1-\exp\left(-2a_{k,\epsilon}\right))-t$. This involves $\displaystyle n\exp\left(-2a_{k,\epsilon}\right)+t$ $\displaystyle=$ $\displaystyle n\exp\left(-2a_{k,\epsilon}\right)+2n\exp\left(-a_{k,\epsilon}\right)$ $\displaystyle\leq$ $\displaystyle 3n\exp\left(-a_{k,\epsilon}\right)$ $\displaystyle=$ $\displaystyle\frac{n(b-1)}{\ell}\cdot\frac{6\ell}{(b-1)\delta\epsilon^{k}\exp\left(a_{k,\epsilon}\right)}\cdot\frac{1}{2}\epsilon^{k}\delta$ $\displaystyle=$ $\displaystyle\frac{n(b-1)}{\alpha\ell}\cdot\frac{1}{2}\epsilon^{k}\delta$ $\displaystyle\overset{\textrm{Lem }~{}\ref{lemma:random}(iv)}{\leq}$ $\displaystyle\frac{1}{2}\epsilon^{k}\delta\cdot\mathrm{Opt}^{*}.$ And by Lemma 3(ii) we have $\frac{t^{2}}{{\rm Var}(Y)}\geq\frac{4n^{2}\exp\left(-2a_{k,\epsilon}\right)}{2n^{2}\exp\left(-2a_{k,\epsilon}\right)}=2$. Therefore $\displaystyle\Pr\left(\mathcal{A}\right)$ $\displaystyle\leq$ $\displaystyle\Pr\left(Y\leq\mathbb{E}(Y)-t\right)$ $\displaystyle\overset{\textrm{Th }~{}\ref{Che-Can}}{\leq}$ $\displaystyle\frac{{\rm Var}(Y)}{{\rm Var}(Y)+t^{2}}$ $\displaystyle=$ $\displaystyle\frac{1}{1+\frac{t^{2}}{{\rm Var}(Y)}}$ $\displaystyle\leq$ $\displaystyle\frac{1}{3}.$ Consider now $\mathcal{B}$ the event $X\geq\left(1-\left(1-\frac{1}{2}\epsilon\right)\epsilon^{k}\right)\delta\mathrm{Opt}^{*}$. Our basic assumption is to consider $\delta$, $k$ and $\epsilon$ constants, and we can certainly assume that $n\geq\frac{16\exp\left(a_{k,\epsilon}\right)}{\epsilon^{k+2}}$, since otherwise we obtain an optimal solution for the set multicover problem in polynomial time. Choosing $\beta=\frac{1}{2}\epsilon^{k+1}$ we have $\displaystyle(1+\beta)(1-\epsilon^{k})$ $\displaystyle=$ $\displaystyle 1-\epsilon^{k}+\frac{1}{2}\epsilon^{k+1}-\frac{1}{2}\epsilon^{2k+1}$ $\displaystyle=$ $\displaystyle 1-\epsilon^{k}\left(1-\frac{1}{2}\epsilon+\frac{1}{2}\epsilon^{k+1}\right)$ $\displaystyle\leq$ $\displaystyle 1-\left(1-\frac{1}{2}\epsilon\right)\epsilon^{k}.$ Note that $\epsilon\in\left[\frac{\delta-1}{2\delta},\left(\frac{\delta-1}{2\delta}\right)^{\frac{1}{k}}\right]$ therewith $1-\epsilon^{k}\geq 1-\frac{\delta-1}{2\delta}=\frac{\delta+1}{2\delta}>\frac{1}{2}$. We thus get $\displaystyle\Pr\left(\mathcal{B}\right)$ $\displaystyle\leq$ $\displaystyle\Pr\left(X\geq(1+\beta)\cdot(1-\epsilon^{k})\delta\mathrm{Opt}^{*}\right)$ $\displaystyle\overset{\textrm{Th }~{}\ref{Doerr}}{\leq}$ $\displaystyle\exp\left(-\frac{\beta^{2}(1-\epsilon^{k})\delta\mathrm{Opt}^{*}}{3}\right)$ $\displaystyle\overset{\textrm{Lem.}\ref{lemma:random}(iv)}{\leq}$ $\displaystyle\exp\left(-\frac{\epsilon^{2k+2}(1-\epsilon^{k})\delta n(b-1)}{12\ell}\cdot\frac{6\ell}{(b-1)\delta\epsilon^{k}\times\exp\left(a_{k,\epsilon}\right)}\right)$ $\displaystyle\leq$ $\displaystyle\exp\left(-\frac{\epsilon^{k+2}(1-\epsilon^{k})n}{2\exp\left(a_{k,\epsilon}\right)}\right)$ $\displaystyle\leq$ $\displaystyle\exp\left(-\frac{\epsilon^{k+2}n}{4\exp\left(a_{k,\epsilon}\right)}\right)$ $\displaystyle\leq$ $\displaystyle\exp\left(-4\right).$ Therefore it holds that $\displaystyle\Pr\left(\overline{\mathcal{A}}\cap\overline{\mathcal{B}}\right)$ $\displaystyle\geq$ $\displaystyle 1-\left(\frac{1}{3}+\exp\left(-4\right)\right),$ (5) where $\overline{\mathcal{A}}$ and $\overline{\mathcal{B}}$ denote the complement events of $\mathcal{A}$ and $\mathcal{B}$ respectively. We conclude that $\displaystyle\Pr\left(\mathcal{C}\right)$ $\displaystyle=$ $\displaystyle\Pr\left(|C|\leq\left(1-\frac{1}{2}\left(1-\epsilon\right)\epsilon^{k}\right)\delta\mathrm{Opt}^{*}\right)$ $\displaystyle=$ $\displaystyle\Pr\left(|C|\leq\left(1-\left(1-\frac{1}{2}\epsilon\right)\epsilon^{k}+\frac{1}{2}\epsilon^{k}\right)\delta\mathrm{Opt}^{*}\right)$ $\displaystyle\overset{(\ref{expection})}{\geq}$ $\displaystyle\Pr\left(X+n-Y\leq\left(1-\left(1-\frac{1}{2}\epsilon\right)\epsilon^{k}+\frac{1}{2}\epsilon^{k}\right)\delta\mathrm{Opt}^{*}\right)$ $\displaystyle\geq$ $\displaystyle\Pr\left(X\leq\left(1-\left(1-\frac{1}{2}\epsilon\right)\epsilon^{k}\right)\delta\mathrm{Opt}^{*}\text{\ and\ }n-Y\leq\frac{1}{2}\epsilon^{k}\delta\mathrm{Opt}^{*}\right)$ $\displaystyle\geq$ $\displaystyle\Pr\left(X\leq\left(1-\left(1-\frac{1}{2}\epsilon\right)\epsilon^{k}\right)\delta\mathrm{Opt}^{*}\text{\ and\ }Y\geq n-\frac{1}{2}\epsilon^{k}\delta\mathrm{Opt}^{*}\right)$ $\displaystyle\geq$ $\displaystyle\Pr\left(X\leq\left(1-\left(1-\frac{1}{2}\epsilon\right)\epsilon^{k}\right)\delta\mathrm{Opt}^{*}\text{\ and\ }Y\geq n-n\exp\left(-2a_{k,\epsilon}\right)-t\right)$ $\displaystyle\geq$ $\displaystyle\Pr\left(X\leq\left(1-\left(1-\frac{1}{2}\epsilon\right)\epsilon^{k}\right)\delta\mathrm{Opt}^{*}\text{\ and\ }Y\geq n(1-\exp\left(-2a_{k,\epsilon}\right))-t\right)$ $\displaystyle\overset{(\ref{Intersection})}{\geq}$ $\displaystyle 1-\left(\frac{1}{3}+\exp\left(-4\right)\right)$ $\displaystyle\geq$ $\displaystyle 0.65.$ $\Box$ Remark 2. The proof above gives for $k=2$ and $\epsilon=\frac{1}{2}$ an approximation ratio of $\max\left\\{\frac{15}{16}\delta,\left(1-\frac{(b-1)\exp\left(\frac{3\delta+1}{8}\right)}{72\ell}\right)\delta\right\\}$. Note that $\frac{\delta-1}{2\delta}<\frac{1}{2}<\left(\frac{\delta-1}{2\delta}\right)^{\frac{1}{2}}$ therewith the condition of Theorem 5 on $\epsilon$ is satisfied. As mentioned above our performance guaranty improves over the ratio presented by Srivastav et al [7], and this without restriction on the parameter $\ell$. Namely, for $\delta\geq 13$ we have $\displaystyle 11(\delta-1)<\exp\left(\frac{3\delta+1}{8}\right)$ $\displaystyle\Rightarrow$ $\displaystyle\frac{11(\delta-1)}{72\ell}<\frac{\exp\left(\frac{3\delta+1}{8}\right)}{72\ell}$ $\displaystyle\overset{b-1\geq 1}{\Rightarrow}$ $\displaystyle\frac{11(\Delta-b)}{72\ell}<\frac{(b-1)\exp\left(\frac{3\delta+1}{8}\right)}{72\ell}$ $\displaystyle\Rightarrow$ $\displaystyle\left(1-\frac{(b-1)\exp\left(\frac{3\delta+1}{8}\right)}{72\ell}\right)\delta<\left(1-\frac{11(\Delta-b)}{72\ell}\right)\delta.$ ## 4 Lower Bound One of the features of the proof is the duality of hypergraphs. In dual hypergraphs, vertices and edges just swap the roles. So the set multicover problem in dual hypergraphs becomes as follows: find a minimum cardinality set $C\subseteq V$ such that for every $E\in\mathcal{E}$ it holds $|E\cap C|\geq b$. This problem is known as the $b$-vertex cover problem and we have that the set multicover problem in $\Delta$-regular hypergraphs is equivalent to the $b$-vertex cover problem in $\Delta$-uniform hypergraphs. ###### Theorem 6 Let $\epsilon>0$, $\Delta$ and $\bf b\in\mathbb{N}_{0}^{n}$ be given and $b=\min_{i}b_{i}$. Then, it is NP-hard to approximate the set multicover problem on $\Delta$-regular hypergraphs within a factor of $\Delta-b-\epsilon$. _Proof_. Assume, for a contradiction, that the theorem is false. Then there exists an algorithm $\mathcal{A}$ that returns a $(\Delta-b-\epsilon)$-approximation in polynomial time for the $b$-vertex cover problem on $\Delta$-uniform hypergraphs. We give a reduction of the minimum vertex cover problem on $\Delta$-uniform hypergraphs to the $b$-vertex cover problem on $\Delta+b-1$-uniform hypergraphs. Let $\tilde{\mathcal{H}}=(V,\mathcal{E})$ be a $\Delta$-uniform hypergraph and let $\alpha=\frac{2}{\epsilon}(\Delta-1-\epsilon)(b-1)$. Now we consider the following algorithm: * 1. Consider all subsets $T\subseteq V$ with $|T|\leq\alpha$. Check if any of these subsets is a vertex cover in $\tilde{\mathcal{H}}$. If it’s the case then return the smallest one of them, else go to step 2. * 2. Add $b-1$ vertices $v_{1},\ldots,v_{b-1}$ to $V$. Define for every hyper-edge $E$ a new edge $E^{\ast}:=E\cup\\{v_{1},\ldots,v_{b-1}\\}$ and the set $\mathcal{E}^{\ast}:=\\{E^{\ast}|E\in\mathcal{E}\\}$. Finally set $\mathcal{H}=(V\cup\\{v_{1},\ldots,v_{b-1}\\},\mathcal{E}^{\ast})$. We execute $\mathcal{A}$ on $\mathcal{H}$. Return $T:=\mathcal{A}(\mathcal{H})\cap V$. #### Claim The algorithm given above returns a vertex cover in $\tilde{\mathcal{H}}$ in polynomial-time with an approximation ratio of $\Delta-1-\frac{\epsilon}{2}$. _Proof_. Correctness and approximation ratio. If $T$ is selected by the algorithm in step $1$ then $T$ is an optimal vertex cover in $\tilde{\mathcal{H}}$. If $T$ is selected by the algorithm in step $2$ then $\mathcal{A}(\mathcal{H})=T\cup K$ for some $K\subset\\{v_{1},\ldots,v_{b-1}\\}$. Note that $T$ and $K$ are disjoint sets. Consider an edge $E\in\mathcal{E}$. Because $\mathcal{A}(\mathcal{H})$ is a $b$-vertex cover in $\mathcal{H}$, we have $|\mathcal{A}(\mathcal{H})\cap E^{\ast}|=|T\cap E^{\ast}|+|K\cap E^{\ast}|\geq b$. Since $T\cap E^{\ast}=T\cap E$ and $|K\cap E^{\ast}|\leq b-1$, it follows that $|T\cap E|\geq 1$. Hence $T$ is a vertex cover in $\tilde{\mathcal{H}}$. Now, let $C$ and $C^{\prime}$ denote a minimum vertex cover in $\tilde{\mathcal{H}}$ and a minimum $b$-vertex cover in $\mathcal{H}$, respectively. Since $D^{{}^{\prime}}:=C\cup\\{v_{1},\ldots,v_{b-1}\\}$ is a feasible $b$-vertex cover in $\mathcal{H}$, it holds that $|C^{\prime}|\leq|C|+b-1$. On the other hand, it is clear that $\mathcal{H}$ is a $\left(\Delta+b-1\right)$-uniform hypergraph, and by the assumption we get $\displaystyle|\mathcal{A}(\mathcal{H})|$ $\displaystyle\leq\left((\Delta+b-1)-b-\epsilon\right)|C^{\prime}|$ $\displaystyle\leq(\Delta-1-\epsilon)|C|+(\Delta-1-\epsilon)(b-1)=(\Delta-1-\epsilon)|C|+\frac{\epsilon}{2}\alpha$ $\displaystyle\overset{|C|\geq\alpha}{\leq}(\Delta-1-\frac{\epsilon}{2})|C|.$ Since $|T|=|\mathcal{A}(\mathcal{H})\cap V|\leq|\mathcal{A}(\mathcal{H})|$, it follows that $|T|\leq(\Delta-1-\frac{\epsilon}{2})|C|$. Running time. In step 1 we test at most $n^{\alpha}$ sets of vertices to be a vertex cover in $\tilde{\mathcal{H}}$. Since $\alpha=\frac{2}{\epsilon}(\Delta-1-\epsilon)(b-1)$ is a constant, the running time in this step is polynomial. In step 2 we add a constant number of vertices to $V$ and execute the algorithm $\mathcal{A}$. Hence the algorithm runs in polynomial time in both steps. With Claim 1 there is a factor $\Delta-1-\frac{\epsilon}{2}$ approximation algorithm for the minimum vertex cover problem on $\Delta$-uniform hypergraphs, which contradicts the statement of Theorem 3. $\Box$ ## 5 The $\frac{\ln_{2}(n+1)}{2b}$-Integrality Gap The integrality gap for set multicover problem is defined as the supremum of the ratio $\frac{{\rm Opt}_{{\bf b}}(\mathcal{H})}{{\rm Opt}^{*}_{{\bf b}}(\mathcal{H})}$ over all instances $\cal{H}$ of the problem. In this section we give a slight modification of the proof presented in [22] for the integrality gap. We present in the following a specific class of instances of the set multicover problem, where ${\bf b}:=(b,\ldots,b)\in\mathbb{N}^{n}$ for which the integrality gap is at least $\frac{\ln_{2}(n+1)}{2b}$. ###### Theorem 7 let ${\bf b}:=(b,\ldots,b)\in\mathbb{N}^{n}$. The integrality gap of the set multicover problem is at least $\frac{\ln_{2}(n+1)}{2b}$. Define $V=F_{2}^{k}\backslash\\{0\\}$ as the set of all $k$-dimensional non- zero vectors with component values of $\mathbb{Z}_{2}=\\{0,1\\}$ for a fixed integer $k$ and we define ${\cal E}$ as a collection of the sets $E_{v}=\\{u\in V:<v,u>\equiv 1[2]\\}$ for each $v\in V$, where $<.\,,.>$ is the usual dot product in $V$. We remark that each element $v\in V$ is contained in exactly half of the sets of ${\cal E}$ therewith the hypergraph ${\cal H}=(V,{\cal E})$ is regular and $n=|V|=2^{k}-1$. ###### Lemma 4 Let ${\cal H}=(V,{\cal E})$ the hypergraph defined and ${\bf b}\in\mathbb{N}_{\geq 1}^{n}$. It holds that the vector $x=(\frac{2b}{|{\cal E}|},\ldots,\frac{2b}{|{\cal E}|})$ is a feasible solution for LP($\Delta,\,{\bf b}$). _Proof_. It is clear that $x=(\frac{2b}{|{\cal E}|},\ldots,\frac{2b}{|{\cal E}|})$ is a feasible solution for lP$(\Delta,{\bf b})$, namely since $\mathcal{H}$ is regular with $\Delta=\frac{|{\cal E}|}{2}$ we have for every $i\in\\{1,\ldots,n\\}$ $\sum_{E\in\Gamma(v_{i})}\frac{2b}{|{\cal E}|}=\frac{2b}{|{\cal E}|}\cdot\Delta=\frac{2b}{|{\cal E}|}\cdot\frac{|{\cal E}|}{2}\geq b$ therewith $\mathrm{Opt}^{*}\leq 2b$. $\Box$ ###### Lemma 5 The optimal integral solution to the previous LP formulation of the set multicover problem requires at least $k$ sets. _Proof_. Let $\\{E_{v_{1}},E_{v_{2}}\ldots E_{v_{t}}\\}$ a collection of sets such that ${\bigcup}_{i\in[t]}E_{v_{i}}=F_{2}^{k}\backslash\\{0\\}$. This implies that the intersection of their complements contains exactly the zero vector, i.e., ${\bigcap}_{i\in[t]}E_{v_{i}}^{C}=\\{0\\}$. It follows that $0$ is the only solution in $F_{2}^{k}$ of the system $<x,v_{i}>\equiv 0[2],\quad\forall i\in[t]$ Then it holds that $t\geq k$, since the dimension of $F_{2}^{k}$ is $k$ while the number of the equations in the system is $t$. From this we conclude $\mathrm{Opt}\geq\ln_{2}(n+1)$. $\Box$ Proof of Theorem 7. Theorem 7 follows from Lemma 4 and Lemma 5. $\Box$ ## 6 Future Work We believe now that the conjecture of Peleg et al. holds in the general setting. Hence proving the trueness of the conjecture remains a big challenge for our future works. ## References * [1] R. Bar-Yehuda. _Using Homogeneous Weights for Approximating the Partial Cover Problem._ Journal of Algorithms, 39(2):137–144, 2001. * [2] I. Dinur, V. Guruswami, S. Khot, O. Regev, _A new multilayered PCP and the hardness of hypergraph vertex cover_ , SIAM J. Comput. 34 (5) 1129–1146, 2005. * [3] B. Doerr, F. (Eds.) Neumann. _Theory of Evolutionary Computation: Recent Developments in Discrete Optimization_. Springer Nature 2019. * [4] R. Duh, M. Fürer. _Approximating k-set cover by semi-local optimization._ in: Proc. 29th Annual Symposium on Theory on Computing, May, pp. 256–264, 1997. * [5] M. El Ouali, H. Fohlin, A. Srivastav. _An approximation algorithm for the partial vertex cover problem in hypergraphs_. Journal of Combinatorial Optimization, 31(2): 846–864, 2016. * [6] M. El Ouali, H. Fohlin, A. Srivastav. _A Randomised approximation algorithm for the hitting set problem_. Theoretical Computer Science 555: 23–34, 2014. * [7] M. El Ouali, P. Munstermann, A. Srivastav. _Randomized approximation for set multicover in hypergraphs_. Algorithmica, 74(2): 574–588, 2016. * [8] U. Feige. _A threshold of $\ln n$ for approximating set cover._ J. ACM, 45(4):634–652, 1998. * [9] T. Fujito, H. Kurahashi. _A Better-Than-Greedy Algorithm for k-Set Multicover._ In: 3rd International Workshop on Approximation and Online Algorithms, pp. 176–189, 2006. * [10] R. Gandhi, S. Khuller and A. Srinivasan. _Approximation Algorithms for Partial Covering Problems._ Journal of Algorithms, 53(1):55–84, 2004. * [11] N.G. Hall, D.S. Hochbaum. _A Fast Approximation Algorithm for the Multicovering Problem._ Discrete Applied Mathematics, 15:35–40, 1986. * [12] D.S. Hochbaum. _Approximation Algorithms for the Set Covering and Vertex Cover Problems._ SIAM J. Comput, 11(3):555–556, August 1982. * [13] D. S. Johnson. _Approximation Algorithms for Combinatorial Problems._ Journal of Computer and System Sciences, 9:256–278, 1974. * [14] R. KARP, _Reducibility among combinatorial problems._ In R.E. Miller and J.W. Thatcher, editors, Complexity of Computer Computations, pp. 85–103. Plenum Press, New York, NY, 1972. * [15] S. Khot and O. Regev. _Vertex Cover Might be Hard to Approximate to Within 2-epsilon._ Journal of Computer and System Sciences, 74(3):335–349, 2008. * [16] C. Koufogiannakis, N.E. Young. _Greedy $\Delta$-approximation algorithm for covering with arbitrary constraints and submodular cost_. Algorithmica, 66(1), 113–152, 2013. * [17] L. Lovász. _On the Ratio of Optimal Integral and Fractional Covers._ Discrete Mathematics, 13(4):383–390, 1975. * [18] R. Motwani, P. Raghavan. _Randomized Algorithms_. Cambridge University Press 1995. * [19] D. Peleg, G. Schechtman, A. Wool. _Randomized Approximation of Bounded Multicovering Problems._ Algorithmica, 18(1):44–66, 1997. * [20] D. Peleg, G. Schechtman, A. Wool. _Approximating bounded 0-1 integer linear programs._ In Proc. 2nd Israel Symp. on Theory of Computing Systems, pp. 69–77, Netanya, 1993. * [21] S. Rajagopalan, V. V. Vazirani. _Primal-dual RNC approximation algorithms for set cover and covering integer programs_. SIAM J. Comput., 28(2), 525–540, 1998. * [22] V. V. Vazirani. _Approximation Algorithms_ , pp. 108–112, Springer 2001.
# Convergence Analysis of Projection Method for Variational Inequalities Yekini Shehu111Department of Mathematics, Zhejiang Normal University, Jinhua, 321004, People’s Republic of China; Institute of Science and Technology (IST), Am Campus 1, 3400, Klosterneuburg, Vienna, Austria; e-mail: <EMAIL_ADDRESS>Olaniyi. S. Iyiola222Department of Mathematics, Minnesota State University-Moorhead, Minnesota, USA; e-mail: <EMAIL_ADDRESS>Xiao-Huan Li333College of Science, Civil Aviation University of China, Tianjin 300300, China.; e-mail<EMAIL_ADDRESS>and Qiao-Li Dong444College of Science, Civil Aviation University of China, Tianjin 300300, China.; e-mail<EMAIL_ADDRESS> (January 22, 2021) ###### Abstract The main contributions of this paper are the proposition and the convergence analysis of a class of inertial projection-type algorithm for solving variational inequality problems in real Hilbert spaces where the underline operator is monotone and uniformly continuous. We carry out a unified analysis of the proposed method under very mild assumptions. In particular, weak convergence of the generated sequence is established and nonasymptotic $O(1/n)$ rate of convergence is established, where $n$ denotes the iteration counter. We also present some experimental results to illustrate the profits gained by introducing the inertial extrapolation steps. ## 1 Introduction We first state the formal definition of some classes of functions that play an essential role in this paper. Let $H$ be a real Hilbert space and $X\subseteq H$ be a nonempty subset. ###### Definition 1.1. A mapping $F:X\to H$ is called * (a) monotone on $X$ if $\langle F(x)-F(y),x-y\rangle\geq 0$ for all $x,y\in X$; * (b) Lipschitz continuous on $X$ if there exists a constant $L>0$ such that $\|F(x)-F(y)\|\leq L\|x-y\|,\ \forall x,y\in X.$ * (c) sequentially weakly continuous if for each sequence $\\{x_{n}\\}$ we have: $\\{x_{n}\\}$ converges weakly to $x$ implies $\\{F(x_{n})\\}$ converges weakly to $F(x)$. Let $C$ be a nonempty, closed and convex subset of $H$ and $F:C\rightarrow H$ be a continuous mapping. The variational inequality problem (for short, VI($F,C$)) is defined as: find $x\in C$ such that $\displaystyle\langle F(x),y-x\rangle\geq 0,\quad\forall y\in C.$ (1) Let SOL denote the solution set of VI($F,C$) (1). Variational inequality theory is an important tool in economics, engineering mechanics, mathematical programming, transportation, and so on (see, for example, [7, 8, 22, 29, 30, 31, 38]). A well-known projection-type method for solving VI($F,C$) (1) is the extragradient method introduced by Korpelevich in [32]. It is well known that the extragradient method requires two projections onto the set $C$ and two evaluations of $F$ per iteration. One important hallmark in the design of numerical methods related to the extragradient method is to minimize the number of evaluations of $P_{C}$ per iteration because if $C$ is a general closed and convex set, then a minimal distance problem has to be solved (twice) in order to obtain the next iterate. This has the capacity to seriously affect the efficiency of the extragradient method in a situation, where a projection onto $C$ is hard to evaluate and therefore computationally costly. An attempt in this direction was initiated by Censor et al.[18], who modified extragradient method by replacing the second projection onto the closed and convex subset $C$ with the one onto a subgradient half-space. Their method, which therefore uses only one projection onto $C$, is called the subgradient extragradient method: $x_{1}\in H$, $\displaystyle\left\\{\begin{array}[]{llll}&y_{n}=P_{C}(x_{n}-\lambda F(x_{n})),\\\ &T_{n}:=\\{w\in H:\langle x_{n}-\lambda F(x_{n})-y_{n},w-y_{n}\rangle\leq 0\\},\\\ &x_{n+1}=P_{T_{n}}(x_{n}-\lambda F(y_{n})).\end{array}\right.$ (5) Using (5), Censor et al. [18] proved weak convergence result for VI($F,C$) (1) with a monotone and $L$-Lipschitz-continuous mapping $F$ where $\lambda\in(0,\frac{1}{L})$. Several other related methods to extragradient method and (5) for solving VI($F,C$) (1) in real Hilbert spaces when $F$ is monotone and $L$-Lipschitz-continuous mapping have been studied in the literature (see, for example, [15, 16, 17, 21, 26, 35, 36, 37, 40, 46]). Motivated the result of Alvarez and Attouch in [2] and Censor et al. in [18], Thong and Hieu [45] introduced an algorithm which is a combination of (5) and inertial method for solving VI($F,C$) (1) in real Hilbert space: $x_{0},x_{1}\in H$, $\displaystyle\left\\{\begin{array}[]{llll}&w_{n}=x_{n}+\alpha_{n}(x_{n}-x_{n-1}),\\\ &y_{n}=P_{C}(w_{n}-\lambda F(w_{n})),\\\ &T_{n}:=\\{w\in H:\langle w_{n}-\lambda F(w_{n})-y_{n},w-y_{n}\rangle\leq 0\\},\\\ &x_{n+1}=P_{T_{n}}(w_{n}-\lambda F(y_{n})).\end{array}\right.$ (10) Thong and Hieu [45] proved that the sequence $\\{x_{n}\\}$ generated by (10) converges weakly to a solution of VI($F,C$) (1) with a monotone and $L$-Lipschitz-continuous mapping $F$ where $0<\lambda L\leq\frac{\frac{1}{2}-2\alpha-\frac{1}{2}\alpha^{2}-\delta}{\frac{1}{2}-\alpha+\frac{1}{2}\alpha^{2}}$ for some $0<\delta<\frac{1}{2}-2\alpha-\frac{1}{2}\alpha^{2}$ and $\\{\alpha_{n}\\}$ is a non-decreasing sequence with $0\leq\alpha_{n}\leq\alpha<\sqrt{5}-2$. One the main features of the above mentioned methods (5), (10) and other related methods is the computational issue, for example, step-sizes. The step- sizes in these above methods are bounded by the inverse of the Lipschitz constant which is quite inefficient, because in most cases a global Lipschitz constant (if it indeed exists) of $F$ cannot be accurately estimated, and is usually overestimated, thereby resulting in too small step-sizes. This, of course, is not practical. Therefore, algorithms (5) and (10) are not applicable in most cases of interest. The usual approach to overcome this difficulty consists in some prediction of a step-size with its further correction (see [29, 38]) or in a usage of an Armijo type line search procedure along a feasible direction (see [43]). In terms of computations, the latter approach is more effective, since very often the former approach requires too many projections onto the feasible set per iteration. This paper focuses on the analysis and development of computational projection-type algorithm with inertial extrapolation step for solving VI($F,C$) (1) when the underline operator $F$ is monotone and uniformly continuous when the feasible set $C$ is a nonempty closed affine subset. We obtain weak convergence of the sequence generated by our method. We provide theoretical analysis of our result with weaker assumption on the underline operator $F$ unlike [17, 18, 35, 36] and many other related results on monotone variational inequalities. We also establish the nonasymptotic $O(1/n)$ rate of convergence, which is not given before in other previous inertial type projection methods for VI($F,C$) (1) (see, e.g.,[21, 45]) and give carefully designed computational experiments to illustrate our results. Our computational results show that our proposed methods outperform the iterative methods (5) and (10). Furthermore, our result complements some recent results on inertial type algorithms (see, e.g., [2, 3, 4, 5, 6, 10, 12, 13, 19, 33, 34, 41, 42]). The paper is organized as follows: We first recall some basic definitions and results in Section 2. Some discussions about the proposed inertial projection- type method are given in Section 3. The weak convergence analysis of our algorithm is then investigated in Section 4. We give the rate of convergence of our proposed method in Section 5 and some numerical experiments can be found in Section 6. We conclude with some final remarks in Section 7. ## 2 Preliminaries First, we recall some properties of the projection, cf. [9] for more details. For any point $u\in H$, there exists a unique point $P_{C}u\in C$ such that $\|u-P_{C}u\|\leq\|u-y\|,\leavevmode\nobreak\ \forall y\in C.$ $P_{C}$ is called the metric projection of $H$ onto $C$. We know that $P_{C}$ is a nonexpansive mapping of $H$ onto $C$. It is also known that $P_{C}$ satisfies $\langle x-y,P_{C}x-P_{C}y\rangle\geq\|P_{C}x-P_{C}y\|^{2},\leavevmode\nobreak\ \leavevmode\nobreak\ \forall x,y\in H.$ (11) In particular, we get from (11) that $\langle x-y,x-P_{C}y\rangle\geq\|x-P_{C}y\|^{2},\leavevmode\nobreak\ \leavevmode\nobreak\ \forall x\in C,y\in H.$ (12) Furthermore, $P_{C}x$ is characterized by the properties $P_{C}x\in C\quad\text{and}\quad\langle x-P_{C}x,P_{C}x-y\rangle\geq 0,\leavevmode\nobreak\ \forall y\in C.$ (13) Further properties of the metric projection can be found, for example, in Section 3 of [23]. The following lemmas will be used in our convergence analysis. ###### Lemma 2.1. The following statements hold in $H$: * (a) $\|x+y\|^{2}=\|x\|^{2}+2\langle x,y\rangle+\|y\|^{2}$ for all $x,y\in H$; * (b) $2\langle x-y,x-z\rangle=\|x-y\|^{2}+\|x-z\|^{2}-\|y-z\|^{2}$ for all $x,y,z\in H$; * (c) $\|\alpha x+(1-\alpha)y\|^{2}=\alpha\|x\|^{2}+(1-\alpha)\|y\|^{2}-\alpha(1-\alpha)\|x-y\|^{2}$ for all $x,y\in H$ and $\alpha\in\mathbb{R}$. ###### Lemma 2.2. (see [1, Lem. 3]) Let $\\{\psi_{n}\\}$, $\\{\delta_{n}\\}$ and $\\{\alpha_{n}\\}$ be the sequences in $[0,+\infty)$ such that $\psi_{n+1}\leq\psi_{n}+\alpha_{n}(\psi_{n}-\psi_{n-1})+\delta_{n}$ for all $n\geq 1$, $\sum_{n=1}^{\infty}\delta_{n}<+\infty$ and there exists a real number $\alpha$ with $0\leq\alpha_{n}\leq\alpha<1$ for all $n\geq 1$. Then the following hold: $(i)\leavevmode\nobreak\ \leavevmode\nobreak\ \sum_{n\geq 1}[\psi_{n}-\psi_{n-1}]_{+}<+\infty$, where $[t]_{+}=\max\\{t,0\\}$; (ii) there exists $\psi^{*}\in[0,+\infty)$ such that $\lim_{n\rightarrow+\infty}\psi_{n}=\psi^{*}$. ###### Lemma 2.3. (see [9, Lem. 2.39]) Let $C$ be a nonempty set of $H$ and $\\{x_{n}\\}$ be a sequence in $H$ such that the following two conditions hold: (i) for any $x\in C$, $\lim_{n\rightarrow\infty}\|x_{n}-x\|$ exists; (ii) every sequential weak cluster point of $\\{x_{n}\\}$ is in $C$. Then $\\{x_{n}\\}$ converges weakly to a point in $C$. The following lemmas were given in $\mathbb{R}^{n}$ in [25]. The proof of the lemmas are the same if given in infinite dimensional real Hilbert spaces. Hence, we state the lemmas and omit the proof in real Hilbert spaces. ###### Lemma 2.4. Let $C$ be a nonempty closed and convex subset of $H$. Let $h$ be a real- valued function on $H$ and define $K:=\\{x:h(x)\leq 0\\}$. If $K$ is nonempty and $h$ is Lipschitz continuous on $C$ with modulus $\theta>0$, then ${\rm dist}(x,K)\geq\theta^{-1}\max\\{h(x),0\\},\leavevmode\nobreak\ \forall x\in C,$ where ${\rm dist}(x,K)$ denotes the distance function from $x$ to $K$. ###### Lemma 2.5. Let $C$ be a nonempty closed and convex subset of $H$, $y:=P_{C}(x)$ and $x^{*}\in C$. Then $\|y-x^{*}\|^{2}\leq\|x-x^{*}\|^{2}-\|x-y\|^{2}.$ (14) The following lemma was stated in [28, Prop. 2.11], see also [27, Prop. 4]. ###### Lemma 2.6. Let $H_{1}$ and $H_{2}$ be two real Hilbert spaces. Suppose $F:H_{1}\rightarrow H_{2}$ is uniformly continuous on bounded subsets of $H_{1}$ and $M$ is a bounded subset of $H_{1}$. Then $F(M)$ is bounded. Finally, the following result states the equivalence between a primal and a weak form of variational inequality for continuous, monotone operators. ###### Lemma 2.7. ([44, Lem. 7.1.7]) Let $C$ be a nonempty, closed, and convex subset of $H$. Let $F:C\rightarrow H$ be a continuous, monotone mapping and $z\in C$. Then $z\in{\rm SOL}\Longleftrightarrow\langle F(x),x-z\rangle\geq 0\quad\text{for all }x\in C.$ ## 3 Inertial Projection-type Method Let us first state the assumptions that we will assume to hold for the rest of this paper. ###### Assumption 3.1. Suppose that the following hold: * (a) The feasible set $C$ is a nonempty closed affine subset of the real Hilbert space $H$. * (b) $F:C\to H$ is monotone and uniformly continuous on bounded subsets of $H$. * (c) The solution set SOL of VI$(F,C)$ is nonempty. ###### Assumption 3.2. Suppose the real sequence $\\{\alpha_{n}\\}$ and constants $\beta,\delta,\sigma>0$ satisfy the following conditions: * (a) $\\{\alpha_{n}\\}\subset(0,1)$ with $0\leq\alpha_{n}\leq\alpha_{n+1}\leq\alpha<1$ for all $n$. * (b) $\delta>\frac{\alpha(1+\alpha)(\alpha+\delta\sigma)+\alpha\sigma\delta(\alpha+\delta\sigma)}{\sigma}$ and $\beta<\frac{\delta\sigma}{\alpha+\delta\sigma}-\alpha(1+\alpha)-\alpha\sigma\delta$. Let $r(x):=x-P_{C}(x-F(x))$ stand for the residual equation. Observe that if we take $y=x-F(x)$ in (12), then we have $\langle F(x),r(x)\rangle\geq\|r(x)\|^{2},\leavevmode\nobreak\ \forall x\in C.$ (15) We now introduce our proposed method below. Algorithm 1 Inertial Projection-type Method 1:Choose sequence $\\{\alpha_{n}\\}$ and $\sigma\in(0,1)$ such that the conditions from Assumption 3.2 hold, and take $\gamma\in(0,1)$. Let $x_{0}=x_{1}\in H$ be a given starting point. Set $n:=1$. 2:Set $w_{n}:=x_{n}+\alpha_{n}(x_{n}-x_{n-1}).$ Compute $z_{n}:=P_{C}(w_{n}-F(w_{n}))$. If $r(w_{n})=w_{n}-z_{n}=0$: STOP. 3:Compute $y_{n}=w_{n}-\gamma^{m_{n}}r(w_{n})$, where $m_{n}$ is the smallest nonnegative integer satisfying $\langle F(y_{n}),r(w_{n})\rangle\geq\frac{\sigma}{2}\|r(w_{n})\|^{2}.$ (16) Set $\eta_{n}:=\gamma^{m_{n}}$. 4:Compute $x_{n+1}=P_{C_{n}}(w_{n}),$ (17) where $C_{n}=\\{x:h_{n}(x)\leq 0\\}$ and $h_{n}(x):=\langle F(y_{n}),x-y_{n}\rangle.$ (18) 5:Set $n\leftarrow n+1$ and goto 2. It is clear that $r(w_{n})=0$ implies that we are at a solution of the variational inequality. In our convergence theory, we will implicitly assume that this does not occur after finitely many iterations, so that Algorithm 1 generates an infinite sequence satisfying, in particular, $r(w_{n})\neq 0$ for all $n\in\mathbb{N}$. We will see that this property implies that Algorithm 1 is well defined. ###### Remark 3.3. (a) Algorithm 1 requires, at each iteration, only one projection onto the feasible set $C$ and another projection onto the half-space $C_{n}$ (see [14] for formula for computing projection onto half-space), which is less expensive than the extragradient method especially for the case when computing the projection onto the feasible set $C$ is a dominating task during iteration. (b) Our Algorithm 1 is much more applicable than (5) and (10) in the sense that algorithm (5) and (10) are applicable only for monotone and $L$-Lipschitz-continuous mapping $F$. Thus, the $L$-Lipschitz constant of $F$ or an estimate of it is needed in order to implement the iterative method (5) but our Algorithm 1 is applicable for a much more general class of monotone and uniformly continuous mapping $F$. (c) We observe that the step-size rule in Step 3 involves a couple of evaluations of $F$, but these are often much less expensive than projections onto $C$ which was considered in [29, 38]. Furthermore, using the fact that $F$ is continuous and (15), we can see that Step 3 in Algorithm 1 is well- defined. $\Diamond$ ###### Lemma 3.4. Let the function $h_{n}$ be defined by (18). Then $h_{n}(w_{n})\geq\frac{\sigma\eta_{n}}{2}\|w_{n}-z_{n}\|^{2}.$ In particular, if $w_{n}\neq z_{n}$, then $h_{n}(w_{n})>0$. If $x^{*}\in\text{SOL}$, then $h_{n}(x^{*})\leq 0$. ###### Proof. Since $y_{n}=w_{n}-\eta_{n}(w_{n}-z_{n})$, using (16) we have $\displaystyle h_{n}(w_{n})$ $\displaystyle=\langle F(y_{n}),w_{n}-y_{n}\rangle$ $\displaystyle=\eta_{n}\langle F(y_{n}),w_{n}-z_{n}\rangle\geq\eta_{n}\frac{\sigma}{2}\|w_{n}-z_{n}\|^{2}\geq 0.$ If $w_{n}\neq z_{n}$, then $h_{n}(w_{n})>0$. Furthermore, suppose $x^{*}\in\text{SOL}$. Then by Lemma 2.7 we have $\langle F(x),x-x^{*}\rangle\geq 0\quad\text{for all }x\in C.$ In particular, $\langle F(y_{n}),y_{n}-x^{*}\rangle\geq 0$ and hence $h_{n}(x^{*})\leq 0.$ ∎ Observe that, in finding $\eta_{n}$, the operator $F$ is evaluated (possibly) many times, but no extra projections onto the set $C$ are needed. This is in contrast to a couple of related algorithms for the solution of monotone variational inequalities where the calculation of a suitable step-size requires (possibly) many projections onto $C$, see, e.g., [20, 29, 46]. ## 4 Convergence Analysis We present our main result in this section. To this end, we begin with a result that shows that the sequence $\\{x_{n}\\}$ generated by Algorithm 1 is bounded under the given assumptions. ###### Lemma 4.1. Let $\\{x_{n}\\}$ be generated by Algorithm 1. Then under Assumptions 3.1 and 3.2, we have that $\\{x_{n}\\}$ is bounded. ###### Proof. Let $x^{*}\in\text{SOL}$. By Lemma 2.5 we get (since $x^{*}\in C_{n}$) that $\displaystyle\|x_{n+1}-x^{*}\|^{2}$ $\displaystyle=$ $\displaystyle\|P_{C_{n}}(w_{n})-x^{*}\|^{2}\leq\|w_{n}-x^{*}\|^{2}-\|x_{n+1}-w_{n}\|^{2}$ $\displaystyle=$ $\displaystyle\|w_{n}-x^{*}\|^{2}-{\rm dist}^{2}(w_{n},C_{n}).$ Now, using Lemma 2.1 (c), we have $\displaystyle\|w_{n}-x^{*}\|^{2}$ $\displaystyle=$ $\displaystyle\|(1+\alpha_{n})(x_{n}-x^{*})-\alpha_{n}(x_{n-1}-x^{*})\|^{2}$ (20) $\displaystyle=$ $\displaystyle(1+\alpha_{n})\|x_{n}-x^{*}\|^{2}-\alpha_{n}\|x_{n-1}-x^{*}\|$ $\displaystyle+\alpha_{n}(1+\alpha_{n})\|x_{n}-x_{n-1}\|^{2}.$ Substituting (20) into (4), we have $\displaystyle\|x_{n+1}-x^{*}\|^{2}$ $\displaystyle\leq$ $\displaystyle(1+\alpha_{n})\|x_{n}-x^{*}\|^{2}-\alpha_{n}\|x_{n-1}-x^{*}\|^{2}+\alpha_{n}(1+\alpha_{n})\|x_{n}-x_{n-1}\|^{2}$ (21) $\displaystyle-\|x_{n+1}-w_{n}\|^{2}.$ We also have (using Lemma 2.1 (a)) $\displaystyle\|x_{n+1}-w_{n}\|^{2}$ $\displaystyle=$ $\displaystyle\|(x_{x_{n}+1}-x_{n})-\alpha_{n}(x_{n}-x_{n-1})\|^{2}$ (22) $\displaystyle=$ $\displaystyle\|x_{x_{n}+1}-x_{n}\|^{2}+\alpha^{2}\|x_{n}-x_{n-1}\|^{2}$ $\displaystyle-2\alpha_{n}\langle x_{x_{n}+1}-x_{n},x_{n}-x_{n-1}\rangle$ $\displaystyle\geq$ $\displaystyle\|x_{x_{n}+1}-x_{n}\|^{2}+\alpha^{2}\|x_{n}-x_{n-1}\|^{2}$ $\displaystyle+\alpha_{n}\Big{(}-\rho_{n}\|x_{x_{n}+1}-x_{n}\|^{2}-\frac{1}{\rho_{n}}\|x_{n}-x_{n-1}\|^{2}\Big{)},$ where $\rho_{n}:=\frac{1}{\alpha_{n}+\delta\sigma}$. Combining (21) and (22), we get $\displaystyle\|x_{n+1}-x^{*}\|^{2}-(1+\alpha_{n})\|x_{n}-x^{*}\|^{2}+\alpha_{n}\|x_{n-1}-x^{*}\|^{2}$ (23) $\displaystyle\leq$ $\displaystyle(\alpha_{n}\rho_{n}-1)\|x_{n+1}-x_{n}\|^{2}+\lambda_{n}\|x_{n}-x_{n-1}\|^{2},$ where $\displaystyle\lambda_{n}:=\alpha_{n}(1+\alpha_{n})+\alpha_{n}\frac{1-\alpha_{n}\rho_{n}}{\rho_{n}}\geq 0$ (24) since $\alpha_{n}\rho_{n}<1$. Taking into account the choice of $\rho_{n}$, we have $\delta=\frac{1-\alpha_{n}\rho_{n}}{\sigma\rho_{n}}$ and from (24), it follows that $\displaystyle\lambda_{n}$ $\displaystyle=$ $\displaystyle\alpha_{n}(1+\alpha_{n})+\alpha_{n}\frac{1-\alpha_{n}\rho_{n}}{\rho_{n}}$ (25) $\displaystyle\leq$ $\displaystyle\alpha(1+\alpha)+\alpha\sigma\delta.$ Following the same arguments as in [1, 2, 11], we define $\varphi_{n}:=\|x_{n}-x^{*}\|^{2},n\geq 1$ and $\varepsilon_{n}:=\varphi_{n}-\alpha_{n}\varphi_{n-1}+\lambda_{n}\|x_{n}-x_{n-1}\|^{2},n\geq 1.$ By the monotonicity of $\\{\alpha_{n}\\}$ and the fact that $\varphi_{n}\geq 0$, we have $\varepsilon_{n+1}-\varepsilon_{n}\leq\varphi_{n+1}-(1+\alpha_{n})\varphi_{n}+\alpha_{n}\varphi_{n-1}+\lambda_{n+1}\|x_{n+1}-x_{n}\|^{2}-\lambda_{n}\|x_{n}-x_{n-1}\|^{2}.$ Using (23), we have $\displaystyle\varepsilon_{n+1}-\varepsilon_{n}$ $\displaystyle\leq$ $\displaystyle(\alpha_{n}\rho_{n}-1)\|x_{n+1}-x_{n}\|^{2}+\lambda_{n}\|x_{n}-x_{n-1}\|^{2}$ (26) $\displaystyle+\lambda_{n+1}\|x_{n+1}-x_{n}\|^{2}-\lambda_{n}\|x_{n}-x_{n-1}\|^{2}$ $\displaystyle=$ $\displaystyle(\alpha_{n}\rho_{n}-1+\lambda_{n+1})\|x_{n+1}-x_{n}\|^{2}.$ We now claim that $\displaystyle\alpha_{n}\rho_{n}-1+\lambda_{n+1}\leq-\beta.$ (27) Indeed by the choice of $\rho_{n}$, we have $\displaystyle\alpha_{n}\rho_{n}-1+\lambda_{n+1}\leq-\beta$ $\displaystyle\Leftrightarrow$ $\displaystyle\alpha_{n}\rho_{n}-1+\lambda_{n+1}+\beta\leq 0$ $\displaystyle\Leftrightarrow$ $\displaystyle\lambda_{n+1}+\beta+\frac{\alpha_{n}}{\alpha_{n}+\delta\sigma}-1\leq 0$ $\displaystyle\Leftrightarrow$ $\displaystyle\lambda_{n+1}+\beta-\frac{\delta\sigma}{\alpha_{n}+\delta\sigma}\leq 0$ $\displaystyle\Leftrightarrow$ $\displaystyle(\alpha_{n}+\delta\sigma)(\lambda_{n+1}+\beta)\leq\delta\sigma$ Now, using (25), we have $(\alpha_{n}+\delta\sigma)(\lambda_{n+1}+\beta)\leq((\alpha+\delta\sigma)(\alpha(1+\alpha)\alpha\delta\sigma+\beta)\leq\delta\sigma,$ where the last inequality follows from Assumption 3.2 (b). Hence, the claim in (27) is true. Thus, it follows from (26) and (27) that $\displaystyle\varepsilon_{n+1}-\varepsilon_{n}\leq-\beta\|x_{n+1}-x_{n}\|^{2}.$ (28) The sequence $\\{\varepsilon_{n}\\}$ is non-increasing and the bounds of $\\{\alpha_{n}\\}$ delivers $\displaystyle-\alpha\varphi_{n-1}\leq\varphi_{n}-\alpha\varphi_{n-1}\leq\varepsilon_{n}\leq\varepsilon_{1},n\geq 1.$ (29) It then follows that $\displaystyle\varphi_{n}\leq\alpha^{n}\varphi_{0}+\varepsilon_{1}\sum_{k=0}^{n-1}\alpha^{k}\leq\alpha^{n}\varphi_{0}+\frac{\varepsilon_{1}}{1-\alpha},n\geq 1.$ (30) Combining (28) and (34), we get $\displaystyle\beta\sum_{k=1}^{n}\|x_{k+1}-x_{k}\|^{2}$ $\displaystyle\leq$ $\displaystyle\varepsilon_{1}-\varepsilon_{n+1}$ (31) $\displaystyle\leq$ $\displaystyle\varepsilon_{1}+\alpha\varphi_{n}$ $\displaystyle\leq$ $\displaystyle\alpha^{n+1}\varphi_{0}+\frac{\varepsilon_{1}}{1-\alpha}$ $\displaystyle\leq$ $\displaystyle\varphi_{0}+\frac{\varepsilon_{1}}{1-\alpha},$ which shows that $\displaystyle\sum_{k=1}^{\infty}\|x_{k+1}-x_{k}\|^{2}<\infty.$ (32) Thus, $\underset{n\rightarrow\infty}{\lim}\|x_{n+1}-x_{n}\|=0$. From $w_{n}=x_{n}+\alpha_{n}(x_{n}-x_{n-1})$, we have $\displaystyle\|w_{n}-x_{n}\|$ $\displaystyle\leq$ $\displaystyle\alpha_{n}\|x_{n}-x_{n-1}\|$ $\displaystyle\leq$ $\displaystyle\alpha\|x_{n}-x_{n-1}\|\rightarrow 0,n\rightarrow\infty.$ Similarly, $\|x_{n+1}-w_{n}\|\leq\|x_{n+1}-x_{n}\|+\|x_{n}-w_{n}\|\rightarrow 0,n\rightarrow\infty.$ Using Lemma 2.2, (23), (25) and (32), we have that $\underset{n\rightarrow\infty}{\lim}\|x_{n}-x^{*}\|$ exists. Hence, $\\{x_{n}\\}$ is bounded. ∎ In the next two lemmas, we show that certain subsequences obtained in Algorithm 1 are null subsequences. These two lemmas are necessary in order to show that the weak limit of $\\{x_{n}\\}$ is an element of $SOL$ and for our weak convergence in Theorem 4.4 below. ###### Lemma 4.2. Let $\\{x_{n}\\}$ generated by Algorithm 1 above and Assumptions 3.1 and 3.2 hold. Then * (a) $\displaystyle\lim_{n\rightarrow\infty}\eta_{n}\|w_{n}-z_{n}\|^{2}=0$; * (b) $\displaystyle\lim_{n\rightarrow\infty}\|w_{n}-z_{n}\|=0.$ ###### Proof. Let $x^{*}\in\text{SOL}$. Since $F$ is uniformly continuous on bounded subsets of $X$, then $\\{F(x_{n})\\},\\{z_{n}\\},\\{w_{n}\\}$ and $\\{F(y_{n})\\}$ are bounded. In particular, there exists $M>0$ such that $\|F(y_{n})\|\leq M$ for all $n\in\mathbb{N}$. Combining Lemma 2.4 and Lemma 3.4, we get $\displaystyle\|x_{n+1}-x^{*}\|^{2}$ $\displaystyle=$ $\displaystyle\|P_{C_{n}}(w_{n})-x^{*}\|^{2}\leq\|w_{n}-x^{*}\|^{2}-\|x_{n+1}-w_{n}\|^{2}$ (33) $\displaystyle=$ $\displaystyle\|w_{n}-x^{*}\|^{2}-{\rm dist}^{2}(w_{n},C_{n})$ $\displaystyle\leq$ $\displaystyle\|w_{n}-x^{*}\|^{2}-\Big{(}\frac{1}{M}h_{n}(w_{n})\Big{)}^{2}$ $\displaystyle\leq$ $\displaystyle\|w_{n}-x^{*}\|^{2}-\Big{(}\frac{1}{2M}\sigma\eta_{n}\|r(w_{n})\|^{2}\Big{)}^{2}$ $\displaystyle=$ $\displaystyle\|w_{n}-x^{*}\|^{2}-\Big{(}\frac{1}{2M}\sigma\eta_{n}\|w_{n}-z_{n}\|^{2}\Big{)}^{2}.$ Since $\\{x_{n}\\}$ is bounded, we obtain from (33) that $\displaystyle\Big{(}\frac{1}{2M}\sigma\eta_{n}\|w_{n}-z_{n}\|^{2}\Big{)}^{2}$ $\displaystyle\leq$ $\displaystyle\|w_{n}-x^{*}\|^{2}-\|x_{n+1}-x^{*}\|^{2}$ (34) $\displaystyle=$ $\displaystyle\Big{(}\|w_{n}-x^{*}\|-\|x_{n+1}-x^{*}\|\Big{)}\Big{(}\|w_{n}-x^{*}\|+\|x_{n+1}-x^{*}\|\Big{)}$ $\displaystyle\leq$ $\displaystyle\|w_{n}-x^{*}\|-\|x_{n+1}-x^{*}\|M_{1}$ $\displaystyle\leq$ $\displaystyle\|w_{n}-x_{n+1}\|M_{1},$ where $M_{1}:=\sup_{n\geq 1}\\{\|w_{n}-x^{*}\|+\|x_{n+1}-x^{*}\|\\}$. This establishes (a). To establish (b), We distinguish two cases depending on the behaviour of (the bounded) sequence of step-sizes $\\{\eta_{n}\\}$. Case 1: Suppose that $\liminf_{n\to\infty}\eta_{n}>0$. Then $0\leq\|r(w_{n})\|^{2}=\frac{\eta_{n}\|r(w_{n})\|^{2}}{\eta_{n}}$ and this implies that $\displaystyle\limsup_{n\to\infty}\|r(w_{n})\|^{2}$ $\displaystyle\leq\limsup_{n\to\infty}\bigg{(}\eta_{n}\|r(w_{n})\|^{2}\bigg{)}\bigg{(}\limsup_{n\to\infty}\frac{1}{\eta_{n}}\bigg{)}$ $\displaystyle=\bigg{(}\limsup_{n\to\infty}\eta_{n}\|r(w_{n})\|^{2}\bigg{)}\frac{1}{\liminf_{n\to\infty}\eta_{n}}$ $\displaystyle=0.$ Hence, $\limsup_{n\to\infty}\|r(w_{n})\|=0$. Therefore, $\lim_{n\rightarrow\infty}\|w_{n}-z_{n}\|=\lim_{n\rightarrow\infty}\|r(w_{n})\|=0.$ Case 2: Suppose that $\liminf_{n\to\infty}\eta_{n}=0$. Subsequencing if necessary, we may assume without loss of generality that $\lim_{n\to\infty}\eta_{n}=0$ and $\lim_{n\rightarrow\infty}\|w_{n}-z_{n}\|=a\geq 0$. Define $\bar{y}_{n}:=\frac{1}{\gamma}\eta_{n}z_{n}+\Big{(}1-\frac{1}{\gamma}\eta_{n}\Big{)}w_{n}$ or, equivalently, $\bar{y}_{n}-w_{n}=\frac{1}{\gamma}\eta_{n}(z_{n}-w_{n})$. Since $\\{z_{n}-w_{n}\\}$ is bounded and since $\lim_{n\to\infty}\eta_{n}=0$ holds, it follows that $\lim_{n\to\infty}\|\bar{y}_{n}-w_{n}\|=0.$ (35) From the step-size rule and the definition of $\bar{y}_{k}$, we have $\langle F(\bar{y}_{n}),w_{n}-z_{n}\rangle<\frac{\sigma}{2}\|w_{n}-z_{n}\|^{2},\ \forall n\in\mathbb{N},$ or equivalently $2\langle F(w_{n}),w_{n}-z_{n}\rangle+2\langle F(\bar{y}_{n})-F(w_{n}),w_{n}-z_{n}\rangle<\sigma\|w_{n}-z_{n}\|^{2},\ \forall n\in\mathbb{N}.$ Setting $t_{n}:=w_{n}-F(w_{n})$, we obtain form the last inequality that $2\langle w_{n}-t_{n},w_{n}-z_{n}\rangle+2\langle F(\bar{y}_{n})-F(w_{n}),w_{n}-z_{n}\rangle<\sigma\|w_{n}-z_{n}\|^{2},\ \forall n\in\mathbb{N}.$ Using Lemma 2.1 (b) we get $2\langle w_{n}-t_{n},w_{n}-z_{n}\rangle=\|w_{n}-z_{n}\|^{2}+\|w_{n}-t_{n}\|^{2}-\|z_{n}-t_{n}\|^{2}.$ Therefore, $\|w_{n}-t_{n}\|^{2}-\|z_{n}-t_{n}\|^{2}<(\sigma-1)\|w_{n}-z_{n}\|^{2}-2\langle F(\bar{y}_{n})-F(w_{n}),w_{n}-z_{n}\rangle\ \forall n\in\mathbb{N}.$ Since $F$ is uniformly continuous on bounded subsets of $H$ and (35), if $a>0$ then the right hand side of the last inequality converges to $(\sigma-1)a<0$ as $n\to\infty$. From the last inequality we have $\limsup_{n\to\infty}\left(\|w_{n}-t_{n}\|^{2}-\|z_{n}-t_{n}\|^{2}\right)\leq(\sigma-1)a<0.$ For $\epsilon=-(\sigma-1)a/2>0$, there exists $N\in\mathbb{N}$ such that $\|w_{n}-t_{n}\|^{2}-\|z_{n}-t_{n}\|^{2}\leq(\sigma-1)a+\epsilon=(\sigma-1)a/2<0\quad\forall n\in\mathbb{N},n\geq N,$ leading to $\|w_{n}-t_{n}\|<\|z_{n}-t_{n}\|\quad\forall n\in\mathbb{N},n\geq N,$ which is a contradiction to the definition of $z_{n}=P_{C}(w_{n}-F(w_{n}))$. Hence $a=0$, which completes the proof. ∎ The boundedness of the sequence $\\{x_{n}\\}$ implies that there is at least one weak limit point. We show that such weak limit point belongs to $SOL$ in the next result. ###### Lemma 4.3. Let Assumptions 3.1 and 3.2 hold. Furthermore let $\\{x_{n_{k}}\\}$ be a subsequence of $\\{x_{n}\\}$ converging weakly to a limit point $p$. Then $p\in\text{SOL}$. ###### Proof. By the definition of $z_{n_{k}}$ together with (13), we have $\langle w_{n_{k}}-F(w_{n_{k}})-z_{n_{k}},x-z_{n_{k}}\rangle\leq 0,\ \forall x\in C,$ which implies that $\langle w_{n_{k}}-z_{n_{k}},x-z_{n_{k}}\rangle\leq\langle F(w_{n_{k}}),x-z_{n_{k}}\rangle,\ \forall x\in C.$ Hence, $\langle w_{n_{k}}-z_{n_{k}},x-z_{n_{k}}\rangle+\langle F(w_{n_{k}}),z_{n_{k}}-w_{n_{k}}\rangle\leq\langle F(w_{n_{k}}),x-w_{n_{k}}\rangle,\ \forall x\in C.$ (36) Fix $x\in C$ and let $k\rightarrow\infty$ in (47). Since $\lim_{k\to\infty}\|w_{n_{k}}-z_{n_{k}}\|=0$, we have $0\leq\liminf_{k\to\infty}\langle F(w_{n_{k}}),x-w_{n_{k}}\rangle$ (37) for all $x\in C$. It follows from (47) and the monotonicity of $F$ that $\displaystyle\langle w_{n_{k}}-z_{n_{k}},x-z_{n_{k}}\rangle+\langle F(w_{n_{k}}),z_{n_{k}}-w_{n_{k}}\rangle$ $\displaystyle\leq$ $\displaystyle\langle F(w_{n_{k}}),x-w_{n_{k}}\rangle$ $\displaystyle\leq$ $\displaystyle\langle F(x),x-w_{n_{k}}\rangle\quad\forall x\in C.$ Letting $k\to+\infty$ in the last inequality, remembering that $\lim_{k\to\infty}\|w_{n_{k}}-z_{n_{k}}\|=0$ for all $k$, we have $\langle F(x),x-p\rangle\geq 0\quad\forall x\in C.$ In view of Lemma 2.7, this implies $p\in\text{SOL}$. ∎ All is now set to give the weak convergence result in the theorem below. ###### Theorem 4.4. Let Assumptions 3.1 and 3.2 hold. Then the sequence $\\{x_{n}\\}$ generated by Algorithm 1 weakly converges to a point in SOL. ###### Proof. We have shown that (i) $\lim_{n\to\infty}\|x_{n}-x^{*}\|$ exists; (ii) $\omega_{w}(x_{n})\subset\text{SOL}$, where $\omega_{w}(x_{n}):=\\{x:\exists x_{n_{j}}\rightharpoonup x\\}$ denotes the weak $\omega$-limit set of $\\{x_{n}\\}$. Then, by Lemma 2.3, we have that $\\{x_{n}\\}$ converges weakly to a point in SOL. ∎ We give some discussions on further contributions of this paper in the remark below. ###### Remark 4.5. (a) Our iterative Algorithm 1 is more applicable than some recent results on projection type methods with inertial extrapolation step for solving VI($F,C$) (1) in real Hilbert spaces. For instance, the proposed method in [21] can only be applied for a case when $F$ is monotone and $L$-Lipschitz continuous. Moreover, the Lipschitz constant or an estimate of it has to be known when implementing the Algorithm 3.1 of [21]. In this result, Algorithm 1 is applicable when $F$ is uniformly continuous and monotone operator. (b) In finite-dimensional spaces, the assumption that $F$ is uniformly continuous on bounded subsets of $C$ automatically holds when $F$ is continuous. Moreover, in this case, only continuity of $F$ is required and our weak convergence in Theorem 4.4 coincides with global convergence of sequence of iterates $\\{x_{n}\\}$ in $\mathbb{R}^{n}$. (c) Lemmas 3.5, 4.1, 4.2 and Theorem 4.4 still hold for a more general case of $F$ pseudo-monotone (i.e., for all $x,y\in H$, $\langle F(x),y-x\rangle\geq 0\Longrightarrow\langle F(y),y-x\rangle\geq 0;$). We give a version of Lemma 4.3 for the case of $F$ pseudo-monotone in the Appendix. $\Diamond$ ## 5 Rate of Convergence In this section we give the rate of convergence of the iterative method 1 proposed in Section 3. We show that the proposed method has sublinear rate of convergence and establish the nonasymptotic $O(1/n)$ convergence rate of the proposed method. To the best of our knowledge, there is no convergence rate result known in the literature without stronger assumptions for inertial projection-type Algorithm 1 for VI$(F,C)$ (1) in infinite dimensional Hilbert spaces. ###### Theorem 5.1. Let Assumptions 3.1 and 3.2 hold. Let the sequence $\\{x_{n}\\}$ be generated by Algorithm 1 and $x_{0}=x_{1}$. Then for any $x^{*}\in\text{SOL}$ and for any positive integer $n$, it holds that $\underset{1\leq i\leq n}{\min}\|x_{i+1}-w_{i}\|^{2}\leq\frac{\Big{[}1+\Big{(}\frac{\alpha}{1-\alpha}+\frac{\alpha(1-\alpha)}{\beta}\Big{)}\Big{(}1+\frac{1-\alpha_{0}}{1-\alpha}\Big{)}\Big{]}\|x_{0}-x^{*}\|^{2}}{n}.$ ###### Proof. From (21), we have $\displaystyle\|x_{n+1}-x^{*}\|^{2}-\|x_{n}-x^{*}\|^{2}-\alpha_{n}(\|x_{n}-x^{*}\|^{2}-\|x_{n-1}-x^{*}\|^{2})$ (38) $\displaystyle\leq$ $\displaystyle\alpha_{n}(1+\alpha_{n})\|x_{n}-x_{n-1}\|^{2}-\|x_{n+1}-w_{n}\|^{2}.$ This implies that $\displaystyle\|x_{n+1}-w_{n}\|^{2}$ $\displaystyle\leq$ $\displaystyle\varphi_{n}-\varphi_{n+1}+\alpha_{n}(\varphi_{n}-\varphi_{n-1})+\delta_{n}$ (39) $\displaystyle\leq$ $\displaystyle\varphi_{n}-\varphi_{n+1}+\alpha[V_{n}]_{+}+\delta_{n},$ where $\delta_{n}:=\alpha_{n}(1+\alpha_{n})\|x_{n}-x_{n-1}\|^{2}$, $V_{n}:=\varphi_{n}-\varphi_{n-1}$, $[V_{n}]_{+}:=\max\\{V_{n},0\\}$ and $\varphi_{n}:=\|x_{n}-x^{*}\|^{2}$. Observe from (31) that $\sum_{n=1}^{\infty}\|x_{n+1}-x_{n}\|^{2}\leq\frac{1}{\beta}\Big{[}\varphi_{0}+\frac{\varepsilon_{1}}{1-\alpha}\Big{]}.$ So, $\displaystyle\sum_{n=1}^{\infty}\delta_{n}$ $\displaystyle=$ $\displaystyle\sum_{n=1}^{\infty}\alpha_{n}(1+\alpha_{n})\|x_{n}-x_{n-1}\|^{2}$ (40) $\displaystyle\leq$ $\displaystyle\sum_{n=1}^{\infty}\alpha(1+\alpha)\|x_{n}-x_{n-1}\|^{2}$ $\displaystyle=$ $\displaystyle\alpha(1+\alpha)\sum_{n=1}^{\infty}\|x_{n}-x_{n-1}\|^{2}$ $\displaystyle\leq$ $\displaystyle\frac{\alpha(1+\alpha)}{\beta}\Big{[}\varphi_{0}+\frac{\varepsilon_{1}}{1-\alpha}\Big{]}:=C_{1}.$ The inequality (38) implies that $\displaystyle V_{n+1}$ $\displaystyle\leq$ $\displaystyle\alpha_{n}V_{n}+\delta_{n}$ $\displaystyle\leq$ $\displaystyle\alpha[V_{n}]_{+}+\delta_{n}.$ Therefore, $\displaystyle[V_{n+1}]_{+}$ $\displaystyle\leq$ $\displaystyle\alpha[V_{n}]_{+}+\delta_{n}$ (41) $\displaystyle\leq$ $\displaystyle\alpha^{n}[V_{1}]_{+}+\sum_{j=1}^{n}\alpha^{j-1}\delta_{n+1-j}.$ Note that by our assumption $x_{0}=x_{1}$. This implies that $V_{1}=[V_{1}]_{+}=0$ and $\delta_{1}=0$. From (41), we get $\displaystyle\sum_{n=2}^{\infty}[V_{n}]_{+}$ $\displaystyle\leq$ $\displaystyle\frac{1}{1-\alpha}\sum_{n=1}^{\infty}\delta_{n}$ (42) $\displaystyle=$ $\displaystyle\frac{1}{1-\alpha}\sum_{n=2}^{\infty}\delta_{n}.$ From (39), we get $\displaystyle\sum_{i=1}^{n}\|x_{i+1}-w_{i}\|^{2}$ $\displaystyle\leq$ $\displaystyle\varphi_{1}-\varphi_{n}+\alpha\sum_{i=1}^{n}[V_{i}]_{+}+\sum_{i=2}^{n}\delta_{i}$ (43) $\displaystyle\leq$ $\displaystyle\varphi_{1}+\alpha C_{2}+C_{1},$ where $C_{2}=\frac{C_{1}}{1-\alpha}\geq\frac{1}{1-\alpha}\sum_{i=2}^{\infty}\delta_{i}\geq\sum_{i=2}^{\infty}[V_{i}]_{+}$ by (42). Now, since $\varepsilon_{1}=\varphi_{1}-\alpha_{1}\varphi_{0}=(1-\alpha_{1})\varphi_{1}$, we have $\displaystyle\varphi_{1}+\alpha C_{2}+C_{1}$ $\displaystyle=$ $\displaystyle\varphi_{0}+\frac{\alpha C_{1}}{1-\alpha}$ (44) $\displaystyle+\frac{\alpha(1+\alpha)}{\beta}\Big{[}\varphi_{0}+\frac{\varepsilon_{1}}{1-\alpha}\Big{]}$ $\displaystyle=$ $\displaystyle\varphi_{0}+\frac{\alpha C_{1}}{1-\alpha}$ $\displaystyle+\frac{\alpha(1+\alpha)}{\beta}\Big{[}1+\frac{1-\alpha_{0}}{1-\alpha}\Big{]}\varphi_{0}$ $\displaystyle=$ $\displaystyle\varphi_{0}+\frac{\alpha}{1-\alpha}\Big{[}1+\frac{1-\alpha_{0}}{1-\alpha}\Big{]}$ $\displaystyle+\frac{\alpha(1+\alpha)}{\beta}\Big{[}1+\frac{1-\alpha_{0}}{1-\alpha}\Big{]}\varphi_{0}$ $\displaystyle=$ $\displaystyle\Big{[}1+\Big{(}\frac{\alpha}{1-\alpha}+\frac{\alpha(1-\alpha)}{\beta}\Big{)}\Big{(}1+\frac{1-\alpha_{0}}{1-\alpha}\Big{)}\Big{]}\varphi_{0}.$ From (43) and (44), we obtain $\displaystyle\underset{1\leq i\leq n}{\min}\|x_{i+1}-w_{i}\|^{2}\leq\frac{\Big{[}1+\Big{(}\frac{\alpha}{1-\alpha}+\frac{\alpha(1-\alpha)}{\beta}\Big{)}\Big{(}1+\frac{1-\alpha_{0}}{1-\alpha}\Big{)}\Big{]}\|x_{0}-x^{*}\|^{2}}{n}.$ (45) ∎ ###### Remark 5.2. (a) Note that $x_{n+1}=w_{n}$ implies that $w_{n}\in C_{n}$, where $C_{n}$ is as defined in Algorithm 1 and hence $h_{n}(w_{n})\leq 0$. By Lemma 3.4, we get $\frac{\sigma\eta_{n}}{2}\|w_{n}-z_{n}\|^{2}\leq h_{n}(w_{n})$. Therefore, $0\leq\frac{\sigma\eta_{n}}{2}\|w_{n}-z_{n}\|^{2}\leq h_{n}(w_{n})\leq 0,$ which implies that $w_{n}=z_{n}$. Thus, the equality $x_{n+1}=w_{n}$ implies that $x_{n+1}$ is already a solution of VI$(F,C)$ (1). In this sense, the error estimate given in Theorem 5.1 can be viewed as a convergence rate result of the inertial projection-type method 1. In particular, (45) implies that, to obtain an $\epsilon$-optimal solution in the sense that $\|x_{n+1}-w_{n}\|^{2}<\epsilon$, the upper bound of iterations required by inertial projection-type method 1 is $\frac{\Big{[}1+\Big{(}\frac{\alpha}{1-\alpha}+\frac{\alpha(1-\alpha)}{\beta}\Big{)}\Big{(}1+\frac{1-\alpha_{0}}{1-\alpha}\Big{)}\Big{]}\|x_{0}-x^{*}\|^{2}}{\epsilon}$. We note that with the ” $\min_{1\leq i\leq n}$”, a nonasymptotic $O(1/n)$ convergence rate implies that an $\epsilon$-accuracy solution, in the sense that $\|x_{n+1}-w_{n}\|^{2}<\epsilon$, is obtainable within no more than $O(1/\epsilon)$ iterations. Furthermore, if $\alpha_{n}=0$ for all $n$, then the ”$\min_{1\leq i\leq n}$” can be removed by setting $i=n$ in Theorem 5.1. $\Diamond$ ## 6 Numerical Experiments In this section, we discuss the numerical behaviour of Algorithm 1 using different test examples taken from the literature which are describe below and compare our method with (5), (10) and the original Algorithm (when $\alpha_{n}=0$) of Algorithm 1. ###### Example 6.1. This first example (also considered in [35, 36]) is a classical example for which the usual gradient method does not converge. It is related to the unconstrained case of VI($F,C$) (1) where the feasible set is $C:=\mathbb{R}^{m}$ (for some positive even integer $m$) and $F:=(a_{ij})_{1\leq i,j\leq m}$ is the square matrix $m\times m$ whose terms are given by $\displaystyle a_{ij}=\left\\{\begin{array}[]{llll}&-1,\leavevmode\nobreak\ \leavevmode\nobreak\ {\rm if}\leavevmode\nobreak\ \leavevmode\nobreak\ j=m+1-i\leavevmode\nobreak\ \leavevmode\nobreak\ {\rm and}\leavevmode\nobreak\ \leavevmode\nobreak\ j>i\\\ &1,\leavevmode\nobreak\ \leavevmode\nobreak\ {\rm if}\leavevmode\nobreak\ \leavevmode\nobreak\ j=m+1-i\leavevmode\nobreak\ \leavevmode\nobreak\ {\rm and}\leavevmode\nobreak\ \leavevmode\nobreak\ j<i\\\ &0\leavevmode\nobreak\ \leavevmode\nobreak\ {\rm otherwise}\end{array}\right.$ The zero vector $z=(0,\ldots,0)$ is the solution of this test example. The initial point $x_{0}$ is the unit vector. We choose $\gamma=0.1$, $\sigma=0.8$ and $\alpha_{n}=0.6$, $m=500$. Figure 1: Comparison of Algorithm 1 with the original Algorithm. The numerical result is listed in Figure 1, which illustrates that Algorithm 1 highly improves the original Algorithm. ###### Example 6.2. This example is taken from [24] and has been considered by many authors for numerical experiments (see, for example, [26, 37, 43]). The operator $A$ is defined by $A(x):=Mx+q$, where $M=BB^{T}+S+D$, where $B,S,D\in\mathbb{R}^{m\times m}$ are randomly generated matrices such that $S$ is skew-symmetric (hence the operator does not arise from an optimization problem), $D$ is a positive definite diagonal matrix (hence the variational inequality has a unique solution) and $q=0$. The feasible set $C$ is described by linear inequality constraints $Bx\leq b$ for some random matrix $B\in\mathbb{R}^{k\times m}$ and a random vector $b\in\mathbb{R}^{k}$ with nonnegative entries. Hence the zero vector is feasible and therefore the unique solution of the corresponding variational inequality. These projections are computed by solving a quadratic optimization problem using the MATLAB solver quadprog. Hence, for this class of problems, the evaluation of $A$ is relatively inexpensive, whereas projections are costly. We present the corresponding numerical results (number of iterations and CPU times in seconds) using four different dimensions $m$ and two different numbers of inequality constraints $k$. We compare our proposed Algorithm 1, original Algorithm, subgradient extragradient method (5) and the inertial subgradient extragradient method (10) using Example 6.2 and the numerical results are listed in Tables 1 -4 and shown in Figures 2-5 below. We take the initial point $x_{0}$ to be the unit vector in these algorithms. We use “OPM” to denote the original Algorithm, “SPM” to denote the subgradient extragradient method (5) and “iSPM” to denote inertial subgradient extragradient method (10). We choose the stopping criterion as $\|x^{k}\|\leq\epsilon=0.001.$ The size $k=30,50,80$ and $m=20,50,80,100$. The matrices $B,S,D$ and the vector $b$ are generated randomly. We choose $\gamma=0.1$, $\sigma=0.8$ and $\alpha_{n}=0.1$ in Algorithm (1). In (5), we choose $\sigma=0.8$, $\rho=0.1$, $\mu=0.2$. In iSPM (10), $\alpha=0.2,L=\|M\|,\tau=0.5\frac{\frac{1}{2}-2\alpha-\frac{1}{2}\alpha^{2}}{\frac{1}{2}-\alpha+\frac{1}{2}\alpha^{2}}$. We denote by “Iter.” the number of iterations and “InIt.” the number of total iterations of finding suitable step size in Tables 1 -4 below. Table 1: Comparison of Algorithm 1, original Algorithm and methods (5) and (10) for $k=50,\alpha_{n}=0.1.$ | | Iter. | | | | | InIt. | | | | CPU in second | ---|---|---|---|---|---|---|---|---|---|---|---|--- $m$ | Alg.1 | OPM | SPM | iSPM | | Alg.1 | OPM | SPM | | Alg.1 | OPM | SPM | iSPM | 20 | 1044 | 1162 | 2891 | 4783 | | 2376 | 5022 | 4283 | | 0.0781 | 0.2188 | 0.6563 | 0.5781 | 50 | 4829 | 5912 | 25544 | 29639 | | 13809 | 30885 | 123982 | | 0.5625 | 0.6094 | 7.5938 | 2.1875 | 80 | 19803 | 22129 | 19736 | 61322 | | 74066 | 157061 | 93047 | | 6.6094 | 7.0313 | 15.9063 | 13.4844 | 100 | 26821 | 31149 | 35520 | 92579 | | 101925 | 220297 | 173670 | | 12.1406 | 15.8438 | 67.4219 | 34.6250 | Table 2: Comparison of Algorithm 1, the original Algorithm and methods (5) and (10) for $k=80,\alpha_{n}=0.1.$ | | Iter. | | | | InIt. | | | | CPU in second | ---|---|---|---|---|---|---|---|---|---|---|--- $m$ | Alg.1 | OPM | SPM | iSPM | | Alg.1 | OPM | SPM | | Alg.1 | OPM | SPM | iSPM 20 | 1479 | 1651 | 5096 | 5306 | | 3676 | 7783 | 19071 | | 0.1875 | 0.2188 | 6.7869 | 0.7500 50 | 4152 | 5088 | 8144 | 28033 | | 11955 | 26594 | 36342 | | 1.0625 | 1.2344 | 8.2344 | 6.2813 80 | 22711 | 25864 | 22281 | 64588 | | 85176 | 182177 | 105237 | | 8.2188 | 9.3438 | 20.6719 | 16.1563 100 | 26314 | 30568 | 37588 | 88138 | | 99998 | 216162 | 185430 | | 13.0625 | 16.1250 | 118.4375 | 36.0625 Table 3: Comparison of Algorithm 1, the original Algorithm and methods (5) and (10) for $k=30,\alpha_{n}=0.6.$ | | Iter. | | | | InIt. | | | | CPU in second | ---|---|---|---|---|---|---|---|---|---|---|--- $m$ | Alg.1 | OPM | SPM | iSPM | | Alg.1 | OPM | SPM | | Alg.1 | OPM | SPM | iSPM 10 | 197 | 469 | 509 | 1327 | | 341 | 1370 | 1460 | | 1.21884 | 0.0313 | 0.2969 | 0.2500 30 | 1548 | 4745 | 4347 | 13697 | | 4243 | 17534 | 16880 | | 10.8750 | 0.4063 | 1.0938 | 0.8750 50 | 1581 | 4899 | 8269 | 26393 | | 4762 | 18859 | 37237 | | 12.9688 | 0.5000 | 2.4688 | 1.6094 70 | 6192 | 6256 | 7826 | 56491 | | 22381 | 81445 | 31406 | | 62.4844 | 5.5781 | 9.3281 | 9.1813 Table 4: Comparison of Algorithm 1, the original Algorithm and methods (5) and (10) for $k=50,\alpha_{n}=0.6.$ | | Iter. | | | | InIt. | | | | CPU in second | ---|---|---|---|---|---|---|---|---|---|---|--- $m$ | Alg.1 | OPM | SPM | iSPM | | Alg.1 | OPM | SPM | | Alg.1 | OPM | SPM | iSPM 10 | 147 | 419 | 579 | 1117 | | 253 | 1024 | 1119 | | 3 | 0.0313 | 0.5781 | 0.3594 30 | 1715 | 5110 | 6373 | 13934 | | 4705 | 19018 | 25006 | | 30.0469 | 0.4844 | 1.5469 | 1.1563 50 | 1308 | 4798 | 9227 | 31585 | | 4062 | 17873 | 41084 | | 41.4375 | 0.4844 | 3.5156 | 2.0469 70 | 5673 | 14944 | 8205 | 52124 | | 20548 | 74974 | 33716 | | 393 | 4.5469 | 7.6406 | 10.4375 Figure 2: Comparison of Algorithm 1, original Algorithm and methods (5) and (10). $k=50,\alpha_{n}=0.1.$ Figure 3: Comparison of Algorithm 1, original Algorithm and methods (5) and (10). $k=80,\alpha_{n}=0.1.$ Figure 4: Comparison of Algorithm 1, original Algorithm and methods (5) and (10). $k=30,\alpha_{n}=0.6.$ Figure 5: Comparison of Algorithm 1, original Algorithm and methods (5) and (10). $k=50,\alpha_{n}=0.6.$ Tables 1 -4 and Figures 2-5 show that Algorithm 1 improves the original Algorithm with respect to “Iter.”, “InIt.” and CPU time. It is also observed from Tables 1 -4 and Figures 2-5 that our proposed Algorithm 1 outperform the subgradient extragradient method (5) and the inertial subgradient extragradient method (10) with respect to the CPU time and the number of iterations when the feasible set $C$ is nonempty closed affine subset of $H$. ## 7 Final Remarks This paper presents a weak convergence result with inertial projection-type method for monotone variational inequality problems in real Hilbert spaces under very mild assumptions. This class of method is of inertial nature because at each iteration the projection-type is applied to a point extrapolated at the current iterate in the direction of last movement. Our proposed algorithm framework is not only more simple and intuitive, but also more general than some already proposed inertial projection type methods for solving variational inequality. Based on some pioneering analysis and Algorithm 1, we established certain nonasymptotic $O(1/n)$ convergence rate results. Our preliminary implementation of the algorithms and experimental results have shown that inertial algorithms are generally faster than the corresponding original un-accelerated ones. In our experiments, the extrapolation step-length $\alpha_{n}$ was set to be constant. How to select $\alpha_{n}$ adaptively such that the overall performance is stable and more efficient deserves further investigation. Interesting topics for future research may include relaxing the conditions on $\\{\alpha_{n}\\}$, improving the convergence results, and proposing modified inertial-type algorithms so that the extrapolation step-size can be significantly enlarged. Acknowledgements The project of the first author has received funding from the European Research Council (ERC) under the European Union’s Seventh Framework Program (FP7 - 2007-2013) (Grant agreement No. 616160) ## References * [1] F. Alvarez, Weak convergence of a relaxed and inertial hybrid projection-proximal point algorithm for maximal monotone operators in Hilbert space. SIAM J. Optim. 14 (2004), 773-782. * [2] F. Alvarez and H. Attouch; An inertial proximal method for maximal monotone operators via discretization of a nonlinear oscillator with damping, Set-Valued Anal. 9 (2001), 3-11. * [3] H. Attouch, X. Goudon and P. Redont; The heavy ball with friction. I. The continuous dynamical system, Commun. Contemp. Math. 2 (1) (2000), 1-34. * [4] H. Attouch and M.O. Czarnecki; Asymptotic control and stabilization of nonlinear oscillators with non-isolated equilibria, J. Differential Equations 179 (1) (2002), 278-310. * [5] H. Attouch, J. Peypouquet and P. Redont; A dynamical approach to an inertial forward-backward algorithm for convex minimization, SIAM J. Optim. 24 (2014), 232-256. * [6] H. Attouch and J. Peypouquet; The rate of convergence of Nesterov’s accelerated forward-backward method is actually faster than $\frac{1}{k^{2}}$, SIAM J. Optim. 26 (2016), 1824-1834. * [7] J.-P. Aubin and I. Ekeland; Applied Nonlinear Analysis, Wiley, New York, 1984. * [8] C. Baiocchi and A. Capelo; Variational and Quasivariational Inequalities; Applications to Free Boundary Problems, Wiley, New York (1984). * [9] H.H. Bauschke and P.L. Combettes; Convex Analysis and Monotone Operator Theory in Hilbert Spaces, CMS Books in Mathematics, Springer, New York (2011). * [10] A. Beck and M. Teboulle; A fast iterative shrinkage-thresholding algorithm for linear inverse problems, SIAM J. Imaging Sci. 2 (1) (2009), 183-202. * [11] R. I. Bot, E. R. Csetnek and C. Hendrich; Inertial Douglas-Rachford splitting for monotone inclusion, Appl. Math. Comput. 256 (2015), 472-487. * [12] R. I. Bot and E. R. Csetnek; An inertial alternating direction method of multipliers, Minimax Theory Appl. 1 (2016), 29-49. * [13] R. I. Bot and E. R. Csetnek; An inertial forward-backward-forward primal-dual splitting algorithm for solving monotone inclusion problems, Numer. Alg. 71 (2016), 519-540. * [14] A. Cegielski; Iterative Methods for Fixed Point Problems in Hilbert Spaces, Lecture Notes in Mathematics 2057, Springer, Berlin, 2012. * [15] L.C. Ceng, N. Hadjisavvas, and N.-C. Wong; Strong convergence theorem by a hybrid extragradient-like approximation method for variational inequalities and fixed point problems, J. Glob. Optim. 46 (2010), 635-646. * [16] L. C. Ceng and J. C. Yao; An extragradient-like approximation method for variational inequality problems and fixed point problems, Appl. Math. Comput. 190 (2007), 205-215. * [17] Y. Censor, A. Gibali, and S. Reich; Strong convergence of subgradient extragradient methods for the variational inequality problem in Hilbert space, Optim. Methods Softw. 26 (2011), 827-845. * [18] Y. Censor, A. Gibali and S. Reich; The subgradient extragradient method for solving variational inequalities in Hilbert space, J. Optim. Theory Appl., 148 (2011), 318-335. * [19] C. Chen, R. H. Chan, S. Ma and J. Yang; Inertial Proximal ADMM for Linearly Constrained Separable Convex Optimization, SIAM J. Imaging Sci. 8 (2015), 2239-2267. * [20] S. Denisov, V. Semenov and L. Chabak; Convergence of the modified extragradient method for variational inequalities with non-Lipschitz operators, Cybernet. Systems Anal. 51 (2015), 757-765. * [21] Q. L. Dong, Y. J. Cho, L. L. Zhong and Th. M. Rassias; Inertial projection and contraction algorithms for variational inequalities, J. Global Optim. 70 (2018), 687-704. * [22] R. Glowinski, J.-L. Lions, and R. Trémolières; Numerical Analysis of Variational Inequalities, North-Holland, Amsterdam (1981). * [23] K. Goebel and S. Reich; Uniform convexity, hyperbolic geometry, and nonexpansive mappings, Marcel Dekker, New York, (1984). * [24] P. T. Harker, and J.-S. Pang; A damped-Newton method for the linear complementarity problem, in Computational Solution of Nonlinear Systems of Equations, Lectures in Appl. Math. 26, G. Allgower and K. Georg, eds., AMS, Providence, RI, 1990, pp. 265-284. * [25] Y. R. He; A new double projection algorithm for variational inequalities, J. Comput. Appl. Math. 185 (2006), 166-173. * [26] D. V. Hieu, P. K. Anh and L. D. Muu; Modified hybrid projection methods for finding common solutions to variational inequality problems, Comput. Optim. Appl. 66 (2017), 75-96 * [27] A.N. Iusem and R. Gárciga Otero; Inexact versions of proximal point and augmented Lagrangian algorithms in Banach spaces, Numer. Funct. Anal. Optim. 22 (2001), 609-640. * [28] A.N. Iusem and M. Nasri; Korpelevich’s method for variational inequality problems in Banach spaces, J. Global Optim. 50 (2011), 59-76. * [29] E.N. Khobotov; Modification of the extragradient method for solving variational inequalities and certain optimization problems, USSR Comput. Math. Math. Phys. 27 (1989), 120-127. * [30] D. Kinderlehrer and G. Stampacchia; An Introduction to Variational Inequalities and Their Applications, Academic Press, New York (1980). * [31] I.V. Konnov; Combined Relaxation Methods for Variational Inequalities, Springer-Verlag, Berlin (2001). * [32] G.M. Korpelevich; The extragradient method for finding saddle points and other problems, Ékon. Mat. Metody 12 (1976), 747-756. * [33] D. A. Lorenz and T. Pock; An inertial forward-backward algorithm for monotone inclusions, J. Math. Imaging Vis., 51 (2015), 311-325. * [34] P. E. Maing$\acute{e}$; Regularized and inertial algorithms for common fixed points of nonlinear operators, J. Math. Anal. Appl. 344 (2008), 876-887. * [35] P.-E. Maingé and M.L. Gobinddass; Convergence of one-step projected gradient methods for variational inequalities, J. Optim. Theory Appl. 171 (2016), 146-168. * [36] Yu.V. Malitsky; Projected reflected gradient methods for monotone variational inequalities, SIAM J. Optim. 25 (2015), 502-520. * [37] Yu.V. Malitsky and V.V. Semenov; A hybrid method without extrapolation step for solving variational inequality problems, J. Global Optim. 61 (2015), 193-202. * [38] P. Marcotte; Applications of Khobotov’s algorithm to variational and network equlibrium problems, Inf. Syst. Oper. Res. 29 (1991), 258-270. * [39] J. Mashreghi and M. Nasri; Forcing strong convergence of Korpelevich’s method in Banach spaces with its applications in game theory, Nonlinear Analysis 72 (2010), 2086-2099. * [40] N. Nadezhkina and W. Takahashi; Strong convergence theorem by a hybrid method for nonexpansive mappings and Lipschitz-continuous monotone mappings, SIAM J. Optim. 16 (2006), 1230-1241. * [41] P. Ochs, T. Brox and T. Pock; iPiasco: Inertial Proximal Algorithm for strongly convex Optimization, J. Math. Imaging Vis. 53 (2015), 171-181. * [42] B. T. Polyak; Some methods of speeding up the convergence of iterarive methods, Zh. Vychisl. Mat. Mat. Fiz. 4 (1964), 1-17. * [43] M.V. Solodov and B.F. Svaiter; A new projection method for variational inequality problems, SIAM J. Control Optim. 37 (1999), 765-776. * [44] W. Takahashi; Nonlinear Functional Analysis, Yokohama Publishers, Yokohama, (2000). * [45] D. V. Thong and D. V. Hieu; Modified subgradient extragradient method for variational inequality problems. In press: Numer. Algor. doi:10.1007/s11075-017-0452-4. * [46] P. Tseng; A modified forward-backward splitting method for maximal monotone mappings, SIAM J. Control Optim. 38 (2000), 431-446. ## 8 Appendix In this case, we present a version of Lemma 4.3 for the case when $F$ is pseudo-monotone. ###### Lemma 8.1. Let $F$ be pseudo-monotone, uniformly continuous and sequentially weakly continuous on $H$. Assume that Assumption 3.2 holds. Furthermore let $\\{x_{n_{k}}\\}$ be a subsequence of $\\{x_{n}\\}$ converging weakly to a limit point $p$. Then $p\in\text{SOL}$. ###### Proof. By the definition of $z_{n_{k}}$ together with (13), we have $\langle w_{n_{k}}-F(w_{n_{k}})-z_{n_{k}},x-z_{n_{k}}\rangle\leq 0,\ \forall x\in C,$ which implies that $\langle w_{n_{k}}-z_{n_{k}},x-z_{n_{k}}\rangle\leq\langle F(w_{n_{k}}),x-z_{n_{k}}\rangle,\ \forall x\in C.$ Hence, $\langle w_{n_{k}}-z_{n_{k}},x-z_{n_{k}}\rangle+\langle F(w_{n_{k}}),z_{n_{k}}-w_{n_{k}}\rangle\leq\langle F(w_{n_{k}}),x-w_{n_{k}}\rangle,\ \forall x\in C.$ (47) Fix $x\in C$ and let $k\rightarrow\infty$ in (47). Since $\lim_{k\to\infty}\|w_{n_{k}}-z_{n_{k}}\|=0$, we have $0\leq\liminf_{k\to\infty}\langle F(w_{n_{k}}),x-w_{n_{k}}\rangle$ (48) for all $x\in C$. Now we choose a sequence $\\{\epsilon_{k}\\}_{k}$ of positive numbers decreasing and tending to $0$. For each $\epsilon_{k}$, we denote by $N_{k}$ the smallest positive integer such that $\left\langle F(w_{n_{j}}),x-w_{n_{j}}\right\rangle+\epsilon_{k}\geq 0\quad\forall j\geq N_{k},$ (49) where the existence of $N_{k}$ follows from (48). Since $\left\\{\epsilon_{k}\right\\}$ is decreasing, it is easy to see that the sequence $\left\\{N_{k}\right\\}$ is increasing. Furthermore, for each $k$, $F(w_{N_{k}})\not=0$ and, setting $v_{N_{k}}=\frac{F(w_{N_{k}})}{\|F(w_{N_{k}}\|^{2}},$ we have $\left\langle F(w_{N_{k}}),v_{N_{k}}\right\rangle=1$ for each $k$. Now we can deduce from (49) that for each $k$ $\left\langle F(w_{N_{k}}),x+\epsilon_{k}v_{N_{k}}-w_{N_{k}}\right\rangle\geq 0,$ and, since $F$ is pseudo-monotone, that $\left\langle F(x+\epsilon_{k}v_{N_{k}}),x+\epsilon_{k}v_{N_{k}}-w_{N_{k}}\right\rangle\geq 0.$ (50) On the other hand, we have that $\left\\{x_{n_{k}}\right\\}$ converges weakly to $p$ when $k\to\infty$. Since $F$ is sequentially weakly continuous on $C$, $\left\\{F(w_{n_{k}})\right\\}$ converges weakly to $F(p)$. We can suppose that $F(p)\not=0$ (otherwise, $p$ is a solution). Since the norm mapping is sequentially weakly lower semicontinuous, we have $0<\|F(p)\|\leq\lim\inf_{k\to\infty}\|F(w_{n_{k}})\|.$ Since $\left\\{w_{N_{k}}\right\\}\subset\left\\{w_{n_{k}}\right\\}$ and $\epsilon_{k}\to 0$ as $k\to\infty$, we obtain $\displaystyle 0$ $\displaystyle\leq$ $\displaystyle\lim\sup_{k\to\infty}\|\epsilon_{k}v_{N_{k}}\|=\lim\sup_{k\to\infty}\Big{(}\frac{\epsilon_{k}}{\|F(w_{n_{k}})\|}\Big{)}$ $\displaystyle\leq$ $\displaystyle\frac{\lim\sup_{k\to\infty}\epsilon_{k}}{\lim\inf_{k\to\infty}\|F(w_{n_{k}})\|}\leq\frac{0}{\|F(p)\|}=0,$ which implies that $\lim_{k\to\infty}\|\epsilon_{k}v_{N_{k}}\|=0$. Hence, taking the limit as $k\to\infty$ in (50), we obtain $\left\langle F(x),x-p\right\rangle\geq 0.$ Now, using Lemma 2.2 of [39], we have that $p\in\text{SOL}$. ∎
# Orthogonal subspace based fast iterative thresholding algorithms for joint sparsity recovery Ningning Han, Shidong Li, and Jian Lu _Member, IEEE_ This work was supported by the National Natural Science Foundation of China under grants 61972265, 11871348 and 61373087, by the Natural Science Foundation of Guangdong Province of China under grant 2020B1515310008, by the Educational Commission of Guangdong Province of China under grant 2019KZDZX1007, and by the Guangdong Key Laboratory of Intelligent Information Processing, China, and the NSF of USA (DMS-1615288). Ningning Han<EMAIL_ADDRESS>and Jian Lu (corresponding author<EMAIL_ADDRESS>are with Shenzhen Key Laboratory of Advanced Machine Learning and Applications, College of Mathematics and Statistics, Shenzhen University, Shenzhen, 518060.Shidong Li <EMAIL_ADDRESS>is with Department of Mathematics, San Francisco State University, San Francisco, CA94132. ###### Abstract Sparse signal recoveries from multiple measurement vectors (MMV) with joint sparsity property have many applications in signal, image, and video processing. The problem becomes much more involved when snapshots of the signal matrix are temporally correlated. With signal’s temporal correlation in mind, we provide a framework of iterative MMV algorithms based on thresholding, functional feedback and null space tuning. Convergence analysis for exact recovery is established. Unlike most of iterative greedy algorithms that select indices in a measurement/solution space, we determine indices based on an orthogonal subspace spanned by the iterative sequence. In addition, a functional feedback that controls the amount of energy relocation from the “tails” is implemented and analyzed. It is seen that the principle of functional feedback is capable to lower the number of iteration and speed up the convergence of the algorithm. Numerical experiments demonstrate that the proposed algorithm has a clearly advantageous balance of efficiency, adaptivity and accuracy compared with other state-of-the-art algorithms. ###### Index Terms: Multiple measurement vectors, null space tuning, thresholding, feedback, orthogonal subspace. ## I Introduction In sparse reconstruction signal models with joint sparsity property, signals are sampled at $L$ time instances, resulting in the multiple measurement vector (MMV) model: $\begin{array}[]{l}Y=\Phi X+E,\end{array}$ (1) where $Y\in\mathbb{C}^{M\times L}$ is the observation matrix containing $L$ measurement/snapshot (column) vectors, $\Phi\in\mathbb{C}^{M\times N}$ is the measurement matrix governed by the specific physical system, and $X\in\mathbb{C}^{N\times L}$ is the underlying source signal matrix, to be recovered. $E\in\mathbb{C}^{M\times L}$ is an additive measurement noise matrix. In this system, $L$ measurements share the same row support and elements in each nonzero row of $X$ are temporally correlated. The solution problem to a noiseless MMV model can be formulated as $\begin{array}[]{l}\min\limits_{X}\|X\|_{0}~{}\text{s.t.}~{}Y=\Phi X,\end{array}$ (2) where $\|X\|_{0}=|\text{supp}(X)|$, $\text{supp}(X)=\\{1\leq i\leq N:X_{i\cdot}\neq 0\\}$, $X_{i\cdot}$ is the $i$-th row of $X$. In [1], the authors have shown that $X$ is the unique solution of (2) if $\begin{array}[]{l}\|X\|_{0}<\frac{\text{spark}(\Phi)+\text{rank}(Y)-1}{2},\end{array}$ (3) where spark$(\Phi)$ is the smallest number of linearly dependent columns of $\Phi$. A large majority of effective algorithms for solving (2) are based on two strategies: extending single measurement vector (SMV) algorithms or exploiting signal subspaces. Well-known algorithms of the first class include simultaneous orthogonal matching pursuit (SOMP) [2]-[5], mixed norm minimization techniques [6]-[14], simultaneous greedy algorithms [15, 16]. However, these algorithms, without exploiting subspace structures or temporal correlations, have not offered realistic improvements over performances than that of SMV cases. Recently, a multiple sparse Bayesian learning (MSBL) algorithm [17]-[21], as an extension of sparse Bayesian SMV algorithms, is seen to improve recovery performances by modeling temporal correlation of sparse vectors. Another strategy is to exploit subspace structures spanned by measurement vectors. Representative algorithms include, e.g., sequential compressive MUSIC (SeqCS-MUSIC) [22, 23], subspace-augmented MUSIC (SA- MUSIC+OSMP) [24], rank aware order recursive matching pursuit (RA-ORMP) [25, 26, 27], semi-supervised MUSIC (SS-MUSIC) [28] etc. In this report, we provide a computationally efficient “greedy” algorithm for joint sparsity signal recoveries from their multiple measurement vectors. The proposed algorithm combines procedures of hard thresholding (HT), functional feedback ($f$-FB) for “tail” energy shrinkage and enhanced feasibility, the null space tuning (NST), and a novel variable selection mechanism. The novel criterion of variable selection is based on estimations of significant coefficients in an orthogonal subspace of the iterative sequence. The cardinality of selected variables is determined by the feedback function $f$. Experimental results show that the proposed algorithm provides superior performances in terms of the efficiency and the critical sparsity (i.e., the maximum sparsity level at which the perfect recovery is guaranteed [29]). In fact, the rate of successful recovery of our algorithm has broken through the algebraic upper bound given in (3). --- Figure 1: Left: Frequency of exact recovery as a function of sparsity; right: running time as a function of sparsity. ## II Orthogonal subspace NST+HT+$f$-FB algorithm ### II-A Notations A submatrix of $\Phi$ with columns indexed by a set $I$ is denoted by $\Phi_{I}$ and a submatrix of $\Phi$ with rows indexed by a set $J$ is denoted by $\Phi_{(J)}$. We denote the $i$-th row and the $j$-th column of a matrix $\Phi$ by $\Phi_{i\cdot}$, and $\Phi_{\cdot j}$, respectively. $T\triangle T^{\prime}$ is the symmetric difference of $T$ and $T^{\prime}$, i.e., $T\triangle T^{\prime}=(T\setminus T^{\prime})\cup(T^{\prime}\setminus T)$. $\bm{H}_{T}(X)$ is a linear operator that sets all but elements belong to rows indexed by $T$ of $X$ to zero. Algorithm $1$ OSNST+HT+$f$-FB --- | Input: $\Phi$, $Y$, $\epsilon$, $f(\cdot)$, $K$; | Output: $W$; | Initialize: $k=1$, $W^{0}=0$; | While $\|Y-\Phi W^{k-1}\|_{2}>\epsilon$ and $k<K$ do | $X^{k}=W^{k-1}+\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}(Y-\Phi W^{k-1})$; | $Q^{k}=$orth$(X^{k})$; | $T_{k}=\\{$Indices of $f(k)$ largest $\|Q^{k}_{i\cdot}\|_{2}\\}$; | $W_{T_{k}}^{k}=X_{T_{k}}^{k}+(\Phi_{T_{k}}^{\ast}\Phi_{T_{k}})^{-1}\Phi_{T_{k}}^{\ast}\Phi_{T^{c}_{k}}X_{T^{c}_{k}}^{k}$; | $W_{T^{c}_{k}}^{k}=0$; | $k=k+1$; | end while; ### II-B Algorithm framework The iterative framework of approximation and null space tuning (NST) algorithms is as follows $\left\\{\begin{aligned} \begin{aligned} &W^{k}=\mathbb{D}(X^{k}),\\\ &X^{k+1}=X^{k}+\mathbb{P}(W^{k}-X^{k}).\\\ \end{aligned}\end{aligned}\right.$ Here $\mathbb{D}(X^{k})$ approximates the desired solution by various principles, and $\mathbb{P}:=I-\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}\Phi$ is the orthogonal projection onto ker$(\Phi)$. Since the sequence $\\{X^{k}\\}$ is always feasible (i.e., $Y=\Phi X^{k}$) under the NST principle, one may split $Y$ as $Y=\Phi X=\Phi_{T_{k}}X_{(T_{k})}^{k}+\Phi_{T^{c}_{k}}X_{(T^{c}_{k})}^{k},$ where $T_{k}$ includes indices of $f(k)$ largest $\|Q^{k}_{i\cdot}\|_{2}$ ($i\in\\{1,\ldots,N\\}$), $f(\cdot)\geq 0$ is a non-decreasing function and columns of $Q^{k}$ are an orthonormal basis for the column space of $X^{k}$, i.e., $Q^{k}$=orth$(X^{k})$. The mechanism of feedback is to feed the contribution of $\Phi_{T^{c}_{k}}X_{(T^{c}_{k})}^{k}$ to $Y$ back to im($\Phi_{T_{k}}$), the image of $\Phi_{T_{k}}$. A straightforward way is to set $\Lambda^{k}=\arg\min\limits_{\Lambda}\|\Phi_{T_{k}}\Lambda-\Phi_{T^{c}_{k}}X_{(T^{c}_{k})}^{k}\|_{2},$ which has the best/least-square solution $\Lambda^{k}=(\Phi_{T_{k}}^{\ast}\Phi_{T_{k}})^{-1}\Phi_{T_{k}}^{\ast}\Phi_{T_{k}^{c}}X_{(T_{k}^{c})}^{k}.$ The orthogonal subspace iterative thresholding algorithm with functional feedback and null space tunning (OSNST+HT+$f$-FB) is then established in Algorithm 1. ### II-C Convergence analysis In this paper, we assume the number of snapshots is smaller than the dimension of measurement, i.e., $L<M$, and the measurement matrix $Y$ is full column rank, i.e., rank$(Y)=L$. We now turn to the convergence of OSNST+HT+$f$-FB. ###### Definition 1. [30]. For each integer $s=1,2,\cdots$, the restricted isometry constant (RIC) $\delta_{s}$ of a matrix $\Phi$ is defined as the smallest number $\delta_{s}$ such that $\begin{array}[]{l}(1-\delta_{s})\|X\|_{F}^{2}\leq\|\Phi X\|_{F}^{2}\leq(1+\delta_{s})\|X\|_{F}^{2}\end{array}$ holds for all $s$ row-sparse matrix $X$. Equivalently, it is given by $\begin{array}[]{l}\delta_{s}=\max\limits_{|S|\leq s}\|I-\Phi_{S}^{\ast}\Phi_{S}\|_{2}.\end{array}$ ###### Definition 2. [31]. For each integer $s=1,2,\cdots$ the preconditioned restricted isometry constant $\gamma_{s}$ of a matrix $A$ is defined as the smallest number $\gamma_{s}$ such that $\begin{array}[]{l}(1-\gamma_{s})\|X\|_{F}^{2}\leq\|(\Phi\Phi^{\ast})^{-\frac{1}{2}}\Phi X\|_{F}^{2}\end{array}$ holds for all $s$ row-sparse matrix $X$. In fact, the preconditioned restricted isometry constant $\gamma_{s}$ represents the restricted isometry property of the preconditioned matrix $(\Phi\Phi^{\ast})^{-\frac{1}{2}}\Phi$. Since $\begin{array}[]{l}\|(\Phi\Phi^{\ast})^{-\frac{1}{2}}\Phi X\|_{F}\leq\|(\Phi\Phi^{\ast})^{-\frac{1}{2}}\Phi\|_{2}\|X\|_{F}=\|X\|_{F},\end{array}$ $\gamma_{s}$ is actually the smallest number such that, for all $s$ row-sparse matrix $X$, $\begin{array}[]{l}(1-\gamma_{s})\|X\|_{F}^{2}\leq\|(\Phi\Phi^{\ast})^{-\frac{1}{2}}\Phi X\|_{F}^{2}\leq(1+\gamma_{s})\|X\|_{F}^{2}.\end{array}$ It indicates $\gamma_{s}(\Phi)=\delta_{s}((\Phi\Phi^{\ast})^{-\frac{1}{2}}\Phi)$. Equivalently, it is given by $\begin{array}[]{l}\gamma_{s}=\max\limits_{|S|\leq s}\|I-\Phi_{S}^{\ast}(\Phi\Phi^{\ast})^{-1}\Phi_{S}\|_{2}.\end{array}$ ###### Definition 3. Let the feasible solution space of (2) be $\mathcal{X}=\\{X\in\mathbb{C}^{N\times L}:Y=\Phi X\\}$. Define the modified matrix condition number of $\mathcal{X}$ by $\alpha=\max\limits_{X\in\mathcal{X}}\frac{\sigma_{\max}(X)}{\sigma_{\min}(X)}$, where $\sigma_{\min}(X)$ and $\sigma_{\max}(X)$ denote the smallest and the largest nonzero singular values of $X$, respectively. ###### Lemma 4. Let $U,V\in\mathbb{C}^{N\times L}$ with $|supp(U)\cup supp(V)|\leq t$, then $|\langle U,(I-\Phi^{\ast}\Phi)V\rangle|\leq\delta_{t}\|U\|_{F}\|V\|_{F}$. Suppose $|R\cup supp(V)|\leq t$, then $\|[(I-\Phi^{\ast}\Phi)V]_{(R)}\|_{F}\leq\delta_{t}\|V\|_{F}$. ###### Proof. Let $T=supp(U)\cup supp(V)$, we then have $\begin{array}[]{l}|\langle U,(I-\Phi^{\ast}\Phi)V\rangle|=|\langle U,V\rangle-\langle\Phi U,\Phi V\rangle|\\\ =|\langle U_{(T)},V_{(T)}\rangle-\langle\Phi_{T}U_{(T)},\Phi_{T}V_{(T)}\rangle|\\\ =|\langle U_{(T)},(I-\Phi_{T}^{\ast}\Phi_{T})V_{(T)}\rangle|\\\ \leq\|U_{(T)}\|_{F}\|(I-\Phi_{T}^{\ast}\Phi_{T})V_{(T)}\|_{F}\\\ \leq\|U_{(T)}\|_{F}\|I-\Phi_{T}^{\ast}\Phi_{T}\|_{2}\|V_{(T)}\|_{F}\\\ \leq\delta_{t}\|U\|_{F}\|V\|_{F}.\\\ \end{array}$ The first and the second inequalities are due to the Cauchy-Schwarz inequality, and the sub-multiplicativity of matrix norms, respectively. The last step is by Definition 1. It then follows that $\begin{array}[]{l}\|[(I-\Phi^{\ast}\Phi)V]_{(R)}\|_{F}^{2}=\langle(\bm{H}_{R}\left((I-\Phi^{\ast}\Phi)V\right),(I-\Phi^{\ast}\Phi)V\rangle\leq\delta_{t}\|[(I-\Phi^{\ast}\Phi)V]_{(R)}\|_{F}\|V\|_{F}.\end{array}$ Therefore, $\|[(I-\Phi^{\ast}\Phi)V]_{(R)}\|_{F}\leq\delta_{t}\|V\|_{F}$. ∎ ###### Remark 5. Let $\gamma_{t}$ be the P-RIP constant of $\Phi$ and $U,V\in\mathbb{C}^{N\times L}$ with $|supp(U)\cup supp(V)|\leq t$, then $|\langle U,(I-\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}\Phi)V\rangle|\leq\gamma_{t}\|U\|_{F}\|V\|_{F}$. Suppose $|R\cup supp(V)|\leq t$, then $\|[(I-\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}\Phi)V]_{(R)}\|_{F}\leq\gamma_{t}\|V\|_{F}$. ###### Lemma 6. For $E\in\mathbb{C}^{M\times L}$, $\|[\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}E]_{(T)}\|_{F}\leq\sqrt{1+\theta_{t}}\|E\|_{F}$, where $\theta_{t}=\delta_{t}((\Phi\Phi^{\ast})^{-1}\Phi)$ and $\delta_{t}((\Phi\Phi^{\ast})^{-1}\Phi)$ is RIC of matrix $(\Phi\Phi^{\ast})^{-1}\Phi$. ###### Proof. $\begin{array}[]{l}\|[\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}E]_{(T)}\|_{F}^{2}=\langle\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}E,\bm{H}_{T}(\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}E)\rangle\\\ =\langle E,(\Phi\Phi^{\ast})^{-1}\Phi\bm{H}_{T}(\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}E)\rangle\\\ \leq\|E\|_{F}\sqrt{1+\theta_{t}}\|[\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}E]_{(T)}\|_{F}.\end{array}$ Applying Definition 1 to the matrix $\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}$ obtains the last step. Hence, for all $E\in\mathbb{C}^{M\times L}$, we have $\|[\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}E]_{(T)}\|_{F}\leq\sqrt{1+\theta_{t}}\|E\|_{F}$. ∎ ###### Lemma 7. Let $Y=\Phi X+E$, where $X\in\mathbb{C}^{N\times L}$ is $s$ row-sparse with $S=$supp$(X)$ and $E\in\mathbb{C}^{M\times L}$ is the measurement error. If $\widetilde{W}\in\mathbb{C}^{N\times L}$ is $\widetilde{s}$ row-sparse, $\widetilde{X}=\widetilde{W}+\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}(Y-\Phi\widetilde{W})$, $\widetilde{Q}=orth(\widetilde{X})$, and $T$ is an index set of $t\geq s$ largest $\|\widetilde{Q}_{i\cdot}\|_{2}$, then $\begin{array}[]{l}\|X_{(T^{c})}\|_{F}\leq\sqrt{2}\alpha(\gamma_{s+\widetilde{s}+t}\|X-\widetilde{W}\|_{F}+\sqrt{1+\theta_{t+s}}\|E\|_{F}),\end{array}$ where $\theta_{t+s}(\Phi)=\delta_{t+s}((\Phi\Phi^{\ast})^{-1}\Phi)$. ###### Proof. Since rank$(Y)=L$ and $Y=\Phi\widetilde{X}$, it is obvious that rank$(\widetilde{X})=L$. Consequently, the singular value decomposition of $\widetilde{X}$ can be denoted as $\widetilde{X}=\widetilde{U}_{\ell}\widetilde{\Sigma}_{(\ell)}\widetilde{V}^{\ast}$, where $\widetilde{U}_{\ell}$ is the first $L$ columns of $\widetilde{U}$ and $\widetilde{\Sigma}_{(\ell)}$ denotes the first $L$ rows of $\widetilde{\Sigma}$. Since $\widetilde{U}_{\ell}$ can be regarded as an orthonormal basis for the range of $\widetilde{X}$, without loss of generality, let $\widetilde{Q}=\widetilde{U}_{\ell}$, we have $\begin{array}[]{l}\|[\widetilde{X}\widetilde{V}\widetilde{\Sigma}_{(\ell)}^{-1}]_{(T)}\|_{F}\geq\|[\widetilde{X}\widetilde{V}\widetilde{\Sigma}_{(\ell)}^{-1}]_{(S)}\|_{F}.\\\ \end{array}$ It then follows that $\begin{array}[]{l}\widetilde{\sigma}_{\min}^{-1}\|\widetilde{X}_{(T)}\|_{F}\geq\widetilde{\sigma}_{\max}^{-1}\|\widetilde{X}_{(S)}\|_{F},\\\ \end{array}$ where $\widetilde{\sigma}_{\min}$ and $\widetilde{\sigma}_{\max}$ denote the smallest and the largest singular value of $\widetilde{\Sigma}_{(\ell)}$. Eliminating the common terms over $T\bigcap S$, we obtain $\begin{array}[]{l}\widetilde{\sigma}_{\min}^{-1}\|[\widetilde{W}+\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}(Y-\Phi\widetilde{W})]_{(T\setminus S)}\|_{F}\geq\widetilde{\sigma}_{\max}^{-1}\|[\widetilde{W}+\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}(Y-\Phi\widetilde{W})]_{(S\setminus T)}\|_{F}.\\\ \end{array}$ For the left hand, $\begin{array}[]{l}\widetilde{\sigma}_{\min}^{-1}\|[\widetilde{W}+\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}(Y-\Phi\widetilde{W})]_{(T\setminus S)}\|_{F}\\\ =\widetilde{\sigma}_{\min}^{-1}\|[\widetilde{W}-X+\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}(\Phi X+E-\Phi\widetilde{W})]_{(T\setminus S)}\|_{F}\\\ =\widetilde{\sigma}_{\min}^{-1}\|[(I-\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}\Phi)(\widetilde{W}-X)+\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}E]_{(T\setminus S)}\|_{F}.\\\ \end{array}$ The right hand satisfies $\begin{array}[]{l}\widetilde{\sigma}_{\max}^{-1}\|[\widetilde{W}+\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}(Y-\Phi\widetilde{W})]_{(S\setminus T)}\|_{F}\\\ =\widetilde{\sigma}_{\max}^{-1}\|[\widetilde{W}+\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}(\Phi X+E-\Phi\widetilde{W})+X-X]_{(S\setminus T)}\|_{F}\\\ \geq\widetilde{\sigma}_{\max}^{-1}\|X_{(S\setminus T)}\|_{F}-\widetilde{\sigma}_{\max}^{-1}\|[(I-\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}\Phi)(\widetilde{W}-X)+\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}E]_{(S\setminus T)}\|_{F}.\\\ \end{array}$ Therefore, we obtain $\begin{array}[]{l}\widetilde{\sigma}_{\max}^{-1}\|X_{(S\setminus T)}\|_{F}\\\ \leq\widetilde{\sigma}_{\max}^{-1}\|[(I-\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}\Phi)(\widetilde{W}-X)+\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}E]_{(S\setminus T)}\|_{F}\\\ +\widetilde{\sigma}_{\min}^{-1}\|[(I-\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}\Phi)(\widetilde{W}-X)+\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}E]_{(T\setminus S)}\|_{F}\\\ \leq\sqrt{2}\widetilde{\sigma}_{\min}^{-1}\|[(I-\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}\Phi)(\widetilde{W}-X)+\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}E]_{(T\triangle S)}\|_{F}\\\ \leq\sqrt{2}\widetilde{\sigma}_{\min}^{-1}\|[(I-\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}\Phi)(\widetilde{W}-X)]_{(T\triangle S)}\|_{F}\sqrt{2}\widetilde{\sigma}_{\min}^{-1}\|[\Phi^{\ast}(\Phi\Phi^{\ast})^{-1}E]_{(T\triangle S)}\|_{F}\\\ \leq\sqrt{2}\widetilde{\sigma}_{\min}^{-1}(\gamma_{s+\widetilde{s}+t}\|X-\widetilde{W}\|_{F}+\sqrt{1+\theta_{t+s}}\|E\|_{F}).\end{array}$ The last step is due to Remark 5 and Lemma 6. In view of Definition 3, we derive $\begin{array}[]{l}\|X_{(S\setminus T)}\|_{F}\leq\sqrt{2}\alpha(\gamma_{s+\widetilde{s}+t}\|X-\widetilde{W}\|_{F}+\sqrt{1+\theta_{t+s}}\|E\|_{F}).\end{array}$ ∎ ###### Lemma 8. Let $Y=\Phi X+E$, where $X\in\mathbb{C}^{N\times L}$ is $s$ row-sparse signal matrix, and $E\in\mathbb{C}^{M\times L}$ is the measurement error. Let $S=$supp$(X)$ be the index set of the $s$ sparse rows of $X$. Denote by $\widetilde{Q}=$orth$(\widetilde{X})$ the orthogonal basis of the row-space of $X$, and $T$ the index set of $t\geq s$ largest values of $\|\widetilde{Q}_{i\cdot}\|_{2}$. If $\overline{W}$ is the feedback of $\widetilde{X}$ given by $\overline{W}_{(T)}=\widetilde{X}_{(T)}+(\Phi_{T}^{\ast}\Phi_{T})^{-1}\Phi_{T}^{\ast}\Phi_{T^{c}}\widetilde{X}_{(T^{c})}$ and $\overline{W}_{(T^{c})}=0$, then $\begin{array}[]{l}\|(X-\overline{W})\|_{F}\leq\frac{\|X_{(T^{c})}\|_{F}}{\sqrt{1-\delta_{s+t}^{2}}}+\frac{\sqrt{1+\delta_{t}}\|E\|_{F}}{1-\delta_{s+t}}.\end{array}$ ###### Proof. For any $Z\in\mathbb{C}^{N\times L}$ supported on $T$, $\begin{array}[]{l}\langle\Phi\overline{W}-Y,\Phi Z\rangle\\\ =\langle\Phi_{T}\widetilde{X}_{(T)}+\Phi_{T}(\Phi_{T}^{\ast}\Phi_{T})^{-1}\Phi_{T}^{\ast}\Phi_{T^{c}}\widetilde{X}_{(T^{c})}-Y,\Phi_{T}Z_{(T)}\rangle\\\ =\langle\Phi_{T}^{\ast}(\Phi_{T}\widetilde{X}_{(T)}+\Phi_{T^{c}}\widetilde{X}_{(T^{c})}-Y),Z_{(T)}\rangle\\\ =\langle\Phi_{T}^{\ast}(\Phi\widetilde{X}-Y),Z_{(T)}\rangle\\\ =0.\\\ \end{array}$ The last step is due to the feasibility of $\widetilde{X}$. The inner product can also be written as $\langle\Phi\overline{W}-Y,\Phi Z\rangle=\langle(\Phi\overline{W}-\Phi X-E),\Phi Z\rangle=0$. Therefore, $\langle(\overline{W}-X),\Phi^{\ast}\Phi Z\rangle=\langle E,\Phi Z\rangle,~{}\forall~{}Z\in\mathbb{C}^{N\times L}$ supported on $T$. Since $(\overline{W}-X)_{T}$ is supported on $T$, one has $\langle(\overline{W}-X),\Phi^{\ast}\Phi_{T}(\overline{W}-X)_{(T)}\rangle=\langle E,\Phi_{T}(\overline{W}-X)_{(T)}\rangle.$ Consequently, $\begin{array}[]{l}\|(\overline{W}-X)_{(T)}\|_{F}^{2}=\langle(\overline{W}-X),\bm{H}_{T}(\overline{W}-X)\rangle\\\ =|\langle(X-\overline{W}),(I-\Phi^{\ast}\Phi)\bm{H}_{T}(X-\overline{W})\rangle+|\langle E,\Phi\bm{H}_{T}(X-\overline{W})\rangle|\\\ \leq\delta_{s+t}\|X-\overline{W}\|_{F}\|(X-\overline{W})_{(T)}\|_{F}+\sqrt{1+\delta_{t}}\|E\|_{F}\|(X-\overline{W})_{(T)}\|_{F}.\\\ \end{array}$ The last step is due to Lemma 4 and Definition 1. We can obtain $\begin{array}[]{l}\|(X-\overline{W})_{(T)}\|_{F}\leq\delta_{s+t}\|X-\overline{W}\|_{F}+\sqrt{1+\delta_{t}}\|E\|_{F}.\end{array}$ It then follows that $\begin{array}[]{l}\|(X-\overline{W})\|_{F}^{2}=\|(X-\overline{W})_{(T)}\|_{F}^{2}+\|(X-\overline{W})_{(T^{c})}\|_{F}^{2}\\\ \leq(\delta_{s+t}\|X-\overline{W}\|_{F}+\sqrt{1+\delta_{t}}\|E\|_{F})^{2}+\|X_{(T^{c})}\|_{F}^{2}.\\\ \end{array}$ This in turn implies $p(\|X-\widetilde{W}\|_{F})\leq 0$, where $p(\cdot)$ is a quadratic polynomial, defined by $\begin{array}[]{l}p(x)=(1-\delta_{s+t}^{2})x^{2}-2\delta_{s+t}\sqrt{1+\delta_{t}}\|E\|_{F}x-(1+\delta_{t})\|E\|_{F}^{2}-\|X_{(T^{c})}\|_{F}^{2}.\end{array}$ Since $(1-\delta_{s+t}^{2})\geq 0$, it means that $\|(X-\overline{W})\|_{F}$ is smaller than the largest root of $p(\cdot)$ $\begin{array}[]{l}\|(X-\overline{W})\|_{F}\leq\frac{\delta_{s+t}\sqrt{1+\delta_{t}}\|E\|_{F}+\sqrt{(1+\delta_{t})\|E\|_{F}^{2}+({1-\delta_{s+t}^{2})\|X_{(T^{c})}\|_{F}^{2}}}}{1-\delta_{s+t}^{2}}\\\ \leq\frac{\|X_{(T^{c})}\|_{F}}{\sqrt{1-\delta_{s+t}^{2}}}+\frac{\sqrt{1+\delta_{t}}\|E\|_{F}}{1-\delta_{s+t}}.\\\ \end{array}$ ∎ ###### Theorem 9. Let $Y=\Phi X+E$, where $X$ is the $s$ row-sparse signal matrix. Then the sequence $\\{W^{k}\\}$ produced by OSNST+HT+$f$-FB satisfies $\begin{array}[]{l}\|(X-W^{k})\|_{F}\leq\rho_{s+f(k)+f(k-1)}^{k}\|X-W^{0}\|_{F}+\frac{\kappa_{s+f(k)+f(k-1)}(1-\rho_{s+f(k)+f(k-1)}^{k})}{1-\rho_{s+f(k)+f(k-1)}}\|E\|_{F},\\\ \end{array}$ where $\rho_{\ell}=\sqrt{\frac{2\alpha^{2}\gamma_{\ell}^{2}}{1-\delta_{\ell}^{2}}}$ and $\kappa_{\ell}=(\frac{\sqrt{1+\delta_{\ell}}}{1-\delta_{\ell}}+\frac{\sqrt{2\alpha^{2}(1+\theta_{\ell})}}{\sqrt{1-\delta_{\ell}^{2}}})$. ###### Proof. Applying Lemma 7 to $\widetilde{W}=W^{k-1}$ and $T=T_{k}$ gives $\begin{array}[]{l}\|X_{(T^{c}_{k})}\|_{F}\leq\sqrt{2}\alpha(\gamma_{s+f(k-1)+f(k)}\|X-W^{k-1}\|_{F}+\sqrt{1+\theta_{s+f(k)}}\|E\|_{F}),\end{array}$ and setting $\overline{W}=W^{k}$ and $T=T_{k}$ in Lemma 8 obtains $\begin{array}[]{l}\|(X-W^{k})\|_{F}\leq\frac{\|X_{(T^{c}_{k})}\|_{F}}{\sqrt{(1-\delta_{s+f(k)}^{2})}}+\frac{\sqrt{1+\delta_{f(k)}}\|E\|_{F}}{1-\delta_{s+f(k)}}.\end{array}$ Combining these two inequalities, we have $\begin{array}[]{l}\|(X-W^{k})\|_{F}\leq\sqrt{\frac{2\alpha^{2}\gamma_{s+f(k)+f(k-1)}^{2}}{(1-\delta_{s+f(k)}^{2})}}\|X-W^{k-1}\|_{F}+(\frac{\sqrt{1+\delta_{f(k)}}}{1-\delta_{s+f(k)}}+\frac{\sqrt{2\alpha^{2}(1+\theta_{s+f(k)})}}{\sqrt{1-\delta_{s+f(k)}^{2}}})\|E\|_{F}.\end{array}$ Since $\delta_{\ell}$ and $\gamma_{\ell}$ are all non-decreasing [30], $\rho_{\ell}$ and $\kappa_{\ell}$ are also all non-decreasing as $\ell$ increases for all integer $\ell$. Note that $f(\ell)$ is also a nondecreasing function, it then follows that $\begin{array}[]{l}\|(X-W^{k})\|_{F}\leq\rho_{s+f(k)+f(k-1)}^{k}\|X-W^{0}\|_{F}+\frac{\kappa_{s+f(k)+f(k-1)}(1-\rho_{s+f(k)+f(k-1)}^{k})}{1-\rho_{s+f(k)+f(k-1)}}\|E\|_{F}.\end{array}$ ∎ Consequently, if the RIP and the P-RIP of the matrix $\Phi$ obeys $2\alpha^{2}\gamma_{s+f(k)+f(k-1)}^{2}+\delta_{s+f(k)+f(k-1)}^{2}<1$, the OSNST+HT+FB algorithm is guaranteed to converge. --- Figure 2: Left: Frequency of exact recovery as a function of sparsity; right: running time as a function of sparsity. ## III EXPERIMENTS In this experiment, the measurement matrix $\Phi$ is an $300\times 1000$ Gaussian random matrix and the number of snapshots is $10$. To model the temporal correlation of MMV problem, we employ an autoregressive process of order $1$, AR(1). As a result, the $j$-th snapshot $X_{\cdot j}$ is generated according to the model $\begin{array}[]{l}X_{\cdot j}=\beta X_{\cdot(j-1)}+(1-\beta)\epsilon_{j},\end{array}$ where $\beta$ is the AR model parameter controlling the temporal correlation and $\epsilon_{j}$ is the level of white Gaussian perturbation. The support of a sparse signal is also chosen randomly and the nonzero entries of Gaussian sparse signals are drawn independently from the Gaussian distribution with zero mean and unit variance. A successful recovery is recorded when $\|X-\widehat{X}\|_{F}/\|X\|_{F}\leq 10^{-4}$, where $X$ is the exact signal matrix and $\widehat{X}$ denotes the recovered signal. Each experiment is tested for $100$ (random) trials. A matlab implementation of the proposed algorithm is also available at https://www.dropbox.com/s/2avudk770m4c6rz/OSNST%2BHT%2Bf-FB.zip?dl=0. We first study the mechanisms of $f$-feedback by introducing six particular index selection functions: $f(x)=x$, $f(x)=3x$, $f(x)=6x$, $f(x)=9x$, $f(x)=12x$ and $f(x)=x^{2}$. As discussed, higher critical sparsity represents better empirical recovery performance. Figure 1 shows the frequency of exact recovery and the running time as functions of the sparsity levels $s$. As shown, linear functions with modest gradients present similar performance, which is better than the quadratic function $f(x)=x^{2}$. In addition, one can accelerate the convergence of the class of OSNST+HT+$f$-FB algorithms by adjusting the cardinality of indices per iteration. Also presented are comparisons among our OSNST+HT+$f$-FB and state-of-the-art techniques such as SOMP [2], $\ell_{2,1}$ norm [8], SHTP [15, 16], RA-ORMP [10], TMSBL [17], SA-MUSIC+OSMP [24], SeqCS-MUSIC [22, 23] in terms of frequency of exact recovery and running time. In this experiment, we adopt a modest setting $f(x)=6x$, which can be applied to other applications. In Figure 2, experimental results show that OSNST+HT+$f$-FB still delivers reasonable performance better than that of SOMP, $\ell_{2,1}$ norm, SHTP, TMSBL, SA-MUSIC+OSMP, and SeqCS-MUSIC, though slightly under-performs that of RA-ORMP. For the execution-time comparison, our algorithm achieves the best performance. Numerical experiments show that our algorithm has a clearly advantageous balance of efficiency, adaptivity and accuracy compared with other state-of-the-art algorithms. ## References * [1] J. Chen and X. Huo, “Theoretical results on sparse representations of multiple-measurement vectors,” IEEE Trans. Signal Process., vol. 54, no. 12, pp. 4634-4643, Dec. 2006. * [2] J. A. Tropp, A. C. Gilbert, and M. J. Strauss, “Algorithms for simultaneous sparse approximation. Part I: Greedy pursuit,” Signal Process., vol. 86, no. 3, pp. 572-588, 2006. * [3] J. F. Determeand, L. Louveaux, L. Jacques and F. Horlin, “On the exact recovery condition of simultaneous orthogonal matching pursuit,” IEEE Signal Process. Lett., vol. 23, no.1, pp. 164-168, Jan. 2016. * [4] J. F. Determeand, L. Louveaux, L. Jacques and F. Horlin, “Improving the correlation lower bound for simultaneous orthogonal matching pursuit,” IEEE Signal Process. Lett., vol. 23, no. 11, pp. 1642-1646, Nov. 2018. * [5] J. Determe, J. Louveaux, L. Jacques, and F. Horlin, “On the noise robustness of simultaneous orthogonal matching pursuit,” IEEE Trans. Signal Process., vol. 65, no. 4, pp. 864-875, Feb. 2017. * [6] S. F. Cotter, B. D. Rao, K. Engang, and K. Kreutz-Delgado, “Sparse solutions to linear inverse problems with multiple measurement vectors,” IEEE Trans. Signal Process., vol. 53, no. 7, pp. 2477-2488, Jul. 2005. * [7] J. A. Tropp, “Algorithms for simultaneous sparse approximation-Part II: Convex relaxation,” Signal Process., vol. 86, pp. 589-602, 2006. * [8] E. V. Berg and M. P. Friedlander, “Theoretical and empirical results for recovery from multiple measurements,” IEEE Trans. Inf. Theory, vol. 56, no. 5, pp. 2516-2527, May 2010. * [9] M. Mishali and Y. C. Eldar, “Reduce and boost: Recovering arbitrary sets of jointly sparse vectors,” IEEE Trans. Signal Process., vol. 56, pp. 4692-4702, Oct. 2008. * [10] Y. C. Eldar and H. Rauhut, “Average case analysis of multichannel sparse recovery using convex relaxation,” IEEE Trans. Inform. Theory, vol. 56, no. 1, pp. 505-519, Jan. 2010. * [11] M. M. Hyder and K. Mahata, “A robust algorithm for joint-sparse recovery,” IEEE Signal Process. Lett., vol. 16, no. 12, pp. 1091-1094, Dec. 2009. * [12] D.Wipf and S. Nagarajan, “Iterative reweighted and methods for finding sparse solutions,” IEEE J. Sel. Topics Signal Process., vol. 4, no. 2, pp. 317-329, Apr. 2010. * [13] X. Du, L. Cheng, and L. Liu, “A swarm intelligence algorithm for joint sparse recovery,” IEEE Signal Process. Lett., vol. 20, no. 6, pp. 611-614, Jun. 2013. * [14] S. Khanna and C.R. Murthy, “Sparse recovery from multiple measurement vectors using exponentiated gradient updates,” IEEE Signal Process. Lett., vol. 25, no. 10, pp. 1485-1489, Oct. 2018. * [15] S. Foucart, “Recovering jointly sparse vectors via hard thresholding pursuit,” in Proc. SAMPTA, 2011. * [16] J. D. Blanchard, M. Cermak, D. Hanle, and Y. Jing, “Greedy algorithms for joint sparse recovery,” IEEE Trans. Signal Process., vol. 62, no. 7, pp. 1694-1704, Apr. 2014. * [17] Z. L. Zhang and B. D. Rao, “Sparse signal recovery with temporally corre-lated source vectors using sparse Bayesian learning,” IEEE Trans. Signal Process., vol. 5, no. 5, pp. 912-926, Sep. 2011. * [18] J. Ziniel and P. Schniter, “Efficient high-dimensional inference in the multiple measurement vector problem,” IEEE Trans. Signal Process., vol. 61, no. 2, pp. 340-354, Jan. 2013. * [19] Q. Wu, Y. D. Zhang, M. G. Amin, and B. Himed, “Multi-task Bayesiancompressive sensing exploiting intra-task dependency,” IEEE Trans. Signal Process., vol. 22, no. 4, pp. 430-434, Apr. 2015. * [20] G. Joseph and C. R. Murthy, “A noniterative online Bayesian algorithm forthe recovery of temporally correlated sparse vectors,” IEEE Trans. Signal Process., vol. 65, no. 20, pp. 5510-5525, Oct. 2017. * [21] J. Shang, Z. Wang, and Q. Huang, “A robust algorithm for joint sparserecovery in presence of impulsive noise,” IEEE Signal Process. Lett., vol. 22, no. 8, pp. 1166-1170, Aug. 2015. * [22] J. M. Kim, O. K. Lee, and J. C. Ye, “Compressive MUSIC: Revisiting the link between compressive sensing and array signal processing,” IEEE Trans. Inf. Theory, vol. 58, no. 1, pp. 278-301, Jan. 2012. * [23] J. M. Kim, O. K. Lee, and J. C. Ye, “Improving noise robustness insubspace-based joint sparse recovery,” IEEE Trans. Signal Process., vol. 60, no. 11, pp. 5799-5809, Nov. 2012 * [24] K. Lee, Y. Bresler, and M. Junge, “Subspace methods for joint sparse recovery,” IEEE Trans. Inf. Theory, vol. 58, no. 6, pp. 3613-3641, Jun. 2012. * [25] M. E. Davies and Y. C. Eldar,“ Rank awareness in joint sparse recovery,” IEEE Trans. Inf. Theory, vol. 58, no. 2, pp. 1135-1146, Feb. 2012. * [26] J. Blanchard and M. Davies, “Recovery guarantees for rank aware pursuits,” IEEE Signal Process. Lett., vol. 19, no. 7, pp. 427–430, Jul. 2012. * [27] J. Kim, J, Wang, and B. Shim, “Nearly Optimal Restricted Isometry Condition for Rank Aware Order Recursive Matching Pursuit,” IEEE Trans. Signal Process., vol. 67, no. 17, pp. 4449-4463, Sep. 2019. * [28] Z. Wen, B. Hou, and L. Jiao, “Joint sparse recovery with semisupervised MUSIC,” IEEE Signal Process. Lett., vol. 24, no. 5, pp. 629-633, May. 2017. * [29] W. Dai and O. Milenkovic, “Subspace pursuit for compressive sensing signal reconstruction,” IEEE Trans. Inf. Theory, vol. 55, no. 5, pp. 2230-2249, May 2009. * [30] E. J. Candes and T. Tao, “Decoding by linear programming,” IEEE Trans. Inf. Theory, vol. 51, no. 12, pp. 4203-4215, Dec. 2005. * [31] S. D. Li, Y. L. Liu, and T. B. Mi, “Iterative hard thresholding for compressed sensing,” Appl. Comput. Harmon. Anal., vol. 37, no. 1, pp. 69-88, Jul. 2014.
11institutetext: Radboud University Nijmegen, Email: 11email<EMAIL_ADDRESS>22institutetext: University of Groningen # Privacy Friendly E-Ticketing For Public Transport††thanks: Version: Fri Jan 22 10:58:06 2021 +0100 / arXiv / ov-pet.tex Jaap-Henk Hoepman 1122 ###### Abstract This paper studies how to implement a privacy friendly form of ticketing for public transport in practice. The protocols described are inspired by current (privacy invasive) public transport ticketing systems used around the world. The first protocol emulates paper based tickets. The second protocol implements a pay-as-you-go approach, with fares determined when users check-in and check-out. Both protocols assume the use of a smart phone as the main user device to store tickets or travel credit. We see this research as a step towards investigating how to design commonly used infrastructure in a privacy friendly manner in practice, paying particular attention to how to deal with failures. ## 1 Introduction At the turn of the century, several countries transitioned from paper based tickets for public transport to electronic forms of ticketing, either for public transport in a metropolitan area like the London Underground (the so- called Oyster card111https://oyster.tfl.gov.uk/ ) and Hong Kong public transport (the Octopus Card222https://www.octopus.com.hk/en ), or for all public transport in an entiry country (like the OV-chipkaart333https://www.ov- chipkaart.nl/ in the Netherlands). These e-ticketing systems are typically based on contactless smart cards. Some of these systems exhibit significant weaknesses in terms of security [11] (because the smart cards used need to be cheap and therefore contain weaker security features) and privacy [14] (because these smart cards often contain a unique fixed identifier [12, 19]).444Note that these security weaknesses are not necessarily easily exploitable in practice [24]. In this paper we focus on the privacy issues in e-ticketing for public transport. This problem has been studied in the past for e-ticketing in particular [15, 16], but also for related problems like electronic toll collection systems [10] or more general road pricing systems [17]. Compared to the work of Heydt-Benjamin et al. [15, 16] (which relies on anonymous e-cash and anonymous credentials as building blocks) we use (partially) blind signatures instead to create either unlinkable travel tickets or unlinkable travel credit. This makes the protocols less complex and more efficient as there is no need to spend several e-cash coins to pay the exact fare. Moreover, our approach is inspired by the work of Stubblebine et al. [22], which studied unlikable transactions in practice, with a particular focus on dealing with failures. We present two privacy friendly e-ticketing protocols for public transport. The first protocol emulates paper based tickets. The second protocol implements a pay-as-you-go approach, with fares determined when users check-in and check-out. Both protocols assume the use of a smart phone as the main user device to store tickets or travel credit, without relying on any tamper-proof component. We wish to stress that this means that we put no trust assumptions on the device the user uses to pay for public transport. In both cases we pay particular attention to possible failures and how to graciously deal with them. The remainder of the paper is structured as follows. We introduce the system model, requirements, threat model and other assumptions in section 2. We discuss protecting privacy in practice in section 3, especially the assumption to use smartphones as the primary user device. Section 4 discusses the primitives used in our protocols, that are then presented in section 5 and section 6. Our conclusions are presented in section 7. ## 2 Problem statement We assume a system that supports many modes of transportation. This means we distinguish several _public transport operators (PTOs)_ that offer public transport services. _Users (U)_ travelling by public transport make trips that may consist of several legs each using a different mode of transportation offered by a different PTO. _Inspectors_ on the trains and busses verify that everybody on board has a valid ticket. A central _public transport clearinghouse (PTC)_ provides the public transport ticketing infrastructure (or at least the APIs to connect to this infrastructure), and distributes financial compensation to the PTOs for services rendered. Separate _payment service providers (PSPs)_ handle payments from banks initiated by users. Users have a digital _token_ that enables them to travel by public transport. Instead of relying on a smart card (to store tickets or other information needed to verify whether someone is entitled to a certain mode of public transportation) we assume most people own a sufficiently modern and capable smartphone, and are willing to use it for public transport (we will discuss this further in section 2.2). Any entity (in particular any of the PTOs) can offer an app for this purpose. ### 2.1 Requirements Any public transport ticketing scheme for the model outlined above should satisfy the following requirements. * • Users should pay for trips, where the fare depends on when and where a user travels, the distance travelled, and whether the user has subscribed to a public transport pass that offers reduced fares (during certain times of the day, or on certain tracks). * • Public transport operators should receive compensation for their services, which (partly) depends on all the actual trips made bu users that traversed part of their infrastructure. In other words, the amount of compensation can depend on how many passengers travelled on which particular track on which day. * • The scheme should be privacy friendly: no party should be able to link any number of trips to each other (as belonging to one, unknown, person) or to any one particular person. In other words, the previous requirement only allows the PTO to learn _how many_ passengers travelled a certain track at a certain time of the day, not _who_ they were, or whether they were the _same people_ that set out on some other trip earlier. * • The scheme should be secure: it should prevent or detect fraud by users (e.g. creating fake tickets, paying less than the required fare) and prevent fraud by operators (e.g. claiming more trips than actually took place over their infrastructure). Travelling without a valid ‘ticket’ should be discouraged by regular inspection and appropriate fines. * • The scheme should be fast enough to process large volumes of travellers at peak hours. Checking travellers by conductors should be fast, e.g., take less than a second. Checking in or out to enter or exit public transport (like in the London Oystercard system or the Dutch OV-chipkaart system) should take only a few hundred milliseconds at most. We note that the timing constraints mentioned in the last requirement are important in practice, but performance measurements are unfortunately out of scope for this paper. ### 2.2 Threat model We assume users may try to actively defraud the system (travelling by public transport while paying less than required or nothing at all), if the probability of being caught is low. They will root their smartphones and install fraudulent apps if there is a clear benefit. This means that in terms of smartphones we cannot assume any trusted environment to store secrets (we rule out the possibility that a public transport app gets to use a secure enclave on the smartphone). In other words: the device and the app are untrusted from the PTO perspective. This means our system is weaker than one relying on smart cards that _can_ be used to store secrets and keep them confidential, preventing their users from accessing (and perhaps copying) them. We assume banks, PSPs, PTOs and the PTC will actively (and collectively) try to break privacy and recover trip details from their users, using any information they can get their hands on. They are untrusted from the user perspective. We do assume however that PTOs do _not_ try to break privacy by writing their apps in such a way that the information provided by the user through the app, but shielded from the central PTO servers by the protocol, is surreptitiously sent to the PTOs regardless. The PTOs could in theory do this. We can mitigate this by offering third party (open source) apps, requiring external audits and analysis, or through the vetting procedures enforced by the smartphone app stores. (This, by the way, is another reason why we cannot assume that the smartphone or app can store secrets.) We assume that PTOs will try to defraud the system and claim more compensation from the PTC than warranted. The PTC is trusted, in the sense that it does not favour one PTO over the other, and that at the of the day all money received must be spent (on compensating PTOs or on the cost of running the clearinghouse and its ticketing infrastructure). Audits can be used to ensure this. We assume the cryptographic primitives used cannot be broken, and that entities keep their secrets secret (unless they could benefit from not doing so). ### 2.3 Other assumptions We assume secure, i.e. authenticated and encrypted, connections between all entities. Clearly the user is not authenticated. We assume fares are course enough to ensure that the price associated with a trip does not reveal the actual trip itself. For example, trip prices could be set at fixed amounts for every ten kilometres travelled, with a fixed ceiling fare for all trips longer than a certain distance. (Care should be taken to ensure that for every possible fare the number of different trips with that fare is sufficiently high to guarantee a reasonable degree of anonymity.) We also assume that local device to device communication is using only ephemeral identifiers (if any) to prevent linking devices over longer periods of time. This means WiFi or Bluetooth are using properly randomised MAC addresses, or random anti-collision identifiers if NFC is used. This also implies that we assume apps do not have access to any other permanent, unique, device specific identifier.555The operating systems of these smartphones should, could and sometimes actually do prevent apps from having access to such a persistent identifier. Preventing the app itself to generate such an identifier itself and store it locally is of course not possible (although audits may reveal this). For normal (long range) internet connections between the smartphone and the servers of the other entities we cannot make such an assumption: it is rather trivial to track users based on their often fixed IP addresses. We discuss this in the next section. ## 3 Protecting privacy in practice Protecting privacy in practice is a major challenge, for several reasons. First of all, practical considerations may rule out certain solutions or may make it impossible to make simplifying assumptions. For example a complex tariff system may lead to a situation where particular fares correspond to one or perhaps only a few particular trips. This is not the case when the tariff system is very simple (e.g., two or three different zones in a metro network). This issue is exacerbated when people are forced to pay for individual trips separately (see the first protocol that emulates paper tickets in section 5) while the payment protocol is not anonymous. Except for cash payments (and certain privacy friendly crypto currencies perhaps) existing and widely accepted payment methods (credit card, debit card or e-banking apps) are account based and thus identifying. Even if this were not the case, the protocols detailed below rely extensively on Internet connectivity that by its very nature is identifying. Strategies to shield the user’s IP address from the other parties involved in the ticketing system (like using Tor [9] or mix networks [7])) should be used, but are probably impractical to use extensively and reliably at scale. Then again, letting users use a trusted VPN would solve most of the problems as this would hide all users behind the IP address of the VPN provider. We will have something more to say about this later on. The biggest paradox, from a privacy perspective, is of course the use of a smartphone as the basic user device for buying, storing and using tickets. On the one hand it is an entirely personal device, capable enough to orchestrate the interactions with all other parties using complex privacy friendly protocols, with the possibility of a nice user friendly interface to boot. Moreover, people expect their smartphone to support their day to day activities, like paying in shops, these days. This makes a smartphone the natural, in fact unavoidable, choice as the user device.666Although a fall back option should always be available those people that cannot afford to own a (recently modern) smartphone. But clearly the use of smartphones comes with severe privacy risks. By design, mobile phone operators know the approximate location of all their subscribers (and can zoom in using a process called triangulation). With GPS, standard on smartphones, location is also readily available to the phone itself as well as all apps that were granted permission to location services. With the increasing complexity of smartphones and the huge app ecosystem, users have very little reason to trust their smartphone or to expect it to protect their privacy. With these caveats in mind we follow a pragmatic approach in this paper, aiming for a strong enough technical protection of privacy under reasonable assumptions. No coalition of PSPs, PTOs and the PTC can link trips777Either as bought in the protocol that emulates paper based tickets, or as implied by check-in and corresponding check-out events to users, beyond what can be ascertained by observing the financial transactions of the users, knowledge of the tariff structure, and (partial) apriori knowledge of the travel patterns of a subset of these users. We do not solely rely on technical mechanisms however, but also depend on legal, societal and market incentives to keep the different stakeholders in check. All measures combined should ensure that the cost of obtaining privacy sensitive information in general outweighs the (business) benefit. The tacit assumption in this work is that it is much safer, from a privacy perspective, to collect personal data locally on the user device, instead of centrally on the servers of the service providers. Clearly a malicious public transport app can collect and upload all this personal data surreptitiously. The assumption is that this cannot or will not happen. ## 4 Primitives Our protocols for privacy friendly ticketing for public transport are based on three primitives, that we will describe in this section: partially blind signatures, attribute based credentials, and a mechanism to implement a form of privacy friendly payment with receipt. ### 4.1 Partially blind signatures Blind signatures were introduced by David Chaum almost four decades ago [6], as the fundamental building block to implement a form of untraceable digital cash. His proposal was to represent each digital coin as a unique serial number blindly signed by the issuing bank. The unique serial number embedded in the coin would prevent double spending, while the blind signature over the coin would guarantee both _untraceability_ (by not knowing which coin was signed) and _unforgeability_ (by signing the coins in the first place). In the protocols below we use a generalisation of this idea called partially blind signatures, introduced by Abe and Fujisaki [1] and further investigated and optimised by Abe and Okamoto [2, 20]. In a partially blind signature scheme the messages to be signed consists of a secret part (only known to the user) and a public part (known to both the user and the signer). Issuing a blind signature involves an interactive protocol between the user and the signer, where the user blinds the secret in order to hide it from the signer. In the protocols below we use these partially blind signatures to issue receipts and/or tickets where the receipt number or the trip details are kept secret. Because such receipts and tickets are only used once (in fact, we need to enforce that they are not used more than once), using simple signatures instead of full blown attribute based credentials (to be discussed further on) suffices. When describing our protocols we write $[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}$\mathit{secret}$}\,|\,\mathit{public}]_{{k_{\mathit{}}}}$ when issuing a blind signature over secret part $\mathit{secret}$ and public part $\mathit{public}$ (using private key ${k_{\mathit{}}}$ of the signer to sign it), and write $[\mathit{secret}\,|\,\mathit{public}]_{{k_{\mathit{}}}}$ when subsequently using it (revealing both the secret and the public part). #### Two faces of blindness Chaum explained blind signatures intuitively by showing how a blind signature could be implemented in a traditional, non digital, setting using carbon paper inside paper envelopes. To obtain a blind signature on a secret message, a user could send the message inside a sealed envelope to the signer, with the inside of the envelope covered with carbon paper. The carbon paper ensures that if the signer signs the envelope from the outside, the carbon paper transfers this signature to the secret message inside the envelope. When the signer returns the still sealed envelope (proving it didn’t see the message) all the user needs to do is to open the envelope to obtain the blindly signed message. This intuitive explanation clearly shows that the message stays hidden from the signer. But this by itself is not enough to prevent a bank from tracing a digital coin signed this way, even if it prevents the bank from learning its serial number. In fact, if the bank signs each envelope in a slightly different way, and remembers which way of signing it used to sign each envelope, it can link actual signatures on messages to the particular envelope on which he put the exact same signature. In other words, in order to guarantee untraceability (sometimes also called _unlinkability_), blind signatures need to guarantee two separate blindness properties: message hiding The message to be signed is hidden from the signer. signature unlinkability Given a final blind signature on a message, the signer cannot determine when it generated that particular signature. To see that these are indeed different properties, observe that a scheme where signing the cryptographic hash of message $m$ (without revealing $m$ itself to the signer) is message hiding but clearly not unlinkable. In the protocols below we rely on both these properties to hold. Most (partially) blind signature schemes in fact satisfy both of them. This is in particular the case888The (partial) blindness property is defined using a game where two messages $m_{0}$ and $m_{1}$ are randomly assigned to two users (based on a random bit $b$). Each user then requests a blind signature on its message from the signer. The signer is then given both signatures (and for each the corresponding message) and asked to guess the value of $b$. If it could distinguish which signature corresponds to which user, it could for sure determine the value of $b$. for the schemes of Abe and Okamoto [2, 20] (but not for the blind signature scheme underlying the Idemix attribute based credential scheme [18, 4]). #### Dealing with failures In the protocols below, partially blind signatures are used to represent receipts received after a successful payment, or as public transport tickets received in exchange for a valid receipt. In both cases a kind of ‘fair exchange’ [21] is required between a user and a signer, and there should be a way to recover from errors in case messages are dropped, connections fail, or system components crash, to ensure that either the exchange takes place completely, or that the exchange is cancelled and both parties return to the state before they started the exchange. Recall from section 2.2 that we assume users to be malicious while service providers (the signers in this case) are honest (but curious). This assumption makes it possible and relatively easy to implement a fair exchange in this particular case. Details will vary depending on the particular blind signature scheme used. For example, the partial blind signature scheme of Okamoto [20] consists of the following phases when creating the blind signature $[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}$s$}\,|\,p]_{{k_{\mathit{}}}}$. 1. 1. The user blinds the secret part $s$ using some randomness $r_{u}$ as $b=\mathit{blind}(s,r_{u})$ and sends this to the signer. 2. 2. The user proves to the signer that she knows $s$ and $r_{u}$ used to construct $\mathit{blind}(s,r_{u})$ using a three messages zero-knowledge protocol. The signing protocol aborts if this proof fails. 3. 3. The signer generates some randomness $r_{s}$ and creates an intermediate signature $i=\mathit{intermediate}(b,p,r_{s},{k_{\mathit{}}})$ using its private key ${k_{\mathit{}}}$ over the blinded information $b$ received from the user, the public part $p$ of the to be signed message, and the randomness $r_{s}$ it just generated. The signer sends this intermediate signature to the user. 4. 4. The user transforms this intermediate signature $i$ to the final partially blind signature $[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}$s$}\,|\,p]_{{k_{\mathit{}}}}$. The user acknowledges this to the signer Both the user and the signer keep a record of the values of all local variables used and messages exchanged during the signing protocol, and keep track of when they aborted the protocol. Current values of local variables must be safely stored before sending any message that depends on them. If both parties successfully complete the protocol, both can destroy the record for the protocol. Observe that the only dispute that can occur is when a user claims not to have received a blind signature in return for a payment or a receipt.999This uses the fact that the signer is honest. The idea is that if the signer claims not to have received the payment or the receipt, then any clearing and settlement of the payment or use of such a receipt will be detected later, and would lead to legal measures. Then the following cases have to be considered. * • If the signer aborts before sending $i=\mathit{intermediate}(b,p,r_{s},{k_{\mathit{}}})$ (the intermediate signature), then the protocol can be restarted from scratch. This results in a different blind signature, possibly for a different blind secret input $s$, but the same public input $p$. But since it is guaranteed that the sender never sent the intermediate signature, we are certain the user was never able to obtain a blind signature in the aborted run. * • If the signer aborted after generating the intermediate signature $i=\mathit{intermediate}(b,p,r_{s},{k_{\mathit{}}})$ (and this intermediate signature may or may not have been received by the user), then the protocol must be picked up from this point, with the user using the stored values for the variables used in step 1 and 2 (which should exist by assumption that local variables must be safely stored before sending messages that depend on them. This means that the previously generated intermediate signature $\mathit{intermediate}(b,p,r_{s},{k_{\mathit{}}})$ is sent to the user. This results in possibly a different blind signature, but for the same blind secret input $s$ and same public input $p$ that were used in the aborted run. We conclude that the above sketched dispute resolution protocol allows the user to obtain a valid blind signature of her choice (if the dispute resolution protocol itself does not abort of course), while guaranteeing that a (dishonest) user is never able to obtain two different blind signatures for two different values $s$ and $s^{\prime}$. ### 4.2 Attribute based credentials Partially blind signatures allow the user to hide (part of) the contents of a message to be signed, but must always reveal the full contents of the signed message to allow the signature to be verified. This means that such signatures _only_ break the link between the signing and the verification of the messages, meaning that the act of signing and the act of verifying is unlinkable. Unfortunately, any two acts of verification can still be linked (using the unique data embedded in each signature). For so-called multi-show unlinkability full blown attributed based credentials are required [18, 4]. We will not go into the details here, but only describe the functionality offered by such credentials, and the privacy properties they entertain. Such attributed based credentials are used in the protocols below to implement travel passes and seasonal tickets that offer reduced fares and that, by their very nature, are on the one hand tied to a particular person while on the other hand need to be presented continually to claim a reduced fare. An attribute based credential is a secure container for one or more attributes $a_{1},\ldots,a_{m}$. Credentials are bound to a particular person, and the attribute(s) it contains describe certain properties of that person. (In the current context, it describes the eligibility to certain fare reductions, for example because the person is more than 65 years old, or because the person is a student.) The values for the attributes are negotiated by the requesting person and the _issuer_ $I$ (under the assumption that the issuer knows or can verify that a particular property holds for the person to which the credential is being issued). The issuer also signs the credential, to prevent fraud. We write $C_{I}(a_{1},\ldots,a_{m})$ for the resulting credential that the person obtains. Typically the credential also contains a hidden private ${k_{\mathit{U}}}$ key known only to the user that is hidden from the issuer when the credential is being issued, somewhat similar to how partially blind signatures work. Tying this private key to the credential and requiring its use when showing the credential later (see below) aims to prevent users from sharing their credentials to commit fraud (e.g., when a student allows her younger, non student, brother to use her credential to obtain a reduced fare ticket). We note that such techniques to bind people to their credentials are not fool proof [4]. To prove a certain attribute, the user engages in a so called interactive _showing_ protocol with a verifier using one or more of such credentials. This showing protocol is typically _selective_ : the user can decide which attributes to reveal to (and which ones to hide from) the verifier. This means that the verifier never gets to see the full credential, which would be a bad idea anyway as every credential signature is unique and therefore would allow subsequent uses of the same credential by the same user to be linked. As we want multi show unlinkability, the user and the verifier instead engage in an (interactive) zero knowledge protocol where the user proves to the verifier that she owns a credential signed by a certain issuer, containing a selection of the revealed attributes $A_{r}\subseteq\\{a_{1},\ldots,a_{m}\\}$. This proof also requires the user to know the embedded private key ${k_{\mathit{U}}}$ (without revealing it of course). This reveals the issuer and the attribute values, and nothing more, to the verifier. We write ${k_{\mathit{U}}},C_{I}(a_{1},\ldots,a_{m})\leftrightarrow I,A_{r}\subseteq\\{a_{1},\ldots,a_{m}\\}$ (where the left hand side shows the input of the user, and the right hand side shows what the verifier learns (provided it knows the public key of the issuer needed to verify the proof). ### 4.3 Privacy friendly payment with receipt A basic mechanism used throughout our protocols is the possibility to pay a certain fare ${f}$ to a payment service provider (PSP) and to receive a receipt ${R}$ for this payment in return.101010The PSP could be your bank (provided it knows how to issue receipts as explained below), or a separate entity that lets bank process the payment and generates a receipt when the payment was successful. The receipt can subsequently be used at a (public transport) service provider to pay for transport. The idea is that such a payment mechanism can be implemented in many different (more or less privacy friendly) ways, with only the receipt being standardised for use in the protocols below. To maximise privacy protection in case the payment itself is less privacy friendly, the receipt ${R}=[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}${\textit{rs}}$}\,|\,{f}]_{{k_{\mathit{\textrm{PSP}}}}}$ is a blind signature over the public fare ${f}$ paid as well as a blind receipt sequence number rs provided by the user, signed by the payment service provider PSP that processed the payment. To make explicit at which particular service the receipt can be used, the name of the service can be added, blindly, by the user as well. The user should ensure that each receipt has a different sequence number. This sequence number is used to prevent reuse of receipts: the sequence numbers in redeemed receipts are recorded as spent. (Which also shows why users have every reason to ensure that sequence numbers are indeed different.) Using a blind signature in this way guarantees that users cannot create fake receipts, while the receipt sequence number cannot be linked to the payment (and hence to the user making the payment). Users are expected to properly protect their receipts and keep them securely stored until use. In the protocols below, the paid fare is first collected by the PSP, then forwarded to a public transport clearinghouse (PTC) that later redistributes the paid fares to the PTOs based on submitted receipts the PTOs have collected. Each fare is recorded by the payment service provider (PSP) as a separate payment transaction for the specified amount with the clearinghouse as the recipient. If the payment transaction involves the bank account of the user (see below), care should be taken to _not_ include the bank account details of the user in the transaction towards the clearinghouse. This happens more or less automatically if the PSP is a separate entity independent of the bank (in which case the transaction will transfer the fare amount from the user bank account to that of the PSP). If the bank itself serves as PSP, an internal bank offset account should be used that aggregates individual payments to the PTC with only the daily or weekly totals being transferred to the actual PTC account. This prevents the clearinghouse from learning the bank account (and hence the identity) of all people travelling with public transport, including how often they travel and an indication of the distance they travel (given that the fare is often a good indication of this). One possible way to implement payment when using a smartphone based public transport ticketing app is to redirect the payment phase to a separate payment app on the user’s smartphone, and let PSP forward the resulting receipt back to the transport ticketing app. A more privacy friendly option is to allow travellers to pay with cash at designated kiosks at public transport stations. Or to support the payment of fares using some kind of online privacy friendly payment scheme (like Digicash [8], or Zcash [3]). ### 4.4 Notation When describing the knowledge acquired by parties involved in the (figures depicting the) protocols below, we use expressions like $(a,b,c)$ to denote that a party learns the values $a$, $b$, and $c$, and moreover learns that they are linked and thus belong together. Values in different tuples are not linked, but can however be correlated based on their actual values: if a party learns a specific fare ${f}$ was paid by user $U$ (i.e., it knows $(U,{f})$) and later sees a ticket with that particular fare for a trip $T$ (i.e., it also knows $(\langle r,d\rangle,{f})$, then it may conclude user $U$ travelled route $r$ on date $d$. We use $\hat{U}$ to denote the possibly static IP network address of the user visible to the other parties.111111This equals the VPN server address or the Tor exit node address in case any of these services are used by the user. \begin{overpic}[abs,unit=1pt]{./fig/prot-papertickets-x.eps} \put(233.81453,4.66342){learns $({\textit{rs}},{f})$} \put(60.2531,39.82378){$[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}$T,\textit{ts}$}\,|\,{f}]_{{k_{\mathit{\textrm{PTO}}}}}$} \put(145.611,75.79517){$[T,\textit{ts}\,|\,{f}]_{{k_{\mathit{\textrm{PTO}}}}}$} \put(111.2948,98.93964){\hbox to0.0pt{\hss$[T,\textit{ts}\,|\,{f}]_{{k_{\mathit{\textrm{PTO}}}}}$}} \put(233.09183,49.10245){\hbox to0.0pt{\hss$[{\textit{rs}}\,|\,{f}]_{{k_{\mathit{\textrm{PSP}}}}}$}} \put(226.93683,167.32111){sender anonymous} \put(227.17271,151.09951){not anonymous} \put(6.56553,13.4382){knows $(U,\hat{U},{f},T,{\textit{rs}},\textit{ts})$} \put(3.49004,167.45662){learns $({f},U,\hat{U})$} \put(6.00142,97.24129){${f}$} \put(55.83259,134.22948){${k_{\mathit{\textrm{PSP}}}}$} \put(40.33569,86.21409){$[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}${\textit{rs}}$}\,|\,{f}]_{{k_{\mathit{\textrm{PSP}}}}}$} \put(123.97216,124.16488){learns $(T,\textit{ts},{f})$} \put(175.13129,18.64164){${k_{\mathit{\textrm{PTO}}}}$} \put(110.73671,62.74341){\hbox to0.0pt{\hss$[{\textit{rs}}\,|\,{f}]_{{k_{\mathit{\textrm{PSP}}}}}$}} \put(95.89024,4.79793){learns $(\hat{U},{\textit{rs}},{f}),(T,\textit{ts},{f})$} \end{overpic} Figure 1: Protocol emulating paper tickets ## 5 Solution 1: Emulating paper tickets One way to achieve privacy in public transport ticketing is to emulate the traditional use of paper tickets in public transport. The basic idea is to first buy the ticket online, and subsequently use it for public transport later, in such a way that the financial transaction used to pay for the ticket cannot be linked to the actual trip being made. The protocol assumes that the public transport app on the user’s smartphone contains a database with all possible trips that can be made by public transportation, together with the corresponding fares to be paid. ### 5.1 Detailed protocol The protocol, graphically represented in figure 1, runs as follows. Phase 1: Obtaining a ticket * • The user selects the route $r$ she wants to travel, and the day $d$ on which she wishes to travel. This defines the trip $T=\langle r,d\rangle$. * • The user calculates the fare ${f}=\textit{fare}(T)$ for the trip. (Incorrectly calculated fares will be detected later.) * • The user starts a payment for this fare, and receives a receipt ${R}=[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}${\textit{rs}}$}\,|\,{f}]_{{k_{\mathit{\textrm{PSP}}}}}$ in return. (See section 4.3 above for details.) The paid fare is credited to the PTC account. * • The user sends this receipt to the PTO. The PTO verifies the signature on the receipt, and submits it for clearing and settlement to the PTC. The PTC also checks the signature on the receipt, and checks whether a receipt with sequence number rs has been submitted before. If so, the receipt is rejected. Otherwise, the PTC accepts the receipt and records rs as submitted. * • The user engages in a partially blind signature issuing protocol with the PTO in order to obtain a ticket $T$ for the trip. The user blindly provides the trip $T$ as well as a blind and fresh ticket sequence number ts. The PTO provides the (unblinded) fare ${f}$ present in the receipt it received in the previous step. As a result the user receives the ticket $T=[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}$T,\textit{ts}$}\,|\,{f}]_{{k_{\mathit{\textrm{PTO}}}}}$, signed by the PTO. Phase 2: Travelling by public transport Public transport operators need to verify that all users that travel with them have a valid ticket, with the correct fare. The traveller and the ticket inspector engage in the following protocol for that purpose. * • The user sends the ticket $T=[T,\textit{ts}\,|\,{f}]_{{k_{\mathit{\textrm{PTO}}}}}$ (revealing all its contents) to the ticket inspector. One way to do so in a sender anonymous fashion is to let the public transport app display the ticket as a QR code on the smartphone display, and let the inspector scan this QR code. Many public transport operators use similar schemes to inspect ‘home print’ paper based tickets. * • The ticket inspector verifies the signature on the ticket, whether fare ${f}$ is correct for trip $T=\langle r,d\rangle$, whether the ticket sequence number ts is not invalidated, whether the date $d$ in trip $T$ is today, and whether the route $r$ in trip $T$ covers the leg (of the total trip) where the ticket inspector asks the user to provide a ticket. If the user cannot provide a valid ticket, a fine is issued. * • The ticket inspector verifies the ticket with the PTO. The PTO also checks the signature on the ticket, and checks whether a ticket with sequence number ts has been submitted before. If so, the ticket is rejected. Otherwise, the PTO accepts the ticket and records ts as submitted. (If trips consist of several legs, the same ticket should be accepted for different legs of the trip.) Note that the fact that ticket sequence numbers must be verified and invalidated in real time implies that the equipment of the inspector must be online. As tickets are only valid for a single day, PTOs may choose to forfeit on this strict form of checking (hence relaxing system requirements), relying on the fact that ticket can still only be used (perhaps multiple times) on a single day. Phase 3: Clearing and settlement PTOs are reimbursed based on the payment receipts received in phase 1, after submitting them to the clearinghouse PTC. For each receipt, the PTC verifies the signature, and verifies that the sequence number in the receipt rs is yet unclaimed. If so, the sequence number is recorded as claimed, and the PTC proceeds to pay the fare specified in the receipt to the PTO. Otherwise the claim is rejected. ### 5.2 Analysis To what extent does this solution fit the requirements set out above? Users obviously pay have to for their trips, and the fare depends on the distance travelled. Inspectors and sufficiently high fines are necessary to keep users honest and disincentivise travelling without a valid ticket. Public transport operators get paid based on the payment receipts they collect when issuing tickets. To get (statistical) information about actual trips made they need to have enough conductors to check the tickets of all their passengers when travelling (as this is the only time when the actual trip details are revealed). If multiple PTOs are involved in a particular trip, proper reimbursement can only be achieved if the app splits up the trip in different legs, one for each PTO the user needs to travel with. The level of privacy protection is reasonable, depending on the properties of the network being used. The protocol prevents trip details to be linked to users, in the following sense: for all the tickets a particular PTO sells for a particular fare ${f}$ it learns the set of IP addresses of users that bought a ticket for this particular fare on the one hand, and the set of trips made for this fare (through inspection) on the other hand, but it can never link a particular user address to a particular trip. In the setup described, the PSP and PTOs could however learn how many tickets you buy, and for which amount (ie for which distance), if they would try to identify you based on your (fixed) IP address. Note that this problem becomes much less significant if the payment receipts issued by the PSP can be used for many different types of purchases, i.e., if they are used as a type of generic digital currency. Even if PTOs and PSP collude, they would not be able to link users with actual trips made, but timing analysis linking payment times with ticket issuing times could be used by the PTO to be more certain about your identity. If you usually buy your tickets on the same day or the day before your trip, your PTO could learn when you travel. The PTO could learn whether you are using public transport a lot, or not. Many short trips on the same day may reveal you are in a city; certain patterns of distances may correspond to popular tourist routes (and hence reveal the city you are in). This limited level of privacy protection may already be a threat for people that engage in protests or civil disobedience, like the Hong Kong protesters or the Extinction Rebellion activists. All these problems can be avoided if users can use cash to buy tickets, at special digital kiosks. Communication between the user and inspector is sender anonymous. This means ticket inspection reveals no personal information. The system is secure: tickets are only issued by the PTO when given a payment receipt for a certain amount, signed by a bank. Only banks can create such a signed proof of payment. The amount paid for a ticket is checked by the conductor when inspecting a ticket. Only PTOs can create a valid ticket (signed in partially blind fashion). This signature is also checked by the conductor. Finally the conductor checks whether the ticket entitles a person to travel when and where the conductor inspected her ticket. Failure of one of these tests means the ticket is invalid. The sequence number of the ticket (embedded to guarantee one-time use) is checked in real-time with an online database of sequence numbers of already inspected tickets. If the sequence number is already in the database, the ticket is invalid. Otherwise, the sequence number is added to the database. ### 5.3 Dealing with failures and disputes Dealing with failures is always a challenge, but this is particularly the case in privacy friendly protocols where often the link between a user and her actions is deliberately broken. This means extra care needs to be taken to create some evidence that allows an entity to challenge a failure, while not eroding the privacy of the users. Below we describe some possible failures, and how they could be dealt with. See also [22] for additional measures that can be taken, and the general strategy to deal with failures during the issuing of blind signatures (like users not receiving a payment receipt after payment, or not receiving a ticket after submitting a payment receipt) outlined in section 4.1. The user wants to cancel a payment The user can return the payment receipt (which contains a unique sequence number) to the PSP to rewind the transaction. The PSP then forwards the payment receipt to the PTC signalling not to accept this payment receipt when a user requests a ticket to be issued. Also, the transfer of money from the bank to the PTC will be reversed. The user receives a valid but incorrect ticket This can happen if the user entered the wrong trip details, or if some internal error caused the wrong ticket to be issued. The user can ’return’ the ticket to the PTO, essentially running the showing protocol normally run when a conductor inspects the ticket. This invalidates the ticket. Using the same payment receipt she can start restart the issuing step, now with the correct trip details (assuming the fare is the same). The user wishes to cancel a ticket issued to her After ’returning’ the ticket the PTO as described in the previous case, she can then proceed to cancel the payment to the bank. The user receives an invalid ticket This is more tricky. Ideally the issuing protocol should guarantee that a valid ticket is issued. If this is impossible, at least the issuer should somehow be able to tell, from the logs, that the user indeed did not receive a valid ticket. Otherwise bogus claims for invalid tickets could be submitted. This all very much depends on the particular issuing protocol used. A valid ticket fails conductor inspection Ideally this should not happen. However, the user or conductor device may malfunction, and the communication between the two devices may be erroneous. If the ticket is valid, and the user app operates correctly, the user should at some point be able to convince the PTO she had a valid ticket when travelling. The app crashes or malfunctions This can be mitigated by ensuring that the app can be reinstalled without loosing any stored tickets, or turning them invalid. This requires operating system support, e.g., allowing data to be restored from data associated with a previous install of the application. The user looses or deletes a ticket There is no way to recover from this situation. (Loosing a ticket could happen when inadvertently deleting the whole app together with all its data.) ### 5.4 Variations and extensions #### Using actual paper tickets, or smart cards Instead of relying on users having smartphones, tickets could actually be printed on paper,121212This may sound pedantic, but in fact when trying to emulate something digitally based on how it was done physically, one always has to consider the option that the original, physical, approach simply works better. or be stored on contactless smart cards instead. In this case, a ticket kiosk needs to be used to allow users to select the ticket they need, allow them to pay (by cash or card), and to print the ticket or issue the ticket to the smart card. In the first case, the ticket (with its signature) is printed as QR code, which the inspectors can scan with their smartphone. In the second case, inspectors need to carry NFC enabled smartphones that allow them to scan the smart card and read the ticket (with its signature) from the smart card. This is certainly possible even with cheap smart cards (as it is not involved in any complex cryptographic operation: the inspector checks the signature locally on the device, and the blind signature is generated by the kiosk where the user buys the ticket). Paper tickets can also be obtained at home through a website (web app) that essentially emulates the functionality of the user smartphone app with respect to obtaining a ticket, but at the end of this phase prints the ticket as a QR code instead of storing it. #### Supporting seasonal tickets Reduced fares for public transport pass subscribers or holders of seasonal tickets can be catered for in a privacy friendly manners using attribute based credentials, in which case the attributes in the credential encode the fare reductions the holder is entitled to. The user can obtain such a credential using a protocol similar to that of buying a single ticket, except that in the last step the PTO issues a full blown credential instead.131313A simple blind signature as used for ordinary tickets will not do as the credential will have to be shown multiple times while retaining the desired privacy properties. The issuing protocol would run like this. * • The user selects which type of seasonal ticket she wishes to buy. This defines a set of attributes $a_{1},\ldots,a_{m}$ that define which type of reduction she is entitled to. * • The user calculates the total price ${f}$ for this seasonal ticket. * • The user starts a payment for this amount, and receives a receipt ${R}=[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}${\textit{rs}}$}\,|\,{f}]_{{k_{\mathit{\textrm{PSP}}}}}$ in return. (Again see section 4.3 above for details.) The paid amount is credited to the PTC account. * • The user sends this receipt to the PTO. The user also sends the list of attributes $a_{1},\ldots,a_{m}$ to the PTO. The PTO verifies that the price ${f}$ present in the receipt corresponds to the amount due for this particular set of attributes. The PTO verifies the signature on the receipt, and submits it for clearing and settlement to the PTC. The PTC also checks the signature on the receipt, and checks whether a receipt with sequence number rs has been submitted before. If some of these tests fail, the receipt is rejected. Otherwise, the PTC accepts the receipt and records rs as submitted. * • The user and the PTO engage in a credential issuing protocol for this set of attributes. As a result the user receives the credential $C_{\textrm{PTO}}(a_{1},\ldots,a_{m})$, signed by the PTO. Such a credential can subsequently be used to travel by public transport with a reduced fare. Interestingly enough, the credential is actually irrelevant when buying a ticket (except that the user needs to apply the correct fare reduction based on the particular credential she owns), because the PTO blindly issues a ticket for a particular fare without learning the actual trip the ticket is for. Correctness of the fare paid is only verified at inspection time, when the inspector gets to see the full ticket containing both the trip and the corresponding fare. To prove that the user is entitled to a reduced fare, the inspection protocol needs to incorporate verification of the necessary attributes in the credential as well. As the original inspection protocol is sender anonymous, the credential verification protocol needs to be sender anonymous as well. This can be achieved by using a fully non- interactive credential showing protocol. Idemix [18], for example, uses a non- interactive proof of knowledge, but relies on a verifier generated nonce to guarantee freshness of the proof. Such a verifier generated nonce would be hard to incorporate in our setting, as it would require the equipment of the inspector to send something to the user device (which would either break sender anonymity or would require cumbersome approaches where the user needs to also scan a QR code on the inspector device). Luckily there is a way out of this dilemma: we can use the cryptographic hash of the randomly chosen ticket sequence number ts already present in the ticket as the nonce instead. The fact that in this case the nonce is generated by the prover is not a problem but actually a feature: the proof is now neatly tied to the ticket for which a reduced fare is claimed, and the original ticket inspection protocol already ensures that the same ticket sequence number cannot be used twice. This forces the user to pick a fresh sequence number.141414Note that the user is by no means forced to select the ticker sequence number _randomly_. Hashing it to derive the actual nonce to be used in the credential showing protocol however ensures that the protocol remains secure when the underlying credential showing protocol relies on actual randomness (and not merely freshness) of the nonce. A problem with the approach outlined above is that there is nothing inherently preventing users to pool and share a single credential (offering reduced fares) with a group of users that each ’prove’ possession of the credential to the inspector when necessary. Unless the private key associated with the credential is securely embedded in the user device (using e.g., a piece of trusted hardware to ensure that even the device owner cannot get access to it), this by itself does prevent such credential pooling attacks. This is a general problem of attribute based credentials, and indeed a problem of online digital identity management in general as securely binding actual persons to their online credentials is hard [4].151515Even embedding the private key in a secure enclave does not strictly speaking prevent the owner of the smartphone to share the phone itself with others (although it is surely not an enticing proposition to be without your private phone for several hours). ## 6 Solution 2: Pay as you go, with credit on device A fundamentally different, and increasingly popular approach for letting people pay for public transportation is to store credit on a contactless smart card serving as a public transport pass. People can (re)charge their passes at special kiosk (essentially transferring money from their bank account to their public transport pass) and subsequently pay when entering or leaving their chosen mode of transportation. This typically involves ‘checking in’ at a gate or turnstile when entering the station, or on the platform or in the bus itself, and ‘checking out’ when arriving at the destination or when changing connections. When people check-in, a check is performed to see whether there is enough credit left on the card. If so the location of the check-in is recorded on the card. When checking out, this check-in location is retrieved, and based on the check-out location the fare is computed and deducted from the credit on the card. To detect fare dodgers that travel without checking in, inspection on the trains or the bus is often still necessary, because it is hard to enforce an air-tight system that forces people to check-in or check- out at all times. The main challenge in implementing such a scheme is to ensure that the check-out operation is performed as fast and reliably as possible (given that at busy transportation hubs many people have to the check-out at the same time, and that a transaction involving a contactless public transport pass is prone to interference and failures). If current public transportation pass systems would actually work as just described, there would be no need to study privacy friendly forms of public transport ticketing: if all that the cards contain is user credit, there would not by any privacy issues with such a system. Unfortunately, this is not the case. All systems mentioned above involve cards with unique serial numbers that are recorded when checking in and when checking out, and stored in a central database. As these serial numbers are static, this allows users to be singled out and their public transportation travel patterns to be recorded over the years. What’s worse: these passes are almost always bound to a particular user (either because they are tied to a personal public transport account, or simply because they were recharged using the bank account of the user). The main reason for adding such tracing of passes is to be able to detect fraud and block passes that appear to be spending more credit than they should be spending based on the amounts used to charge them. Here we aim to emulate such a credit-based system in a privacy friendly manner, without needing to rely on tamper proof hardware or secure execution environments to prevent users from committing fraud by tampering with the credit on their tokens (i.e., their smartphones) in their possession. ### 6.1 Detailed protocol \begin{overpic}[abs,unit=1pt]{./fig/prot-PAYG-v2-x.eps} \put(187.93512,11.81414){${k_{\mathit{\textrm{PTO}}}}$} \put(284.3413,30.66255){${k_{\mathit{\textrm{PTC}}}}$} \put(117.1577,149.26766){\hbox to0.0pt{\hss$[\textit{cs}_{i}\,|\,v_{i}]_{{k_{\mathit{\textrm{PTC}}}}}$}} \put(62.03376,69.41835){$[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}$\textit{cs}_{i+1}$}\,|\,v_{i+1}]_{{k_{\mathit{\textrm{PTC}}}}}$} \put(62.48745,126.33398){$[\textit{cs}_{i},\ell,t]_{{k_{\mathit{\textrm{PTO}}}}}$} \put(75.93469,92.36206){\vbox{$[\textit{cs}_{i},\ell,t]_{{k_{\mathit{\textrm{PTO}}}}}$\\\ $\quad[{\textit{rs}}\,|\,v]_{{k_{\mathit{\textrm{PSP}}}}}$}} \put(187.32685,126.33398){accept/reject} \put(235.47272,259.36298){sender anonymous} \put(235.7086,243.14238){not anonymous} \put(186.9926,69.41835){$[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}$\textit{cs}_{i+1}$}\,|\,v_{i+1}]_{{k_{\mathit{\textrm{PTC}}}}}$} \put(250.88631,171.17952){\vbox{learns\\\ $(\ell,t)$ \\\ $(\textit{cs}_{i},v_{i},{f},{\textit{rs}},v,v_{i+1})$ }} \put(9.1803,261.63647){learns $(U,\hat{U},v)$} \put(14.53731,191.77547){$v$} \put(65.50573,228.91522){${k_{\mathit{\textrm{PSP}}}}$} \put(48.87158,180.39395){$[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}${\textit{rs}}$}\,|\,v]_{{k_{\mathit{\textrm{PSP}}}}}$} \put(131.68397,210.16317){learns $(\textit{cs}_{i},\ell,t)$} \put(240.34894,149.09602){\hbox to0.0pt{\hss$[\textit{cs}_{i}\,|\,v_{i}]_{{k_{\mathit{\textrm{PTC}}}}}$}} \put(240.34894,92.05391){\hbox to0.0pt{\hss${f},\textit{cs}_{i},[{\textit{rs}}\,|\,v]_{{k_{\mathit{\textrm{PSP}}}}}$}} \put(181.21301,188.80437){$[\textit{cs}_{i},\ell,t]_{{k_{\mathit{\textrm{PTO}}}}}$} \put(98.19686,171.85806){$[\textit{cs}_{i},\ell,t]_{{k_{\mathit{\textrm{PTO}}}}}$} \put(3.49004,32.43116){\vtop{learns\\\ $(\textit{cs}_{i},v_{i},\ell,t,{f},\ldots$\\\ $\ldots,\text{checkout location},{\textit{rs}},v,v_{i+1})$ }} \end{overpic} Figure 2: Protocol “pay as you go” Each user maintains travel credit on their own device. As the device is not assumed to be trusted or tamper resistant, care must be taken to ensure that users cannot create counterfeit credit, or spend more than they have credit. Therefore, travel credit $C=[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}$\textit{cs}$}\,|\,v]_{{k_{\mathit{\textrm{PTC}}}}}$ is represented by a blind signature of the Public Transport Clearinghouse PTC over the secret (blinded) sequence number cs and the known credit value $v$. Once the credit token is used, the sequence number cs becomes known. In the protocol below the PTC uses this to record the ’state’ of such a token as either checked-in or spent. We assume in this protocol that check-in and check-out use a sender anonymous form of communication, for example by using near field communication with a randomised anti-collision identifier. The protocol runs as follows. Phase 1: Obtaining credit * • The user starts a payment for the amount $v$ she wishes to obtain credit for, and receives a receipt ${R}=[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}${\textit{rs}}$}\,|\,v]_{{k_{\mathit{\textrm{PSP}}}}}$ in return. (See section 4.3 above for details.) The paid amount is credited to the PTC account. * • The user can use this receipt to add the credit to her device when checking out (see phase 3 below).161616A separate protocol between the user device and the PTC to add credit is also possible, but is not discussed here. Phase 2: Check-in to start a trip * • The user sends her credit token $C=[\textit{cs}_{i}\,|\,v_{i}]_{{k_{\mathit{\textrm{PTC}}}}}$ to the check-in device. * • The check-in device verifies the signature on the credit token, checks that the stored value $v_{i}$ is larger than some minimum credit required,171717This is necessary to prevent users to accrue (too much) negative credit by checking in with hardly any credit and going on an expensive trip. and submits it to the PTC. The PTC verifies the signature on the credit token, and checks whether $\textit{cs}_{i}$ is recorded as spent or checked-in. If so, the credit token is rejected. Otherwise it is accepted and the PTC records $\textit{cs}_{i}$ as checked-in, and records the associated value $v_{i}$ necessary when issuing a new credit token at check out. * • The check-in device sends the user a check-in token $\textrm{I}=[\textit{cs}_{i},\ell,t]_{{k_{\mathit{\textrm{PTO}}}}}$, containing the check-in location $\ell$, the check-in time $t$ and the credit token sequence number $\textit{cs}_{i}$, all signed by the PTO. The PTO logs the tokens for bookkeeping purposes. The user stores the check-in token. Phase 3: Check out to finish a trip * • The user sends her check-in token $\textrm{I}=[\textit{cs}_{i},\ell,t]_{{k_{\mathit{\textrm{PTO}}}}}$ to the check-out device. * • If the user wants to add additional credit to her device, she also submits a receipt ${R}=[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}${\textit{rs}}$}\,|\,v]_{{k_{\mathit{\textrm{PSP}}}}}$ obtained earlier. * • The check-out device verifies the signatures on both tokens, and validates the time on the check-in token (i.e., checks whether the check-in time $t$ and check-in location $\ell$ make sense given the check-out time and the check-out location). * • Given the check-in time $t$, the check-in location $\ell$, the check-out time and the check-out location, the check-out device computes the fare $f$. * • The check-out device then submits the fare ${f}$, the credit sequence number $\textit{cs}_{i}$, and the (optional) receipt $[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}${\textit{rs}}$}\,|\,v]_{{k_{\mathit{\textrm{PSP}}}}}$ to the PTC. (The PTO signs this transfer). The PTC also verifies the signature on the receipt, and checks whether $\textit{cs}_{i}$ is recorded as checked- in. If not, the check in is rejected. Otherwise it is accepted and the PTC records $\textit{cs}_{i}$ as spent. The PTC retrieves the associated credit $v_{i}$ (stored when the credit token was submitted at check-in) and computes the new credit $v_{i+1}=v_{i}+v-{f}$. (Negative credit is possible, but controlled through the credit check at check-in.) * • The user engages in a partially blind signature issuing protocol with the PTC, using the check-out device as a relay, in order to obtain an updated credit token $C_{i+1}$. The user provides a blind and fresh credit sequence number $\textit{cs}_{i+1}$. The PTC provides the (unblinded) credit value $v_{i+1}$ it just computed. As a result the user receives the new credit token $C_{i+1}=[\hbox{\pagecolor{lightgray}\rule[-1.5pt]{0.0pt}{8.5pt}$\textit{cs}_{i+1}$}\,|\,v_{i+1}]_{{k_{\mathit{\textrm{PTC}}}}}$. * • The PTC proceeds to pay the fare to the PTO. (This can be done in bulk.) * • The user stores the new credit token. The user also logs the check-in in a local trip history (that can be consulted to resolve disputes). It may verify locally whether the deducted fare is correct. Phase 4: Inspection The inspector needs to verify that every person travelling has a valid check- in token. * • The user sends her check-in token $[\textit{cs}_{i},\ell,t]_{{k_{\mathit{\textrm{PTO}}}}}$ to the inspector. * • The inspector verifies the signatures on the token, validates the time on the check-in token (i.e., checks whether the check-in time $t$ and check-in location $\ell$ make sense given the inspection location, and submits the check-in token to the PTC. The PTC also verifies the signature on the check-in token, and checks whether $\textit{cs}_{i}$ is recorded as checked-in. If not, the credit token is rejected. Otherwise it is accepted and the PTC records $\textit{cs}_{i}$ as inspected. (If this particular token is encountered by a different inspector, on a leg of the trip that is inconsistent with earlier inspections of the same token, then fraud is assumed.) Note that the fact that credit must be verified in real time implies that the equipment of the inspector must be online, communicating with the PTC (and not the PTO). Phase 6: Clearing and settlement The PTO logs all check-in and credit tokens submitted during check out. The PTC pays the fare as soon as it receives the check-out token and computes the new credit token. (It may accumulate fares to pay the total amount every day or week.) The PTO verifies the payments it receives with the logs it keeps. ### 6.2 Practical considerations As discussed in the introduction of this section, the main challenge in practice of is to make check-in and check-out as fast (and reliable) as possible. Reliability can be improved in the above protocol by adding an acknowledgement message back from the user to the check-in or check-out device whenever the check-in or check-out token have been received in good order, and letting the check-in or check-out device generate an appropriate sound as confirmation. The user device itself could confirm proper check-in or check-out immediately after receiving the token (and sending the acknowledgement), or sound an alarm when the expected token is not received within a short timeout. But an additional message does increase the time needed to check-in or check-out, and adds another point of failure as well: what to do if the acknowledgement message itself is not delivered? Check-in speed is constrained both by the real time connection between the user device and the check-in device, and the real time connection between the check-in device and the PTC which needs to verify that the credit token is not double-spent. This check could be made asynchronous, and the check-in token be issues optimistically, at the expense of ramping up inspection within the public transportation system to detect people that checked in with such a double spent credit token. Alternatively, when there are not too many check out devices, optimistically issued check-in tokens can be revoked when necessary by blacklisting the embedded credit sequence number cs and sending this to all check-out devices. The check-in token itself involves computing a basic signature over the credit sequence number sent by the user device, after verifying the blind signature over the credit token. This should not prove to be an issue in practice. Check-out is more complex as it involves issuing a blind signature over the new credit, where the check-out device works as a relay between the user device and the PTC. Check-out speed can be significantly improved by decoupling the issuing of the check-out token from updating the credit on the user device, doing it ’lazily’ after check-out with a separate protocol that runs between the user device and the PTC. In this case a basic check-out token containing the fare can be issued by the check-out device, with an ordinary signature (instead of a blind one). To protect user privacy however, care needs to be taken to then hide the user address from the PTC to prevent it from linking the previous credit sequence number $\textit{cs}_{i}$ to this used address (as this allows the full trip to be linked to a particular user). ### 6.3 Dealing with failures and disputes Beyond the failures and disputes for protocol 1 discussed in section 5.3, the use of of check-in and check-out devices poses additional challenges. Also the fact that credit is stored on the user device makes the solution more fragile and risky for the user. Dispute resolution depends on clear information about what happened about the time a failure occurred. Unfortunately, due to their privacy friendly nature, the protocols retain very little useful information by themselves. Adding timestamps to local logs of each protocol step, by the PTC, the PTO, and the user device will help compare logs in case of disputes (and detect possible fraud). Creating append only logs (using hash chaining techniques) increases their integrity, especially if occasional public commitments to the current state of the log are recorded. A hash of the log on a user device can be submitted when checking in and checking out, and be included in the check in and check out token (that are signed by the PTO). This poses no linkability as the log will be updated with every check in and check out, provided such updates always contain some private information from the user device (e.g. the serial number used in the next credit token). To aid dispute resolution, the PTO could also issue a separate check-out $[\textit{cs}_{i},\ell,t,{f}]_{{k_{\mathit{\textrm{PTO}}}}}$ to the user when she checks out, containing the credit sequence number, check-out location, time, and fare, signed by the PTO. This allows the user to verify the correctness of the new credit token she receives when checking out (and allows her to check that the correct check-out information is used to compute the fare). Check in fails If it is a communication error in the first step, the user can try again. Otherwise, if the credit token fails to verify the user needs to start a dispute resolution (if she believes the credit token should be valid). If the credit token is accepted, but subsequent steps fail, dispute resolution should clear the recorded serial number for the credit token from the clearinghouse database to ensure it is valid the next time the user checks in. Check out fails If it is a communication error, the user can try again. Otherwise, if the check-in token (or payment receipt) fails to verify the user needs to start a dispute resolution (if she believes the check-in token should be valid). If the check in token is not accepted, dispute resolution needs to determine whether the user actually tried to check in earlier, or did not. If the user did not get an error when checking in, for sure the PTO log will contain the serial number of the current credit token. Fare dispute After checking out user discovers that the fare paid does not correspond to the fare due for the trip made. The user should submit a piece of the log with all entries involving the check in and corresponding check out for this trip (which should follow each other immediately in the user device log, and are thus linked through the internal hash chain). This is then matched with the corresponding logs of the PTO and clearinghouse. Any discrepancy can be compensated by adding it to the current credit on the device by issuing a new credit token. This can be done even after the user has made other, more recent, trips. ### 6.4 Analysis The security analysis is similar to that of the previous protocol, as presented in section 5.2. We therefore focus on the privacy aspects here. A significant improvement over the previous protocol is that neither the PTO nor the PTC obtains information about the user identity: communication with the inspector and the check-in or check-out devices is sender-anonymous. PTOs can link check-in and check-out location (and hence trips) to credit sequence numbers, but PTOs cannot link these to anything else (either on their own, or when colluding with others). Credit sequence numbers are in essence ephemeral identifiers. The situation changes slightly when a user decides to buy additional credit and to add it when checking out. In that case, the PTC learns also the value $v$ of the additional credit which can be linked to a particular user when thew PTC colludes with the PSP and the particular credit bought is more or less unique. This can be mitigated by allowing users only to buy predefined values of credit, thus ensuring a reasonable anonymity set of users all buying the same credit at roughly the same time. (We implicitly assume here that a user buys credit well in advance to prevent timing correlation attacks.) The situation also changes when the credit values stored in a credit token are unique. This would allow the PTO to link a check-in with credit $v_{i}$ with a subsequent check-in with credit value $v_{i+1}$. With a bit of ’luck’ a PTO might be able to link several trips made by the same user this way. This chain is severed as soon as a common credit value is reached,181818Under reasonable assumptions this would not be a concern in practice. Suppose the maximum credit is $€100$ and fares are multiples of $10$ cents, then there are $1000$ different possible credit values. If there are one million users, the anonymity set would on average contain $1000$ people (although the distribution is probably skewed with larger anonymity sets for smaller credit values). The system could also define some default credit value options (like $€25$, $€50$, and $€100$) and nudge users to always top-up their credit to these defaults. The anonymity sets for these particular values would then be much larger. or when a user decides to travel with a different, non colluding, PTO. ### 6.5 Pay as you go, paying later Given the potential benefits of ‘pay as you go’, it would especially be nice to allow users to pay for their trips afterwards, instead of forcing them to lock significant funds on the device itself. However, introducing a pay later option creates a risk for PTOs as users may fail to pay their debts, so mitigation strategies need to be considered. The basic idea is use to the same protocol, but allowing negative credit. The main risk is that users use their device up to the maximum negative credit, then de-install the app from their smartphone, and then reinstall a fresh one with a balance of zero. To counter such sybil-like attacks, reinstalling an app should be hard. One way to do so is to tie the install to your device identity or app store identity. In that case the app provider or even the app store itself could start asking questions when someone repeatedly installs the app. But this is not as straightforward as it seems, because ideally we want to allow arbitrary third parties to provide public transport apps (to increase trust). One idea is that any (third party) app must be ’blessed’, by the clearinghouse, with an ’admission credential’. In other words, a user can install any app he or she desires, but all protocols outlined above first verify whether the user has a valid admission credential. The user can obtain this credential, through the app, by registering the app with the clearinghouse.191919However, there should be a way to tie this credential to the specific device being used, to prevent cloning. This registration process requires the user to prove his or her identity (for example using a government wide digital identity scheme). Note that relying on such an approach is risky, as it undermines the main message that the public transport app is privacy friendly: if that is supposed to be the case, why does it require me to sign in with a government approved digital identity? The admission credential is special, because it can be blacklisted: the clearinghouse keeps information about all credentials it issued so that when a user wants to obtain a new admission credential (because he or she claims to have lost their phone, reinstalled the app or whatever), then the previous admission credential becomes blacklisted. Information about the blacklisted credential is sent to all PTOs so that when they check whether some user has a valid admission credential (in the first step of each protocol), this will fail for all blacklisted credentials. Note however that this will not deteriorate the privacy protection offered by the protocols, at least not for users without blacklisted credentials: for every credential that is _not_ blacklisted, the PTOs have no way to trace or link valid admission credentials that are not yet blacklisted. The exact privacy properties depend on the specific method to blacklist credentials: a naive scheme might allow the clearinghouse to share blacklisting information about _all_ users to the PTO to make them all traceable. The most privacy friendly scheme doesn’t even allow blacklisted users to be linked or identified [5, 23]. ## 7 Conclusions In this paper we explored options how to implement privacy friendly ticketing for public transport in practice. We show that this certainly possible, with certain constraints (or issues that deserve further study, see below). Starting point is the observation that from a privacy perspective it is better to collect personal data locally on the user device, instead of centrally on the servers of the service providers. Two different approaches (buying tickets beforehand, and pay as you go) have been studied. We show that these can be implemented with reasonably good privacy properties, under reasonably practical assumptions. In particular we show that an untrusted smartphone can be used as the ’token’ to carry tickets or travel credit. This allows third parties to provide the apps for that purpose, which should increase the (perceived) trust of the overall system. For the second protocol, there are rather strict requirements on the maximum checking in and checking out time (in the order of 200-300 milliseconds); actual implementations of the protocols proposed are necessary to verify whether these requirements can be met. One meta conclusion of this work is that we need an efficient, frictionless, way to provide sender anonymity on the Internet, similar to the use of randomised MAC addresses on local networks. A VPN is too weak (the VPN provider sees everything its users do), yet Tor is too strong (there is no need to protect against a NSA like adversary) given the impact on performance. If randomised client IP addresses could be used by default to set up a TCP connection between a client and a server, that would already provide a tremendous boost in privacy on the Internet as servers can no longer trace their users based on their IP address. There are some proposals for temporary IPv6 addresses that partially address this issue [13], but these only apply to larger subnets and do nothing to hide the often fixed IP addresses of private xDSL connections. The second meta conclusion of this work is that there is a need to make apps (or data in apps) _uncloneable_ , so that they can be used in similar contexts and with similar properties as smart cards. Moreover, there should be a secure way to establish that the person holding the phone and/or using the app is indeed the owner of the phone (and not someone that uses the phone with permission of the real owner). These properties are also mandatory to increase the security of attribute credentials, in particular to prevent the attributes in them being pooled or shared. This seems challenging if at the same time we want the apps to be open source. One idea is to use either the SIM card present in most smartphones, or to use the secure element present in most modern smartphones. ## References * [1] M. Abe and E. Fujisaki “How to date blind signatures” In _Advances in Cryptology – ASIACRYPT 96_ , Lecture Notes in Computer Science 1163 Springer, 1996, pp. 244–251 * [2] Masayuki Abe and Tatsuaki Okamoto “Provably Secure Partially Blind Signatures” In _Advances in Cryptology – CRYPTO 2000_ , Lecture Notes in Computer Science 1880 Springer, 2000, pp. 271–286 DOI: 10.1007/3-540-44598-6 * [3] Eli Ben-Sasson et al. “Zerocash: Decentralized Anonymous Payments from Bitcoin” In _IACR Cryptol. ePrint Arch._ 2014, 2014, pp. 349 URL: http://eprint.iacr.org/2014/349 * [4] Jan Camenisch and Anna Lysyanskaya “An Efficient System for Non-transferable Anonymous Credentials with Optional Anonymity Revocation” In _Advances in Cryptology – EUROCRYPT 2001_ , Lecture Notes in Computer Science 2045 Springer, 2001, pp. 93–118 DOI: 10.1007/3-540-44987-6 * [5] Jan Camenisch and Anna Lysyanskaya “Dynamic Accumulators and Application to Efficient Revocation of Anonymous Credentials” In _Advances in Cryptology – CRYPTO 2002_ , Lecture Notes in Computer Science 2442 Springer, 2002, pp. 61–76 * [6] David Chaum “Blind Signatures for Untraceable Payments” In _Advances in Cryptology – CRYPTO ’82_ Plenum Press, New York, 1982, pp. 199–203 * [7] David Chaum “Untraceable electronic mail, return addresses, and digital pseudonyms” In _Communications of the ACM_ 24.2, 1981, pp. 84–88 * [8] David Chaum, Amos Fiat and Moni Naor “Untraceable Electronic Cash” In _Advances in Cryptology - CRYPTO ’88, 8th Annual International Cryptology Conference, Santa Barbara, California, USA, August 21-25, 1988, Proceedings_ 403, Lecture Notes in Computer Science Springer, 1988, pp. 319–327 DOI: 10.1007/0-387-34799-2\\_25 * [9] Roger Dingledine, Nick Mathewson and Paul F. Syverson “Tor: The Second-Generation Onion Router” In _13th USENIX Security Symposium_ USENIX Association, 2004, pp. 303–320 * [10] Valerie Fetzer et al. “P4TC – Provably-Secure yet Practical Privacy-Preserving Toll Collection” In _PoPETs_ , 2020, pp. 62–152 * [11] Flavio D. Garcia et al. “Dismantling MIFARE Classic” In _Computer Security - ESORICS 2008, 13th European Symposium on Research in Computer Security, Málaga, Spain, October 6-8, 2008. Proceedings_ 5283, Lecture Notes in Computer Science Springer, 2008, pp. 97–114 DOI: 10.1007/978-3-540-88313-5\\_7 * [12] Simson L. Garfinkel, Ari Juels and Ravi Pappu “RFID Privacy: An Overview of Problems and Proposed Solutions” In _IEEE Security & Privacy_ 3.3, 2005, pp. 34–43 * [13] F. Gont, S. Krishnan, T. Narten and R. Draves “Temporary Address Extensions for Stateless Address Autoconfiguration in IPv6” RFC Editor, Internet Requests for Comments, 2020, pp. 1–22 URL: https://tools.ietf.org/id/draft-ietf-6man-rfc4941bis-10.txt * [14] Ivan Gudymenko “Privacy-preserving E-ticketing Systems for Public Transport Based on RFID/NFC Technologies”, 2015 URL: https://d-nb.info/1073206866/34 * [15] Thomas S. Heydt-Benjamin, Hee-Jin Chae, Benessa Defend and Kevin Fu “Privacy for Public Transportation” In _Privacy Enhancing Technologies, 6th International Workshop, PET 2006, Cambridge, UK, June 28-30, 2006, Revised Selected Papers_ 4258, Lecture Notes in Computer Science Springer, 2006, pp. 1–19 DOI: 10.1007/11957454\\_1 * [16] Gesine Hinterwälder et al. “Efficient E-Cash in Practice: NFC-Based Payments for Public Transportation Systems” In _Privacy Enhancing Technologies - 13th International Symposium, PETS 2013_ , Lecture Notes in Computer Science 7981 Springer, 2013, pp. 40–59 * [17] Jaap-Henk Hoepman and George Huitema “Privacy Enhanced Fraud Resistant Road Pricing” In _What Kind of Information Society? Governance, Virtuality, Surveillance, Sustainability, Resilience_ , IFIP Advances in Information and Communication Technology 328 Springer, 2010, pp. 202–213 * [18] IBM Research Zürich Team “Specification of the Identity Mixer Cryptographic Library”, 2012 * [19] Ari Juels “RFID security and privacy: A research survey” In _IEEE Journal on Selected Areas in Communications_ 24.2, 2006, pp. 381–394 * [20] Tatsuaki Okamoto “Efficient Blind and Partially Blind Signatures Without Random Oracles” In _Theory of Cryptography (TCC) 2006_ , Lecture Notes in Computer Science 3876 Springer, 2006, pp. 80–99 DOI: 10.1007/11681878 * [21] Henning Pagnia, Holger Vogt and Felix C. Gärtner “Fair Exchange” In _The Computer Journal_ 46.1, 2003, pp. 55–75 * [22] Stuart G. Stubblebine, Paul F. Syverson and David M. Goldschlag “Unlinkable serial transactions: protocols and applications” In _ACM Transactions on Information and System Security_ 2.4, 1999, pp. 354–389 * [23] Patrick P. Tsang, Man Ho Au, Apu Kapadia and Sean W. Smith “Blacklistable Anonymous Credentials: Blocking Misbehaving Users without TTPs” In _Int. Conf. on Computer and Communications Security (CCS) 2007_ ACM, 2007, pp. 72–81 * [24] Undisclosed authors “Security analysis of the Dutch OV-Chipkaart”, 2008 URL: http://www.translink.nl/media/bijlagen/nieuws/TNO_ICT_-_Security_Analysis_OV-Chipkaart_-_public_report.pdf
###### Abstract The paper presents the simulation studies of 10 $\mu$m pitch microstrips on a fully depleted monolithic active CMOS technology and describes their potential to provide a new and cost-effective solution for particle tracking and timing applications. The Fully Depleted Monolithic Active Microstrip Sensors (FD- MAMS) described in this work, which are developed within the framework of the ARCADIA project, are compliant with commercial CMOS fabrication processes. A TCAD simulation campaign was performed in the perspective of an upcoming engineering production run with the aim of designing FD-MAMS, studying their electrical characteristics and optimising the sensor layout for enhanced performance in terms of low capacitance, fast charge collection and low-power operation. A very fine pitch of 10 $\mu$m was chosen to provide very high spatial resolution. This small pitch still allows readout electronics to be monolithically integrated in the inter-strip regions, enabling the segmentation of long strips and the implementation of distributed readout architectures. The effects of surface radiation damage expected for total ionising doses of the order of 10 to 105 krad were also modelled in the simulations. The results of the simulations exhibit promising performance in terms of timing and low power consumption and motivate R&D efforts to further develop FD-MAMS; the results will be experimentally verified through measurements on the test structures that will be available at the beginning of 2021. ###### keywords: Particle detectors; silicon detectors; monolithic sensors; microstrip sensors; CMOS; TCAD simulations; fast timing. xx 1 5 Received: date; Accepted: date; Published: date Fully Depleted Monolithic Active Microstrip Sensors: TCAD simulation study of an innovative design concept Lorenzo De Cilladi 1,2,*, Thomas Corradino 3,4, Gian-Franco Dalla Betta 3,4, Coralie Neubüser 4 and Lucio Pancheri 3,4 on behalf of the ARCADIA collaboration Firstname Lastname, Firstname Lastname and Firstname Lastname Correspondence<EMAIL_ADDRESS> ## 1 Introduction Charged particle tracking and timing are fundamental tools for both physics research and for numerous applications. Although a number of detection techniques are available, silicon detectors have become largely employed due to their versatility and to the parallel strong developments of the semiconductor industry. Various flavours of silicon sensors have been developed to meet the specific requirements of different experiments and applications, such as high spatial resolution, fast charge collection, low power consumption, high radiation tolerance and low cost per unit area. Silicon detectors are divided in two categories, namely hybrid and monolithic detectors. The former are made of two separate silicon elements, the sensor and the chip, which are interconnected through external bump or wire bonding. While the sensor hosts the sensing volume only, the chip integrates the front- end readout electronics. On the contrary, monolithic sensors, which are emerging as a valid alternative to hybrid detectors, embed the front-end electronics in the same silicon substrate which hosts the sensing volume, with benefits in terms of material budget, production yield and fabrication cost, as they are produced with commercial microelectronics processes wermes (2009, 2019); garcia (2018). Due to their characteristics, monolithic sensors have recently raised a wide interest in different research fields; studies, proposals and developments have been made for applications in high energy physics (HEP) mager (2016); pernegger (2017); wang (2017), X-ray imaging wunderer (2014); hatsui (2013), medical particle imaging mattiazzo (2018) and space experiments scotti (2019). The state of the art includes three main types of monolithic sensors. The first type, called Depleted Field Effect Transistors (DEPFETs), is capable of low noise operation thanks to low sensor input capacitance rummel (2009). DEPFET detectors have been developed and used for HEP applications marinas (2011), for X-ray imaging in space treis (2010) and for free electron laser experiments lutz (2010). The main limitation of DEPFETs is the need to reset their internal gate which can be quickly saturated by the leakage lutz (2016), thus making this technology not suitable for environments with high levels of non-ionising radiation. A second approach consists in the SOI (Silicon-On-Insulator) monolithic sensors. SOI sensors embed a buried-oxide layer separating a thin low- resistivity silicon layer, which hosts the integrated readout circuitry, from a thicker high-resistivity substrate, which serves as the sensitive detection region kucewicz (2005); lan (2015). This technology allows a low capacitance to be obtained lan (2015); however, SOI sensors suffer from back-gate effect and have a reduced radiation hardness, due to accumulation of positive holes charges in the buried oxide layer after irradiation hagino (2020). Strategies have been found to overcome these limitations and to recover from the Total Ionising Dose (TID) asano (2016), but, as a consequence, the fabrication process of SOI sensors have become highly specialised and not compliant with standard microelectronics production processes. This results in increased cost per unit area, which is a critical issue for large-area detector applications. A third flavour of monolithic sensors is represented by CMOS sensors turchetta (2003). CMOS sensors were already in use for light detection when they were first proposed for charged particle tracking at the beginning of the 2000s turchetta (2001). Over the last years, important advancements in CMOS sensors allowed them to be employed in many applications, eventually leading to very large scale productions for particle trackers at collider experiments. The STAR pixel detector, which took data at the Relativistic Heavy Ion Collider (RHIC) from 2014 to 2016, was the first large area monolithic pixel tracker ever built, for a total of 0.16 m2 contin (2018). These dimensions have been exceeded by the newly-constructed Inner Tracking System of the ALICE experiment at CERN, in which a total detector surface of about 10 m2 is covered by ALPIDE CMOS monolithic active111A monolithic sensors is called “active” if it integrates a signal amplifier inside each pixel or strip. pixel sensors (MAPS) mager (2016). These achievements demonstrate the level of maturity and reliability that CMOS sensors have recently reached. However, there is still room for further improvements, especially in terms of charge collection speed and radiation hardness, and possibility to push previous limits in terms of low power density, high spatial resolution and SNR. Pixel detectors are the first choice for small scale applications and for vertex trackers at collider experiments wermes (2009) as they have an intrinsic capability of providing a two-dimensional position information rossi (2006). On the other hand, microstrip sensors peisert (1992) are largely used as particle detectors for space applications and are a competitive option for particle trackers due to their high spatial resolution, simpler readout and much lower power density (i.e. power consumption per unit area) compared to pixel detectors. Large experiments at particle colliders have largely employed silicon hybrid strip sensors in the past, and are still developing and assembling new large-area trackers based on this technology, as in the case of the Phase-2 Upgrades of the CMS Outer Tracker chowdhury (2020) and of the ATLAS Strip Inner Tracker david (2018). Recent space experiments equipped with silicon hybrid microstrip trackers include FERMI-LAT atwood (2007), DAMPE azzarello (2016), PAMELA adriani (2003) and AMS-02 lubelsmeyer (2011). Strip- like sensors integrated in a monolithic technology have been proposed by combining the outputs of 55 $\mu$m $\times$ 55$\mu$m benhammadi (2019) or 40 $\mu$m $\times$ 600$\mu$m han (2020) pixels in each column or row of a pixel matrix. Spatial resolution of 1.25-1.3 $\mu$m was achieved using hybrid silicon microstrip sensors with 25 $\mu$m pitch straver (1994). However, it has recently been demonstrated with fully depleted double-SOI monolithic pixel sensors that the 1 $\mu$m limit can be exceeded by semiconductor detectors sekigawa (2017). The keys to a high spatial resolution with analogue readout are a fine microstrip pitch, a low sensor thickness to reduce Coulomb scattering and delta-ray emission, and an increased SNR, which can be achieved by reducing the leakage current and the sensor input capacitance to the readout electronics, but which is ultimately limited by the noise of the front-end electronics turchetta (1993); peisert (1992); straver (1994). This paper presents the first investigation, design and simulation studies of CMOS Fully Depleted Monolithic Active Microstrip Sensors with 10 $\mu$m pitch for charged particle detection. Properly optimised sensor layouts may allow sub-micron resolution, improved radiation hardness and fast timing performance thanks to full depletion wang (2017); snoeys (2017) in a power-saving and cost-effective commercial technology. Moreover, a further advantage of monolithic microstrips is the potential complexity reduction of the detector assembly compared to hybrid microstrip detectors. In fact, since many readout functions can be monolithically integrated on the same chip which hosts the sensing volume, 1-by-1 strip bonding to the external readout electronics would not be needed anymore. We have hence studied and designed the FD-MAMS within the framework of the INFN ARCADIA project in order to provide an innovative solution for satellite-based space trackers and for large area particle detectors at future collider experiments. The results of the Technology Computer-Aided Design (TCAD) simulation campaign222The TCAD simulations were produced using the Synopsys ® Sentaurus (Version O-2018.06-SP2) software. which allowed different MAMS design flavours to be compared in terms of sensor capacitance, reference voltage values, leakage current and charge collection time and efficiency are presented; the effects of the inclusion of a silicon dioxide (SiO2) layer on top of the sensor and of surface radiation damage on the sensor operating parameters are explored; the study of charge sharing between groups of adjacent strips when particles with different Linear Energy Transfer (LET) traverse the sensor is reported. A selection of the MAMS presented in this paper is going to be implemented in test structures which were submitted in November 2020 for an engineering production run. The paper is organised as follows: Section 2 presents the sensor concept for the ARCADIA fully depleted CMOS monolithic microstrip sensors and illustrates the simulation campaign that was performed for the sensor design optimisation; Section 3 describes and discusses the results of the simulations; Section 4 presents the conclusions, the future perspectives and the planned tests for the ARCADIA monolithic microstrip sensors. ## 2 The ARCADIA sensor concept The ARCADIA project and its precursor, SEED (Sensor with Embedded Electronics Development), designed an innovative sensor concept pancheri (2019, 2020) based on a modified 110 nm CMOS process developed in collaboration with LFoundry and compatible with their standard 110 nm CMOS process. Up to 6 metal layers can be stacked on top of the sensor, for a total metal and insulator thickness of about 4-5 $\mu$m. The ARCADIA collaboration is developing a scalable event-driven readout architecture to cover large detection surfaces (O(cm2)) while maintaining ultra-low power consumption. The target for pixel sensors is 10-20 mW/cm2 at high rates O(100 MHz), but for less dense particle environments (e.g. in space applications) a dedicated low-power operation mode implements a cyclic pulling of the data packets from each section of the pixel matrix and disables most of the serialisers and data transceivers, further reducing the total power consumption of the chip. In our project, an n-on-n sensor concept enabling full substrate depletion over tens or hundreds of microns and allowing full CMOS electronics to be implemented was employed. A simplified view of the sensor cross section is visible in Figure 1. The process allows to achieve sensor thicknesses from 50 to 400 $\mu$m. A high resistivity n-type substrate was used and constitutes the active volume. The sensing n-well node, located on top of the sensor, collects the electrons produced by ionisation due to particles traversing the active detection volume. N-doped and p-doped wells intended to host pMOSFETs and nMOSFETs respectively are shielded by a deep p-well, which allows the integration of full CMOS electronics and, hence, more complex digital functions, when necessary. In fact, the deep p-well prevents the n-wells hosting pMOSFETs from competing with the n-doped sensing node in the collection of the charge, thus avoiding loss of charge collection efficiency. Figure 1: ARCADIA monolithic sensor concept. The dotted arrows indicate the drift path of electrons (e) and holes (h) generated by a particle crossing the sensor. The voltages Vnwell and Vback applied to the sensor contacts are shown in green. A p+ boron-doped region sits at the backside of the n-substrate, thus forming a pn-junction; when a negative bias voltage Vback is applied to the backside p+ contact, sensor depletion starts from the pn-junction at the bottom of the sensor and eventually extends to the whole sensor, if the backside voltage is sufficiently large. Since the high voltage needed for sensor depletion is applied at the backside, it is possible to maintain the voltage Vnwell applied to the front n-well electrode below 1 V and to use low-voltage integrated electronics (1.2 V transistors) which is more radiation-resistant and has lower noise. Full sensor depletion allows fast charge collection by drift (beneficial to enhance the timing performance), higher charge collection efficiency, deeper collection depth and larger SNR; it also leads to improved radiation tolerance, as charge losses by trapping are reduced pernegger (2017). Since thicker sensors need higher backside bias voltage to reach full depletion, termination structures composed of multiple floating guard rings are used to avoid early breakdown at the edges of the pn-junction. An additional n-type epitaxial layer, with lower resistivity than the substrate, is integrated between the n-type substrate and the deep p-wells. Its aim is to better control the potential barrier below the deep p-well, in order to delay the onset of the punch through current described in details in Paragraph 2.2.2. The feasibility of this sensor concept and approach to Fully Depleted monolithic CMOS sensors was proven in the framework of the SEED project pancheri (2019, 2020). The upcoming ARCADIA engineering run will include different design flavours of FD-CMOS monolithic sensors, both pixelated and strip-like. Large-area ($1.3\times 1.3$ cm2) pixel demonstrators with embedded CMOS electronics and pixel test structures ($0.5\times 0.5$ and $1.5\times 1.5$ mm2) without integrated readout circuitry neubueser (2020) are foreseen, with pitches ranging from 10 to 50 $\mu$m. The test structures will include as well the innovative MAMS and will allow a detailed characterisation of these sensors. The 3D TCAD simulation campaign performed to design the first FD-MAMS will be presented and discussed in the following. ### 2.1 TCAD simulations 3D TCAD simulations were employed as a tool to optimise the sensor layout and performance. The use of 3D simulations is necessary to have a more realistic domain and results which are more accurate and less affected by boundary conditions. Furthermore, we were also interested in studying the charge collection dynamics after a particle crosses the sensor, and this is more straightforward with 3D simulations. A fine pitch of 10 $\mu$m was chosen for the microstrips in order to explore the characteristics and performance of a sensor layout which pushes the requirements on both spatial and timing resolution. Different sensor thicknesses foreseen for the production runs were simulated. Variations in the sensor layout and operating parameters were tested to study and optimise the sensor response. The simulated sensor flavours take into account the limitations imposed by the foundry’s sensors fabrication process, especially for the n-well and p-well sizes. The strip simulations investigated sensor flavours which pushed the design to the limits of the process requirements. Figure 2: Example TCAD 3D sensor domains for ARCADIA microstrips (top row) and corresponding cross sections (bottom row). (a) Standard simulation domain for sensors with the deep p-well. (b) Addition of n-wells above the deep p-wells. (c) Simulated ARCADIA microstrips without deep p-wells. All the TCAD simulations were performed at a temperature of 300 K. A standard simulation domain including three 50 $\mu$m long, 50 $\mu$m thick, 10 $\mu$m pitch microstrips is shown as an example in Figure 2 (a). The n-doped substrate is shown in green, the epitaxial layer in yellow, the microstrip sensing n-wells in red, the p-wells in blue and the less doped deep p-well in light blue. The default value for Vnwell is 0.8 V. The p-wells, instead, are kept at a voltage V${}_{pwell}=0$ V. One of the simulated sensor flavours has been specifically designed to allow for CMOS digital library cells to be integrated along the strips and is shown in Figure 2 (b). This sensor variant would allow the deployment of complex CMOS digital functions along the strip for distributed signal processing. We observed that the n-wells dedicated to the implementation of PMOS transistors and shielded by the deep p-well do not significantly influence the electrical characteristics of the detector in the TCAD simulation results. Therefore, we did not include them in the simulations. The deep p-well can be removed in the test structures that will be used to characterise the sensor (see Figure 2, c), and the necessary CMOS front-end electronics can be deployed at the end of the strips in the chip periphery. Sensors without the deep p-well were simulated as well. Different n-well, p-well and deep p-well sizes were considered to find the optimal layout in terms of sensor performance. Simulations were also employed to predict the effects that possible production uncertainties can have on the sensor operating parameters and electrical characteristics. For instance, the thickness and resistivity of the epitaxial layer may vary within a confidence range around their typical specified value (see Appendix A). 3D simulations for the different cases were run and compared. Some simulation parameters were fine-tuned using characterization results from a previous set of test structures, produced in the framework of the SEED project pancheri (2019). ### 2.2 Electrical and transient simulation In this Section, the simulations performed to extract the sensor electrical characteristics and to study the charge collection dynamics are briefly illustrated. Shared definitions and conventions on simulation setups and operating parameters were agreed for the whole ARCADIA simulation campaign and are also described in neubueser (2020). The strip length in the upcoming production run will be 1.2 cm. However, MAMS with lengths of 50 $\mu$m were simulated in order to run a large set of TCAD simulations in a reasonable computational time. The results were then scaled to the desired length. #### 2.2.1 Depletion voltage Sensor depletion starts at the backside, where the pn-junction between the n-type substrate and the p+ contact is located. If no negative bias voltage is applied to the backside contact, the sensor is not fully depleted and the collection n-wells are not isolated. This means that a resistive path exists between the n-type sensing nodes (see Figure 3, on the left). Therefore, if a voltage difference is applied between two adjacent n-wells, a current will flow between them. As the negative voltage applied to the backside contact increases, the space charge region enlarges through the high resistivity substrate, eventually merging with the depletion volume which surrounds the pn-junctions formed between the n-type substrate or epitaxial layer and the deep p-wells. At this point, the sensor is fully depleted, the resistive path between the sensing nodes is closed and the collection n-wells are isolated; this is shown in Figure 3, on the right. In this condition, no current (except for the leakage current) will flow among adjacent n-wells even when different voltages are applied to them. Figure 3: Depletion process in ARCADIA microstrips. On the left: cross section of a sensor before full depletion is reached. On the right: cross section of a fully depleted MAMS. The orange lines indicate the edge of the depletion region. This behaviour can be observed in the orange example IV curve in Figure 4 ($I_{nwell,unbalanced}$). The simulated domain shown in Figure 2 (left) was used. In this simulation, a voltage unbalance of 10 mV was applied between adjacent strips: the first n-well was biased at 0.79 V, the central one at 0.8 V and the third one at 0.81 V. The curve shows the current measured at the sensing node of the central strip as a function of $|V_{back}|$. A current of about 1 nA is measured at $V_{back}=0$ V. As the backside voltages increases and the space charge region enlarges, the current starts decreasing, and eventually reaches a plateau at a current of about $10^{-5}$ nA. This baseline corresponds to the leakage current (green IV curve, $I_{nwell,leakage}$). The backside voltage at which the single microstrips become isolated and the plateau is reached is the sensor depletion voltage $V_{dpl}$; this voltage is evaluated as the intersection point between the exponential decay fitting of the IV curve decreasing segment and the baseline. Figure 4: Example characteristic IV and CV curves extracted from TCAD simulations of ARCADIA monolithic sensors. The red vertical axis refer to the sensor capacitance ($C_{sens}$) CV curve. Figure 5 shows the simulated electrostatic potential and electric field maps at $V_{back}=V_{dpl}$ in a cross section of a 3-strip domain with all the n-wells at $V_{nwell}=0.8$ V. Electric field lines are plotted on top of both the electrostatic potential and the electric field maps. Figure 5: Electrostatic potential map (left) and electric field map (right) for a group of three ARCADIA microstrip sensors at $V_{back}=V_{dpl}$. The electric field lines are plotted on top of both maps. #### 2.2.2 Punch-through If $V_{back}$ exceeds a certain value, a hole current flowing between the shallow p-doped backside region and the (deep) p-well exponentially increases. This condition is known as punch-through and the hole current is the punch- through current chu (1972). We define the voltage corresponding to the onset of the punch-through as $V_{pt}$. The onset of the punch-through currents can be observed from the blue IV curve in Figure 4 ($I_{pwell}$), which shows the absolute value of the current measured at the top p-well contacts as a function of $|V_{back}|$. The dip in the curve, corresponding to the point of sign inversion of the current, was defined as $V_{pt}$. The simulation domain includes three 50 $\mu$m long, 50 $\mu$m thick, 10 $\mu$m pitch microstrips. In this case, the n-wells are all biased at $V_{nwell}=0.8$ V, which is the default value. Sensor operation in low punch-through regime can be tolerated, whereas a too large punch-through current ought to be avoided, as it determines a substantial increase in the power consumption of the whole detector. For this reason, we chose $V_{back}$ = $V_{pt}$ as a safe reference sensor operating voltage; this is the operating point for all the results shown in the following, if not stated differently. The sensor power density can be defined as $pd=\frac{V_{back}\cdot(I_{pwell}+I_{nwell})}{A}$, where $I_{nwell}$ and $I_{pwell}$ are the currents flowing at the sensing node and at the top p-well contacts respectively, and $A$ is the top surface area of the simulated microstrip domain. In order to quantify the maximum acceptable backside bias voltage that limits the absorbed power density, the value $V_{pd}$ at which $pd=0.1$ mW/cm2 was extracted from the simulated IV curves (see Figure 4). Figure 6 shows the hole current density at two different $|V_{back}|$ > $|V_{pt}|$ in the simulation domain used to extract the $I_{pwell}$ curve of Figure 4. On the left, a backside voltage exceeding $V_{pt}$ by 1 V was chosen, while on the right $V_{back}$ was set to $V_{pd}$. An increase in the hole current density of several orders of magnitude can be observed below the deep p-wells and in the substrate. Figure 6: Hole current density in a simulated sensor domain including three microstrips in punch-through condition at two different $V_{back}$. Care had to be taken to ensure that $|V_{dpl}|$ < $|V_{pt}|$ in the designed sensors. In this way, full depletion is reached before the onset of the punch- through. Moreover, the voltage operating range between $V_{dpl}$ and $V_{pt}$, defined as $\Delta V_{op}=|V_{pt}-V_{dpl}|$, should be large enough to ensure safe operation in full depletion before the onset of the punch-through even if deviations from the simulated design occur in the sensor fabrication process. #### 2.2.3 Leakage current The same sensor domain and n-well voltage configuration used for the extraction of $V_{pt}$ was also used to evaluate the sensor leakage current $I_{leak}$. The leakage current is defined as the current flowing at the collection nodes in full depletion and in absence of external stimuli, such as particles or radiation. The leakage current as a function of the backside bias voltage is shown in Figure 4 as a green curve ($I_{leak}$). In the example shown in Figure 4, a value of 10 fA was extracted for $I_{leak}$ at $V_{back}=V_{pt}$. #### 2.2.4 Sensor capacitance The sensor CV curve was simulated through AC simulations with a frequency of 10 kHz using the same sensor domain employed for $V_{pt}$ and $I_{leak}$ evaluation, with $V_{nwell}=0.8$ V. The major contribution to the sensor capacitance $C_{sens}$, which is the input capacitance seen by the DC-coupled front-end electronics, originates from the lateral capacitance between the collection n-well and the surrounding p-wells. It is thus important to minimize this contribution by a careful selection of the distance between the edge of the collection n-well and the p-wells; we call this distance "gap" (see Figure 1). An example CV curve is shown in red in Figure 4, with the capacitance per unit length considered. In the example of Figure 4, a value of about 0.33 fF/$\mu$m was obtained at $V_{back}=V_{pt}$. It has to be mentioned that in these sensors the depletion voltage does not necessarily correspond to the voltage of minimum capacitance. The reason for this is the presence of the epitaxial layer, which is located far from the backside pn-junction and has a lower resistivity than the substrate. Therefore, the depletion of the epitaxial layer begins after the depletion of the substrate and progresses more slowly with voltage. Full depletion of the whole sensor, including the epitaxial layer, and minimum capacitance are only reached at $|V_{back}|>|V_{dpl}|$. From this point, both capacitance and leakage current values will be intended at $V_{back}=V_{pt}$. A central focus of the layout optimisation was the minimisation of the sensor capacitance. In fact, low input capacitance to the DC-coupled CMOS readout electronics allows for low-noise readout, low analog power pernegger (2017) and, in particular, SNR maximisation. Large input capacitance worsens the noise levels and the speed of the front-end electronics wang (2017). #### 2.2.5 Surface radiation damage In the simulation campaign performed to study the properties of MAMS, a silicon dioxide (SiO2) layer was added on the top-side of the sensor. In addition to this, surface damage was modeled to evaluate the effects of Total Ionising Dose (TID) on the sensor electrical properties. The impact of surface radiation damage was modeled following the AIDA-2020-D7.4 report passeri (2019). The model introduces fixed positive oxide charges and band-gap acceptor/donor defect levels (trap states) at the Si-SiO2 interface. The concentrations of oxide charges and defect levels start from a fixed value before irradiation (i.e. with the only inclusion of the SiO2 surface layer, at $dose=0$) and increase with the dose provided to the sensors. The dependence of the oxide charge density $Q_{ox}$ [charges $\cdot$ cm-2], of the acceptor integrated interface trap state density $N_{int}^{acc}$ [cm-2] and of the donor integrated interface trap state density $N_{int}^{don}$ [cm-2] on the dose is shown in Figure 7. Pre-irradiation values, shown as dotted horizontal lines in Figure 7, are $Q_{ox}=6.5\cdot 10^{10}$ charges $\cdot$ cm-2, $N_{int}^{acc}=2.0\cdot 10^{9}$ cm-2 and $N_{int}^{don}=2.0\cdot 10^{9}$ cm-2. Figure 7: Dependence of the oxide charge density $Q_{ox}$, acceptor integrated interface trap state density $N_{int}^{acc}$ and donor integrated interface trap state density $N_{int}^{don}$ on the dose for the surface radiation damage model described in passeri (2019). Pre-irradiation values are shown as horizontal dotted lines. In the simulation campaign, the effects of the inclusion of the SiO2 layer and of the radiation damage on the leakage current, sensor capacitance, depletion voltage and punch-through voltage were investigated and will be discussed in Section 3. #### 2.2.6 Transient simulations TCAD transient simulations were run to study the sensor charge collection process in response to particles traversing the simulated microstrip domain. These simulations also let us identify the most relevant layout parameters to be optimised for improving the sensor performance in terms of fast and uniform charge collection irrespective of the particle incidence position. The transient simulations employ the Synopsys ® Sentaurus TCAD HeavyIon model, described in sentaurus (2018). The HeavyIon model gives an analytical description of the amount of charge generated within a 3D cylindrical distribution along the incident particle track. Two main parameters have to be passed to the HeavyIon model: the linear Energy Transfer (LET), defined as the average deposited charge per unit length, and the transverse size of the charge deposition volume generated around the particle trajectory. We chose the charge transverse distribution profile to be gaussian around the particle track. Figure 8: Best-case and worst-case scenarios considered in the TCAD transient simulations. The microstrips are represented as adjacent grey blocks and the particle traversing the domain is shown as an orange cylinder. The nomenclature used to identify the microstrips (from 1 to 5) is illustrated. Two extreme cases in terms of particle impact position were studied to evaluate the uniformity of charge collection time and charge collection efficiency. Particle trajectories perpendicular to the sensor surface were considered. In the best-case scenario, the particle impact point corresponds to the centre of a microstrip, which is the centre of a collection n-well. On the contrary, in the worst-case scenario, the particle traverses the sensor at the edge between two adjacent microstrips, i.e. in the middle of a p-well. In Figure 8 the two cases and the corresponding numbering of the strips are illustrated. This conventional strip nomenclature will be used in the following when referring to transient simulations. In order to save computational time, a reduced TCAD simulation domain that employs the symmetries was used. This reduced domain corresponds to a quarter of the full domain, with the particle incident in the corner of the domain instead of in the centre. An example for the best-case scenario is shown in Figure 9. The collected charge and current signals were then scaled to reproduce the full domain case, which includes nine or ten 100$\mu$m long microstrips in the best-case and worst-case scenario respectively (Figure 8). These numbers and size of strips guarantee that that the amount of deposited charge reaching the borders of the simulation domain is negligible. The correctness of this strategy was verified and confirmed by comparing the results of a simulation with a quarter domain and of a simulation with full domain. Figure 9: Example reduced TCAD domain used in transient simulations (best-case scenario). The microstrips are labelled following the nomenclature illustrated in Figure 8. A crossing particle is represented as an orange cylinder hitting the corner of the simulated reduced domain. An example of current signals $I_{nwell}(t)$ measured at the microstrip sensing nodes when a particle crosses the microstrip domain is shown in Figure 10 (left). We defined as charge collection efficiency for the i-th strip (CCEi) the integral of the current signal $I_{nwell,i}(t)$ extracted from the i-th strip and normalised at the total charge $Q_{tot}$ deposited in the sensor by the particle, according to the formula $CCE_{i}(t)=\frac{\int_{0}^{t}I_{nwell,i}(t\textquoteright)\,dt\textquoteright}{Q_{tot}}=\frac{\int_{0}^{t}I_{nwell,i}(t\textquoteright)\,dt\textquoteright}{LET\cdot d_{Si}}$ (1) where $d_{Si}$ is the sensor thickness. The total charge collection efficiency CCE for the whole simulated domain is defined as $CCE(t)=\sum_{i=1}^{N_{strips}}CCE_{i}(t)$ (2) where $N_{strips}$ is the total number of strips in the simulated domain. The total CCE at the end of the charge collection process (i.e. at $t=t_{end}=30$ ns, which was observed to be large enough for complete charge collection) has to be equal to 100% in the absence of recombination: $CCE(t=t_{max})=100\%$ (3) The CCEi as a function of time is shown in Figure 10, on the right, for strip number 1. The times needed for collecting the 95% and 99% of the total deposited charge were evaluated and referred to as $t_{95}$ and $t_{99}$, respectively. These values were compared for different design options and used to select the layouts of the fastest sensor flavours. The spatial mesh of the transient simulations was forced to be finer around the particle trajectory to more accurately simulate the charge deposition and the drift of electrons and holes from their generation points along the particle track towards the electrodes. Additionally, the time step of the transient simulations was fine tuned to guarantee the necessary accuracy while keeping the computational time requirement economical. We observed that these adjustments prevented the simulations from giving unphysical results. Figure 10: Simulated current signals (left) and corresponding charge collection efficiency CCEi (right) in the best-case and worst-case scenarios for strip number 1 in an example 50 $\mu$m thick microstrip domain. A particle track with an LET of $1.28\cdot 10^{-5}$ pC/$\mu$m was simulated. ### 2.3 Determination of the LET for heavy nuclei Since MAMS are an interesting candidate for tracking detectors in space applications, the charge collection was studied not only for minimum ionising particles (MIPs) but also for heavy nuclei of interest for in-orbit astroparticle experiments. The LET values of carbon and oxygen ions were studied in Geant4 (version 10.6 patch 01) simulations, and the typically used olive (2014) LET of 80 electron-hole (e-h) pairs per $\mu$m, or $1.28\cdot 10^{-5}$ pC per $\mu$m in silicon for MIPs could be reproduced. The Geant4 simulation setup included a 50 $\mu$m thick silicon layer immersed in air and with a transverse size of 1 $\times$ 1 cm2. The particle gun was positioned 15 cm in front of the centre of the silicon layer. The G4EmPenelopePhysics physics list was used to model the electromagnetic processes and the necessary precision on the energy deposited within the silicon was achieved with a maximum step size of 1 $\mu$m muonsilicon (2020). The LETs for carbon (C12+) and oxygen (O16+) ions at their minimum ionisation were computed from their most probable energy loss (i.e. the most probable value of the straggling or Landau functions tanabashi (2018); bichsel (2006)). Figure 11 shows the LET as a function of the particle energy obtained for C and O ions traversing 50 $\mu$m of silicon. The energies $E_{min}$ at which C and O ions are at the minimum of ionisation were found to be 35 GeV and 60 GeV, respectively. The corresponding LETs are $45.6\cdot 10^{-5}$ pC/$\mu$m and $83.0\cdot 10^{-5}$ pC/$\mu$m, which results in 36 and 65 times the MIP value. This is consistent with the expected scaling from the Bethe-Bloch formula. We were especially interested in studying the charge sharing among the microstrips surrounding the particle impact point and the charge collection time at different LETs. This will be reported and discussed in Section 3. Figure 11: Dependence of the LET on the energy of carbon ions (C12+, blue) and oxygen ions (O16+, orange) incident on 50 $\mu$m thick silicon. The LET values were evaluated through Geant4 simulations. The red vertical lines indicate the minimum ionisation energies for the two particle species. ## 3 Results and discussions In this section, the results of the TCAD simulation campaign will be presented. Their implications will be discussed and their connections to the design objectives will be highlighted. As mentioned in Section 1, the main targets of the FD-MAMS design were the following. 1. 1. To enhance the spatial resolution. A very fine pitch of 10 $\mu$m was chosen to reach this goal. Intrinsic spatial resolution in case of digital readout would be equal to $\frac{pitch}{\sqrt{12}}=\frac{10\,\mu m}{\sqrt{12}}\simeq 2.9$ $\mu$m, which can be further improved thanks to charge sharing and with an analog readout. 2. 2. To minimise the sensor capacitance $C_{sens}$ at $V_{back}=V_{pt}$. A low sensor capacitance is particularly important to keep low electronic noise and, consequently, to maximise the SNR. 3. 3. To obtain fast and uniform charge collection, irrespective of the particle incidence position. This will enhance the sensor timing capabilities and will reduce the dead-time between successive particle detections. For reasons of space available for MAMS in the first ARCADIA engineering run, only a few sensor flavours could be included. Hence, a simulation campaign was needed to identify the best performing sensor layouts. The deep p-well, when present, was kept the same size as the p-well. The expression "p-well and deep p-well" will be contracted and referred to as "(deep) p-well". In the legends of the figures, the abbreviation "dpw" will be used for deep p-well. ### 3.1 SiO2 layer and surface damage A first group of TCAD simulation studies was aimed at investigating the effects of the SiO2 layer and of surface TID damage on the FD-MAMS characteristics. The model that we employed was presented in Paragraph 2.2.5. As can be seen from Figure 12, for one of the selected 50 $\mu$m thick microstrip layouts, the inclusion of the SiO2 layer with a minimum concentration of traps and oxide charges ($dose=0$) determines a small increase of about 5% in the leakage current $I_{leak}$ from 20.8 fA to 22.0 fA. The sensor capacitance $C_{sens}$ is strongly affected by the inclusion of the SiO2 layer, as it increases by 31% from 0.26 fF/$\mu$m to 0.34 fF/$\mu$m. Both $I_{leak}$ and $C_{sens}$ are found to rise with increasing dose. The minimum dose that we considered is 50 krad, as the model is not validated for lower doses passeri (2019). Figure 13, instead, shows the effect of the SiO2 layer and of the TID on $V_{dpl}$ and on $V_{pt}$. The effect of the dose on these two values is smaller than in the case of $I_{leak}$ and $C_{sens}$. Furthermore, $V_{dpl}$ and $V_{pt}$ are influenced by the dose in opposite directions, which results in a slight increase in the operating range $\Delta$Vop with increasing dose. Figure 12: Leakage current $I_{leak}$ (green) and sensor capacitance $C_{sens}$ (red) as a function of the total ionising dose for a 50 $\mu$m thick microstrip sensor. The values obtained in simulations with and without the SiO2 layer in the absence of irradiation are shown as horizontal lines and referred to as "dose = 0" and "no SiO2 layer" respectively. Figure 13: Depletion voltage $V_{dpl}$ (blue) and punch-through voltage $V_{pt}$ (orange) as a function of the total ionising dose for a 50 $\mu$m thick microstrip sensor. The values obtained in simulations with and without the silicon dioxide layer in the absence of irradiation are shown as horizontal lines. #### 3.1.1 Effect on sensor capacitance The reason for the significant capacitance increase even after the simple inclusion of the SiO2 layer was found to be due to the introduction of positive oxide charges at the Si-SiO2 interface neubueser (2020). In fact, the model that we adopted foresees a significant positive oxide charge concentration $Q_{ox}$ = 6.5 $\cdot$ $10^{10}$ charges/cm-2 already at dose = 0. These positive oxide charges attract free electrons from the n-type silicon epitaxial layer towards the Si-SiO2 interface in the gap and determine an increase in the electron concentration around the heavily n-doped collection well, as illustrated in Figures 14 and 15. This electron accumulation behaves as an extension of the collection n-well. Figure 14: Schematic illustration of the electron accumulation in the gap between the collection n-well and the surrounding p-wells due to the positive oxide charges introduced at the Si-SiO2 interface. Figure 15: Electron density in an example microstrip simulation domain without (left) and with (right) the SiO2 layer on top of the sensors. ### 3.2 Capacitance minimisation The sizes of both the collection n-well and of the gap were found to contribute to $C_{sens}$. Therefore, both n-well and (deep) p-well sizes were adjusted to find the optimal layout for $C_{sens}$ minimisation. It was observed the inclusion of the SiO2 layer influences $C_{sens}$ in different ways for different gap sizes. Hence, $C_{sens}$ with and without the SiO2 layer was evaluated. Figure 16 shows the trend of $C_{sens}$ as a function of the gap size for 50 $\mu$m thick microstrips. The different sensor thicknesses considered (50, 100 and 300 $\mu$m) were found not to influence the sensor capacitance. Both the case with fixed minimum-size n-well and variable (deep) p-well (blue curves) and the case with fixed minimum-size (deep) p-well and variable n-well (orange curve) were studied. The dash-dotted lines refer to simulations without the surface SiO2 layer, whereas solid lines to the case with SiO2 layer included with minimal oxide charge and trap concentration. The reason for which smaller gaps with fixed n-wells could not be investigated is referred to as channel choking, a condition that inhibits sensor operation; this condition is explained in Section 3.3. The vertical grey band in Figure 16 and in the following ones corresponds to the forbidden region due to the constraints on n-well and (deep) p-well minimum sizes imposed by the fabrication process. The leftmost limit of the grey band is still permitted. Variations of n-well and of (deep) p-well size do not lead to the same $C_{sens}$ for the same gap size. A fixed n-well size with SiO2 layer included shows a trend that is not monotonic, but has a minimum at slightly less than 0.34 fF/$\mu$m. This effect is caused by the electron accumulation in the gap at the Si-SiO2 interface. However, the difference in $C_{sens}$ between the minimum-capacitance option and the sensor layout at the edge of the forbidden region is lower than 2%. There was, as expected, no benefit found from having large n-wells. The sensor capacitance increases with the n-well size, as can be seen from the blue curve in Figure 16. Therefore, we chose the best layout for minimum $C_{sens}$ to have the smallest possible n-well size and sufficiently small (deep) p-well. Figure 16: $C_{sens}$ as a function of the gap size for different sensor layout configuration. The vertical grey band is the forbidden region due to fabrication constraints; its leftmost limit is still permitted. Figure 17 compares the sensor capacitance for layouts with deep p-well (orange curve) and without deep p-well (green curve). All the sensor flavours feature the minimum n-well size permitted by the fabrication process. On the one hand, removing the deep p-well could help in further reducing the sensor capacitance. On the other hand, this choice would strongly affect the sensor bias voltage operating range, as discussed in the following section. Figure 17: Sensor capacitance $C_{sens}$ as a function of the gap size for different sensor layout configurations with and without the deep p-well. ### 3.3 Reference and operating voltages We found the influence of the n-well size on the operating voltages to be negligible compared to the effect of the (deep) p-well size. Therefore, for the sake of capacitance minimisation, we fixed the n-well size at the smallest possible value. With this assumption, Figure 18 presents the effect of the (deep) p-well size effect on $V_{dpl}$ and on $V_{pt}$ for 50 $\mu$m thick sensors. Both the cases with (orange curves) and without deep p-well (green curves) were considered and compared. The voltage values are reported for the case of dose = 0. In all the layouts considered in Figure 18, the onset of the punch through happens at voltages sufficiently larger than the depletion voltage. Outside of the forbidden region (grey band), the operating range $\Delta V_{op}$ is always between 4.2 V and 6.2 V, or between the 23% and the 41% of $V_{dpl}$. This is a sufficiently large operating range for safe sensor operation, even in the hypothesis of possible doping inhomogeneities among adjacent microstrips or slight deviations from the doping design values. Similar observations on $\Delta V_{op}$ have been made for 100 $\mu$m thick and 300 $\mu$m thick sensors. Figure 18: Sensor depletion voltage $V_{dpl}$ and punch-through voltage $V_{pt}$ as a function of the gap size for different sensor layout configurations. The orange region indicates the forbidden region due to the observed channel choking. As a general trend, it can be observed in Figure 18 that smaller (deep) p-wells result in larger $V_{dpl}$ and $V_{pt}$. This can be interpreted as follows. Large p-doped surfaces below the (deep) p-wells create wider pn- junctions with the n-doped epitaxial layer, thus facilitating the depletion of the underlying epitaxial layer at lower voltages. On the other hand, large (deep) p-wells also lower the potential barrier that prevents the direct flow of holes towards the substrate. This results in the earlier onset of the punch through hole current between the (deep) p-wells and the backside p+ region. Sensors without the deep p-well showed higher reference voltages. In fact, the presence of a deep p-well reduces the epitaxial layer thickness below the p-wells, thus requiring a lower voltage to achieve both full depletion and the onset of punch-through currents. Finally, for sensors with too large deep p-well, a phenomenon that we defined as channel choking was observed. This consists in the closure of the conductive channel below the collection n-well due to the lateral merging of the closely adjacent depletion regions formed at the junctions between the deep p-wells and the n-epitaxial layer. In this situation, in the simulations performed to extract $V_{dpl}$, no current flows among the n-wells at low values of $V_{back}$, even though the space charge region of the backside junction has not reached the surface yet. In this condition, the $I_{nwell,unbalanced}$ curve, that corresponds to the orange curve shown in Figure 4, appears flat and no $V_{dpl}$ can be extracted. This means that the n-wells are already isolated from one another at $V_{back}=0$ V and that the process of charge collection, which generates the current $I_{nwell}$ measured at the sensing node, is inhibited by the strong potential barrier present below the n-wells. No channel choking was observed for sensor layouts without the deep p-well. For completeness, Figure 19 (left) illustrates the dependence of $V_{pt}$ and $V_{dpl}$ on the sensor thickness for the sensor layout with minimum sizes for the n-well and for the (deep) p-well. The trend is linear over a wide range of thicknesses, both with and without the deep p-well. Also the operating voltage $\Delta V_{op}=V_{pt}-V_{dpl}$ linearly increases with the sensor thickness, as shown in Figure 19 (right). The sensor thickness investigated was extended down to 20 $\mu$m, well below the smallest thickness (i.e. 50 $\mu$m) of the sensors that will be produced in the first ARCADIA engineering run. The reason for this will become clear in Paragraph 3.5.1, as the study of very thin sensors was functional for enhancing the speed of the charge collection process and, consequently, for improving the sensor timing performance. Figure 19: Dependence of $V_{dpl}$ and $V_{pt}$ (left) and of the operating voltage range $\Delta V_{op}$ (right) on the sensor thickness. The voltage $V_{pd}$ at which the power density is 0.1 mW/cm2 was found to be about 4-5 V above $V_{pt}$ for 50 $\mu$m thick microstrips, 7-8 V for 100 $\mu$m thick microstrips and 18-20 V for 300 $\mu$m thick microstrips when the deep p-well was included. ### 3.4 Effects of Vnwell The $V_{nwell}$ voltage was varied with the aim of finding possible improvements in the sensor performances. The results are shown in Figure 20 (left), where the vertical red line indicates the default value of 0.8 V. A minimum $V_{nwell}$ of about 0.5 V is necessary to satisfy the condition $|V_{pt}|>|V_{dpl}|$. Moreover, an increase in $V_{nwell}$ has several interesting effects. First of all, it allows the sensor full depletion to be reached at lower (in absolute value) backside voltages. Secondly, it also shifts the onset of the punch through towards larger $|V_{back}|$, thus increasing the operating range $\Delta V_{op}$. Finally, as shown in Figure 20 (right), larger $V_{nwell}$ implies lower sensor capacitance. Figure 20: $V_{dpl}$ and $V_{pt}$ (left) and sensor capacitance (right) as a function of $V_{nwell}$. The vertical red line indicates the default value of $V_{nwell}=0.8$ V. ### 3.5 Charge collection studies As described in Paragraph 2.2.6, TCAD transient simulations were used to study the charge collection dynamics. In order to select the layouts with the optimal performance in terms of fast and uniform charge collection, the effect of the (deep) p-well size on the charge collection time at $V_{back}=V_{pt}$ was evaluated. The time $t_{95}$ needed to collect 95% of the total charge deposited in the simulated sensor domain is plotted in Figure 21 for 50 $\mu$m thick sensors and LET = $1.28\cdot 10^{-5}$ pC/$\mu$m (1 MIP) as a function of the gap size and with fixed minimum n-well size. Figure 21: $t_{95}$ as a function of the gap size for best-case and worst-case scenarios. Microstrips with large gaps, hence small (deep) p-wells, are to be preferred for fast charge collection. The reason for this is a higher $|V_{pt}|$, which enables sensor operation at a larger $|V_{back}|$. The consequent stronger electric field in the sensor results in higher charge velocity in the silicon substrate. For the same reason, microstrip sensors without deep p-well revealed a significantly faster charge collection in both the best-case and the worst-case scenario. Flavours with small (deep) p-wells also show very uniform charge collection for different particle incidence positions. The difference in $t_{95}$ for the best-case and worst-case scenarios is below 0.1 ns for the fastest permitted options. This result is also achieved thanks to the fine microstrip pitch of 10 $\mu$m. The channel choking, as described in Section 3.3, limits the deep p-well size as the potential barrier below the sensing node slows down the electron collection. This problem, as shown in Figure 21, can be avoided by removing the deep p-well. Figure 22 demonstrates that the proposed MAMS guarantee fast sensor response also under heavily ionising particles. The charge collection time is only weakly proportional to the charge deposited by the incident particle within an LET range of [1.28; 128] $\cdot$ 10-5 pC/$\mu$m. A 50 $\mu$m thick sensor was considered in Figure 22, and the LET values corresponding to 1 MIP, carbon (C) ion and oxygen (O) ion at their minimum of ionisation are highlighted as vertical green lines. Moreover, $t_{99}$ is added to show that the time needed for complete charge collection is only slightly larger than $t_{95}$, due to a small fraction of charge collected by the strips adjacent to the central one. However, $t_{95}$ and $t_{99}$ were never found to exceed 2 ns and 3 ns respectively in 50 $\mu$m thick sensors. Figure 22: $t_{95}$ (blue) and $t_{99}$ (red) as a function of the LET for best-case and worst-case scenarios. #### 3.5.1 Further enhancements for fast timing performance As we discussed in Section 3.5, the first strategy for improving the timing performance of the proposed microstrip sensors is to remove the deep p-well in order to obtain larger $|V_{pt}|$. However, we also investigated other ways to increase $|V_{pt}|$ and to speed up the charge collection. In particular, as shown in Section 3.4, a larger $V_{nwell}$ is capable of shifting the onset of the punch-through current towards larger $|V_{back}|$. Therefore, we explored the effects of $V_{nwell}$ on the charge collection time. In a strip readout system, timing information can be retrieved only from the strips collecting most of the charge (i.e. strip(s) number 1, following the nomenclature of Figure 8), as they provide a signal with sufficiently large SNR. Therefore, in order to study the sensor timing performance and after verifying through $t_{95}$ that the total deposited charge is quickly collected in the whole simulation domain, we considered the time $t_{95}^{central}$ needed to collect 95% of the charge in the central strip(s). Figure 23 shows the dependence of $t_{95}^{central}$ at $V_{back}=V_{pt}$ and with LET = $1.28\cdot 10^{-5}$ pC/$\mu$m on the voltage applied to the sensing node. A 50 $\mu$m thick sensor with a layout optimised for fast charge collection was considered. A significant improvement could be reached at larger $V_{nwell}$. For the option without deep p-well and at $V_{nwell}=3$ V, $t_{95}^{central}$ is 0.84 ns in the best-case and 0.94 ns in the worst-case scenario. If we assume an electron drift saturation velocity of $\sim 1\cdot 10^{7}$ cm/s in silicon at a temperature of 300 K canali (1975), the minimum drift time for electrons that have to cover a 50 $\mu$m distance is 0.5 ns. This explains the saturation observed in Figure 23 and demonstrates the fast charge collection and the promising timing capabilities of the proposed MAMS. Figure 23: $t_{95}^{central}$ as a function of the voltage $V_{nwell}$ applied to the sensing node for best-case and worst-case scenarios. The vertical red line indicates the default value of $V_{nwell}=0.8$ V. A way to further reduce the collection time is to explore thinner sensors. Figure 24 demonstrates that the charge collection time $t_{95}^{central}$ is proportional to the sensor thickness. For these simulations, $V_{nwell}$ was set to the 0.8 V and a 1 MIP LET was considered. Even at thicknesses as large as 300 $\mu$m, $t_{95}^{central}$ does not exceed 6 ns. In the best-case scenario without the deep p-well, reducing the sensor thickness from 50 $\mu$m to 40 $\mu$m, 30 $\mu$m and 20 $\mu$m results in a decrease in $t_{95}^{central}$ of 15%, 33% and 50%, respectively. Analogous proportionality was observed for $t_{95}$. Therefore, for future production runs, thinner sensors could be considered for the enhancement of the timing performance. Figure 24: $t_{95}^{central}$ as a function of the sensor thickness for best- case and worst-case scenarios. ### 3.6 Charge sharing A set of TCAD simulations was dedicated to study the charge sharing among adjacent microstrips when particles with different LETs traverse the sensor. Charge sharing is relevant for improving the spatial resolution, especially with analog readout, and is enhanced by fine microstrip pitches and large sensor thicknesses. On the contrary, it is reduced at higher $V_{back}$ for a fixed sensor thickness. In Figure 25, the case of a 300 $\mu$m thick sensor at $V_{back}=V_{pt}$ is presented for the best-case scenario. The total charge collected by each strip (identified using the nomenclature of Figure 8) is plotted versus the LET. The black horizontal line indicates a possible charge threshold corresponding to 10% of a MIP at the single strip level. A comparison with the sensors that will be produced in the first ARCADIA engineering run will allow deeper investigation on the charge sharing, a fine tuning of the simulations and studies aimed at evaluating the spatial resolution of 10 $\mu$m pitch MAMS. Figure 25: Charge sharing among adjacent microstrips. The total charge collected by strips 1 to 5 (following the nomenclature illustrated in Figure 8) is shown as a function of the LET. ## 4 Conclusions In this work, we presented detailed TCAD simulations of CMOS-based FD-MAMS, which may find use for tracking and timing in particle and nuclear physics, space and medical applications. The results of the TCAD simulation campaign, performed to design the 10 $\mu$m pitch FD-MAMS, demonstrate their very fast and uniform charge collection, which encourages their practicality for various applications, even under heavily ionizing particles. The effect of surface ionizing radiation damage was investigated, and the layout parameters were optimized to achieve a minimum capacitance, beneficial for electronic noise reduction. The possibility to operate the sensor in full depletion and at low- power density (i.e. before the onset of the punch through current) was verified in the simulations. A preference for small collection diodes and small (deep) p-wells emerged for obtaining lower capacitance and faster sensor response. Additionally, these simulations confirmed the possibility of monolithically integrating readout architectures in the inter-strip regions for strips of 10 $\mu$m pitch. The first FD-MAMS samples will be produced in the upcoming ARCADIA engineering production run at the beginning of 2021 and will allow the simulation results to be compared with experimental data from electrical characterisation, laser and beam irradiation tests. The promising results of the first simulation campaign on FD-MAMS will translate into further R&D activities to enhance the sensor performance in terms of low capacitance and high timing and spatial resolution. yes ## Appendix A Expected effects from epitaxial layer thicknesses Possible variations in the epitaxial layer thickness of [-15%; +30%] communicated by the foundry with respect to the reference value induced us to investigate their effect on the operating parameters. While the sensor capacitance was observed not to be influenced, both $V_{dpl}$ and $V_{pt}$ showed a linear dependence on the epitaxial layer thickness. This behaviour is presented in Figure 26. Figure 26: $V_{dpl}$ and $V_{pt}$ as a function of the epitaxial layer thickness, expressed as percentage variation with respect to the reference thickness. ###### Acknowledgements. The research activity presented in this article has been carried out in the framework of the ARCADIA experiment funded by the Istituto Nazionale di Fisica Nucleare (INFN), CSN5. The activity has also been supported by the project "Dipartimento di Eccellenza", Physics Department of the University of Torino (Dipartimento di Fisica - Università degli Studi di Torino), Italy, funded by MUR. Data curation, Lorenzo de Cilladi; Formal analysis, Lorenzo de Cilladi; Investigation, Lorenzo de Cilladi; Supervision, Coralie Neubüser and Lucio Pancheri; Writing – original draft, Lorenzo de Cilladi; Writing – review & editing, Thomas Corradino, Gian-Franco Dalla Betta, Coralie Neubüser and Lucio Pancheri. The authors declare no conflict of interest. References ## References * garcia (2018) Garcia-Sciveres, M., and Wermes, N., A review of advances in pixel detectors for experiments with high rate and radiation. Rep. Prog. Phys. 2018, 81(6), pp. 066101, DOI: https://doi.org/10.1088/1361-6633/aab064 * wermes (2009) Wermes, N. et al., Pixel detectors for charged particles. Nucl. Instrum. Methods Phys. Res. Sect. A 2009, 604(1-2), pp. 370–379, DOI: https://doi.org/10.1016/j.nima.2009.01.098 * wermes (2019) Wermes, N., Pixel detectors… where do we stand?. Nucl. Instrum. Methods Phys. Res. Sect. A 2019, 924, pp. 44–50, DOI: https://doi.org/10.1016/j.nima.2018.07.003 * mager (2016) Mager, M., and the ALICE collaboration, ALPIDE, the Monolithic Active Pixel Sensor for the ALICE ITS upgrade. Nucl. Instrum. Methods Phys. Res. Sect. A 2016, 824, pp. 434–438, DOI: https://doi.org/10.1016/j.nima.2015.09.057 * pernegger (2017) Pernegger, H., et al., First tests of a novel radiation hard CMOS sensor process for Depleted Monolithic Active Pixel Sensors. J. Instrum. 2017, 12(06), pp. P06008, DOI: https://doi.org/10.1088/1748-0221/12/06/P06008 * wang (2017) Wang, T., et al., Development of a depleted monolithic CMOS sensor in a 150 nm CMOS technology for the ATLAS inner tracker upgrade. J. Instrum. 2017, 12(01), pp. C01039, DOI: https://doi.org/10.1088/1748-0221/12/01/C01039 * wunderer (2014) Wunderer, C.B., et al., The PERCIVAL soft X-ray imager. J. Instrum. 2014, 9(03), pp. C03056, DOI: https://doi.org/10.1088/1748-0221/9/03/C03056 * hatsui (2013) Hatsui, T., et al., A direct-detection X-ray CMOS image sensor with 500 $\mu$m thick high resistivity silicon. In Proc. Int. Image Sensor Workshop, June 12–16 2013, 3, p. 4, URL: http://www.imagesensors.org/Past%20Workshops/2013%20Workshop/2013%20Papers/03-5_058_hatsui_paper.pdf * mattiazzo (2018) Mattiazzo, S., et al., iMPACT: An innovative tracker and calorimeter for proton computed tomography. IEEE Tran. Rad. Plasma Med. Sci. 2017, 2(4), pp. 345–352, DOI: https://doi.org/10.1109/TRPMS.2018.2825499 * scotti (2019) Scotti, V., and Osteria, G., for the CSES-LIMADOU collaboration, The High Energy Particle Detector onboard CSES-02 satellite. In Proc. 36th Int. Cosmic Ray Conf. (ICRC), Jul 24 – Aug 1 2019, DOI: https://pos.sissa.it/358/135/pdf * rummel (2009) Rummel, S., et al., Intrinsic properties of DEPFET active pixel sensors. J. Instrum. 2009, 4(03), pp. P03003, DOI: https://doi.org/10.1088/1748-0221/4/03/P03003 * marinas (2011) Mariñas, C., and Vos, M., The Belle-II DEPFET pixel detector: A step forward in vertexing in the superKEKB flavour factory. Nucl. Instrum. Methods Phys. Res. Sect. A 2011, 650(1), pp. 59–63, DOI: https://doi.org/10.1016/j.nima.2010.12.116 * treis (2010) Treis, J., et al., MIXS on BepiColombo and its DEPFET based focal plane instrumentation. Nucl. Instrum. Methods Phys. Res. Sect. A 2010, 624(2), pp. 540–547, DOI: https://doi.org/10.1016/j.nima.2010.03.173 * lutz (2010) Lutz, G., et al., DEPFET sensor with intrinsic signal compression developed for use at the XFEL free electron laser radiation source. Nucl. Instrum. Methods Phys. Res. Sect. A 2010, 624(2), pp. 528–532, DOI: https://doi.org/10.1016/j.nima.2010.03.002 * lutz (2016) Lutz, G., et al., The DEPFET sensor-amplifier structure: A method to beat 1/f noise and reach sub-electron noise in pixel detectors. Sensors 2016, 16(5), p. 608, DOI: https://doi.org/10.3390/s16050608 * kucewicz (2005) Kucewicz, W., et al., Development of monolithic active pixel detector in SOI technology. Nucl. Instrum. Methods Phys. Res. Sect. A 2005, 541(1–2), pp. 172–177, DOI: https://doi.org/10.1016/j.nima.2005.01.054 * lan (2015) Lan, H., et al., SoI Monolithic Active Pixel Sensors for Radiation Detection Applications: A Review. IEEE Sensors Journal 2015, 15(5), pp. 2732–2746, DOI: https://doi.org/10.1109/JSEN.2015.2389271 * hagino (2020) Hagino, K., et al., Radiation damage effects on double-SOI pixel sensors for X-ray astronomy. Nucl. Instrum. Methods Phys. Res. Sect. A 2020, 978, p. 164435, DOI: https://doi.org/10.1016/j.nima.2020.164435 * asano (2016) Asano, M., et al., Characteristics of non-irradiated and irradiated double SOI integration type pixel sensor. Nucl. Instrum. Methods Phys. Res. Sect. A 2016, 831, pp. 315–321, DOI: https://doi.org/10.1016/j.nima.2016.03.095 * turchetta (2003) Turchetta, R., et al., Monolithic active pixel sensors (MAPS) in a VLSI CMOS technology. Nucl. Instrum. Methods Phys. Res. Sect. A 2003, 501(1), pp. 251–259, DOI: https://doi.org/10.1016/S0168-9002(02)02043-0 * turchetta (2001) Turchetta, R., et al., A monolithic active pixel sensor for charged particle tracking and imaging using standard VLSI CMOS technology. Nucl. Instrum. Methods Phys. Res. Sect. A 2001, 458(3), pp. 677–689, DOI: https://doi.org/10.1016/S0168-9002(00)00893-7 * contin (2018) Contin, G., et al., The STAR MAPS-based PiXeL detector. Nucl. Instrum. Methods Phys. Res. Sect. A 2018, 907, pp. 60–80, DOI: https://doi.org/10.1016/j.nima.2018.03.003 * rossi (2006) Rossi, L., Fischer, P., Rohe, T., and Wermes, N., Pixel detectors: From fundamentals to applications. Springer Science & Business Media 2006, ISBN: 978-3-540-28333-1 * peisert (1992) Peisert, A., Silicon microstrip detectors. In: Instrumentation in High Energy Physics, World Scientific 1992, pp. 1–79, DOI: https://doi.org/10.1142/9789814360333_0001 * chowdhury (2020) Chowdhury, S. R., and the CMS collaboration, The Phase-2 Upgrade of the CMS Outer Tracker. Nucl. Instrum. Methods Phys. Res. Sect. A 2020, 979, pp. 164432, DOI: https://doi.org/10.1016/j.nima.2020.164432 * david (2018) David, C., A new strips tracker for the upgraded ATLAS ITk detector. J. Instrum. 2018, 13(01), pp. C01003, DOI: https://doi.org/10.1088/1748-0221/13/01/C01003 * atwood (2007) Atwood, W. B., et al., Design and initial tests of the Tracker-converter of the Gamma-ray Large Area Space Telescope. Astropart. Phys. 2007, 28(4-5), pp. 422–434, DOI: https://doi.org/10.1016/j.astropartphys.2007.08.010 * azzarello (2016) Azzarello, P., et al., The DAMPE silicon–tungsten tracker. Nucl. Instrum. Methods Phys. Res. Sect. A 2016, 831, pp. 378–384, DOI: https://doi.org/10.1016/j.nima.2016.02.077 * adriani (2003) Adriani, O., et al., The magnetic spectrometer of the PAMELA satellite experiment. Nucl. Instrum. Methods Phys. Res. Sect. A 2003, 511(1-2), pp. 72–75, DOI: https://doi.org/10.1016/S0168-9002(03)01754-6 * lubelsmeyer (2011) Lübelsmeyer, K., et al., Upgrade of the Alpha Magnetic Spectrometer (AMS-02) for long term operation on the International Space Station (ISS). Nucl. Instrum. Methods Phys. Res. Sect. A 2011, 654(1), pp. 639–648, DOI: https://doi.org/10.1016/j.nima.2011.06.051 * benhammadi (2019) Benhammadi, S., et al., DECAL: A Reconfigurable Monolithic Active Pixel Sensor for use in Calorimetry and Tracking. In Topical Workshop on Electronics for Particle Physics, Sep. 2019, 2, p. 6, URL: https://s3.cern.ch/inspire-prod-files-4/4f56e107d64ecc5b165344c1695f60c1 * han (2020) Han, Y., et al., Study of CMOS strip sensor for future silicon tracker. Nucl. Instrum. Methods Phys. Res. Sect. A 2020, 981, pp. 164520, DOI: https://doi.org/10.1016/j.nima.2020.164520 * straver (1994) Straver, J., et al., One micron spatial resolution with silicon strip detectors. Nucl. Instrum. Methods Phys. Res. Sect. A 1994, 348(2-3), pp. 485–490, DOI: https://doi.org/10.1016/0168-9002(94)90785-4 * sekigawa (2017) Sekigawa, D., et al., Fine-pixel detector FPIX realizing sub-micron spatial resolution developed based on FD-SOI technology. In Int. Conf. Tech. Instrum. Part. Phys. 2017, Springer Singapore, pp. 331–338, URL: https://link.springer.com/chapter/10.1007/978-981-13-1316-5_62 * turchetta (1993) Turchetta, R., Spatial resolution of silicon microstrip detectors. Nucl. Instrum. Methods Phys. Res. Sect. A 1993, 335(1-2), pp. 44–58, DOI: https://doi.org/10.1016/0168-9002(93)90255-G * snoeys (2017) Snoeys, W., et al., A process modification for CMOS monolithic active pixel sensors for enhanced depletion, timing performance and radiation tolerance. Nucl. Instrum. Methods Phys. Res. Sect. A 2017, 871, pp. 90–96, DOI: https://doi.org/10.1016/j.nima.2017.07.046 * pancheri (2019) Pancheri, L., et al., A 110 nm CMOS process for fully-depleted pixel sensors. J. Instrum. 2019, 14(06), pp. C06016, DOI: https://doi.org/10.1088/1748-0221/14/06/C06016 * pancheri (2020) Pancheri, L., et al., Fully Depleted MAPS in 110-nm CMOS Process With 100–-300 $\mu$m Active Substrate. IEEE Trans. Electron Devices 2020, 67(6), pp. 2393–2399, DOI: https://doi.org/10.1109/TED.2020.2985639 * neubueser (2020) Neubüser, C., et al., Sensor design optimization of innovative low-power, large area MAPS for HEP and applied science. arXiv preprint arXiv:2011.09723 2020, URL: https://arxiv.org/abs/2011.09723 * chu (1972) Chu, J. L., Persky, G., and Sze, S. M., Thermionic injection and space-charge-limited current in reach-through p+np+ structures. J. Appl. Phys. 1972, 43(8), pp. 3510–3515, DOI: https://doi.org/10.1063/1.1661745 * passeri (2019) Passeri, D., et al., TCAD radiation damage model. AIDA-2020-D7.4 deliverable report 2019, URL: http://cds.cern.ch/search?p=AIDA-2020-D7.4 * sentaurus (2018) Sentaurus Device User Guide, version O-2018.06. Synopsys Inc., June 2018, pp. 681–683. * olive (2014) Olive, K. A., et al., (Particle Data Group), Particle Detectors at Accelerators. Chin. Phys. C 2014, 38, 090001 * muonsilicon (2020) Centis Vignali, M., https://github.com/mcentis/muonOnSilicon * tanabashi (2018) Tanabashi, M., et al., (Particle Data Group), Passage of Particles Through Matter. Phys. Rev. D 2018 and 2019 update, 98, 030001 * bichsel (2006) Bichsel, H., A method to improve tracking and particle identification in TPCs and silicon detectors. Nucl. Instrum. Methods Phys. Res. Sect. A 2006, 562(1), pp. 154–197, DOI: https://doi.org/10.1016/j.nima.2006.03.009 * canali (1975) Canali, C., and the CMS collaboration, Electron and hole drift velocity measurements in silicon and their empirical relation to electric field and temperature. IEEE Trans. Electron Devices 1975, 22(11), pp. 1045–1047, DOI: https://doi.org/10.1109/T-ED.1975.18267
# Recurrent Sums and Partition Identities Roudy El Haddad Université La Sagesse, Faculté de génie, Polytech ###### Abstract Sums of the form $\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(m);N_{m}}\cdots a_{(1);N_{1}}}}$ where the $a_{(k);N_{k}}$’s are same or distinct sequences appear quite often in mathematics. We will refer to them as recurrent sums. In this paper, we introduce a variety of formulas to help manipulate and work with this type of sums. We begin by developing variation formulas that allow the variation of a recurrent sum of order $m$ to be expressed in terms of lower order recurrent sums. We then proceed to derive theorems (which we will call inversion formulas) which show how to interchange the order of summation in a multitude of ways. Later, we introduce a set of new partition identities in order to then prove a reduction theorem which permits the expression of a recurrent sum in terms of a combination of non-recurrent sums. Finally, we apply this reduction theorem to a recurrent form of two famous types of sums: The $p$-series and the sum of powers. ###### Keywords. Recurrent sums, Partitions, Stirling numbers of the first kind, Bell polynomials, Multiple harmonic series, Riemann zeta function, Bernoulli numbers, Faulhaber formula. MSC 2020: primary 11P84 secondary 11B73, 11M32, 05A18 ## 1 Introduction and Notation The harmonic series was first studied and proven to diverge in the 14th century by Nicole Oresme [32]. Later, in the 17th century, new proofs for this divergence were provided by Pietro Mengoli [29], Johann Bernoulli [5], and Jacob Bernoulli [3, 4]. However, a more general form of this series does converge. Euler was the first to study such sums of the form $\zeta(s)=\sum_{n=1}^{\infty}{\frac{1}{n^{s}}}$ where $s$ is a real number. In the famous Basel problem, Euler proved that $\zeta(2)=\frac{\pi^{2}}{6}$ (see [12, 13, 15]. Fourteen additional proofs can be found in [11]). He later provided a general formula for this zeta function for positive even values of $s$. Euler’s definition was then extended to a complex variable $s$ by Riemann in his 1859 article “On the Number of Primes Less Than a Given Magnitude”. More recently, the multiple harmonic series, an even more general form of the zeta function, has been introduced and studied. Note that Euler was the first to study these multiple harmonic series for length 2 in [14]. A multiple harmonic series (MHS) or multiple zeta values (MZV) is defined as: $\zeta(s_{1},s_{2},\ldots,s_{k})=\sum_{1\leq N_{1}<N_{2}<\cdots<N_{k}}{\frac{1}{N_{1}^{s_{1}}N_{2}^{s_{2}}\cdots N_{k}^{s_{k}}}}.$ A very important variant of the MHS (see [27, 25, 30]) often referred to as multiple zeta star values MZSV or multiple harmonic star series MHSS (or simply multiple zeta values) is defined by: $\zeta^{\star}(s_{1},s_{2},\ldots,s_{k})=\sum_{1\leq N_{1}\leq N_{2}\leq\cdots\leq N_{k}}{\frac{1}{N_{1}^{s_{1}}N_{2}^{s_{2}}\cdots N_{k}^{s_{k}}}}.$ This variant of the multiple harmonic series is directly related to the Riemann zeta function $\zeta(s)$ [23, 18]. Additionally, it is involved in a variety of sums and series including the Arakawa–Kaneko zeta function [37] and Euler sums. Such sums have tremendous importance in number theory. They have been of interest to mathematicians for a long time and have been systematically studied since the 1990s with the work of Hoffman [23, 24] and Zagier [38]. However, their importance is not limited to Number Theory. In fact, such sums/series have appeared in physics even before the phrase “multiple zeta values” had been coined. As an example, the number $\zeta(\overline{6},\overline{2})$ appeared in the quantum field theory literature in 1986 [8]. They play a major role in the connection of knot theory with quantum field theory [9, 26]. MZVs and MZSVs became even more important after they became needed for higher order calculations in quantum electrodynamics (QED) and quantum chromodynamics (QCD) [7, 6]. These sums are a particular case of what we called recurrent sums as they are of the form $\sum_{1\leq N_{1}\leq\cdots\leq N_{m}\leq n}{a_{(m);N_{m}}\cdots a_{(1);N_{1}}}$ with $a_{(i);N_{i}}=\frac{1}{N_{i}^{s_{i}}}$ for all $i$. The particular case has been extensively studied while the general case received much less interest. Although there are hundreds if not thousands of formulae to help in the study of multiple harmonic star sums and multiple zeta star values, barely any formulae can be found for its general counterpart. In this article, we are interested in studying this more general form which is expressed as follows: $\sum_{1\leq N_{1}\leq\cdots\leq N_{m}\leq n}{a_{(m);N_{m}}\cdots a_{(1);N_{1}}}.$ We will also consider the particular case where all sequences are the same: $\sum_{1\leq N_{1}\leq\cdots\leq N_{m}\leq n}{a_{N_{m}}\cdots a_{N_{1}}}.$ This structure of sums appears in a variety of areas of mathematics. The objective is to develop formulae to improve and facilitate the way we work with recurrent sums. This includes deriving formulae to calculate the variation of such sums, formulae to interchange the order of summation as well as formulae to represent recurrent sums in terms of a combination of non- recurrent sums. Note that this type of sums is intimately related to partitions as they appear in the representation of recurrent sums as a combination of simple non-recurrent sums. Therefore, this article will also focus on partition identities that are needed to prove the previously stated theorems as well as the ones that can be derived from these same theorems. Among these partition identities that can be found through these theorems, a definition of binomial coefficients in terms of a sum over partitions will be presented. Similarly, we produce some identities involving special sums, over partitions, of Bernoulli numbers. Furthermore, we are also interested in applying the formulae develop for the general case to some particular cases. First, we will apply our results to the multiple sums of powers in order to generalize Faulhaber’s formula. Then, we will go back to the most famous particular case which is the MZSV and show how our results on the general case can improve in this case. A particularly beautiful identity that we will present is the following which relates the recurrent sum of $\frac{1}{N^{2}}$ to the zeta function for positive even values: $\sum_{N_{m}=1}^{\infty}{\cdots\sum_{N_{1}=1}^{N_{2}}{\frac{1}{N_{m}^{2}\cdots N_{1}^{2}}}}=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\left(\zeta(2i)\right)^{y_{k,i}}}}=\left(2-\frac{1}{2^{2(m-1)}}\right)\zeta(2m).$ Although this paper focuses on the generalized version of the multiple zeta star values, the multiple zeta values itself is a particular form of a type of sums presented in [20] and which is closely related to the recurrent sums by the relations also presented in that cited article. The main theorems of this paper have potential applications such as the following: Surprisingly, this form appears in the general formula for the $n$-th integral of $x^{m}(\ln x)^{m^{\prime}}$. In the unpublished paper [19], the relations presented in this paper will be used to derive and prove this general formula for the $n$-th integral of $x^{m}(\ln x)^{m^{\prime}}$. In paper [20], the partition identities here presented are combined with additional partition identities in order to produce identities for odd and even partitions. Let us now introduce some notation in order to facilitate the representation of such sums in this paper: For any $m,q,n\in\mathbb{N}$ where $n\geq q$ and for any set of sequences $a_{(1);N_{1}},\cdots,a_{(m);N_{m}}$ defined in the interval $[q,n]$, let $R_{m,q,n}(a_{(1);N_{1}},\cdots,a_{(m);N_{m}})$ represent the general recurrent sum of order $m$ for the sequences $a_{(1);N_{1}},\cdots,a_{(m);N_{m}}$ with lower an upper bounds respectively $q$ and $n$. For simplicity, however, we will denote it simply as $R_{m,q,n}$. $\begin{split}R_{m,q,n}&=\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}\\\ &=\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{2}=q}^{N_{3}}{\sum_{N_{1}=q}^{N_{2}}{a_{(m);N_{m}}\cdots a_{(2);N_{2}}a_{(1);N_{1}}}}}\\\ &=\sum_{q\leq N_{1}\leq\cdots\leq N_{m}\leq n}{a_{(m);N_{m}}\cdots a_{(2);N_{2}}a_{(1);N_{1}}}.\\\ \end{split}$ (1) The most common case of a recurrent sum is that where all sequences are the same, $\begin{split}R_{m,q,n}(a_{N_{1}}\cdots a_{N_{m}})&=\sum_{N_{m}=q}^{n}{a_{N_{m}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}}\\\ &=\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{2}=q}^{N_{3}}{\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{2}}a_{N_{1}}}}}\\\ &=\sum_{q\leq N_{1}\leq\cdots\leq N_{m}\leq n}{a_{N_{m}}\cdots a_{N_{2}}a_{N_{1}}}.\\\ \end{split}$ (2) For simplicity, we will denote it as $\hat{R}_{m,q,n}$. This type of sums is described as recurrent because they can also be expressed using the following recurrent form: $\begin{cases}R_{m,q,n}=\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}R_{m-1,q,N_{m}}}\\\ R_{0,i,j}\,\,\,=1\,\,\forall i,j\in\mathbb{N}.\end{cases}$ (3) ###### Remark. A recurrent sum of order $0$ is always equal to $1$. It is not equivalent to an empty sum $($which is equal to $0)$. In this paper, this type of sums will be studied. In Section 2, formulas for the calculation of variation of these sums in terms of lower order recurrent sums will be presented. Then, in Section 3, inversion formulas will be presented, which will allow the interchange of the order of summation in such sums. Finally, in Section 4, we will present a reduction formula that allows the representation of a recurrent sum as a combination of simple (non- recurrent) sums. These relations will be, then, used to calculate certain special sums such as the recurrent harmonic sum and the recurrent equivalent to the Faulhaber formula. ## 2 Variation Formulas In this section, we will develop formulas to express the variation of a recurrent sum of order $m$ ($R_{m,q,n+1}-R_{m,q,n}$) in terms of lower order recurrent sums. Equivalently, these formulas can be used to express $R_{m,q,n+1}$ in terms of $R_{m,q,n}$ and lower order recurrent sums. ### 2.1 Simple expression We start by proving the most basic form for the variation formula as illustrated by the following Lemma. This is needed in order to prove the general form of this formula. ###### Lemma 2.1. For any $m,q,n\in\mathbb{N}$, we have that $R_{m+1,q,n+1}=a_{(m+1);n+1}R_{m,q,n+1}+R_{m+1,q,n}.$ ###### Proof. $\begin{split}R_{m+1,q,n+1}&=\sum_{N_{m+1}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(m+1);N_{m+1}}\cdots a_{(1);N_{1}}}}\\\ &=a_{(m+1);n+1}\sum_{N_{m}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(m);N_{m}}\cdots a_{(1);N_{1}}}}+\sum_{N_{m+1}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(m+1);N_{m+1}}\cdots a_{(1);N_{1}}}}\\\ &=a_{(m+1);n+1}R_{m,q,n+1}+R_{m+1,q,n}.\end{split}$ ∎ Now we apply the basic case from Lemma 2.1 to show the general variation formula that allows $R_{m,q,n+1}$ to be expressed in terms of $R_{m,q,n}$ and of recurrent sums of order going from $0$ to $(m-1)$. ###### Theorem 2.1. For any $m,q,n\in\mathbb{N}$ where $n\geq q$ and for any set of sequences $a_{(1);N_{1}},\cdots,a_{(m);N_{m}}$ defined in the interval $[q,n+1]$, we have that $\sum_{N_{m}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(m);N_{m}}\cdots a_{(1);N_{1}}}}=\sum_{k=0}^{m}{\left(\prod_{j=0}^{m-k-1}{a_{(m-j);n+1}}\right)\left(\sum_{N_{k}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(k);N_{k}}\cdots a_{(1);N_{1}}}}\right)}.$ Using the notation from Eq. (1), this theorem can be written as $R_{m,q,n+1}=\sum_{k=0}^{m}{\left(\prod_{j=0}^{m-k-1}{a_{(m-j);n+1}}\right)R_{k,q,n}}.$ ###### Proof. 1\. Base Case: verify true for $m=1$. $\begin{split}\sum_{k=0}^{1}{\left(\prod_{j=0}^{-k}{a_{(1-j);n+1}}\right)R_{k,q,n}}&=\left(\prod_{j=0}^{0}{a_{(1-j);n+1}}\right)R_{0,q,n}+\left(\prod_{j=0}^{-1}{a_{(1-j);n+1}}\right)R_{1,q,n}\\\ &=(a_{(1);n+1})(1)+(1)\left(\sum_{N_{1}=q}^{n}{a_{(1);N_{1}}}\right)\\\ &=R_{1,q,n+1}.\end{split}$ 2\. Induction hypothesis: assume the statement is true until $m$. $R_{m,q,n+1}=\sum_{k=0}^{m}{\left(\prod_{j=0}^{m-k-1}{a_{(m-j);n+1}}\right)R_{k,q,n}}.$ 3\. Induction step: we will show that this statement is true for ($m+1$). We have to show the following statement to be true: $R_{m+1,q,n+1}=\sum_{k=0}^{m+1}{\left(\prod_{j=0}^{m-k}{a_{(m+1-j);n+1}}\right)R_{k,q,n}}.$ From Lemma 2.1, $R_{m+1,q,n+1}=a_{(m+1);n+1}R_{m,q,n+1}+R_{m+1,q,n}.$ By applying the induction hypothesis, $\begin{split}R_{m+1,q,n+1}&=a_{(m+1);n+1}\sum_{k=0}^{m}{\left(\prod_{j=0}^{m-k-1}{a_{(m-j);n+1}}\right)R_{k,q,n}}+R_{m+1,q,n}\\\ &=a_{(m+1);n+1}\sum_{k=0}^{m}{\left(\prod_{j=1}^{m-k}{a_{(m+1-j);n+1}}\right)R_{k,q,n}}+R_{m+1,q,n}\\\ &=\sum_{k=0}^{m}{\left(\prod_{j=0}^{m-k}{a_{(m+1-j);n+1}}\right)R_{k,q,n}}+R_{m+1,q,n}.\end{split}$ Noticing that $\sum_{k=m+1}^{m+1}{\left(\prod_{j=0}^{m-k}{a_{(m+1-j);n+1}}\right)R_{k,q,n}}=R_{m+1,q,n}$ hence, $R_{m+1,q,n+1}=\sum_{k=0}^{m+1}{\left(\prod_{j=0}^{m-k}{a_{(m+1-j);n+1}}\right)R_{k,q,n}}.$ Hence, the theorem is proven by induction. ∎ ###### Corollary 2.1. If all sequences are the same, Theorem 2.1 will be reduced to the following form, $\sum_{N_{m}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}=\sum_{k=0}^{m}{\left(a_{n+1}\right)^{m-k}\left(\sum_{N_{k}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{k}}\cdots a_{N_{1}}}}\right)}.$ Using the notation from Eq. (2), this theorem can be written as $\hat{R}_{m,q,n+1}=\sum_{k=0}^{m}{\left(a_{n+1}\right)^{m-k}\hat{R}_{k,q,n}}.$ ###### Example 2.1. Consider that $m=2$, we have the two following cases: * • If all sequences are distinct, $\sum_{N_{2}=q}^{n+1}{b_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}-\sum_{N_{2}=q}^{n}{b_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}=(b_{n+1})\sum_{N_{1}=q}^{n}{a_{N_{1}}}+(b_{n+1})(a_{n+1}).$ * • If all sequences are the same, $\sum_{N_{2}=q}^{n+1}{a_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}-\sum_{N_{2}=q}^{n}{a_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}=(a_{n+1})\sum_{N_{1}=q}^{n}{a_{N_{1}}}+(a_{n+1})^{2}.$ ###### Remark. Set $a_{(m);N}=\cdots=a_{(2);N}=1$, Theorem 2.1 becomes $\sum_{N_{m}=q}^{n+1}{\sum_{N_{m-1}=q}^{N_{m}}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}}=\sum_{k=1}^{m}{\left(\sum_{N_{k}=q}^{n}{\sum_{N_{k-1}=q}^{N_{k}}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}}\right)}+a_{n+1}.$ ### 2.2 Simple recurrent expression A recursive form of Theorem 2.1 can be obtained by expanding and factoring the theorem’s expression. ###### Theorem 2.2. For any $m,q,n\in\mathbb{N}$ where $n\geq q$ and for any set of sequences $a_{(1);N_{1}},\cdots,a_{(m);N_{m}}$ defined in the interval $[q,n+1]$, we have that $\sum_{N_{m}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(m);N_{m}}\cdots a_{(1);N_{1}}}}-\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(m);N_{m}}\cdots a_{(1);N_{1}}}}=a_{(m);n+1}\left\\{a_{(m-1);n+1}\left[\cdots a_{(2);n+1}\left(a_{(1);n+1}(1)+\sum_{N_{1}=q}^{n}{a_{(1);N_{1}}}\right)+\sum_{N_{2}=q}^{n}{\sum_{N_{1}=q}^{N_{2}}{a_{(2);N_{2}}a_{(1);N_{1}}}}\right]+\sum_{N_{m-1}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(m-1);N_{m-1}}\cdots a_{(1);N_{1}}}}\right\\}.$ Using the notation from Eq. (1), this theorem can be written as $R_{m,q,n+1}=a_{(m);n+1}\left\\{a_{(m-1);n+1}\left[\cdots a_{(2);n+1}\left(a_{(1);n+1}\left(R_{0,q,n}\right)+R_{1,q,n}\right)+R_{2,q,n}\right]+R_{m-1,q,n}\right\\}+R_{m,q,n}$ where $R_{0,q,n}=1$. ###### Proof. 1\. Base Case: verify true for $m=1$. $a_{(1);n+1}(R_{0,q,n})+R_{1,q,n}=a_{(1);n+1}(1)+\sum_{N_{1}=q}^{n}{a_{(1);N_{1}}}=\sum_{N_{1}=q}^{n+1}{a_{(1);N_{1}}}=R_{1,q,n+1}.$ 2\. Induction Hypothesis: assume the statement is true until $m$. $R_{m,q,n+1}=a_{(m);n+1}\left\\{a_{(m-1);n+1}\left[\cdots a_{(2);n+1}\left(a_{(1);n+1}\left(R_{0,q,n}\right)+R_{1,q,n}\right)+R_{2,q,n}\right]+R_{m-1,q,n}\right\\}+R_{m,q,n}.$ 3\. Induction Step: we will show that this statement is true for $(m+1)$. We have to show the following statement to be true: $R_{m+1,q,n+1}=a_{(m+1);n+1}\left\\{a_{(m);n+1}\left[\cdots a_{(2);n+1}\left(a_{(1);n+1}\left(R_{0,q,n}\right)+R_{1,q,n}\right)+R_{2,q,n}\right]+R_{m,q,n}\right\\}+R_{m+1,q,n}.$ From Lemma 2.1, $R_{m+1,q,n+1}=a_{(m+1);n+1}R_{m,q,n+1}+R_{m+1,q,n}.$ By applying the induction hypothesis, $R_{m+1,q,n+1}=a_{(m+1);n+1}\left\\{a_{(m);n+1}\left[\cdots a_{(2);n+1}\left(a_{(1);n+1}\left(R_{0,q,n}\right)+R_{1,q,n}\right)+R_{2,q,n}\right]+R_{m,q,n}\right\\}+R_{m+1,q,n}.$ Hence, the theorem is proven by induction. ∎ ###### Corollary 2.2. If all sequences are the same, Theorem 2.2 will be reduced to the following form, $\sum_{N_{m}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}-\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}=a_{n+1}\left\\{a_{n+1}\left[\cdots a_{n+1}\left(a_{n+1}(1)+\sum_{N_{1}=q}^{n}{a_{N_{1}}}\right)+\sum_{N_{2}=q}^{n}{\sum_{N_{1}=q}^{N_{2}}{a_{N_{2}}a_{N_{1}}}}\right]+\sum_{N_{m-1}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m-1}}\cdots a_{N_{1}}}}\right\\}.$ Using the notation from Eq. (2), this theorem can be written as $\hat{R}_{m,q,n+1}=a_{n+1}\left\\{a_{n+1}\left[\cdots a_{n+1}\left(a_{n+1}\left(\hat{R}_{0,q,n}\right)+\hat{R}_{1,q,n}\right)+\hat{R}_{2,q,n}\right]+\hat{R}_{m-1,q,n}\right\\}+\hat{R}_{m,q,n}$ where $\hat{R}_{0,q,n}=1$. ###### Example 2.2. Consider that $m=2$, we have the two following cases: * • If all sequences are distinct, $\sum_{N_{2}=q}^{n+1}{b_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}-\sum_{N_{2}=q}^{n}{b_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}=(b_{n+1})\left\\{a_{n+1}(1)+\sum_{N_{1}=q}^{n}{a_{N_{1}}}\right\\}.$ * • If all sequences are the same, $\sum_{N_{2}=q}^{n+1}{a_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}-\sum_{N_{2}=q}^{n}{a_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}=(a_{n+1})\left\\{a_{n+1}(1)+\sum_{N_{1}=q}^{n}{a_{N_{1}}}\right\\}.$ ### 2.3 General expression The variation of a recurrent sum can also be expressed in terms of only a certain range of lower order recurrent sums. In other words, $R_{m,q,n+1}$ can be expressed in terms of $R_{m,q,n}$ and of recurrent sums of order going only from $p$ to $(m-1)$. To do so, we develop the following theorem. ###### Theorem 2.3. For any $m,q,n\in\mathbb{N}$ where $n\geq q$, for any $p\in[0,m]$, and for any set of sequences $a_{(1);N_{1}},\cdots,a_{(m);N_{m}}$ defined in the interval $[q,n+1]$, we have that $\sum_{N_{m}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(m);N_{m}}\cdots a_{(1);N_{1}}}}=\sum_{k=p+1}^{m}{\left(\prod_{j=0}^{m-k-1}{a_{(m-j);n+1}}\right)\left(\sum_{N_{k}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(k);N_{k}}\cdots a_{(1);N_{1}}}}\right)}+\left(\prod_{j=0}^{m-p-1}{a_{(m-j);n+1}}\right)\left(\sum_{N_{p}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(p);N_{p}}\cdots a_{(1);N_{1}}}}\right).$ Using the notation from Eq. (1), this theorem can be written as $R_{m,q,n+1}=\sum_{k=p+1}^{m}{\left(\prod_{j=0}^{m-k-1}{a_{(m-j);n+1}}\right)R_{k,q,n}}+\left(\prod_{j=0}^{m-p-1}{a_{(m-j);n+1}}\right)R_{p,q,n+1}.$ ###### Proof. By applying Theorem 2.1, $\begin{split}R_{m,q,n+1}&=\sum_{k=0}^{m}{\left(\prod_{j=0}^{m-k-1}{a_{(m-j);n+1}}\right)R_{k,q,n}}\\\ &=\sum_{k=p+1}^{m}{\left(\prod_{j=0}^{m-k-1}{a_{(m-j);n+1}}\right)R_{k,q,n}}+\sum_{k=0}^{p}{\left(\prod_{j=0}^{m-k-1}{a_{(m-j);n+1}}\right)R_{k,q,n}}\\\ &=\sum_{k=p+1}^{m}{\left(\prod_{j=0}^{m-k-1}{a_{(m-j);n+1}}\right)R_{k,q,n}}+\left(\prod_{j=0}^{m-p-1}{a_{(m-j);n+1}}\right)\sum_{k=0}^{p}{\left(\prod_{j=m-p}^{m-k-1}{a_{(m-j);n+1}}\right)R_{k,q,n}}.\end{split}$ From Theorem 2.1, with $m$ substituted by $p$, we have $R_{p,q,n+1}=\sum_{k=0}^{p}{\left(\prod_{j=0}^{p-k-1}{a_{(p-j);n+1}}\right)R_{k,q,n}}=\sum_{k=0}^{p}{\left(\prod_{j=m-p}^{m-k-1}{a_{(m-j);n+1}}\right)R_{k,q,n}}.$ Hence, by substituting, we get $R_{m,q,n+1}=\sum_{k=p+1}^{m}{\left(\prod_{j=0}^{m-k-1}{a_{(m-j);n+1}}\right)R_{k,q,n}}+\left(\prod_{j=0}^{m-p-1}{a_{(m-j);n+1}}\right)R_{p,q,n+1}.$ ∎ ###### Corollary 2.3. If all sequences are the same, Theorem 2.3 will be reduced to the following form, $\sum_{N_{m}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}=\sum_{k=p+1}^{m}{\left(a_{n+1}\right)^{m-k}\left(\sum_{N_{k}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{k}}\cdots a_{N_{1}}}}\right)}+\left(a_{n+1}\right)^{m-p}\left(\sum_{N_{p}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{p}}\cdots a_{N_{1}}}}\right).$ Using the notation from Eq. (2), this theorem can be written as $\hat{R}_{m,q,n+1}=\sum_{k=p+1}^{m}{\left(a_{n+1}\right)^{m-k}\hat{R}_{k,q,n}}+\left(a_{n+1}\right)^{m-p}\hat{R}_{p,q,n+1}.$ ###### Example 2.3. For $p=2$ and if the sequences are the same: $\sum_{N_{m}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}=\sum_{k=3}^{m}{\left(a_{n+1}\right)^{m-k}\left(\sum_{N_{k}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{k}}\cdots a_{N_{1}}}}\right)}+\left(a_{n+1}\right)^{m-2}\left(\sum_{N_{2}=q}^{n+1}{\sum_{N_{1}=q}^{N_{2}}{a_{N_{2}}a_{N_{1}}}}\right).$ ###### Example 2.4. For $p=m-2$ and if the sequences are the same: $\sum_{N_{m}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}=\left(\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}\right)+\left(a_{n+1}\right)\left(\sum_{N_{m-1}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m-1}}\cdots a_{N_{1}}}}\right)+\left(a_{n+1}\right)^{2}\left(\sum_{N_{m-2}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m-2}}\cdots a_{N_{1}}}}\right).$ ###### Remark. Set $a_{(m);N}=\cdots=a_{(2);N}=1$, Theorem 2.3 becomes $\sum_{N_{m}=q}^{n+1}{\sum_{N_{m-1}=q}^{N_{m}}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}}=\sum_{k=p+1}^{m}{\left(\sum_{N_{k}=q}^{n}{\sum_{N_{k-1}=q}^{N_{k}}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}}\right)}+\sum_{N_{p}=q}^{n+1}{\sum_{N_{p-1}=q}^{N_{p}}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}}.$ ### 2.4 General recurrent expression Similarly, the theorem introduced in the previous section can be reformulated in a recursive form by expanding and factoring the expression of Theorem 2.3 to obtain the following expression. ###### Theorem 2.4. For any $m,q,n\in\mathbb{N}$ where $n\geq q$, for any $p\in[0,m]$, and for any set of sequences $a_{(1);N_{1}},\cdots,a_{(m);N_{m}}$ defined in the interval $[q,n+1]$, we have that $R_{m,q,n+1}=a_{(m);n+1}\left\\{a_{(m-1);n+1}\left[\cdots a_{(p+2);n+1}\left(a_{(p+1);n+1}\left(R_{p,q,n+1}\right)+R_{p+1,q,n}\right)+R_{p+2,q,n}\right]+R_{m-1,q,n}\right\\}+R_{m,q,n}.$ ###### Proof. From Theorem 2.2, with $m$ substituted by $p$, we have $R_{p,q,n+1}=a_{(p);n+1}\left\\{a_{(p-1);n+1}\left[\cdots a_{(2);n+1}\left(a_{(1);n+1}\left(R_{0,q,n}\right)+R_{1,q,n}\right)+R_{2,q,n}\right]+R_{p-1,q,n}\right\\}+R_{p,q,n}$ where $R_{0,q,n}=1$. Substituting into the expression of Theorem 2.2, the inner part becomes $R_{p,q,n+1}$ and we get the desired formula. ∎ ###### Corollary 2.4. If all sequences are the same, Theorem 2.4 will be reduced to the following form, $\hat{R}_{m,q,n+1}=a_{n+1}\left\\{a_{n+1}\left[\cdots a_{n+1}\left(a_{n+1}\left(\hat{R}_{p,q,n+1}\right)+\hat{R}_{p+1,q,n}\right)+\hat{R}_{p+2,q,n}\right]+\hat{R}_{m-1,q,n}\right\\}+\hat{R}_{m,q,n}.$ ###### Example 2.5. For $p=1$ and if the sequences are the same: $\sum_{N_{m}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}-\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}=a_{n+1}\left\\{a_{n+1}\left[\cdots a_{n+1}\left(\sum_{N_{1}=q}^{n+1}{a_{N_{1}}}\right)+\sum_{N_{2}=q}^{n}{\sum_{N_{1}=q}^{N_{2}}{a_{N_{2}}a_{N_{1}}}}\right]+\sum_{N_{m-1}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m-1}}\cdots a_{N_{1}}}}\right\\}.$ ###### Example 2.6. For $p=m-2$ and if the sequences are the same: $\sum_{N_{m}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}-\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}=a_{n+1}\left\\{a_{n+1}\left[\sum_{N_{m-2}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m-2}}\cdots a_{N_{1}}}}\right]+\sum_{N_{m-1}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m-1}}\cdots a_{N_{1}}}}\right\\}.$ ## 3 Inversion Formulas In this section, we will develop formulas to interchange the order of summation in a recurrent sum. ### 3.1 Particular case (for 2 sequences) We start by proving the inversion formula with $2$ sequences which is required in order to prove the more general inversion formula with $m$ sequences. ###### Theorem 3.1. For $m,q,n\in\mathbb{N}$ where $n\geq q$ and for any 2 sequences $a_{N_{1}}$ and $b_{N_{2}}$ defined in the interval $[q,n]$, we have that $\sum_{N_{2}=q}^{n}{b_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}=\sum_{N_{1}=q}^{n}{a_{N_{1}}\sum_{N_{2}=N_{1}}^{n}{b_{N_{2}}}}.$ ###### Proof. By expanding the sum, we get $\sum_{N_{2}=q}^{n}{b_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}=b_{q}\left(\sum_{N_{1}=q}^{q}{a_{N_{1}}}\right)+b_{q+1}\left(\sum_{N_{1}=q}^{q+1}{a_{N_{1}}}\right)+\cdots+b_{n-1}\left(\sum_{N_{1}=q}^{n-1}{a_{N_{1}}}\right)+b_{n}\left(\sum_{N_{1}=q}^{n}{a_{N_{1}}}\right)=b_{q}\left(a_{q}\right)+b_{q+1}\left(a_{q}+a_{q+1}\right)+\cdots+b_{n-1}\left(a_{q}+\cdots+a_{n-1}\right)+b_{n}\left(a_{q}+\cdots+a_{n}\right).$ By regrouping the $b_{N}$ terms instead of the $a_{N}$ terms, the expression becomes $\sum_{N_{2}=q}^{n}{b_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}=a_{q}\left(b_{q}+\cdots+b_{n}\right)+a_{q+1}\left(b_{q+1}+\cdots+b_{n}\right)+\cdots+a_{n-1}\left(b_{n-1}+b_{n}\right)+a_{n}\left(b_{n}\right)=a_{q}\left(\sum_{N_{2}=q}^{n}{b_{N_{2}}}\right)+a_{q+1}\left(\sum_{N_{2}=q+1}^{n}{b_{N_{2}}}\right)+\cdots+a_{n-1}\left(\sum_{N_{2}=n-1}^{n}{b_{N_{2}}}\right)+a_{n}\left(\sum_{N_{2}=n}^{n}{b_{N_{2}}}\right)=\sum_{N_{1}=q}^{n}{a_{N_{1}}\sum_{N_{2}=N_{1}}^{n}{b_{N_{2}}}}.$ ∎ ###### Corollary 3.1. If all sequences are the same, Theorem 3.1 becomes $\sum_{N_{2}=q}^{n}{a_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}=\sum_{N_{1}=q}^{n}{a_{N_{1}}\sum_{N_{2}=N_{1}}^{n}{a_{N_{2}}}}.$ ### 3.2 General case (for m sequences) We now prove the more general inversion formula with $m$ sequences which allows us to invert the order of summation for a recurrent sum of order $m$. ###### Theorem 3.2. For any $m,q,n\in\mathbb{N}$ where $n\geq q$ and for any set of sequences $a_{(1);N_{1}},\cdots,a_{(m);N_{m}}$ defined in the interval $[q,n]$, we have that $\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{1}=q}^{n}{a_{(1);N_{1}}\sum_{N_{2}=N_{1}}^{n}{a_{(2);N_{2}}\cdots\sum_{N_{m}=N_{m-1}}^{n}{a_{(m);N_{m}}}}}.$ ###### Proof. 1\. Base Case: verify true for $m=2$. This statement is true as proven in Theorem 3.1. 2\. Induction hypothesis: assume the statement is true until $m$. $\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{1}=q}^{n}{a_{(1);N_{1}}\sum_{N_{2}=N_{1}}^{n}{a_{(2);N_{2}}\cdots\sum_{N_{m}=N_{m-1}}^{n}{a_{(m);N_{m}}}}}.$ 3\. Induction step: we will show that this statement is true for $(m+1)$. We have to show the following statement to be true: $\sum_{N_{m+1}=q}^{n}{a_{(m+1);N_{m+1}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{1}=q}^{n}{a_{(1);N_{1}}\sum_{N_{2}=N_{1}}^{n}{a_{(2);N_{2}}\cdots\sum_{N_{m+1}=N_{m}}^{n}{a_{(m+1);N_{m+1}}}}}.$ $\sum_{N_{m+1}=q}^{n}{a_{(m+1);N_{m+1}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{m+1}=q}^{n}{a_{(m+1);N_{m+1}}\left(\sum_{N_{m}=q}^{N_{m+1}}{a_{(m);N_{m}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}\right)}.$ Let $b_{N_{m}}$ be the following sequence (that dependents only on $N_{m}$), $b_{N_{m}}=a_{(m);N_{m}}\sum_{N_{m-1}=q}^{N_{m}}{a_{(m-1);N_{m-1}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}.$ By applying this substitution in the previous expression, we obtain a recurrent sum of order 2 that contains the 2 sequences $a_{(m+1);N_{m+1}}$ and $b_{N_{m}}$. Then, we apply the inversion formula for the case of 2 sequences (Theorem 3.1) to get the following, $\sum_{N_{m+1}=q}^{n}{a_{(m+1);N_{m+1}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{m+1}=q}^{n}{a_{(m+1);N_{m+1}}\left(\sum_{N_{m}=q}^{N_{m+1}}{b_{N_{m}}}\right)}=\sum_{N_{m}=q}^{n}{b_{N_{m}}\left(\sum_{N_{m+1}=N_{m}}^{n}{a_{(m+1);N_{m+1}}}\right)}=\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}\left(\sum_{N_{m+1}=N_{m}}^{n}{a_{(m+1);N_{m+1}}}\right)}}}.$ The sum of $a_{(m+1);N_{m+1}}$ has $N_{m}$ and $n$ as lower and upper bounds. Thus, knowing that $n$ is a constant, the sum of $a_{(m+1);N_{m+1}}$ depends only on $N_{m}$. This allows us to extract this sum from the inner sums to get $\sum_{N_{m+1}=q}^{n}{a_{(m+1);N_{m+1}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{m}=q}^{n}{\left(a_{(m);N_{m}}\sum_{N_{m+1}=N_{m}}^{n}{a_{(m+1);N_{m+1}}}\right)\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}.$ Let $A_{N_{m}}$ be the following sequence (that only depends on $N_{m}$), $A_{N_{m}}=a_{(m);N_{m}}\sum_{N_{m+1}=N_{m}}^{n}{a_{(m+1);N_{m+1}}}.$ By substituting $A_{N_{m}}$ into the previous expression, we get a recurrent sum of order $m$ in terms of the following $m$ sequences: $A_{N_{m}},a_{(m-1);N_{m-1}},\cdots,a_{(1);N_{1}}$. Then the inversion formula for the case of $m$ sequences (which was assumed to be true in the induction hypothesis) is applied, $\sum_{N_{m+1}=q}^{n}{a_{(m+1);N_{m+1}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{m}=q}^{n}{A_{N_{m}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{1}=q}^{n}{a_{(1);N_{1}}\sum_{N_{2}=N_{1}}^{n}{a_{(2);N_{2}}\cdots\sum_{N_{m}=N_{m-1}}^{n}{A_{N_{m}}}}}=\sum_{N_{1}=q}^{n}{a_{(1);N_{1}}\sum_{N_{2}=N_{1}}^{n}{a_{(2);N_{2}}\cdots\sum_{N_{m}=N_{m-1}}^{n}{a_{(m);N_{m}}\sum_{N_{m+1}=N_{m}}^{n}{a_{(m+1);N_{m+1}}}}}}.$ We conclude that it must hold for all $m\geq 2$. ∎ ###### Corollary 3.2. If all sequences are the same, Theorem 3.2 becomes $\sum_{N_{m}=q}^{n}{a_{N_{m}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}}=\sum_{N_{1}=q}^{n}{a_{N_{1}}\sum_{N_{2}=N_{1}}^{n}{a_{N_{2}}\cdots\sum_{N_{m}=N_{m-1}}^{n}{a_{N_{m}}}}}.$ Similarly, the innermost summation can be turned into the outermost summation as illustrated by Theorem 3.3. ###### Theorem 3.3. For any $m,q,n\in\mathbb{N}$ where $n\geq q$ and for any set of sequences $a_{(1);N_{1}},\cdots,a_{(m);N_{m}}$ defined in the interval $[q,n]$, we have that $\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{1}=q}^{n}{a_{(1);N_{1}}\sum_{N_{m}=N_{1}}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{3}=N_{1}}^{N_{4}}{a_{(3);N_{3}}\sum_{N_{2}=N_{1}}^{N_{3}}{a_{(2);N_{2}}}}}}.$ ###### Proof. From Theorem 3.2, $\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{1}=q}^{n}{a_{(1);N_{1}}\sum_{N_{2}=N_{1}}^{n}{a_{(2);N_{2}}\cdots\sum_{N_{m}=N_{m-1}}^{n}{a_{(m);N_{m}}}}}.$ Applying Theorem 3.2 to the inner part of the right side sum would transform it as follows $\sum_{N_{2}=N_{1}}^{n}{a_{(2);N_{2}}\cdots\sum_{N_{m}=N_{m-1}}^{n}{a_{(m);N_{m}}}}=\sum_{N_{m}=N_{1}}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{3}=N_{1}}^{N_{4}}{a_{(3);N_{3}}\sum_{N_{2}=N_{1}}^{N_{3}}{a_{(2);N_{2}}}}}.$ Hence, substituting back into Theorem 3.2 would give us the desired formula. ∎ ###### Corollary 3.3. If all sequences are the same, Theorem 3.3 becomes $\sum_{N_{m}=q}^{n}{a_{N_{m}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}}=\sum_{N_{1}=q}^{n}{a_{N_{1}}\sum_{N_{m}=N_{1}}^{n}{a_{N_{m}}\cdots\sum_{N_{3}=N_{1}}^{N_{4}}{a_{N_{3}}\sum_{N_{2}=N_{1}}^{N_{3}}{a_{N_{2}}}}}}.$ ### 3.3 Inversion of p sequences from m sequences Finally, as we will show in this section, it is possible to partially invert the order of summation for a recurrent sum. In other words, as shown by the following theorem, it is possible to invert the order of summation of only the $p$ innermost summations from $m$ summations. ###### Theorem 3.4. For any $m,q,n\in\mathbb{N}$ where $n\geq q$, for any $p\in[0,m]$, and for any set of sequences $a_{(1);N_{1}},\cdots,a_{(m);N_{m}}$ defined in the interval $[q,n]$, we have that $\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{p}=q}^{N_{p+1}}{a_{(p);N_{p}}\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{p+1}=q}^{N_{p+2}}{a_{(p+1);N_{p+1}}\sum_{N_{1}=q}^{N_{p+1}}{a_{(1);N_{1}}\sum_{N_{2}=N_{1}}^{N_{p+1}}{a_{(2);N_{2}}\cdots\sum_{N_{p}=N_{p-1}}^{N_{p+1}}{a_{(p);N_{p}}}}}}}.$ ###### Proof. By replacing $m$ by $p$ and $n$ by $N_{p+1}$ in Theorem 3.2, we get the following relation, $\sum_{N_{p}=q}^{N_{p+1}}{a_{(p);N_{p}}\cdots\sum_{N_{2}=q}^{N_{3}}{a_{(2);N_{2}}\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{1}=q}^{N_{p+1}}{a_{(1);N_{1}}\sum_{N_{2}=N_{1}}^{N_{p+1}}{a_{(2);N_{2}}\cdots\sum_{N_{p}=N_{p-1}}^{N_{p+1}}{a_{(p);N_{p}}}}}.$ Thus, $\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{p}=q}^{N_{p+1}}{a_{(p);N_{p}}\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{p+1}}^{N_{p+2}}{a_{(p+1);N_{p+1}}\left(\sum_{N_{p}=q}^{N_{p+1}}{a_{(p);N_{p}}\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}\right)}}=\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{p+1}=q}^{N_{p+2}}{a_{(p+1);N_{p+1}}\left(\sum_{N_{1}=q}^{N_{p+1}}{a_{(1);N_{1}}\sum_{N_{2}=N_{1}}^{N_{p+1}}{a_{(2);N_{2}}\cdots\sum_{N_{p}=N_{p-1}}^{N_{p+1}}{a_{(p);N_{p}}}}}\right)}}.$ ∎ ###### Corollary 3.4. If all sequences are the same, Theorem 3.4 becomes $\sum_{N_{m}=q}^{n}{a_{N_{m}}\cdots\sum_{N_{p}=q}^{N_{p+1}}{a_{N_{p}}\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}}=\sum_{N_{m}=q}^{n}{a_{N_{m}}\cdots\sum_{N_{p+1}=q}^{N_{p+2}}{a_{N_{p+1}}\sum_{N_{1}=q}^{N_{p+1}}{a_{N_{1}}\sum_{N_{2}=N_{1}}^{N_{p+1}}{a_{N_{2}}\cdots\sum_{N_{p}=N_{p-1}}^{N_{p+1}}{a_{N_{p}}}}}}}.$ Similarly, the innermost summation can be pulled back to the $p$-th position as illustrated by Theorem 3.5. ###### Theorem 3.5. For any $m,q,n\in\mathbb{N}$ where $n\geq q$, for any $p\in[0,m]$, and for any set of sequences $a_{(1);N_{1}},\cdots,a_{(m);N_{m}}$ defined in the interval $[q,n]$, we have that $\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{p}=q}^{N_{p+1}}{a_{(p);N_{p}}\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(1);N_{1}}}}}=\sum_{N_{m}=q}^{n}{a_{(m);N_{m}}\cdots\sum_{N_{p+1}=q}^{N_{p+2}}{a_{(p+1);N_{p+1}}\sum_{N_{1}=q}^{N_{p+1}}{a_{(1);N_{1}}\sum_{N_{p}=N_{1}}^{N_{p+1}}{a_{(p);N_{p}}\sum_{N_{p-1}=N_{1}}^{N_{p}}{a_{(p-1);N_{p-1}}\cdots\sum_{N_{2}=N_{1}}^{N_{3}}{a_{(2);N_{2}}}}}}}}.$ ###### Proof. By applying Theorem 3.3 (with $m$ substituted by $p$ and $n$ substituted by $N_{p+1}$) to Theorem 3.4, we get the desired theorem. ∎ ###### Corollary 3.5. If all sequences are the same, Theorem 3.5 becomes $\sum_{N_{m}=q}^{n}{a_{N_{m}}\cdots\sum_{N_{p}=q}^{N_{p+1}}{a_{N_{p}}\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{1}}}}}=\sum_{N_{m}=q}^{n}{a_{N_{m}}\cdots\sum_{N_{p+1}=q}^{N_{p+2}}{a_{N_{p+1}}\sum_{N_{1}=q}^{N_{p+1}}{a_{N_{1}}\sum_{N_{p}=N_{1}}^{N_{p+1}}{a_{N_{p}}\sum_{N_{p-1}=N_{1}}^{N_{p}}{a_{N_{p-1}}\cdots\sum_{N_{2}=N_{1}}^{N_{3}}{a_{N_{2}}}}}}}}.$ ## 4 Reduction Formulas The objective of this section is to introduce formulas which can be used to reduce recurrent sums from their originally recurrent form $\left(\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}\right)$ to a form containing only simple non-recurrent sums $\left(\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)^{y}\right)$. ### 4.1 A brief introduction to partitions In this paper, partitions are involved in the reduction formula for a recurrent sum. For this reason, in this section, we will present a brief introduction to partitions. ###### Definition. A partition of a non-negative integer $m$ is a set of positive integers whose sum equals $m$. We can represent a partition of $m$ as a vector $(y_{k,1},\cdots,y_{k,m})$ that verifies $\displaystyle\begin{pmatrix}y_{k,1}\\\ \vdots\\\ y_{k,m}\\\ \end{pmatrix}\cdot\begin{pmatrix}1\\\ \vdots\\\ m\\\ \end{pmatrix}=y_{k,1}+2y_{k,2}+\cdots+my_{k,m}=m.$ (4) The set of partitions of a non-negative integer $m$ is the set of vectors $(y_{k,1},\cdots,y_{k,m})$ that verify the previous identity. We will denote this set by $P$. The cardinality of this set is equal to the number of partitions of $m$ (which is the partition function denoted by $p(m)$), $\text{Card}(P)=p(m).$ (5) Hence, the set of partitions of $m$ is the set of vectors $\\{(y_{1,1},\cdots,y_{1,m}),(y_{2,1},\cdots,y_{2,m}),\cdots\\}$ which consists of $p(m)$ vectors. The value of $p(m)$ is obtained from the generating function developed by Euler in the mid-eighteen century [16], $\sum_{m=0}^{\infty}{p(m)x^{m}}=\prod_{j=1}^{\infty}{\frac{1}{1-x^{j}}}.$ (6) Euler also showed that this relation implies the following recurrent definition for $p(m)$, $p(m)=\sum_{j=1}^{\infty}{(-1)^{j-1}\left(p\left(m-\frac{j(3j-1)}{2}\right)-p\left(m-\frac{j(3j+1)}{2}\right)\right)}.$ (7) In 1918, Hardy and Ramanujan provided an asymptotic expression for $p(m)$ in [22]. Later, in 1937, Rademacher was able to improve on Hardy and Ramanujan’s formula by proving the following expression for $p(m)$ in [34], $p(m)=\frac{1}{\pi\sqrt{2}}\sum_{k=1}^{\infty}{\sqrt{k}A_{k}(m)\frac{d}{dm}\left[\frac{\sinh\left(\frac{\pi}{k}\sqrt{\frac{2}{3}\left(m-\frac{1}{24}\right)}\right)}{\sqrt{m-\frac{1}{24}}}\right]}$ (8) where $A_{k}(m)$ is a Kloosterman type sum, $A_{k}(m)=\sum_{\begin{subarray}{c}0\leq h<k\\\ gcd(h,k)=1\end{subarray}}{e^{\pi i(s(h,k)-2mh/k)}}$ (9) and where the notation $s(m,k)$ represents a Dedekind sum. However, this formula has the disadvantage of being an infinite sum. This formula remained the only exact explicit formula for $p(m)$ until Ono and Bruinier presented a new formula for $p(m)$ as a finite sum [10]. Additionally, two of the most famous ways of representing a partition are using Ferrers diagrams or using Young diagrams. Similarly, there exists some variants of Ferrers diagrams that are used (see [33]). ###### Remark. For readers intrested in a more detailed explanation of partition, see [1]. ### 4.2 Reduction Theorem and Partition Identities We will start this section by proving several lemmas which are needed in order to prove the main theorem of this section (Theorem 4.1, which we will call the reduction theorem). However, some of these lemmas are important on their own as they provide relations governing partitions. We start by proving the following trivial lemma. ###### Lemma 4.1. No partition of a non-negative integer $m$ constructed from a sum of $r$ terms (positive integers) can contain an integer larger or equal to $m-r+2$. ###### Proof. The smallest sum of $r$ positive integers containing $i$ is $i+\underbrace{(1+\cdots+1)}_{(r-1)}=i+(r-1)$. If $i\geq m-r+2$ then $i+(r-1)\geq m-r+2+r-1=m+1>m$. Hence, such a sum, being strictly larger than $m$, cannot be a partition of $m$. ∎ Before we can proceed to prove the other needed lemmas, we need to define the following notation: Let $[x^{r}]\left(P(x)\right)$ represent the coefficient of $x^{r}$ in $P(x)$. Let $x^{\overline{m}}=x(x+1)\cdots(x+m-1)$ represent the Rising factorial. Let $(x)_{m}=x(x-1)\cdots(x-m+1)$ represent the Falling factorial. The original definition of Stirling numbers of the first kind $S(m,r)$ was as the coefficients in the expansion of $(x)_{m}$: $(x)_{m}=\sum_{k=0}^{m}{S(m,k)x^{k}}\,\,\,\,\,\,\,\,or\,\,\,\,\,\,\,\,S(m,r)=[x^{r}](x)_{m}.$ (10) In a similar way, the unsigned Stirling numbers of the first kind, denoted $|S(m,r)|$ or ${m\brack r}$, can be expressed in terms of the Rising factorial $x^{\overline{m}}$: $x^{\overline{m}}=\sum_{k=0}^{m}{{m\brack k}x^{k}}\,\,\,\,\,\,\,\,or\,\,\,\,\,\,\,\,{m\brack r}=[x^{r}]\left(x^{\overline{m}}\right).$ (11) From this definition, the famous finite sum of the unsigned Stirling numbers of the first kind can be directly deduced by substituting $x$ by 1 to get $\sum_{k=0}^{m}{{m\brack k}}=1(1+1)\cdots(1+m-1)=m!.$ (12) Note that $|S(m,r)|$ can also be defined as the number of permutations of $m$ elements with $r$ disjoint cycles. Similarly, the previous relation can be obtained by noticing that permutations are partitioned by number of cycles. ###### Remark. More details on Stirling numbers of the first kind can be found in [28]. For simplicity, we define $\sum{f(i)}$ to mean $\sum_{i=1}^{m}{f(i)}$. In particular, $\sum{i.y_{k,i}}=\sum_{i=1}^{m}{i.y_{k,i}}$ and $\sum{y_{k,i}}=\sum_{i=1}^{m}{y_{k,i}}$. Additionally, let a partition of $m$ of length $r$ refer to a partition $(y_{k,1},\cdots,y_{k,m})$ of $m$ such that $\sum{y_{k,i}}=r$. Now that we have defined the needed notation, we can continue proving the required lemmas. ###### Lemma 4.2. Let $m$ and $r$ be two non-negative integers with $r\leq m$, the following sum over partitions of $m$ of length $r$ can be expressed in terms of the unsigned Stirling numbers of the first kind as follows, $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}}(y_{k,i})!}}}=\frac{1}{m!}{m\brack r}.$ ###### Proof. A Bell polynomial is defined as follows $B_{m,r}(x_{1},x_{2},\cdots,x_{m-r+1})=\sum_{\begin{subarray}{c}y_{1}+2y_{2}+\cdots+(m-r+1)y_{m-r+1}=m\\\ y_{1}+y_{2}+\cdots+y_{m-r+1}=r\end{subarray}}{\frac{m!}{y_{1}!y_{2}!\cdots y_{m-r+1}!}\left(\frac{x_{1}}{1!}\right)^{y_{1}}\left(\frac{x_{2}}{2!}\right)^{y_{2}}\cdots\left(\frac{x_{m-r+1}}{(m-r+1)!}\right)^{y_{m-r+1}}}.$ These polynomials can also be rewritten more compactly as $B_{m,r}(x_{1},x_{2},\cdots,x_{m-r+1})=m!\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m-r+1}{\frac{1}{y_{k,i}!}\left(\frac{x_{i}}{i!}\right)^{y_{k,i}}}}.$ A property of the Bell polynomial, shown in [36], is that the value of the Bell polynomial on the sequence of factorials equals an unsigned Stirling number of the first kind, $B_{m,r}(0!,1!,\cdots,(m-r)!)=|S(m,r)|={m\brack r}.$ Likewise, by a numerical substitution into the definition of Bell polynomials, we have $B_{m,r}(0!,1!,\cdots,(m-r)!)=m!\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m-r+1}{\frac{1}{i^{y_{k,i}}(y_{k,i})!}}}.$ Hence, by equating, we get $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m-r+1}{\frac{1}{i^{y_{k,i}}(y_{k,i})!}}}=\frac{1}{m!}{m\brack r}.$ From Lemma 4.1, we know that the biggest integer that can appear in a partition of an integer $m$ using $r$ terms is $m-r+1$ (which means that $y_{k,m-r}=\cdots=y_{k,m}=0$). Thus, we get $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}}(y_{k,i})!}}}=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m-r+1}{\frac{1}{i^{y_{k,i}}(y_{k,i})!}}}=\frac{1}{m!}{m\brack r}.$ ∎ By Adding the arguments of the sum from Lemma 4.2 for all possible partition lengths, we obtain the following identity. ###### Lemma 4.3. Let $m$ be a non-negative integer, the following sum over all partitions of $m$ can be shown to equal $1$ independently of the value of $m$, $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}}(y_{k,i})!}}}=1.$ ###### Proof. From Lemma 4.2, we have $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}}(y_{k,i})!}}}=\frac{1}{m!}{m\brack r}.$ Hence, $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}}(y_{k,i})!}}}=\sum_{r=0}^{m}{\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}}(y_{k,i})!}}}}=\sum_{r=0}^{m}{\frac{1}{m!}{m\brack r}}=\frac{1}{m!}\sum_{r=0}^{m}{{m\brack r}}.$ However, we have already shown that the finite sum of ${m\brack r}$ is equal to $m!$. Hence, $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}}(y_{k,i})!}}}=1.$ ∎ A more general form of Lemma 4.2 is illustrated in the following lemma. ###### Lemma 4.4. Let $(\varphi_{1},\cdots,\varphi_{m})$ be a partition of $\varphi\leq m$ such that $\sum{\varphi_{i}}=r_{\varphi}$. Let $(y_{k,1},\cdots,y_{k,m})=\\{(y_{1,1},\cdots,y_{1,m}),(y_{2,1},\cdots,y_{2,m}),\cdots\\}$ be the set of all partitions of $m$. $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\\\ y_{k,i}\geq\varphi_{i}\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\frac{1}{(m-\varphi)!}{m-\varphi\brack r-r_{\varphi}}\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}(\varphi_{i})!}}.$ ###### Remark. Knowing that the largest element of a partition of $\varphi$ is $\varphi$, we can rewrite it as follows $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\\\ y_{k,i}\geq\varphi_{i}\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\frac{1}{(m-\varphi)!}{m-\varphi\brack r-r_{\varphi}}\prod_{i=1}^{\varphi}{\frac{1}{i^{\varphi_{i}}(\varphi_{i})!}}.$ ###### Proof. Knowing that $\binom{n}{k}$ is zero if $n<k$, then $\binom{y_{k,i}}{\varphi_{i}}=0$ if $\exists i\in\mathbb{N},y_{k,i}<\varphi_{i}$. Hence, $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\\\ \exists i,y_{k,i}<\varphi_{i}\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}+\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\\\ y_{k,i}\geq\varphi_{i}\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\\\ y_{k,i}\geq\varphi_{i}\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}.$ The first part of the proof is complete. $\begin{split}\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}&=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}}(y_{k,i})!}.\frac{y_{k,i}!}{\varphi_{i}!(y_{k,i}-\varphi_{i})!}}}\\\ &=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}}}.\frac{1}{\varphi_{i}!(y_{k,i}-\varphi_{i})!}}}\\\ &=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}.\frac{1}{i^{y_{k,i}-\varphi_{i}}(y_{k,i}-\varphi_{i})!}}}\\\ &=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}-\varphi_{i}}(y_{k,i}-\varphi_{i})!}}}.\end{split}$ As $\varphi_{1},\cdots,\varphi_{m}$ are all constants then $\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}$ is constant. This factor is constant and is common to all terms of the sum, therefore, we can factor it and take it outside the sum. $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\left(\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}\right)\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}-\varphi_{i}}(y_{k,i}-\varphi_{i})!}}}.$ Having that $(\varphi_{1},\cdots,\varphi_{m})$ is a partition of $\varphi\leq m$, hence, $\sum{i.\varphi_{i}}=\varphi\leq m$. Thus, the condition $\sum{i.y_{k,i}}=m$ can be replaced by $\sum{i.(y_{k,i}-\varphi_{i})}=\sum{i.y_{k,i}}-\sum{i.\varphi_{i}}=m-\varphi$. Similarly, $r_{\varphi}=\sum{\varphi_{i}}$, hence, the condition $\sum{y_{k,i}}=r$ can be replaced by $\sum{(y_{k,i}-\varphi_{i})}=\sum{y_{k,i}}-\sum{\varphi_{i}}=r-r_{\varphi}$. Hence, $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\left(\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}\right)\sum_{\begin{subarray}{c}k\\\ \sum{i.(y_{k,i}-\varphi_{i})}=m-\varphi\\\ \sum{(y_{k,i}-\varphi)}=r-r_{\varphi}\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}-\varphi_{i}}(y_{k,i}-\varphi_{i})!}}}.$ Let $Y_{k,i}=y_{k,i}-\varphi_{i}$, $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\left(\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}\right)\sum_{\begin{subarray}{c}k\\\ \sum{i.Y_{k,i}}=m-\varphi\\\ \sum{Y_{k,i}}=r-r_{\varphi}\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{Y_{k,i}}Y_{k,i}!}}}.$ Knowing that the largest element of a partition of $(m-\varphi)$ is $(m-\varphi)$, hence, $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\left(\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}\right)\sum_{\begin{subarray}{c}k\\\ \sum{i.Y_{k,i}}=m-\varphi\\\ \sum{Y_{k,i}}=r-r_{\varphi}\end{subarray}}{\prod_{i=1}^{m-\varphi}{\frac{1}{i^{Y_{k,i}}Y_{k,i}!}}}.$ Applying Lemma 4.2, with $y_{k,i}$ substituted by $Y_{k,i}$, $m$ substituted by $m-\varphi$, and $r$ substituted by $r-r_{\varphi}$, we get $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\left(\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}\right)\frac{1}{(m-\varphi)!}{m-\varphi\brack r-r_{\varphi}}.$ The proof is complete. ∎ ###### Remark. If $\varphi>m$, then $\sum{i.Y_{k,i}}=m-\varphi<0$ which makes Lemma 4.2 invalid which then makes this lemma invalid. Similarly, a more general form of Lemma 4.3 is illustrated in the following lemma. ###### Lemma 4.5. Let $(y_{k,1},\cdots,y_{k,m})=\\{(y_{1,1},\cdots,y_{1,m}),(y_{2,1},\cdots,y_{2,m}),\cdots\\}$ be the set of all partitions of $m$. Let $(\varphi_{1},\cdots,\varphi_{m})$ be a partition of $r\leq m$. $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ y_{k,i}\geq\varphi_{i}\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}(\varphi_{i})!}}.$ ###### Remark. Knowing that the largest element of a partition of $r$ is $r$, we can rewrite it as follows $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ y_{k,i}\geq\varphi_{i}\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\prod_{i=1}^{r}{\frac{1}{i^{\varphi_{i}}(\varphi_{i})!}}.$ ###### Proof. Knowing that $\binom{n}{k}$ is zero if $n<k$, then $\binom{y_{k,i}}{\varphi_{i}}=0$ if $\exists i\in\mathbb{N},y_{k,i}<\varphi_{i}$. Hence, $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ \exists i,y_{k,i}<\varphi_{i}\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}+\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ y_{k,i}\geq\varphi_{i}\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\\\ y_{k,i}\geq\varphi_{i}\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}.$ The first part of the proof is complete. $\begin{split}\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}&=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}}(y_{k,i})!}.\frac{y_{k,i}!}{\varphi_{i}!(y_{k,i}-\varphi_{i})!}}}\\\ &=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}}}.\frac{1}{\varphi_{i}!(y_{k,i}-\varphi_{i})!}}}\\\ &=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}.\frac{1}{i^{y_{k,i}-\varphi_{i}}(y_{k,i}-\varphi_{i})!}}}\\\ &=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}-\varphi_{i}}(y_{k,i}-\varphi_{i})!}}}.\end{split}$ As $\varphi_{1},\cdots,\varphi_{m}$ are all constants then $\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}$ is constant. This factor is constant and is common to all terms of the sum, therefore, we can factor it and take it outside the sum. $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\left(\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}\right)\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}-\varphi_{i}}(y_{k,i}-\varphi_{i})!}}}.$ Having that $(\varphi_{1},\cdots,\varphi_{m})$ is a partition of $r\leq m$, hence, $\sum{i.\varphi_{i}}=r\leq m$. Thus, the condition $\sum{i.y_{k,i}}=m$ can be replaced by $\sum{i.(y_{k,i}-\varphi_{i})}=\sum{i.y_{k,i}}-\sum{i.\varphi_{i}}=m-r(\geq 0)$. Hence, $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\left(\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}\right)\sum_{\begin{subarray}{c}k\\\ \sum{i.(y_{k,i}-\varphi_{i})}=m-r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{y_{k,i}-\varphi_{i}}(y_{k,i}-\varphi_{i})!}}}.$ Let $Y_{k,i}=y_{k,i}-\varphi_{i}$, $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\left(\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}\right)\sum_{\begin{subarray}{c}k\\\ \sum{i.Y_{k,i}}=m-r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{i^{Y_{k,i}}Y_{k,i}!}}}.$ Knowing that the largest element of a partition of $(m-r)$ is $(m-r)$, hence, $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\left(\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}\right)\sum_{\begin{subarray}{c}k\\\ \sum{i.Y_{k,i}}=m-r\end{subarray}}{\prod_{i=1}^{m-r}{\frac{1}{i^{Y_{k,i}}Y_{k,i}!}}}.$ Applying Lemma 4.3, with $y_{k,i}$ substituted by $Y_{k,i}$ and $m$ substituted by $m-r$, we get $\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{\binom{y_{k,i}}{\varphi_{i}}}{i^{y_{k,i}}(y_{k,i})!}}}=\left(\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}\varphi_{i}!}}\right).$ The proof is complete. ∎ ###### Remark. If $r>m$, then $\sum{i.Y_{k,i}}=m-r<0$ which makes Lemma 4.3 invalid which then makes this lemma invalid. ###### Proposition 4.1. Let $B_{m,r}(x_{1},\cdots,x_{m-r+1})$ be the partial Bell polynomial and $B_{m}(x_{1},\cdots,x_{m})$ be the complete Bell polynomial, $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\frac{1}{m!}\sum_{r=0}^{m}{B_{m,r}(x_{1},\cdots,x_{m-r+1})}=\frac{1}{m!}B_{m}(x_{1},\cdots,x_{m})$ where $x_{i}=(i-1)!(\sum_{N=q}^{n}{(a_{N})^{i}})$. ###### Proof. From Lemma 4.1, we can write $\sum_{\begin{subarray}{c}\sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{\begin{subarray}{c}\sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m-r+1}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n}{(a_{N})^{i}}\right)^{y_{k,i}}}}.$ We can notice that the right side term of the previous expression corresponds to a multiple of a special value of the partial Bell polynomial where $x_{i}=(i-1)!(\sum_{N=q}^{n}{(a_{N})^{i}}),\forall i\in[1,m]$. Hence, $\sum_{\begin{subarray}{c}\sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\frac{1}{m!}B_{m,r}(x_{1},\cdots,x_{m-r+1}).$ Additionally, the sum over the partitions of $m$ is equivalent to the sum for $r$ going from $0$ to $m$ of the sums over the partitions of $m$ of length $r$. Thus, $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{r=0}^{m}{\sum_{\begin{subarray}{c}\sum{i.y_{k,i}}=m\\\ \sum{y_{k,i}}=r\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n}{(a_{N})^{i}}\right)^{y_{k,i}}}}}=\frac{1}{m!}\sum_{r=0}^{m}{B_{m,r}(x_{1},\cdots,x_{m-r+1})}.$ Applying the definition of a complete Bell polynomial, we get $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\frac{1}{m!}B_{m}(x_{1},\cdots,x_{m}).$ ∎ Now that all the required lemmas have been proven, we show the following theorem which allows the representation of a recurrent sum in terms of non- recurrent sums. ###### Theorem 4.1 (Reduction Theorem). Let $m$ be a non-negative integer, $k$ be the index of the $k$-th partition of $m$ $(1\leq k\leq p(m))$, $i$ be an integer between $1$ and $m$, and $y_{k,i}$ be the multiplicity of $i$ in the $k$-th partition of $m$. The reduction theorem for recurrent sums is stated as follow: $\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n}{(a_{N})^{i}}\right)^{y_{k,i}}}}.$ ###### Proof. 1\. Base Case: verify true for $n=q$, $\forall m\in\mathbb{N}$. $\begin{split}\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{q}{(a_{N})^{i}}\right)^{y_{k,i}}}}&=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\left(a_{q}\right)^{i.y_{k,i}}}}\\\ &=\sum_{\sum{i.y_{k,i}}=m}{\left(a_{q}\right)^{\sum{i.y_{k,i}}}\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}}}\\\ &=\left(a_{q}\right)^{m}\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}}}.\end{split}$ By applying Lemma 4.3, we get $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{q}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\left(a_{q}\right)^{m}.$ Likewise, $\sum_{N_{m}=q}^{q}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}=a_{q}\cdots a_{q}=\left(a_{q}\right)^{m}.$ 2\. Induction hypothesis: assume the statement is true until $n$, $\forall m\in\mathbb{N}$. $\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n}{(a_{N})^{i}}\right)^{y_{k,i}}}}.$ 3\. Induction step: we will show that this statement is true for $(n+1)$, $\forall m\in\mathbb{N}$. We have to show the following statement to be true: $\sum_{N_{m}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n+1}{(a_{N})^{i}}\right)^{y_{k,i}}}}.$ $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n+1}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\left(\sum_{N=q}^{n}{(a_{N})^{i}}+(a_{n+1})^{i}\right)^{y_{k,i}}}}.$ The binomial theorem states that $(a+b)^{n}=\sum_{\varphi=0}^{n}{\binom{n}{\varphi}a^{n-\varphi}b^{\varphi}}.$ Hence, $\begin{split}\left(\sum_{N=q}^{n}{(a_{N})^{i}}+(a_{n+1})^{i}\right)^{y_{k,i}}&=\sum_{\varphi=0}^{y_{k,i}}{\binom{y_{k,i}}{\varphi}{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi}{\left((a_{n+1})^{i}\right)}^{y_{k,i}-\varphi}}.\end{split}$ Thus, $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n+1}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\sum_{\varphi=0}^{y_{k,i}}{\binom{y_{k,i}}{\varphi}{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi}{\left((a_{n+1})^{i}\right)}^{y_{k,i}-\varphi}}}}=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\sum_{\varphi=0}^{y_{k,i}}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\binom{y_{k,i}}{\varphi}{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi}{\left(a_{n+1}\right)}^{i.y_{k,i}-i.\varphi}}}}.$ Let $A_{\varphi,i,k}={\frac{1}{(y_{k,i})!i^{y_{k,i}}}\binom{y_{k,i}}{\varphi}{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi}{\left(a_{n+1}\right)}^{i.y_{k,i}-i.\varphi}}$. By expanding then regrouping, it can be seen that $\prod_{i=1}^{m}{\sum_{\varphi=0}^{y_{k,i}}{A_{\varphi,i,k}}}=\sum_{\varphi_{m}=0}^{y_{k,m}}{\cdots\sum_{\varphi_{1}=0}^{y_{k,1}}{\prod_{i=1}^{m}{A_{\varphi_{i},i,k}}}}.$ This is because, for any given $k$, by expanding the product of sums (the left hand side term), we will get a sum of products of the form $A_{\varphi_{1},1}A_{\varphi_{2},2}\cdots A_{\varphi_{m},m}$ ($\prod_{i=1}^{m}{A_{\varphi_{i},i}}$) for all combinations of $\varphi_{1},\varphi_{2},\cdots,\varphi_{m}$ such that $0\leq\varphi_{1}\leq y_{k,1},\cdots,0\leq\varphi_{m}\leq y_{k,m}$, which is equivalent to the right hand side term. Hence, $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n+1}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{\sum{i.y_{k,i}}=m}{\sum_{\varphi_{m}=0}^{y_{k,m}}{\cdots\sum_{\varphi_{1}=0}^{y_{k,1}}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\binom{y_{k,i}}{\varphi_{i}}{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi_{i}}{\left(a_{n+1}\right)}^{i.y_{k,i}-i.\varphi_{i}}}}}}.$ A more compact way of writing the repeated sum over the $\varphi_{i}$’s is by expressing it with one sum that combines all the conditions. The set of conditions $0\leq\varphi_{1}\leq y_{k,1},\cdots,0\leq\varphi_{m}\leq y_{k,m}$ can be expressed as the condition $0\leq\varphi_{i}\leq y_{k,i}$ for $i\in[1,m]$. $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n+1}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{\sum{i.y_{k,i}}=m}{\sum_{0\leq\varphi_{i}\leq y_{k,i}}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\binom{y_{k,i}}{\varphi_{i}}{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi_{i}}{\left(a_{n+1}\right)}^{i.y_{k,i}-i.\varphi_{i}}}}}.$ Similarly, let $j$ represent $\sum{i.\varphi_{i}}$. Hence, we can add the trivial condition that is $j=\sum{i.\varphi_{i}}$ to the sum over $\varphi_{i}$. Additionally, $\sum{i.\varphi_{i}}=j$ is minimal when $\varphi_{1}=0,\cdots,\varphi_{m}=0$. Hence $j_{min}=0$. $\sum{i.\varphi_{i}}=j$ is maximal when $\varphi_{1}=y_{k,1},\cdots,\varphi_{m}=y_{k,m}$. Hence $j_{max}=\sum{i.y_{k,i}}=m$. Therefore, we have that $0\leq j\leq m$ or equivalently that $j$ can go from $0$ to $m$. Hence, knowing that adding a true statement to a condition does not change the condition, we can add this additional condition to get $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n+1}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{\sum{i.y_{k,i}}=m}{\sum_{\begin{subarray}{c}j=0\\\ \sum{i.\varphi_{i}}=j\\\ 0\leq\varphi_{i}\leq y_{k,i}\end{subarray}}^{m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\binom{y_{k,i}}{\varphi_{i}}{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi_{i}}{\left(a_{n+1}\right)}^{i.y_{k,i}-i.\varphi_{i}}}}}.$ Knowing that $\binom{y_{k,i}}{\varphi_{i}}=0$ if $\varphi_{i}>y_{k,i}$, hence, the terms produced for $\varphi_{i}>y_{k,i}$ would be zero. Thus, we can remove the condition $0\leq\varphi_{i}\leq y_{k,i}$ because terms that do not satisfy this condition will be zeros and, therefore, would not change the value of the sum. $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n+1}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{\sum{i.y_{k,i}}=m}{\sum_{\begin{subarray}{c}j=0\\\ \sum{i.\varphi_{i}}=j\end{subarray}}^{m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\binom{y_{k,i}}{\varphi_{i}}{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi_{i}}{\left(a_{n+1}\right)}^{i.y_{k,i}-i.\varphi_{i}}}}}.$ We expand the expression then, from all values of $k$ (from every partitions $(y_{k,1},\cdots,y_{k,m})$ of $m$), we regroup together the terms having a combination of exponents $(\varphi_{1},\cdots,\varphi_{m})$ that forms a partition of the same integer $j$ and we do so $\forall j\in[0,m]$. Hence, performing this manipulation allows us to interchange the sum over $k$ (over $\sum_{i}{i.y_{k,i}}=m$) with the sums over $j$. Thus, the expression becomes as follows, $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n+1}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{\begin{subarray}{c}j=0\\\ \sum{i.\varphi_{i}}=j\end{subarray}}^{m}{\sum_{\begin{subarray}{c}\sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\binom{y_{k,i}}{\varphi_{i}}{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi_{i}}{\left(a_{n+1}\right)}^{i.y_{k,i}-i.\varphi_{i}}}}}=\sum_{\begin{subarray}{c}j=0\\\ \sum{i.\varphi_{i}}=j\end{subarray}}^{m}{\sum_{\begin{subarray}{c}\sum{i.y_{k,i}}=m\end{subarray}}{{\left(a_{n+1}\right)}^{\sum{i.y_{k,i}}-\sum{i.\varphi_{i}}}\left[\prod_{i=1}^{m}{{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi_{i}}}\right]\left[\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\binom{y_{k,i}}{\varphi_{i}}}\right]}}=\sum_{\begin{subarray}{c}j=0\\\ \sum{i.\varphi_{i}}=j\end{subarray}}^{m}{{\left(a_{n+1}\right)}^{m-j}\left[\prod_{i=1}^{m}{{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi_{i}}}\right]\left(\sum_{\begin{subarray}{c}\sum{i.y_{k,i}}=m\end{subarray}}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\binom{y_{k,i}}{\varphi_{i}}}}\right)}.$ Applying Lemma 4.5, we get $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n+1}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{\begin{subarray}{c}j=0\\\ \sum{i.\varphi_{i}}=j\end{subarray}}^{m}{{\left(a_{n+1}\right)}^{m-j}\left[\prod_{i=1}^{m}{{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi_{i}}}\right]\left(\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}(\varphi_{i})!}}\right)}=\sum_{\begin{subarray}{c}j=0\\\ \sum{i.\varphi_{i}}=j\end{subarray}}^{m}{{\left(a_{n+1}\right)}^{m-j}\left(\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}(\varphi_{i})!}}{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi_{i}}\right)}.$ Knowing that for any given value of $j$ there is multiple combinations of $\varphi_{1},\cdots,\varphi_{m}$ that satisfy $\sum{i.\varphi_{i}}=j$. Hence, every value of $j$ corresponds to a sum of the sum’s argument for all partitions of $j$ (for all combinations of $\varphi_{1},\cdots,\varphi_{m}$ satisfying $\sum{i.\varphi_{i}}=j$). Therefore, we can split the outer sum with two conditions into two sums each with one of the conditions as follows, $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n+1}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{j=0}^{m}{{\left(a_{n+1}\right)}^{m-j}\sum_{\sum{i.\varphi_{i}}=j}{\left(\prod_{i=1}^{m}{\frac{1}{i^{\varphi_{i}}(\varphi_{i})!}}{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi_{i}}\right)}}.$ Knowing that the largest element of a partition of $j$ is $j$, $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n+1}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{j=0}^{m}{{\left(a_{n+1}\right)}^{m-j}\left(\sum_{\sum{i.\varphi_{i}}=j}{\prod_{i=1}^{j}{\frac{1}{i^{\varphi_{i}}(\varphi_{i})!}}{\left(\sum_{N=q}^{n}{(a_{N})^{i}}\right)}^{\varphi_{i}}}\right)}.$ By using the induction hypothesis, the expression becomes $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n+1}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{j=0}^{m}{\left(a_{n+1}\right)^{m-j}\left(\sum_{N_{j}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{j}}\cdots a_{N_{1}}}}\right)}.$ Using Corollary 2.1, we get $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=q}^{n+1}{(a_{N})^{i}}\right)^{y_{k,i}}}}=\sum_{N_{m}=q}^{n+1}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}.$ The theorem is proven by induction. ∎ ###### Corollary 4.1. If the recurrent sum starts at 1, Theorem 4.1 becomes $\sum_{N_{m}=1}^{n}{\cdots\sum_{N_{1}=1}^{N_{2}}{a_{N_{m}}\cdots a_{N_{1}}}}=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=1}^{n}{(a_{N})^{i}}\right)^{y_{k,i}}}}.$ An additional partition identity that can be deduced from Theorem 4.1 is as follows. ###### Corollary 4.2. For any $m,n\in\mathbb{N}$, we have that $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{n}{i}\right)^{y_{k,i}}}}=\binom{n+m-1}{m}.$ ###### Proof. From paper [21], we have the following relation, $\sum_{N_{m}=1}^{n}{\cdots\sum_{N_{1}=1}^{N_{2}}{1}}=\binom{n+m-1}{m}.$ By applying Theorem 4.1, we get $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{1}{i}\sum_{N=1}^{n}{1}\right)^{y_{k,i}}}}=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!}\left(\frac{n}{i}\right)^{y_{k,i}}}}=\binom{n+m-1}{m}.$ ∎ ###### Example 4.1. For $n=1$, Corollary 4.2 gives $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}}}=\binom{m}{m}=1.$ ###### Example 4.2. For $n=2$, Corollary 4.2 gives $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{2^{y_{k,i}}}{(y_{k,i})!i^{y_{k,i}}}}}=\binom{m+1}{m}=m+1.$ ###### Example 4.3. For $n=3$, Corollary 4.2 gives $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{3^{y_{k,i}}}{(y_{k,i})!i^{y_{k,i}}}}}=\binom{m+2}{m}=\frac{(m+1)(m+2)}{2}.$ ### 4.3 Particular cases In this section, we will apply the reduction formula for the cases of $m$ from $1$ to $4$. These cases were independently proven using two distinct methods (which are omitted here for simplicity). Similarly, these formulas were verified for a certain range of $n$ using a computer program which calculated the right expression as well as the left expression then checks that they are equal. * • For $m=1$ $\sum_{N_{1}=1}^{n}{a_{N_{1}}}=\sum_{N=1}^{n}{a_{N}}.$ * • For $m=2$ $\sum_{N_{2}=1}^{n}{\sum_{N_{1}=1}^{N_{2}}{a_{N_{2}}a_{N_{1}}}}=\frac{1}{2}\left(\sum_{N=1}^{n}{a_{N}}\right)^{2}+\frac{1}{2}\left(\sum_{N=1}^{n}{\left(a_{N}\right)^{2}}\right).$ * • For $m=3$ $\sum_{N_{3}=1}^{n}{\sum_{N_{2}=1}^{N_{3}}{\sum_{N_{1}=1}^{N_{2}}{a_{N_{3}}a_{N_{2}}a_{N_{1}}}}}=\frac{1}{6}\left(\sum_{N=1}^{n}{a_{N}}\right)^{3}+\frac{1}{2}\left(\sum_{N=1}^{n}{a_{N}}\right)\left(\sum_{N=1}^{n}{\left(a_{N}\right)^{2}}\right)+\frac{1}{3}\left(\sum_{N=1}^{n}{\left(a_{N}\right)^{3}}\right).$ * • For $m=4$ $\sum_{N_{4}=1}^{n}{\sum_{N_{3}=1}^{N_{4}}{\sum_{N_{2}=1}^{N_{3}}{\sum_{N_{1}=1}^{N_{2}}{a_{N_{4}}a_{N_{3}}a_{N_{2}}a_{N_{1}}}}}}=\frac{1}{24}\left(\sum_{N=1}^{n}{a_{N}}\right)^{4}+\frac{1}{4}\left(\sum_{N=1}^{n}{a_{N}}\right)^{2}\left(\sum_{N=1}^{n}{\left(a_{N}\right)^{2}}\right)+\frac{1}{3}\left(\sum_{N=1}^{n}{a_{N}}\right)\left(\sum_{N=1}^{n}{\left(a_{N}\right)^{3}}\right)+\frac{1}{8}\left(\sum_{N=1}^{n}{\left(a_{N}\right)^{2}}\right)^{2}+\frac{1}{4}\left(\sum_{N=1}^{n}{\left(a_{N}\right)^{4}}\right).$ ### 4.4 General Reduction Theorem We define the notation $|A|$ as the number of elements in the set $A$. Note that if $A$ is a set of sets then $|A|$ represents the number of sets in $A$ as they are considered the elements of $A$. Let $m$ be a non-negative integer and let $\\{(y_{1,1},\cdots,y_{1,m}),(y_{2,1},\cdots,y_{2,m}),\cdots\\}$ be the set of all partitions of $m$. Let us consider the set $M=\\{1,\cdots,m\\}$. The permutation group $S_{m}$ is the set of all permutations of the set $\\{1,\cdots,m\\}$. Let $\sigma\in S_{m}$ be a permutation of the set $\\{1,\cdots,m\\}$ and let $\sigma(i)$ represent the $i$-th element of this given permutation. The number of such permutations is given by $|S_{m}|=m!.$ (13) The cycle-type of a permutation $\sigma$ is the ordered set where the $i$-th element represents the number of cycles of size $i$ in the cycle decomposition of $\sigma$. The number of ways of arranging $i$ elements cyclically is $(i-1)!$. The number of possible combinations of $y_{k,i}$ cycles of size $i$ is $[(i-1)!]^{y_{k,i}}$. Hence, the number of permutations having cycle-type $(y_{k,1},\cdots,y_{k,m})$ is given by $\prod_{i=1}^{m}{[(i-1)!]^{y_{k,i}}}.$ (14) A partition $P$ of a set $M$ is a set of non-empty disjoint subsets of $M$ such that every element of $M$ is present in exactly one of the subsets. Let $P=\\{\underbrace{P_{1,1},\cdots,P_{1,y_{1}}}_{y_{1}\,\,sets},\cdots,\underbrace{P_{m,1},\cdots,P_{m,y_{m}}}_{y_{m}\,\,sets}\\}$ represent a partition of a set of $m$ elements (for our purpose let it be the set $\\{1,\cdots,m\\}$). $P_{i,y}$ represents the $y$-th subset of order (size) $i$. $y_{i}$ represents the number of subsets of size $i$ contained in this partition of the set. It is interesting to note that $(y_{1},\cdots,y_{m})$ will always form a partition of the non-negative integer $m$. However, the number of partitions of $m$ is different from the number of partitions of a set of $m$ elements because there are more than one partition of the set of $m$ elements that can be associated with a given partition of $m$. In fact, we can easily determine that the number of partitions of a set of $m$ elements associated with the partition $(y_{1},\cdots,y_{m})$ is given by $|\Omega_{k}|=\frac{m!}{1!^{y_{k,1}}\cdots m!^{y_{k,m}}(y_{k,1})!\cdots(y_{k,m})!}=\frac{m!}{\prod_{i=1}^{m}{i!^{y_{k,i}}y_{k,i}!}}.$ (15) where $\Omega_{k}$ is the set of all partitions of the set of $m$ elements associated the partition $(y_{k,i},\cdots,y_{k,m})$. This is because the number of ways to divide $m$ objects into $l_{1}$ groups of $1$ element, $l_{2}$ groups of $2$ elements, $\cdots$, and $l_{m}$ groups of $m$ elements is given by $\frac{m!}{1!^{l_{1}}\cdots m!^{l_{m}}l_{1}!\cdots l_{m}!}=\frac{m!}{\prod_{i=1}^{m}{i!^{l_{i}}l_{i}!}}.$ (16) We will denote by $\Omega$ the set of all partitions of the set of $m$ elements. Finally, a partition $P$ of a set $M$ is a refinement of a partition $\rho$ of the same set $M$ if every element in $P$ is a subset of an element in $\rho$. We denote this as $P\succeq\rho$. Using the notation introduced, we can formulate a generalization of Theorem 4.1 where all sequences are distinct. ###### Theorem 4.2. Let $m,n,q\in\mathbb{N}$ such that $n\geq q$. Let $a_{(1);N},\cdots,a_{(m);N}$ be $m$ sequences defined in the interval $[q,n]$. we have that $\begin{split}\sum_{\sigma\in S_{m}}{\left(\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(\sigma(m));N_{m}}\cdots a_{(\sigma(1));N_{1}}}}\right)}=\sum_{\begin{subarray}{c}P\in\Omega\end{subarray}}{\prod_{i=1}^{m}{[(i-1)!]^{y_{k,i}}\left[\prod_{g=1}^{y_{k,i}}{\left(\sum_{N=q}^{n}{\prod_{h\in P_{i,g}}{a_{(h);N}}}\right)}\right]}}.\end{split}$ ###### Remark. The theorem can also be written as $\begin{split}&\sum_{\sigma\in S_{m}}{\left(\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(\sigma(m));N_{m}}\cdots a_{(\sigma(1));N_{1}}}}\right)}\\\ &\,\,=\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\sum_{\Omega_{k}}{\prod_{i=1}^{m}{[(i-1)!]^{y_{k,i}}\left[\prod_{g=1}^{y_{k,i}}{\left(\sum_{N=q}^{n}{\prod_{h\in P_{i,g}}{a_{(h);N}}}\right)}\right]}}}\\\ &\,\,=|S_{m}|\sum_{\begin{subarray}{c}k\\\ \sum{i.y_{k,i}}=m\end{subarray}}{\frac{1}{|\Omega_{k}|}\sum_{\Omega_{k}}{\prod_{i=1}^{m}{\frac{1}{y_{k,i}!i^{y_{k,i}}}\left[\prod_{g=1}^{y_{k,i}}{\left(\sum_{N=q}^{n}{\prod_{h\in P_{i,g}}{a_{(h);N}}}\right)}\right]}}}.\end{split}$ As every partition of a set of $m$ elements is associated with a given partition of $m$, hence, adding up all the partitions of the set for ever given partition of $m$ is equivalent to adding up all partitions of the set. The first form is obtained by regrouping together, from the set of all partitions of the set $\\{1,\cdots,m\\}$, those who are associated with a given partition of $m$. The second expression is obtained by noting that $\frac{|S_{m}|}{|\Omega_{k}|}\prod_{i=1}^{m}{\frac{1}{y_{k,i}!i^{y_{k,i}}}}=\prod_{i=1}^{m}{{[(i-1)!]^{y_{k,i}}}}$. These forms are shown as they can be more easily used to show that this theorem reduces to Theorem 4.1 if all sequences are the same. ###### Proof. Both sides of the equation produce all combinations of terms which are products of the $m$ sequences. Hence, the strategy of this proof is to show that every combination appear with the same multiplicity on both sides. We can assume the sequences to all be distinct without lost of generality. We can write $\sum_{\sigma\in S_{m}}{\left(\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(\sigma(m));N_{m}}\cdots a_{(\sigma(1));N_{1}}}}\right)}=\sum_{\sigma\in S_{m}}{\left(\sum_{N_{m}=q}^{n}{\cdots\sum_{N_{1}=q}^{N_{2}}{a_{(m);N_{\sigma(m)}}\cdots a_{(1);N_{\sigma(1)}}}}\right)}.$ Hence, we can consider the symmetric group $S_{m}$ as acting on $N=(N_{1},\cdots,N_{m})$. $N=(N_{1},\cdots,N_{m})$ has an isotropy group $S_{m}(N)$ and an associated partition $\rho$ of the set of $m$ elements. The partition $\rho$ is the set of all equivalence classes of the relation given by $a\sim b$ if and only if $N_{a}=N_{b}$ and $S_{m}(N)=\\{\sigma\in S_{m}\,\,|\,\,\sigma(i)\sim i\,\,\forall i\\}$. Thus, $a_{(m);N_{m}}\cdots a_{(1);N_{1}}$ (17) appears $|S_{m}(N)|$ times in the expansion of the left hand side of the theorem. Likewise, in the right hand side, (17) can only appears in the terms corresponding to partitions $P$ which are refinements of $\rho$. (17) appears $\sum_{P\succeq\rho}{\prod_{i=1}^{m}{[(i-1)!]^{y_{k,i}}}}$ (18) times in the right hand side of the theorem. Also let us notice that $[(i-1)!]^{y_{k,i}}$ corresponds to $(|P_{i,1}|-1)!\cdots(|P_{i,y_{k,i}}|-1)!$ because $|P_{i,1}|=\cdots=|P_{i,y_{k,i}}|=i$. Hence, $\prod_{i=1}^{m}{[(i-1)!]^{y_{k,i}}}$ corresponds to $\prod_{P_{h,g}\subset P}{(|P_{h,g}|-1)!}$ which is equal to the number of permutations having cycle- type specified by $P$. Knowing that any element of $S_{m}(N)$ has a unique cycle-type specified by a partition that refines $\rho$, hence, we conclude that $\sum_{P\succeq\rho}{\prod_{i=1}^{m}{[(i-1)!]^{y_{k,i}}}}=|S_{m}(N)|.$ (19) As both sides of the theorem produce the same terms and with the same multiplicity, we can say that these sides are equal to each other. ∎ ###### Example 4.4. For $m=2$, Theorem 4.2 gives the following, $\sum_{N_{2}=q}^{n}{\sum_{N_{1}=q}^{N_{2}}{a_{N_{2}}b_{N_{1}}}}+\sum_{N_{2}=q}^{n}{\sum_{N_{1}=q}^{N_{2}}{b_{N_{2}}a_{N_{1}}}}=\left(\sum_{N=q}^{n}{a_{N}}\right)\left(\sum_{N=q}^{n}{b_{N}}\right)+\left(\sum_{N=q}^{n}{a_{N}b_{N}}\right).$ ###### Example 4.5. For $m=3$, Theorem 4.2 gives the following, $\sum_{\sigma\in S_{3}}{\left(\sum_{N_{3}=q}^{n}{\sum_{N_{2}=q}^{N_{3}}{\sum_{N_{1}=q}^{N_{2}}{a_{(\sigma(3));N_{3}}a_{(\sigma(2));N_{2}}a_{(\sigma(1));N_{1}}}}}\right)}=\left(\sum_{N=q}^{n}{a_{(1);N}}\right)\left(\sum_{N=q}^{n}{a_{(2);N}}\right)\left(\sum_{N=q}^{n}{a_{(3);N}}\right)+\left(\sum_{N=q}^{n}{a_{(1);N}}\right)\left(\sum_{N=q}^{n}{a_{(2);N}a_{(3);N}}\right)+\left(\sum_{N=q}^{n}{a_{(2);N}}\right)\left(\sum_{N=q}^{n}{a_{(1);N}a_{(3);N}}\right)+\left(\sum_{N=q}^{n}{a_{(3);N}}\right)\left(\sum_{N=q}^{n}{a_{(1);N}a_{(2);N}}\right)+2\left(\sum_{N=q}^{n}{a_{(1);N}a_{(2);N}a_{(3);N}}\right).$ ### 4.5 Example applications In this section, we will apply the reduction formula presented in Theorem 4.1 to simplify certain special recurrent sums. The first special sum that we will simplify is a recurrent sum of $N^{p}$ which will produce a recurrent form of the Faulhaber formula. The second special sum is the recurrent harmonic series as well as the recurrent $p$-series for positive even values of $p$. #### 4.5.1 Recurrent Faulhaber Formula The Faulhaber formula is a formula developed by Faulhaber in a 1631 edition of Academia Algebrae [17] to calculate sums of powers $(N^{p})$. The Faulhaber formula is as follows $\sum_{N=1}^{n}{N^{p}}=\frac{1}{p+1}\sum_{j=0}^{p}{(-1)^{j}\binom{p+1}{j}B_{j}n^{p+1-j}}$ (20) where $B_{j}$ are the Bernoulli numbers of the first kind. ###### Remark. See [31] for details on the history of Bernoulli numbers. In this section, we will use the reduction formula for recurrent sums to develop a formula for a recurrent form of the Faulhaber formula. ###### Theorem 4.3. For any $m,n,p\in\mathbb{N}$, we have that $\begin{split}\sum_{N_{m}=1}^{n}{\cdots\sum_{N_{1}=1}^{N_{2}}{{N_{m}}^{p}\cdots{N_{1}}^{p}}}&=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\left(\sum_{N=1}^{n}{N^{ip}}\right)^{y_{k,i}}}}\\\ &=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\left(\frac{n^{ip+1}}{ip+1}\sum_{j=0}^{ip}{(-1)^{j}\binom{ip+1}{j}\frac{B_{j}}{n^{j}}}\right)^{y_{k,i}}}}\\\ \end{split}$ where $B_{j}$ are the Bernoulli numbers of the first kind. ###### Proof. This theorem is obtained by applying Theorem 4.1 and then applying Faulhaber’s formula. ∎ ###### Corollary 4.3. For $p=1$, Theorem 4.3 becomes $\begin{split}\sum_{N_{m}=1}^{n}{\cdots\sum_{N_{1}=1}^{N_{2}}{{N_{m}}\cdots{N_{1}}}}&=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\left(\sum_{N=1}^{n}{N^{i}}\right)^{y_{k,i}}}}\\\ &=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\left(\frac{n^{i+1}}{i+1}\sum_{j=0}^{i}{(-1)^{j}\binom{i+1}{j}\frac{B_{j}}{n^{j}}}\right)^{y_{k,i}}}}.\end{split}$ Where $B_{j}$ are the Bernoulli numbers of the first kind. Let us now consider a few particular cases: * • For $m=2$ $\begin{split}&\sum_{N_{2}=1}^{n}{\sum_{N_{1}=1}^{N_{2}}{{N_{2}}^{p}{N_{1}}^{p}}}\\\ &\,\,=\frac{1}{2}\left(\sum_{N=1}^{n}{{N}^{p}}\right)^{2}+\frac{1}{2}\left(\sum_{N=1}^{n}{{N}^{2p}}\right)\\\ &\,\,=\frac{1}{2}\left[\left(\frac{n^{p+1}}{p+1}\sum_{j=0}^{p}{(-1)^{j}\binom{p+1}{j}\frac{B_{j}}{n^{j}}}\right)^{2}+\left(\frac{n^{2p+1}}{2p+1}\sum_{j=0}^{2p}{(-1)^{j}\binom{2p+1}{j}\frac{B_{j}}{n^{j}}}\right)\right].\end{split}$ ###### Example 4.6. For $p=1$, by applying this theorem and exploiting Faulhaber’s formula, we can get the following formula $\sum_{N_{2}=1}^{n}{\sum_{N_{1}=1}^{N_{2}}{{N_{2}}{N_{1}}}}=\frac{n(n+1)(n+2)(3n+1)}{24}.$ ###### Example 4.7. For $p=2$, by applying this theorem and exploiting Faulhaber’s formula, we can get the following formula $\sum_{N_{2}=1}^{n}{\sum_{N_{1}=1}^{N_{2}}{{N_{2}}^{2}{N_{1}}^{2}}}=\frac{n(n+1)(n+2)(2n+1)(2n+3)(5n-1)}{360}.$ * • For $m=3$ $\sum_{N_{3}=1}^{n}{\sum_{N_{2}=1}^{N_{3}}{\sum_{N_{1}=1}^{N_{2}}{{N_{3}}^{p}{N_{2}}^{p}{N_{1}}^{p}}}}=\frac{1}{6}\left(\sum_{N=1}^{n}{a_{N}}\right)^{3}+\frac{1}{2}\left(\sum_{N=1}^{n}{a_{N}}\right)\left(\sum_{N=1}^{n}{\left(a_{N}\right)^{2}}\right)+\frac{1}{3}\left(\sum_{N=1}^{n}{\left(a_{N}\right)^{3}}\right).$ ###### Example 4.8. For $p=1$, by applying this theorem and exploiting Faulhaber’s formula, we can get the following formula $\sum_{N_{3}=1}^{n}{\sum_{N_{2}=1}^{N_{3}}{\sum_{N_{1}=1}^{N_{2}}{{N_{3}}{N_{2}}{N_{1}}}}}=\frac{n^{2}(n+1)^{2}(n+2)(n+3)}{48}=\left(\sum_{N=1}^{n}{N}\right)\left[\frac{n(n+1)(n+2)(n+3)}{4!}\right].$ #### 4.5.2 Recurrent p-series and harmonic series In this section, using the formula developed by Euler and the reduction theorem (Theorem 4.1), we will prove an expression which can be used to calculate a recurrent form of the zeta function for positive even values. Then we will conjecture a solution for a more general form of the Basel problem. We start by applying Theorem 4.1 and using the expression of the zeta function for positive even values to get an expression for the recurrent series of $\frac{1}{N^{2p}}$ (or recurrent harmonic series). ###### Theorem 4.4. For any $m,p\in\mathbb{N}$, we have that $\begin{split}\sum_{N_{m}=1}^{\infty}{\cdots\sum_{N_{1}=1}^{N_{2}}{\frac{1}{N_{m}^{2p}\cdots N_{1}^{2p}}}}&=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\left(\zeta(2ip)\right)^{y_{k,i}}}}\\\ &=(-1)^{pm}(2\pi)^{2pm}\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{(-1)^{y_{k,i}}}{(y_{k,i})!}\left(\frac{B_{2ip}}{(2i)(2ip)!}\right)^{y_{k,i}}}}.\end{split}$ ###### Proof. By applying Theorem 4.1, $\begin{split}\sum_{N_{m}=1}^{\infty}{\cdots\sum_{N_{1}=1}^{N_{2}}{\frac{1}{N_{m}^{2p}\cdots N_{1}^{2p}}}}&=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\left(\sum_{N=1}^{\infty}{\left(\frac{1}{N^{2p}}\right)^{i}}\right)^{y_{k,i}}}}\\\ &=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\left(\zeta(2ip)\right)^{y_{k,i}}}}.\end{split}$ Euler proved that, for $m\geq 1$ (see [2] for a proof), $\zeta(2m)=\frac{(-1)^{m+1}(2\pi)^{2m}}{2(2m)!}B_{2m}.$ Hence, $\begin{split}\sum_{N_{m}=1}^{\infty}{\cdots\sum_{N_{1}=1}^{N_{2}}{\frac{1}{N_{m}^{2p}\cdots N_{1}^{2p}}}}&=\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\left((-1)^{ip+1}\frac{B_{2ip}(2\pi)^{2ip}}{2(2ip)!}\right)^{y_{k,i}}}}\\\ &=(-1)^{pm}(2\pi)^{2pm}\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{(-1)^{y_{k,i}}}{(y_{k,i})!}\left(\frac{B_{2ip}}{(2i)(2ip)!}\right)^{y_{k,i}}}}.\end{split}$ ∎ The following table summarizes some values of the zeta function for positive even arguments, $\zeta(2)=\frac{\pi^{2}}{6}\,\,\,\,\,\,\zeta(4)=\frac{\pi^{4}}{90}\,\,\,\,\,\,\zeta(6)=\frac{\pi^{6}}{945}\,\,\,\,\,\,\zeta(8)=\frac{\pi^{8}}{9450}\,\,\,\,\,\,\zeta(10)=\frac{\pi^{10}}{93555}\,\,\,\,\,\,$ $\zeta(12)=\frac{691\pi^{12}}{638512875}\,\,\,\,\,\,\zeta(14)=\frac{2\pi^{14}}{18243225}\,\,\,\,\,\,\zeta(16)=\frac{3617\pi^{16}}{325641566250}\,\,\,\,\,\,.$ By using the values in the above table as well as Theorem 4.4 and playing with different values, we can notice some identities. In particular, we can conjecture the following statement for the recurrent sum of $\frac{1}{N^{2}}$ (recurrent harmonic series with $2p=2$) for different values of $m$ (for different numbers of summations). This represents a generalization of the Basel Problem solved by Euler. However, this conjecture has already been proven, hence, we will directly use it to develop additional identities. ###### Theorem 4.5. For any $m\in\mathbb{N}$, we have that $\sum_{N_{m}=1}^{\infty}{\cdots\sum_{N_{1}=1}^{N_{2}}{\frac{1}{N_{m}^{2}\cdots N_{1}^{2}}}}=\frac{(-1)^{m+1}2\left(2^{2m-1}-1\right)B_{2m}\pi^{2m}}{(2m)!}=\left(2-\frac{1}{2^{2(m-1)}}\right)\zeta(2m)$ or identically (from Theorem 4.1), $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{1}{(y_{k,i})!i^{y_{k,i}}}\left(\zeta(2i)\right)^{y_{k,i}}}}=\frac{(-1)^{m+1}2\left(2^{2m-1}-1\right)B_{2m}\pi^{2m}}{(2m)!}=\left(2-\frac{1}{2^{2(m-1)}}\right)\zeta(2m).$ ###### Proof. In [35], the following relation was proven but in another notation, $\sum_{1\leq N_{1}\leq\cdots\leq N_{m}}{\frac{1}{N_{m}^{2}\cdots N_{1}^{2}}}=\left(\frac{2^{2m-1}-1}{2^{2m-2}}\right)\zeta(2m).$ Hence, $\sum_{N_{m}=1}^{\infty}{\cdots\sum_{N_{1}=1}^{N_{2}}{\frac{1}{N_{m}^{2}\cdots N_{1}^{2}}}}=\left(\frac{2^{2m-1}-1}{2^{2m-2}}\right)\zeta(2m)=\left(2-\frac{1}{2^{2(m-1)}}\right)\zeta(2m).$ Euler proved that, for $m\geq 1$, $B_{2m}=\frac{(-1)^{m+1}2(2m)!}{(2\pi)^{2m}}\zeta(2m)\,\,\,\,\,\,\,\,or\,\,\,\,\,\,\,\,\zeta(2m)=\frac{(-1)^{m+1}(2\pi)^{2m}}{2(2m)!}B_{2m}.$ Hence, by substituting, we get $\sum_{N_{m}=1}^{\infty}{\cdots\sum_{N_{1}=1}^{N_{2}}{\frac{1}{N_{m}^{2}\cdots N_{1}^{2}}}}=\frac{(-1)^{m+1}2\left(2^{2m-1}-1\right)B_{2m}\pi^{2m}}{(2m)!}.$ The proof of the first equation is completed. Applying Theorem 4.1, we get the second equation. ∎ ###### Corollary 4.4. For any $m\in\mathbb{N}$, we have that $\sum_{\sum{i.y_{k,i}}=m}{\prod_{i=1}^{m}{\frac{(-1)^{y_{k,i}}}{(y_{k,i})!}\left(\frac{B_{2ip}}{(2i)(2ip)!}\right)^{y_{k,i}}}}=\left(\frac{1}{2^{2m-1}}-1\right)\frac{B_{2m}}{(2m)!}$ ###### Proof. By applying Theorem 4.4 with $p=1$ to Theorem 4.5, we obtain the theorem. ∎ We will use this to prove that this recurrent harmonic series (or recurrent $p$-series) with $2p=2$ will converge to 2 as the number of summations $m$ goes to infinity. ###### Theorem 4.6. For any $m\in\mathbb{N}$, we have that $\lim_{m\to\infty}{{\left(\sum_{N_{m}=1}^{\infty}{\cdots\sum_{N_{1}=1}^{N_{2}}{\frac{1}{N_{m}^{2}\cdots N_{1}^{2}}}}\right)}}=2.$ ###### Proof. It is known that $\lim_{m\to\infty}\zeta(2m)=1$. By applying Theorem 4.5, $\lim_{m\to\infty}{{\left(\sum_{N_{m}=1}^{\infty}{\cdots\sum_{N_{1}=1}^{N_{2}}{\frac{1}{N_{m}^{2}\cdots N_{1}^{2}}}}\right)}}=\lim_{m\to\infty}{\left(2-\frac{1}{2^{2(m-1)}}\right)}\times\lim_{m\to\infty}{\zeta(2m)}=2.$ ∎ ###### Example 4.9. For $m=4$, we have $\sum_{N_{4}=1}^{\infty}{\sum_{N_{3}=1}^{N_{4}}{\sum_{N_{2}=1}^{N_{3}}{\sum_{N_{1}=1}^{N_{2}}{\frac{1}{N_{4}^{2}N_{3}^{2}N_{2}^{2}N_{1}^{2}}}}}}=\frac{1}{24}\left(\sum_{N=1}^{\infty}{\frac{1}{N^{2}}}\right)^{4}+\frac{1}{4}\left(\sum_{N=1}^{\infty}{\frac{1}{N^{2}}}\right)^{2}\left(\sum_{N=1}^{\infty}{\frac{1}{N^{4}}}\right)+\frac{1}{3}\left(\sum_{N=1}^{\infty}{\frac{1}{N^{2}}}\right)\left(\sum_{N=1}^{\infty}{\frac{1}{N^{6}}}\right)+\frac{1}{8}\left(\sum_{N=1}^{\infty}{\frac{1}{N^{4}}}\right)^{2}+\frac{1}{4}\left(\sum_{N=1}^{\infty}{\frac{1}{N^{8}}}\right)=\frac{127\pi^{8}}{604800}{=\left(2-\frac{1}{2^{2(3)}}\right)\zeta(8)\approx 1.992466004.}$ Similarly, we will use this to show that the sum (over all non-negative values of $m$) of the recurrent harmonic series with $2p=2$ will diverge. ###### Theorem 4.7. We have that, $\sum_{m=0}^{\infty}{\left(\sum_{N_{m}=1}^{\infty}{\cdots\sum_{N_{1}=1}^{N_{2}}{\frac{1}{N_{m}^{2}\cdots N_{1}^{2}}}}\right)}\to\infty.$ ###### Proof. Applying Theorem 4.5, $\sum_{m=0}^{\infty}{\left(\sum_{N_{m}=1}^{\infty}{\cdots\sum_{N_{1}=1}^{N_{2}}{\frac{1}{N_{m}^{2}\cdots N_{1}^{2}}}}\right)}=\sum_{m=0}^{\infty}{\left(2-\frac{1}{2^{2(m-1)}}\right)\zeta(2m)}.$ For $m=0$, we have $\left(2-\frac{1}{2^{2(m-1)}}\right)\zeta(2m)=(2-4)(-1/2)=1$. Knowing that $\zeta(2m)\geq 1$ for $m\geq 1$ and noticing the following identity for $m\geq 1$, $1\leq\left(2-\frac{1}{2^{2(m-1)}}\right)\leq 2.$ Hence, for $m\geq 0$, $\left(2-\frac{1}{2^{2(m-1)}}\right)\zeta(2m)\geq 1.$ Thus, $\lim_{n\to\infty}{\sum_{m=0}^{n}{\left(\sum_{N_{m}=1}^{\infty}{\cdots\sum_{N_{1}=1}^{N_{2}}{\frac{1}{N_{m}^{2}\cdots N_{1}^{2}}}}\right)}}\geq\lim_{n\to\infty}{\sum_{m=0}^{n}{1}}=\infty.$ Hence, this sums is infinite. ∎ ## References * Andrews [1998] Andrews, G. E. (1998). The theory of partitions. Number 2. Cambridge university press. * Arfken and Weber [1999] Arfken, G. B. and Weber, H. J. (1999). Mathematical methods for physicists. * Bernoulli [1689] Bernoulli, J. (1689). Propositiones arithmeticae de seriebus infinitis earumque summa finita [arithmetical propositions about infinite series and their finite sums]. basel: J. conrad. * Bernoulli [1713] Bernoulli, J. (1713). Ars Conjectandi, Opus Posthumum; Accedit Tractatus De Seriebus Infinitis, Et Epistola Gallicè scripta De Ludo Pilae Reticularis [Theory of inference, posthumous work. With the Treatise on infinite series…]. Thurnisii. * Bernoulli [1742] Bernoulli, J. (1742). ”corollary iii of de seriebus varia”. opera omnia. lausanne & basel: Marc-michel bousquet & co. 4:8. * Blümlein et al. [2010] Blümlein, J., Broadhurst, D., and Vermaseren, J. A. (2010). The multiple zeta value data mine. Computer Physics Communications, 181(3):582–625. * Blümlein and Kurth [1999] Blümlein, J. and Kurth, S. (1999). Harmonic sums and mellin transforms up to two-loop order. Physical Review D, 60(1):014018. * Broadhurst [1986] Broadhurst, D. (1986). Exploiting the 1, 440-fold symmetry of the master two-loop diagram. Zeitschrift für Physik C Particles and Fields, 32(2):249–253. * Broadhurst [2013] Broadhurst, D. (2013). Multiple zeta values and modular forms in quantum field theory. In Computer algebra in quantum field theory, pages 33–73. Springer. * Bruinier and Ono [2013] Bruinier, J. H. and Ono, K. (2013). Algebraic formulas for the coefficients of half-integral weight harmonic weak maass forms. Advances in Mathematics, 246:198–219. * Chapman [1999] Chapman, R. (1999). Evaluating $\zeta$ (2). Preprint. * [12] Euler, L. De summis serierum reciprocarum, commentarii academiae scientiarum petropolitanae 7 (1740), 123–134. Opera Omnia, Series, 1:73–86. * Euler [1743] Euler, L. (1743). Demonstration de la somme de cette suite 1+ 1/4+ 1/9+ 1/16+… Journal litteraire d’Allemagne, de Suisse et du Nord, pages 115–127. * Euler [1776] Euler, L. (1776). Meditationes circa singulare serierum genus. Novi commentarii academiae scientiarum Petropolitanae, pages 140–186. [Reprinted in “Opera Omnia,” Ser. I, Vol. 15, pp. 217-267, Teubner, Berlin, 1927]. * Euler [1811] Euler, L. (1811). De summatione serierum in hac forma contentarum $a/1+a^{2}/4+a^{3}/9+a^{4}/16+a^{5}/25+a^{6}/36+$ etc. Memoires de l’academie des sciences de St.-Petersbourg, pages 26–42. * Euler [1988] Euler, L. (1988). Introduction to analysis of the infinite: Book i, translation of introductio in analysin infinitorum (1748) to english from the original latin by j. d. blanton. * [17] Faulhaber, J. Academia algebrae. Darinnen die miraculosische Inventiones zu den höchsten weiters continuirt und profitiert werden, call number QA154, 8:F3. * Granville [1997] Granville, A. (1997). A decomposition of riemann’s zeta-function. London Mathematical Society Lecture Note Series, pages 95–102. * Haddad [a] Haddad, R. E. Explicit formula for the integral of order $n$ of $x^{m}(\ln x)^{m^{\prime}}$. unpublished. * Haddad [b] Haddad, R. E. Multiple sums and partition identities. unpublished. * Haddad [c] Haddad, R. E. Repeated sums and binomial coefficients. unpublished. * Hardy and Ramanujan [1918] Hardy, G. H. and Ramanujan, S. (1918). Asymptotic formulaae in combinatory analysis. Proceedings of the London Mathematical Society, s2-17(1):75–115. * Hoffman [1992] Hoffman, M. (1992). Multiple harmonic series. Pacific Journal of Mathematics, 152(2):275–290. * Hoffman [1997] Hoffman, M. E. (1997). The algebra of multiple harmonic series. Journal of Algebra, 194(2):477 – 495. * Hoffman and Moen [1996] Hoffman, M. E. and Moen, C. (1996). Sums of triple harmonic series. journal of number theory, 60(2):329–331. * Kassel [2012] Kassel, C. (2012). Quantum groups, volume 155. Springer Science & Business Media. * Kuba and Panholzer [2019] Kuba, M. and Panholzer, A. (2019). A note on harmonic number identities, stirling series and multiple zeta values. International Journal of Number Theory, 15(07):1323–1348. * Loeb [1992] Loeb, D. E. (1992). A generalization of the stirling numbers. Discrete mathematics, 103(3):259–269. * Mengoli [1650] Mengoli, P. (1650). ”praefatio [preface]”. novae quadraturae arithmeticae, seu de additione fractionum [new arithmetic quadrature (i.e., integration), or on the addition of fractions]. bologna: Giacomo monti. * Murahara and Ono [2019] Murahara, H. and Ono, M. (2019). Interpolation of finite multiple zeta and zeta-star values. arXiv preprint arXiv:1908.09307. * Nielsen [1923] Nielsen, N. (1923). Traité élémentaire des nombres de Bernoulli. Gauthier-Villars. * Oresme [1961] Oresme, N. (1961). Quaestiones super geometriam Euclidis, volume 3. Brill Archive. * Propp [1989] Propp, J. (1989). Some variants of ferrers diagrams. Journal of Combinatorial Theory, Series A, 52(1):98–128. * Rademacher [1938] Rademacher, H. (1938). On the partition function p(n). Proceedings of the London Mathematical Society, s2-43(1):241–254. * Schneider [2016] Schneider, R. (2016). Partition zeta functions. Research in Number Theory, 2(1):9. * Wang and Wang [2009] Wang, W. and Wang, T. (2009). General identities on bell polynomials. Computers & Mathematics with Applications, 58(1):104–118. * [37] Xu, C. Duality formulas for arakawa-kaneko zeta values and related variants. * Zagier [1994] Zagier, D. (1994). Values of zeta functions and their applications. In First European Congress of Mathematics Paris, July 6–10, 1992, pages 497–512. Springer.
15mm15mm15mm15mm # Thermally driven fission of protocells Romain Attal Cité des Sciences et de l’Industrie 30, avenue Corentin-Cariou 75019 Paris, France<EMAIL_ADDRESS> ###### Abstract. We propose a simple mechanism for the self-replication of protocells. Our main hypothesis is that the amphiphilic molecules composing the membrane bilayer are synthesised inside the protocell through globally exothermic chemical reactions. The slow increase of the inner temperature forces the hottest molecules to move from the inner leaflet to the outer leaflet of the bilayer. This asymmetric translocation process makes the outer leaflet grow faster than the inner leaflet. This differential growth increases the mean curvature and amplifies any local shrinking of the protocell until it splits in two. ###### Key words and phrases: protocell, bilayer, translocation, self-replication, thermodynamical instability. ###### Contents 1. 1 Protocells and metabolism 2. 2 Hypotheses of our model 3. 3 Flows, forces and energy dissipation 1. 3.1 Main irreversible processes 2. 3.2 Flows associated to each irreversible process 3. 3.3 Thermodynamical forces 4. 3.4 Conductance matrix 5. 3.5 Entropy production and stability 4. 4 Membrane geometry and growth equation 1. 4.1 Conservation of matter and exponential growth 2. 4.2 Cylindrical growth in steady state 5. 5 Thermal instability of cylindrical growth 1. 5.1 The Squeezed Sausage Theorem (SST) 2. 5.2 Fluctuations, translocation and heat transfer 3. 5.3 Thermal balance 6. 6 Translocation between leaflets 1. 6.1 An effective potential for translocation 2. 6.2 Computation of $L_{m\theta}$ 3. 6.3 Computation of $L_{mm}$ 4. 6.4 Estimation of $L_{\theta\theta}$ 5. 6.5 Destabilisation 7. 7 Conclusions and perspectives 8. A The mean curvature of the membrane 9. B Solutions of the growth equation 1. B.1 Case 1 : $\eta\neq 1$ and $\tau>\tau_{+}(\eta)$ or $\tau<\tau_{-}(\eta)$ 2. B.2 Case 2 : $\eta\neq 1$ and $\tau\in\\{\tau_{+}(\eta),\tau_{-}(\eta)\\}$ 3. B.3 Case 3 : $\tau_{-}(\eta)<\tau<\tau_{+}(\eta)$ 10. C Smooth perturbation of cylindrical growth 1. C.1 Isovolumic variation of the area 2. C.2 Isovolumic variation of the total mean curvature 11. D Asymptotic expansion of $F(a)$ ## 1\. Protocells and metabolism The objects modeled in the present article are protocells, the putative ancestors of modern living cells [23, 34]. In the absence of fossils [38], we ignore their detailed properties. However, we can sketch a minimalist functional diagram of protocells (FIG. 1). FoodWasteMetabolismHeatBiomass Protocells initiate the fundamental process of life : Food $\to$ Biomass + Heat + Waste. The protocell is a vesicle bounded by a bilayer made of amphiphilic molecules. Nutrient molecules (food) enter by mere diffusion, since they are consumed inside, where their concentration is lower than outside. Conversely, waste molecules have a larger concentration inside and therefore diffuse passively to the outside. The metabolism is a network of unknown chemical reactions taking place only inside the protocell. The net reaction is supposed to be exothermic, since living matter is hotter than abiotic matter (under the same external thermodynamical conditions). Let us compare this scheme to actual evolved cells. The growth of bacteria in a nutrient rich medium follows a species dependent periodic process [5, 12]. At regular time intervals, each cell splits to form two daughter cells. This requires the synchronization of numerous biochemical and mechanical processes inside the cell, involving cytoskeletal structures positioned at the locus of the future cut (septum). However, in the history of life, such complex structures are a high-tech luxury and must have appeared much later than the ability to split. Protocells must have used a simple splitting mechanism to ensure their reproduction, before the appearance of genes, RNA, enzymes and all the complex organelles present today even in the most rudimentary forms of autonomous life [34]. In this article, we present a simple model for the growth and self-replication of a protocell, following the laws of irreversible thermodynamics near equilibrium. Our guide is the rate of entropy production, which is minimal in a steady state [32, 20]. A key point of our approach is that the heat produced by the metabolism of the protocell is approximately proportional to its volume, whereas the heat flow that it can loose is proportional to the area of its membrane. In a rod-shaped cell (bacillus) growing linearly, these two quantities are approximately proportional so that each increment of the membrane area should be sufficient, ideally, to evacuate the heat produced by the corresponding increment of the cell volume. However, the irreversible physical and chemical processes produce heat more quickly than the growing tubular membrane can dissipate to the outside. This increases slowly the inner temperature and enhances the fluctuations of the shape of the membrane, of the various concentrations and of the local electric field. In a growing spherical protocell, the maximal heat flow that the membrane can expell to the outside without overheating the inside puts an upper limit to the radius of the protocell. Indeed, the formation of two small protocells from a big one releases work [33], so that large protocells are mechanically unstable. However, neither [33] nor [8] provides a path to follow to realise this deformation. In our model, we start from a cylindrical shape to simplify the computations. As the inner temperature increases, the growth of the outer leaflet of the membrane becomes more probable than the growth of the inner leaflet. If a random thermal fluctuation lowers slightly the radius of this cylinder, then its area increases more quickly than during the steady state cylindrical growth (FIG. 1). $\downarrow$$\downarrow$$\downarrow$hotcold Splitting a cylindrical protocell. This reduction of the radius induces a loss of convexity of the membrane. This favors the outflow of heat and the ratio area/volume increases slightly, compared to a convex cylindrical shape. The plan of the article goes as follows. In Section II, in order to formulate these ideas mathematically, we state all the physical hypotheses of our model of protocells. In Section III, we define the various flows of matter and energy and their associated thermodynamical forces. In the linear approximation, the rate of entropy production is the scalar product of these flows and forces and is minimal in a steady state [32]. In Section IV, we derive a differential equation for the evolution of the area and the integral of the mean curvature of the membrane, starting from the advancement of the chemical reaction for the synthesis of the membrane molecules. This linear equation admits a solution growing exponentially. In Section V, we use variational calculus [10] to prove that the local reduction of the radius of the cell increases its length and its area, if its volume is kept constant. This intuitive property implies that heat is more easily released when the protocell is squeezed. In Section VI, we propose a molecular mechanism for the increase of the mean curvature of the membrane associated to this squeezing. The position of each amphiphilic membrane molecule is reduced to a single degree of freedom : the distance from the polar head to the middle of the hydrophobic slice. We use a double well effective potential to describe the trapping of these molecules in the membrane. Due to the temperature difference between the inner and outer sides, the membrane molecules go from the inner leaflet to the outer leaflet more often than in the opposite direction. This asymmetry forces the membrane to curve and shrink around the middle of the protocell and initiates its splitting. Our main mathematical result (Proposition VI.I) states that a stability condition, $L_{m\theta}^{2}<L_{mm}L_{\theta\theta}$, can not be satisfied at high temperature, because the squared crossed conductance, $L_{m\theta}^{2}$, increases more quickly than the product of the diffusion coefficients, $L_{mm}$ for membrane molecules and $L_{\theta\theta}$ for heat. Hence, the cylindrical growth process is unstable when the temperature difference is sufficiently high. We conclude in Section VII with a proposition of an experimental test for our model. The appendices contain the detailed computations of our model. The mathematical notions involved are elementary (linear differential equations and geometry of surfaces). ## 2\. Hypotheses of our model Let us state more precisely the hypotheses underlying our model : 1. (1) Our protocells are made of a membrane of average thickness $2\varepsilon$, bounding a cytosol of finite volume $\mathcal{V}(t)$. 2. (2) The cytosol contains unknown specific molecules (reactants, catalysers, chromophores, …) which participate to a network of chemical reactions. We suppose that the concentrations are constant and uniform in the volume $\mathcal{V}$. 3. (3) The protocell starts with a cylindrical shape closed by two hemispherical caps of fixed radius, $R_{0}$. The total length, $\ell(t)+2R_{0}+2\varepsilon$, increases with time due to the synthesis of membrane molecules (FIG. 3). $\ell(t)+2R_{0}+2\varepsilon$$2\varepsilon$${2(R_{0}-\varepsilon)}$${2(R_{0}+\varepsilon)}$$\ell(t)$ Geometry of an idealised cylindrical protocell. This may seem a rather drastic hypothesis, but the computations could be made for a generic, approximately spherical shape using an expansion in spherical harmonics. This would add to the model an unnecessary mathematical complexity that would hide the main physical phenomena. The use of cylindrical, rotation invariant shapes allows us to reduce the problem to one dimension. Moreover, this is a best case scenario for the release of heat in steady state, since the ratio volume/area can be held constant in a steady growth. 4. (4) Due to the surface tension of the membrane, its mean curvature has an upper bound, $H_{\max}=\mfrac{1}{R_{0}}$. Indeed, due to the attractive forces between the polar heads of the membrane molecules, and due to their geometry, they can not form structures arbitrarily small [27]. 5. (5) Food (nutrients and water) enters the protocell by mere passive diffusion through the membrane. Waste and heat also diffuse passively but in the opposite direction. Protocells did not use specialized membrane molecules for an active transport through the membrane. 6. (6) The membrane molecules are synthesised inside the protocell in an unknown network of chemical reactions. It might use some encapsulated catalyzers or chromophores trapped in the volume and catching part of the ambient light [23], but we will make no hypothesis on the details of this network. 7. (7) These metabolic reactions generate heat to be evacuated and increase slowly the internal temperature, $T_{1}$, whereas the external temperature, $T_{0}$, remains fixed. 8. (8) The characteristic time of the variations of $T_{1}(t)$ is much larger than the characteristic times of chemical reactions and diffusion processes across the membrane. 9. (9) The cytosol is homogenous and contains no organelles, no cytoskeleton, no enzymes, no RNA/DNA. Just simple chemical reactants uniformly distributed. (Rashevsky’s model [33] allows for a slight radial variation of concentrations due to the diffusion of food and waste through the membrane). 10. (10) The membrane is a bilayer made of unspecified amphiphilic molecules. We presume that their hydrophobic tails are long enough (10-12 carbon atoms) to form a stable bilayer, but not too bulky in order to allow flip-flop (or translocation) processes between the two leaflets. We do not include sterol molecules because they are the product of a long biochemical selection process [27], and a high-tech luxury for protocells. 11. (11) The inner leaflet (L1) is at temperature $T_{1}$ whereas the outer leaflet (L0) is at temperature $T_{0}<T_{1}$. This temperature drop allows the bilayer to undergo coupled transport phenomena (food and waste diffusion, including water leaks, heat diffusion, flip-flop, etc.). 12. (12) The membrane may contain other molecules, in small concentrations, but we don’t need them to transport food, waste or any molecule through the membrane. The validity of these hypotheses will depend on the agreement of their predictions with the results of future experiments made with real protocells. ## 3\. Flows, forces and energy dissipation In any living system, some processes release energy whereas other processes consume energy. Globally, the system takes usable energy from the outside and rejects unusable energy, in the form of heat and waste, that can be used by other living systems. In order to describe such a system, we must define the various flows of matter and energy and the forces causing these flows. Any gradient of concentration, pressure, temperature, etc. will cause a current of particles, fluid, heat, etc. These processes are generally irreversible and dissipate energy to inaccessible degrees of freedom. This dissipation of a conserved quantity is measured by the entropy function, which increases as time passes. The study of irreversible thermodynamical processes near equilibrium [29, 30, 35, 20] is based on the rate of entropy production, represented by a bilinear function of flows (chemical reaction speed, thermal current, particle current, electric current, etc.) and forces (chemical affinity, temperature gradient, concentration gradient, electric tension, etc.). In a first approximation, flows and forces are related linearly, as in Ohm’s law : (1) $\displaystyle\text{electric current }=\text{ conductivity }\times\text{ electric field}$ and the power dissipated is a quadratic function of the tension : (2) power dissipated $\displaystyle=\text{ tension }\times\text{ current}$ $\displaystyle=\text{ conductance }\times\text{ tension}^{2}.$ Similarly, in viscous fluids : (3) power dissipated $\displaystyle=\text{ friction coefficient }\times\text{ velocity}^{2}.$ We suppose that the protocell metabolism is in a steady state not too far from equilibrium, so that the various flows, $J_{i}$, and the thermodynamic forces, $X_{k}$, are linearly related : (4) $\displaystyle J_{i}=\sum_{k}L_{ik}X_{k}$ and the entropy rate is a quadratic function of $X$ : (5) $\displaystyle\sigma:=XJ=\sum_{ik}X_{i}L_{ik}X_{k}.$ The coefficients $L_{ik}$ are called phenomenological because their computation depends on the chosen model of microscopic dynamics (kinetic theory) and their numerical value has to be compared to a measurement in the real world to (in)validate this model and the linearity hypothesis. An important property of the phenomenological coefficients is provided by Onsager’s relations [29, 30, 20, 35]. Under the hypotheses of microscopic reversibility and parity of the variables under time reversal (in particular, in the absence of magnetic coupling and vorticity), the matrix $L$ is symmetric : (6) $\displaystyle L_{ik}=L_{ki}.$ This important law has been checked experimentally for various systems near equilibrium and is satisfied quite accurately in many cases. ### 3.1. Main irreversible processes To each irreversible physical or chemical process are associated a flow of matter or energy and a thermodynamical force, just as an electric current and an electric tension correspond to each branch of an electric network. If we identify the main processes that take place during the growth of a protocell, we can compute the global rate of dissipation of energy, or entropy creation. According to Prigogine’s Theorem [32, 11], this rate reaches a minimum when the system is in a steady state. In order to compute this dissipation, we need to define the various compartiments containing energy. In the sequel of this article, the subscript $0$ (resp. $1$) will denote the variables outside (resp. inside) the protocell. The physical and chemical processes are grouped as follows : $f_{0}\to f_{1}\,$ : food molecules (nutrients $+$ water) diffuse into the protocell through the membrane. $f\to m+c+w\,$ : food is transformed into membrane, cytosol and waste, inside the protocell. This is a global process, a superposition of catabolism and anabolism. Taking into account the stoichiometric coefficients, we can write more precisely : (7) $\displaystyle\sum_{i}\nu_{f_{i}}f_{i}\ \longrightarrow\ \sum_{j}\nu_{m_{j}}m_{j}+\sum_{k}\nu_{c_{k}}c_{k}+\sum_{l}\nu_{w_{l}}w_{l}$ where $f_{i}$ denotes the food molecules of type $i$, $m_{j}$ the membrane molecules of type $j$, $c_{k}$ the cytosol molecules of type $k$ and $w_{l}$ the waste molecules of type $l$. If $N_{\alpha}$ is the number of molecules of type $\alpha$, the advancement of this reaction, $\xi$, is defined by : (8) $\displaystyle\mathrm{d}\xi:=\frac{\mathrm{d}N_{\alpha}}{\pm\nu_{\alpha}}$ where the stoichiometric coefficients, $\nu_{\alpha}$, are counted positively for the products and negatively for the reactants. Note that our definition of $\xi$ involves $N_{\alpha}$ instead of the volumic concentration, $C_{\alpha}=N_{\alpha}/\mathcal{V}$, because the volume is not fixed. $w_{1}\to w_{0}\,$ : waste molecules diffuse out of the protocell through the membrane. $m_{1}\leftrightarrows m_{0}\,$ : molecules of the membrane bilayer go from one side to the other. In modern cells, this process is catalysed by enzymes (flippase for $0\to 1$ and floppase for $1\to 0$), but in protocells such a complex machinery did not exist yet [34]. If we suppose that the first membranes were not as thick as today (most phospholipids in modern and healthy cell walls have hydrophobic chains made of $\sim$ 16-22 atoms of carbon [27]), the exchange of molecules between the two leaflets could have been possible in a reasonable time to allow spontaneous splitting. Medium length lipids (10-14 atoms of carbon) could be good candidates to make stable, flippable and not too porous protocells. We isolate the process of translocation ($\restrictwand\xleftrightharpoons{}\restrictwandup$) because the ratio ${N_{m0}}/{N_{m1}}$ of the numbers of membrane molecules on each side is related to the mean curvature of the bilayer, which is the geometric parameter monitoring the splitting process. $q_{1}\to q_{0}\,$ : electric charges can be transfered from one side of the membrane to the other, by an ionic bound on the polar head of the membrane molecules. This electric current builds up an electric tension, $U_{01}$, counteracted by possible ionic leaks through the membrane. If we suppose that the membrane molecules are monovalent fatty acids, each one can carry a monocation (H+, Na+, K+, …). This cotransport process could be the ancestor of the modern sodium-potassium pump. Anions also can participate to this transmembrane electric current, by leaking throuh water pores [14]. ### 3.2. Flows associated to each irreversible process The main processes of our model are described by the following flows in the protocell (see FIG. 3.2) : $J_{f}\,$ : the flow of food entering the protocell through its membrane (molecules per unit time per unit area). $J_{w}\,$ : the flow of waste exiting the protocell through its membrane (molecules per unit time per unit area). $J_{\theta}\,$ : the heat flow exiting the protocell by diffusion through its membrane (energy per unit time per unit area). $J_{mab}\,$ : the flow of membrane molecules from $a$ to $b$ (molecules per unit time per unit area). The possible values of $a$ and $b$ are : $c\,$ : the cytosol ; $1\,$ : the inner leaflet of the membrane (L1) ; $0\,$ : the outer leaflet of the membrane (L0) ; The net flow of membrane molecules is usually unidirectional, ${\mathbf{C}}\to{\mathbf{L}}_{1}\to{\mathbf{L}}_{0}$, hence $J_{mc1}>0$ and $J_{m}:=J_{m10}-J_{m01}>0$. $J_{r}\,$ : the speed of the synthesis reaction inside the cytosol (molecules per unit time per unit volume). $\xi$ being the advancement of the reaction $f\to m+c+w$, defined above, then $J_{r}$ is the time derivative of $\xi$ : (9) $\displaystyle J_{r}:=\frac{\mathrm{d}\xi}{\mathrm{d}t}.$ $J_{q}\,$ : some ions can be transported from one side to the other, bounded to the polar head of the membrane molecules. L0L1$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwand$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$\restrictwandup$$J_{m}$$J_{q}$ions$J_{f}$${\it{f}=}$food$f\,$$\xlongrightarrow{J_{r}}\ m+c+w$$J_{mc1}$${\it{w}=}$waste$J_{w}$$J_{\theta}$ heat diffusion$\textbf{E}=\text{environment}$$T_{0}$$\boxed{\textbf{C}=\text{cytosol}}$$T_{1}(t)$ Main flows of energy and matter in our model. We then have the following linear flow diagram for the synthesis and motion of membrane molecules : (10) $\displaystyle{\mathbf{E}}\xlongrightarrow{J_{f}}{\mathbf{C}}\xlongrightarrow{J_{r}}{\mathbf{C}}\xlongrightarrow{\color[rgb]{.1,.6,1}J_{mc1}}{\mathbf{L}}_{1}\xlongrightarrow{\color[rgb]{.1,.6,1}J_{m10}}{\mathbf{L}}_{0}.$ This picture is however slightly misleading. Indeed, the amphiphilic molecules being in a liquid phase, their positions fluctuate in each leaflet (transversal diffusion) and they undergo perpendicular motions (protrusion) and translocations from one leaflet to the other. The pictures obtained by molecular dynamics simulations [14, 15, 4, 1] give us a more precise representation of real world membranes. ### 3.3. Thermodynamical forces The thermodynamical forces associated to these processes are defined as follows : $X_{\theta}\,$ : the thermal force is the difference of the inverse temperatures inside and outside the protocell : (11) $\displaystyle X_{\theta}:=\frac{1}{T_{0}}-\frac{1}{T_{1}}>0.$ $X_{f_{i}}\,$ : the chemical force driving the food molecules, $f_{i}$, is the difference of the ratios $-\mu_{f_{i}}/T$ outside and inside the protocell : (12) $\displaystyle X_{f_{i}}:=\frac{\mu_{f_{i}0}}{T_{0}}-\frac{\mu_{f_{i}1}}{T_{1}}.$ The influx of food is guided by mere diffusion through the membrane (dedicated channel and intrinsic proteins did not exist yet in protocells). Since food is consumed inside the protocell, $[f_{i}]_{1}<[f_{i}]_{0}$. For a spherical protocell, the profile of the concentration of each molecule (as a function of the distance to the center) can be computed by solving the diffusion equation [33]. An important result of this computation is the existence of a discontinuity in the concentration of each molecule, $f_{i}$, proportional to the radius, $R$, of the protocell, to the rate of the reaction, $q_{i}$ (concentration/time), and inversely proportional to the permeability, $h_{i}$ (length/time), of the membrane for this molecule : $[f_{i}]_{1}-[f_{i}]_{0}\propto\mfrac{q_{i}R}{h_{i}}$. $X_{w_{j}}\,$ : the force driving the waste molecules to the outside of the protocell is the difference of chemical potentials divided by the temperature : (13) $\displaystyle X_{w_{j}}:=\frac{\mu_{w_{j}0}}{T_{0}}-\frac{\mu_{w_{j}1}}{T_{1}}.$ Note that $X_{w_{j}}$ and $X_{f_{i}}$ must have different signs for waste and food to go in opposite directions. $X_{r}\,$ : the chemical force driving the synthesis reactions (metabolism) is the chemical reaction affinity, $A_{r}$, of the global process $(f\to m+c+w)$, divided by the inner temperature of the protocell : (14) $\displaystyle X_{r}:=\frac{A_{r}}{T_{1}}.$ This affinity is a linear combination of the chemical potentials of the synthesis equation, weighted by the stoichiometric coefficients, counted positively for the reactants $(f)$ and negatively for the products $(m,c,w)$ : (15) $\displaystyle A_{r}=\sum_{i}\nu_{f_{i}}\mu_{f_{i}}-\sum_{j}\nu_{m_{j}}\mu_{m_{j}}-\sum_{k}\nu_{c_{k}}\mu_{c_{k}}-\sum_{l}\nu_{w_{l}}\mu_{w_{l}}.$ $X_{m^{\prime}}\,$ : The membrane molecules are synthesised in the cytosol at temperature $T_{1}$. Their hydrophobic tail enforces the spontaneous organisation of these molecules into a bilayer. We suppose that the temperature varies only across the membrane. The driving force of this isothermal process is the affinity of the reaction $m_{c}\to m_{1}$, divided by the inner temperature, $T_{1}$ : (16) $\displaystyle X_{m^{\prime}}=\frac{A_{mc1}}{T_{1}}=\frac{\mu_{mc}-\mu_{m1}}{T_{1}}.$ Here, $\mu_{mc}$ is the chemical potential of the free membrane molecules inside the cytosol and $\mu_{m1}$ is their chemical potential in the inner leaflet. The heat released to the inner leaflet during this process is : (17) $\displaystyle Q_{mc1}=\mu_{mc}-\mu_{m1}=T_{1}X_{m^{\prime}}.$ $X_{m}\,$ : the membrane molecules are transfered from the inner layer, at temperature $T_{1}$, to the outer leaflet, at temperature $T_{0}<T_{1}$, releasing the heat $Q_{m10}$ into the environmental thermostat, at temperature $T_{0}$. The thermodynamical force of this process is : (18) $\displaystyle X_{m}=\frac{\mu_{m1}}{T_{1}}-\frac{\mu_{m0}}{T_{0}}.$ $X_{q}\,$ : the thermodynamical force driving the ions of species $i$, of charge $z_{i}e$, across the membrane is the difference of electrochemical potentials [2] : (19) $\displaystyle X_{qi}$ $\displaystyle=\tilde{\mu}_{i1}-\tilde{\mu}_{i0}$ $\displaystyle=\big{(}\mu_{i1}+z_{i}e\psi_{1}\big{)}-\big{(}\mu_{i0}+z_{i}e\psi_{0}\big{)}$ $\displaystyle=\mu_{i1}-\mu_{i0}+z_{i}eU_{10}.$ where $\psi$ denotes the electrostatic potential and $U_{10}:=\psi_{1}-\psi_{0}$ is the electric tension across the membrane. Among these forces, only $X_{\theta}$ is a linear function of the small temperature difference, $\Delta T=T_{1}-T_{0}$. The others have, generically, a supplementary constant term, of order $0$ in $\Delta T$. ### 3.4. Conductance matrix The phenomenological coefficients, $L_{ik}$, which couple all the irreversible processes of our linear model, can be put in a $7\times 7$ matrix : (20) $\displaystyle L=\begin{pmatrix}L_{\theta\theta}&L_{\theta f}&L_{\theta w}&L_{\theta m}&L_{\theta m^{\prime}}&L_{\theta q}&L_{\theta r}\\\ L_{f\theta}&L_{ff}&L_{fw}&L_{fm}&L_{fm^{\prime}}&L_{fq}&L_{fr}\\\ L_{w\theta}&L_{wf}&L_{ww}&L_{wm}&L_{wm^{\prime}}&L_{wq}&L_{wr}\\\ L_{m\theta}&L_{mf}&L_{mw}&L_{mm}&L_{mm^{\prime}}&L_{mq}&L_{mr}\\\ L_{m^{\prime}\theta}&L_{m^{\prime}f}&L_{m^{\prime}w}&L_{m^{\prime}m}&L_{m^{\prime}m^{\prime}}&L_{m^{\prime}q}&L_{m^{\prime}r}\\\ L_{q\theta}&L_{qf}&L_{qw}&L_{qm}&L_{qm^{\prime}}&L_{qq}&L_{qr}\\\ L_{r\theta}&L_{rf}&L_{rw}&L_{rm}&L_{rm^{\prime}}&L_{rq}&L_{rr}\\\ \end{pmatrix}.$ In a first approximation, some coefficients can be set equal to zero : (21) $\displaystyle L\simeq\begin{pmatrix}L_{\theta\theta}&L_{\theta f}&L_{\theta w}&L_{\theta m}&0&L_{\theta q}&0&\\\ L_{f\theta}&L_{ff}&0&0&0&0&0\\\ L_{w\theta}&0&L_{ww}&0&0&0&0\\\ L_{m\theta}&0&0&L_{mm}&0&L_{mq}&0\\\ 0&0&0&0&L_{m^{\prime}m^{\prime}}&0&0\\\ L_{q\theta}&0&0&L_{qm}&0&L_{qq}&0\\\ 0&0&0&0&0&0&L_{rr}\\\ \end{pmatrix}.$ The diagonal coefficients of $L$ are positive but we let $L_{\bullet r}=0=L_{r\bullet}$ because the synthesis reactions take place in the cytosol and are decoupled from the transport processes across the membrane. Similarly, we let $L_{\bullet m^{\prime}}=0=L_{m^{\prime}\bullet}$, because the transfer of membrane molecules from the cytosol to the inner leaflet is decoupled from the other processes. Since the diffusion processes of different molecules (food, waste, ions or membrane constituents) across the membrane are supposed to be decoupled, we put $L_{fw}=0=L_{wf}$, $L_{fm}=0=L_{mf}$ and $L_{wm}=0=L_{mw}$. $L_{\theta\theta}$ is the thermal diffusion coefficient across the membrane. $L_{m^{\prime}m^{\prime}}$ is the diffusion coefficient for the transport of membrane molecules from the cytosol to the inner leaflet of the membrane. $L_{ff}$, $L_{ww}$, and $L_{mm}$, are the conductance coefficients of food, waste and membrane molecules through the membrane. We suppose that all these diagonal coefficients are constant and uniform across the cytosol or the membrane, because protocells could not rely on local specialised channel molecules (intrinsic proteins, in evolved cells) to supply their food and evacuate their waste. We also suppose that food and waste molecules are electrically neutral and that the electric current is entirely due to the transport of small ions with the help of the translocation process and water pores. The off-diagonal coefficients, $L_{f\theta}=L_{\theta f}$, $L_{w\theta}=L_{\theta w}$, $L_{m\theta}=L_{\theta m}$ and $L_{q\theta}=L_{\theta q}$, depend on the heat capacity of the molecules transported and on the rate constants of this transport. They couple the transport of matter and the heat flow. For our purpose, the most interesting off-diagonal coefficient is $L_{\theta m}$. It can be viewed as the ratio of heat flow, $J_{\theta}$, to the affinity $X_{m}$ when $T_{0}=T_{1}$ and in the absence of food and waste driving forces : (22) $\displaystyle L_{\theta m}=\left(\frac{J_{\theta}}{X_{m}}\right)_{(X_{\theta},X_{f},X_{w},X_{m^{\prime}})=0}.$ In this case, the thermal flow is due only to the asymmetry of the membrane, induced by its bending. This phenomenon is similar to the Dufour effect [20]. If one can prove experimentally that a bending of the membrane induces a heat flow through it, this means that $L_{\theta m}\neq 0$, hence, by Onsager’s reciprocity relations, $L_{m\theta}\neq 0$, i.e. a heat flow modifies the bending. Indeed, we also have the relation : (23) $\displaystyle L_{m\theta}=\left(\frac{J_{m}}{X_{\theta}}\right)_{(X_{m},X_{f},X_{w},X_{m^{\prime}})=0}.$ Hence, $L_{m\theta}$ measures the effect of a slight temperature difference (between both sides of the membrane) on the induced flow of molecules between the leaflets, which implies a modification of its mean curvature. This phenomenon is similar to the thermodiffusion or Soret effect [20]. It is reciprocal to the previous effect and might be easier to observe and measure. ### 3.5. Entropy production and stability Just as the power dissipated by Joule effect in an ohmic conductor is (24) Power dissipated $\displaystyle=\text{Current}\times\text{Voltage}$ $\displaystyle=\text{Conductance}\times\text{Voltage}^{2},$ the rate of dissipation of energy, or entropy creation, in a general chemical system out of equilibrium is a quadratic function of the thermodynamical forces acting in the system [32, 20] : (25) Rate of entropy produced $\displaystyle=\text{Flows}\times\text{Forces}$ $\displaystyle=\text{Forces}\times\text{Conductance matrix}\times\text{Forces}.$ This relation rests on a linearity hypothesis supposed to be valid only in the neighbourhood of an equilibrium state. The main difference between the ohmic conductor and the chemical system is that, in the latter, the conductance is not a single number but a matrix which, in the general case, couples all the currents. Taking into account the various thermodynamical forces defined previously, the rate of entropy production inside the protocell has to the following expression : (26) $\displaystyle\sigma(X)$ $\displaystyle=L_{ff}X_{f}^{2}+L_{ww}X_{w}^{2}+L_{mm}X_{m}^{2}+L_{m^{\prime}m^{\prime}}X_{m^{\prime}}^{2}+L_{rr}X_{r}^{2}+L_{\theta\theta}X_{\theta}^{2}+2X_{\theta}(L_{\theta f}X_{f}+L_{\theta w}X_{w}+L_{\theta m}X_{m})$ $\displaystyle=L_{ff}\left(\frac{\mu_{f0}}{T_{0}}-\frac{\mu_{f1}}{T_{1}}\right)^{2}+L_{ww}\left(\frac{\mu_{w0}}{T_{0}}-\frac{\mu_{w1}}{T_{1}}\right)^{2}+L_{mm}\left(\frac{\mu_{m0}}{T_{0}}-\frac{\mu_{m1}}{T_{1}}\right)^{2}$ $\displaystyle\qquad+L_{\theta\theta}\left(\frac{1}{T_{0}}-\frac{1}{T_{1}}\right)^{2}+L_{m^{\prime}m^{\prime}}\left(\frac{\mu_{mc}-\mu_{m1}}{T_{1}}\right)^{2}+L_{rr}\left(\frac{A_{r}}{T_{1}}\right)^{2}$ $\displaystyle\qquad+2\left(\frac{1}{T_{0}}-\frac{1}{T_{1}}\right)\left(L_{\theta f}\left(\frac{\mu_{f0}}{T_{0}}-\frac{\mu_{f1}}{T_{1}}\right)+L_{\theta w}\left(\frac{\mu_{w0}}{T_{0}}-\frac{\mu_{w1}}{T_{1}}\right)+L_{\theta m}\left(\frac{\mu_{m0}}{T_{0}}-\frac{\mu_{m1}}{T_{1}}\right)\right).$ The stability of this steady state is equivalent to the positivity of the matrix $L$, which is also the matrix of second order derivatives of $\sigma$ in the coordinate system $X=(X_{\theta},X_{f},X_{w},X_{m},X_{m^{\prime}},X_{r})$ : (27) $\displaystyle L_{ik}=\frac{1}{2}\,\frac{\partial^{2}\sigma}{\partial X_{i}\partial X_{k}}.$ If $P$ is a $n\times n$ matrix with real coefficients, the positivity of $P$, defined by : (28) $\displaystyle u^{\text{t}}Pu>0\qquad\forall u\in\mathbb{R}^{n}$ implies the following inequalities : (29) $\displaystyle P_{ii}>0\quad\forall\,i\qquad\text{and}\qquad P_{ii}P_{jj}>\left(\frac{P_{ij}+P_{ji}}{2}\right)^{2}\quad\forall\,i,j.$ These conditions are necessary but not sufficient to ensure the positivity of $P$. In the present case, $L$ being symmetric, we have, in particular : (30) $\displaystyle L_{ii}$ $\displaystyle>0$ $\displaystyle L_{\theta\theta}L_{ff}$ $\displaystyle>L_{\theta f}^{2}$ $\displaystyle L_{\theta\theta}L_{ww}$ $\displaystyle>L_{\theta w}^{2}$ $\displaystyle L_{\theta\theta}L_{qq}$ $\displaystyle>L_{\theta q}^{2}$ $\displaystyle L_{mm}L_{qq}$ $\displaystyle>L_{mq}^{2}$ $\displaystyle L_{\theta\theta}L_{mm}$ $\displaystyle>L_{\theta m}^{2}.$ If one of these inequalities is not satisfied, the growth process is destabilized. In Section VI, we will prove that the last one can be reversed as the inner temperature of the protocell increases. In order to prove this proposition, we must first write down evolution equations for the geometry of the cell. ## 4\. Membrane geometry and growth equation Just as the growth of a child depends on his diet, the evolution of the geometric parameters of a protocell depends on the flow of molecules to its membrane. This flow is determined by the food intake and by the rate of the synthesis of these structural molecules. In this section, we establish the differential equations governing the growth of the volume and area of a cylindrical protocell by relating them to the flows of matter. ### 4.1. Conservation of matter and exponential growth The advancement, $\xi$, of the overall synthesis reaction, $f\to m+c+w$, is the internal clock of the protocell. The corresponding flow of matter, $J_{r}=\mfrac{\mathrm{d}\xi}{\mathrm{d}t}$, is channeled to all the other processes in the protocell. In particular, it determines the flux of matter to the inner leaflet and the growth speed of the membrane. By writing the equations of conservation of matter, we can then determine the evolution of the size of the protocell. Let $a\in\\{c,1,0\\}$ denote the possible position of a membrane molecule : either in the cytosol $(c)$, or the inner leaflet $(1)$ or the outer leaflet $(0)$. Let $N_{ma}$ be the number of membrane molecules in each of them. The time derivatives of these functions are related to the flows defined previously : (31) $\displaystyle\frac{\mathrm{d}N_{mc}}{\mathrm{d}t}$ $\displaystyle=-J_{mc1}\mathcal{A}_{1}+J_{rm}\mathcal{V}$ $\displaystyle\frac{\mathrm{d}N_{m1}}{\mathrm{d}t}$ $\displaystyle=J_{mc1}\mathcal{A}_{1}-J_{m10}\mathcal{A}$ $\displaystyle\frac{\mathrm{d}N_{m0}}{\mathrm{d}t}$ $\displaystyle=J_{m10}\mathcal{A}.$ Similarly, the number of food (resp. cytosol and waste) molecules, $N_{f}$ (resp. $N_{c}$ and $N_{w}$), evolves according to the following relations : (32) $\displaystyle\frac{\mathrm{d}N_{f}}{\mathrm{d}t}$ $\displaystyle=J_{f}\mathcal{A}_{0}-J_{rf}\mathcal{V}$ $\displaystyle\frac{\mathrm{d}N_{c}}{\mathrm{d}t}$ $\displaystyle=J_{rc}\mathcal{V}$ $\displaystyle\frac{\mathrm{d}N_{w}}{\mathrm{d}t}$ $\displaystyle=-J_{w}\mathcal{A}_{1}+J_{rw}\mathcal{V}$ where the flows $J_{r\bullet}$ are defined by : (33) $\displaystyle J_{rm}$ $\displaystyle:=\nu_{m}\frac{\mathrm{d}\xi}{\mathrm{d}t}$ $\displaystyle J_{rf}$ $\displaystyle:=\nu_{f}\frac{\mathrm{d}\xi}{\mathrm{d}t}\,=\,\frac{\nu_{f}}{\nu_{m}}\,J_{rm}$ $\displaystyle J_{rc}$ $\displaystyle:=\nu_{c}\frac{\mathrm{d}\xi}{\mathrm{d}t}\,=\,\frac{\nu_{c}}{\nu_{m}}\,J_{rm}$ $\displaystyle J_{rw}$ $\displaystyle:=\nu_{w}\frac{\mathrm{d}\xi}{\mathrm{d}t}\,=\,\frac{\nu_{w}}{\nu_{m}}\,J_{rm}.$ In a steady state, the concentration of membrane molecules in the cytosol is constant : (34) $\displaystyle C_{mc}:=\frac{N_{mc}}{\mathcal{V}}=\text{cst.}$ Let $c_{m0}$ and $c_{m1}$ be the average number of membrane molecules per unit area in each leaflet : (35) $\displaystyle c_{m0}:=\frac{N_{m0}}{\mathcal{A}_{0}}\qquad\text{and}\qquad c_{m1}:=\frac{N_{m1}}{\mathcal{A}_{1}}.$ The conservation equations for $m$ imply the evolution equations of the geometry of the protocell : (36) $\displaystyle c_{m0}\frac{\mathrm{d}\mathcal{A}_{0}}{\mathrm{d}t}$ $\displaystyle=J_{m10}\,\frac{\mathcal{A}_{0}+\mathcal{A}_{1}}{2}$ $\displaystyle c_{m1}\frac{\mathrm{d}\mathcal{A}_{1}}{\mathrm{d}t}$ $\displaystyle=J_{mc1}\mathcal{A}_{1}-c_{m0}\frac{\mathrm{d}\mathcal{A}_{0}}{\mathrm{d}t}$ $\displaystyle c_{mc}\frac{\mathrm{d}\mathcal{V}}{\mathrm{d}t}$ $\displaystyle=J_{rm}\mathcal{V}-J_{mc1}\mathcal{A}_{1}.$ Let us introduce the following parameters : (37) $\displaystyle 2\varepsilon$ $\displaystyle:=\text{average thickness of the membrane}$ $\displaystyle\eta$ $\displaystyle:=\frac{c_{m1}}{c_{m0}}\quad\text{(layer density ratio $\simeq 1$)}$ $\displaystyle\tau$ $\displaystyle:=\frac{J_{m10}}{J_{mc1}}\quad\text{(transmission rate {through} the membrane)}$ $\displaystyle t_{1}$ $\displaystyle:=\frac{c_{m1}}{J_{mc1}}\quad\text{(inner leaflet characteristic time)}$ $\displaystyle\tau_{c}$ $\displaystyle:=\frac{J_{mc1}}{J_{rm}}\quad\text{(transmission rate {to} the membrane)}$ $\displaystyle t_{c}$ $\displaystyle:=\frac{c_{mc}}{J_{rm}}\quad\text{(cytosol characteristic time)}.$ The transmission ratio, $\tau$, can be written in terms of thermodynamical forces : (38) $\displaystyle\tau:=\frac{J_{m}}{J_{mc1}}=\frac{L_{mm}X_{m}+L_{m\theta}X_{\theta}+{\ldots}}{L_{m^{\prime}m^{\prime}}X_{m^{\prime}}}.$ Let $\mathcal{U}={\mathcal{V}}/{\varepsilon}$ and $\dot{X}=t_{1}\mfrac{\mathrm{d}X}{\mathrm{d}t}$. We obtain the following system of differential equations : (39) $\displaystyle\dot{\mathcal{A}}_{0}$ $\displaystyle=\frac{\eta\tau}{2}\,(\mathcal{A}_{0}+\mathcal{A}_{1})\,=\,\eta\tau\mathcal{A}$ $\displaystyle\dot{\mathcal{A}}_{1}$ $\displaystyle=-\frac{\tau}{2}\,\mathcal{A}_{0}+\left(1-\frac{\tau}{2}\right)\,\mathcal{A}_{1}\,=\,(1-\tau)\mathcal{A}-\mathcal{B}$ $\displaystyle\dot{\mathcal{U}}$ $\displaystyle=\frac{t_{1}}{t_{c}}\,\mathcal{U}-\frac{c_{m1}}{\varepsilon c_{mc}}\,\mathcal{A}_{1}.$ In matrix form : (40) $\displaystyle\dot{X}$ $\displaystyle=\begin{pmatrix}\dot{\mathcal{A}}_{0}\\\ \dot{\mathcal{A}}_{1}\\\ \dot{\mathcal{U}}\\\ \end{pmatrix}=\begin{pmatrix}\frac{\eta\tau}{2}&\frac{\eta\tau}{2}&0\\\ -\frac{\tau}{2}&\frac{2-\tau}{2}&0\\\ 0&-\frac{c_{m1}}{\varepsilon c_{mc}}&\frac{t_{1}}{t_{c}}\\\ \end{pmatrix}\begin{pmatrix}\mathcal{A}_{0}\\\ \mathcal{A}_{1}\\\ \mathcal{U}\\\ \end{pmatrix}=MX$ $\displaystyle M$ $\displaystyle:=\begin{pmatrix}\frac{\eta\tau}{2}&\frac{\eta\tau}{2}&0\\\ -\frac{\tau}{2}&\frac{2-\tau}{2}&0\\\ 0&-\frac{c_{m1}}{\varepsilon c_{mc}}&\frac{t_{1}}{t_{c}}\\\ \end{pmatrix}\quad\text{and}\quad X:=\begin{pmatrix}\mathcal{A}_{0}\\\ \mathcal{A}_{1}\\\ \mathcal{U}\\\ \end{pmatrix}.$ This growth equation is solved in Appendix B. The matrix $M$ has a block diagonal form, hence $\mathcal{A}_{0}$ and $\mathcal{A}_{1}$ evolve independently of $\mathcal{U}$, whereas the equation for $\mathcal{U}$ contains terms linear in $\mathcal{A}_{0}$ and $\mathcal{A}_{1}$. The upper left $2\times 2$ block is not diagonal, hence $\mathcal{A}_{0}$ and $\mathcal{A}_{1}$ are linear combinations of exponential functions of time (multiplied by an affine function of $t$ in the degenerate, non diagonalisable case). The rates of growth of these exponential functions are the eigenvalues of this $2\times 2$ block, plus an exponential of growth rate $\mfrac{t_{1}}{t_{c}}$ for $\mathcal{U}$. ### 4.2. Cylindrical growth in steady state When we meet an ordinary differential equation, describing the time evolution of a dynamical system, a first reflex is to search for constant solutions or at least steady state solutions, where the speed is constant. In the present case, we can look for a solution where the length increases steadily whereas the radius is constant. This corresponds to the observed growth of some bacterial species in difficult environments [28]. When the sludge content of wastewater is too high or when the composition is lopsided, a higher percentage of bacteria adopt a filamentous growth strategy which allows them to survive in harsher conditions, by catching food more easily. If the protocell grows like a cylinder of radius $R_{0}$, we have $\varepsilon\mathcal{A}=R_{0}\mathcal{B}$, hence $\mfrac{\mathrm{d}\mathcal{A}}{\mathcal{A}}=\mfrac{\mathrm{d}\mathcal{B}}{\mathcal{B}}$ and (41) $\displaystyle x$ $\displaystyle:=\frac{R_{0}}{\varepsilon}=\frac{\mathcal{A}}{\mathcal{B}}=\frac{\mathrm{d}\mathcal{A}}{\mathrm{d}\mathcal{B}}=\frac{\dot{\mathcal{A}}}{\dot{\mathcal{B}}}$ $\displaystyle=\frac{\big{(}(\eta-1)\tau+1\big{)}\mathcal{A}-\mathcal{B}}{\big{(}(\eta+1)\tau-1\big{)}\mathcal{A}+\mathcal{B}}$ $\displaystyle=\frac{\alpha_{+}\mathcal{A}-\mathcal{B}}{\alpha_{-}\mathcal{A}+\mathcal{B}}$ where $\alpha_{\pm}=(\eta\mp 1)\tau\pm 1$. Therefore, $x$ satisfies the fixed point equation : (42) $\displaystyle x=\frac{\alpha_{+}x-1}{\alpha_{-}x+1}\qquad\text{{\it i.e. }}\qquad\alpha_{-}x^{2}-(\alpha_{+}-1)x+1=0.$ The discriminant of this quadratic equation is (43) $\displaystyle(\alpha_{+}-1)^{2}-4\alpha_{-}$ $\displaystyle=(\eta-1)^{2}\tau^{2}-4(\eta+1)\tau+4$ $\displaystyle=4\Delta(\eta,\tau)$ (cf. Appendix B) and its roots, $x_{\pm}$, are related to the eigenvalues, $\lambda_{\pm}$, of the matrix $M$ (Eq. 40) : (44) $\displaystyle x_{\pm}$ $\displaystyle=\frac{1}{2\alpha_{-}}\left(\alpha_{+}-1\pm\sqrt{(\alpha_{+}-1)^{2}-4\alpha_{-}}\right)$ $\displaystyle=\frac{(\alpha_{+}-1)\pm 2\sqrt{\Delta(\eta,\tau)}}{2\alpha_{-}}$ $\displaystyle=\frac{2\lambda_{\pm}-1}{\alpha_{-}}.$ Consequently, the radius, $R_{0}$, of the cylinder whose length increases in a steady state is determined by the flows $(J_{mc1},J_{m10},J_{rm})$ and the concentrations $(C_{mc},c_{m1},c_{m0})$, via the coefficients $(\varepsilon,\eta,\tau)$ : (45) $\displaystyle R_{0}$ $\displaystyle=\varepsilon x_{\pm}=\varepsilon\,\frac{\lambda_{\pm}-1}{2\alpha_{-}}=\frac{\varepsilon\big{(}(\eta+1)\tau\pm 2\sqrt{\Delta}\big{)}}{2\big{(}(\eta+1)\tau-1\big{)}}$ $\displaystyle=\frac{\varepsilon}{2}\,\frac{(\eta+1)\tau\pm\sqrt{(\eta-1)^{2}\tau^{2}-(\eta+1)\tau+1}}{(\eta+1)\tau-1}$ ## 5\. Thermal instability of cylindrical growth As long as the protocell grows by increasing only its length, keeping a cylindrical shape of fixed radius, $R_{0}$, its volume and its membrane area grow proportionally, i.e. $\dot{\mathcal{A}}=\text{cst.}\times\dot{\mathcal{B}}$. If the heat generated by the metabolic reactions were exactly proportional to the volume increment, the increase of the area of the membrane would be sufficient to evacuate steadily the heat generated by the chemical reactions taking place inside the newly created volume. However, the heat generated by all these irreversible processes adds up to that coming from the exothermic metabolic reactions and the inner temperature must therefore increase. This overheating generates larger fluctuations of all the physical parameters which destabilize the initial steady state of cylindrical growth. We will see below that the geometrical parameters $(\mathcal{A},\mathcal{B},\mathcal{V})$ can follow a path leading to a more efficient release of heat, by reducing the radius $R_{0}$. ### 5.1. The Squeezed Sausage Theorem (SST) When we squeeze a sausage, its length increases as well as its area. Indeed, the stuffing being incompressible, the squeezing is an isovolumic deformation. The stuffing is pushed longitudinally, away from the squeezed zone, and increases the length of the sausage, thanks to the elasticity of the gut. The area of the slice of reduced radius increases consequently to bound the same volume. Let us prove this mathematically. A length $\delta x$ of cylinder of radius $R_{0}$ has volume $\delta\mathcal{V}$ and boundary area $\delta A$ given by : (46) $\displaystyle\delta\mathcal{V}$ $\displaystyle=\pi R_{0}^{2}\,\delta x$ $\displaystyle\delta\mathcal{A}$ $\displaystyle=2\pi R_{0}\,\delta x$ Let us suppose that this cylindrical growth is perturbed by a small, local radius variation, which can be positive (anevrism) or negative (stenosis). We study here a triangular perturbation and, in the appendix, a smooth $(\mathcal{C}^{2})$, rotation invariant perturbation of the cylinder. To keep it simple, we suppose that this perturbation is piecewise linear and symmetric, with an extremum $\delta R$ at $x=0$, and vanishes outside of the interval $\left[-\mfrac{\delta x^{\prime}}{2},\mfrac{\delta x^{\prime}}{2}\right]$. FIG. 5.1 represents the resulting isovolumic deformation according with the sign of $\delta R$. $2R_{0}$$\delta x$$\delta\mathcal{V}$$\delta\mathcal{A}$$\delta x^{\prime}>\delta x$ $2(R_{0}+\delta R)<2R_{0}$ $\delta\mathcal{V}^{\prime}=\delta\mathcal{V}$ $\delta\mathcal{A}^{\prime}>\delta A$ $2(R_{0}+\delta R)>2R_{0}$ $\delta\mathcal{V}^{\prime}=\delta V$ $\delta x^{\prime}<\delta x$$\delta\mathcal{A}^{\prime}<\delta A$ Isovolumic variation of the area of a cylinder under a small triangular deformation. In the second and third pictures of FIG. 5.1, the Gaussian curvature is concentrated on the circular sections at $x=0$ and at $x=\pm\mfrac{\delta x^{\prime}}{2}$ (dotted lines), where the mean curvature has a finite discontinuity. The volume and lateral membrane area of this slice of thickness $\delta x^{\prime}$ (contained between the dotted lines) are therefore : (47) $\displaystyle\delta\mathcal{V}^{\prime}$ $\displaystyle=\pi\left(R_{0}+\frac{\delta R}{2}\right)^{2}\,\delta x^{\prime}$ $\displaystyle\delta\mathcal{A}^{\prime}$ $\displaystyle=2\pi\left(R_{0}+\frac{\delta R}{2}\right)\,\delta x^{\prime}.$ The straight slice and the deformed slice have equal volumes $(\delta\mathcal{V}=\delta\mathcal{V}^{\prime})$ if their thicknesses satisfy : (48) $\displaystyle\frac{\delta x^{\prime}}{\delta x}=\left(1+\frac{\delta R}{2R_{0}}\right)^{-2}\simeq\left(1-\frac{\delta R}{R_{0}}\right).$ Hence the ratio of their areas is (49) $\displaystyle\frac{\delta\mathcal{A}^{\prime}}{\delta\mathcal{A}}\simeq\left(1+\frac{\delta R}{2R_{0}}\right)\left(1-\frac{\delta R}{R_{0}}\right)\simeq 1-\frac{\delta R}{2R_{0}}.$ The heat flows through these surfaces are, respectively : (50) $\displaystyle\delta q$ $\displaystyle=L_{\theta\theta}\left(\frac{1}{T_{0}}-\frac{1}{T_{1}}\right)\,\delta\mathcal{A}$ $\displaystyle\delta q^{\prime}$ $\displaystyle=L_{\theta\theta}\left(\frac{1}{T_{0}}-\frac{1}{T_{1}}\right)\,\delta\mathcal{A}^{\prime}$ hence their ratio is the same as for the areas : (51) $\displaystyle\frac{\delta q^{\prime}}{\delta q}=\frac{\delta\mathcal{A}^{\prime}}{\delta\mathcal{A}}=1-\frac{\delta R}{2R_{0}}.$ When $\delta R<0$, this ratio is larger than $1$. Consequently, the inner volume being held fixed, a small stenosis of a cylindrical protocell evacuates heat more efficiently than a small anevrism. This local reduction of the radius of the protocell increases its mean curvature. For this deformation to happen, the outer leaflet must grow more rapidly than the inner leaflet. Therefore, the equilibrium $m_{1}\leftrightarrows m_{0}$ must be shifted towards $m_{0}$ in order to have $\delta R<0$. This is possible if $T_{1}$ increases slightly and $m_{1}\to m_{0}$ is exothermic. We propose that the translocation of membrane molecules to the outer leaflet [14, 15, 4, 1] can be triggered by the increase of the inner temperature, $T_{1}(t)$. The area of the outer leaflet then increases more quickly than the area of the inner leaflet, which leads to the bending of the membrane until the total splitting of the protocell into two daughters. ### 5.2. Fluctuations, translocation and heat transfer In order to increase $L_{m\theta}$ and destabilise the cylindrical growth, the transfer coefficient, $\tau$, must also increase. In [14, 15], the authors present a detailed mechanism for the transfer of membrane molecules between the leaflets. Due to the fluctuations of ionic densities in the neighbourhood of the membranes, the local electric field fluctuates strongly enough to push molecules of water into the membrane, via the field-dipole interaction force (dielectrophoresis). When it is sufficiently strong, this force can create a transient water pore that is stable enough to let some membrane molecules dive into this water pore and join the other side. The increase of the inner temperature can also enhance these ionic density fluctuations and favor this translocation process from the hot side to the cold side, since the hottest, most agitated molecules have a higher probability to dive into the water pore than the colder molecules. This asymmetric flow of hot molecules to the cold side enhances the outgoing heat flow and cools down the protocell. During this process, the shape of the hydrophobic tails is not important, as long as they remain in the hydrophobic zone, surrounded by siblings. The only energetic cost is for the hydrophilic head surrounded by these aliphatic chains, and some clandestine water molecules forming the water pore (not represented below). The shape of the tail is irrelevant since the energy depends only on the position of the polar head (FIG. 5.2). $\circ$$\circ$$\circ$$\circ$$\circ$$\circ$$\circ$$\circ$hydrophilic zonehydrophilic zonehydrophobic zone Translocation of a membrane molecule from one leaflet to the other. ### 5.3. Thermal balance Let us make a thermal balance of the whole growth process. After heating its cold nutrient molecules from $T_{0}$ to $T_{1}$ and processing the isothermal inner chemical reactions $(J_{r})$, our protocell disposes of its hot waste (including some water flowing through the water pores) and loses heat by translocation of membrane molecules from the inside to the outside, and by diffusion $(J_{\theta})$ without mass transfer. Let $q_{i}$ be the heat exported out of the protocell by each molecule of type $i$. Cold entering molecules and hot outgoing molecules both have $q_{i}>0$. Let $\kappa_{i}$ be the heat capacity of the molecules of type $i$. Let $J_{h}$ be the outgoing heat flow (energy/(time $\times$ area)). The heat flow exported by the cold entering food and water molecules is : (52) $\displaystyle J_{f}q_{f}$ $\displaystyle=J_{f}\kappa_{f}(T_{1}-T_{0}).$ Similarly, the heat flow exported by the outgoing waste and water molecules is : (53) $\displaystyle J_{w}q_{w}$ $\displaystyle=J_{w}\kappa_{w}(T_{1}-T_{0}).$ And the heat flow exported by the net translocation of membrane molecules is : (54) $\displaystyle J_{m}q_{m}$ $\displaystyle=J_{m}\kappa_{m}(T_{1}-T_{0})$ if we suppose that they immediately thermalise from $T_{1}$ to $T_{0}$ once they reach the outer leaflet. The contact of the hydrophobic tails inside the membrane allows for a diffusive heat flow : (55) $\displaystyle J_{\theta}$ $\displaystyle=\sum_{i}L_{\theta k}X_{k}.$ The total heat flow is the sum of these terms : (56) $\displaystyle J_{h}$ $\displaystyle:=(J_{f}q_{f}+J_{w}q_{w}+J_{m}q_{m})+J_{\theta}$ $\displaystyle=\big{(}J_{f}\kappa_{f}+J_{w}\kappa_{w}\big{)}(T_{1}-T_{0})$ $\displaystyle\quad+(L_{mm}X_{m}+L_{m\theta}X_{\theta}+{\ldots})\kappa_{m}(T_{1}-T_{0})$ $\displaystyle\quad+(L_{\theta\theta}X_{\theta}+L_{\theta m}X_{m}+{\ldots}).$ Since (57) $\displaystyle T_{1}-T_{0}=T_{0}T_{1}X_{\theta}=\frac{T_{0}^{2}X_{\theta}}{1-T_{0}X_{\theta}}$ $L_{m\theta}$ appears as a factor of $X_{\theta}^{2}$ in the convective term, $J_{m}$, whereas $L_{\theta m}$ is a factor of $X_{m}$ in the diffusive term, $J_{\theta}$. Moreover, $X_{m}$ increases linearly with $X_{\theta}$ : (58) $\displaystyle X_{m}$ $\displaystyle=\frac{\mu_{m0}}{T_{0}}-\frac{\mu_{m1}}{T_{1}}$ $\displaystyle=\frac{\mu_{m}^{\circ}}{T_{0}}-\frac{\mu_{m}^{\circ}}{T_{1}}+k_{B}\ln\left(\frac{a_{m0}}{a_{m1}}\right)$ $\displaystyle=\mu_{m}^{\circ}X_{\theta}+k_{B}\ln\left(\frac{a_{m0}}{a_{m1}}\right)$ where $\mu^{\circ}$ denotes the standard chemical potential, at temperature $298$ K and pressure $1$ atm [2]. The $L_{m\theta}$-dependent term in $J_{h}$ becomes : (59) $\displaystyle J_{h}=L_{m\theta}\left(\frac{\kappa_{m}T_{0}^{2}X_{\theta}^{2}}{1-T_{0}X_{\theta}}+\mu_{m}^{\circ}X_{\theta}\right)+{\ldots}$ Consequently, as the cytosol heats up, $J_{h}$ increases more quickly by translocation ($\kappa_{m}$ term) than by diffusion ($\mu_{m}^{\circ}$ term). Translocation is a particular kind of heat convection and by analogy with the Rayleigh-Bénard instability [35], we conjecture the existence of a transition from a diffusive regime to a convective regime, where translocation overtakes diffusion and expells heat more efficiently. ## 6\. Translocation between leaflets The energetic barrier, of width $2\varepsilon^{\prime}$ and height $E_{\ast}$, is difficult to penetrate for the hydrophilic head since this guarantees the stability of the bilayer under ordinary thermal fluctuations. When the ratio of concentrations, $\eta=\mfrac{c_{m1}}{c_{m0}}$, becomes too large compared to unity, the mechanical constraint on the inner leaflet is released by pushing molecules to the outer leaflet. Conversely, when the outer leaflet is stretched and the inner leaflet compressed, $\eta$ is slightly greater than unity (FIG. 6). $\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup\restrictwandup$$c_{m1}$compressed$c_{m0}<c_{m1}$stretched$\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand\restrictwand$ Mechanical constraints modify the ratio, $\eta$, of leaflet concentrations. To facilitate this process, some water molecules can leak through the hydrophobic zone and ease the passage of the hydrophilic head. This leakage of water lowers the activation energy, $E_{\ast}$, and realizes an aqueous catalysis of the translocation process [14, 15, 4, 1]. If we suppose that the density, $n_{p}$, of water pores in the membrane is constant for fixed temperatures, $T_{0}$ and $T_{1}$, then $J_{m10}$ depends only on this density and on the net number, $j_{mp}$, of membrane molecules translocated from $\mathbf{L}_{1}$ to $\mathbf{L}_{0}$ during the lifetime of the pores : (60) $\displaystyle n_{p}$ $\displaystyle:=\text{ number of water pores per unit area }$ $\displaystyle j_{mp}$ $\displaystyle:=\text{ net number of translocations}$ through each water pore $\displaystyle J_{m}$ $\displaystyle=n_{p}j_{mp}.$ This first approximation is based on the hypothesis that the pores have the same size, the same lifetime and the same number of net translocations during their short life. However, to be more realistic, we must take into account the fact that larger pores live longer and leak more (over the same duration) than smaller short lived pores. We integrate over the interval of possible lifetimes $(t_{p})$ the density of water pores of lifetime $t_{p}$ created per unit time $(n_{p}(t_{p}))$ multiplied by the net number $(\nu_{mp}(t_{p}))$ of molecules each pore of lifetime $t_{p}$ translocates from the inside to the outside during its existence : (61) $\displaystyle J_{m}=\int_{0}^{\infty}\mathrm{d}t_{p}\,n_{p}(t_{p})\nu_{mp}(t_{p}).$ The increase of $X_{\theta}$ enhances at the same time the rate of formation of pores, hence $n_{p}$, and the net number of translocated molecules, due to larger thermal fluctuations. Therefore, $J_{m10}$ increases more than linearly as a function of $X_{\theta}$. Consequently, the crossed conductivity coefficient, $L_{m\theta}$, increases with $X_{\theta}$. On the other side of the inequality, $L_{\theta\theta}$ and $L_{mm}$ depend more weakly on the temperature. Indeed, the heat diffusion coefficient, $L_{\theta\theta}$, involves the (temperature independant) number of interacting degrees of freedom between the hydrophobic tails inside the hydrophobic layer, and the molecular diffusion coefficient : (62) $\displaystyle L_{mm}=T_{0}\left(\frac{J_{m}}{\mu_{m1}-\mu_{m0}}\right)_{(X_{\theta},X_{f},X_{w},X_{m^{\prime}})=0}$ depends mainly on the ratio of concentrations between the two leaflets, i.e. on $\eta$. In order to know if the initial inequality, $L_{m\theta}^{2}<L_{\theta\theta}L_{mm}$, can be reversed, the temperature dependance of the convective coefficient, $L_{m\theta}$, must be computed and compared to that of the diffusion coefficients, $L_{\theta\theta}$ and $L_{mm}$. This necessitates a microscopic model of the interactions of membrane molecules and water and a precise description of the translocation process, to go beyond the linear response theory. In the sequel, we adopt a simple mean field approach where each molecule evolves in the same energetic landscape as the others. ### 6.1. An effective potential for translocation The exact shape and position of each membrane molecule is described by dozens of parameters specifying the position of each atom and the orientation of each interatomic bond. It would be cumbersome to take them all into account to describe mathematically the evolution of a single molecule inside the membrane. However, we can make a simplifying approximation by remarking that the main energetic cost is in the displacement of the hydrophilic head into the hydrophobic layer or the protrusion of this head outside of the membrane, which forces the tail to go into the hydrophilic zone. We can make a mean- field approximation by considering only the position, $z$, of the hydrophilic head as a dynamical variable, and defining an adequate effective potential energy, $U(z)$, that traps the head inside the membrane. In the sequel of this article, we will use a double well effective potential to compute the net flow, $J_{m}$, across a plane bilayer subject to a difference of temperatures. By differentiation, we obtain the coefficients $L_{m\theta}$ and $L_{mm}$ and, in particular, their dependence on temperature. This model suggests that the inequality $L_{m\theta}^{2}<L_{mm}L_{\theta\theta}$ can be reversed if the inner temperature increases sufficiently. Our hypotheses are the following ones : 1. (1) The membrane molecules have length $\varepsilon=\varepsilon^{\prime}+\varepsilon^{\prime\prime}$, where $\varepsilon^{\prime\prime}$ is the size of the hydrophilic head and $\varepsilon^{\prime}$ is the length of the hydrophobic tail. 2. (2) The translocation process is described by only one parameter : the position of the center of mass of the hydrophilic head, varying between $-\varepsilon$ and $+\varepsilon$. 3. (3) On each side of the membrane, the distribution of velocities of the heads follows a Maxwell-Boltzmann law [35]. The probability of finding a molecule with velocity $v$ perpendicularly to the membrane is : (63) $\displaystyle p_{i}(v)=\sqrt{\frac{m}{2\pi k_{B}T_{i}}}\exp\left(-\frac{mv^{2}}{2k_{B}T_{i}}\right).$ 4. (4) The translocation requires an energy $E^{\ast}$ and the head of the molecule evolves in an effective double well potential (FIG. 4). hydrophilic zonepolar heads\+ water\+ ionshydrophobic zonealiphatic chainshydrophilic zonepolar heads\+ water\+ ions$z$$-\varepsilon$$-\varepsilon^{\prime}$$T_{1}$$\varepsilon$$\varepsilon^{\prime}$$T_{0}$$E^{\ast}$$\color[rgb]{1,0,0}U(z)$ Potential energy of the hydrophilic head 5. (5) The hydrophilic heads trapped in the well $[-\varepsilon,-\varepsilon^{\prime}]$ have temperature $T_{1}$, whereas those trapped in the well $[\varepsilon^{\prime},\varepsilon]$ have temperature $T_{0}$. The thermalisation processes for the motion along the $z$ axis occur only once the head is trapped in the arrival well. This drastic hypothesis simplifies the computations and should be refined in a more realistic model. In reality, the motions of the hydrophobic tails between $z=-\varepsilon^{\prime}$ and $z=+\varepsilon^{\prime}$ can thermalise the molecule during the travel across the membrane and this affects the translocation time. Only half of the molecules of kinetic energy $E>E_{\ast}$ can escape from a well to the other side. The time it takes them to go through the barrier is given by : (64) $\displaystyle t_{f}$ $\displaystyle=\int_{-\varepsilon^{\prime}}^{+\varepsilon^{\prime}}\,\mathrm{d}z\sqrt{\frac{m}{2(E-E_{\ast})}}$ $\displaystyle=2\varepsilon^{\prime}\sqrt{\frac{m}{2(E-E_{\ast})}}$ $\displaystyle=2\varepsilon^{\prime}\sqrt{\frac{m}{mv^{2}-2E_{\ast}}}.$ The flow of molecules of velocity belonging to the interval $[v,v+\mathrm{d}v]$, with $v>v_{\ast}:=\sqrt{\mfrac{2E_{\ast}}{m}}$, going from side $1$ to side $0$, is proportional to the surface density of molecules, $c_{m1}$, to the Maxwell-Boltzmann weight, $p_{1}(v)\mathrm{d}v$, of this velocity interval, and to the reciprocal of the translocation time : (65) $\displaystyle J_{m10}$ $\displaystyle=\int_{v_{\ast}}^{+\infty}\mathrm{d}v\,\frac{c_{m1}p_{1}(v)}{t_{f}}$ $\displaystyle=\int_{E_{\ast}}^{+\infty}\frac{\mathrm{d}E}{\sqrt{2mE}}\,\frac{1}{2\varepsilon^{\prime}}\,\sqrt{\frac{2(E-E_{\ast})}{m}}\,\frac{c_{m1}e^{-E/k_{B}T_{1}}}{\sqrt{\frac{2\pi k_{B}T_{1}}{m}}}$ $\displaystyle=\frac{1}{2\varepsilon^{\prime}\sqrt{\pi}}\int_{E_{\ast}}^{+\infty}\frac{\mathrm{d}E}{\sqrt{2mE}}\sqrt{\frac{E-E_{\ast}}{k_{B}T_{1}}}\,c_{m1}e^{-E/k_{B}T_{1}}.$ The net flow of molecules from leaflet $1$ to leaflet $0$ is : (66) $\displaystyle J_{m}$ $\displaystyle:=J_{m10}-J_{m01}$ $\displaystyle=\frac{1}{2\varepsilon^{\prime}\sqrt{\pi}}\int_{E_{\ast}}^{+\infty}\frac{\mathrm{d}E}{\sqrt{2mE}}\sqrt{\frac{E-E_{\ast}}{k_{B}T_{1}}}\,c_{m1}e^{-E/k_{B}T_{1}}$ $\displaystyle\quad-\frac{1}{2\varepsilon^{\prime}\sqrt{\pi}}\int_{E_{\ast}}^{+\infty}\frac{\mathrm{d}E}{\sqrt{2mE}}\sqrt{\frac{E-E_{\ast}}{k_{B}T_{0}}}\,c_{m0}e^{-E/k_{B}T_{0}}.$ ### 6.2. Computation of $L_{m\theta}$ The temperature $T_{0}$ being fixed, we have : (67) $\displaystyle L_{m\theta}$ $\displaystyle=\frac{\partial J_{m}}{\partial X_{\theta}}=-\frac{\partial J_{m}}{\partial T_{1}^{-1}}$ $\displaystyle=-\frac{c_{m1}}{2\varepsilon^{\prime}\sqrt{\pi}}\int_{E_{\ast}}^{+\infty}\mathrm{d}E\,\sqrt{\frac{E-E_{\ast}}{2mE}}\,\frac{\partial}{\partial T_{1}^{-1}}\left(\frac{e^{-E/k_{B}T_{1}}}{\sqrt{k_{B}T_{1}}}\right)$ $\displaystyle=\frac{c_{m1}}{2\varepsilon^{\prime}k_{B}\sqrt{2\pi mk_{B}T_{1}}}\int_{E_{\ast}}^{+\infty}\mathrm{d}E\,\sqrt{\frac{E-E_{\ast}}{E}}\,\left(E-\frac{k_{B}T_{1}}{2}\right)\,{e^{-E/k_{B}T_{1}}}.$ We set $u_{\ast 1}:=\mfrac{E_{\ast}}{k_{B}T_{1}}$ and change the variable of integration from $E$ to $s:=\mfrac{E}{E_{\ast}}$ : (68) $\displaystyle L_{m\theta}$ $\displaystyle=\frac{c_{m1}E_{\ast}\sqrt{k_{B}T_{1}}}{4\varepsilon^{\prime}k_{B}\sqrt{2\pi m}}\int_{1}^{+\infty}\mathrm{d}s\,\sqrt{1-\frac{1}{s}}\,(2su_{\ast 1}-1)\,{e^{-su_{\ast 1}}}$ $\displaystyle=\alpha_{1}F(u_{\ast 1})$ $\displaystyle\alpha_{1}$ $\displaystyle:=\frac{c_{m1}E_{\ast}\sqrt{k_{B}T_{1}}}{4\varepsilon^{\prime}k_{B}\sqrt{2\pi m}}$ where the function $F$ is defined by : (69) $\displaystyle F(a):=\int_{1}^{+\infty}\mathrm{d}s\,\sqrt{1-\frac{1}{s}}\,(2as-1)e^{-as}.$ We can now compute the relative variations of $L_{m\theta}$ with respect to relative variations of temperature. Since $L_{m\theta}$ depends on $T_{1}$ through $E_{\ast}$ and $F(u_{\ast 1})$, we have : (70) $\displaystyle\frac{\partial\ln L_{m\theta}}{\partial\ln T_{1}}$ $\displaystyle=\frac{\partial\ln\alpha_{1}}{\partial\ln T_{1}}+\frac{\partial\ln F}{\partial\ln T_{1}}$ $\displaystyle=\frac{1}{2}+\frac{\partial\ln E_{\ast}}{\partial\ln T_{1}}+\frac{\partial\ln u_{\ast 1}}{\partial\ln T_{1}}\,\frac{\partial\ln F}{\partial\ln u_{\ast 1}}$ $\displaystyle=\frac{1}{2}+\frac{\partial\ln E_{\ast}}{\partial\ln T_{1}}+\left(\frac{\partial\ln E_{\ast}}{\partial\ln T_{1}}-1\right)\,\frac{\partial\ln F}{\partial\ln u_{\ast 1}}$ $\displaystyle=\frac{1}{2}+\frac{\partial\ln E_{\ast}}{\partial\ln T_{1}}\left(1+\frac{\partial\ln F}{\partial\ln u_{\ast 1}}\right)-\frac{\partial\ln F}{\partial\ln u_{\ast 1}}.$ $\mfrac{\partial\ln E_{\ast}}{\partial\ln T_{1}}$ can not be computed in the present model, because it depends on the microscopic details of the formation of water pores. However, we know that $E_{\ast}$ diminishes as $T_{1}$ increases, since the water pores become more frequent (and, probably, larger and more durable) when the ionic density fluctuations increase [14, 15]. Consequently, we have : (71) $\displaystyle\frac{\partial\ln E_{\ast}}{\partial\ln T_{1}}<0.$ In Appendix C, we prove that $1+\mfrac{\partial\ln F}{\partial\ln u_{\ast 1}}$ is slightly negative at high temperature. Since $\mfrac{\partial\ln E_{\ast}}{\partial\ln T_{1}}$ is also negative, we obtain the following estimate : (72) $\displaystyle\frac{\partial\ln L_{m\theta}}{\partial\ln T_{1}}\gtrsim\frac{3}{2}\qquad\text{at high temperature.}$ ### 6.3. Computation of $L_{mm}$ $L_{mm}$ is obtained by differentiating $J_{m}$ with respect to $X_{m}=\mfrac{\mu_{m1}-\mu_{m0}}{T_{0}}$ while keeping the other thermodynamical forces equal to zero : (73) $\displaystyle L_{mm}$ $\displaystyle=\left(\frac{\partial J_{m}}{\partial X_{m}}\right)_{(X_{\theta},X_{f},X_{w},X_{m^{\prime}})=0}$ $\displaystyle=T_{0}\left(\frac{\partial J_{m}}{\partial(\mu_{m1}-\mu_{m0})}\right)_{(X_{\theta},X_{f},X_{w},X_{m^{\prime}})=0}$ $\displaystyle=\frac{1}{k_{B}}\left(\frac{\partial J_{m}}{\partial\ln(a_{1}/a_{0})}\right)_{(X_{\theta},X_{f},X_{w},X_{m^{\prime}})=0}.$ In our model, based on the double well effective potential, the activities of the membrane molecules in each leaflet are equal to their respective concentrations. A more accurate model, taking into account the attractive interactions inside each leaflet, is necessary to improve this first approximation. Replacing $\mfrac{a_{1}}{a_{0}}$ by $\mfrac{c_{m1}}{c_{m0}}=\eta$, we obtain : (74) $\displaystyle L_{mm}=\frac{1}{k_{B}}\left(\frac{\partial J_{m}}{\partial\ln\eta}\right)_{T_{1}=T_{0}}.$ $J_{m}$ is a linear combination of the leaflet concentrations : (75) $\displaystyle J_{m}$ $\displaystyle=\zeta(T_{1})c_{m1}-\zeta(T_{0})c_{m0}$ $\displaystyle\zeta(T)$ $\displaystyle:=\frac{E_{\ast}e^{-u_{\ast}}}{2\varepsilon^{\prime}\sqrt{2\pi mk_{B}T}}\int_{0}^{+\infty}\mathrm{d}x\,e^{-u_{\ast}x}\sqrt{\frac{x}{x+1}}$ $\displaystyle u_{\ast}$ $\displaystyle:=\frac{E_{\ast}}{k_{B}T}.$ If the temperatures of both leaflets are equal, then $J_{m}$ is simply proportional to the difference of their concentrations : (76) $\displaystyle\big{(}J_{m}\big{)}_{T_{0}=T_{1}=T}=\zeta(T)(c_{m1}-c_{m0})$ and its derivative with respect to $\ln\eta$, while $c_{m0}$ is held fixed, is : (77) $\displaystyle\left(\frac{\partial J_{m}}{\partial\ln\eta}\right)_{c_{m0}=\text{cst.}}=\zeta(T_{1})c_{m1}=k_{B}L_{mm}.$ Since (78) $\displaystyle\int_{0}^{+\infty}\mathrm{d}x\,e^{-ax}\sqrt{\frac{x}{x+1}}=\frac{1}{a}-\frac{\ln(a)}{2}+\mathcal{O}(1)\qquad(a\to 0^{+})$ the high temperature expansion of $L_{mm}$ gives : (79) $\displaystyle\left(\frac{\partial\ln L_{mm}}{\partial\ln T}\right)_{T_{1}=T_{0}=T}=\left(\frac{\partial\ln\zeta}{\partial\ln T}\right)_{T_{1}=T_{0}=T}=\frac{1}{2}+o(1).$ ### 6.4. Estimation of $L_{\theta\theta}$ The heat diffusion coefficient, $L_{\theta\theta}$, depends only on the number of degrees of freedom that interact in the membrane bilayer. As long as the structure of the membrane is unchanged, the same hydrophobic tails interact similarly at any temperature. Therefore, we conjecture that $L_{\theta\theta}$ is independant of the temperature in the liquid disordered phase [27]. Therefore : (80) $\displaystyle\frac{\partial\ln L_{\theta\theta}}{\partial\ln T}\simeq 0.$ ### 6.5. Destabilisation Putting together the scaling laws for $L_{m\theta}$, $L_{mm}$ and $L_{\theta\theta}$, we obtain : (81) $\displaystyle\frac{\partial}{\partial\ln T_{1}}\left(\frac{L_{m\theta}^{2}}{L_{mm}L_{\theta\theta}}\right)=3-\frac{1}{2}-0=\frac{5}{2}$ The main mathematical proposition of the present article is the following. ###### Proposition 6.1. Since $\mfrac{L_{m\theta}^{2}}{L_{mm}L_{\theta\theta}}$ grows as $T_{1}^{5/2}$, the stability condition, $L_{m\theta}^{2}<L_{mm}L_{\theta\theta}$, can not be satisfied at high temperature. The exact value of $T_{1}$ for which this transition occurs can not be computed in our simple model, but the only characteristic temperature being $\mfrac{E_{\ast}}{k_{B}}$, the critical temperature must be of this order of magnitude. This destabilisation of the steady growth regime is comparable with the onset of heat convection in a fluid subject to a strong temperature gradient. In fine, the self-replication of protocells could be interpreted as a convective phenomenon inside their membrane, triggered by their metabolic activity. ## 7\. Conclusions and perspectives We have proposed a toy model of protocell growth, fission and reproduction. The scenario thus described can be viewed as the ancestor of mitosis. The main force driving this irreversible process is the temperature difference between the inside and the outside of the protocell, due to the inner chemical activity. We propose that the increase of the inner temperature, due to a rudimentary inner metabolism, enhances the transfer of membrane molecules from the inner leaflet to the outer leaflet, as described in silico by models of molecular dynamics [14, 15]. Due to this transfer of molecules, coupled to a heat transfer, the difference of their areas and the total mean curvature of the median surface increase. The cylindrical growth becomes unstable and any slight local reduction of the radius of the initial cylinder increases until the protocell is cut into two daughter protocells, each one containing reactants and catalysers to continue the growth and fission process. The cut occurs near the hottest zone, around the middle. This model is based on the idea [23] that the early forms of life were simple vesicles containing a particular network of chemical reactions, precursor of modern cellular metabolism : Protolife = Cellularity + Inner Metabolism. With a large supply of reactants in the so-called prebiotic soup [31, 16, 23], and with an optimal salinity and pH, these ingredients are sufficient to induce an exponential growth of prebiomass and make possible the exploration of a large number of chemical reactions in these miniature chemical factories. The possibility to sythesize complex molecules (sterols, RNA, DNA, proteins, etc.) comes later, once these factories self-replicate and thrive. In order to test our model experimentally, we have to manipulate vesicles that can be heated from within in a controled way. Let us imagine, in a solution maintained at temperature $T_{0}$, vesicles containing molecules of type $A$ able to absorb visible radiation, with which the surrounding molecules do not interact. Let us suppose that $A$ re-emits radiation in the near infrared. The heat thus generated inside the vesicle creates a controled temperature difference, $T_{1}-T_{0}>0$, between both sides of the membrane. If $L_{m\theta}$ is large enough, we should observe a bending of the membrane of the vesicles due to the transfer of the hottest molecules from the inner leaflet to the outer leaflet. Another experimental test of our model can be made by observing eukaryotic cells, where the mitochondria are the main source of heat. It seems possible to measure their temperature variations using fluorescent molecules [3]. Although the very notion of temperature at this scale and far from a thermodynamical equilibrium is not clear, the measurement of the temperature variations inside the cell during its life cycle could be correlated with the onset of mitosis and with the shape of mitochondrial network [21]. Our model is obviously oversimplified since the polar heads of membrane molecules are treated as an ideal gas in a box. In particular, we haven’t taken into account the interaction between these molecules and the surrounding solution. This calls for the development of a better model to treat the effect of these interactions on the temperature dependence of the conductance coefficients. The scaling law of the ratio $\mfrac{L_{m\theta}^{2}}{L_{mm}L_{\theta\theta}}$ at temperatures higher than $\mfrac{E_{\ast}}{k_{B}}$ is the key argument that explains the splitting of the protocell. Future investigations and experiments will decide of the plausibility of this proposition. Acknowledgments : We thank Jorgelindo Da Veiga Moreira (Université de Montréal), Marc Henry (Université de Strasbourg), Olivier Lafitte (Institut Galilée, Université Paris XIII), Kirone Mallick (Institut de Physique Théorique, CEA, Saclay), Laurent Schwartz (AP-HP) and Jean-Yves Trosset (SupBiotech, Villefuif) for their advice and helpful discussions. ## Appendix A The mean curvature of the membrane Let $\Sigma_{t}$ be a family of surfaces, indexed by a time parameter $t\in[t_{0},+\infty[$. We suppose that each $\Sigma_{t}$ is a smooth, orientable and closed (compact, without boundary) hence diffeomorphic to the standard $2$-sphere. At each point $P\in\Sigma_{t}$, the Taylor expansion of the distance from $Q\in\Sigma_{t}$ to the tangent plane, $T_{P}\Sigma_{t}$, defines a quadratic form whose eigenvalues (homogenous to the inverse of a length) do not depend on the coordinate system in the neighbourhood of $P$. We denote them $R_{-}$ and $R_{+}$. The mean curvature of $\Sigma_{t}$ at $P$ is the arithmetic mean of the principal curvatures : (82) $\displaystyle H:=\frac{1}{2}\left(\frac{1}{R_{+}}+\frac{1}{R_{-}}\right)$ and the gaussian curvature is their product : (83) $\displaystyle K:=\frac{1}{R_{+}R_{-}}.$ In the case of a cylinder, $R_{+}=+\infty$ and $R_{-}=R_{0}=$ its radius, hence $H_{\text{cyl.}}(P)=\mfrac{1}{2R_{0}}$ and $K(P)=0$ at every point $P\in\Sigma_{t}$ (except on the end hemispheres). Let $\Sigma_{t0}$ and $\Sigma_{t1}$ be the surfaces obtained by shifting $\Sigma_{t}$ in the normal direction, over an infinitesimal distance $\varepsilon$ on both sides of $\Sigma_{t}$. Let $\mathcal{A}_{0}(t)$ (resp. $\mathcal{A}_{1}(t)$) be the average area of the outer (resp. inner) layer of the membrane, measured at the hydrophilic heads, and $\mathcal{A}=\mfrac 12(\mathcal{A}_{0}+\mathcal{A}_{1})$ the average area of the median surface, where the hydrophobic tails join. The difference of their areas, $\mathcal{A}_{1}-\mathcal{A}_{0}$, is given by the first term of Weyl’s Tube Formula [13] : (84) $\displaystyle\mathcal{A}_{0}-\mathcal{A}_{1}=4\varepsilon\int_{\Sigma_{t}}H\,\mathrm{d}A+\mathcal{O}(\varepsilon^{2}).$ Let $\mathcal{B}$ be the infinitesimal variation of area along the outer normal : (85) $\displaystyle\mathcal{B}:=2\varepsilon\int_{\Sigma_{t}}H\,\mathrm{d}A.$ $\mathcal{A}_{0}$ and $\mathcal{A}_{1}$ can also be written as functions of $\mathcal{A}$ and $\mathcal{B}$ : (86) $\displaystyle\mathcal{A}_{0}=\mathcal{A}+\mathcal{B}\qquad\text{and}\qquad\mathcal{A}_{1}=\mathcal{A}-\mathcal{B}.$ Our dynamical variables are the area of the median surface, $\mathcal{A}(t)=\int_{\Sigma_{t}}\mathrm{d}A$, the volume of the cytosol, $\mathcal{V}(t)$, and the variation of area, $\mathcal{B}(t)=2\varepsilon\int_{\Sigma_{t}}H\,\mathrm{d}A$. In the next section, we will establish their evolution equations as a consequence of the balance equations for the number of membrane molecules. Remark : In the case of a cylinder of radius $R_{0}$, we have $H=\mfrac{1}{2R_{0}}$ and (87) $\displaystyle\mathcal{A}_{0}-\mathcal{A}_{1}=2\mathcal{B}=4\varepsilon H\mathcal{A}=\frac{2\varepsilon\mathcal{A}}{R_{0}}.$ Since $2\varepsilon\,\mfrac{\mathcal{A}_{0}+\mathcal{A}_{1}}{2}=2\varepsilon\mathcal{A}$ is also the volume, $v$, of this normal thickening of $\Sigma_{t}$, we have : (88) $\displaystyle v=\frac{2\mathcal{B}}{H}=4\mathcal{B}R_{0}.$ ## Appendix B Solutions of the growth equation In this appendix, we solve the growth equation, using basic linear algebra and standard results about linear differential equations [19]. The matrix form of the growth equation is (89) $\displaystyle\dot{X}=\begin{pmatrix}\dot{\mathcal{A}}_{0}\\\ \dot{\mathcal{A}}_{1}\\\ \dot{\mathcal{U}}\\\ \end{pmatrix}=\begin{pmatrix}\frac{\eta\tau}{2}&\frac{\eta\tau}{2}&0\\\ -\frac{\tau}{2}&\frac{2-\tau}{2}&0\\\ 0&-\frac{c_{m1}}{\varepsilon c_{mc}}&\frac{t_{1}}{t_{c}}\\\ \end{pmatrix}\begin{pmatrix}\mathcal{A}_{0}\\\ \mathcal{A}_{1}\\\ \mathcal{U}\\\ \end{pmatrix}=MX.$ As long as no flow vanishes, the determinant of $M$ is non-zero : (90) $\displaystyle\det(M)=\frac{\eta\tau t_{1}}{2t_{c}}=\frac{c_{m1}^{2}J_{m10}J_{rm}}{2c_{m0}c_{mc}J_{mc1}^{2}}$ and the protocell grows exponentially : (91) $\displaystyle X(t)=e^{\frac{t}{t_{1}}M}X(0).$ In general, the two leaflets of the membrane grow at different speeds. Indeed, the characteristic polynomial of $M$ is : (92) $\displaystyle\det(M-\lambda\,\mathrm{Id})$ $\displaystyle=\begin{vmatrix}\frac{\eta\tau}{2}-\lambda&\frac{\eta\tau}{2}&0\\\ -\frac{\tau}{2}&\frac{2-\tau}{2}-\lambda&0\\\ 0&-\frac{c_{m1}}{\varepsilon c_{mc}}&\frac{t_{1}}{t_{c}}-\lambda\\\ \end{vmatrix}$ $\displaystyle=\left(\lambda^{2}-\lambda\left(1+\frac{(\eta-1)\tau}{2}\right)+\frac{\eta\tau}{2}\right)\left(\frac{t_{1}}{t_{c}}-\lambda\right).$ Its roots are $\mfrac{t_{1}}{t_{c}}$ and the two roots, $\lambda_{\pm}(\eta,\tau)$, of the polynomial $\lambda^{2}-\lambda\big{(}1+(\eta-1)\mfrac{\tau}{2})\big{)}+\mfrac{\eta\tau}{2}$ : (93) $\displaystyle\lambda_{\pm}(\eta,\tau)$ $\displaystyle:=\frac{1}{2}\left(1+\frac{(\eta-1)\tau}{2}\pm\sqrt{\Delta(\eta,\tau)}\right)$ $\displaystyle\Delta(\eta,\tau)$ $\displaystyle:=\frac{1}{4}(\eta-1)^{2}\tau^{2}-(\eta+1)\tau+1.$ $\bullet$ If $\eta=1$, i.e. if both leaflets have the same density, then $\Delta$ is an affine function of $\tau$ : (94) $\displaystyle\Delta(1,\tau)=1-2\tau\qquad\text{and}\qquad\lambda_{\pm}(1,\tau)=\frac{1\pm\sqrt{1-2\tau}}{2}.$ If, moreover, $\tau=\mfrac 12$, i.e. the inner leaflet transmits half of the incoming membrane molecules to the outer leaflet, then (95) $\displaystyle\Delta\left(1,\frac{1}{2}\right)=0\qquad\text{and}\qquad\lambda_{\pm}\left(1,\frac{1}{2}\right)=\frac{1}{2}$ and both leaflets grow at the same speed. $\bullet$ If $\eta\neq 1$, then $\Delta(\eta,\tau)$ is a quadratic function of $\tau$, bounded from below, of discriminant (96) $\displaystyle\delta=(\eta+1)^{2}-(\eta-1)^{2}=4\eta>0$ and has distinct roots : (97) $\displaystyle\tau_{\pm}(\eta)=2\,\frac{\eta+1\pm\sqrt{4\eta}}{(\eta-1)^{2}}=2\left(\frac{\sqrt{\eta}\pm 1}{\eta-1}\right)^{2}=\frac{2}{(\sqrt{\eta}\mp 1)^{2}}.$ Physically, $\eta\simeq 1$ and $\tau\simeq\mfrac 12$. If $\eta=1+h$, with $0<h\ll 1$ then $\tau_{+}\simeq\mfrac{8}{h^{2}}\gg 1>\tau_{-}$ and (98) $\displaystyle\tau_{-}\simeq\frac{2}{\left(2+\frac{h}{2}\right)^{2}}\simeq\frac{1}{2}-\frac{h}{4}.$ Consequently, $\tau$ stays $<\tau_{-}$ (FIG. B). $\tau_{-}$$\mfrac 12$physicalregionnon-physical region$\Delta>0$$\Delta<0$$\Delta>0$$\tau_{+}\gg\tau_{-}$ Exponential growth necessitates to keep $\tau<\tau_{-}$. Mathematically, we have three possibilities : $\displaystyle(1)$ $\displaystyle\tau<\tau_{-}(\eta)\quad\text{or}\quad\tau>\tau_{+}(\eta)\ \to\ \Delta(\eta,\tau)>0$ $\displaystyle\quad\lambda_{+}(\eta,\tau)\neq\lambda_{-}(\eta,\tau)\quad\text{(real numbers) ;}$ $\displaystyle(2)$ $\displaystyle\tau=\tau_{-}(\eta)\quad\text{or}\quad\tau=\tau_{+}(\eta)\ \to\ \Delta(\eta,\tau_{\pm})=0$ $\displaystyle\quad\lambda_{+}\big{(}\eta,\tau_{\pm}(\eta)\big{)}=\lambda_{-}\big{(}\eta,\tau_{\pm}(\eta)\big{)}=\frac{\sqrt{\eta}}{\sqrt{\eta}\pm 1};$ $\displaystyle(3)$ $\displaystyle\tau_{-}(\eta)<\tau<\tau_{+}(\eta)\ \to\ \Delta(\eta,\tau)<0$ $\displaystyle\quad\lambda_{+}(\eta,\tau)=\overline{\lambda}_{-}(\eta,\tau)\quad\text{(complex numbers).}$ ### B.1. Case 1 : $\eta\neq 1$ and $\tau>\tau_{+}(\eta)$ or $\tau<\tau_{-}(\eta)$ In these intervals, $N$ is diagonalisable and a basis of eigenvectors of $N$ is given by : (99) $\displaystyle\mathcal{B}_{\pm}$ $\displaystyle=\begin{pmatrix}\frac{1}{2}\\\ \frac{\lambda_{\pm}(\eta,\tau)}{\eta\tau}-\frac{1}{2}\end{pmatrix}$ $\displaystyle=\frac{\mathcal{A}_{0}-\mathcal{A}_{1}}{2}+\frac{\lambda_{\pm}(\eta,\tau)}{\eta\tau}\,\mathcal{A}_{1}$ $\displaystyle=\mathcal{B}+\frac{\lambda_{\pm}(\eta,\tau)}{\eta\tau}\,\mathcal{A}_{1}$ i.e. $\mathcal{B}_{+}$ and $\mathcal{B}_{-}$ grow exponentially, with a rate of growth $\lambda_{\pm}/t_{1}$, respectively : (100) $\displaystyle\mathcal{B}_{\pm}(t)=\mathcal{B}_{\pm}(0)\,\exp\left(\frac{\lambda_{\pm}(\eta,\tau)}{t_{1}}\,t\right).$ The area of the inner leaflet is : (101) $\displaystyle\mathcal{A}_{1}(t)$ $\displaystyle=\frac{\mathcal{B}_{+}(t)-\mathcal{B}_{-}(t)}{\frac{\lambda_{+}}{\eta\tau}-\frac{\lambda_{-}}{\eta\tau}}$ $\displaystyle=\frac{\eta\tau}{\sqrt{\Delta}}\left(\mathcal{B}_{+}(0)e^{t\lambda_{+}/t_{1}}-\mathcal{B}_{-}(0)e^{t\lambda_{-}/t_{1}}\right).$ The area of the outer leaflet is : (102) $\displaystyle\mathcal{A}_{0}(t)$ $\displaystyle=\frac{(2\lambda_{+}-\eta\tau)\mathcal{B}_{-}(t)-(2\lambda_{-}-\eta\tau)\mathcal{B}_{+}(t)}{\lambda_{+}-\lambda_{-}}$ $\displaystyle=\frac{2\lambda_{+}-\eta\tau}{\sqrt{\Delta}}\mathcal{B}_{-}(0)e^{t\lambda_{-}/t_{1}}-\frac{2\lambda_{-}-\eta\tau}{\sqrt{\Delta}}\mathcal{B}_{+}(0)e^{t\lambda_{+}/t_{1}}.$ And $\mathcal{U}(t)$ is obtained from $\mathcal{A}_{1}(t)$ : (103) $\displaystyle e^{tt_{1}/t_{c}}\,\frac{\mathrm{d}}{\mathrm{d}t}\left(\mathcal{U}(t)e^{-tt_{1}/t_{c}}\right)$ $\displaystyle=-\frac{c_{m1}}{\varepsilon c_{mc}}\,\mathcal{A}_{1}(t)$ (104) $\displaystyle\mathcal{U}(t)\,e^{-tt_{1}/t_{c}}$ $\displaystyle=-\frac{\eta\tau c_{m1}\mathcal{B}_{+}(0)}{\varepsilon c_{mc}\sqrt{\Delta}\left(\frac{\lambda_{+}}{t_{1}}-\frac{t_{1}}{t_{c}}\right)}\,e^{t\,\left({\frac{\lambda_{+}}{t_{1}}-\frac{t_{1}}{t_{c}}}\right)}$ $\displaystyle+\frac{\eta\tau c_{m1}\mathcal{B}_{-}(0)}{\varepsilon c_{mc}\sqrt{\Delta}\left(\frac{\lambda_{-}}{t_{1}}-\frac{t_{1}}{t_{c}}\right)}\,e^{t\,\left({\frac{\lambda_{-}}{t_{1}}-\frac{t_{1}}{t_{c}}}\right)}+\text{cst.}$ $\displaystyle\mathcal{U}(t)$ $\displaystyle=\frac{\eta\tau c_{m1}}{\varepsilon c_{mc}\sqrt{\Delta}}\left(\frac{e^{t\lambda_{-}/t_{1}}}{\frac{\lambda_{-}}{t_{1}}-\frac{t_{1}}{t_{c}}}-\frac{e^{t\lambda_{+}/t_{1}}}{\frac{\lambda_{+}}{t_{1}}-\frac{t_{1}}{t_{c}}}\right)$ $\displaystyle\ +\text{cst.}\,e^{tt_{1}/t_{c}}$ where the integration constant is determined by $\mathcal{U}(0)$. ### B.2. Case 2 : $\eta\neq 1$ and $\tau\in\\{\tau_{+}(\eta),\tau_{-}(\eta)\\}$ In this singular case, the upper-left $2\times 2$ submatrix is not diagonalisable but conjugate to a lower triangular matrix of Jordan form : (105) $\displaystyle N:=\frac{1}{2}\begin{pmatrix}\eta\tau&\eta\tau\\\ -\tau&2-\tau\\\ \end{pmatrix}=T\begin{pmatrix}\Lambda_{\pm}(\eta)&0\\\ 1&\Lambda_{\pm}(\eta)\\\ \end{pmatrix}T^{-1}$ where $\Lambda_{\pm}(\eta)$ is the single eigenvalue of $N$ when $\tau$ is fixed equal to $\tau_{+}(\eta)$ or $\tau_{-}(\eta)$ : (106) $\displaystyle\Lambda_{\pm}(\eta)$ $\displaystyle:=\lambda\big{(}\eta,\tau_{\pm}(\eta)\big{)}=\frac{2+(\eta-1)\tau_{\pm}(\eta)}{4}$ $\displaystyle=\frac{2+\frac{2(\eta-1)}{(\sqrt{\eta}\mp 1)^{2}}}{4}=\frac{\eta\pm\sqrt{\eta}}{\eta-1}=\frac{\sqrt{\eta}}{\sqrt{\eta}\pm 1}.$ An easy computation gives us : (107) $\displaystyle\frac{2\Lambda_{+}}{\eta\tau}-1=\frac{1-3\sqrt{\eta}}{\eta+\sqrt{\eta}}\quad\text{and}\quad\frac{2\Lambda_{-}}{\eta\tau}-1=\frac{1+3\sqrt{\eta}}{\eta-\sqrt{\eta}}.$ $N$ has a unique proper line, generated by the vector (108) $\displaystyle\mathcal{B}_{\ast}^{\pm}$ $\displaystyle:=\begin{pmatrix}\frac{1}{2}\\\ \frac{\Lambda_{\pm}}{\eta\tau_{\pm}}-\frac{1}{2}\end{pmatrix}=\frac{1}{2}\begin{pmatrix}1\\\ \frac{1\mp 3\sqrt{\eta}}{\eta\pm\sqrt{\eta}}\end{pmatrix}$ $\displaystyle=\frac{\mathcal{A}_{0}}{2}+\left(\frac{1\mp 3\sqrt{\eta}}{\eta\pm\sqrt{\eta}}\right)\frac{\mathcal{A}_{1}}{2}$ where the lower $\ast$ means that $\lambda_{+}=\lambda_{-}$, whereas the upper $\pm$ depends on the choice between $\tau=\tau_{+}(\eta)$ and $\tau=\tau_{-}(\eta)$. Since (109) $\displaystyle\mathcal{B}_{\ast}^{\pm}(t)=\mathcal{B}_{\ast}^{\pm}(0)\,e^{t\Lambda_{\pm}/t_{1}}$ we obtain : (110) $\displaystyle\mathcal{A}_{0}(t)+\left(\frac{3\eta\mp 1}{\eta\pm\sqrt{\eta}}\right)\,\mathcal{A}_{1}(t)=2\,\mathcal{B}_{\ast}^{\pm}(0)\,e^{t\Lambda_{\pm}/t_{1}}.$ Since $\mathcal{B}_{\ast}^{\pm}=T{0\choose 1}$, the vector $\mathcal{B}_{\ast}^{\pm}$ is the right column of $T$. The left column of $T$ is the vector $\mathcal{B}_{\bullet}^{\pm}={x\choose y}$ which satisfies the equation $(N-\Lambda_{\pm})\mathcal{B}_{\bullet}^{\pm}=\mathcal{B}_{\ast}^{\pm}$, or in extenso : (111) $\displaystyle\left(\frac{\eta\tau}{2}-\Lambda_{\pm}\right)x+\left(\frac{\eta\tau}{2}\right)y$ $\displaystyle=\frac{1}{2}$ $\displaystyle-\frac{\tau}{2}\,x+\frac{2-\tau-2\Lambda_{\pm}}{2}\,y$ $\displaystyle=\frac{\Lambda_{\pm}}{\eta\tau}-\frac{1}{2}.$ Taking $x=0$ and $y=\mfrac{1}{\eta\tau}$ gives a solution : (112) $\displaystyle\big{(}N-\Lambda_{\pm}\big{)}\mathcal{B}_{\bullet}^{\pm}$ $\displaystyle=\begin{pmatrix}\frac{\eta\tau}{2}-\Lambda_{\pm}&\frac{\eta\tau}{2}\\\ -\frac{\tau}{2}&\frac{2-\tau}{2}-\Lambda_{\pm}\end{pmatrix}\begin{pmatrix}0\\\ \frac{1}{\eta\tau}\end{pmatrix}$ $\displaystyle=\begin{pmatrix}\frac{1}{2}\\\ \frac{2-\tau-2\Lambda_{\pm}}{\eta\tau}\end{pmatrix}=\begin{pmatrix}\frac{1}{2}\\\ \frac{2\Lambda_{\pm}-\eta\tau}{\eta\tau}\end{pmatrix}=\mathcal{B}_{\ast}^{\pm}.$ The matrix $T$ and its inverse, $T^{-1}$, are therefore : (113) $\displaystyle T=\begin{pmatrix}0&1\\\ \frac{1}{\eta\tau}&\frac{2\Lambda-\eta\tau}{2\eta\tau}\end{pmatrix}\qquad\text{and}\qquad T^{-1}=\begin{pmatrix}\frac{\eta\tau-2\Lambda}{2}&\eta\tau\\\ 1&0\end{pmatrix}.$ Since $\mathcal{B}_{\bullet}^{\pm}=\mfrac{\mathcal{A}_{1}}{\eta\tau_{\pm}(\eta)}$, we have $\mathcal{B}_{\bullet}^{\pm}(t)=\mathcal{B}_{\ast}(0)\,\mfrac{t}{t_{1}}\,e^{t\Lambda_{\pm}/t_{1}}$ and : (114) $\displaystyle\mathcal{A}_{1}(t)=\eta\tau_{\pm}(\eta)\,\mathcal{B}_{\ast}^{\pm}(0)\,\frac{t}{t_{1}}\,e^{t\Lambda_{\pm}/t_{1}}$ ### B.3. Case 3 : $\tau_{-}(\eta)<\tau<\tau_{+}(\eta)$ In this interval, $\Delta<0$ and $M$ has two distinct complex conjugated eigenvalues, $\lambda$ and $\overline{\lambda}$, functions of $\eta$ and $\tau$. Let $\alpha,\beta\in\mathbb{R}$ be the real and imaginary parts of $\lambda$ : (115) $\displaystyle\alpha$ $\displaystyle:=\frac{2+(\eta-1)\tau}{4}\,>0$ $\displaystyle\beta$ $\displaystyle:=\frac{\sqrt{-\Delta}}{2}\,>0$ $\displaystyle\lambda(\eta,\tau)$ $\displaystyle=\alpha+\mathbf{i}\,\beta\qquad(\mathbf{i}\,^{2}=-1).$ Let $V$ (resp. $\overline{V}$) be a complex eigenvector of $N$, of eigenvalue $\lambda$ (resp. $\overline{\lambda}$), for instance : (116) $\displaystyle V:=\mathcal{B}+\frac{\lambda}{\eta\tau}\,\mathcal{A}_{1}\qquad\text{and}\qquad\overline{V}:=\mathcal{B}+\frac{\overline{\lambda}}{\eta\tau}\,\mathcal{A}_{1}$ then the real and imaginary parts of $V$, defined by $V^{\prime}:=\mfrac 12(V+\overline{V})$ and $V^{\prime\prime}:=\mfrac{1}{2\mathbf{i}\,}(V-\overline{V})$, form a basis of $\mathbb{R}^{2}$ on which $N$ acts as an orthogonal matrix [19] : (117) $\displaystyle\frac{1}{2}\begin{pmatrix}\eta\tau&\eta\tau\\\ -\tau&2-\tau\\\ \end{pmatrix}$ $\displaystyle=U\begin{pmatrix}\alpha&\beta\\\ -\beta&\alpha\\\ \end{pmatrix}U^{-1}$ $\displaystyle V^{\prime}=U\begin{pmatrix}1\\\ 0\end{pmatrix}=\begin{pmatrix}\frac{1}{2}\\\ \frac{\alpha}{\eta\tau}-\frac{1}{2}\end{pmatrix}$ $\displaystyle\quad V^{\prime\prime}=U\begin{pmatrix}0\\\ 1\end{pmatrix}=\begin{pmatrix}0\\\ \frac{\beta}{\eta\tau}\end{pmatrix}$ i.e. the matrix $U$ has $V^{\prime}$ and $V^{\prime\prime}$ as columns : (118) $\displaystyle U=\begin{pmatrix}\frac{1}{2}&0\\\ \frac{\alpha}{\eta\tau}-\frac{1}{2}&\frac{\beta}{\eta\tau}\end{pmatrix}=\begin{pmatrix}\frac{1}{2}&0\\\ \frac{2-(\eta+1)\tau}{4\eta\tau}&\frac{\sqrt{-\Delta}}{2\eta\tau}\end{pmatrix}.$ Let $s=\mfrac{t}{t_{1}}$. Since our evolution operator, the exponential of $sN$, is : (119) $\displaystyle e^{sN}=e^{\alpha s}U\begin{pmatrix}\cos(\beta s)&\sin(\beta s)\\\ -\sin(\beta s)&\cos(\beta s)\end{pmatrix}U^{-1}$ we have : (120) $\displaystyle V^{\prime}(s)$ $\displaystyle=e^{sN}V^{\prime}(0)=e^{\alpha s}U\begin{pmatrix}\cos(\beta s)&\sin(\beta s)\\\ -\sin(\beta s)&\cos(\beta s)\end{pmatrix}\begin{pmatrix}\frac{1}{2}\\\ \frac{\alpha}{\eta\tau}-\frac{1}{2}\end{pmatrix}$ $\displaystyle=\frac{e^{\alpha s}}{2\eta\tau}\begin{pmatrix}\frac{1}{2}&0\\\ \frac{\alpha}{\eta\tau}-\frac{1}{2}&\frac{\beta}{\eta\tau}\end{pmatrix}\begin{pmatrix}\eta\tau\cos(\beta s)+(2\alpha-\eta\tau)\sin(\beta s)\\\ -\eta\tau\sin(\beta s)+(2\alpha-\eta\tau)\cos(\beta s)\end{pmatrix}$ $\displaystyle=\frac{e^{\alpha s}}{2\eta\tau}\begin{pmatrix}\frac{\eta\tau}{2}\cos(\beta s)+\frac{2\alpha-\eta\tau}{2}\sin(\beta s)\\\ \frac{(2\alpha-\eta\tau)(2\beta+\eta\tau)}{2\eta\tau}\cos(\beta s)+\left(\frac{(2\alpha-\eta\tau)^{2}}{2\eta\tau}-\beta\right)\sin(\beta s)\end{pmatrix}.$ Similarly, we have the expression of $V^{\prime\prime}(s)$ : (121) $\displaystyle V^{\prime\prime}(s)$ $\displaystyle=e^{sN}V^{\prime\prime}(0)$ $\displaystyle=\frac{\beta e^{\alpha s}}{\eta\tau}\begin{pmatrix}\frac{1}{2}\sin(\beta s)\\\ \left(\frac{2\alpha-\eta\tau}{2\eta\tau}\right)\sin(\beta s)+\frac{\beta}{\eta\tau}\cos(\beta s)\end{pmatrix}.$ Finally, $\mathcal{A}_{1}$ and $\mathcal{B}$ are obtained from $V^{\prime}$ and $V^{\prime\prime}$ by the linear relations : (122) $\displaystyle\mathcal{B}(s)$ $\displaystyle=\frac{\beta V^{\prime}(s)-\alpha V^{\prime\prime}(s)}{\beta-\alpha}$ $\displaystyle\mathcal{A}_{1}(s)$ $\displaystyle=\frac{\eta\tau}{\alpha-\beta}\big{(}V^{\prime}(s)-V^{\prime\prime}(s)\big{)}.$ ## Appendix C Smooth perturbation of cylindrical growth In this appendix, we compute the variation of the area and of the total mean curvature of a surface of revolution under a small variation of its generating curve. We will work in an orthonormal system of coordinates $(x,y,z)$. Let us suppose now that $\Sigma$ is a revolution surface whose generating curve, rotated around the axis $\\{y=0=z\\}$, is given by : (123) $\displaystyle\sqrt{y^{2}+z^{2}}=R(x)=R_{0}+\delta R(x)$ with $|\delta R(x)|\ll R_{0}$. The function $\delta R$ represents an infinitesimal normal perturbation around the cylindrical shape. The variable $x$ satisfies $0\leq x\leq\ell$ and the deformed cylinder is glued smoothly with two hemispherical caps of radius $R_{0}$. In other words, we suppose that (124) $\displaystyle\delta R(0)=\delta R(\ell)$ $\displaystyle=0$ $\displaystyle\delta R^{\prime}(0)=\delta R^{\prime}(\ell)$ $\displaystyle=0.$ Let us compute the variations of area, $\delta\mathcal{A}$, of length, $\delta\ell$, and of total mean curvature, $\delta\mathcal{H}$, for a fixed volume. ### C.1. Isovolumic variation of the area $\mathcal{A}$ is a functional of the length, $\ell$, the radius, $R$, and its derivative, $R^{\prime}$ : (125) $\displaystyle\mathcal{A}(\ell,R,R^{\prime})=\int_{0}^{\ell}\mathrm{d}x\,2\pi R\sqrt{1+R^{\prime 2}}.$ Its variation under infinitesimal changes of $\ell$ and $R$ is : (126) $\displaystyle\delta\mathcal{A}$ $\displaystyle=2\pi R_{0}\,\delta\ell$ $\displaystyle+2\pi\int_{0}^{\ell}\mathrm{d}x\,\delta R\left(\sqrt{1+R^{\prime 2}}-\frac{\mathrm{d}}{\mathrm{d}x}\left(\frac{RR^{\prime}}{\sqrt{1+R^{\prime 2}}}\right)\right).$ Since (127) $\displaystyle\frac{\mathrm{d}}{\mathrm{d}x}\left(\frac{RR^{\prime}}{\sqrt{1+R^{\prime 2}}}\right)$ $\displaystyle=\frac{RR^{\prime\prime}+R^{\prime 2}}{\sqrt{1+R^{\prime 2}}}-R^{\prime}R^{\prime\prime}\frac{RR^{\prime}}{\big{(}1+R^{\prime 2}\big{)}^{3/2}}$ $\displaystyle=(1+R^{\prime 2})^{-3/2}\left(\big{(}RR^{\prime\prime}+R^{\prime 2}\big{)}\big{(}1+R^{\prime 2}\big{)}-RR^{\prime 2}R^{\prime\prime}\right)$ $\displaystyle=(1+R^{\prime 2})^{-3/2}\big{(}RR^{\prime\prime}+R^{\prime 2}+R^{\prime 4}\big{)}$ we obtain (128) $\displaystyle\delta\mathcal{A}$ $\displaystyle=2\pi R_{0}\,\delta\ell$ $\displaystyle+2\pi\int_{0}^{\ell}\mathrm{d}x\,\delta R\big{(}1+R^{\prime 2}\big{)}^{-3/2}\big{(}1-RR^{\prime\prime}-R^{\prime 4}\big{)}.$ Similarly, the volume, $\mathcal{V}$, is a functional of $\ell$ and $R$ : (129) $\displaystyle\mathcal{V}(\ell,R)=\frac{4\pi R_{0}^{3}}{3}+\int_{0}^{\ell}\mathrm{d}x\,\pi R^{2}$ and its variation under infinitesimal changes of $\ell$ and $R$ is : (130) $\displaystyle\delta\mathcal{V}=\pi R_{0}^{2}\,\delta\ell+2\pi\int_{0}^{\ell}\mathrm{d}x\,R\,\delta R.$ If $\mathcal{V}$ is held constant, then $\delta\mathcal{V}=0$ and : (131) $\displaystyle\big{(}\delta\ell\big{)}_{\mathcal{V}=\text{cst.}}=-\frac{2}{R_{0}^{2}}\int_{0}^{\ell}\mathrm{d}x\,R\,\delta R.$ Inserting this expression of $\delta\ell$ into that of $\delta\mathcal{A}$, we obtain the isovolumic variation of area : (132) $\displaystyle(\delta\mathcal{A})_{\mathcal{V}=\text{cst.}}$ $\displaystyle=2\pi\int_{0}^{\ell}\mathrm{d}x\,\delta R\left(\big{(}1+R^{\prime 2}\big{)}^{-3/2}\big{(}1-RR^{\prime\prime}-R^{\prime 4}\big{)}-\frac{2R}{R_{0}}\right).$ ###### Theorem C.1. The isovolumic variational derivatives of the length and of the area of a (nearly cylindrical) closed revolution surface are negative : (133) $\displaystyle\left(\frac{\delta\ell}{\delta R}\right)_{\mathcal{V}=\text{cst.}}<0\qquad\text{and}\qquad\left(\frac{\delta\mathcal{A}}{\delta R}\right)_{\mathcal{V}=\text{cst.}}<0.$ In other words, since the stuffing is incompressible whereas the gut is elastic, the length and the area of a squeezed sausage increase. We call this simple statement the Squeezed Sausage Theorem (SST). ### C.2. Isovolumic variation of the total mean curvature The circles $\\{x=\text{cst.}\\}$ and the meridians, obtained by rotating the generating curve of equation $z^{2}=R^{2}(x)$, form an orthogonal system of geodesics [7], and the mean curvature of $\Sigma$ is given by : (134) $\displaystyle H=\frac{1}{2}\left(\frac{1}{R\sqrt{1+R^{\prime 2}}}+\frac{R^{\prime\prime}}{\big{(}1+R^{\prime 2}\big{)}^{3/2}}\right).$ The lateral area of a slice of width $\mathrm{d}x$, perpendicular to the axis of the surface, is : (135) $\displaystyle\mathrm{d}A=2\pi R\sqrt{1+R^{\prime 2}}\,\mathrm{d}x$ and the total mean curvature is : (136) $\displaystyle\mathcal{H}$ $\displaystyle:=\int_{\Sigma}H\,\mathrm{d}A$ $\displaystyle=\int_{\text{caps}}H\,\mathrm{d}A+\int_{0}^{\ell}2\pi R\sqrt{1+R^{\prime 2}}\,\mathrm{d}x$ $\displaystyle=4\pi R_{0}^{2}\cdot\frac{1}{R_{0}}+2\pi\int_{0}^{\ell}\frac{1}{2}\left(1+\frac{R\,R^{\prime\prime}}{1+R^{\prime 2}}\right)\mathrm{d}x$ $\displaystyle=4\pi R_{0}+\pi\ell+\pi\int_{0}^{\ell}\frac{R\,R^{\prime\prime}}{1+R^{\prime 2}}\,\mathrm{d}x.$ Since $\mathcal{H}$ is a functional of $\ell$, $R$, $R^{\prime}$ and $R^{\prime\prime}$, its variation under a change $\delta R$ of the radius of gyration and a change of length $\delta\ell$, is obtained after a double integration by parts [10] : (137) $\displaystyle\delta\mathcal{H}$ $\displaystyle=\pi\delta\ell+\pi\int_{0}^{\ell}\mathrm{d}x\ \frac{R^{\prime\prime}}{1+R^{\prime 2}}+\frac{\mathrm{d}}{\mathrm{d}x}\left(\frac{2RR^{\prime}R^{\prime\prime}}{\big{(}1+R^{\prime 2}\big{)}^{2}}\right)$ $\displaystyle\hskip 85.35826pt+\frac{\mathrm{d}^{2}}{\mathrm{d}x^{2}}\left(\frac{R}{1+R^{\prime 2}}\right)\,\delta R.$ Instead of computing each term of the integrand, let us make the approximation $R^{\prime 2}\ll 1$, valid when the initial cylinder is only slightly deformed. The expression of $\delta\mathcal{H}$ then simplifies to (138) $\displaystyle\delta\mathcal{H}$ $\displaystyle\simeq\pi\delta\ell+\pi\delta\int_{0}^{\ell}\mathrm{d}x\,RR^{\prime\prime}$ $\displaystyle\simeq\pi\delta\ell+2\pi\int_{0}^{\ell}\mathrm{d}x\,R^{\prime\prime}\,\delta R.$ Using the expression of $\delta\ell=-\mfrac{2}{R_{0}^{2}}\int_{0}^{\ell}\mathrm{d}x\,R\,\delta R$ when $\mathcal{V}$ is held constant, we obtain : (139) $\displaystyle\big{(}\delta\mathcal{H}\big{)}_{\mathcal{V}=\text{cst.}}\simeq 2\pi\int_{0}^{\ell}\mathrm{d}x\left(R^{\prime\prime}-\frac{R}{R_{0}^{2}}\right)\,\delta R.$ As long as $R_{0}^{2}|R^{\prime\prime}|\ll R$, the isovolumic variational derivative of $\mathcal{H}$ with respect to $R$ is negative : (140) $\displaystyle\left(\frac{\delta\mathcal{H}}{\delta R}\right)_{\mathcal{V}=\text{cst.}}<0\qquad\text{if}\quad R^{\prime 2}\ll 1\quad\text{and}\quad R_{0}^{2}|R^{\prime\prime}|\ll R.$ When $\delta R$ approaches $-R_{0}$ and the protocell is ready to split, the two radii of curvature are small compared to $R_{0}$ but have opposite sign, hence the Gaussian curvature around the septum is large and negative. After the cut, when the two caps are formed, the mean curvature and the Gaussian curvature are positive again. ## Appendix D Asymptotic expansion of $F(a)$ The change of variable $t=\sqrt{a(s-1)}$ in the integral defining $F$ gives us : (141) $\displaystyle F(a)$ $\displaystyle=\frac{e^{-a}}{a}\int_{0}^{+\infty}\mathrm{d}t\,f(a,t)$ $\displaystyle f(a,t)$ $\displaystyle:=2t^{2}\,e^{-t^{2}}\left(2\sqrt{t^{2}+a}-\frac{1}{\sqrt{t^{2}+a}}\right).$ Let (142) $\displaystyle G(a):=\int_{0}^{+\infty}\mathrm{d}t\,f(a,t)=ae^{a}F(a).$ The function $f(0,\cdot)$ is integrable over the half line $[0,+\infty[$ and (143) $\displaystyle G(0)$ $\displaystyle=\int_{0}^{+\infty}\mathrm{d}t\,f(0,t)$ $\displaystyle=\int_{0}^{+\infty}\mathrm{d}t\,2t\,e^{-t^{2}}(2t^{2}-1)$ $\displaystyle=\int_{0}^{+\infty}\mathrm{d}u\,e^{-u}(2u-1)=1.$ Let us compute the asymptotic expansion of $G(a)$ when $a\to 0^{+}$ : (144) $\displaystyle{\,}G(a)-G(0)$ $\displaystyle=\int_{0}^{+\infty}\mathrm{d}t\,\big{(}f(a,t)-f(0,t)\big{)}$ $\displaystyle=2\int_{0}^{+\infty}\mathrm{d}t\,t^{2}e^{-t^{2}}\left(2\big{(}\sqrt{t^{2}+a}-t\big{)}-\left(\frac{1}{\sqrt{t^{2}+a}}-\frac{1}{t}\right)\right)$ $\displaystyle=\int_{0}^{+\infty}2t\,\mathrm{d}t\,e^{-t^{2}}\big{(}\sqrt{t^{2}+a}-t\big{)}\left(2t+\frac{1}{\sqrt{t^{2}+a}}\right)$ $\displaystyle=\int_{0}^{+\infty}\mathrm{d}u\,e^{-u}\big{(}\sqrt{u+a}-\sqrt{u}\big{)}\left(2\sqrt{u}+\frac{1}{\sqrt{u+a}}\right)$ $\displaystyle=2\int_{0}^{+\infty}\mathrm{d}u\,e^{-u}\sqrt{u(u+a)}+\int_{0}^{+\infty}\mathrm{d}u\,e^{-u}(1-2u)-\int_{0}^{+\infty}\mathrm{d}u\,e^{-u}\sqrt{\frac{u}{u+a}}$ $\displaystyle=2\int_{0}^{+\infty}\mathrm{d}u\,e^{-u}\sqrt{u(u+a)}-1-\int_{0}^{+\infty}\mathrm{d}u\,e^{-u}\sqrt{\frac{u}{u+a}}.$ Hence : (145) $\displaystyle G(a)=\varphi(a)-\varphi^{\prime}(a)$ where (146) $\displaystyle\varphi(a)$ $\displaystyle:=2\int_{0}^{+\infty}\mathrm{d}u\,e^{-u}\sqrt{u(u+a)}$ $\displaystyle=2a^{2}\int_{0}^{+\infty}\mathrm{d}x\,e^{-ax}\sqrt{x(x+1)}.$ $\varphi(a)$ being the Laplace transform of the function $x\mapsto 2a^{2}\sqrt{x(x+1)}$, its expansion as $0^{+}$ is given by integrating the expansion of $\sqrt{x(x+1)}$ at $+\infty$ term by term : (147) $\displaystyle\sqrt{x(x+1)}$ $\displaystyle=x+\frac{1}{2}-\frac{1}{8x}+\mathcal{O}(x^{-2})$ $\displaystyle\varphi(a)$ $\displaystyle=2a^{2}\left(\frac{1}{a^{2}}+\frac{1}{2a}-\frac{1}{8}\int_{1}^{+\infty}\mathrm{d}x\,\frac{e^{-ax}}{x}+\mathcal{O}(1)\right)$ $\displaystyle=2+a-\frac{a^{2}}{4}\ln(a)+\mathcal{O}(a^{2}).$ Similarly, for $\varphi^{\prime}(a)$ we have : (148) $\displaystyle\sqrt{\frac{x}{x+1}}$ $\displaystyle=1-\frac{1}{2x}+\frac{3}{8x^{2}}+\mathcal{O}(x^{-3})\qquad(x\to+\infty)$ $\displaystyle\varphi^{\prime}(a)$ $\displaystyle=a\int_{0}^{1}\mathrm{d}x\,e^{-ax}\sqrt{\frac{x}{x+1}}$ $\displaystyle+a\int_{1}^{+\infty}\mathrm{d}x\,e^{-ax}\sqrt{\frac{x}{x+1}}$ $\displaystyle=a\int_{0}^{1}+a\left(\frac{e^{-a}}{a}-\frac{1}{2}\int_{1}^{+\infty}\mathrm{d}x\,\frac{e^{-ax}}{x}+\mathcal{O}(1)\right)$ $\displaystyle=1-\frac{a\ln(a)}{2}+\mathcal{O}(a).$ Consequently : (149) $\displaystyle G(a)$ $\displaystyle=1+\frac{a\ln(a)}{2}+\mathcal{O}(a)$ and (150) $\displaystyle F(a)$ $\displaystyle=\frac{e^{-a}}{a}+\frac{e^{-a}\ln(a)}{2}+\mathcal{O}(a)$ $\displaystyle=\frac{1}{a}+\frac{1}{2}\ln(a)+o(1).$ The asymptotic expansion of $\frac{\partial\ln F}{\partial\ln a}$ is therefore : (151) $\displaystyle\frac{\partial\ln F}{\partial\ln a}=-1+\frac{a\ln(a)}{2}+\mathcal{O}(a).$ In particular, since $a\ln(a)<0$ for $0<a<1$, we have (152) $\displaystyle\frac{\partial\ln F}{\partial\ln a}<-1\qquad(a\to 0^{+}).$ ## References * [1] J. S. Allhusen & J. C. Conboy : The Ins and Outs of Lipid Flip-Flop (Acc. Chem. Res., 2017, 50, 1, 58-65). * [2] P. W. Atkins & J. de Paula : Physical chemistry (Oxford University Press, 11th edition, 2017). * [3] D. Chrétien et al. : Mitochondria are physiologically maintained at close to 50 $\circ$C (PLoS Biol 16(1): e2003992. https://doi.org/ 10.1371/journal.pbio.2003992 ; January 25, 2018) * [4] F.-X. Contreras, L. Sánchez-Magraner, A. Alonso, F. M. Goñi : Transbilayer (flip-flop) lipid motion and lipid scrambling in membranes (FEBS Letters, 584 (2010) 1779–1786). * [5] S. Cooper : Bacterial growth and division (Academic Press, 1991). * [6] J. da Veiga Moreira, S. Peres, J.-M. Steyaert, E. Bigan, L. Paulevé, M.-L. Nogueira & L. Schwartz : Cell cycle progression is regulated by intertwined redox oscillators. (Theoretical Biology and Medical Modelling, 12(1), 1-14, 2015). * [7] M. Do Carmo : Differential Geometry of Curves and Surfaces (Prentice hall, 1976). * [8] J. England : Statistical physics of self-replication (J. Chem. Phys., 139 , 121923 (2013)). * [9] A. Erdélyi : Asymptotic expansions (Dover Publications, 1956). * [10] I. M. Gel’fand & S. V. Fomin : Calculus of variations (Prentice-Hall, 1963). * [11] P. Glansdorff & I. Prigogine : Thermodynamic theory of structure, stability and fluctuations (Wiley Interscience, New York, 1971). * [12] G. Gottschalk : Bacterial metabolism (2nd edition, Springer-Verlag, 1986). * [13] A. Gray : Tubes (2nd edition, Birkhäuser, 2004). * [14] A. A. Gurtovenko & I. Vattulainen : Ion Leakage through Transient Water Pores in Protein-free Lipid Membranes Driven by Transmembrane Ionic Charge Imbalance (Biophys. J. , 92, March 2007, 1878-1890). * [15] A. A. Gurtovenko & I. Vattulainen : Molecular Mechanism for Lipid Flip-Flops (J. Phys. Chem. B 2007, 111, 13554-13559). * [16] J. B. S. Haldane : The origin of Life (1929). * [17] M. Henry : Thermodynamics of life (Unpublished notes). * [18] T. L. Hill : Free energy transduction and biochemical cycle kinetics (Dover Publications, 2005). * [19] M. W. Hirsch, S. Smale & R. L. Devaney : Differential Equations, Dynamical Systems and an Introduction to Chaos (Academic Press, 3rd edition, 2013). * [20] A. Katchalsky & P. F. Curran : Nonequilibrium Thermodynamics in Biophysics (Harvard University Press, 1965). * [21] K. Mitra, C. Wunder, B. Roysam, G. Lin & J. Lippincott-Schwartz : A hyperfused mitochondrial state achieved at G1-S regulates cyclin E buildup and entry into S phase (PNAS, 2009, 106, 29, 11960-11965). * [22] H. J. Morowitz : Energy Flow in Biology (Academic Press, 1968). * [23] H. J. Morowitz : Beginnings of Cellular Life (Yale University Press, 1992). * [24] H. J. Morowitz & E. Smith : The origin and nature of life on Earth (Cambridge University Press, 2016). * [25] H. J. Morowitz & E. Smith : Universality in intermediate metabolism (PNAS, vol. 101, n∘ 36, 2004, 13168-13173). * [26] H. J. Morowitz & E. Smith : Energy flow and the organization of life (2007). * [27] O. Mouritsen : Life as a matter of fat (2nd edition, Springer-Verlag, 2016). * [28] P. H. Nielsen, C. Kragelund, R. J. Seviour & J. Lund Nielsen : Identity and ecophysiology of filamentous bacteria in activated sludge (FEMS Microbiol. Rev. 33 (2009) 969-998). * [29] L. Onsager : Reciprocal relations in irreversible processes. I (Phys. Rev., 37, pp. 405-426, 1931). * [30] L. Onsager : Reciprocal relations in irreversible processes. II (Phys. Rev., 38, pp. 2265-2279, 1931). * [31] I. Oparin : The Origin of Life on Earth (Oliver and Boyd, Edinburgh, 1957). * [32] I. Prigogine : Introduction to thermodynamics of irreversible processes (John Wiley and Sons, 1962). * [33] N. Rashevsky : Mathematical biophysics, vol. 1 (Dover Publications, 3rd edition, 1960). * [34] S. Rasmussen et al., editors : Protocells. Bridging Nonliving and Living Matter (MIT Press, 2009). * [35] L. Reichl : A Modern Course in Statistical Physics (Arnold, 1980). * [36] M. Salazar-Roa & M. Malumbres : Fueling the cell division cycle (Trends Cell Biol. 2017 Jan. ; 27(1):69-81). * [37] E. Schrödinger : What is life ? (Cambridge University Press, 1945). * [38] S. M. Stanley : Exploring Earth and Life through Time (W. H. Freeman, 1993).
 # Quantum lock-in detection of a vector light shift Kosuke Shibata<EMAIL_ADDRESS>Naota Sekiguchi Takuya Hirano Department of Physics, Gakushuin University, Tokyo, Japan ###### Abstract We demonstrate detection of a vector light shift (VLS) using the quantum lock- in method. The method offers precise and accurate VLS measurement without being affected by real magnetic field fluctuations. We detect a VLS on a Bose–Einstein condensate (BEC) of 87Rb atoms caused by an optical trap beam with a resolution less than 1 Hz. We also demonstrate elimination of a VLS by controlling the beam polarization to realize a long coherence time of a transversally polarized $F$ = 2 BEC. Quantum lock-in VLS detection should find wide application, including the study of spinor BECs, electric-dipole moment searches, and precise magnetometry. ## I Introduction The a.c. Stark shift or light shift plays significant roles in atomic physics. One example is the optical trap Grimm et al. (2000), which has been extensively used in cold atom experiments and has been the subject of intriguing and important research, including low-dimensional Görlitz et al. (2001) and uniform gases Gaunt et al. (2013), and atoms in an optical lattice with applications to quantum simulation Bloch et al. (2008) and atomic clocks Derevianko and Katori (2011); Katori (2011). It has also enabled the study of multi-component gases and, in particular, spinor Bose–Einstein condensates (BECs) Stamper-Kurn and Ueda (2013). The light shift has vector and tensor components and hence is state-dependent in general Deutsch and Jessen (1998); Geremia et al. (2006); Deutsch and Jessen (2010). The state dependence has been exploited for realizing state- selective transport Mandel et al. (2003a, b) and confinement Heinz et al. (2020). However, a state-dependent shift is often undesirable for situations in which well-controlled spin evolution is required. Escaping from a vector light shift (VLS), which is equivalent to a fictitious magnetic field, has been an important issue in precise measurements, such as the search for an atomic electric-dipole moment Romalis and Fortson (1999) and exotic spin- dependent interactions Jackson Kimball et al. (2017a). Reducing the VLS is also important in atomic magnetometers, in which the VLS introduces systematic errors. The quantum noise associated with the light shift due to the probe field ultimately limits the sensitivity Fleischhauer et al. (2000). The VLS restricts the potential use of optically trapped atoms for magnetically sensitive experiments. While its effect can be diminished by applying a bias magnetic field in a direction orthogonal to the wavevector, the VLS can still be a significant noise source in precise measurements Romalis and Fortson (1999). It is necessary to reduce the VLS when an ultralow magnetic field is required. In addition, the relative direction cannot be chosen satisfactorily in some situations, such as in 3D optical lattice experiments. In order to eliminate the VLS caused by optical trapping beams, the light polarization should be precisely controlled, because the VLS is proportional to the intensity of a circularly polarized component Grimm et al. (2000); Deutsch and Jessen (1998); Geremia et al. (2006); Deutsch and Jessen (2010). However, it is a formidable task to precisely extinguish the circular component at the atom position located in a vacuum cell. Polarization measurements and control outside the cell do not assure the degree of linear polarization due to the stress-induced birefringence of the vacuum windows Jellison (1999). Therefore, a sensitive and robust polarization measurement method using atoms themselves as a probe is important. Most effective polarization measurements are accomplished by using atoms themselves as a probe. Polarization measurements with an atomic gas have been performed with various methods including Larmor precession measurement Zhu et al. (2013), precise microwave spectroscopy Steffen et al. (2013), and frequency modulation nonlinear magneto-optical rotation Jackson Kimball et al. (2017b). Differential Ramsey interferometry has been developed for spinor condensates Wood et al. (2016). Polarization measurements by fluorescence detection have been recently demonstrated for ions Yuan et al. (2019). In this paper, we demonstrate VLS detection by applying the quantum lock-in method Kotler et al. (2011); de Lange et al. (2011). The measurement is immune to environmental magnetic field noise, and thus achieves excellent precision and accuracy. We detect a VLS induced by an optical trap beam on a BEC of 87Rb atoms with a resolution less than 1 Hz. This detection method is feasible to implement and should have wide applications in various research areas involved with optical fields. The paper is organized as follows. In Sec. II, our experimental method and setup are presented. The experimental results are described in Sec. III. We discuss the applications and potential performance of the quantum lock-in VLS detection in Sec. IV. We conclude the paper in Sec. V. ## II Experimental method and setup We produce a BEC in a vacuum glass cell. A BEC of $3\times 10^{5}$ atoms in the hyperfine spin $F=2$ state is trapped in a crossed optical trap. The trap consists of an axial beam at the wavelength of 850 nm and a radial beam at 1064 nm. The axial and radial beam waists are $\approx$ 30 $\mu$m and 70 $\mu$m, respectively. A magnetic bias field $B$ of 15 $\mu$T is applied along the axial beam to define the quantization axis, as shown in Fig. 1(a). The atoms are initially in the $|F,m_{F}\rangle=|2,2\rangle$ state, where $m_{F}$ denotes the magnetic sublevel. The ellipticity of the axial beam at the atomic position is controlled with a quarter waveplate (QWP) in the VLS measurement described below. The QWP is located between a polarization beam splitter for polarization cleaning and the cell. The angle of the QWP is adjusted with a precise manual rotation stage. The minimum scale of the rotation stage is 0.28 mrad. Figure 1: (color online) (a) Experimental configuration. A BEC is trapped in the axial trap beam along the $z$ axis and the radial trap beam along the $x$ axis (not shown). (b) Typical TOF image of a BEC measured after rf pulses for the detection. The spin components ($m_{F}=-2,-1,0,1,2$) are spatially resolved by the Stern–Gerlach method. (c) Time sequence for the quantum lock- in detection of a VLS. The beam power, $P(t)$, is modulated with a frequency $\omega_{m}$. The phase of the spin vector evolves with an angular frequency of $\omega(t)$. The accumulated phase, $\Phi=\int_{0}^{T}\omega(t)dt$, is finally measured. The time sequence for the quantum lock-in detection of a VLS is shown in Fig. 1(c). The lock-in technique enables enhanced sensitivity at the modulation frequency while reducing the effect of unwanted noise. We measure a VLS induced by the axial optical trap beam with multiple rf pulses. The trap beam power, $P(t)$, is modulated with a frequency $\omega_{m}$ during the pulse application as $P(t)=P_{0}\left(1+p\sin(\omega_{m}t)\right)\equiv P_{0}+P_{1}\sin(\omega_{m}t),$ (1) where $P_{0}$ is the mean power and $p$ is the modulation index. $P_{1}$ can be negative by changing the modulation phase by $\pi$. $\omega_{m}$ is set to be sufficiently higher than twice the trapping frequency to avoid parametric heating of the atoms. The modulation generates an a.c. fictitious magnetic field to be measured, given by $B_{\mathrm{fic}}=-\frac{1}{4}\alpha^{(\mathrm{1})}\mathcal{C}I_{1}\sin(\omega_{m}t)\equiv B_{1}\sin(\omega_{m}t),$ (2) where $\alpha^{(\mathrm{1})}$ is the a.c. vector polarizability, $\mathcal{C}$ is the degree of the circularity and $I_{1}$ is the beam intensity corresponding to $P_{1}$. The pulse set consists of an initial $\pi/2$ pulse at $t=0$, an odd number ($N$) of $\pi$-pulses, and a readout $\pi/2$ pulse. The pulses are equally spaced by $\Delta T$. The spacing satisfies $\omega_{m}$$=\pi/\Delta T$ so that the evolved phase due to the fictitious field is constructively accumulated. The relative phase, $\Delta\varphi$, between the initial and read-out pulses is set to $\pi/2$ for maximum sensitivity to small changes in the accumulated phase, $\Phi$. $\Phi$ is explicitly given by $\Phi=\frac{2}{\pi}\frac{g_{F}\mu_{B}B_{1}}{\hbar}T\equiv\frac{2}{\pi}\omega_{1}T,$ (3) where $g_{F}$ is the Landé g-factor, $\mu_{B}$ is the Bohr magneton, $\hbar$ is the reduced Planck constant and $T=(N+1)\Delta T$ is the phase accumulation time. $\omega_{1}/(2\pi)$ represents the VLS corresponding to $B_{1}$ in units of frequency. The read-out pulse converts $\Phi$ into the magnetization, $m$, as $m\equiv\frac{\sum_{i}iN_{i}}{N_{\mathrm{tot}}}=\mathcal{V}F\sin\Phi,$ (4) where $N_{i}$ is the atom number in the $|F,m_{F}=i\rangle$ state $(i=-2,-1,0,1,2)$ after the read-out pulse, $N_{\mathrm{tot}}=\sum_{i}N_{i}$ is the total atom number, and $\mathcal{V}$ is the visibility. $\mathcal{V}$ is ideally $1$, but in practice it is less than $1$ due to magnetic field noise Kotler et al. (2011). Imperfections in the initial state preparation and spin manipulation also decrease $\mathcal{V}$. The magnetization is measured by standard absorption imaging after a time-of-flight with Stern–Gerlach spin separation (see Fig. 1(b)). ## III Results Figure 2: (color online) Detection of the VLS. The error bars represent the sample standard deviation. The red solid line is the fitting curve by $\mathcal{V}F\sin(ap)$. The right axis represents $B_{1}$. It should be noted that the right axis scale is not linear since $B_{1}$ is proportional to $\arcsin(\frac{m}{\mathcal{V}F})$. We first confirm the validity of the detection scheme. We perform a lock-in detection with $\omega_{m}=2\pi\times 2$ kHz ($\Delta T=0.25$ ms) and $N=27$, and hence $T=7$ ms. $P_{0}$ is fixed to 11 mW. The change in $m$ is observed as $p$ is varied. The result is plotted in Fig. 2. Here, the angle of the QWP axis, $\theta$, is approximately 4∘ apart from the optimal angle, $\theta^{*}$, minimizing the VLS. The experimental determination of $\theta^{*}$ is described below. $m$ is well fitted by a sinusoidal function $\mathcal{V}F\sin(ap)$, indicating the VLS was successfully detected. The visibility in this detection setting is found to be $\mathcal{V}=0.746(42)$ from an independent measurement with no modulation ($p=0$) where $\Delta\varphi$ is scanned. The detection is used to minimize the VLS. We control the VLS by changing $\theta$ with $p$ fixed to 0.32. The $\theta$-dependence of $m$ is shown in Fig. 3(a). Because $\mathcal{C}\approx\sin 2(\theta-\theta^{*})\equiv\sin 2\Delta\theta$ when the birefringence in the optical path is small Wood et al. (2016) and $|\Delta\theta|\ll 1$, we fit $m$ by $\mathcal{V}F\sin(\beta_{1}(\theta-\theta^{*}))$. The fit gives $\beta_{1}=6.2(2)$, which is in reasonable agreement with the calculation. $\theta^{*}$ is found to be -6.6(1.1) mrad. The VLS resolution is evaluated as $\delta\omega=\beta_{1}\delta\theta^{*}/T=2\pi\times 0.16$ Hz, where $\delta\theta^{*}$ is the uncertainty in the $\theta^{*}$ estimation. Figure 3: (color online) Polarization dependence of the signal. (a) Measurement result with $T$ = 7 ms. (b) Measurement result with $T$ = 28.2 ms. The blue circles and red squares represent $m_{+}$ and $m_{-}$, respectively. (c) $\Delta m$ as a function of $\theta$. $\Delta B_{1}$ is the difference between the fictitious magnetic fields for positive and negative $P_{1}$. The solid lines in (a) and (c) are the fitting curves. We perform a fine estimation of $\theta^{*}$ by extending $T$ to 27.2 ms and applying a larger modulation. In this experiment, $\omega_{m}$ and $N$ are $2\pi\times 625$ Hz and $33$, respectively. We measure $m$ for $P_{1}=\pm 13$ mW, referred to as $m_{\pm}$, respectively. In finding $\theta^{*}$, we use $\Delta m=m_{+}-m_{-}$ to cancel the offset due to the background field and the systematic error in the spin measurement. The results are shown in Figs. 3(b) and (c). $\Delta m$ is fitted by $4\mathcal{V}F\sin(\beta_{2}(\theta-\theta^{*}))$, giving $\beta_{2}=117(16)$ and $\theta^{*}=0.06$ mrad with $\delta\theta^{*}$ = 0.40 mrad. The angle resolution is improved 2.8 times. We observe a larger variance in $m$ in the experiments for the fine $\theta^{*}$ estimation. The standard deviation of $\Delta m$ is on average 0.64, while that for the reference data without modulation is $0.09$. Therefore, a further improvement by a factor of at least 7 is possible, because $\Delta m$ should ideally be independent of $T$ and the modulation strength. We ascribe the increased variance to the actual variation of the vector shift over the experimental runs, caused by beam polarization fluctuation. The result of the sensitive detection implies that the beam circularity varies with the standard deviation of approximately $3\times 10^{-3}$. On the other hand, from an independent experiment, we expect that the retardance of the QWP should vary by several mrad due to the temperature change in our experimental room (within $\approx$ 0.6 K with a period of around 20 minutes). The BEC is subject to a fictitious magnetic field gradient without the VLS cancellation, because it is located at the shoulder of the optical trap beam due to gravity sag. While the observed fictitious magnetic field is small, the gradient in the fictitious field can be on the order of 100 $\mu$T/m. The gradient displaces the trap potential for each spin state other than the $m_{F}$ = $0$ state, thereby driving the spin dependent motion. We observe an actual motion in a transversally-spin-polarized BEC in the hyperfine spin $F=2$ state, prepared after the initial $\pi/2$ pulse. We plot the vertical displacement of the spin components in the TOF image, which reflects the momentum, in Figs. 4(a)–(d). The direction of the motion inverts depending on the sign of $\Delta\theta$ and the motion becomes small at $\Delta\theta\approx 0$. These observations indicate that the motion is induced by the fictitious magnetic field. Figure 4: (color online) Effects of the fictitious magnetic field on a transversally polarized BEC. (a)–(d) Vertical displacement of the center of mass in the TOF image. The panels show the data for $\Delta\theta$ = (-1, +0.02, +1, +4) degrees, respectively. The solid and dashed lines are guides for the eyes. (e)–(h) Population evolution corresponding to (a)–(d). The fictitious magnetic field gradient also causes nonlinear spin evolution and thus a population change, as does the real magnetic field gradient Eto et al. (2014). The initial polarized atomic spin state breaks due to the spin mixing seeded by the nonlinear spin evolution. We show $p_{0}=N_{0}/N_{\mathrm{tot}}$, $p_{1}=(N_{-1}+N_{+1})/(2N_{\mathrm{tot}})$, and $p_{2}=(N_{-2}+N_{+2})/(2N_{\mathrm{tot}})$ in Figs. 4(e)–(h). The population changes are observed at an earlier time ($t<100$ ms) except for the case $\Delta\theta\approx 0$. These changes can be attributed to the fictitious magnetic field gradient. The faster population change for $\Delta\theta=4^{\circ}$ is consistent with a qualitative estimation of the characteristic time for the change of $t_{*}\propto b^{-2/3}$, where $b$ is the magnetic field gradient Eto et al. (2014). A slow population change, which occurs regardless of $\Delta\theta$, is caused by a residual axial magnetic field gradient, $\partial B_{z}/\partial z$. The existence of the axial gradient in these data is confirmed by the fact that the spin components separate in the axial direction at later times. Figure 5: (color online) Atom number losses. The solid line is an exponential fit to the data with the optimized field gradient. The inset shows the population evolution for the optimized case. The dotted lines are the prediction curves of the mean-field driven evolution without including the inelastic losses Kronjäger et al. (2008). We next observed the change in the atom loss rate. Figure 5 shows the evolution of the atom numbers, corresponding to the data in Figs. 4(b) and (d). The decay is faster when $\Delta=4^{\circ}$ than for $\Delta\approx 0^{\circ}$. In the latter case, the decay rate starts to increase from around $t$ = 150 ms, where the population changes occur (see Fig. 4(f)). No increase in the loss rate is observed at the later time when we optimize $\theta$ and reduce the axial real magnetic field gradient. The $1/e$ time for the optimized condition is found to be 742 (31) ms. The change of the loss rate can be understood from the property of the inelastic collisions in the $F=2$ state Tojo et al. (2009). Note that the inelastic collisional loss in the polarized state is inhibited due to the restriction of the angular momentum conservation. The break of the polarized state due to the field gradient results in rapid atom losses. However, the loss still occurs for the optimized condition. Although the remaining loss may be due to the residual field inhomogeneity, it is associated with the spin mixing induced by the quadratic Zeeman energy Kronjäger et al. (2006, 2008). The model presented in Kronjäger et al. (2006, 2008), however, needs to be modified to explain the observed population conservation, shown in the inset of Fig.5. According to Kronjäger et al. (2008), the population evolution in the limit of small quadratic Zeeman energy, $q$, is approximately given by $\displaystyle p_{0}=$ $\displaystyle\frac{3}{8}\left[1+\frac{q}{2g_{1}n}(1-\cos(4g_{1}nt/\hbar))\right],$ (5) $\displaystyle p_{1}=$ $\displaystyle\frac{1}{4},$ (6) $\displaystyle p_{2}=$ $\displaystyle\frac{1}{16}\left[1-\frac{3q}{2g_{1}n}(1-\cos(4g_{1}nt/\hbar))\right],$ (7) where $g_{1}=\frac{4\pi\hbar^{2}}{m}\frac{a_{4}-a_{2}}{7}$ is the interaction strength with $a_{\mathcal{F}}$ being the $s$-wave scattering length for the collisional channel of the total angular momentum $\mathcal{F}$ and $n$ is the mean atomic density. Following these equations, $p_{0}$ and $p_{2}$ would undergo oscillations, which is not in agreement with the observed experimental result. We therefore attribute the population conservation to polarization purification by inelastic collisional losses Eto et al. (2019). It should be noted that the observed population conservation contrasts with the case of the $F=1$ state, in which the magnitude of the polarization modulates Jasperse et al. (2017). ## IV Discussion The quantum lock-in VLS detection is of practical use in cold atom experiments. It can be used for evaluating the degree of circular polarization of an optical trap beam at the atomic position, as we have shown. As the vacuum window birefringence introduces a maximum ellipticity of $10^{-2}$ or $10^{-1}$ Steffen et al. (2013), a beam with no special care taken with respect to the in vacuo polarization may generate a fictitious field of several nT or a VLS of tens of Hz, even with a shallow trap for ultracold atom experiments. Quantum lock-in detection is sensitive enough to ensure better linear polarization at the atomic position and therefore will greatly improve the magnetic conditions in cold atom experiments. The sensitivity is sufficient to suppress the VLS below the requirements for magnetically sensitive experiments, including studies of spinor BECs. Although a homogeneous linear Zeeman shift does not affect the spinor physics due to spin conservation Stamper-Kurn and Ketterle (2001), a magnetic field gradient below several $\mu$T/m is typically required to prevent magnetic polarization and observe the intrinsic magnetic ground state Stenger et al. (1998) or dynamics. The sub-Hz VLS resolution of quantum lock-in detection meets this challenging demand. Reducing the VLS is also important for precise measurements. In addition to a direct energy shift, an inhomogeneous fictitious field is also detrimental to measurement accuracy Cates et al. (1988). VLS reduction leads to a long coherence time, which is a mandatory requirement for highly sensitive measurements. We have constructed a precise BEC magnetometer using a transversally polarized $F=2$ BEC with a long coherence time, realized using VLS elimination as we have shown. The detail of the $F=2$ BEC magnetometer will be presented elsewhere Sekiguchi et al. (2020). We finally discuss the sensitivity limitations. The sensitivity of the quantum lock-in detection is essentially the same as that of a Ramsey interferometer with an equal phase accumulation time. As the atom shot noise is dominant over the photon shot noise in typical absorption imaging, the standard quantum limit in the VLS measurement is given by Giovannetti et al. (2004, 2006) $\delta\omega=\frac{1}{T\sqrt{N_{\mathrm{tot}}}}.$ (8) Here we replace the factor $\frac{2}{\pi}$ in Eq. (3) due to the sinusoidal modulation with the maximal value of 1, which is realized with a rectangular waveform modulation. Substituting $N_{\mathrm{tot}}=3\times 10^{5}$ and $T$ = $30$ ms into Eq. (8), we obtain $\delta\omega=2\pi\times 10$ mHz. This is equivalent to a single shot field sensitivity of $\sim$ 1 pT. ## V Conclusions We have demonstrated precise detection of a VLS due to an optical trap using the quantum lock-in method. We have applied the detection to eliminating the VLS, to observe the extension of the lifetime of transversally polarized $F=2$ BEC. The attained resolution of sub Hz is sufficient to suppress the VLS below the required level for magnetically sensitive research, including the study of spinor BECs. Although our demonstration was performed with a BEC, the scope of the detection method is not limited to cold atom gases; the proposed method can be applied to spin systems such as trapped ions and diamond NV centers, where coherent spin control is possible. ###### Acknowledgements. This work was supported by the MEXT Quantum Leap Flagship Program (MEXT Q-LEAP) Grant Number JPMXS0118070326 and JSPS KAKENHI Grant Number JP19K14635. ## References * Grimm et al. (2000) R. Grimm, M. Weidemüller, and Y. B. Ovchinnikov, Adv. At. Mol. Opt. Phys. 42, 95 (2000). * Görlitz et al. (2001) A. Görlitz, J. M. Vogels, A. E. Leanhardt, C. Raman, T. L. Gustavson, J. R. Abo-Shaeer, A. P. Chikkatur, S. Gupta, S. Inouye, T. Rosenband, et al., Phys. Rev. Lett. 87, 130402 (2001). * Gaunt et al. (2013) A. L. Gaunt, T. F. Schmidutz, I. Gotlibovych, R. P. Smith, and Z. Hadzibabic, Phys. Rev. Lett. 110, 200406 (2013). * Bloch et al. (2008) I. Bloch, J. Dalibard, and W. Zwerger, Rev. Mod. Phys. 80, 885 (2008). * Derevianko and Katori (2011) A. Derevianko and H. Katori, Rev. Mod. Phys. 83, 331 (2011). * Katori (2011) H. Katori, Nat. Photon. 5, 203 (2011). * Stamper-Kurn and Ueda (2013) D. M. Stamper-Kurn and M. Ueda, Rev. Mod. Phys. 85, 1191 (2013). * Deutsch and Jessen (1998) I. H. Deutsch and P. S. Jessen, Phys. Rev. A 57, 1972 (1998). * Geremia et al. (2006) J. M. Geremia, J. K. Stockton, and H. Mabuchi, Phys. Rev. A 73, 042112 (2006). * Deutsch and Jessen (2010) I. H. Deutsch and P. S. Jessen, Opt. Commun. 283, 681 (2010). * Mandel et al. (2003a) O. Mandel, M. Greiner, A. Widera, T. Rom, T. W. Hänsch, and I. Bloch, Phys. Rev. Lett. 91, 010407 (2003a). * Mandel et al. (2003b) O. Mandel, M. Greiner, A. Widera, T. Rom, T. W. Hänsch, and I. Bloch, Nature (London) 425, 937 (2003b). * Heinz et al. (2020) A. Heinz, A. J. Park, N. Šantić, J. Trautmann, S. G. Porsev, M. S. Safronova, I. Bloch, and S. Blatt, Phys. Rev. Lett. 124, 203201 (2020). * Romalis and Fortson (1999) M. V. Romalis and E. N. Fortson, Phys. Rev. A 59, 4547 (1999). * Jackson Kimball et al. (2017a) D. F. Jackson Kimball, J. Dudley, Y. Li, D. Patel, and J. Valdez, Phys. Rev. D 96, 075004 (2017a). * Fleischhauer et al. (2000) M. Fleischhauer, A. B. Matsko, and M. O. Scully, Phys. Rev. A 62, 013808 (2000). * Jellison (1999) G. E. Jellison, Appl. Opt. 38, 4784 (1999). * Zhu et al. (2013) K. Zhu, N. Solmeyer, C. Tang, and D. S. Weiss, Phys. Rev. Lett. 111, 243006 (2013). * Steffen et al. (2013) A. Steffen, W. Alt, M. Genske, D. Meschede, C. Robens, and A. Alberti, Rev. Sci. Instrum. 84, 126103 (2013). * Jackson Kimball et al. (2017b) D. F. Jackson Kimball, J. Dudley, Y. Li, and D. Patel, Phys. Rev. A 96, 033823 (2017b). * Wood et al. (2016) A. A. Wood, L. D. Turner, and R. P. Anderson, Phys. Rev. A 94, 052503 (2016). * Yuan et al. (2019) W. H. Yuan, H. L. Liu, W. Z. Wei, Z. Y. Ma, P. Hao, Z. Deng, K. Deng, J. Zhang, and Z. H. Lu, Rev. Sci. Instrum. 90, 113001 (2019). * Kotler et al. (2011) S. Kotler, N. Akerman, Y. Glickman, A. Keselman, and R. Ozeri, Nature (London) 473, 61 (2011). * de Lange et al. (2011) G. de Lange, D. Ristè, V. V. Dobrovitski, and R. Hanson, Phys. Rev. Lett. 106, 080802 (2011). * Eto et al. (2014) Y. Eto, M. Sadgrove, S. Hasegawa, H. Saito, and T. Hirano, Phys. Rev. A 90, 013626 (2014). * Kronjäger et al. (2008) J. Kronjäger, C. Becker, P. Navez, K. Bongs, and K. Sengstock, Phys. Rev. Lett. 100, 189901 (2008). * Tojo et al. (2009) S. Tojo, T. Hayashi, T. Tanabe, T. Hirano, Y. Kawaguchi, H. Saito, and M. Ueda, Phys. Rev. A 80, 042704 (2009). * Kronjäger et al. (2006) J. Kronjäger, C. Becker, P. Navez, K. Bongs, and K. Sengstock, Phys. Rev. Lett. 97, 110404 (2006). * Eto et al. (2019) Y. Eto, H. Shibayama, K. Shibata, A. Torii, K. Nabeta, H. Saito, and T. Hirano, Phys. Rev. Lett. 122, 245301 (2019). * Jasperse et al. (2017) M. Jasperse, M. J. Kewming, S. N. Fischer, P. Pakkiam, R. P. Anderson, and L. D. Turner, Phys. Rev. A 96, 063402 (2017). * Stamper-Kurn and Ketterle (2001) D. M. Stamper-Kurn and W. Ketterle, in _Coherent atomic matter waves_ , edited by R. Kaiser, C. Westbrook, and F. David (Springer, Berlin, Heidelberg, 2001), pp. 139–217. * Stenger et al. (1998) J. Stenger, S. Inouye, D. M. Stamper-Kurn, H.-J. Miesner, A. P. Chikkatur, and W. Ketterle, Nature (London) 396, 345 (1998). * Cates et al. (1988) G. D. Cates, S. R. Schaefer, and W. Happer, Phys. Rev. A 37, 2877 (1988). * Sekiguchi et al. (2020) N. Sekiguchi, A. Torii, H. Toda, R. Kuramoto, D. Fukuda, T. Hirano, and K. Shibata (2020), eprint arXiv:2009.07569. * Giovannetti et al. (2004) V. Giovannetti, S. Lloyd, and L. Maccone, Science 306, 1330 (2004). * Giovannetti et al. (2006) V. Giovannetti, S. Lloyd, and L. Maccone, Phys. Rev. Lett. 96, 010401 (2006).
# Integral morphisms and log blow-ups Fumiharu Kato ###### Abstract. This paper is a revision of the author’s old preprint “Exactness, integrality, and log modifications”. We will prove that any quasi-compact morphism of fs log schemes can be modified locally on the base to an integral morphism by base change by fs log blow-ups. ###### 2010 Mathematics Subject Classification: Primary: 14A99, Secondary: 14E05 ## 1\. Introduction The aim of this paper is to prove the following theorem: ###### Theorem 1.1. Let $f\colon X\rightarrow Y$ be a quasi-compact morphism of fs $(=$ fine and saturated$)$ log schemes. Then for any $y\in Y$ there exists an étale neighborhood $V\rightarrow Y$ of $\overline{y}$ $(=$ a separable closure of $y)$ and an fs log blow-up $V^{\prime}:=\operatorname{Bl}_{\mathcal{K}}(V)\rightarrow V$ along a coherent log ideal $\mathcal{K}\subset\mathcal{M}_{V}$, by which the fs base change $f_{V^{\prime}}\colon X_{V^{\prime}}\rightarrow V^{\prime}$ is integral. Here, by fs log blow-up (resp. fs base change) we mean a log blow-up (resp. base change) in the category of fs log schemes; cf. Remark 3.6. This theorem has been announced and proved in somewhat incomplete and inaccurate form in the author’s old preprint [5], a first draft of which has actually been written in 1997, and afterwards put in the arXiv in 1999. Since then, mainly because the author has been away from log geometry, the paper has been kept unpublished; the author apologizes for all inconvenience caused thereby. While there have been much progress and many new results in log geometry last two decades, the paper has sometimes been referred to. Moreover, it seems, to the best of the author’s knowledge, that the theorem itself has not yet been written anywhere, even in the foundational book [11] by Ogus, and became folklore among experts. In fact, the theorem is nowadays a consequence of combination of known results. For example, Luc Illusie, Kazuya Kato, and Chikara Nakayama proved in [4] (see also [11, III.2.6.7]) a weak version of the theorem, where “integral” is replaced by “$\mathbb{Q}$-integral”. Then by a further fs log blow-up of the base to make the log structure free (i.e., to make each stalk of $\overline{\mathcal{M}}=\mathcal{M}/\mathcal{O}^{\times}$ a free monoid), the resulting map becomes integral (cf. [11, I.4.7.5]). Since, due to Nizioł [10, 4.11], the composition of fs log blow-ups is again an fs log blow-up, this actually suffices to prove the theorem. In the mean time, in August 2020, Michael Temkin asked the author some questions on the preprint, and suggested the final form of the theorem presented as above. Based upon the fact that the theorem has to be referred to in a recent work [2] of him and his coauthors, and that the theorem has not yet been presented in published form, he encouraged the author to revise the old preprint for publishing. This is the situation from whence the present paper comes out, where we keep the original proof based on the technique of toric flattening, the original idea of which is attributed to Takeshi Kajiwara. Let us mention some consequences of Theorem 1.1. Tsuji [13, II.3.4] proved that any integral and quasi-compact morphism between fs log schemes can be made saturated by fs base change by “multiplication-by-$n$” map. Combined with our result, this yields the following: ###### Corollary 1.2. Let $f\colon X\rightarrow Y$ be a quasi-compact morphism between fs log schemes. Then for any $y\in Y$ there exists a diagram $\lx@xy@svg{\hbox{\raise 0.0pt\hbox{\kern 14.38945pt\hbox{\ignorespaces\ignorespaces\ignorespaces\hbox{\vtop{\kern 0.0pt\offinterlineskip\halign{\entry@#!@&&\entry<EMAIL_ADDRESS>0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\raise 0.0pt\hbox{$\textstyle{V^{\prime\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces{\hbox{\kern 44.98112pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces{\hbox{\kern 0.0pt\raise-29.49998pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern 44.98112pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\raise 0.0pt\hbox{$\textstyle{V^{\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{\hbox{\kern 60.57668pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\lx@xy@hook{1}}}}}}{\hbox{\lx@xy@droprule}}\ignorespaces{\hbox{\kern 91.16835pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces{\hbox{\kern 52.7789pt\raise-29.49998pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern 91.16835pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\raise 0.0pt\hbox{$\textstyle{\operatorname{Bl}_{K}(V)\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces{\hbox{\kern 109.59393pt\raise-30.38885pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces{\hbox{\kern 158.6112pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern 158.6112pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\raise 0.0pt\hbox{$\textstyle{V\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces{\hbox{\kern 165.63898pt\raise-30.38885pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces{\hbox{\kern 203.25844pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern 203.25844pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\raise 0.0pt\hbox{$\textstyle{Y}$}}}}}}}{\hbox{\kern-14.38945pt\raise-40.88885pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\raise 0.0pt\hbox{$\textstyle{\operatorname{S}(Q^{\prime})\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces\ignorespaces{\hbox{\kern 21.28053pt\raise-46.07634pt\hbox{{}\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\hbox{\hbox{\kern 0.0pt\raise-0.8264pt\hbox{$\scriptstyle{\mu}$}}}\kern 3.0pt}}}}}}\ignorespaces{\hbox{\kern 38.38945pt\raise-40.88885pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern 38.38945pt\raise-40.88885pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\raise 0.0pt\hbox{$\textstyle{\operatorname{S}(Q^{\prime})\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{\hbox{\kern 67.16837pt\raise-40.88885pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\lx@xy@hook{1}}}}}}{\hbox{\lx@xy@droprule}}\ignorespaces{\hbox{\kern 91.24336pt\raise-40.88885pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern 91.24336pt\raise-40.88885pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\raise 0.0pt\hbox{$\textstyle{\operatorname{Bl}_{K}(Q)\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces\ignorespaces{\hbox{\kern 132.62137pt\raise-45.3958pt\hbox{{}\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\hbox{\hbox{\kern 0.0pt\raise-1.50694pt\hbox{$\scriptstyle{\pi}$}}}\kern 3.0pt}}}}}}\ignorespaces{\hbox{\kern 152.01952pt\raise-40.88885pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern 152.01952pt\raise-40.88885pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\raise 0.0pt\hbox{$\textstyle{\operatorname{S}(Q)}$}}}}}}}\ignorespaces}}}}\ignorespaces,$ with all vertical arrows strict and all squares cartesian in the category of fs log schemes, such that * • $V\rightarrow Y$ is an étale neighborhood of $\overline{y}$; * • $\pi$ is an fs log blow-up of $\operatorname{S}(Q)=\operatorname{Spec}\mathbb{Z}[Q]$ $($with the log structure by $Q\rightarrow\mathbb{Z}[Q]$$)$ along an ideal $K\subset Q$; * • $\operatorname{S}(Q^{\prime})\hookrightarrow\operatorname{Bl}_{K}(Q)$ is an arbitrary affine patch of $\operatorname{Bl}_{K}(Q)$; * • $\mu$ is the morphism induced from the multiplication-by-$n$ homomorphism $Q^{\prime}\rightarrow Q^{\prime}$ for some positive integrer $n$, and that the fs base change $f_{V^{\prime\prime}}\colon X\times_{Y}V^{\prime\prime}\rightarrow V^{\prime\prime}$ is saturated. Our second application is to log flatness. ###### Corollary 1.3. Let $f\colon X\rightarrow Y$ be a log flat and finitely presented morphism between fs log schemes. Then for any $y\in Y$ there exists an étale neoghborhood $V\rightarrow Y$ of $\overline{y}$ and an fs log blow-up $V^{\prime}\rightarrow V$ such that the underlying morphism of the fs base change $f_{V^{\prime}}\colon X\times_{Y}V^{\prime}\rightarrow V^{\prime}$ is flat. Indeed, we may assume that $f$ is integral, log flat, and of finite presentation. We may further assume that $f$ has a global chart by $h\colon Q\rightarrow P$, which is “neat” at a point $x\in X$ (as in Lemma 2.4 below) such that $Q\cong\overline{\mathcal{M}}_{Y,\overline{y}}$ ($y=f(x)$) and $\overline{P}\cong\overline{\mathcal{M}}_{X,\overline{x}}$. Then, since $\overline{h}\colon\overline{Q}=Q\rightarrow\overline{P}$ is integral, so is $h$ (cf. Lemma 2.2 (5)). Since $Q$ is sharp (i.e., $Q^{\times}=\\{1\\}$) and $h$ is local (i.e., $h^{-1}(P^{\times})=Q^{\times}$), the ring homomorphism $\mathbb{Z}[P]\rightarrow\mathbb{Z}[Q]$ is flat (cf. Lemma 2.2 (6)). Since the log flatness implies that $X$ is flat over $Y\otimes_{\mathbb{Z}[Q]}\mathbb{Z}[P]$ ([12, 4.15]), we deduce that $f$ itself is flat. As for the interaction between log flatness and usual flatness, much has been studied recently by some authors. Among them, we refer to a preprint by Gillam [3]. It seems that, combining our result with many of the results therein, we can deduce several useful consequences. Finally, let us remark that, if $Y$ in Theorem 1.1 is log regular, then $V$ can be equal to $Y$ itself, i.e., one can find an fs log blow-up of $Y$ itself that makes the morphism $f$ integral by fs base change. This version of the theorem has been proven independently in [2, 3.6.11], which we will include, with a few comments, at the end of this paper (Proposition 4.2). ###### Remark 1.4. The original paper [5] included an “exactness version” of the theorem, where “integral” is replaced by “exact”, which we do not include in the present paper, since it follows immediately from the result in [4] mentioned above. The composition of this paper is as follows. In the next section, we will collect some basics of integral morphisms and neat charts. In Section 3, we will overview log blow-ups. We will then prove the main theorem in Section 4. The original version of the paper owes much to Richard Pink, Takeshi Kajiwara, and Max Planck Institute für Mathematik in Bonn, Germany. In addition, the preparation of the present version owes much to Michael Temkin for his encouragement, and to Chikara Nakayama for valuable discussions. ### 1.5. Notation and conventions All rings and monoids are assumed to be commutative. For a monoid $M$, we denote by $M^{\times}$ and $M^{\mathrm{gp}}$ the subgroup of invertible elements and the associated group, respectively, and write $\overline{M}=M/M^{\times}$. All sheaves on schemes are considered with respect to the étale topology. For a point $x$ of a scheme, we denote by $\overline{x}$ a separable closure of $x$. For a log scheme $X$, we denote by $\alpha_{X}\colon\mathcal{M}_{X}\rightarrow\mathcal{O}_{X}$ the log structure of $X$, and write $\overline{\mathcal{M}}_{X}=\mathcal{M}_{X}/\mathcal{O}^{\times}_{X}$. We denote by $\underline{X}$ the underlying scheme of $X$, which is also considered as a log scheme with the trivial log structure. For a monoid $P$, we denote by $\operatorname{S}(P)$ the log scheme whose underlying scheme is the affine scheme $\operatorname{Spec}\mathbb{Z}[P]$ with the log structure induced from $P\rightarrow\mathbb{Z}[P]$. ($\operatorname{S}(P)$ is denoted by $\mathsf{A}_{P}$ in [11].) ## 2\. Integral homomorphisms Let us first recall the definition of integral homomorphisms. ###### Definition 2.1. A homomorphism $h\colon Q\rightarrow P$ of integral monoids is said to be integral if, for any integral monoid $Q^{\prime}$ and any homomorphism $Q\rightarrow Q^{\prime}$, the push-out $P\oplus_{Q}Q^{\prime}$ in the category of monoids is an integral monoid. It can be shown ([6, (4.1)][11, I.4.6.2]) that $h\colon Q\rightarrow P$ is integral if and only if it has the following property: if $h(a_{1})b_{1}=h(a_{2})b_{2}$ for $a_{1},a_{2}\in Q$ and $b_{1},b_{2}\in P$, there exists $a_{3},a_{4}\in Q$ and $b\in P$ such that $b_{1}=h(a_{3})b$, $b_{2}=h(a_{4})b$ and $a_{1}a_{3}=a_{2}a_{4}$. ###### Lemma 2.2. $(1)$ The composition of integral homomorphisms is integral. For a diagram $Q\stackrel{{\scriptstyle h}}{{\rightarrow}}P\stackrel{{\scriptstyle k}}{{\rightarrow}}R$ of integral monoids, if $k\circ h$ is integral and $k$ is exact, then $h$ is integral; if $k\circ h$ is integral and $h$ is surjective, then $k$ is integral. $(2)$ The pushout of an integral homomorphism in the category of monoids is integral. $(3)$ If $P$ is an integral monoid, and $N\subset P$ is a submonoid, then the canonical map $P\rightarrow P/N$ is integral. $(4)$ An integral and local homomorphism of integral monoids is exact. $(5)$ A homomorphism $h\colon Q\rightarrow P$ of integral monoids is integral if and only if $\overline{h}\colon\overline{Q}\rightarrow\overline{P}$ is integral. $(6)$ A homomorphism $h\colon Q\rightarrow P$ of integral monoids is integral if the homomorphism of monoid algebras $\mathbb{Z}[Q]\rightarrow\mathbb{Z}[P]$ is flat. The converse is true, if $h$ is local and $Q$ is sharp. Recall that a homomorphism $h\colon Q\rightarrow P$ of integral monoids said to be exact if $(h^{\mathrm{gp}})^{-1}(P)=Q$, where $h^{\mathrm{gp}}\colon Q^{\mathrm{gp}}\rightarrow P^{\mathrm{gp}}$ is the associated group homomorphism. Recall also that, for a monoid $P$ and a submonoid $N\subset P$, the quotient monoid $P/N$ is given by $P/\sim$ (endowed with the natural monoid structure), where $a\sim b$ if and only if $ac=bd$ for some $c,d\in N$. ###### Proof. For (1), (4), and (6), see [11, I.4.6.3 & I.4.6.7]. (2) is immediate from the definition of integral homomorphisms. (5) follows from (1), (3), and the fact that a homomorphism of integral monoids of the form $Q\rightarrow\overline{Q}$ is always exact. Hence it suffices to show (3). To show that $\pi\colon P\rightarrow P/N$ ($a\mapsto\overline{a}$) is integral, take $a_{1},a_{2},b_{1},b_{2}\in P$ such that $\overline{a_{1}}\overline{b_{1}}=\overline{a_{2}}\overline{b_{2}}$; the last equality means $a_{1}b_{1}c_{1}=a_{2}b_{2}c_{2}$ for $c_{1},c_{2}\in N$, and if we set $a_{3}=b_{1}c_{1}$, $a_{4}=b_{2}c_{2}$ and $b=\overline{1}$, then we have $a_{1}a_{3}=a_{2}a_{4}$, $\overline{b_{1}}=\overline{a_{3}}b$ and $\overline{b_{2}}=\overline{a_{4}}b$, which shows that $\pi$ is integral. ∎ ###### Definition 2.3. A morphism $f\colon X\rightarrow Y$ of integral log schemes is said to be integral at $x\in X$ if the monoid homomorphism $\overline{\mathcal{M}}_{Y,\overline{y}}\rightarrow\overline{\mathcal{M}}_{X,\overline{x}}$, where $y=f(x)$, is integral, or equivalently, $\mathcal{M}_{Y,\overline{y}}\rightarrow\mathcal{M}_{X,\overline{x}}$ is integral (cf. Lemma 2.2 (6)). We say $f$ is integral if it is integral at all points of $X$. By Lemma 2.2 (2), integral morphisms are stable under base change in the category of fine log schemes. It is, however, not true that integral morphisms are stable under base change in the category of fs log schemes, cf. [11, I.4.6.5]. So it is often convenient to refer to a base change (or a fiber product) in the category of fs log schemes as fs base change for emphasis. Let us finally mention some technical facts on charts. ###### Lemma 2.4. Let $f\colon X\rightarrow Y$ be a morphism of fs log schemes. Then, for any $x\in X$ and $y=f(x)$, there exists commutative diagram $\textstyle{X\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{f}$$\textstyle{U\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{g}$$\textstyle{\operatorname{S}(P)\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{Y}$$\textstyle{V\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{\operatorname{S}(Q)}$ comprised of an étale neighborhood $V\rightarrow Y$ of $\overline{y}$, an fppf neighborhood $U\rightarrow X$ of $\overline{x}$, and a chart $Q\rightarrow P$ of $g\colon U\rightarrow V$ such that the following conditions are satisfied: * (a) $P$ and $Q$ are fs monoids; * (b) $Q\cong\overline{\mathcal{M}}_{Y,\overline{y}}$, and $\overline{P}\cong\overline{\mathcal{M}}_{X,\overline{x}}$; * (c) $Q^{\mathrm{gp}}\rightarrow P^{\mathrm{gp}}$ is injective and $P^{\mathrm{gp}}/Q^{\mathrm{gp}}\cong\mathcal{M}^{\mathrm{gp}}_{X/Y,\overline{x}}$. $($Note that, in this situation, $Q$ is sharp, and $Q\rightarrow P$ is local.$)$ ###### Proof. Since $Y$ is fs, one can take on an étale neighborhood of $\overline{y}$ a chart subordinate to the fs monoid $Q=\overline{\mathcal{M}}_{Y,\overline{y}}$. Then one can construct a local chart of $f$ as above according to the recipe as in [11, III.1.2.7], where one can take $P$ to be an fs monoid by the construction as in [11, II.2.4.4] ∎ ###### Lemma 2.5. Let $f\colon X\rightarrow Y$ be a morphism of fine log schemes. $(1)$ Suppose $f$ has a $($global$)$ chart by a homomorphism $h\colon Q\rightarrow P$ of fine monoids, such that the conditions (b) and (c) in Lemma 2.4 is satisfied. Then $f$ is integral if and only if the homomorphism $h$ is integral. $(2)$ If $f$ is integral at $x$, then it is integral at all points in an open neighborhood of $x$. ###### Proof. See [11, III.2.5.2]. ∎ ## 3\. Log blow-ups In this section, we briefly recall the notion of log blow-ups and their basic properties. Recall first that an ideal of a monoid $M$ is a subset $K\subset M$ such that $x\in K$ and $a\in M$ imply $ax\in K$. Trivial ideals are $\emptyset$ and $M$ itself. It follows from Dickson’s lemma that any ideal of a finitely generated monoid is finitely generated. If $\pi\colon M\rightarrow\overline{M}$ is the canonical map, the map $K\mapsto\pi(K)$ gives a bijection from the set of ideals of $M$ to the set of ideals of $\overline{M}$. Let $X$ be a log scheme. A log ideal of $X$ is a sheaf of ideals $\mathcal{K}$ of $\mathcal{M}_{X}$. We denote by $\overline{\mathcal{K}}$ the corresponding ideal of $\overline{\mathcal{M}}_{X}$. For a morphism $f\colon X\rightarrow Y$ of fine log schemes and a log ideal $\mathcal{K}$ of $Y$, one has the extension of the log ideal $\mathcal{K}\mathcal{M}_{X}=(f^{-1}\mathcal{K})\mathcal{M}_{X}$, which is a log ideal of $X$. ###### Example 3.1. Let $P$ be a monoid and $K\subset P$ an ideal. One has the log ideal $\widetilde{K}$ associated to $K$ on $X=\operatorname{S}(P)$, constructed as follows. For any open subset $U\subset X$, we have a monoid homomorphism $P\rightarrow\mathcal{M}_{X}(U)$, and hence the extension ideal $K\mathcal{M}_{X}(U)$ of $\mathcal{M}_{X}(U)$. Then $\widetilde{K}$ is the sheafification of the subpresheaf of ideals of $\mathcal{M}_{X}$ given by $U\mapsto K\mathcal{M}_{X}(U)$. Note that, for any $x\in X$, we have $\widetilde{K}_{\overline{x}}=K\mathcal{M}_{X,\overline{x}}$. ###### Definition 3.2. A log ideal $\mathcal{K}$ of $X$ is called coherent at $x\in X$ if there exists a local chart $U\rightarrow\operatorname{S}(P)$, where $U$ is an étale neighborhood around $\overline{x}$, and an ideal $K\subset P$ such that $\mathcal{K}|_{U}=\widetilde{K}\mathcal{M}_{U}$ (let us say, in this situation, that $K$ is a chart of $\mathcal{K}$ over $U$). A log ideal is called coherent if it is coherent at all points. ###### Remark 3.3 (cf. [11, II.2.6.2]). If a log ideal $\mathcal{K}$ of $X$ is coherent at $x\in X$, then, for any local chart $U\rightarrow\operatorname{S}(P)$ around $\overline{x}$, the pullback ideal $K\subset P$ of $\overline{\mathcal{K}}_{\overline{x}}$ by $P\rightarrow\mathcal{M}_{X,\overline{x}}\rightarrow\overline{\mathcal{M}}_{X,\overline{x}}$ generates $\mathcal{K}$ around $\overline{x}$; i.e., $\widetilde{K}\mathcal{M}_{U^{\prime}}=\mathcal{K}|_{U^{\prime}}$ over an étale neighborhood $U^{\prime}$ of $\overline{x}$ contained in $U$. A coherent log ideal $\mathcal{K}$ of a log scheme $X$ is said to be invertible if, for any $x\in X$, $\overline{\mathcal{K}}_{\overline{x}}$ is a principal (i.e. generated by a single element) ideal of $\overline{\mathcal{M}}_{X,\overline{x}}$, or equivalently, there exist étale locally a chart $U\rightarrow\operatorname{S}(P)$ and an ideal $K\subset P$ as in Definition 3.2 with $K$ being principal. ###### Definition 3.4. A morphism $f\colon X^{\prime}\rightarrow X$ of fine (resp. fs) log schemes is called a log blow-up along a coherent log ideal $\mathcal{K}$ if it has the following universal mapping property: * (a) $\mathcal{K}\mathcal{M}_{X^{\prime}}$ is an invertible log ideal of $\mathcal{M}_{X^{\prime}}$; * (b) If $g\colon T\rightarrow X$ is a morphism of fine (resp. fs) log schemes such that $\mathcal{K}\mathcal{M}_{T}$ is invertible, then there exists a uniquely morphism $T\rightarrow X^{\prime}$ such that the diagram $\textstyle{T\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{g}$$\textstyle{X^{\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{f}$$\textstyle{X}$ commutes. The log blow-up along a coherent log ideal is unique up to isomorphisms. Since every extension of invertible log ideal is again invertible, we have: ###### Lemma 3.5. The family of log blow-ups is stable under base change. More precisely, if $X_{\mathcal{K}}\rightarrow X$ is a log blow-up of a fine log scheme $X$ along a coherent log ideal $\mathcal{K}$, and if $Y\rightarrow X$ is a morphism of fine log schemes, then $X_{\mathcal{K}}\times_{X}Y\rightarrow Y$ is a log blow-up of $Y$ along $\mathcal{K}\mathcal{M}_{Y}$. If $X$ is an fs log scheme, and $f\colon X^{\prime}\rightarrow X$ is a log blow-up in the category of fine log schemes, then the saturation $f^{\mathrm{sat}}\colon X^{\prime\mathrm{sat}}\rightarrow X$ gives a log blow- up of $X$ along the same coherent log ideal in the category of fs log schemes. Hence, to show the existence of log blow-ups, it suffices to deal with the case of fine log schemes. ###### Remark 3.6. As indicated above, the log blow-ups in the category of fs log schemes are rather similar to normalized blow-ups, i.e., blow-up followed by normalization. To make clear the distinction, we will call log blow-ups taken in the category of fs log schemes fs log blow-ups. The following construction of log blow-ups is due to Kazuya Kato [7, (1.3.3)] (cf. [11, III.2.6]): We first construct the log blow-up $\operatorname{Bl}_{K}(P)\longrightarrow\operatorname{S}(P)$ of $\operatorname{S}(P)$, where $P$ is a fine monoid, along the coherent log ideal $\widetilde{K}$ constructed from an ideal $K\subset P$. Let $I(K)$ be the ideal of $\mathbb{Z}[P]$ generated by $K$, and consider the natural morphism $\operatorname{Proj}\bigoplus_{n}I(K)^{n}\rightarrow\operatorname{Spec}\mathbb{Z}[P]$. $\operatorname{Proj}\bigoplus_{n}I(K)^{n}$ has the affine open covering $\operatorname{Proj}\bigoplus_{n}I(K)^{n}=\bigcup_{t\in K}\operatorname{Spec}\mathbb{Z}[P\langle t^{-1}K\rangle].$ Here, $P\langle E\rangle$ for a subset $E$ of $P^{\mathrm{gp}}$ denotes the smallest fine submonoid in $P^{\mathrm{gp}}$ that contains $P$ and $E$. The canonical log structures given by $P\langle t^{-1}K\rangle\rightarrow\mathbb{Z}[P\langle t^{-1}K\rangle]$ glue to a fine log structure on $\operatorname{Proj}\bigoplus_{n}I(K)^{n}$. Then it follows that $\operatorname{Bl}_{K}(P):=\operatorname{Proj}\bigoplus_{n}I(K)^{n}\rightarrow\operatorname{S}(P)$ gives a log blow-up of $\operatorname{S}(P)$ along $\widetilde{K}$. To give a more explicit local description, take generators $t_{0},\ldots,t_{r}$ of $K$. Then $\operatorname{Bl}_{K}(P)$ is the union of the affine log schemes $\operatorname{Spec}\mathbb{Z}\big{[}P\big{\langle}\textrm{{\footnotesize$\frac{t_{0}}{t_{i}},\ldots,\frac{t_{r}}{t_{i}}$}}\big{\rangle}\big{]},$ with the log structure induced from $P\langle t_{0}/t_{i},\ldots,t_{r}/t_{i}\rangle\rightarrow\mathbb{Z}[P\langle t_{0}/t_{i},\ldots,t_{r}/t_{i}\rangle]$, i.e., the affine log schemes $\operatorname{S}(P\langle t_{0}/t_{i},\ldots,t_{r}/t_{i}\rangle)$, for $i=0,\ldots,r$. Let $X$ be a fine log scheme, and $\mathcal{K}$ a coherent log ideal of $X$. Suppose there exist a chart $\lambda\colon X\rightarrow\operatorname{S}(P)$ modeled on a fine monoid $P$ and an ideal $K\subset P$ such that $\mathcal{K}=\widetilde{K}\mathcal{M}_{X}$. Then, by Lemma 3.5, $\operatorname{Bl}_{\mathcal{K}}(X)=X\times_{\operatorname{S}(P)}\operatorname{Bl}_{K}(P)\longrightarrow X$ gives a log blow-up of $X$ along $\mathcal{K}$. In general, we take an étale covering $\\{U_{\alpha}\\}_{\alpha\in L}$ of $X$ such that each $U_{\alpha}$ allow a chart $U_{\alpha}\rightarrow\operatorname{S}(P_{\alpha})$ with an ideal $K_{\alpha}\subset P_{\alpha}$ satisfying $\mathcal{K}|_{U_{\alpha}}=\widetilde{K}_{\alpha}\mathcal{M}_{U_{\alpha}}$. Then, by the universality of log blow-ups, the local log blow-ups $\operatorname{Bl}_{\mathcal{K}_{\alpha}}(U_{\alpha})\rightarrow U_{\alpha}$ constructed as above glue to a log blow-up of $X$ along $\mathcal{K}$. ###### Example 3.7. Let $P$ be a sharp fs monoid, and set $X=\operatorname{S}(P)$. Set $M=P^{\mathrm{gp}}$ and $N=\operatorname{Hom}_{\mathbb{Z}}(M,\mathbb{Z})$. The scheme $X$ is an affine toric variety corresponding to the corn $\sigma$ in $N_{\mathbb{R}}$ such that $\sigma^{\vee}\cap M=P$; i.e., $\sigma$ is the dual corn of the corn in $M_{\mathbb{R}}$ generated by $P$. Let $\phi\colon\sigma\rightarrow\mathbb{R}_{\geq 0}$ be a continuous convex piecewise linear function satisfying the following conditions (cf. [9, p.27]): * (a) $\phi(\lambda x)=\lambda\phi(x)$ for $x\in\sigma$ and $\lambda\in\mathbb{R}_{\geq 0}$; * (b) $\phi(N\cap\sigma)\subset\mathbb{Z}$. The function $\phi$ induces an ideal $K_{\phi}$ of $P$ given by $K_{\phi}=\\{m\in M\mid\langle x,m\rangle\geq\phi(x)\ \textrm{for all}\ x\in\sigma\\}.$ Then the fs log blow-up of $X$ along the log ideal $\widetilde{K}_{\phi}$ is the normalization of the blow-up of the toric variety $X=X_{\sigma}$ obtained from the coarsest subdividing fan $\Sigma_{\phi}$ of the cone $\sigma$ such that $\phi$ is linear on each cone in $\Sigma_{\phi}$; cf. [9, p.31, Theorem 10]. ## 4\. Proof of the theorem ###### Lemma 4.1. Let $P,Q$ be sharp fs monoids, and $h\colon Q\hookrightarrow P$ an injective homomorphism. Consider the induced morphism $f\colon X=\operatorname{S}(P)\rightarrow Y=\operatorname{S}(Q)$ of fs log schemes. Then there exists an ideal $K\subset Q$ such that the following conditions are satisfied: if $\textstyle{X^{\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{f^{\prime}}$$\textstyle{X\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{f}$$\textstyle{\operatorname{Bl}_{K}(Y)\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{Y}$ is the fs base change of $f$ by the fs log blow-up $\operatorname{Bl}_{K}(Y)\rightarrow Y$, then the underlying scheme-theoretic morphism of $f^{\prime}$ is equidimensional. Note that $X^{\prime}\rightarrow X$ is isomorphic to the fs log blow-up along $\widetilde{KP}=\widetilde{K}\mathcal{M}_{X}$, i.e., $X^{\prime}\cong\operatorname{Bl}_{KP}(X)$. ###### Proof. In this proof, we follow the original argument in [5, 3.16] based on the idea of T. Kajiwara, which we note is similar to the argument in [1, Lemma 4.3]. We use the following notation: * • $M_{Q}=Q^{\mathrm{gp}}$, $M_{P}=P^{\mathrm{gp}}$; * • $N_{Q}=\operatorname{Hom}_{\mathbb{Z}}(M_{Q},\mathbb{Z})$, $N_{P}=\operatorname{Hom}_{\mathbb{Z}}(M_{P},\mathbb{Z})$; * • $\sigma_{Q}$ $($resp. $\sigma_{P})$ $=$ the cone in $N_{Q}$ $($resp. $N_{P})$ such that $Q=\sigma^{\vee}_{Q}\cap M_{Q}$ $($resp. $P=\sigma^{\vee}_{P}\cap M_{P})$. * • $\Sigma_{Q}$ (resp. $\Sigma_{P}$) $=$ the fan made up from the faces of the cone $\sigma_{Q}$ (resp. $\sigma_{P}$). Note that we have a map $\Phi\colon\Sigma_{P}\rightarrow\Sigma_{Q}$ of fans that induces the morphism of affine toric schemes $\operatorname{Spec}\mathbb{Z}[P]\rightarrow\operatorname{Spec}\mathbb{Z}[Q]$. Consider the subset $\Sigma_{P,1}\subset\Sigma_{P}$ (resp. $\Sigma_{Q,1}\subset\Sigma_{Q}$) of rays, i.e., cones of dimension $1$. Each $\rho\in\Sigma_{P,1}$ is mapped by $\Phi$ onto a ray in $\Sigma_{Q}$ or to a point (i.e. the origin of $N_{Q}$). If $\rho$ is mapped onto a ray, then take the primitive base $n_{1}\in N_{Q}$ of $\Phi(\rho)$, and extend it to a $\mathbb{Z}$-base $n_{1},\ldots,n_{r}$ of $N_{Q}$. The $r+1$-rays spanned by $n_{1},\ldots,n_{r},-(n_{1}+\cdots+n_{r})$ defines the projective $r$-space $\mathbb{P}^{r}_{\mathbb{Z}}$, and hence the ideal $\mathcal{O}(-1)$ gives rise to a support function, denoted by $\phi_{\rho}$, i.e., a continuous convex piecewise linear function $N_{Q,\mathbb{R}}\rightarrow\mathbb{R}_{\geq 0}$ satisfying the conditions (a) and (b) in Example 3.7; we denote the restriction of $\phi_{\rho}$ onto $\sigma_{Q}$ by the same symbol. If $\Phi(\rho)$ is a point, then set $\phi_{\rho}=0$. Set $\phi=\sum_{\rho\in\Sigma_{P,1}}\phi_{\rho},$ and let $\Sigma^{\prime}_{Q}$ be the coarsest fan that subdivides $\Sigma_{Q}$ such that $\phi$ is linear on each cone in $\Sigma^{\prime}_{Q}$. (The author learned this way of constructing $\phi$ from T. Kajiwara.) Now, let $K\subset Q$ be an ideal constructed from $\phi$ as in Example 3.7. Consider the fs log blow-up $\operatorname{Bl}_{K}(Y)\rightarrow Y$, and the fs base change $f^{\prime}\colon X^{\prime}:=X\times_{Y}\operatorname{Bl}_{K}(Y)\rightarrow\operatorname{Bl}_{K}(Y)$ of $f$. Then, $X^{\prime}\rightarrow X$ is the log blow-up of $X$ along $\widetilde{K}\mathcal{M}_{X}$, which is the normalized toric blow-up induced from the piecewise linear function on $N_{P,\mathbb{R}}$ given by the pull- back of $\phi$. If we denote the corresponding fan of $X\times_{Y}\operatorname{Bl}_{K}(Y)$ by $\Sigma^{\prime}_{P}$, then the induced map $\Phi^{\prime}\colon\Sigma^{\prime}_{P}\rightarrow\Sigma^{\prime}_{Q}$ maps each ray onto either a ray or a point (the origin), and hence mapping each cone onto a cone. Therefore, the morphism $f^{\prime}\colon X^{\prime}\rightarrow\operatorname{Bl}_{K}(Y)$ is equidimensional by [1, Lemma 4.1]. ∎ ###### Proof of Theorem 1.1. Let $Q=\overline{\mathcal{M}}_{Y,\overline{y}}$, and take an étale local chart $Y\leftarrow V\rightarrow\operatorname{S}(Q)$ around $\overline{y}$. For any $x\in X_{V}=X\times_{Y}V$, take an fppf local chart $X_{V}\leftarrow U_{x}\rightarrow\operatorname{S}(P_{x})$ around $\overline{x}$, where $P_{x}$ is an fs monoid, which extends to a local chart of $f$ as in Lemma 2.4. Since $f$ is quasi-compact, one can take finitely many $x_{1},\ldots,x_{n}\in X_{V}$ such that $X_{V}$ is covered by the union of $U_{i}:=U_{x_{i}}$ for $i=1,\ldots,n$. We set $P_{i}=P_{x_{i}}$ for $i=1,\ldots,n$. For $i=1,\ldots,n$, there exists by Lemma 4.1 an ideal $K_{i}\subset Q$ such that the fs base change $\operatorname{Bl}_{K_{i}P_{i}}(P_{i})=\operatorname{S}(P_{i})\times_{\operatorname{S}(Q)}\operatorname{Bl}_{K_{i}}(Q)\rightarrow\operatorname{Bl}_{K_{i}}(Q)$ by the corresponding log blow-up is equidimensional. Set $K=K_{1}\cdots K_{n}$. Then $\operatorname{Bl}_{KP_{i}}(P_{i})=\operatorname{S}(P_{i})\times_{\operatorname{S}(Q)}\operatorname{Bl}_{K}(Q)\rightarrow\operatorname{Bl}_{K}(Q)$ is equidimensional for any $i=1,\ldots,n$. One can further perform a toric blow-up of the toric scheme $\operatorname{Bl}_{K}(Q)$ so that the resulting toric scheme is smooth over $\mathbb{Z}$ (cf. [9, p.32, Theorem 11]). Since the composition of fs log blow-ups is again an fs log blow-up ([10, 4.11]), we may assume that there exists an ideal $K\subset Q$ such that * (a) the induced morphism $None$ $\operatorname{Bl}_{KP_{i}}(P_{i})=\operatorname{S}(P_{i})\times_{\operatorname{S}(Q)}\operatorname{Bl}_{K}(Q)\longrightarrow\operatorname{Bl}_{K}(Q)$ is equidimensional for each $i=1,\ldots,n$; * (b) $\operatorname{Bl}_{K}(Q)$ is smooth over $\mathbb{Z}$. We claim that $(\ast)_{i}$ is integral for each $i=1,\ldots,n$. Since toric schemes are Cohen-Macaulay, the properties (a) and (b) imply that the underlying scheme-theoretic morphism of $(\ast)_{i}$ is flat. Thus, for any cones $\sigma$ from the fan of $\operatorname{Bl}_{K}(Q)$ and $\tau$ from the fan of $\operatorname{Bl}_{KP_{i}}(P_{i})$ such that $\tau$ is mapped to $\sigma$, the affine portion of $(\ast)_{i}$ $\operatorname{S}(P^{\prime}_{i})\longrightarrow\operatorname{S}(Q^{\prime})$ where $Q^{\prime}=\sigma^{\prime}\cap M_{Q}$ and $P^{\prime}_{i}=\tau^{\prime}\cap M_{P}$, is flat, and hence is integral by Lemma 2.2 (6) and Lemma 2.5 (2). This means $(\ast)_{i}$ is integral. Now, by Lemma 2.5 (2), we deduce that $U_{i}\rightarrow V$ is integral for any $i=1,\ldots,n$, and hence $X_{V}\rightarrow V$ is integral. ∎ Let us finally remark that, the argument of the above proofs shows that, if we start from a toroidal morphism (in the sense as in [1, §1]) $f\colon X\rightarrow Y$, then, since $f$ is described globally by a morphism of polyhedral complexes $f_{\Delta}\colon\Delta_{X}\rightarrow\Delta_{Y}$ of K. Kato’s fans (cf. [8]), one can actually do the above argument globally on $Y$; cf. [1, 4.4]. Since the only question here lies as to whether one can take a global log blow-up of $Y$, one can slightly generalize the situation to $Y$ being log regular but without assuming $f$ to be toroidal. This situation has been treated in [2], which we include here for the reader’s convenience: ###### Proposition 4.2 ([2, 3.6.11]). Let $f\colon X\rightarrow Y$ be a quasi-compact morphism of fs log schemes, where $Y$ is log regular. Then there exists an fs log blow-up $Y^{\prime}:=\operatorname{Bl}_{\mathcal{K}}(Y)\rightarrow Y$ along a coherent log ideal $\mathcal{K}\subset\mathcal{M}_{Y}$ such that the fs base change $f_{Y^{\prime}}\colon X_{Y^{\prime}}\rightarrow Y^{\prime}$ is integral. ## References * [1] Abramovich, D.; Karu, K.: Weak semistable reduction in characteristic $0$, Invent. Math. 139 (2000), no. 2, 241–273. * [2] Abramovich, D.; Temkin, M.; Wlodarczyk, J.: Relative Desingularization and principalization of ideals, preprint, arXiv:2003.03659. * [3] Gillam, W.D.: Logarithmic flatness, preprint, arXiv:1601.02422. * [4] Illusie, L.; Kato, K.; Nakayama, C.: Quasi-unipotent Logarithmic Riemann-Hilbert Correspondences, J. Math. Sci. Univ. Tokyo 12 (2005), 1–66. * [5] Kato, F.: Exactness, integrality, and log modifications, preprint, arXiv:math/9907124. * [6] Kato, K.: Logarithmic structures of Fontaine–Illusie, in Algebraic Analysis, Geometry and Number Theory (J.-I. Igusa, ed.). Johns Hopkins Univ., 1988, 191–224. * [7] Kato, K.: Logarithmic degeneration and Dieudonne theory, preprint. * [8] Kato, K.: Toric singularities, Amer. J. Math. 116 (1994), no. 5, 1073–1099. * [9] Kempf, G., Knudsen, F., Mumford, D., Saint-Donat, B.: Toroidal Embeddings I, Lecture Note in Math. 339, Springer-Verlag, Berlin, Heidelberg, New York (1973). * [10] Nizioł, W.: Toric singularities: log-blow-ups and global resolutions, Journal of Algebraic Geometry 15 (2006), 1–29. * [11] Ogus, A.: Lectures on Logarithmic Algebraic Geometry, Cambridge University Press, Nov. 2018. * [12] Olsson, M.C.: Logarithmic geometry and algebraic stacks, Ann. Sci. École Norm. Sup. (4) 36 (2003), no. 5, 747–791. * [13] Tsuji, T.: Saturated morphisms of logarithmic schemes, Tunis. J. Math. 1 (2019), no. 2, 185–220. Department of Mathematics, Tokyo Institute of Technology, 2-12-1 Ookayama, Meguro, Tokyo 152-8551, Japan (e-mail<EMAIL_ADDRESS>
Decay width modeling of Higgs boson within THDM model T.V. Obikhod and I.A. Petrenko Institute for Nuclear Research National Academy of Sciences of Ukraine 03068 Kiev, Ukraine e-mail<EMAIL_ADDRESS> Abstract As part of the search for new physics beyond the Standard Model, we chose the determination of the Higgs boson decay width as one of the least experimentally determined values. The decay widths into the four fermions of the lightest and heaviest CP-even Higgs bosons of the THDM model were calculated, taking into account QCD and electroweak corrections in the NLO approximation. To achieve this goal, the program Monte Carlo Prophecy 4f with special scenarios of parameters, 7B1 and 5B1 were used. It was found that the decay width of the heavier CP-even Higgs boson, H differs from HSM by 1227.93 times and changes to a negative value when deviating from the standard scenarios. Scale factors k${}^{2}_{Z}$ and k${}^{2}_{W}$ showed the predominance of the associated with Z boson production cross section of CP- even Higgs boson over the associated with W production cross section. 1\. Introduction In light of the latest experimental data on the searches for new physics beyond the Standard model (SM), Higgs boson remains the only candidate for a window into new physics, [1]. This task is related to the experimental study and theoretical predictions of the properties of the Higgs boson: the production cross sections, the partial decay width, coupling measurements, $k_{i}$. The crucial role for the investigation of the Higgs boson properties is played by the Higgs branching ratios and decay widths, [2]. The Higgs particle is a massive scalar boson with zero spin, no electric charge, and no colour charge is very unstable, decaying immediately into other particles. As all the channels of decay of the Higgs boson as well as possible new particles with certain masses have not yet been studied, there are uncertainties in the properties of the coupling constants and, accordingly, in the decay width of this particle. This fact is demonstrated by the deviation of the predicted SM Higgs decay width of about $4.07\cdot 10^{-3}$ GeV from the experimental data, which are presented in Table 1, [3, 4]. _Table 1._ Run 1 observed (expected) direct 95% CL constraints on the width of the 125 GeV resonance from fits to the $\gamma\gamma$ and ZZ mass spectra. The CMS measurement from the 4l mass line-shape was performed using Run 2 data. Experiment | $M_{\gamma\gamma}$ | $M_{4l}$ | ---|---|---|--- ATLAS | $\prec 5.0(6.2)$ GeV | $\prec 2.6(6.2)$ GeV | CMS | $\prec 2.4(3.1)$ GeV | $\prec 1.1(1.6)$ GeV | The purpose of our paper is to calculate decay widths of lightest, h, and CP- even, H, Higgs bosons of Two Higgs doublet model (THDM), [5] as well as the value of the deviation from SM of the sum of the partial Higgs decay widths compared to the SM, $\kappa_{H}^{2}$, through computer modeling with the help of Monte Carlo program Prophecy 4f 3.0 [6]. 2\. The calculations of decay width and scale factors The Standard Model predicts a very small width of about $4$ MeV for a $126$ GeV Higgs boson. But the error of the energy measurement at the LHC is hundreds of times greater, of the order of $1$ GeV, and it will not be possible to significantly reduce it. As a result, measuring the width of the Higgs boson directly is unrealistic. However, it is possible to accumulate data on the production and decay of the Higgs boson at significantly higher energies - not in the vicinity of $126$ GeV, but, say, above $300$ GeV, [7]. This process will look like the birth and decay of a virtual Higgs boson in this mass range. It is, of course, strongly weakened in comparison with the main process at the resonance peak, but it can be quite measurable. As THDM model predicts the existence of five Higgs bosons, we will carry out our calculations for two bosons: lightest Higgs boson, h and CP-even Higgs boson, H as the analog of virtual Higgs boson described above. Thus, the idea of theorists - to accumulate data on the production and decay of the Higgs boson at significantly higher energies can be realized. The efficiency of this method can be estimated by comparing the calculations of decay widths for the lightest and heaviest bosons. The precise experimental investigation of the Higgs boson and theoretical searches for deviations from the predictions of $SM$ requires precise Monte Carlo computer modeling. Prophecy 4f computes the inclusive partial decay widths and differential distributions of the decay products, where unweighted events for leptonic final states are provided. The advantage of the Prophecy4f program is that it allows the calculations for the Higgs decays into four fermions including full electroweak and QCD next-to-leading order (NLO) corrections with interference contributions between different $WW$/$ZZ$ channels, and inclusion of all off-shell effects of intermediate $W/Z$ bosons. We’ll consider the processes, LO Feynman diagram of which is in the form of Fig. 1 _Fig.1._ _Generic diagram for decay of $H\rightarrow 4f$ where $V=W,Z$, from [6]._ The total state width of Higgs boson is equal to the sum of the partial channel widths [6]: $\Gamma_{H\rightarrow 4f}=\Gamma^{{total}}=\Gamma^{{leptonic}}+\Gamma^{{semi- leptonic}}+\Gamma^{{hadronic}},$ The total width can be presented via $ZZ$, $WW$ decays and their interference: $\Gamma_{H\rightarrow 4f}=\Gamma_{H\rightarrow W^{*}W^{*}\rightarrow 4f}+\\\ +\Gamma_{H\rightarrow Z^{*}Z^{*}\rightarrow 4f}+\Gamma_{WW/ZZ-int}\ ,$ where the components are defined in terms of specific final states: $\Gamma_{H\rightarrow W^{\star}W^{\star}\rightarrow 4f}=9\cdot\Gamma_{H\rightarrow\nu_{e}\overline{e}\mu^{-}\nu_{\mu}}+12\cdot\Gamma_{H\rightarrow\nu_{e}\overline{e}d\overline{u}}+4\cdot\Gamma_{H\rightarrow u\overline{d}s\overline{c}}\ ,$ $\Gamma_{H\rightarrow Z^{*}Z^{*}\rightarrow 4f}=3\cdot\Gamma_{H\rightarrow\nu_{e}\overline{\nu_{e}}\nu_{\mu}\overline{\nu_{\mu}}}+3\cdot\Gamma_{H\rightarrow e\overline{e}\mu\mu^{+}}\\\ \hskip 73.97733pt+9\cdot\Gamma_{H\rightarrow\nu_{e}\overline{\nu_{e}}\mu\mu^{+}}+3\cdot\Gamma_{H\rightarrow\nu_{e}\overline{\nu_{e}}\nu_{e}\overline{\nu_{e}}}\\\ \hskip 73.97733pt+3\cdot\Gamma_{H\rightarrow e\overline{e}e\overline{e}}+6\cdot\Gamma_{H\rightarrow\nu_{e}\overline{\nu_{e}}u\overline{u}}\\\ \hskip 73.97733pt+9\cdot\Gamma_{H\rightarrow\nu_{e}\overline{\nu_{e}}d\overline{d}}+6\cdot\Gamma_{H\rightarrow u\overline{u}e\overline{e}}\\\ \hskip 73.97733pt+9\cdot\Gamma_{H\rightarrow d\overline{d}e\overline{e}}+1\cdot\Gamma_{H\rightarrow u\overline{u}c\overline{c}}\\\ \hskip 73.97733pt+3\cdot\Gamma_{H\rightarrow d\overline{d}s\overline{s}}+6\cdot\Gamma_{H\rightarrow u\overline{u}s\overline{s}}\\\ \hskip 73.97733pt+2\cdot\Gamma_{H\rightarrow u\overline{u}u\overline{u}}+3\cdot\Gamma_{H\rightarrow d\overline{d}d\overline{d}}\ ,$ $\Gamma_{WW/ZZ- int}=3\cdot\Gamma_{H\rightarrow\nu_{e}\overline{e}e\overline{\nu_{e}}}-3\cdot\Gamma_{H\rightarrow\nu_{e}\overline{\nu_{e}}\mu\mu^{+}}-3\cdot\Gamma_{H\rightarrow\nu_{e}\overline{e}\mu\overline{\nu_{\mu}}}+\\\ \hskip 88.2037pt2\cdot\Gamma_{H\rightarrow u\overline{d}d\overline{u}}-2\cdot\Gamma_{H\rightarrow u\overline{u}s\overline{s}}-2\cdot\Gamma_{H\rightarrow u\overline{d}s\overline{c}}\ .$ Using scenarios obtained from the experimental measurements, [8], we presented the calculated NLO results on the four-fermion decays of light $CP$-even Higgs boson, $h$, Table 2 _Table 2._ Decay widths of lightest Higgs boson, $h$. | Full decay width of lightest Higgs boson, $h$, (MeV) | $\Gamma\rightarrow WW$ | $\Gamma\rightarrow ZZ$ | $\Gamma^{int}$ ---|---|---|---|--- 5-B1 | 0.92852 | 0.8326 | 0.1007 | -0.00478 7-B1 | 0.93026 | 0.8311 | 0.104 | -0.00484 We also perform calculations of CP-even Higgs boson with the different parameters presented below, in Table 3 and decay widths, Table 4 _Table 3._ THDM input parameters. | $M_{H}$, GeV | $M_{H+}$, GeV | $M_{A}$, GeV | $\lambda_{5}$ | $\tan\beta$ | $c_{\alpha\beta}$ ---|---|---|---|---|---|--- I | 360 | 690 | 420 | -1.9 | 4.5 | 0.15 II | 600 | 690 | 690 | -1.9 | 4.5 | 0.15 _Table 4._ Decay width of $CP$-even Higgs boson, $H$. Full decay width of lightest Higgs boson, $h$, (MeV) | $\Gamma\rightarrow WW$ | $\Gamma\rightarrow ZZ$ | $\Gamma^{int}$ ---|---|---|--- -54.487 | -71.47 | 17.203 | -0.22 1176.36 | 789.98 | 385.74 | 0.64 The calculations of SM Higgs boson decay width give us the following result, Table 5 _Table 5._ Decay width of SM Higgs boson, $H_{SM}$ Full decay width of lightest Higgs boson, $h$, (MeV) | $\Gamma\rightarrow WW$ | $\Gamma\rightarrow ZZ$ | $\Gamma^{int}$ ---|---|---|--- 0.958 | 0.858 | 0.10724 | -0.00724 In the absence of beyond SM (BSM) Higgs decay modes, total scale factor $\kappa_{H}^{2}$ is the value of the deviation of the sum of the partial Higgs decay widths compared to the SM total width $\Gamma^{SM}_{H}$, [9, 10]: $\kappa_{H}^{2}\left(\kappa_{i},m_{H}\right)=\sum\limits_{{j=WW^{\star},ZZ^{\star},b\overline{b},\tau^{-}\tau^{+},\gamma\gamma,Z\gamma,gg,t\overline{t},c\overline{c},s\overline{s},\mu^{-}\mu^{+}}}\frac{\Gamma_{j}(\kappa_{i},m_{H})}{\Gamma^{SM}_{H}(m_{H})}\ .$ Since the identification of four leptons is the most detectable decay mode in comparison with other decay channels, the optimal direction of the search for new physics will be finding and comparison of the factor $\kappa_{H}^{2}$ for the decays of two Higgs bosons — the lightest and the heaviest one into $WW$ or $ZZ$ bosons. So, the scale factor $\kappa_{H}^{2}$ in this case is the following: $\kappa_{H}^{2}\left(\kappa_{i},m_{H}\right)=\sum\limits_{{j=WW^{\star},\ ZZ^{\star}}}\frac{\Gamma_{j}(\kappa_{i},m_{H})}{\Gamma^{SM}_{H}(m_{H})}\ .$ It is also interesting to calculate scale factors $\kappa_{W}^{2}$ and $\kappa_{Z}^{2}$ $\frac{\Gamma_{WW^{*}}}{\Gamma^{SM}_{WW^{*}}}=\kappa_{W}^{2},$ $\frac{\Gamma_{ZZ^{*}}}{\Gamma^{SM}_{ZZ^{*}}}=\kappa_{Z}^{2},$ which allow probing for BSM contributions in the loops for each channel separately. Moreover, these factors make it possible to calculate the deviations from the SM of the associated production cross sections in accordance with the formulas: $\frac{\sigma_{WH}}{\sigma^{SM}_{WH}}=\kappa_{W}^{2},$ $\frac{\sigma_{ZH}}{\sigma^{SM}_{ZH}}=\kappa_{Z}^{2}.$ The results of our calculations are performed in the Table 6: _Table 6._ Scaling factors of two Higgs bosons Higgs boson | Scenario | $\kappa_{W}^{2}$ | $\kappa_{Z}^{2}$ | $\kappa_{H}^{2}$ | ---|---|---|---|---|--- | $\kappa_{H}^{2}$ | | | w/o int | | w int | | | | h | 5-B1 | 0.97 | 0.939 | 0.967 | 0.969 h | 7-B1 | 0.968 | 0.97 | 0.969 | 0.971 H | II | 921 | 3597 | 1218 | 1228 From the comparison of the data from Table 6 we see the slight change in $\kappa_{H}^{2}$ factor for 5-B1 and 7-B1 scenarios and huge increase compared to SM one for scenario II. Moreover, we can see the increasing of $\kappa_{H}^{2}$ factor for all scenarios with inclusion of interference. The BSM contributions in the loops for $WW$ channel are larger in 5-B1 scenario but for 7-B1 scenario the larger contribution in the loops are for $ZZ$ channel. Therefore, the chose of renormalization schema is also essential to the final result. The sharp jump in the $\kappa_{H}^{2}$ factor for heavier $CP$-even Higgs boson indicates about significant deviation from the $SM$ for scenario 7-B1. The difference in factor $\kappa_{Z}^{2}$ compared to $\kappa_{W}^{2}$ by almost four times indicates the predominance of the associated with $Z$ boson production cross section of $CP$-even Higgs boson over the associated with $W$ production cross section. 3\. Conclusions The searches for BSM physics are connected with studying of Higgs boson properties. The way of the realization of this purpose is connected with the decay widths measurements and theoretical predictions of Higgs boson properties. The most perspective and convenient Higgs boson decay channel into four fermions is one of the interesting way of its investigation. For the precise measurements of the decay width is proposed THDM model in the paper. We have considered lightest and CP-even heavier Higgs bosons, h and H correspondingly and modeled their decay widths into four fermions with the help of Monte Carlo program Prophecy 4f 3.0. The results of our calculations led us to the following conclusions connected with the searches of deviations from SM: * • decay widths of lightest Higgs boson, $h$ and $H_{SM}$ almost do not differ from each other; * • the scale factor $\kappa_{H}^{2}$ of $CP$-even Higgs boson, $H$ equal to 1228; * • the calculations of decay widths strongly depend on the parameter space and can take negative values as the masses of the $CP$-even and $CP$-odd Higgs bosons decrease by almost two times from the parameters of the 7B1 scenario; * • the interference account leads to an insignificant increase in decay widths; * • the difference in factor $\kappa_{Z}^{2}$ compared to $\kappa_{W}^{2}$ by almost four times indicates the predominance of the associated with $Z$ boson production cross section of $CP$-even Higgs boson, $H$ over the associated with $W$ production cross section. * • BSM contributions in the loops for $WW$ and $ZZ$ channels are vary depending on renormalization schema. ## References * [1] CERNweb. The Higgs boson as a probe for new physics // URL: https://ep-news.web.cern.ch/higgs-boson-probe-new-physics. * [2] M. Spira. Higgs boson production and decay at hadron colliders // Progress in Particle and Nuclear Physics 2017, V.95, p. 98-159. * [3] ATLAS Collaboration. Measurement of the Higgs boson mass from the $H{\rightarrow}{\gamma}{\gamma}$ and $H{\rightarrow}Z{Z}^{*}{\rightarrow}4l$ channels in $pp$ collisions at center-of-mass energies of 7 and 8 TeV with the ATLAS detector // Phys. Rev. D 2014, V.90, p. 052004. * [4] CMS Collaboration. Measurements of properties of the Higgs boson decaying into four leptons in pp collisions at $\sqrt{s}=13$ TeV // CMS-PAS-HIG-16-041, 2017. * [5] G.C. Branco et al. Theory and phenomenology of two-Higgs-doublet models // Phys. Rep. 2012, V.156, p. 1-102. * [6] A. Denner, S. Dittmaier, A. Muck. Prophecy4f 3.0: A Monte Carlo program for Higgs-boson decays into four-fermion final states in and beyond the Standard Model // Comput. Phys. Commun. 2020, V. 254, p. 107336. * [7] elementy$\\_ru$ // URL:https://elementy.ru/novosti$\\_nauki/432235.$ * [8] A. Denner, S. Dittmaier, J.-N. Lang. Renormalization of mixing angles // JHEP 2018, V. 11, p. 104, arXiv:1808.03466 [hep-ph]. * [9] The LHC Higgs Cross Section Working Group. Handbook of LHC Higgs Cross Sections: 3. Higgs Properties // CERN-2013-004, arXiv:1307.1347 [hep-ph] * [10] D. de Florian et al. Handbook of LHC Higgs Cross Sections: 4. Deciphering the Nature of the Higgs Sector // CERN Yellow Reports: Monographs Volume 2/2017 (CERN–2017–002-M), arXiv:1610.07922 [hep-ph].
# Quantum theory cannot violate a causal inequality Tom Purves<EMAIL_ADDRESS>H.H. Wills Physics Laboratory, University of Bristol, Tyndall Avenue, Bristol, BS8 1TL, U.K. Anthony J. Short <EMAIL_ADDRESS>H.H. Wills Physics Laboratory, University of Bristol, Tyndall Avenue, Bristol, BS8 1TL, U.K. ###### Abstract Within quantum theory, we can create superpositions of different causal orders of events, and observe interference between them. This raises the question of whether quantum theory can produce results that would be impossible to replicate with any classical causal model, thereby violating a causal inequality. This would be a temporal analogue of Bell inequality violation, which proves that no local hidden variable model can replicate quantum results. However, unlike the case of non-locality, we show that quantum experiments _can_ be simulated by a classical causal model, and therefore cannot violate a causal inequality. _Introduction_.– A fascinating aspect of quantum theory that has been investigated recently is the possibility for the causal order of events to be placed into superposition Chiribella2013 ; Chiribella2012 ; Branciard2016 ; Zych2019 ; Barrett2020 , leading to ‘causal indefiniteness’ about the order with which events have taken place. This phenomenon has been tested experimentally Procopio2015 ; White2018 ; Goswami2020 , and can be exploited to gain advantages within quantum theory. For example, setups based on the quantum switch Chiribella2013 can help to determine whether unknown unitaries commute or anticommute Chiribella2012 . An interesting question is whether quantum theory can generate results which could not be simulated by any classical causal model. Such results would violate a Causal Inequality Oreshkov2012 ; Brukner2014 ; Branciard2015 ; Abbott2016 . These are the temporal analogues of Bell Inequalities Bell1964 , and the violation of such an inequality in nature would call into question the elementary properties that scientists regularly invoke when talking about cause and effect relationships. In this paper we focus on the relationship between the type of causal indefiniteness present in quantum theory and the type needed to violate causal inequalities. We show that despite allowing causally indefinite processes, the correlations generated by quantum theory can be simulated by a classical causal model. This means that quantum theory cannot violate causal inequalities, and hence cannot yield an advantage over classical causal processes for tasks defined in a theory-independent way (such as ‘guess your neighbour’s input’ Almeida2010 ; Branciard2015 ). Previous works in this direction have shown that particular switch-type scenarios cannot violate causal inequalities Arajo2015 , and that causal order cannot be placed in a pure superposition Costa2020 ; Yokojima2021 . It has also been shown that causal inequality violations are possible when we condition on measurement outcomes of one party Milz2020 . However, our results imply that such violations are not possible for general quantum setups without conditioning. Indefinite causal structure is often studied via process matrices Oreshkov2012 , which assume that local laboratories obey standard quantum theory, but allow any connections between them consistent with this. This may include processes which are not achievable in standard quantum theory, or in nature more generally. Here we focus on what is possible in standard quantum theory, using quantum control of different parties’ operations to generate superpositions of causal order, in a similar way to Colnaghi2012 ; Araujo2014 . As process matrices can yield causal inequality violation, a corollary of our result is that all process matrices cannot be implemented in standard quantum theory. _Results_.– Before considering quantum processes, we first define causal processes, which are those which could be realised classically by a set of parties in separate laboratories passing systems between them pearl ; Barrett2020 . First consider two parties, Alice and Bob, with measurement settings $x$ and $y$ and measurement results $a$ and $b$ respectively. During the experiment, depicted in figure 1, each party sees a system enter their laboratory exactly once, performs a measurement on it with their corresponding measurement setting (which may also modify the system), and records their result. They then pass the system out of their laboratory. Apart from the systems entering and leaving their laboratories, the two parties cannot communicate with each other, but the system leaving one laboratory may be later sent into the other. Alice and Bob’s joint measurement results can be described by a conditional probability distribution $p(ab|xy)$. However, not all such probability distributions can be achieved by a causal process. Figure 1: An example of a causal process in which Alice goes before Bob. Note that the system which is passed from Alice’s to Bob’s laboratory could encode information about $a$ and $x$. The most general causal process in this case would be to first choose randomly whether Alice or Bob would go first (with probabilities $p(\textrm{Alice first})$ or $p(\textrm{Bob first})$). If Alice goes first, then her measurement result can depend on her measurement setting but not on Bob’s, who hasn’t acted yet, so is given by $p(a|x)$. She can then encode her measurement setting and result in the system and pass it out of her laboratory. This system then enters Bob’s laboratory, where his measurement result can depend on all of the other variables, given by $p(b|a,x,y)$. Considering the other causal order in which Bob goes first in the same way, we obtain Branciard2015 $\displaystyle p^{\mathrm{causal}}(ab|xy)=$ $\displaystyle p(\textrm{Alice first})p(a|x)p(b|a,x,y)$ $\displaystyle+p(\textrm{Bob first})p(b|y)p(a|b,x,y)$ (1) For the multiparty generalisation Oreshkov2016multi ; Abbott2016 , observe that the above causal probability contains two types of terms. The first, such as $p(\textrm{Alice first})$, determines the order in which the parties act, and the second, such as $p(a|x)$ or $p(b|a,x,y)$, determines the outcome probabilities of their measurements, constrained by their causal order. We now extend these ‘who is next?’ and ‘what did they see?’ type probabilities to an arbitrary number of parties. We use $l_{k}$ to denote the $k{\textrm{th}}$ party that receives the system (or equivalently, the $k{\textrm{th}}$ laboratory the system enters), and denote the probability for this to occur by $p_{k}(l_{k}|H_{k-1})$. The conditional on $H_{k-1}$ represents the history (including all previous parties that have measured, and their inputs and outputs) for it should be permitted for parties in the causal past of $l_{k}$ to affect who is the next party to act. As a simple example of this, consider a tripartite experiment, with Alice, Bob and Charlie participating. If Charlie comes first, the system could be passed to Alice or Bob next, based on the outcome of his measurement. Here, $p_{2}(\text{Alice next}|\text{Charlie got outcome = }1)$ may not be equal to $p_{2}(\text{Alice next}|\text{Charlie got outcome = }0)$. Scenarios of this form this are what $p_{k}(l_{k}|H_{k-1})$ accounts for. The probability for $l_{k}$ to obtain given results may also depend on this history (but, importantly, not on the causal future), and of course on the measurement setting, denoted $x_{l_{k}}$. We write this probability as $p_{k}(a_{l_{k}}|H_{k-1},x_{l_{k}})$. A causal model is then the summation over all available parties at all stages of the measurement procedure, under the assumption that each party only acts once in the entire procedure. ###### Definition 1 A causal probabilistic model can be written as $\displaystyle p^{\mathrm{causal}}(\vec{a}|\vec{x})=\sum_{l_{1}\notin\mathcal{L}_{0}}...\sum_{l_{N}\notin\mathcal{L}_{N}-1}$ $\displaystyle p_{1}(l_{1}|H_{0})p_{1}(a_{l_{1}}|H_{0},x_{l_{1}})...$ $\displaystyle\quad...p_{N}(l_{N}|H_{N-1}$ $\displaystyle)p_{N}(a_{l_{N}}|H_{N-1},x_{l_{N}})$ (2) where the $p_{k}(l_{k}|H_{k-1})$ terms represent probabilities for party $l_{k}$ to act at stage $k$ of the causal order, and $p_{k}(a_{l_{k}}|H_{k-1},x_{l_{k}})$ terms represent probabilities for party $l_{k}$, who has acted at stage $k$ of the causal order to obtain measurement result $a_{l_{k}}$. Both of the above probabilities are conditional on a history, $H_{k-1}$, which contains all of the information about previous inputs, outputs and party order. In particular, the history $H_{k}=(h_{1},...,h_{k})$ is the ordered list of triples $h_{i}=(l_{i},a_{l_{i}},x_{l_{i}})$. The summations are performed over all possible next parties, excluding parties who have already acted, which are stored in the unordered sets $\mathcal{L}_{k}=\\{l_{1},...,l_{k}\\}$. To emphasise the symmetry between the terms we include $H_{0}$ and $\mathcal{L}_{0}$, which are defined as empty sets, as no parties have acted at that point. This definition leads to a convex polytope of causal probability distributions $p^{\mathrm{causal}}(\vec{a}|\vec{x})$. Note that although the notation differs, this generates the same set of probabilities as was previously defined in Abbott2016 ; Oreshkov2016multi . Linear constraints on these probabilities which are satisfied by all $p^{\mathrm{causal}}(\vec{a}|\vec{x})$ but which could be violated by some arbitrary probability distribution $p(\vec{a}|\vec{x})$ are known as ‘causal inequalities’, and are a temporal analogue of the Bell inequalities which have been widely studied in the context of quantum non-locality. By definition, any $p^{causal}(a,b|x,y)$ cannot violate a causal inequality. A violation of a causal inequality, by observation in experiment or by calculation in theory, proves that those experimental results or predictions do not have a causal explanation of the type defined above. _Quantum Processes_.– A general representation of quantum theory is provided by the quantum circuit model. However, if we construct a circuit with the parties’ actions at fixed locations, then there is no causal indefiniteness and a causal inequality cannot be violated 111We could space out the circuit such that there is at most one lab at each time-step, and then pass the full quantum state between the labs as a classical hidden variable which would allow us to recover the same correlations in a causal way. Even to capture all classical causal processes, we need to be able to alter when different parties act. This can be achieved in the circuit model by representing the parties’ actions by controlled quantum gates. Such gates could be constructed within standard quantum theory (e.g. by sending a system into the lab when the control is in the appropriate state and not otherwise), and are effectively what has been used in experiments probing quantum causality White2018 . Quantum circuits involving controlled lab gates appear sufficient to represent any processes achievable within standard quantum theory. For simplicity, we consider a setup involving a single quantum control which can trigger any of the labs. However this is equivalent to considering any quantum circuit which can be constructed from any number of individual controlled lab operations and other unitary gates (see the appendix for more details). The key idea is to consider $N$ parties, each of whom will interact with a quantum system exactly once, but in an order that is controlled coherently via the quantum control. We allow arbitrary unitary transformations of the system and control between each party’s action, so that the ordering of later parties can be modified by earlier actions. Figure 2: Illustration of the quantum protocol. The system interacts with the different parties via a sequence of controlled entangling unitaries. Quantum control of causal order is achieved by a series of unitaries $U_{n}$ on the control and system wires. To allow the maximum possible interference, and avoid ‘collapses’ which would prevent interference between different causal orders, we model each party’s measurement as a unitary interaction between the system and a local measurement register. This corresponds to the case in which there is no record in the measuring device of the time at which the measurement was performed. At the end of the experiment, all parties read off their measurement results from their local measurement registers (which can be modelled by a standard projective measurement). Each party also has a ‘flag’ which keeps track of how many times they have interacted with the system. At the end of the protocol we require that each party has interacted with the system exactly once. Formally, the Hilbert space can be decomposed into the following components * • An arbitrary quantum system $\mathcal{H}_{s}$, which is passed between parties. * • A quantum control $\mathcal{H}_{c}$ which has dimension $N+1$. The basis states $\ket{1}\ldots\ket{N}$ denote which party will measure next, while $\ket{0}$ is treated as a ‘do nothing’ command. By considering superpositions of these basis states, we can superpose different causal orders. * • A result register $\mathcal{H}_{r_{i}}$ for each of the $N$ parties. The different results are represented by orthonormal basis states $\ket{a_{i}}$ with $a_{i}\in\mathcal{A}_{i}$, leading to the result register having dimension $|\mathcal{A}_{i}|$. We choose one of these states as a starting state for the results register and denote it by $\ket{0}_{r_{i}}$. * • A ‘flag’ $\mathcal{H}_{f_{i}}$ for each of the $N$ parties, indicating how many times they have interacted with the system. For simplicity, we take each of these to be infinite dimensional, with basis states labelled by the integers. When the party interacts with the system the value of the flag is unitarily raised by the operator $\Gamma=\sum_{n}\ket{n+1}_{f_{i}}\bra{n}_{f_{i}}$. Each flag starts in the $\ket{0}_{f_{i}}$ state, and at the end of the protocol, we require them all to be in the $\ket{1}_{f_{i}}$ state. Note that we do not include separate local quantum ancillas for the parties, as these can always be incorporated in $\mathcal{H}_{s}$. We denote the combined result and flag spaces by $\mathcal{H}_{r}=\bigotimes{\mathcal{H}_{r_{i}}}$ and $\mathcal{H}_{f}=\bigotimes{\mathcal{H}_{f_{i}}}$ respectively. We consider quantum protocols as follows. Firstly, the initial state $\displaystyle\ket{0}=\ket{0}_{s}\ket{0}_{c}\ket{0}_{r}^{N}\ket{0}_{f}^{N}\in\mathcal{H}_{s}\otimes\mathcal{H}_{c}\otimes\mathcal{H}_{r}\otimes\mathcal{H}_{f}.$ (3) is prepared, and each party $l$ either chooses or is distributed their individual classical measurement setting $x_{l}$. The protocol then consists of $T$ time-steps, each of which is composed of two operations. Firstly, an arbitrary unitary transformation $U_{t}$ is applied to the system and control, which can depend on the time $t$. Secondly, a fixed controlled lab-activation unitary $V$ is applied, which activates whichever party is specified by the control. This is given by $V=\ket{0}\bra{0}_{c}\otimes I+\sum_{l=1}^{N}\ket{l}\bra{l}_{c}\otimes V_{s,r_{l}}(x_{l})\otimes\Gamma_{f_{l}}\otimes I$ (4) where the identities are over all remaining subsystems. $V_{s,r_{l}}(x_{l})$ is a unitary which implements the measurement of party $l$ on the system specified by the measurement setting $x_{l}$, and stores the result in the register $r_{l}$. For example, two different values of $x_{l}$ could correspond to party $l$ measuring the system in either the computational or the Fourier basis. Note that by incorporating ancillas within the system, any local quantum measurement (i.e a POVM) is realisable within this paradigm. Ancillas can also be used to generate arbitrary mixed states if required (via purification). The unitary operator $\Gamma_{f_{l}}$ raises the flag system of the party making their measurement. At the end of the protocol, we require that the flags are in the state $\ket{1}_{f}^{N}$ (i.e. that each party has measured the system once). This places constraints on the possible protocols which can be constructed. Note that each party does not have access to an operation which resets the flag, aside from the initialisation operation at the start of the protocol. They therefore always ‘remember’ if they have made a measurement or not. Also, we do not allow circuits involving the controlled inverse of a party’s action (which would lower their flag and erase their memory), as this would enlarge the set of causal possibilities even classically. The total unitary for the protocol is given by $\mathcal{U}=VU_{T}VU_{T-1}...VU_{1}.$ (5) At the end of the protocol, each party performs a projective measurement on their results register to obtain their final result 222Note that from a many- worlds perspective Everett such an additional step would not be necessary. However, we include it here to maintain connection with standard quantum theory and give an explicit formula for the outcome probabilities. The output probability distribution of the quantum protocol is therefore given by $\displaystyle p^{\mathrm{quantum}}(\vec{a}|\vec{x})=|(\ket{\vec{a}}\bra{\vec{a}}_{r}\otimes I)\mathcal{U}\ket{0}|^{2}.$ (6) The full protocol is illustrated as a quantum circuit in figure 2. The main result of this paper is that any probability distribution which can be generated within quantum theory, as described above, can also be obtained via a classical causal process. ###### Theorem 1 Any quantum probability distribution $p^{\mathrm{quantum}}(\vec{a}|\vec{x})$ can be exactly replicated by a classically causal process $p^{\mathrm{causal}}(\vec{a}|\vec{x})$. Hence quantum theory cannot violate a causal inequality. In particular, we now show how to construct an explicit classical causal process which replicates the results of any quantum protocol, together with a sketch of the proof of Theorem 1. The full proof of the theorem can be found in the appendix. We first define notation for describing states at each stage of the quantum protocol, and then show how to use these to construct the probabilities in the corresponding classical model. ###### Definition 2 The (un-normalised) state with a History $H_{k-1}$, at a time $t$, with the control set to trigger the action of party $l_{k}$ is given by $\displaystyle\ket{\psi_{(l_{k},t,H_{k-1})}}=(\ket{l_{k}}\bra{l_{k}}_{c}\\!\otimes\\!\pi^{H_{k-1}}_{rf}\\!\otimes\\!I_{s})U_{t}VU_{t-1}...VU_{1}\ket{0}.$ (7) The projector onto the result and flag spaces is given by $\pi^{H_{k-1}}_{rf}=\bigotimes_{i=1}^{N}\left(\pi^{H_{k-1}}_{r_{i}f_{i}}\right),$ where $\displaystyle\pi^{H_{k-1}}_{r_{i}f_{i}}=\begin{cases}\ket{a_{i}}\bra{a_{i}}_{r_{i}}\otimes\ket{1}\bra{1}_{f_{i}}&\text{ if }(i,a_{i},x_{i})\in H_{k-1},\\\ I_{r_{i}}\otimes\ket{0}\bra{0}_{f_{i}}&\text{ otherwise }.\end{cases}$ (8) This notation describes states which are about to be measured by the parties (i.e., a $V$ type operator is about to act on them). We also set up some notation for states which have just been measured, in a similar fashion. ###### Definition 3 The (un-normalised) state with a History $H_{k}$, at a time $t$, in which party $l_{k}$ has just acted is given by $\displaystyle\ket{\phi_{(l_{k},t,H_{k})}}=(\ket{a_{l_{k}}}\bra{a_{l_{k}}}_{r_{k}}\otimes I)V\ket{\psi_{(l_{k},t,H_{k-1})}}.$ (9) With these definitions, we can associate the states in this quantum process with the probabilities in our classical causal model. ###### Definition 4 The probability for party $l_{k}$ to act next, given a history $H_{k-1}$ is given by: $\displaystyle p_{k}(l_{k}|H_{k-1})=\frac{\sum_{t_{k}=1}^{T}|\ket{\psi_{(l_{k},t_{k},H_{k-1})}}|^{2}}{\sum_{l_{k}^{\prime}\notin\mathcal{L}_{k-1}}\sum_{t_{k}^{\prime}=1}^{T}|\ket{\psi_{(l_{k}^{\prime},t_{k}^{\prime},H_{k-1})}}|^{2}}.$ (10) We have summed over time Note2 , because it is possible within the quantum paradigm to conduct the $k^{th}$ measurement at different times according to a background clock (which we note the labs have no access to). Note that states at different times combine incoherently, but different sequences leading to the same set of historical measurement results combine coherently inside $\ket{\psi_{(l_{k},t_{k},H_{k-1})}}$. The form of equation (10) makes it a valid probability distribution, as it is non-negative, and obeys the correct normalisation that $\sum_{l_{k}\notin\mathcal{L}_{k-1}}p_{k}(l_{k}|H_{k-1})=1$. Also note that it depends on only those input variables $x_{i}$ which appear in the history $H_{k-1}$. Next, we specify similar probabilities for seeing measurement results based on a given history. ###### Definition 5 The probability for party $l_{k}$ to obtain the measurement result $a_{l_{k}}$, given a history $H_{k-1}$, and an input variable $x_{l_{k}}$ is given by: $\displaystyle p_{k}(a_{l_{k}}|H_{k-1},x_{l_{k}})=\frac{\sum_{t_{k}=1}^{T}|\ket{\phi_{(l_{k},t_{k},H_{k})}}|^{2}}{\sum_{a^{\prime}_{l_{k}}\in\mathcal{A}_{l_{k}}}\sum_{t_{k}^{\prime}=1}^{T}|\ket{\phi_{(l_{k},t_{k}^{\prime},H_{k}^{\prime})}}|^{2}},$ (11) where $H_{k}=(H_{k-1},(l_{k},a_{l_{k}},x_{l_{k}}))$ and $H_{k}^{\prime}=(H_{k-1},(l_{k},a^{\prime}_{l_{k}},x_{l_{k}}))$ This is again a valid probability distribution, since $\sum_{a_{l_{k}}\in\mathcal{A}_{l_{k}}}p_{k}(a_{l_{k}}|H_{k-1},x_{l_{k}})=1$. In the numerator, we have simply taken sum of the modulus squared of all of the states which have the correct historical results, the control in the correct state, and the results register containing the result we want to calculate the probability for. To prove Theorem 1, We begin by inserting $p_{k}(l_{k}|H_{k-1})$ (from (10)) and $p_{k}(a_{l_{k}}|H_{k-1},x_{l_{k}})$ (from (11)) into the definition of a causal model (1). We are then able to straightforwardly cancel the numerator of the ‘who is next?’ type probabilities with the denominator of the ‘what did they see?’ probabilities for the probabilities evaluated at the same stage of the causal order. Next, we show that a sum over the last party to measure in the numerator at one stage of the causal order, cancels with the denominator at the next stage of the causal order. We then make the observation that for the first stage of the causal order, the denominator of $p_{1}(l_{1}|H_{0})$ is equal to one (which corresponds to the fact that someone must measure first in the quantum circuit). Finally, we note that the numerator of the final term, summed over all parties, represents exactly the probabilities $p^{\mathrm{quantum}}(\vec{a}|\vec{x})$ arising from the quantum protocol. This allows us to simulate the results of the quantum protocol via the classically causal model given in (1). Given that it can be replicated by a causal model, it follows that quantum theory cannot violate a causal inequality. In the appendix, we give an example of how these results can be applied in practice, based on the quantum switch Chiribella2013 . This involves the causal order of two parties becoming entangled with the control. A third party then performs a measurement which leads to interference between the two causal orders. It has already been shown that this simple setup cannot be used to violate a causal inequality Arajo2015 ; Oreshkov2012 . However, it is instructive to see how it fits into our framework. Despite the quantum setup including interference, our results give an explicit classical causal process which generates the same behaviour (i.e. the same $p(a,b,c|x,y,z)$). _Conclusions_.– By using a quantum control to determine when different parties measure, and treating these measurements as coherent unitary processes, quantum theory allows us to generate superpositions of causal orders and to observe interference between them. At the level of the theory, such processes do not arise from a single causal order, or even a mixture of orders. However, we have shown that the probabilities $p(\vec{a}|\vec{x})$ generated by any quantum protocol _can_ be simulated by a classical causal process. This means that quantum theory cannot violate a causal inequality, and thus one could not convince a sceptic that nature deviates from classical notions of causality. This is in sharp distinction to non-locality, where not only does the theory appear non-local (e.g. via entangled states) but we can also prove that some quantum probabilities cannot be replicated by any local hidden variable model. By violating a Bell inequality we can therefore prove non-locality experimentally. Although our framework is very general, one key requirement is that each party interacts with the system once (which leads to a requirement on the final flag state). This is the normal setup for causal inequalities, and allows us to assign a single input and output to each party, and to represent the experimental results via $p(\vec{a}|\vec{x})$. However, it would be interesting to lift this assumption in future research. For example, could we obtain a violation of causality if parties are allowed to measure twice, or a variable number of times, or to forget they have measured? We also have a technical assumption that the protocol takes finite time (i.e. that it terminates after a finite number of steps). This seems physically reasonable, but it might be interesting to investigate lifting this assumption, as well as to consider extending the results to continuous time. Finally, it would be interesting to consider a network structure in the causal scenario, in the non-local case this is known to generate non-linear Bell inequalities, and sets of non convex probability polytopes Brunner2020 . Investigation of causal indefiniteness and causal inequalities in these type of scenarios might prove of general interest. Finally, our framework assumes standard quantum theory. If the theory changes significantly to incorporate quantum gravity we might expect new possibilities for causal inequality violation, although not necessarily Zych2019 (note that even classical general relativity allows for the existence of closed time-like curves, which appear to violate the simple classical causal models we have considered here Lobo2010 ; Barrett2020cyclic ; Arajo2017 ). We hope that the framework and tools developed here will prove helpful in discussing these interesting issues, and in highlighting differences from the standard case. T.Purves acknowledges support from the EPSRC. _Note added -_ Independently obtained related results using the process matrix formalism Wechs2021 appeared on the ArXiv on the same day as this paper. ## References * (1) Giulio Chiribella, Giacomo Mauro D’Ariano, Paolo Perinotti, and Benoit Valiron. Quantum computations without definite causal structure. Physical Review A, 88(2), 2013. * (2) Giulio Chiribella. Perfect discrimination of no-signalling channels via quantum superposition of causal structures. Phys. Rev. A, 86:040301(R), Oct 2012. * (3) Cyril Branciard. Witnesses of causal nonseparability: an introduction and a few case studies. Scientific Reports, 6(1):26018, May 2016. * (4) Magdalena Zych, Fabio Costa, Igor Pikovski, and Časlav Brukner. Bell’s theorem for temporal order. Nature Communications, 10(1):3772, Aug 2019. * (5) Jonathan Barrett, Robin Lorenz, and Ognyan Oreshkov. Quantum causal models, 2020. arXiv:1906.10726. * (6) Lorenzo M. Procopio, Amir Moqanaki, Mateus Araújo, Fabio Costa, Irati Alonso Calafell, Emma G. Dowd, Deny R. Hamel, Lee A. Rozema, Časlav Brukner, and Philip Walther. Experimental superposition of orders of quantum gates. Nature Communications, 6(1), Aug 2015. * (7) K. Goswami, C. Giarmatzi, M. Kewming, F. Costa, C. Branciard, J. Romero, and A. G. White. Indefinite causal order in a quantum switch. Phys. Rev. Lett., 121:090503, 2018. * (8) K. Goswami and J. Romero. Experiments on quantum causality. AVS Quantum Science, 2(3):037101, Oct 2020. * (9) Aram W. Harrow, Avinatan Hassidim, Debbie W. Leung, and John Watrous. Adaptive versus nonadaptive strategies for quantum channel discrimination. Phys. Rev. A, 81:032339, Mar 2010. * (10) Daniel Ebler, Sina Salek, and Giulio Chiribella. Enhanced communication with the assistance of indefinite causal order. Phys. Rev. Lett., 120:120502, 2018. * (11) Guérin, Philippe Allard and Rubino, Giulia and Brukner, Časlav Communication through quantum-controlled noise Phys. Rev. A, 99.062317, Jun 2019. * (12) Rubino, Giulia and Rozema, Lee A. and Ebler, Daniel and Kristjánsson, Hlér and Salek, Sina and Allard Guérin, Philippe and Abbott, Alastair A. and Branciard, Cyril and Brukner, Časlav and Chiribella, Giulio and et al. Experimental quantum communication enhancement by superposing trajectories Physical Review Research, 3.013093, Jan 2021. * (13) Yokojima, Wataru and Quintino, Marco Túlio and Soeda, Akihito and Murao, Mio Consequences of preserving reversibility in quantum superchannels Quantum, q-2021-04-26-441, Apr 2021. * (14) Araújo, Mateus and Guérin, Philippe Allard and Baumeler, Ämin Quantum computation with indefinite causal structures Physical Review A, 96.052315, Nov 2017. * (15) Ognyan Oreshkov, Fabio Costa, and Časlav Brukner. Quantum correlations with no causal order. Nature Communications, 3(1):1092, 2012. * (16) Časlav Brukner. Quantum causality. Nature Physics, 10(4):259–263, 2014. * (17) Cyril Branciard, Mateus Araújo, Adrien Feix, Fabio Costa, and Časlav Brukner. The simplest causal inequalities and their violation. New Journal of Physics, 18(1):013008, 2015. * (18) Alastair A. Abbott, Christina Giarmatzi, Fabio Costa, and Cyril Branciard. Multipartite causal correlations: Polytopes and inequalities. Physical Review A, 94(3), Sep 2016. * (19) J. S. Bell. On the einstein podolsky rosen paradox. Physics Physique Fizika, 1:195–200, 1964. * (20) Mafalda L. Almeida, Jean-Daniel Bancal, Nicolas Brunner, Antonio Acín, Nicolas Gisin, and Stefano Pironio. Guess your neighbor’s input: A multipartite nonlocal game with no quantum advantage. Physical Review Letters, 104(23), Jun 2010. * (21) Mateus Araújo, Cyril Branciard, Fabio Costa, Adrien Feix, Christina Giarmatzi, and Časlav Brukner. Witnessing causal nonseparability. New Journal of Physics, 17(10):102001, Oct 2015. * (22) Fabio Costa. A no-go theorem for superpositions of causal orders, 2020. arXiv:2008.06205. * (23) Simon Milz, Dominic Jurkschat, Felix A. Pollock, and Kavan Modi. Quantum chicken-egg dilemmas: Delayed-choice causal order and the reality of causal non-separability, 2020. arXiv:2008.07876. * (24) Timoteo Colnaghi, Giacomo Mauro D’Ariano, Stefano Facchini, and Paolo Perinotti. Quantum computation with programmable connections between gates. Physics Letters A, 376(45):2940 – 2943, 2012. * (25) Mateus Araújo, Fabio Costa, and Časlav Brukner. Computational advantage from quantum-controlled ordering of gates. Phys. Rev. Lett., 113:250402, Dec 2014. * (26) Judea Pearl. Causality: Models, Reasoning and Inference. Cambridge University Press, USA, 2nd edition, 2009. * (27) Ognyan Oreshkov and Christina Giarmatzi. Causal and causally separable processes. New Journal of Physics, 18(9):093020, 2016. * (28) Francisco S. N. Lobo. Closed timelike curves and causality violation, 2010. arXiv:1008.1127. * (29) Jonathan Barrett, Robin Lorenz, and Ognyan Oreshkov. Cyclic quantum causal models, 2020. arXiv:2002.12157. * (30) Hugh Everett. ”relative state” formulation of quantum mechanics. Rev. Mod. Phys., 29:454–462, Jul 1957. * (31) Note that from a many-worlds perspective Everett such an additional step would not be necessary. However, we include it here to maintain connection with standard quantum theory and give an explicit formula for the outcome probabilities. * (32) Here, we summed from $t=1$ to $T$, but we could have equivalently chosen $t=k$ to $T-(N-k)$, and only went with the former to make the subsequent notation easier to read, at the cost of including $0$ amplitude states in our probability definition. * (33) Nicolas Gisin, Jean-Daniel Bancal, Yu Cai, Patrick Remy, Armin Tavakoli, Emmanuel Zambrini Cruzeiro, Sandu Popescu, Nicolas Brunner. Constraints on nonlocality in networks from no-signaling and independence. Nature Communications, 11(1):2378, 2020 * (34) Julian Wechs and Hippolyte Dourdent and Alastair A. Abbott and Cyril Branciard. Quantum circuits with classical versus quantum control of causal order arXiv:2101.08796. ## Appendix A Equivalence of combined and individual controlled lab gates In this section we show that the framework for quantum processes in the main text is equivalent to considering any quantum circuit built up of standard unitary gates and controlled gates for individual laboratories, in terms of the probability distributions they can generate. The key ingredient is to show how to construct individual controlled lab gates from $V$, and conversely how to construct the operation $V$ from individual controlled lab gates. To map any circuit involving individual controlled lab gates into our framework, we first space out the gates in the circuit, so that there is only one gate per time-step (this will increase the depth, but not affect the results). If an individual lab gate acts on only part of the system, we extend it such that it acts on the entire system, taking the action to be trivial (i.e. tensored with the identity) on any part of the system which was not initially included. We can then replace each individual controlled lab gate by a circuit fragment involving one use of $V$, using the approach described below. Finally we merge all unitary gates between instances of $V$ into the unitaries $U_{t}$. This will lead to a circuit in our framework yielding exactly the same results as the original circuit. To go in the other direction, we simply replace each instance of $V$ with its construction in terms of individual controlled lab gates. Figures 1 and 2 show how to construct an individual controlled lab gate using $V$. Figure 3 shows how to construct $V$ from individual controlled lab gates. @C=1.5em @R=1.2em &c_l 1 s 2V_s,r_l(x_l) r_l V_s,r_l(x_l) f_l V_s,r_l(x_l) Figure 3: A controlled lab gate for an individual laboratory @C=1.5em @R=1.2em &—0⟩_c W_l 2 W_l^† c_l -1 -1 s 2V r V f V Figure 4: An equivalent circuit to the individual controlled lab gate above, built from a single instance of $V$, where $W_{l}\ket{0}=\ket{l}$. Note that the individual wires may represent composite subsystems rather than individual qubits. @C=1.2em @R=1.2em &c 1 2 3⋯ 3 3⋯ 3 3⋯ 2 1 —0⟩ 3 —0⟩ 2 —0⟩ 1 s 2V_s,r_1(x_1) 4V_s,r_2(x_2) 7V_s,r_n(x_n) r_1 V_s,r_1(x_1) f_1 V_s,r_1(x_1) r_2 V_s,r_n(x_n) f_2 V_s,r_n(x_n) ⋯ ⋯ r_n V_s,r_n(x_n) f_n V_s,r_n(x_n) Figure 5: V, built from individual controled lab gates. The first and last $CNOT$ gate are controlled from the state $\ket{1}_{c}$, the second and second from last are controlled from $\ket{2}_{c}$, and so on until the $n$’th and $n+1$’th $CNOT$, which are controlled from state $\ket{n}_{c}$. ## Appendix B Proof of the Main Result In this appendix section, we give the full proof of the main result, that the probabilities generated by a quantum protocol can be replicated by a classical causal model, and therefore cannot violate a causal inequality. We begin with recalling a few definitions from the main text, together with some convenient derived quantities. Note that we assume throughout that laboratory labels $l$ are non-zero. ###### Definition 6 A causal probabilistic model can be written as $\displaystyle p(\vec{a}|\vec{x})=\sum_{l_{1}\notin\mathcal{L}_{0}}\sum_{l_{2}\notin\mathcal{L}_{1}}...\sum_{l_{N}\notin\mathcal{L}_{N}-1}p_{1}(l_{1}|H_{0})p_{1}(a_{l_{1}}|H_{0},x_{l_{1}})p_{2}(l_{2}|H_{1})p_{2}(a_{l_{1}}|H_{1},x_{l_{1}})...p_{N}(l_{N}|H_{N-1})p_{N}(a_{l_{N}}|H_{N-1},x_{l_{N}})$ (12) where $p_{k}(l_{k}|H_{k-1})$ terms represent probabilities for party $l_{k}$ to act at stage $k$ of the causal order, and $p_{k}(a_{l_{k}}|H_{k-1},x_{l_{k}})$ terms represent probabilities for party $l_{k}$, who has acted at stage $k$ of the causal order to obtain measurement result $a_{l_{k}}$. Both of the above probabilities are conditional on a history, $H_{k-1}$, which contains all of the information about previous inputs, outputs and party order. In particular, the history $H_{k}=(h_{1},...,h_{k})$ is the ordered collection of triples $h_{k}=(l_{k},a_{l_{k}},x_{l_{k}})$. The summations are performed over all possible next parties, excluding parties who have already acted, which are stored in the unordered sets $\mathcal{L}_{k}=\\{l_{1},...,l_{k}\\}$. To emphasise the symmetry between the terms we include $H_{0}$ and $\mathcal{L}_{0}$, which are defined as empty sets, as no parties have acted at that point. ###### Definition 7 The state of the system with a History $H_{k-1}$, at a time given by $t$, with the control set to trigger the action of party $l_{k}$ is given by $\displaystyle\ket{\psi_{(l_{k},t,H_{k-1})}}=(\ket{l_{k}}\bra{l_{k}}_{c}\otimes\pi^{H_{k-1}}_{rf}\otimes I)U_{t}VU_{t-1}...VU_{1}\ket{0}.$ (13) The projector onto the result and flag spaces is given by $\pi^{H_{k-1}}_{rf}=\bigotimes_{i=1}^{N}\left(\pi^{H_{k-1}}_{r_{i}f_{i}}\right),$ where $\displaystyle\pi^{H_{k-1}}_{r_{i}f_{i}}=\begin{cases}\ket{a_{i}}\bra{a_{i}}_{r_{i}}\otimes\ket{1}\bra{1}_{f_{i}}&\text{ if }(i,a_{i},x_{i})\in H_{k-1},\\\ I_{r_{i}}\otimes\ket{0}\bra{0}_{f_{i}}&\text{ otherwise }.\end{cases}$ (14) We also define the same state evolved to the end of protocol to be $\displaystyle{\ket{\bar{\psi}_{(l_{k},t,H_{k-1})}}}=VU_{T}VU_{T-1}...U_{t+1}V\ket{\psi_{((l_{k},t,H_{k-1}))}}.$ (15) It will also be convenient for the proof to define $\ket{\psi_{(0,t,H_{k-1})}}$ and ${\ket{\bar{\psi}_{(0,t,H_{k-1})}}}$, which are the same as the above states, but with $l_{k}=0$ (i.e. the control in the ‘do nothing’ setting). ###### Definition 8 The state of the system with a History $H_{k}$, at a time given by $t$, in which party $l_{k}$ has just acted is given by $\displaystyle\ket{\phi_{(l_{k},t,H_{k})}}=(\ket{a_{l_{k}}}\bra{a_{l_{k}}}_{r_{k}}\otimes I)V\ket{\psi_{(l_{k},t,H_{k-1})}}.$ (16) We also define the same state evolved to the end of protocol to be $\displaystyle{\ket{\bar{\phi}_{(l_{k},t,H_{k})}}}=VU_{T}VU_{T-1}...U_{t+1}\ket{\phi_{((l_{k},t,H_{k}))}}.$ (17) It will also be convenient for the proof to define $\ket{\phi_{(0,t,H_{k})}}=V\ket{\psi_{(0,t,H_{k})}}$ and ${\ket{\bar{\phi}_{(0,t,H_{k})}}}=VU_{T}VU_{T-1}...U_{t+1}\ket{\phi_{((0,t,H_{k}))}}.$ ###### Definition 9 The probability for party $l_{k}$ to act next, given a history $H_{k-1}$ is given by: $\displaystyle p_{k}(l_{k}|H_{k-1})=\frac{\sum_{t_{k}=1}^{T}|\ket{\psi_{(l_{k},t_{k},H_{k-1})}}|^{2}}{\sum_{l_{k}^{\prime}\notin\mathcal{L}_{k-1}}\sum_{t_{k}^{\prime}=1}^{T}|\ket{\psi_{(l_{k}^{\prime},t_{k}^{\prime},H_{k-1})}}|^{2}}$ (18) ###### Definition 10 The probability for party $l_{k}$ to obtain the measurement result $a_{l_{k}}$, given a history $H_{k-1}$, and an input variable $x_{l_{k}}$ is given by: $\displaystyle p_{k}(a_{l_{k}}|H_{k-1},x_{l_{k}})=\frac{\sum_{t_{k}=1}^{T}|\ket{\phi_{(l_{k},t_{k},H_{k})}}|^{2}}{\sum_{a^{\prime}_{l_{k}}\in\mathcal{A}_{l_{k}}}\sum_{t_{k}^{\prime}=1}^{T}|\ket{\phi_{(l_{k},t_{k}^{\prime},H_{k}^{\prime})}}|^{2}}$ (19) where $H^{\prime}_{k}=(H_{k-1},(l_{k},a^{\prime}_{l_{k}},x_{l_{k}}))$ (i.e. $H_{k}$ with $a_{l_{k}}$ replaced by $a^{\prime}_{l_{k}}$). ###### Definition 11 The quantum protocol consists of preparing an initial state $\ket{0}$, then acting with an alternating sequence of unitaries $U_{t}$ that act on the system and the control, and unitaries $V$ that act on the system, results and flag spaces as specified by the control. The total unitary for the protocol is given by $\displaystyle\mathcal{U}=VU_{T}VU_{T-1}V...VU_{N}V...VU_{1}$ (20) where we note that for an $N$ party protocol, $T\geq N$. Finally, the results registers are measured in the computational basis, giving the outcome probability distribution $\displaystyle p^{\mathrm{quantum}}(\vec{a}|\vec{x})=|(\ket{\vec{a}}\bra{\vec{a}}_{r}\otimes I)\mathcal{U}\ket{0}|^{2}.$ (21) With these definitions in place, we first prove some useful orthogonality lemmas concerning the barred states. ###### Lemma 1 We have that $\displaystyle\Braket{\bar{\psi}_{l^{\prime},t^{\prime},H}}{\bar{\psi}_{l,t,H}}=0$ (22) unless $l=l^{\prime}$ and $t^{\prime}=t$. Proof: consider first that $t=t^{\prime}$ and $l\neq l^{\prime}$. Then we have that $\braket{\bar{\psi}_{l^{\prime},t,H}}{\bar{\psi}_{l,t,H}}=\braket{{\psi}_{l^{\prime},t,H}}{{\psi}_{l,t,H}}=0$, since $\ket{{\psi}_{l^{\prime},t,H}}$ and $\ket{{\psi}_{l,t,H}}$ are orthogonal on the control $\mathcal{H}_{c}$. Next, consider that $t<t^{\prime}$. Then $\braket{\bar{\psi}_{l^{\prime},t^{\prime},H}}{\bar{\psi}_{l,t,H}}=\bra{{\psi}_{l^{\prime},t^{\prime},H}}U_{t^{\prime}}V...U_{t+1}V\ket{{\psi}_{l,t,H}}=0$ since $V\ket{{\psi}_{l,t,H}}$ contains a raised $l$ flag that is not raised in $\bra{{\psi}_{l^{\prime},t^{\prime},H}}$, and there is no operator connecting the two which can lower this flag. The case with $t>t^{\prime}$ follows from the $t<t^{\prime}$ case by noting that $\Braket{\bar{\psi}_{l^{\prime},t^{\prime},H}}{\bar{\psi}_{l,t,H}}=\Braket{\bar{\psi}_{l,t,H}}{\bar{\psi}_{l^{\prime},t^{\prime},H}}^{*}$. ###### Lemma 2 We have that $\displaystyle\braket{\bar{\phi}_{l^{\prime},t^{\prime},H}}{\bar{\phi}_{l,t,H}}=0$ (23) unless $l=l^{\prime}$ and $t=t^{\prime}$. Proof: consider first that $t=t^{\prime}$ and $l\neq l^{\prime}$. Then we have that $\braket{\bar{\phi}_{l^{\prime},t,H}}{\bar{\phi}_{l,t,H}}=\braket{{\phi}_{l^{\prime},t,H}}{{\phi}_{l,t,H}}=0$, since $\ket{{\phi}_{l^{\prime},t,H}}$ and $\ket{{\phi}_{l,t,H}}$ are orthogonal on the control $\mathcal{H}_{c}$. Next, consider that $t<t^{\prime}$. Then $\braket{\bar{\phi}_{l^{\prime},t^{\prime},H}}{\bar{\phi}_{l,t,H}}=\bra{{\phi}_{l^{\prime},t^{\prime},H}}VU_{t^{\prime}}V...U_{t+1}\ket{{\phi}_{l,t,H}}=0$, since the leftmost $V$ either raises a flag not in the history $H$, or the control at this point is set to zero, either of which will give the desired orthogonality. The case with $t>t^{\prime}$ follows from the $t<t^{\prime}$ case by noting that $\Braket{\bar{\phi}_{l^{\prime},t^{\prime},H}}{\bar{\phi}_{l,t,H}}=\Braket{\bar{\phi}_{l,t,H}}{\bar{\phi}_{l^{\prime},t^{\prime},H}}^{*}$. We now move onto proving the main result. This will consist of four stages, the first concerns a cancellation within terms of the same causal order stage, which allows us to rewrite the causal model in a nice way. The second and third results concern the initial and final terms in the inductive proof. The former corresponds to the fact that ‘somebody has to measure first’ in the quantum protocol, and the latter that the final term in the causal model has sufficient expressive power to capture the quantum measurement probabilities in their entirety. Finally, the fourth result concerns cancellations between terms at subsequent stages of the causal order. This leads to our main result which ties all of this together for a full proof that $p(\vec{a}|\vec{x})=|\bra{0}\mathcal{U}\ket{0}|^{2}$ is causal. ###### Result 1 There is an equality between the numerator of the ‘who is next’ type probabilities $p_{k}(l_{k}|H_{k-1})$, and the denominator of the ‘results’ type probabilities $p_{k}(a_{l_{k}}|H_{k-1},x_{l_{k}})$, allowing us to write the product of these probabilities in a nice way as $\displaystyle p_{k}(l_{k}|H_{k-1})p_{k}(a_{l_{k}}|x_{l_{k}},H_{k-1})=\frac{\sum_{t_{k}=1}^{T}|\ket{\phi_{(l_{k},t_{k},H_{k})}}|^{2}}{\sum_{l^{\prime}_{k}\notin\mathcal{L}_{k-1}}\sum_{t^{\prime}_{k}=1}^{T}|\ket{\psi_{(l^{\prime}_{k},t_{k}^{\prime},H_{k-1})}}|^{2}}.$ (24) . Proof: Starting with the denominator of the ‘results’ probability $\displaystyle\sum_{a^{\prime}_{l_{k}}\in\mathcal{A}_{l_{k}}}\sum_{t_{k}=1}^{T}\left|\ket{\phi_{(l_{k},t_{k},H^{\prime}_{k})}}\right|^{2}$ $\displaystyle=\sum_{a^{\prime}_{l_{k}}\in\mathcal{A}_{l_{k}}}\sum_{t_{k}=1}^{T}\left|(\ket{a^{\prime}_{l_{k}}}\bra{a^{\prime}_{l_{k}}}_{r_{k}}\otimes\mathcal{I})V\ket{\psi_{(l_{k},t_{k},H_{k-1})}}\right|^{2}$ $\displaystyle=\sum_{t_{k}=1}^{T}\left|\sum_{a^{\prime}_{l_{k}}\in\mathcal{A}_{l_{k}}}(\ket{a^{\prime}_{l_{k}}}\bra{a^{\prime}_{l_{k}}}_{r_{k}}\otimes\mathcal{I})V\ket{\psi_{(l_{k},t_{k},H_{k-1})}}\right|^{2}.$ $\displaystyle=\sum_{t_{k}=1}^{T}|V\ket{\psi_{(l_{k},t_{k},H_{k-1})}}|^{2}$ (25) $\displaystyle=\sum_{t_{k}=1}^{T}|\ket{\psi_{(l_{k},t_{k},H_{k-1})}}|^{2},$ (26) we obtain the numerator of the ‘who is next’ probabilities. In the second line we have used orthogonality on the results register, in the third line we have used the fact that after a measurement by party $l_{k}$, some result in $\mathcal{A}_{l_{k}}$ must have been obtained, and in the final line we have used unitarity. Using this to cancel the numerator of (18) with the denominator of (19) we obtain the desired result. ###### Result 2 The denominator of the first term $p_{1}(l_{1}|H_{0})$ satisfies $\displaystyle\sum_{l_{1}\notin\mathcal{L}_{0}}\sum_{t_{1}=1}^{T}|\ket{\psi_{(l_{1},t_{1},H_{0})}}|^{2}=1.$ (27) Proof: by first using unitarity and then Lemma 1 we have $\displaystyle\sum_{l_{1}\notin\mathcal{L}_{0}}\sum_{t_{1}=1}^{T}|\ket{\psi_{(l_{1},t_{1},H_{0})}}|^{2}=\sum_{l_{1}\notin\mathcal{L}_{0}}\sum_{t_{1}=1}^{T}|\ket{\bar{\psi}_{(l_{1},t_{1},H_{0}})}|^{2}$ $\displaystyle=|\sum_{l_{1}\notin\mathcal{L}_{0}}\sum_{t_{1}=1}^{T}\ket{\bar{\psi}_{(l_{1},t_{1},H_{0}})}|^{2}.$ (28) To simplify this further, consider evolving the state $\ket{\psi_{0,t_{1}-1,H_{0}}}$ forward for a full time-step using $U_{t_{1}}V$. As the control is in state $0$, no measurement occurs during $V$, and the unitary $U_{1}$ creates a superposition in which the control takes any possible state. Symbolically, $\displaystyle U_{t_{1}}V\ket{\psi_{0,t_{1}-1,H_{0}}}=\sum_{l_{1}\notin\mathcal{L}_{0}}\ket{\psi_{(l_{1},t_{1},H_{0}})}+\ket{\psi_{(0,t_{1},H_{0}})}.$ (29) By applying $VU_{T}V\ldots U_{t_{1}+1}V$ to this equation, we can obtain a similar form for the barred states, $\displaystyle\ket{\bar{\psi}_{0,t_{1}-1,H_{0}}}=\sum_{l_{1}\notin\mathcal{L}_{0}}\ket{\bar{\psi}_{(l_{1},t_{1},H_{0}})}+\ket{\bar{\psi}_{(0,t_{1},H_{0}})}.$ (30) We can rearrange this equation and substitute for the sum over $l_{1}$ on the right-hand side of (B) to obtain $\displaystyle\sum_{l_{1}\notin\mathcal{L}_{0}}\sum_{t_{1}=1}^{T}|\ket{\psi_{(l_{1},t_{1},H_{0})}}|^{2}=|\sum_{t_{1}=1}^{T}\left(\ket{\bar{\psi}_{(0,t_{1}-1,H_{0}})}-\ket{\bar{\psi}_{(0,t_{1},H_{0})}}\right)|^{2}$ (31) By expanding the summation on the right-hand side we find that only the first and last terms remain, giving $\displaystyle\sum_{l_{1}\notin\mathcal{L}_{0}}\sum_{t_{1}=1}^{T}|\ket{\psi_{(l_{1},t_{1},H_{0})}}|^{2}=|\ket{\bar{\psi}_{(0,0,H_{0}})}-\ket{\bar{\psi}_{(0,T,H_{0}})}|^{2}$ (32) Note that it is impossible by the requirements of our protocol that no-one has measured by time $t=T$. Such a scenario would violate the assumption that there are exactly $N$ flags raised at the end of the protocol. Therefore, $\ket{\bar{\psi}_{(0,T,H_{0}})}=0$, and we find $\displaystyle\sum_{l_{1}\notin\mathcal{L}_{0}}\sum_{t_{1}=1}^{T}|\ket{\psi_{(l_{1},t_{1},H_{0})}}|^{2}=|\ket{\bar{\psi}_{(0,0,H_{0}})}|^{2}=|\mathcal{U}\ket{0}|^{2}=1$ (33) as desired. ###### Result 3 The outcome statistics in the numerator of the final term in the causal probabilistic model represent the quantum probabilities arising from the protocol. In other words, $\displaystyle\sum_{l_{N}\in\mathcal{L}_{N}}\sum_{t_{N}=1}^{T}|\ket{\phi_{(l_{N},t_{N},H_{N})}}|^{2}=|\left(\ket{\vec{a}}\bra{\vec{a}}\otimes I\right)\mathcal{U}\ket{0}|^{2}$ (34) Proof: Firstly, note that by unitarity and Lemma 2 we have that $\displaystyle\sum_{l_{N}\in\mathcal{L}_{N}}\sum_{t_{N}=1}^{T}|\ket{\phi_{(l_{N},t_{N},H_{N})}}|^{2}$ $\displaystyle=\sum_{l_{N}\in\mathcal{L}_{N}}\sum_{t_{N}=1}^{T}|\ket{\bar{\phi}_{(l_{N},t_{N},H_{N})}}|^{2}$ $\displaystyle=|\sum_{l_{N}\in\mathcal{L}_{N}}\sum_{t_{N}=1}^{T}\ket{\bar{\phi}_{(l_{N},t_{N},H_{N})}}|^{2}.$ (35) The history $H_{N}$ represents a case in which all parties have already measured. At time $t<T$, what are the possible ways that this history can be filled? Either, nobody has measured in the previous time-step, or the last party to be filled into the history (subject to the requirement every party must enter the history exactly once) has just measured. In any case, evolving the linear combination of these states forward a time-step must produce a state at $t+1$ which contains an empty control (i.e., no-one else is left to measure, so don’t trigger them!). This gives us the key relation $\displaystyle VU_{t+1}\left(\sum_{l_{N}\in\mathcal{L}_{N}}\ket{\phi_{(l_{N},t,H_{N}})}+\ket{\phi_{(0,t,H_{N})}}\right)=\ket{\phi_{(0,t+1,H_{N}})}$ (36) which holds for $t<T$. Applying $VU_{T}V\ldots U_{t+2}$ we obtain $\displaystyle\left(\sum_{l_{N}\in\mathcal{L}_{N}}\ket{\bar{\phi}_{(l_{N},t,H_{N}})}+\ket{\bar{\phi}_{(0,t,H_{N})}}\right)=\ket{\bar{\phi}_{(0,t+1,H_{N}})}$ (37) which we can then rearrange to get $\displaystyle\sum_{l_{N}\in\mathcal{L}_{N}}\ket{\bar{\phi}_{(l_{N},t,H_{N})}}=\ket{\bar{\phi}_{(0,t+1,H_{N})}}-\ket{\bar{\phi}_{(0,t,H_{N})}}.$ (38) By separating out the $t_{N}=T$ term in equation (B) and then substituting this in the remaining terms , we find that $\displaystyle\sum_{l_{N}\in\mathcal{L}_{N}}\sum_{t_{N}=1}^{T}|\ket{\phi_{(l_{N},t_{N},H_{N})}}|^{2}$ $\displaystyle=|\sum_{l_{N}\in\mathcal{L_{N}}}\ket{\bar{\phi}_{(l_{N},T,H_{N})}}+\sum_{t_{N}=1}^{T-1}\sum_{l_{N}\in\mathcal{L_{N}}}\ket{\bar{\phi}_{(l_{N},t_{N},H_{N})}}|^{2}$ $\displaystyle=|\sum_{l_{N}\in\mathcal{L_{N}}}\ket{\bar{\phi}_{(l_{N},T,H_{N})}}+\sum_{t_{N}=1}^{T-1}\left(\ket{\bar{\phi}_{(0,t_{N}+1,H_{N})}}-\ket{\bar{\phi}_{(0,t_{N},H_{N})}}\right)|^{2}$ $\displaystyle=|\sum_{l_{N}\in\mathcal{L_{N}}}\ket{\bar{\phi}_{(l_{N},T,H_{N})}}+\ket{\bar{\phi}_{(0,T,H_{N})}}-\ket{\bar{\phi}_{(0,1,H_{N})}}|^{2}.$ (39) Now we note that $\ket{\bar{\phi}_{(0,1,H_{N})}}=0$ since it would be impossible for all parties to have measured in one time-step, and for the control to be in the zero state. Then we note that $\sum_{l_{N}\in\mathcal{L_{N}}}\ket{\bar{\phi}_{(l_{N},T,H_{N})}}+\ket{\bar{\phi}_{(0,T,H_{N})}}=(\pi^{H_{N}}_{rf}\otimes I)\mathcal{U}\ket{0}$, which is to say that these are simply the possible states at the end of the protocol, containing the measurement results we want to calculate the probabilities for in the history. Therefore $\displaystyle\sum_{l_{N}\in\mathcal{L}_{N}}\sum_{t_{N}=1}^{T}|\ket{\phi_{(l_{N},t_{N},H_{N})}}|^{2}$ $\displaystyle=|(\pi^{H_{N}}_{rf}\otimes I)\mathcal{U}\ket{0}|^{2}$ $\displaystyle=|(\ket{\vec{a}}\bra{\vec{a}}\otimes I)\mathcal{U}\ket{0}|^{2}$ (40) as desired. ###### Result 4 This is a technical result which establishes an equality between the states after measurement at causal order stage $k$ and the states before measurement at the next stage of the causal order. Namely, for $1\leq k<N$ that: $\displaystyle\sum_{l_{k}\in\mathcal{L}_{k}}\sum_{t=1}^{T}|\ket{\phi_{(l_{k},t,H_{k})}}|^{2}=\sum_{t^{\prime}=1}^{T}\sum_{l^{\prime}_{k+1}\notin\mathcal{L}_{k}}|\ket{{\psi}_{(l^{\prime}_{k+1},t^{\prime},H_{k})}}|^{2}.$ (41) Proof: Firstly, by unitarity and Lemma 2 we have that $\displaystyle\sum_{l_{k}\in\mathcal{L}_{k}}\sum_{t=1}^{T}|\ket{\phi_{(l_{k},t,H_{k})}}|^{2}$ $\displaystyle=\sum_{l_{k}\in\mathcal{L}_{k}}\sum_{t=1}^{T}|\ket{\bar{\phi}_{(l_{k},t,H_{k})}}|^{2}$ $\displaystyle=|\sum_{l_{k}\in\mathcal{L}_{k}}\sum_{t=1}^{T}\ket{\bar{\phi}_{(l_{k},t,H_{k})}}|^{2}.$ (42) Consider time-evolving a state just after the $t^{\textrm{th}}$ measurement step, in which history $H_{k}$ has been obtained (either by the last party just having measured, or by all parties in $H_{k}$ having measured previously), by $U_{t+1}$. This links states of the form $\ket{\phi_{(l_{k},t,H_{k})}}$ and $\ket{\psi_{(l^{\prime}_{k},t+1,H_{k})}}$ via $\displaystyle U_{t+1}\left(\sum_{l_{k}\in\mathcal{L}_{k}}\ket{\phi_{(l_{k},t,H_{k})}}+\ket{\phi_{(0,t,H_{k})}}\right)=\sum_{l^{\prime}_{k+1}\notin\mathcal{L}_{k}}\ket{\psi_{(l^{\prime}_{k+1},t+1,H_{k})}}+\ket{\psi_{(0,t+1,H_{k})}}$ (43) for $t<T$. Applying $VU_{T}V\ldots U_{t+2}V$ we obtain a very similar result for barred states; $\displaystyle\sum_{l_{k}\in\mathcal{L}_{k}}\ket{\bar{\phi}_{(l_{k},t,H_{k})}}+\ket{\bar{\phi}_{(0,t,H_{k})}}=\sum_{l^{\prime}_{k+1}\notin\mathcal{L}_{k}}\ket{\bar{\psi}_{(l^{\prime}_{k+1},t+1,H_{k})}}+\ket{\bar{\psi}_{(0,t+1,H_{k})}}.$ (44) We also note that $\ket{\phi_{(0,t,H_{k})}}=V\ket{\psi_{(0,t,H_{k})}}$ and hence that $\ket{\bar{\phi}_{(0,t,H_{k})}}=\ket{\bar{\psi}_{(0,t,H_{k})}}$. Making this substitution and rearranging a little we get $\displaystyle\sum_{l_{k}\in\mathcal{L}_{k}}\ket{\bar{\phi}_{(l_{k},t,H_{k})}}=\sum_{l^{\prime}_{k+1}\notin\mathcal{L}_{k}}\ket{\bar{\psi}_{(l^{\prime}_{k+1},t+1,H_{k})}}+\ket{\bar{\psi}_{(0,t+1,H_{k})}}-\ket{\bar{\psi}_{(0,t,H_{k})}}.$ (45) By substituting (45) into (B) for $t<T$, we arrive at $\displaystyle\sum_{l_{k}\in\mathcal{L}_{k}}\sum_{t=1}^{T}\left|\ket{\phi_{(l_{k},t,H_{k})}}\right|^{2}$ $\displaystyle=\left|\sum_{l_{k}\in\mathcal{L}_{k}}\ket{\bar{\phi}_{(l_{k},T,H_{k})}}+\sum_{t=1}^{T-1}\left(\sum_{l^{\prime}_{k+1}\notin\mathcal{L}_{k}}\ket{\bar{\psi}_{(l^{\prime}_{k+1},t+1,H_{k})}}+\ket{\bar{\psi}_{(0,t+1,H_{k})}}-\ket{\bar{\psi}_{(0,t,H_{k})}}\right)\right|^{2}$ $\displaystyle=\left|\sum_{l_{k}\in\mathcal{L}_{k}}\ket{\bar{\phi}_{(l_{k},T,H_{k})}}+\sum_{t=1}^{T-1}\sum_{l^{\prime}_{k+1}\notin\mathcal{L}_{k}}\ket{\bar{\psi}_{(l^{\prime}_{k+1},t+1,H_{k})}}+\ket{\bar{\psi}_{(0,T,H_{k})}}-\ket{\bar{\psi}_{(0,1,H_{k})}}\right|^{2}$ (46) where for the sums over time in the last two terms only the states with maximal and minimal times remain. Given that $k<N$ and all parties must have measured by the end of the protocol, it must be the case that $\ket{\bar{\phi}_{(l_{k},T,H_{k})}}=0$ and $\ket{\bar{\psi}_{(0,T,H_{k})}}=0$. Also as $k\geq 1$ it must be the case that $\ket{\bar{\psi}_{(0,1,H_{k})}}=0$ and $\ket{\bar{\psi}_{(l^{\prime}_{k+1},1,H_{k})}}=0$, as these states are just before the first measurement and hence must have no history. Using these results in equation (46) and setting $t^{\prime}=t+1$, we obtain $\displaystyle\sum_{l_{k}\in\mathcal{L}_{k}}\sum_{t=1}^{T}|\ket{\phi_{(l_{k},t,H_{k})}}|^{2}=\left|\sum_{t^{\prime}=1}^{T}\sum_{l^{\prime}_{k+1}\notin\mathcal{L}_{k}}\ket{\bar{\psi}_{(l^{\prime}_{k+1},t^{\prime},H_{k})}}\right|^{2}.$ (47) Finally, using Lemma 1 and unitarity we arrive at $\displaystyle\sum_{l_{k}\in\mathcal{L}_{k}}\sum_{t=1}^{T}|\ket{\phi_{(l_{k},t,H_{k})}}|^{2}$ $\displaystyle=\sum_{t^{\prime}=1}^{T}\sum_{l^{\prime}_{k+1}\notin\mathcal{L}_{k}}\left|\ket{\bar{\psi}_{(l^{\prime}_{k+1},t^{\prime},H_{k})}}\right|^{2}$ $\displaystyle=\sum_{t^{\prime}=1}^{T}\sum_{l^{\prime}_{k+1}\notin\mathcal{L}_{k}}\left|\ket{{\psi}_{(l^{\prime}_{k+1},t^{\prime},H_{k})}}\right|^{2}$ (48) as required. ###### Result 5 We will now show that the results of the quantum protocol can be replicated by a causal process. In particular $\displaystyle p(\vec{a}|\vec{x})=|\left(\ket{\vec{a}}\bra{\vec{a}}\otimes I\right)\mathcal{U}\ket{0}|^{2}$ $\displaystyle=$ $\displaystyle\sum_{l_{1}\notin\mathcal{L}_{0}}\sum_{l_{2}\notin\mathcal{L}_{1}}$ $\displaystyle...\sum_{l_{N}\notin\mathcal{L}_{N}-1}p_{1}(l_{1}|H_{0})p_{1}(a_{l_{1}}|H_{0},x_{l_{1}})p_{2}(l_{2}|H_{1})p_{2}(a_{l_{1}}|H_{1},x_{l_{1}})...p_{N}(l_{N}|H_{N-1})p_{N}(a_{l_{N}}|H_{N-1},x_{l_{N}})$ (49) and as such, the outcome statistics $p(\vec{a}|\vec{x})$ cannot violate a causal inequality. Proof: Firstly, substituting definitions 9 and 10 into the causal model (6), and then using Result 1, we can re-write the probability distribution for the entire causal model as $\displaystyle p(\vec{a}|\vec{x})=\sum_{l_{1}\notin\mathcal{L}_{0}}\sum_{l_{2}\notin\mathcal{L}_{1}}...\sum_{l_{N}\notin\mathcal{L}_{N}-1}\frac{\sum_{t_{1}=1}^{T}|\ket{\phi_{(l_{1},t_{1},H_{1})}}|^{2}}{\sum_{l_{1}^{\prime}\notin\mathcal{L}_{0}}\sum_{t_{1}^{\prime}=1}^{T}|\ket{\psi_{(l_{1}^{\prime},t_{1}^{\prime},H_{0})}}|^{2}}$ $\displaystyle\frac{\sum_{t_{2}=1}^{T}|\ket{\phi_{(l_{2},t_{2},H_{2})}}|^{2}}{\sum_{l_{2}^{\prime}\notin\mathcal{L}_{1}}\sum_{t_{2}^{\prime}=1}^{T}|\ket{\psi_{(l_{2}^{\prime},t_{2}^{\prime},H_{1})}}|^{2}}...$ $\displaystyle...$ $\displaystyle\frac{\sum_{t_{N}=1}^{T}|\ket{\phi_{(l_{N},t_{N},H_{N})}}|^{2}}{\sum_{l_{N}^{\prime}\notin\mathcal{L}_{N-1}}\sum_{t_{N}^{\prime}=1}^{T}|\ket{\psi_{(l_{N}^{\prime},t_{N}^{\prime},H_{N-1})}}|^{2}}$ (50) Let us begin by performing a simple reshuffling of (B)’s numerators and denominators, by writing the denominator of the term associated to causal order stage $k$ as the denominator of the term associated to $k-1$. $\displaystyle p(\vec{a}|\vec{x})=\frac{1}{\sum_{l_{1}^{\prime}\notin\mathcal{L}_{0}}\sum_{t_{1}^{\prime}=1}^{T}|\ket{\psi_{(l_{1}^{\prime},t_{1}^{\prime},H_{0})}}|^{2}}\sum_{l_{1}\notin\mathcal{L}_{0}}\sum_{l_{2}\notin\mathcal{L}_{1}}...\sum_{l_{N}\notin\mathcal{L}_{N}-1}$ $\displaystyle\frac{\sum_{t_{1}=1}^{T}|\ket{\phi_{(l_{1},t_{1},H_{1})}}|^{2}}{\sum_{l_{2}^{\prime}\notin\mathcal{L}_{1}}\sum_{t_{2}^{\prime}=1}^{T}|\ket{\psi_{(l_{2}^{\prime},t_{2}^{\prime},H_{1})}}|^{2}}\frac{\sum_{t_{2}=1}^{T}|\ket{\phi_{(l_{2},t_{2},H_{2})}}|^{2}}{\sum_{l_{3}^{\prime}\notin\mathcal{L}_{2}}\sum_{t_{3}^{\prime}=1}^{T}|\ket{\psi_{(l_{3}^{\prime},t_{3}^{\prime},H_{2})}}|^{2}}...$ $\displaystyle...\frac{\sum_{t_{k-1}=1}^{T}|\ket{\phi_{(l_{k-1},t_{k-1},H_{k-1})}}|^{2}}{\sum_{l_{k}^{\prime}\notin\mathcal{L}_{k-1}}\sum_{t_{k}^{\prime}=1}^{T}|\ket{\psi_{(l_{k}^{\prime},t_{k}^{\prime},H_{k-1})}}|^{2}}...\sum_{t_{N}=1}^{T}|\ket{\phi_{(l_{N},t_{N},H_{N})}}|^{2}$ (51) by using Result 2 we have $\displaystyle p(\vec{a}|\vec{x})$ $\displaystyle=\sum_{l_{1}\notin\mathcal{L}_{0}}\sum_{l_{2}\notin\mathcal{L}_{1}}...\sum_{l_{N}\notin\mathcal{L}_{N}-1}\frac{\sum_{t_{1}=1}^{T}|\ket{\phi_{(l_{1},t_{1},H_{1})}}|^{2}}{\sum_{l_{2}^{\prime}\notin\mathcal{L}_{1}}\sum_{t_{2}^{\prime}=1}^{T}|\ket{\psi_{(l_{2}^{\prime},t_{2}^{\prime},H_{1})}}|^{2}}\frac{\sum_{t_{2}=1}^{T}|\ket{\phi_{(l_{2},t_{2},H_{2})}}|^{2}}{\sum_{l_{3}^{\prime}\notin\mathcal{L}_{2}}\sum_{t_{3}^{\prime}=1}^{T}|\ket{\psi_{(l_{3}^{\prime},t_{3}^{\prime},H_{2})}}|^{2}}...\sum_{t_{N}=1}^{T}|\ket{\phi_{(l_{N},t_{N},H_{N})}}|^{2}$ (52) now note that we can rewrite the leftmost sum as $\displaystyle\sum_{l_{1}\notin\mathcal{L}_{0}}=\sum_{\mathcal{L}_{1}}\sum_{l_{1}\in\mathcal{L}_{1}},$ (53) where the sum over $\mathcal{L}_{1}$ is over all singleton sets $\\{l_{1}\\}$ (and hence has $N$ terms), and the subsequent sum over ${l_{1}\in\mathcal{L}_{1}}$ contains just a single term. We can use this to rewrite the probability distribution as $\displaystyle p(\vec{a}|\vec{x})$ $\displaystyle=\sum_{\mathcal{L}_{1}}\frac{\sum_{l_{1}\in\mathcal{L}_{1}}\sum_{t_{1}=1}^{T}|\ket{\phi_{(l_{1},t_{1},H_{1})}}|^{2}}{\sum_{l_{2}^{\prime}\notin\mathcal{L}_{1}}\sum_{t_{2}^{\prime}=1}^{T}|\ket{\psi_{(l_{2}^{\prime},t_{2}^{\prime},H_{1})}}|^{2}}\sum_{l_{2}\notin\mathcal{L}_{1}}...\sum_{l_{N}\notin\mathcal{L}_{N}-1}\frac{\sum_{t_{2}=1}^{T}|\ket{\phi_{(l_{2},t_{2},H_{2})}}|^{2}}{\sum_{l_{3}^{\prime}\notin\mathcal{L}_{2}}\sum_{t_{3}^{\prime}=1}^{T}|\ket{\psi_{(l_{3}^{\prime},t_{3}^{\prime},H_{2})}}|^{2}}...\sum_{t_{N}=1}^{T}|\ket{\phi_{(l_{N},t_{N},H_{N})}}|^{2}$ (54) where we have used the fact that the first numerator and denominator do not depend on $\\{l_{2},\ldots l_{N}\\}$. By application of Result 4 this is just $\displaystyle p(\vec{a}|\vec{x})$ $\displaystyle=\sum_{\mathcal{L}_{1}}\sum_{l_{2}\notin\mathcal{L}_{1}}...\sum_{l_{N}\notin\mathcal{L}_{N}-1}\frac{\sum_{t_{2}=1}^{T}|\ket{\phi_{(l_{2},t_{2},H_{2})}}|^{2}}{\sum_{l_{3}^{\prime}\notin\mathcal{L}_{2}}\sum_{t_{3}^{\prime}=1}^{T}|\ket{\psi_{(l_{3}^{\prime},t_{3}^{\prime},H_{2})}}|^{2}}...\sum_{t_{N}=1}^{T}|\ket{\phi_{(l_{N},t_{N},H_{N})}}|^{2}$ (55) we can iterate the same process again using $\displaystyle\sum_{\mathcal{L}_{1}}\sum_{l_{2}\notin\mathcal{L}_{1}}=\sum_{\mathcal{L}_{2}}\sum_{l_{2}\in\mathcal{L}_{2}}.$ (56) The left-hand side corresponds to first picking $l_{1}$ (with $N$ possibilities) and then picking a different $l_{2}$ ($N-1$ possibilities), whereas the right-hand side corresponds to first picking a pair of distinct labs $\mathcal{L}_{2}$ (with $N(N-1)/2$ possibilities) and then picking which of them was last ($2$ possibilities). This gives $\displaystyle p(\vec{a}|\vec{x})$ $\displaystyle=\sum_{\mathcal{L}_{2}}\sum_{l_{2}\in\mathcal{L}_{2}}...\sum_{l_{N}\notin\mathcal{L}_{N}-1}\frac{\sum_{t_{2}=1}^{T}|\ket{\phi_{(l_{2},t_{2},H_{2})}}|^{2}}{\sum_{l_{3}^{\prime}\notin\mathcal{L}_{2}}\sum_{t_{3}^{\prime}=1}^{T}|\ket{\psi_{(l_{3}^{\prime},t_{3}^{\prime},H_{2})}}|^{2}}...\sum_{t_{N}=1}^{T}|\ket{\phi_{(l_{N},t_{N},H_{N})}}|^{2}$ (57) which is just $\displaystyle p(\vec{a}|\vec{x})$ $\displaystyle=\sum_{\mathcal{L}_{2}}\frac{\sum_{l_{2}\in\mathcal{L}_{2}}\sum_{t_{2}=1}^{T}|\ket{\phi_{(l_{2},t_{2},H_{2})}}|^{2}}{\sum_{l_{3}^{\prime}\notin\mathcal{L}_{2}}\sum_{t_{3}^{\prime}=1}^{T}|\ket{\psi_{(l_{3}^{\prime},t_{3}^{\prime},H_{2})}}|^{2}}\sum_{l_{3}\notin\mathcal{L}_{2}}...\sum_{l_{N}\notin\mathcal{L}_{N}-1}\frac{\sum_{t_{3}=1}^{T}|\ket{\phi_{l_{3},t_{3},H_{3}}}|^{2}}{\sum_{l_{4}^{\prime}\notin\mathcal{L}_{3}}\sum_{t_{4}^{\prime}=1}^{T}|\ket{\psi_{(l_{4}^{\prime},t_{4}^{\prime},H_{3})}}|^{2}}...\sum_{t_{N}=1}^{T}|\ket{\phi_{(l_{N},t_{N},H_{N})}}|^{2}$ (58) application of Result 4 leads to another cancellation, so that we may write now $\displaystyle p(\vec{a}|\vec{x})$ $\displaystyle=\sum_{\mathcal{L}_{2}}\sum_{l_{3}\notin\mathcal{L}_{2}}...\sum_{l_{N}\notin\mathcal{L}_{N}-1}\frac{\sum_{t_{3}=1}^{T}|\ket{\phi_{l_{3},t_{3},H_{3}}}|^{2}}{\sum_{l_{4}^{\prime}\notin\mathcal{L}_{3}}\sum_{t_{4}^{\prime}=1}^{T}|\ket{\psi_{(l_{4}^{\prime},t_{4}^{\prime},H_{3})}}|^{2}}...\sum_{t_{N}=1}^{T}|\ket{\phi_{(l_{N},t_{N},H_{N})}}|^{2}$ (59) We can then iterate this process by applying the general result that $\displaystyle\sum_{\mathcal{L}_{k}}\sum_{l_{k+1}\notin\mathcal{L}_{k}}=\sum_{\mathcal{L}_{k+1}}\;\sum_{l_{k+1}\in\mathcal{L}_{k+1}},$ (60) and cancelling one of the numerators and denominators using Result 4 until we are left with the final term, $\displaystyle p(\vec{a}|\vec{x})$ $\displaystyle=\sum_{\mathcal{L}_{N}}\sum_{l_{n}\in\mathcal{L}_{N}}\sum_{t_{N}=1}^{T}|\ket{\phi_{(l_{N},t_{N},H_{N})}}|^{2}$ (61) the summation $\sum_{\mathcal{L}_{N}}=1$, as the only term corresponds to $\mathcal{L}_{N}=\\{1,2,\ldots N\\}$. Application of Result 3 then shows that this causal model indeed reproduces the quantum probabilities.i.e. that $\displaystyle p(\vec{a}|\vec{x})=|\left(\ket{\vec{a}}\bra{\vec{a}}\otimes I\right)\mathcal{U}\ket{0}|^{2}$ (62) as desired. ## Appendix C Example We now give an example of how our results apply in practice, based on the quantum switch Chiribella2013 . This involves using a quantum control to determine the order in which two operations are applied to another quantum system. The switch can be modelled in number of different ways (e.g. as a process matrix that exhibits causal non-separability Branciard2016 ) but here the basic idea is to prepare a superposition state of the form $\frac{1}{\sqrt{2}}\left(\ket{1}_{c}\otimes U_{A}U_{B}\ket{0}_{sfr}+\ket{2}_{c}\otimes U_{B}U_{A}\ket{0}_{sfr}\right)$ (63) where $U_{A}$ and $U_{B}$ are unitary transformations by Alice and Bob (representing their measurements). If a third party, Charlie, measures the control in a basis consisting of superpositions of $\ket{1}$ and $\ket{2}$ this will introduce interference between the two causal orders in which either Alice or Bob goes first. It has already been shown that this simple setup cannot be used to violate a causal inequality Arajo2015 ; Oreshkov2012 . However, it is instructive to see how it fits into our framework. @C=1.5em @R=1.2em &—0⟩_c U_1 1 U_2 1 2U_3 1 —0⟩_s_1 3V 3V U_3 3V —0⟩_s_2 V V U_3 V —0⟩_r V V V —0⟩_f V V V Figure 6: Realisation of the quantum switch in our framework through a quantum circuit. Because parties cannot directly measure the control in our framework, we transfer the state of the control into the system before Charlie’s measurement, and split the system into two qubits to facilitate this. Overall, the circuit we consider is shown in figure 6, where $\displaystyle U_{1}\ket{0}_{c}$ $\displaystyle=\frac{1}{\sqrt{2}}\left(\ket{1}_{c}+\ket{2}_{c}\right),$ $\displaystyle U_{2}\ket{1}_{c}$ $\displaystyle=\ket{2}_{c}$ $\displaystyle U_{2}\ket{2}_{c}$ $\displaystyle=\ket{1}_{c}$ $\displaystyle U_{3}\ket{1}_{c}\ket{\psi}_{s_{1}}\ket{0}_{s_{2}}$ $\displaystyle=\ket{3}_{c}\ket{0}_{s_{1}}\ket{\psi}_{s_{2}},$ $\displaystyle U_{3}\ket{2}_{c}\ket{\psi}_{s_{1}}\ket{0}_{s_{2}}$ $\displaystyle=\ket{3}_{c}\ket{1}_{s_{1}}\ket{\psi}_{s_{2}},$ (64) and $V$ is as given in equation 4 of the main body text. By considering the outcome statistics generated by this switch setup, we find that they differ from those which would be obtained from an equal mixture of the causal orders $A\rightarrow B\rightarrow C$ and $B\rightarrow A\rightarrow C$, due to the presence of interference. We proceed with an explicit calculation for the setup in figure 6. We assume that all parties have two possible measurements, hence their input variables $x,y,z$ are bits. When their input bit is zero, they measure the first part of the system in the computational basis and output the result in $a,b,c$. When their input bit is one, they instead measure the first part of the system in the Fourier basis (composed of the states $\ket{\pm}=\frac{1}{\sqrt{2}}\left(\ket{0}\pm\ket{1}\right)$, and output zero if they obtain the state $\ket{+}$ and one if they obtain the state $\ket{-}$. All of these measurements are implemented via unitary operations between the system and result register. e.g. $\displaystyle V_{s_{1},r_{1}}(x_{1}=1)$ $\displaystyle=\ket{+}\bra{+}_{s_{1}}\otimes I_{r_{1}}+\ket{-}\bra{-}_{s_{1}}\otimes\left(\ket{0}\bra{1}_{r_{1}}+\ket{1}\bra{0}_{r_{1}}\right)$ (65) Consider the case where the input variables are $x=0,y=1,z=1$. After some calculation, we find the state at the end of the protocol to be $\displaystyle\ket{3}_{c}\bigg{(}$ $\displaystyle\ket{+}_{s_{1}}(\frac{1}{2\sqrt{2}}\ket{+}_{s_{2}}+\frac{1}{4}\ket{0}_{s_{2}})\ket{000}_{r}+\ket{-}_{s_{1}}(\frac{1}{4}\ket{0}_{s_{2}}-\frac{1}{2\sqrt{2}}\ket{+}_{s_{2}})\ket{001}_{r}$ $\displaystyle+$ $\displaystyle\ket{+}_{s_{1}}(\frac{1}{2\sqrt{2}}\ket{-}_{s_{2}}+\frac{1}{4}\ket{0}_{s_{2}})\ket{010}_{r}+\ket{-}_{s_{1}}(\frac{1}{4}\ket{0}_{s_{2}}-\frac{1}{2\sqrt{2}}\ket{-}_{s_{2}})\ket{011}_{r}$ $\displaystyle+$ $\displaystyle\frac{1}{4}\ket{+}_{s_{1}}\ket{1}_{s_{2}}\ket{100}_{r}+\frac{1}{4}\ket{-}_{s_{1}}\ket{1}_{s_{2}}\ket{101}_{r}-\frac{1}{4}\ket{+}_{s_{1}}\ket{1}_{s_{2}}\ket{110}_{r}-\frac{1}{4}\ket{-}_{s_{1}}\ket{1}_{s_{2}}\ket{111}_{r}\bigg{)}\ket{111}_{f},$ (66) where we adopt the convention that $\ket{000}_{r}=\ket{a=0,b=0,c=0}_{r}$, et cetera. The probabilities to observe different outcomes in this measurement setting can then be obtained from this state. For example, $p(000|011)=|\frac{1}{2\sqrt{2}}\ket{+}_{s_{1}}\ket{+}_{s_{2}}+\frac{1}{4}\ket{+}_{s_{1}}\ket{0}_{s_{2}}|^{2}=5/16$. Such probabilities notably involve interference between different causal orders. In particular, they differ from those which would be obtained in the naive classical case, in which we first flip a coin to determine which of the two causal orders we will place ourselves in, and then perform the measurements in this causal order. We now calculate this ‘naive causal’ probability $p^{\text{nc}}(000|011)$. One half of the time, when we are in the causal order $A\rightarrow B\rightarrow C$, Alice measures $0$ with certainty and Bob, and Charlie have each a $50:50$ chance to measure either $0$ or $1$. The other half of the time, we are in the causal order $B\rightarrow A\rightarrow C$ all parties have a $50:50$ chance to measure either $0$ or $1$ (since Bob’s measurement in the Fourier basis, which occurs first, now makes Alice completely uncertain of her outcome). Putting this all together we find $p^{\text{nc}}(000|011)=\frac{1}{2}\times 1\times\frac{1}{2}\times\frac{1}{2}+\frac{1}{2}\times\frac{1}{2}\times\frac{1}{2}\times\frac{1}{2}=3/16\neq p(000|011)$. Nevertheless, our results show that we can find some classical causal model which generates the same outcome distribution $p(abc|xyz)$ as the quantum case. Let’s do this explicitly, to see how our proof translates in practice. We find by direct substitution into the definitions 4, and 5 in the main body that: $\displaystyle p_{1}(l_{1}=\text{Alice}|H_{0})=1/2\quad$ $\displaystyle\qquad\qquad p_{1}(l_{1}=\text{Bob}|H_{0})=1/2$ $\displaystyle p_{2}(l_{2}=\text{Bob}|H_{1}=(1,0,0))=1\quad$ $\displaystyle\qquad\qquad p_{2}(l_{2}=\text{Alice}|H_{1}=(2,0,1))=1$ $\displaystyle p_{3}(l_{3}=\text{Charlie}|H_{2}=((1,0,0),(2,0,1))=1\quad$ $\displaystyle\qquad\qquad p_{3}(l_{3}=\text{Charlie}|H_{2}=((2,0,1),(1,0,0))=1$ $\displaystyle p_{1}(a=0|H_{0},x=0)=1\quad$ $\displaystyle\qquad\qquad p_{1}(b=0|H_{0},y=1)=1/2$ $\displaystyle p_{2}(a=0|H_{1}=(1,0,0),x=0)=1/2\quad$ $\displaystyle\qquad\qquad p_{2}(b=0|H_{1}=(2,0,1),y=1)=1/2$ which are all the same as the naive classical case. However, the results-type probability for Charlie differs from the naive case. In particular we find $\displaystyle p_{3}(c=0|H_{2}=((1,0,0),(2,0,1),z=1)=\frac{5}{6}\quad$ $\displaystyle\qquad\qquad p_{3}(c=0|H_{2}=((2,0,1),(1,0,0),z=1)=\frac{5}{6}.$ (68) Despite the ordering of the history for the classical protocol being different in these two cases, the quantum calculation given by definition 5 is the same for both (as it only depends on the flags raised and results obtained before Charlie measures). This leads to interference between the two causal orders of $A$ and $B$ in $\ket{\phi_{(l_{3}=\text{Charlie},3,H_{2})}}$. This alternative classical procedure which emulates the quantum experiment therefore gives $p^{\text{ac}}(000|011)=\frac{1}{2}\times\frac{1}{2}\times\frac{5}{6}+\frac{1}{2}\times\frac{1}{2}\times\frac{1}{2}\times\frac{5}{6}=5/16$ as desired. Although we have focused on only one probability here, the same method can be used to generate a full classical strategy which replicates the quantum experiment for all input and output choices.
# Uniquely orderable interval graphs Marta Fiori-Carones Instytut Matematyki, Uniwersytet Warszawski, Banacha 2, 02-097 Warszawa — Poland<EMAIL_ADDRESS>and Alberto Marcone Dipartimento di scienze matematiche, informatiche e fisiche, Università di Udine, Via delle Scienze 208, 33100 Udine — Italy<EMAIL_ADDRESS> ###### Abstract. Interval graphs and interval orders are deeply linked. In fact, edges of an interval graphs represent the incomparability relation of an interval order, and in general, of different interval orders. The question about the conditions under which a given interval graph is associated to a unique interval order (up to duality) arises naturally. Fishburn provided a characterisation for uniquely orderable finite connected interval graphs. We show, by an entirely new proof, that the same characterisation holds also for infinite connected interval graphs. Using tools from reverse mathematics, we explain why the characterisation cannot be lifted from the finite to the infinite by compactness, as it often happens. ###### Key words and phrases: Interval graphs, infinite graphs, unique orderability, reverse mathematics ###### 2020 Mathematics Subject Classification: Primary 05C63; Secondary 05C75, 03B30, 05C62 Both authors were partially supported by the Italian PRIN 2017 Grant “Mathematical Logic: models, sets, computability”. ## 1\. Introduction An _interval graph_ is a graph whose vertices can be mapped (by an _interval representation_) to nonempty intervals of a linear order in such a way that two vertices are adjacent if and only if the intervals associated to them intersect (it is thus convenient to assume that the adjacency relation is reflexive). Consequently, if two vertices are incomparable in the graph, the corresponding intervals are placed one before the other in the linear order. The definition of interval graphs leads to an analogous concept for partial orders. In fact, a partial order $<_{P}$ is an _interval order_ if its points can be mapped to nonempty intervals of a linear order in such a way that $x<_{P}y$ if and only if the interval associated to $x$ completely precedes the interval associated to $y$. Thus interval graphs are the incomparability graphs of interval orders, i.e. two vertices are adjacent in the graph if and only if they are incomparable in the partial order. Norbert Wiener was probably the first to pay attention to interval orders, disguised under the less familiar name ‘relations of complete sequence’, in [Wie14]. Interval graphs and interval orders were rediscovered and given the current name in [Fis70]. There is now an extensive literature on the topic: [Tro97] provides a survey for many result in this area, focusing primarily on finite structures. Interval graphs and interval orders are extensively employed in diverse fields like psychology, archaeology and physics, just to mention a few. Wiener himself noticed that interval orders are useful for the analysis of temporal events and in the representation of measures subject to a margin of error. Interval orders actually occur in many digital calendars, where hours and days form a linear order and a rectangle covers the time assigned to an appointment: if two rectangles intersect, we better choose which event we will miss. Intervals are also suitable for representations of measurements of physical properties which are subject to error, since they can take into account the accuracy of the measuring device much better than a representation with points. In psychology and economics the overlap between two intervals often indicates that the corresponding stimuli or preferences are indistinguishable. In the first paragraph we described how to build an interval order from an interval representation of an interval graph. In general, an interval graph leads to many different interval orders on its vertices: an extreme example is a totally disconnected graph which is associated to any total order on its vertices. This paper deals with the situation were the interval graph is _uniquely orderable_ , i.e. there is essentially only one interval order associated to the given interval graph. (The “essentially” in the previous sentence is due to the obvious observation that if an interval order is associated to a graph, then the same is true for the reverse partial order.) Here the extreme example is a complete graph, which is associated to a unique partial order, the antichain of its vertices. The question of which interval graphs are uniquely orderable is easily settled for non connected graphs. It is in fact immediate that a non connected interval graph is uniquely orderable if and only if it has at most two components each of which is complete. We can thus restrict our attention to connected interval graphs. In this context, Fishburn [Fis85, §3.6], building on results proved in [Han82], provides two characterizations of unique orderability for finite graphs. Indeed, some steps of the proof heavily rely on the finiteness of the graph. This is in contrast with the rest of Fishburn’s monograph, where results are systematically proved for arbitrary interval graphs and orders; we thus believe that Fishburn did not know whether his result held for infinite interval graphs as well. The main result of this paper solves this issue by extending Fishburn’s characterizations to arbitrary interval graphs by an entirely different proof (for undefined notions see §2 below): ###### Theorem 1. Let $(V,E)$ be a (possibly infinite) connected interval graph. Let $W=\\{(a,b)\in V\times V\mid\neg a\,E\,b\\}$ and $(a,b)\,Q\,(c,d)\iff a\,E\,c\land b\,E\,d$. The following are equivalent: 1. (1) $(V,E)$ is uniquely orderable; 2. (2) $(V,E)$ does not contain a buried subgraph; 3. (3) the graph $(W,Q)$ has two components. Fishburn’s statement is slightly different from ours, since it is formulated for connected interval graphs without universal vertices. Since universal vertices (i.e. those adjacent to all vertices of the graph) are incomparable to all other vertices in any partial order associated to an interval graph, removing all universal vertices does not change the unique orderability of the graph. We prefer our formulation of the result since it highlights the connectedness of the graph, which is the central property characterising the class of interval graphs for which 1 holds. A typical method to lift a result from finite structures to arbitrary ones is compactness. Hence, once 1 is proved for finite interval graphs, the first attempt to generalise it to the infinite is to argue by compactness. This is not obvious and, using tools from mathematical logic, we are able to show that it is in fact impossible. To this end we work in the framework of reverse mathematics, a research program whose goal is to establish the minimal axioms needed to prove a theorem. In this framework compactness is embodied by the formal system $\mathsf{WKL}_{0}$. We first indicate, with results which parallel those obtained in [Mar07] about interval orders, that all the basic aspects of the theory of interval orders can be developed in $\mathsf{WKL}_{0}$. On the other hand we prove the following: ###### Theorem 2. Over the base system $\mathsf{RCA}_{0}$, the following are equivalent: 1. (1) $\mathsf{ACA}_{0}$, 2. (2) a countable connected interval graph $(V,E)$ is uniquely orderable if and only if does not contain a buried subgraph. Since $\mathsf{ACA}_{0}$ is properly stronger than $\mathsf{WKL}_{0}$ this shows that compactness does not suffice to prove 1. Section 2 establishes notation and terminology, while Section 3 is devoted to the proof of 1. Section 4 gives an overview of the reverse mathematics of interval graphs: the first author’s PhD thesis [FC19] includes full proofs. The last section is devoted to the proof of 2. ## 2\. Preliminaries In this section we establish the terminology used in the paper and underline some properties of interval graphs that turn out to be useful in the next section. All the graphs $(V,E)$ in this paper are such that $E\subseteq V\times V$ is a symmetric relation (we do not ask $E$ to be irreflexive, as in some cases it is convenient to have reflexivity). As usual, we write $v\,E\,u$ to mean $(v,u)\in E$ and, if $V^{\prime}\subseteq V$, we write $(V^{\prime},E)$ in place of $(V^{\prime},E\cap(V^{\prime}\times V^{\prime}))$. We denote by $(V,\overline{E})$ the _complementary graph_ of $(V,E)$: for $u,v\in V$ we have $u\,\overline{E}\,v$ if and only if $u\,E\,v$ does not hold. Paths and cycles are defined as usual, and their length is the number of their edges. A _simple cycle_ $v_{0}\,E\,\dots\,E\,v_{n}$ is a cycle such that the vertices in $v_{0},\dots,v_{n-1}$ are distinct. A _chord_ of a cycle $v_{0}\,E\,v_{1}\,E\,\dots\,E\,v_{n}$ is an edge $(v_{i},v_{j})$ with $2\leq j-i\leq n-2$. The chord is _triangular_ if either $j-i=2$ or $j-i=n-2$. ###### Definition 2.1. If $(V,\prec)$ is a strict partial order, the _comparability graph of $(V,\prec)$_ is the graph $(V,E)$ such that for $v,u\in V$ it holds that $v\,E\,u$ if and only if either $v\prec u$ or $u\prec v$. The _incomparability graph of $(V,\prec)$_ is the complementary graph of the comparability graph, so that two vertices are adjacent if and only if they coincide or are $\prec$-incomparable. While the comparability graph of a strict partial order is irreflexive, its incomparability graph is reflexive. Notice that a graph $(V,E)$ can be the incomparability graph of more than one partial order: we say that each such partial order is _associated to $(V,E)$_. In particular, $\prec$ and the dual of $\prec$ (i.e. $\prec^{\prime}$ such that $u\prec^{\prime}v$ iff $v\prec u$) are associated to the same incomparability graph. ###### Definition 2.2. A graph $(V,E)$ is _uniquely orderable_ if it is the incomparability graph of a partial order $\prec$ and the only other partial order associated to $(V,E)$ is the dual order of $\prec$; in other words, there exists a unique (up to duality) partial order $\prec$ such that for each $v,u\in V$ it holds that $\neg u\,E\,v$ if and only if $u\prec v$ or $v\prec u$. The following definition formalises the intuitive idea of interval graph given in the previous pages. ###### Definition 2.3. A graph $(V,E)$ is an _interval graph_ if it is reflexive and there exist a linear order $(L,<_{L})$ and a map $F\colon V\to\wp(L)$ such that for all $v,u\in V$, $F(v)$ is an interval in $(L,<_{L})$ (i.e. if $\ell<_{L}\ell^{\prime}<_{L}\ell^{\prime\prime}$ and $\ell,\ell^{\prime\prime}\in F(v)$, then also $\ell^{\prime}\in F(v)$) and $v\,E\,u\Leftrightarrow F(v)\cap F(u)\neq\emptyset.$ It is well-known that we may in fact assume that there exist functions $f_{L},f_{R}\colon V\to L$ such that $F(v)=\\{\ell\in L\mid f_{L}(v)\leq_{L}\ell\leq_{L}f_{R}(v)\\}$ for all $v\in V$ (this is the definition given in [Fis85]). We say that $(L,<_{L},f_{L},f_{R})$ (but often only $(f_{L},f_{R})$ or just $F$) is a _representation_ of $(V,E)$. To decide whether two vertices $u$ and $v$ are adjacent in an interval graph with representation $(f_{L},f_{R})$ we can assume without loss of generality that $f_{L}(v)\leq_{L}f_{L}(u)$ and then simply check whether $f_{L}(u)\leq_{L}f_{R}(v)$. In the context of a representation $(f_{L},f_{R})$ of an interval graph, we write $F(v)<_{L}F(u)$ in place of $f_{R}(v)<_{L}f_{L}(u)$. Then $\neg v\,E\,u$ means that either $F(v)<_{L}F(u)$ or $F(u)<_{L}F(v)$. Figure 1 provides an example of interval graph, while the graph in Figure 2 does not have an interval representation (in the figures self loops are not shown for clarity). abcd Figure 1. An example of interval graph with its representation acbxyz? Figure 2. A graph which is not an interval graph, with a partial representation A classical characterization of interval graphs is the following ([LB62], see [Fis85, Theorem 3.6]). ###### Definition 2.4. A graph $(V,E)$ is _triangulated_ if every simple cycle of length four or more has a chord. An _asteroidal triple_ in $(V,E)$ is an independent set of three vertices (i.e. a set of pairwise non adjacent vertices) of $V$ such that any two of them are connected by a path that avoids the vertices adjacent to the third. ###### Theorem 2.5. A reflexive graph $(V,E)$ is an interval graph if and only it is triangulated and has no asteroidal triples. ###### Proposition 2.6. Let $v_{0}\,E\,\dots\,E\,v_{n}$ be a path in the interval graph $(V,E)$ with representation $F$, and suppose $w\in V$ is such that $F(w)\nless_{L}F(v_{0})$ and $F(v_{n})\nless_{L}F(w)$. Then $v_{i}\,E\,w$ for some $i\leq n$, and hence $v_{0}\,E\,\dots\,E\,v_{i}\,E\,w$ and $w\,E\,v_{i}\,E\,\dots\,E\,v_{n}$ are paths. ###### Proof. Let $i\leq n$ be maximum such that $F(w)\nless_{L}F(v_{i})$. If $i=n$, then $F(w)\nless_{L}F(v_{n})$ and $F(v_{n})\nless_{L}F(w)$ imply $v_{n}\,E\,w$. If $i<n$, then $F(w)<_{L}F(v_{i+1})$ and $F(v_{i})\nless_{L}F(v_{i+1})$ (because $v_{i}\,E\,v_{i+1}$) imply $F(v_{i})\nless_{L}F(w)$. This, together with $F(w)\nless_{L}F(v_{i})$, yields $v_{i}\,E\,w$. ∎ ###### Definition 2.7. Let $(V,E)$ be a graph. A path $v_{0}\,E\,\dots\,E\,v_{n}$ is a _minimal path_ if $\neg v_{i}\,E\,v_{j}$ for every $i,j$ such that $i+1<j\leq n$. Notice that if $v_{0}\,E\,\dots\,E\,v_{n}$ is a path of minimal length among the paths connecting $v_{0}$ and $v_{n}$, then it is a minimal path, but the reverse implication does not hold. ###### Property 2.8. Let $(V,E)$ be a graph. Then each path can be refined to a minimal path. ###### Proof. The statement follows immediately from the following observation: if $v_{0}\,E\,\dots\,E\,v_{n}$ is a path and $v_{i}\,E\,v_{j}$ with $i+1<j\leq n$, then $v_{0}\,E\,\dots\,E\,v_{i}\,E\,v_{j}\,E\,\dots\,E\,v_{n}$ is still a path. ∎ ###### Property 2.9. Let $(V,E)$ be an interval graph with representation $(L,<_{L},f_{L},f_{R})$ and suppose that $v_{0}\,E\,\dots\,E\,v_{n}$ is a minimal path with $F(v_{0})<_{L}F(v_{n})$. 1. (i) Then $f_{R}(v_{i})<_{L}f_{R}(v_{i+1})$ for each $i<n-1$ and $f_{L}(v_{j})<_{L}f_{L}(v_{j+1})$ for each $j>0$; 2. (ii) if $F(v)<_{L}F(v_{0})$, then $\neg v_{i}\,E\,v$ for every $i\neq 1$; symmetrically, if $F(v_{n})<_{L}F(v)$, then $\neg v_{i}\,E\,v$ for every $i\neq n-1$. ###### Proof. To check the first conjunct of (i), suppose $i<n-1$ is least such that $f_{R}(v_{i+1})\leq_{L}f_{R}(v_{i})$. Since $i<n-1$ it holds that $\neg v_{j}\,E\,v_{n}$ for each $j\leq i$ by definition of minimal path. An easy induction, starting with our assumption $F(v_{0})<_{L}F(v_{n})$, shows that $F(v_{k})<_{L}F(v_{n})$ for each $k\leq i$. Thus, in particular it holds that $f_{R}(v_{i})<_{L}f_{R}(v_{n})$. Let $m\leq n$ be least such that $f_{R}(v_{i})<_{L}f_{R}(v_{m})$ and notice that $m>i+1$ by choice of $i$. By choice of $m$ it holds that $f_{R}(v_{m-1})\leq_{L}f_{R}(v_{i})<_{L}f_{R}(v_{m})$, and so that $f_{L}(v_{m})\leq_{L}f_{R}(v_{m-1})$ because $v_{m-1}\,E\,v_{m}$. To summarise we get that $f_{L}(v_{m})\leq_{L}f_{R}(v_{i})<_{L}f_{R}(v_{m})$, namely that $v_{i}\,E\,v_{m}$ contrary to the definition of minimal path. The second conjunct of (i) follows from the first considering the interval representation given by the linear order $(L,>_{L})$ and by the maps $f_{L}$ and $f_{R}$. For (ii), let $v_{0}\,E\,\dots\,E\,v_{n}$ be a minimal path and $F(v)<_{L}F(v_{0})<_{L}F(v_{n})$. Assume $v\,E\,v_{i}$, for some $i>1$ (notice that $\neg v\,E\,v_{0}$ by assumption). Since $f_{R}(v)<_{L}f_{L}(v_{0})$ by assumption, $f_{R}(v_{0})<_{L}f_{R}(v_{i})$ by (i), and $f_{L}(v_{i})<_{L}f_{R}(v)$ by $v\,E\,v_{i}$, it holds that $v_{0}\,E\,v_{i}$, contrary to the definition of minimal path. ∎ ## 3\. Uniquely orderable connected interval graphs In this section we prove 1. Suppose $(V,E)$ is a connected incomparability graph. Saying that $(V,E)$ is not uniquely orderable amounts to check that there are two partial orders $\prec$ and $\prec^{\prime}$ associated to $(V,E)$ and three vertices $a,b,c\in V$ such that $a\prec b\prec c$ and $b\prec^{\prime}a\prec^{\prime}c$. The vertices $a$ and $b$ can be reordered regardless, so to speak, the order of $c$. The connected graph pictured (by one of its interval representations) in Figure 3 is an example of a non uniquely orderable connected interval graph (in fact the intervals for $a$ and $b$ can be swapped without changing their relationship with the intervals $c$ and $k$). $\scriptstyle{a}$$\scriptstyle{b}$$\scriptstyle{c}$$\scriptstyle{k}$ Figure 3. An interval representation of a non uniquely orderable connected interval graph The first characterization of uniquely orderable interval graphs exploits the above observation to identify subgraphs which are forbidden in uniquely orderable interval graphs. ###### Definition 3.1. Let $(V,E)$ be a graph. For $B\subseteq V$ let $K(B)=\\{v\in V\mid\forall b\in B\,(v\,E\,b)\\}$ and $R(B)=V\setminus(B\cup K(B))$. We say that $B$ is a _buried subgraph_ of $(V,E)$ if the following hold: 1. (i) there exist $a,b\in B$ such that $\neg a\,E\,b$, 2. (ii) $K(B)\cap B=\emptyset$ and $R(B)\neq\emptyset$, 3. (iii) if $b\in B$ and $r\in R(B)$, then $\neg b\,E\,r$. The last point in the previous definition implies that any path between a vertex in $B$ and a vertex outside $B$ must go through a vertex in $K(B)$. The main consequence of (iii), which we use many times without mention, is that if $v\in V$ is such that there exist $a,b\in B$ such that $v\,E\,a$ and $\neg v\,E\,b$, then $v\in B$ (because $\neg v\,E\,b$ implies $v\notin K(B)$, while $v\,E\,a$ and (iii) imply $v\notin R(B)$). Our definition of buried subgraph is slightly different from the one in [Fis85], but it is equivalent for the class of graphs studied by Fishburn, i.e. connected interval graphs without universal vertices. Since we allow universal vertices, in condition (ii) we substituted $K(B)\neq\emptyset$ with $R(B)\neq\emptyset$ (the former condition implies the latter if there are no universal vertices, the reverse implication holds if the graph is connected by (iii)). Moreover we restated condition (iii) in simpler, yet equivalent, terms. The other main character of 1 is the graph $(W,Q)$. ###### Definition 3.2. If $(V,E)$ is a graph we let $W=\\{(a,b)\in V\times V\mid\neg a\,E\,b\\}$ and (writing $ab$ in place of $(a,b)$ for concision) $ab\,Q\,cd$ if and only if $a\,E\,c$ and $b\,E\,d$. If $ab$ and $cd$ are elements of $W$ which are connected by a path in $(W,Q)$ we write $ab\,\bar{Q}\,cd$. ###### Proposition 3.3. Let $(V,E)$ be an interval graph and $\prec$ a partial order associated to $(V,E)$. If $ab,cd\in W$ are such that $ab\,\bar{Q}\,cd$ and $a\prec b$, then $c\prec d$. In particular we have $\neg ab\,\bar{Q}\,ba$. ###### Proof. Suppose first that $ab\,Q\,cd$, so that $a\,E\,c$ and $b\,E\,d$. Notice that $b\,E\,c$ and $a\,E\,d$ cannot both hold. In fact, if $a,b,c,d$ are not all distinct, then this would contradict $ab\in W$ or $cd\in W$. Otherwise, $a\,E\,c\,E\,b\,E\,d\,E\,a$ would be a simple cycle of length four without chords, against Theorem 2.5. If $\neg b\,E\,c$, then $c\prec b$ because $a\prec b$ and $a\,E\,c$. From this we obtain $c\prec d$, since $d\,E\,b$. If instead $\neg a\,E\,d$ we obtain first $a\prec d$ and then again $c\prec d$. To derive $c\prec d$ from $ab\,\bar{Q}\,cd$ it suffices to apply the transitivity of $\prec$ to a $Q$-path connecting $ab$ with $cd$. ∎ The last part of the previous proposition implies that if $W\neq\emptyset$ (which is equivalent to $(V,E)$ being not complete), then $(W,Q)$ has at least two components. Moreover, if $(W,Q)$ has more than two (and so at least four) components, then for every partial order $\prec$ associated to $(V,E)$ there exist $ab,cd\in W$ such that $a\prec b$ and $c\prec d$, yet $ab\,\bar{Q}\,cd$ fails. We split, as originally done by Fishburn, the proof of 1 in three steps corresponding to (1) implies (2) (Lemma 3.4), (3) implies (1) (Lemma 3.5), and (2) implies (3) (Theorem 3.12). The proof of the first implication in [Fis85] is not completely accurate, and we apply Fishburn’s idea after a preliminary step which is necessary even when the graph is finite. The second implication is straightforward and applies to interval graphs of any cardinality. The proof of the last implication is completely new and requires more work. The connectedness of the graph is not needed in the first two implications. Moreover, the hypotheses of Lemma 3.4 could be further relaxed, as the proof applies to arbitrary incomparability graphs. ###### Lemma 3.4. Every uniquely orderable interval graph does not contain a buried subgraph. ###### Proof. Let $(V,E)$ be an interval graph with a buried subgraph $B$. Fix a partial order $\prec_{0}$ associated to $(V,E)$ and some $b_{0}\in B$. We define a new binary relation $\prec$ on $V$ as follows: when either $u,v\in B$ or $u,v\notin B$ set $u\prec v$ if and only if $u\prec_{0}v$; when $b\in B$ and $v\notin B$ set $b\prec v$ if and only if $b_{0}\prec_{0}v$, and $v\prec b$ if and only if $v\prec_{0}b_{0}$. Thus the whole $B$ is $\prec$-above the elements not in $B$ which are $\prec_{0}$-below $b_{0}$ and $\prec$-below the elements not in $B$ which are $\prec_{0}$-above $b_{0}$. Using the fact that the vertices not in $B$ are either $\prec_{0}$-incomparable to every vertex of $B$ or $\prec_{0}$-comparable to every vertex of $B$, it is straightforward to check that $\prec$ is transitive, and hence a partial order. For the same reason $\prec$ is associated to $(V,E)$. The key feature of $\prec$ (not necessarily shared by $\prec_{0}$) is that $B$ is $\prec$-convex, i.e. if $b\prec v\prec b^{\prime}$ with $b,b^{\prime}\in B$, then $v\in B$ as well. Indeed, if $v\notin B$, then $b\prec v$ implies $b_{0}\prec_{0}v$ and $v\prec b^{\prime}$ implies $v\prec_{0}b_{0}$. Following now [Fis85], let $\prec^{\prime}$ be such that the restrictions of $\prec$ and $\prec^{\prime}$ to $B$ are dual, while $\prec^{\prime}$ and $\prec$ coincide on $V\setminus B$ and between elements of $B$ and $V\setminus B$. Formally, $u\prec^{\prime}v$ if and only if either $u,v\in B$ and $v\prec u$, or if at least one of $u$ and $v$ does not belong to $B$ and $u\prec v$. The transitivity of $\prec^{\prime}$ is a consequence of the $\prec$-convexity of $B$ (an observation lacking in the proof given in [Fis85]) and hence $\prec^{\prime}$ is a partial order associated to $(V,E)$. If $x,y\in B$ are such that $x\prec y$ and $v\in R(B)$ (these elements exist by Definition 3.1) we have either $x\prec y\prec v$ or $v\prec x\prec y$. In the first case $y\prec^{\prime}x\prec^{\prime}v$, in the second case $v\prec^{\prime}y\prec^{\prime}x$, witnessing that $\prec^{\prime}$ is neither $\prec$ nor the dual order of $\prec$. ∎ ###### Lemma 3.5. Let $(V,E)$ be an interval graph. If ($W,Q$) has two components, then $(V,E)$ is uniquely orderable. ###### Proof. This follows easily from Proposition 3.3. ∎ For the proof of Theorem 3.12 we describe a construction that, starting from a pair of non-adjacent vertices, attempts to build the minimal buried subgraph containing those two vertices. We then show that if this attempt always fails, then for any $ab,cd\in W$ either $ab\,\bar{Q}\,cd$ or $ab\,\bar{Q}\,dc$. ###### Construction 3.6. Let $(V,E)$ be a connected interval graph and $v,u\in V$ be such that $\neg v\,E\,u$. We define recursively $B_{n}(v,u)\subseteq V$: $\displaystyle B_{0}(v,u)$ $\displaystyle=\\{v,u\\}$ $\displaystyle B_{n+1}(v,u)$ $\displaystyle=\\{w\in V\mid\exists z,z^{\prime}\in B_{n}(v,u)\,(z\,E\,w\land\neg z^{\prime}\,E\,w)\\}$ We then set $B(v,u)=\bigcup B_{n}(v,u)$. If $w\in B(v,u)$ let $e_{w}$ be the least $n$ such that $w\in B_{n}(v,u)$ (formally we should write $e_{w}^{v,u}$ but we omit the superscript as $v$ and $u$ will always be understood). A straightforward induction shows that $B_{n}(v,u)\subseteq B_{n+1}(v,u)$, for each $n\in\mathbb{N}$ (for the base step recall that interval graphs are reflexive, so that $v$ and $u$ themselves witness that $v,u\in B_{1}(v,u)$). We now show that $B(v,u)$ is close to being a buried subgraph. ###### Property 3.7. In the situation of 3.6, $B(v,u)$ is a buried subgraph if and only if $R(B(v,u))\neq\emptyset$. ###### Proof. Notice that Condition (i) of Definition 3.1 is witnessed by $v$ and $u$. Condition (3) is obvious, because if $r\notin K(B(u,v))$ but $b\,E\,r$ for some $b\in B(u,v)$, then $r\in B(u,v)$. Moreover, if $k\in K(B(v,u))$, then $k\in K(B_{n}(u,v))$ for every $n$ and hence $k\notin B(u,v)$; hence $B(v,u)\cap K(B(v,u))=\emptyset$. Therefore, to verify that $B(v,u)$ is a buried subgraph it suffices that $R(B(v,u))\neq\emptyset$. ∎ In the next propositions, we will always consider a connected interval graph $(V,E)$ with representation $(L,<_{L},f_{L},f_{R})$, fix $v,u\in V$ with $\neg v\,E\,u$ and $F(v)<_{L}F(u)$ and define $B(v,u)$ as in 3.6. For brevity, we call this set of hypotheses $(\maltese)$ and indicate it next to the proposition number. ###### Proposition 3.8 (✠). Let $x,y\in B(v,u)$. If $F(u)<_{L}F(x)$ and $f_{L}(x)\leq_{L}f_{L}(y)$, then $e_{x}\leq e_{y}$. Analogously, if $F(x)<_{L}F(v)$ and $f_{R}(y)\leq_{L}f_{R}(x)$, then $e_{x}\leq e_{y}$ as well. ###### Proof. We prove the first half of the statement by induction on $e_{y}$. The base case is trivially satisfied since there is no $y\in B_{0}(v,u)$ satisfying the hypotheses. Assume $e_{y}>0$ and let $z\in B_{e_{y}-1}(v,u)$ be such that $z\,E\,y$. If $f_{L}(x)\leq_{L}f_{L}(z)$, then $e_{x}\leq e_{z}<e_{y}$ by induction hypothesis. Otherwise, $f_{L}(z)<_{L}f_{L}(x)\leq f_{L}(y)\leq_{L}f_{R}(z)$ given that $z\,E\,y$. This means that $z\,E\,x$, which implies that $x\in B_{e_{z}+1}(v,u)$ since $u\in B_{e_{z}}(v,u)$ is such that $\neg u\,E\,x$. Hence $e_{x}\leq e_{z}+1\leq e_{y}$ The second half of the statement follows considering the representation $(L,>_{L},f_{L},f_{R})$. ∎ ###### Proposition 3.9 (✠). Let $w\in B(v,u)$. If $F(w)\nless_{L}F(u)$, then there exists a path $u\,E\,b_{1}\,E\,\dots b_{k}\,E\,w$ such that $e_{b_{i}}<e_{w}$ for all $i\leq k$. Analogously, if $F(v)\nless_{L}F(w)$, then there exists a path $v\,E\,b_{1}\,E\,\dots b_{k}\,E\,w$ such that $e_{b_{i}}<e_{w}$ for all $i\leq k$. ###### Proof. By definition of $B_{e_{w}}(v,u)$ there exists a path $b_{0}\,E\,b_{1}\,E\,\dots\,E\,b_{k}\,E\,w$ where $b_{i}\in B(v,u)$ and $0=e_{b_{0}}<e_{b_{1}}<\dots<e_{b_{k}}<e_{w}$. Hence $b_{0}\in\\{u,v\\}$ and, since $F(w)\nless_{L}F(u)$ and $F(u)\nless_{L}F(v)$, by Proposition 2.6 we can assume that $b_{0}=u$. The second half of the statement follows from the first one as usual. ∎ ###### Proposition 3.10 (✠). Let $x,z\in B(v,u)$ and $m=\max\\{e_{x},e_{z}\\}$. Assume $F(z)<_{L}F(x)$ and $F(v)\nless_{L}F(x)$ (this implies $m>0$). Then there exists a minimal path $z\,E\,v_{1}\,E\,\dots\,E\,v_{n}\,E\,x$ and $s\in B(v,u)$ with $e_{s}<m$ such that $e_{v_{i}}<m$ and $F(v_{i})<_{L}F(s)$ for each $i\leq n$. Analogously, if $F(x)<_{L}F(z)$ and $F(x)\nless_{L}F(u)$ there exists a minimal path $x\,E\,v_{1}\,E\,\dots\,E\,v_{n}\,E\,z$ and $s\in B(v,u)$ with $e_{s}<m$ such that $e_{v_{i}}<m$ and $F(s)<_{L}F(v_{i})$ for each $i\leq n$. ###### Proof. Notice that once we find the minimal path $z\,E\,v_{1}\,E\,\dots\,E\,v_{n}\,E\,x$ and $s\in B(v,u)$ with $e_{s}<m$ such that $e_{v_{i}}<m$ for all $i\leq n$ it suffices to prove that $F(v_{n})<_{L}F(s)$, since then $F(v_{i})<_{L}F(s)$ for $i<n$ follows from 2.9.i. We can apply Proposition 3.9 to both $x$ and $z$ obtaining paths connecting $v$ to $x$ and $v$ to $z$ and with $e_{b}<m$ for all vertices $b$, distinct from $x$ and $z$, occurring in the paths. Joining these paths and then using 2.8 we obtain a minimal path $z\,E\,v_{1}\,E\,\dots\,E\,v_{n}\,E\,x$ with $e_{v_{i}}<m$. Notice that $n>0$ as $\neg z\,E\,x$. Let $j<m$ be such that $e_{v_{n}}=j$: we may assume $j$ is least for which such a minimal path exists. If $j=0$, then we claim that we can assume $v_{n}=v$ and hence we can choose $s=u$. In fact, if $v_{n}=u$, then $F(x)<_{L}F(v)$ is impossible and we have $v\,E\,x$. The hypotheses imply $F(v)\nless_{L}F(z)$ and, since $F(u)\nless_{L}F(v)$, by Proposition 2.6 we can find $i<n$ such that $v\,E\,v_{i}$ and consider the path $z\,E\,v_{1}\,E\,\dots\,E\,v_{i}\,E\,v\,E\,x$. We now assume $j>0$: there exists $s\in B(v,u)$ such that $e_{s}<j$ and $\neg s\,E\,v_{n}$. We claim that $F(v_{n})<_{L}F(s)$, completing the proof. Suppose on the contrary that $F(s)<_{L}F(v_{n})$ ($F(v_{n})\cap F(s)\neq\emptyset$ cannot hold because $\neg s\,E\,v_{n}$). In this case we have $F(s)<_{L}F(x)$ because $f_{L}(v_{n})<_{L}f_{L}(x)$ by 2.9.i. Hence $F(v)\nless_{L}F(s)$ and we can use Proposition 3.9 and 2.8 to obtain a minimal path $s\,E\,u_{1}\,E\,\dots\,E\,u_{\ell}$, with $u_{\ell}=v$ and $e_{u_{i}}<e_{s}$. Since $F(x)\nless_{L}F(s)$ and $F(v)\nless_{L}F(x)$ by Proposition 2.6 there exists $k\leq\ell$ such that $u_{k}\,E\,x$. We distinguish two cases: $F(z)\nless_{L}F(s)$ and $F(z)<_{L}F(s)$. In the first case we apply Proposition 2.6 to the path $s\,E\,u_{1}\,E\,\dots\,E\,u_{k}\,E\,x$: there exists $h\leq k$ such that $z\,E\,u_{h}$. Since $z\,E\,u_{h}\,E\,\dots\,E\,u_{k}\,E\,x$ can be refined to a minimal path and $e_{u_{i}}<j$, the minimality of $j$ is contradicted. In the second case we apply Proposition 2.6 to the path $z\,E\,v_{1}\,E\,\dots\,E\,v_{n}\,E\,x$: there exists $h<n$ (recall that $\neg v_{n}\,E\,s$) such that $v_{h}\,E\,s$. Then $z\,E\,v_{1}\,E\,\dots\,E\,v_{h}\,E\,s\,E\,u_{1}\,E\,\dots\,E\,u_{k}\,E\,x$ can be refined to a path, which can then be refined to a minimal path $z\,E\,w_{1}\,E\,\dots\,E\,w_{r}\,E\,x$. Notice that $w_{r}=u_{p}$, for some $p\leq k$ because $\neg v_{i}\,E\,x$ for every $i\leq h<n$, by minimality of the path $z\,E\,v_{1}\,E\,\dots\,E\,v_{n}\,E\,x$, and $F(s)<_{L}F(x)$. Since $e_{w_{r}}<j$ we contradict again the minimality of $j$. The second half of the statement follows from the first one as usual. ∎ ###### Lemma 3.11 (✠). If $x,y\in B(v,u)$, $f_{R}(x)\leq_{L}f_{R}(v)$ and $f_{L}(u)\leq_{L}f_{L}(y)$, then $vu\,\bar{Q}\,xy$. ###### Proof. The proof is by induction on $e_{x}+e_{y}$. If $e_{x}+e_{y}=0$, then $x=v$ and $y=u$, so that the conclusion is immediate (recall that $Q$ is reflexive). Now assume that $e_{x}+e_{y}>0$ and suppose $e_{x}\leq e_{y}$ (if $e_{y}<e_{x}$ we can employ the usual trick of reversing the representation) and hence $e_{y}>0$. If $u\,E\,y$, then $xu\,Q\,xy$ and, since the induction hypothesis implies $vu\,\bar{Q}\,xu$ (because $e_{u}=0$), we obtain $vu\,\bar{Q}\,xy$. Thus we assume $\neg u\,E\,y$ and hence $F(u)<_{L}F(y)$. Let $z\in B_{e_{y}-1}(v,u)$ be such that $y\,E\,z$. If $f_{L}(u)\leq_{L}f_{L}(z)$, then we can apply the induction hypothesis to $xz$ obtaining $vu\,\bar{Q}\,xz$. Since $xz\,Q\,xy$, we are done. We thus assume $f_{L}(z)<_{L}f_{L}(u)$ which, together with $F(u)<_{L}F(y)$ and $z\,E\,y$, implies $f_{R}(u)<_{L}f_{R}(z)$ and hence $z\,E\,u$. Notice moreover that $z\neq u$ and hence (since $z\neq v$ is obvious) $e_{y}>1$. If $\neg x\,E\,z$, then $xu\,Q\,xz\,Q\,xy$ and, since by induction hypothesis $vu\,\bar{Q}\,xu$, we have $vu\,\bar{Q}\,xy$. If instead $x\,E\,z$ we must have $f_{L}(z)\leq_{L}f_{R}(x)\leq_{L}f_{R}(v)$. Let $t\in B_{e_{y}-2}(v,u)$ be such that $\neg t\,E\,z$. If $F(z)<_{L}F(t)$, then $F(u)<_{L}F(t)$ and $f_{L}(y)<_{L}f_{L}(t)$, so that Proposition 3.8 implies $y\in B_{e_{y}-2}(v,u)$, which is impossible. Hence $F(t)<_{L}F(z)$. This implies $f_{R}(t)<_{L}f_{R}(x)$. It follows that $x\in B_{e_{y}-1}(v,u)$, either by Proposition 3.8, if $F(x)<_{L}F(v)$, or because $x\in B_{1}(v,u)$ if $x\,E\,v$, given that $\neg x\,E\,u$. Since $vu\,\bar{Q}\,tu$ holds by induction hypothesis and we have also $tu\,Q\,tz\,Q\,ty$ it suffices to show that $ty\,\bar{Q}\,xy$. If $t\,E\,x$ the conclusion is immediate, otherwise $F(t)<_{L}F(x)$. Since $F(v)\nless_{L}F(x)$ we can apply Proposition 3.10 finding a minimal path $t\,E\,u_{1}\,E\,\dots\,E\,u_{n}\,E\,x$ and $s\in B_{e_{y}-2}(v,u)$ such that $u_{i}\in B_{e_{y}-2}(v,u)$ and $F(u_{i})<_{L}F(s)$ for all $i\leq n$. We claim that $\neg u_{i}\,E\,y$, for each $i\leq n$, so that $ty\,Q\,u_{1}y\,Q\,\dots\,Q\,u_{n}y\,Q\,xy$ witnesses $ty\,\bar{Q}\,xy$. Indeed, if $u_{i}\,E\,y$, for some $i\leq n$, we would have $f_{L}(y)<_{L}f_{R}(u_{i})<_{L}f_{L}(s)$ and we could apply Proposition 3.8 to obtain $y\in B_{e_{y}-2}(v,u)$, which is impossible. ∎ ###### Theorem 3.12. Let $(V,E)$ be a connected interval graph. If $(V,E)$ does not contain a buried subgraph, then $(W,Q)$ has two components. ###### Proof. Fix a representation $(L,<_{L},f_{L},f_{R})$ of $(V,E)$ and assume that $(V,E)$ does not contain a buried subgraph. We show that if $ab,cd\in W$ are such that $F(a)<_{L}F(b)$ and $F(c)<_{L}F(d)$, then $ab\,\bar{Q}\,cd$. We can assume without loss of generality that $f_{R}(c)\leq_{L}f_{R}(a)$. We consider three cases: 1. Case 1: $f_{L}(b)<_{L}f_{L}(d)$: $B(a,b)$ (which satisfies the hypotheses of $(\maltese)$) is not a buried subgraph and hence by 3.7 we must have $B(a,b)=V$. In particular $c,d\in B(a,b)$ and we are in the hypotheses of Lemma 3.11: we conclude that $ab\,\bar{Q}\,cd$. 2. Case 2: $f_{R}(a)<_{L}f_{L}(d)\leq_{L}f_{L}(b)$: $B(a,d)$ (which satisfies the hypotheses of $(\maltese)$) is not a buried subgraph and hence by 3.7 $b,c\in B(a,d)$. Lemma 3.11 implies both $ad\,\bar{Q}\,ab$ and $ad\,\bar{Q}\,cd$. It follows that $ab\,\bar{Q}\,cd$. 3. Case 3: $f_{L}(d)\leq_{L}f_{R}(a)$: neither $B(a,b)$ nor $B(c,d)$ (which both satisfy the hypotheses of $(\maltese)$) is a buried subgraph. By 3.7 we have $c\in B(a,b)$, which implies $ab\,\bar{Q}\,cb$, and $b\in B(c,d)$, which together with $f_{L}(d)<_{L}f_{L}(b)$ yields $cd\,\bar{Q}\,cb$ (we use Lemma 3.11 in both cases). Thus $ab\,\bar{Q}\,cd$ also in this case.∎ ## 4\. Reverse mathematics and interval graphs Reverse mathematics is a research program, which dates back to the Seventies, whose goal is to find the exact axiomatic strength of theorems from different areas of mathematics. It deals with statements about countable, or countably representable, structures, using the framework of the formal system of second order arithmetic $\mathsf{Z}_{2}$. We do not introduce reverse mathematics here, but refer the reader to monographs such as [Sim09] and [Hir15]. The subsystems of second order arithmetic are obtained by limiting the comprehension and induction axioms of $\mathsf{Z}_{2}$ to specific classes of formulas. We mention only the subsystems we are going to use in this paper: $\mathsf{RCA}_{0}$ is the weak base theory corresponding to computable mathematics, $\mathsf{WKL}_{0}$ extends $\mathsf{RCA}_{0}$ by adding Weak König’s Lemma (each infinite binary tree has an infinite path), and $\mathsf{ACA}_{0}$ is even stronger allowing for definitions of sets by arithmetical comprehension. It is well-known that $\mathsf{WKL}_{0}$ is equivalent to many compactness principles and thus we can claim that a theorem not provable in $\mathsf{WKL}_{0}$ does not admit a proof by compactness. In particular this applies to 1, as 2 shows that it is not provable in $\mathsf{WKL}_{0}$. The second author studied the equivalence of different characterizations of interval orders from the reverse mathematics perspective in [Mar07]. A similar study can be carried out for interval graphs, and we summarize here the main results: full details and proofs are included in the first author’s PhD thesis [FC19], which includes also results about the subclass of indifference graphs (corresponding to proper interval orders studied in [Mar07]). As customary in reverse mathematics, the system in parenthesis indicates where the definition is given or the statement proved. Notice also that in this and in the next section we deal with countable graphs and orders, the only ones second order arithmetic and its subsystems can speak of. In the literature it is possible to find slightly different definitions of interval graphs and orders, which depend on the notion of interval employed. For example intervals may be required to be closed or not. We thus have five conceptually distinct definitions of interval graphs: ###### Definition 4.1 ($\mathsf{RCA}_{0}$). Let $(V,E)$ be a graph. * • $(V,E)$ is an _interval graph_ if there exist a linear order $(L,<_{L})$ and a relation $F\subseteq V\times L$ such that, abbreviating $\\{x\in L\mid(p,x)\in F\\}$ by $F(p)$, for all $p,q\in V$ the following hold: * (i1) $F(p)\neq\emptyset$ and $\forall x,y\in F(p)\,\forall z\in L\,(x<_{L}z<_{L}y\rightarrow z\in F(p))$, * (i2) $p\,E\,q\Leftrightarrow F(p)\cap F(q)\neq\emptyset$. * • $(V,E)$ is a _1-1 interval graph_ if it also satisfies * (i3) $F(p)\neq F(q)$ whenever $p\neq q$. * • $(V,E)$ is a _closed interval graph_ if there exist a linear order $(L,<_{L})$ and two functions $f_{L},f_{R}\colon V\to L$ such that for all $p,q\in V$ * (c1) $f_{L}(p)<_{L}f_{R}(p)$, * (c2) $p\,E\,q\Leftrightarrow f_{L}(p)\leq_{L}f_{R}(q)\leq_{L}f_{R}(p)\lor f_{L}(q)\leq_{L}f_{R}(p)\leq_{L}f_{R}(q)$ * • A closed interval graph $(V,E)$ is a _1-1 closed interval graph_ if we also have * (c3) $f_{R}(p)\neq f_{R}(q)\lor f_{L}(p)\neq f_{L}(q)$ whenever $p\neq q$. * • $(V,E)$ is a _distinguishing interval graph_ if (c1) and (c2) hold together with * (c4) $f_{i}(p)\neq f_{j}(q)$ whenever $p\neq q\lor i\neq j$. ### 4.1. Definitions and characterizations of interval graph In Definition 2.3 we mentioned that every interval graph is a closed interval graph: in fact all the notions introduced in Definition 4.1 are equivalent in a sufficiently strong theory. Our first results concern the systems where the implications between the notions introduced in Definition 4.1 can be proved. The same investigation for interval orders was carried out in [Mar07] and in this respect interval graphs and interval orders behave similarly. Indeed the proofs of the results we are going to state either mimic the corresponding proofs for interval orders or are easily derived from those results. Definition 4.1 enumerates increasingly strong conditions, so that the implications from a later to an earlier notion are easily proved in $\mathsf{RCA}_{0}$. Regarding the other implications we obtain that, as is the case for interval orders, there are three distinct notions of interval graphs in $\mathsf{RCA}_{0}$, namely that of interval, 1-1 interval and closed interval graph. ###### Theorem 4.2 ($\mathsf{RCA}_{0}$). Every closed interval graph is a distinguishing interval graph. ###### Theorem 4.3 ($\mathsf{RCA}_{0}$). The following are equivalent: 1. (1) $\mathsf{WKL}_{0}$; 2. (2) every interval graph is a 1-1 interval graph; 3. (3) every 1-1 interval graph is a closed interval graph; 4. (4) every interval graph is a closed interval graph. ### 4.2. Structural characterizations of interval graphs Since interval graphs are incomparability graphs (and Definition 2.1 can be given in $\mathsf{RCA}_{0}$) we first look at the most important structural characterization of comparability graphs. The first result is due to Jeff Hirst ([Hir87, Theorem 3.20]). ###### Lemma 4.4 ($\mathsf{RCA}_{0}$). The following are equivalent: 1. (1) $\mathsf{WKL}_{0}$; 2. (2) every irreflexive graph such that every cycle of odd length has a triangular chord is a comparability graph. We then consider two structural characterizations of interval graphs (notice that Definition 2.4 can be given in $\mathsf{RCA}_{0}$). The necessity of both conditions is provable in $\mathsf{RCA}_{0}$, but the sufficiency of one of them requires $\mathsf{WKL}_{0}$. ###### Theorem 4.5 ($\mathsf{RCA}_{0}$). Every interval graph is an incomparability graph such that every simple cycle of length four has a chord. Moreover, every interval graph is triangulated and has no asteroidal triples. Every incomparability graph such that every simple cycle of length four has a chord is an interval graph. ###### Theorem 4.6 ($\mathsf{RCA}_{0}$). The following are equivalent: 1. (1) $\mathsf{WKL}_{0}$; 2. (2) if a reflexive graph is triangulated and has no asteroidal triples, then it is an interval graph. Figure 4 summarizes the results about the different definitions and characterizations of interval graphs. The arrows correspond to provability in $\mathsf{RCA}_{0}$, while every implication from a notion below another is equivalent to $\mathsf{WKL}_{0}$. distinguishing interval1-1 closed intervalclosed interval1-1 intervalfour cycle + incomparabilityinterval graphtriangulated + no asteroidal triples Figure 4. Implications in $\mathsf{RCA}_{0}$ Schmerl [Sch05] claimed that the statement “A graph is an interval graph if and only if each finite subgraph is representable by intervals” is equivalent to $\mathsf{WKL}_{0}$. Theorem 4.6 confirms his claim and shows that compactness is necessary to prove the statement. On the other hand, the corresponding statement for interval orders, i.e. an order is an interval order if and only if each suborders is an interval order, is provable in $\mathsf{RCA}_{0}$ because the structural characterization of interval orders (as the partial orders not containing $2\oplus 2$) is provable in $\mathsf{RCA}_{0}$ [Mar07, Theorem 2.1]. The different strengths of the structural characterizations of interval graphs and orders can be traced to the fact that an interval order carries full information about the relative position of the intervals in its representations, while an interval graph does not. Lekkerkerker and Boland [LB62] provide another characterization of interval graphs listing all the forbidden subgraphs. It is routine to check in $\mathsf{RCA}_{0}$ that those graphs are a complete list of graphs whose cycles of length greater than four do not have chords or which contain an asteroidal triple. ### 4.3. Interval graphs and interval orders Different definitions for interval orders, mirroring those of Definition 4.1, were given and studied in [Mar07]. We give here only the most basic one, as the others can be easily guessed from this. ###### Definition 4.7 ($\mathsf{RCA}_{0}$). A partial order $(V,\preceq)$ is an _interval order_ if there exist a linear order $(L,<_{L})$ and a relation $F\subseteq V\times L$ such that, abbreviating $\\{x\in L\mid(p,x)\in F\\}$ by $F(p)$, for all $p,q\in V$ the following hold: * (i1) $F(p)\neq\emptyset$ and $\forall x,y\in F(p)\,\forall z\in L\,(x<_{L}z<_{L}y\rightarrow z\in F(p))$, * (i2) $p\preceq q\Leftrightarrow\forall x\in F(p)\,\forall y\in F(q)\,(x<_{L}y)$. We explore the strength of the statements that allow moving from interval graphs to interval orders and back. By the previous results (and the corresponding ones in [Mar07]) it suffices to consider three different notions on each side, and we concentrate on the relationship between corresponding notions. In one direction everything goes through in $\mathsf{RCA}_{0}$. ###### Theorem 4.8 ($\mathsf{RCA}_{0}$). Let $(V,E)$ be a graph and let $\mathcal{P}$ be any of “interval”, “1-1 interval”, “closed interval”. $(V,E)$ is a $\mathcal{P}$ graph if and only if there exists a $\mathcal{P}$ order $(V,\prec)$ such that $p\,E\,q\Leftrightarrow p\nprec q\land q\nprec p$ for all $p,q\in V$. The other direction is more interesting, as only in one case $\mathsf{RCA}_{0}$ suffices. The proofs of the reversals to $\mathsf{WKL}_{0}$ are modifications of the proof of [Mar07, Theorem 6.4]. ###### Theorem 4.9 ($\mathsf{RCA}_{0}$). Let $(V,\preceq)$ be a partial order. $(V,\preceq)$ is an interval order if and only if $(V,E)$, where $p\,E\,q\Leftrightarrow p\nprec q\land q\nprec p$ for all $p,q\in V$, is an interval graph. ###### Theorem 4.10 ($\mathsf{RCA}_{0}$). The following are equivalent: 1. (1) $\mathsf{WKL}_{0}$ 2. (2) Let $(V,\preceq)$ be a partial order. $(V,\preceq)$ is a 1-1 interval order if and only if $(V,E)$, where $p\,E\,q\Leftrightarrow p\nprec q\land q\nprec p$ for all $p,q\in V$, is a 1-1 interval graph. 3. (3) Let $(V,\preceq)$ be a partial order. $(V,\preceq)$ is a closed interval order if and only if $(V,E)$, where $p\,E\,q\Leftrightarrow p\nprec q\land q\nprec p$ for all $p,q\in V$, is a closed interval graph. ## 5\. Why compactness does not suffice It is immediate (using Theorem 4.5) that Lemmas 3.4 and 3.5 are provable in $\mathsf{RCA}_{0}$. On the other hand, we now show that Theorem 3.12 is much stronger, and indeed equivalent to $\mathsf{ACA}_{0}$. As mentioned in the introduction of the paper, this result explains why the attempts to prove it by compactness cannot succeed. ###### Lemma 5.1 ($\mathsf{ACA}_{0}$). Let $(V,E)$ be a connected reflexive graph which is triangulated and with no asteroidal triples. Suppose furthermore that $a,b,c,d\in V$ are such that $\neg ab\,\bar{Q}\,cd$ and $\neg ab\,\bar{Q}\,dc$. Then there exists a buried subgraph $B\subseteq V$ such that either $a,b\in B$ or $c,d\in B$, and no subgraph $A\subseteq B$, which contains either $a,b$ or $c,d$ respectively, is a buried subgraph. ###### Proof. By Theorems 4.3 and 4.6 $\mathsf{WKL}_{0}$, and a fortiori $\mathsf{ACA}_{0}$, suffices to prove that any connected graph which is triangulated and with no asteroidal triples has a closed interval representation. We then need to check that the proof of Theorem 3.12, which indeed provides a buried subgraph with the desired properties, goes through in $\mathsf{ACA}_{0}$. The first step is checking that, given $v,u\in V$ with $\neg v\,E\,u$, we can carry out 3.6 and define $B(v,u)$ and the various $B_{n}(v,u)$’s in $\mathsf{ACA}_{0}$. In fact the definition of each $B_{n}(v,u)$ in 3.6 uses an instance of arithmetical comprehension and thus the whole construction, as presented there, appears to require the system known as $\mathsf{ACA}_{0}^{+}$, which is properly stronger than $\mathsf{ACA}_{0}$. This problem can however be overcome in the following way. Given $v,u\in V$ as before, we can characterize $B(v,u)$ as the set of $w\in V$ such that there exists a finite tree $T\subseteq 2^{<\mathbb{N}}$ and a label function $\ell\colon T\to V$ with the following properties: * • $\ell(\emptyset)=w$ (here $\emptyset$ is the root of $T$); * • if $\sigma\in T$ is not a leaf of $T$, then $\sigma{}^{\smallfrown}0,\sigma{}^{\smallfrown}1\in T$, $\ell(\sigma)\,E\,\ell(\sigma{}^{\smallfrown}0)$ and $\neg\ell(\sigma)\,E\,\ell(\sigma{}^{\smallfrown}1)$; * • if $\sigma\in T$ is a leaf of $T$, then $\ell(\sigma)\in\\{v,u\\}$. In fact, the tree and its label function describe the ‘steps’ allowing $w$ to enter $B(v,u)$. Moreover $B_{n}(v,u)$ is the set of $w\in V$ such that there exists $T\subseteq 2^{<n}$ and $\ell$ witnessing $w\in B(v,u)$. These characterizations of $B(v,u)$ and $B_{n}(v,u)$ use $\Sigma^{0}_{1}$-formulas, and show that $\mathsf{ACA}_{0}$ suffices to prove the existence of the sets. Once $B(v,u)$ and each $B_{n}(v,u)$ are defined, it is straightforward to check that all subsequent steps in the proof of Theorem 3.12 can be carried out in $\mathsf{RCA}_{0}$. ∎ To prove that Theorem 3.12 implies $\mathsf{ACA}_{0}$ we use the following notions. Given an injective function $f\colon\mathbb{N}\to\mathbb{N}$ we say that $i$ is true for $f$ when $f(k)>f(i)$ for all $k>i$. It is easy to see that there exist infinitely many $i$ which are true for $f$. If $i$ is not true for $f$, i.e. if $f(k)<f(i)$ for some $k>i$, we say that $i$ is false for $f$. Moreover, we say that $i$ is true for $f$ at stage $s$ if $f(k)>f(i)$ whenever $i<k<s$, and that $i$ is false for $f$ at stage $s$ if $f(k)<f(i)$ for some $k$ with $i<k<s$. If the injective function $f$ is fixed, we omit “for $f$” from this terminology. The following Proposition is well-known (see e.g. the discussion after Definition 4.1 in [FHM+16]). ###### Proposition 5.2 ($\mathsf{RCA}_{0}$). The following are equivalent: 1. (1) $\mathsf{ACA}_{0}$; 2. (2) if $f\colon\mathbb{N}\to\mathbb{N}$ is an injective function there exists an infinite set $T$ such that every $i\in T$ is true for $f$. ###### Theorem 5.3 ($\mathsf{RCA}_{0}$). The following are equivalent: 1. (1) $\mathsf{ACA}_{0}$; 2. (2) let $(V,E)$ be a connected graph, triangulated and with no asteroidal triples; if $a,b,c,d\in V$ are such that $\neg ab\,\bar{Q}\,cd$ and $\neg ab\,\bar{Q}\,dc$, then there exists a buried subgraph $B\subseteq V$ such that either $a,b\in B$ or $c,d\in B$, and no subgraph $A\subseteq B$, which contains either $a,b$ or $c,d$ respectively, is a buried subgraph; 3. (3) let $(V,E)$ be a connected closed interval graph; if $(W,Q)$ has more than two components, then there exists a buried subgraph $B\subseteq V$; 4. (4) let $(V,E)$ be a connected closed interval graph; if $(V,E)$ is not uniquely orderable, then there exists a buried subgraph $B\subseteq V$. ###### Proof. $(1\Rightarrow 2)$ is Lemma 5.1. The implication $(2\Rightarrow 3)$ is trivial, while $(3\Rightarrow 4)$ follows directly from Lemma 3.5, which goes through in $\mathsf{RCA}_{0}$. To prove $(4\Rightarrow 1)$ we fix an injective function $f\colon\mathbb{N}\to\mathbb{N}$ and we define (within $\mathsf{RCA}_{0}$) a connected closed interval graph $(V,E)$ such that $(W,Q)$ has more than two components. We then prove, arguing in $\mathsf{RCA}_{0}$, that the unique buried subgraph $B\subseteq V$ codes the (necessarily infinite) set of numbers which are true for $f$. We let $V=\\{a,b,k,r\\}\cup\\{x_{i},y_{i}\mid i\in\mathbb{N}\\}$. Beside making sure that $(V,E)$ is reflexive, the definition of the edge relation is by stages: at stage $s$ we define $E$ on $V_{s}=\\{a,b,k,r\\}\cup\\{x_{i},y_{i}\mid i<s\\}\subseteq V$. At stage $0$ let $k$ be adjacent to $a$, $b$ and $r$ (and add no other edges). At stage $s+1$ we define the vertices adjacent to $x_{s}$ and $y_{s}$ by the following clauses: 1. (a) $a\,E\,x_{s}\,E\,b\,E\,y_{s}$ and $x_{s}\,E\,k\,E\,y_{s}$, 2. (b) $x_{s}\,E\,x_{i}$ and $y_{s}\,E\,y_{i}$ for each $i<s$, 3. (c) $x_{s}\,E\,y_{i}$ for each $i\leq s$, 4. (d) for $i\leq s$, $y_{s}\,E\,x_{i}$ if and only if $i$ is true for $f$ at stage $s+1$. It is immediate that $(V,E)$ is connected. To check that it is a closed interval graph we define a closed interval representation $f_{L},f_{R}:V\to L$ where $(L,<_{L})$ is a dense linear order. The definition of $f_{L}$ and $f_{R}$ reflects the construction of the graph by stages. At stage $0$ assign to the members of $V_{0}$ elements of $L$ satisfying $f_{L}(r)<_{L}f_{L}(k)<_{L}f_{R}(r)<_{L}f_{L}(a)<_{L}f_{R}(a)<_{L}f_{L}(b)<_{L}f_{R}(b)<_{L}f_{R}(k).$ This ensures that we are representing the restriction of the graph to $V_{0}$. At stage $s+1$, first let $f_{L}(x_{s})=f_{L}(a)$ and $f_{R}(y_{s})=f_{R}(b)$ (since this is done at every stage, we are respecting conditions (a) and (b)). We thus still need to define $f_{R}(x_{s})$ and $f_{L}(y_{s})$; first of all we make sure that $f_{L}(b)<_{L}f_{L}(y_{i})<_{L}f_{L}(y_{s})<_{L}f_{R}(x_{s})<_{L}f_{R}(b)$ for every $i<s$, so that (c) is also respected. To respect condition (d) as well we satisfy the following requirements: * • if $i<s$ is true at stage $s+1$, then $f_{R}(x_{s})<_{L}f_{R}(x_{i})$ (which implies $f_{L}(y_{s})<_{L}f_{R}(x_{i})$); * • if $j<s$ is false at stage $s+1$, then $f_{R}(x_{j})<_{L}f_{L}(y_{s})$. The existence of $f_{L}(y_{s})<_{L}f_{R}(x_{s})$ with these properties follows from the density of $L$ and from the fact that if $i<s$ is true at stage $s+1$ and $j<s$ is false at stage $s+1$, then $f_{R}(x_{j})<_{L}f_{R}(x_{i})$. To see this notice that: * • if $i<j$, then $i$ was also true at stage $j+1$ and we set $f_{R}(x_{j})<_{L}f_{R}(x_{i})$ then; * • if $j<i$, then $j$ was already false at stage $i+1$ (if $j$ was true at stage $i+1$, then $f(j)<f(i)$, and $i$ would be false at stage $s+1$ because $j$ is false at that stage), and hence we set $f_{R}(x_{j})<_{L}f_{L}(y_{i})<_{L}f_{R}(x_{i})$ at that stage. Figure 5 depicts a sample interval representation following this construction. $k$$a$$b$$r$$y_{0}$$x_{0}$$y_{1}$$x_{1}$$y_{2}$$x_{2}$$y_{3}$$x_{3}$ Figure 5. Interval representation of $V_{4}$ in case $1$ (and so $2$) becomes false at stage $3$. To check that $(V,E)$ is not uniquely orderable let $\prec_{1}$ be the partial order induced by the interval representation we just described: $v\prec_{1}u$ if and only if $f_{R}(v)<_{L}f_{L}(u)$. Define $\prec_{2}$ so that $\prec_{1}$ and $\prec_{2}$ coincide on $V\setminus\\{s\\}$ and $u\prec_{2}s$ for all $u\in V\setminus\\{r,k\\}$. It is immediate that both $\prec_{1}$ and $\prec_{2}$ are associated to $(V,E)$, and that $\prec_{2}$ is not dual of $\prec_{1}$. By (4) there exists a buried subgraph $B\subseteq V$. First of all notice that $k\in K(B)$ and hence $k\notin B$. Now observe that $r\in B$ implies, using Conditions (i) and (iii) of Definition 3.1, that either some $x_{n}$ or some $y_{n}$ belongs to $B$. From there, using Condition (iii) again, it is easy to see that $B=V\setminus\\{k\\}$ and hence $R(B)=\emptyset$, contradicting Condition (ii). Thus $r\notin B$. Then, in order to satisfy Condition (i), we must have either $a,b\in B$ or $a,y_{n}\in B$ or $x_{m},y_{n}\in B$, for some $n$ and some $m$ which is false at stage $n+1$. In any case we have $a\in B$: in the first two cases this is obvious, and in the latter case this follows from Condition (iii) because $a\,E\,x_{m}$ and $\neg a\,E\,y_{n}$ for every $n$ and $m$. But then, using $b\,E\,y_{n}$ and $\neg b\,E\,a$ we obtain $b\in B$ even in the second and third case. Thus we can conclude that $a,b\in B$. For each $n$ we have $y_{n}\,E\,b$ and $\neg y_{n}\,E\,a$ and therefore $y_{n}\in B$. Since $b\,E\,x_{n}\,E\,a$ for each $n\in\mathbb{N}$, then either $x_{n}\in K(B)$ or $x_{n}\in B$ depending whether $x_{n}$ is adjacent to every $y_{m}$ or not, namely whether $n$ is true or false for $f$. Therefore we showed $B=\\{a,b\\}\cup\\{y_{n}\mid n\in\mathbb{N}\\}\cup\\{x_{n}\mid n\text{ is false}\\},$ so that $K(B)=\\{k\\}\cup\\{x_{n}\mid n\text{ is true}\\}$ and $R(B)=\\{r\\}$. Then $T=\\{n\mid x_{n}\notin B\\}$ is the (necessarily infinite) set of all $n$ which are true for $f$. ∎ ## References * [FC19] Marta Fiori-Carones. Filling cages. Reverse mathematics and combinatorial principles. PhD thesis, Università di Udine, Italy, 2019. * [FHM+16] Emanuele Frittaion, Matthew Hendtlass, Alberto Marcone, Paul Shafer, and Jeroen Van der Meeren. Reverse mathematics, well-quasi-orders, and Noetherian spaces. Archive for Mathematical Logic, 55(3-4):431–459, 2016. * [Fis70] Peter C Fishburn. Intransitive indifference with unequal indifference intervals. Journal of Mathematical Psychology, 7(1):144–149, 1970. * [Fis85] Peter C. Fishburn. Interval Orders and Interval Graphs. Wiley, 1985. * [Han82] Philip Hanlon. Counting interval graphs. Transactions of the American Mathematical Society, 272(2):383–426, 1982. * [Hir87] Jeffry L. Hirst. Combinatorics in Subsystems of Second Order Arithmetic. PhD thesis, The Pennsylvania State University, 1987. * [Hir15] Denis R. Hirschfeldt. Slicing the Truth. World Scientific, 2015. * [LB62] Cornelis J. Lekkerkerker and Johan C. Boland. Representation of a finite graph by a set of intervals on the real line. Fundamenta Mathematicae, 51:45–64, 1962. * [Mar07] Alberto Marcone. Interval orders and reverse mathematics. Notre Dame Journal of Formal Logic, 48:425–448, 2007. * [Sch05] James H. Schmerl. Reverse mathematics and graph coloring: eliminating diagonalization. In S. Simpson, editor, Reverse Mathematics 2001, volume 21 of Lecture Notes in Logic, pages 331–348. Association of Symbolic Logic, La Jolla, CA, 2005. * [Sim09] Stephen G. Simpson. Subsystems of Second Order Arithmetic. Cambridge University Press, second edition, 2009. * [Tro97] William T. Trotter. New perspectives on interval orders and interval graphs. In Surveys in combinatorics, 1997 (London), volume 241 of London Math. Soc. Lecture Note Ser., pages 237–286. Cambridge Univ. Press, Cambridge, 1997. * [Wie14] Norbert Wiener. A contribution to the theory of relative position. Proc. Camb. Philos. Soc., 17:441–449, 1914.
# Extended theoretical transition data in C i – iv W. Li,1,3 A. M. Amarsi,2 A. Papoulia1,3 J. Ekman1 and P. Jönsson1 1Department of Materials Science and Applied Mathematics, Malmö University, SE-205 06, Malmö, Sweden 2Theoretical Astrophysics, Department of Physics and Astronomy, Uppsala University, Box 516, SE-751 20 Uppsala, Sweden 3Division of Mathematical Physics, Lund University, Post Office Box 118, SE-221 00 Lund, Sweden E-mail<EMAIL_ADDRESS> (Accepted XXX. Received YYY; in original form ZZZ) ###### Abstract Accurate atomic data are essential for opacity calculations and for abundance analyses of the Sun and other stars. The aim of this work is to provide accurate and extensive results of energy levels and transition data for C i – iv. The Multiconfiguration Dirac–Hartree–Fock and relativistic configuration interaction methods were used in the present work. To improve the quality of the wave functions and reduce the relative differences between length and velocity forms for transition data involving high Rydberg states, alternative computational strategies were employed by imposing restrictions on the electron substitutions when constructing the orbital basis for each atom and ion. Transition data, e.g., weighted oscillator strengths and transition probabilities, are given for radiative electric dipole (E1) transitions involving levels up to $\mathrm{1s^{2}2s^{2}2p6s}$ for C i, up to $\mathrm{1s^{2}2s^{2}7f}$ for C ii, up to $\mathrm{1s^{2}2s7f}$ for C iii, and up to $\mathrm{1s^{2}8g}$ for C iv. Using the difference between the transition rates in length and velocity gauges as an internal validation, the average uncertainties of all presented E1 transitions are estimated to be 8.05%, 7.20%, 1.77%, and 0.28%, respectively, for C i – iv. Extensive comparisons with available experimental and theoretical results are performed and good agreement is observed for most of the transitions. In addition, the C i data were employed in a reanalysis of the solar carbon abundance. The new transition data give a line-by-line dispersion similar to the one obtained when using transition data that are typically used in stellar spectroscopic applications today. ###### keywords: Atomic data — Atomic processes — Line: formation — Radiative transfer — Sun: abundances — Methods: numerical ††pubyear: 2020††pagerange: Extended theoretical transition data in C i – iv–A ## 1 Introduction Accurate atomic data are of fundamental importance to many different fields of astronomy and astrophysics. This is particularly true for carbon. As the fourth-most abundant metal in the cosmos (Asplund et al., 2009), carbon is a major source of opacity in the atmospheres and interiors of stars. Complete and reliable sets of atomic data for carbon are essential for stellar opacity calculations, because of their significant impact on stellar structure and evolution (e.g. VandenBerg et al., 2012; Chen et al., 2020). Accurate atomic data for carbon are also important in the context of spectroscopic abundance analyses and Galactic Archaeology. Carbon abundances measured in late-type stars help us to understand the nucleosynthesis of massive stars and AGB stars, and thus the Galactic chemical evolution (e.g. Franchini et al., 2020; Jofré et al., 2020; Stonkutė et al., 2020). In early- type stars, carbon abundances help constrain the present-day Cosmic Abundance Standard (e.g. Nieva & Przybilla, 2008, 2012; Alexeeva et al., 2019). In the Sun, the carbon abundance is precisely measured in order to put different cosmic objects onto a common scale (e.g. Caffau et al., 2010; Amarsi et al., 2019). In all of these cases, oscillator strengths for C i (cool stars) and for C i – iv (hot stars) underpin the spectroscopic analyses; this is especially the case for studies that relax the assumption of local thermodynamic equilibrium (LTE; e.g. Przybilla et al., 2001; Nieva & Przybilla, 2006), in which case much larger sets of reliable atomic data are needed. On the experimental side, a number of studies of transition data have been presented in the literature. Neutral C i transition probabilities for the $\mathrm{2p4p\rightarrow 2p3s}$ transition array have been studied by Miller et al. (1974) using a spectroscopic shock tube and by Jones & Wiese (1984) using a wall-stabilized arc. The measurements of relative oscillator strengths for $\mathrm{2p3p\rightarrow 2p3s}$, $\mathrm{2p3d\rightarrow 2p3p}$ and $\mathrm{2p4s\rightarrow 2p3p}$ have been performed by Musielok et al. (1997); Bacawski et al. (2001); Golly et al. (2003) using a wall-stabilized arc. Older measurements of oscillator strengths are also available using the same technique (Maecker, 1953; Richter, 1958; Foster, 1962; Boldt, 1963; Goldbach & Nollez, 1987; Goldbach et al., 1989). By analysing the high-resolution spectra obtained with the Goddard High Resolution Spectrograph on the Hubble Space Telescope, Federman & Zsargo (2001) derived oscillator strengths for C i lines below 1200 Å. For C ii, a number of measurements have also been performed. Träbert et al. (1999) measured the radiative decay rates for the intercombination (IC) transitions $\mathrm{2s2p^{2}~{}^{4}P\rightarrow 2s^{2}2p~{}^{2}P^{o}}$ at a heavy-ion storage ring, and the total measured radiative decay rates to the ground term were 125.8 $\pm$ 0.9 s-1 for $\mathrm{{}^{4}P_{1/2}}$, 9.61 $\pm$ 0.05 s-1 for $\mathrm{{}^{4}P_{3/2}}$, and 45.35 $\pm$ 0.15 s-1 for $\mathrm{{}^{4}P_{5/2}}$. The aforementioned results are, however, not in agreement with the values measured by Fang et al. (1993) using a radio- frequency ion trap, i.e., 146.4(+8.3, -9.2) s-1 for $\mathrm{{}^{4}P_{1/2}}$, 11.6(+0.8, -1.7) s-1 for $\mathrm{{}^{4}P_{3/2}}$, and 51.2(+2.6, -3.5) s-1 for $\mathrm{{}^{4}P_{5/2}}$. Goly & Weniger (1982) measured the transition probabilities from a helium-carbon arc for some multiplets of $\mathrm{\\{2p^{3},2s^{2}3p\\}\rightarrow 2s2p^{2}}$ and $\mathrm{2s^{2}4s\rightarrow 2s^{2}3p}$ with estimated relative uncertainty of 50%. Using an electric shock tube, Roberts & Eckerle (1967) provided the relative oscillator strengths of some C ii multiplets with relative uncertainties of 7%. Reistad et al. (1986) gave lifetimes for 11 C ii levels using the beam-foil excitation technique and extensive cascade analyses. For C iii, the IC decay rate of the $\mathrm{2s2p~{}^{3}P^{o}_{1}\rightarrow 2s^{2}~{}^{1}S_{0}}$ transition was measured to be 121.0 $\pm$ 7 s-1 by Kwong et al. (1993) using a radio-frequency ion trap and 102.94 $\pm$ 0.14 s-1 by Doerfert et al. (1997) using a heavy-ion storage ring. The discrepancy between the values obtained from the two different methods is quite large, i.e., of the order of 15%. The result given by the latter measurement is closer to earlier $ab~{}initio$ calculations ranging between 100 and 104 s-1 (Fleming et al., 1994; Fischer, 1994; Ynnerman & Fischer, 1995). Several measurements have also been performed for the lifetimes of the low-lying levels of C iii (Reistad & Martinson, 1986; Mickey, 1970; Nandi et al., 1996; Buchet-Poulizac & Buchet, 1973a). For the system of Li-like C iv, the transition probabilities of the $\mathrm{1s^{2}2p~{}^{2}P^{o}_{1/2,3/2}\rightarrow 1s^{2}2s~{}^{2}S_{1/2}}$ transitions were measured by Berkner et al. (1965) using the foil-excitation technique and by Knystautas et al. (1971) using the beam-foil technique, respectively. There are also a number of measurements of lifetimes in C iv using the beam-foil technique (Donnelly et al., 1978; Buchet-Poulizac & Buchet, 1973b; Jacques et al., 1980). On the theoretical side, Froese Fischer et al. have performed detailed studies of C i – iv, focusing on the low-lying levels. They carried out Multiconfiguration Hartree-Fock (MCHF) calculations and used the Breit-Pauli (MCHF-BP) approximation for computing energy levels and transition properties, e.g., transition probabilities, oscillator strengths, and lifetimes, in C i (Tachiev & Fischer, 2001; Fischer, 2006; Fischer & Tachiev, 2004), C ii (Tachiev & Fischer, 2000), C iii (Tachiev & Fischer, 1999; Fischer, 2000), and C iv (Godefroid et al., 2001; Fischer et al., 1998). Hibbert et al. have presented extensive calculations for optical transitions. They used the CIV3 code (Hibbert, 1975) to calculate oscillator strengths and transition probabilities in C i (Hibbert et al., 1993), C ii (Corrégé & Hibbert, 2004), and C iii (Kingston & Hibbert, 2000). In the calculations of Hibbert et al. (1993); Corrégé & Hibbert (2004), empirical adjustments were introduced to the diagonal matrix elements in order to accurately reproduce energy splittings. Their C i oscillator strengths are frequently used in the abundance analyses of cool stars (Sect. 5). A number of other authors have also presented theoretical transition data for carbon. Zatsarinny & Fischer (2002) calculated the oscillator strengths for transitions to high-lying excited states of C i using a spline frozen-cores method. Nussbaumer & Storey (1984) provided the radiative transition probabilities using the $LS$-coupling approximation and intermediate coupling approximation, respectively, for the six energetically lowest configurations of C i. Nussbaumer & Storey (1981) calculated the transition probabilities for C ii, from terms up to $\mathrm{2s^{2}4f~{}^{2}F^{o}}$, using the $LS$-coupling and close coupling (CC) approximation, respectively. In view of the great astrophysical interest for large sets of homogeneous atomic data, extensive spectrum calculations of transition data in the carbon atom and carbon ions were carried out under the umbrella of the Opacity Project using the CC approximation of the R-matrix theory, and the results are available in the Opacity Project online database (TOPbase; Cunto & Mendoza (1992); Cunto et al. (1993)). The latest compilation of C i transition probabilities was made available by Haris & Kramida (2017), and those of C ii- iv can be found in earlier compilations by Wiese & Fuhr (2007b); Wiese & Fuhr (2007a) and Fuhr (2006). In this context, the General-purpose Relativistic Atomic Structure Package (Grasp) has, more recently, been used by Aggarwal & Keenan (2015) to predict the radiative decay rates and lifetimes of 166 levels belonging to the n $\leq$ 5 configurations in C iii. Using an updated and extended version of this code (Grasp2K), Jönsson et al. (2010) determined transition data involving 26 levels in C ii. Although for the past decades a considerable amount of research has been conducted for carbon, there is still a need for extended sets of reliable theoretical transition data. To address this, we have carried out new calculations based on the fully relativistic Multiconfiguration Dirac–Hartree–Fock (MCDHF) and relativistic configuration interaction (RCI) methods, as implemented in the newest version of the Grasp code, Grasp2018 (Jönsson et al., 2013; Fischer et al., 2019). We performed energy spectrum calculations for 100, 69, 114, and 53 states, in C i – iv, respectively. Electric dipole (E1) transition data (wavelengths, transition probabilities, line strengths, and oscillator strengths) were computed along with the corresponding lifetimes of these states. This paper is structured into six sections, including the introduction. Our theoretical methods are described in Sect. 2, and computational details are given in Sect. 3. In Sect. 4, we present our results and the validation of the data. As a complementary method of validation, in Sect. 5, we use the derived data in a reanalysis of the solar carbon abundance. Finally, we present our conclusions in Sect. 6. ## 2 Theory In the Multiconfiguration Dirac-Hartree-Fock (MCDHF) method (Grant, 2007; Fischer et al., 2016), wave functions for atomic states $\gamma^{(j)}\,PJM$, $j=1,2,\ldots,N$ with angular momentum quantum numbers $JM$ and parity $P$ are expanded over ${N_{\mathrm{CSFs}}}$ configuration state functions $\Psi(\gamma^{(j)}\,PJM)=\sum_{i}^{{N_{\mathrm{CSFs}}}}c^{(j)}_{i}\,\Phi(\gamma_{i}\,PJM).$ (1) The configuration state functions (CSFs) are $jj$-coupled many-electron functions, recursively built from products of one-electron Dirac orbitals. As for the notation, $\gamma_{i}$ specifies the occupied subshells of the CSF with their complete angular coupling tree information. The radial large and small components of the one-electron orbitals and the expansion coefficients {$c^{(j)}_{i}$} of the CSFs are obtained, for a number of targeted states, by solving the Dirac-Hartree-Fock radial equations and the configuration interaction eigenvalue problem resulting from applying the variational principle on the statistically weighted energy functional of the targeted states with terms added for preserving the orthonormality of the one-electron orbitals. The energy functional is based on the Dirac-Coulomb (DC) Hamiltonian and accounts for relativistic kinematic effects. Once the radial components of the one-electron orbitals are determined, higher-order interactions, such as the transverse photon interaction and quantum electrodynamic (QED) effects (vacuum polarization and self-energy), are added to the Dirac-Coulomb Hamiltonian. Keeping the radial components fixed, the expansion coefficients {$c^{(j)}_{i}$} of the CSFs for the targeted states are obtained by solving the configuration interaction eigenvalue problem. The evaluation of radiative E1 transition data (transition probabilities, oscillator strengths) between two states: $\gamma^{\prime}P^{\prime}J^{\prime}M^{\prime}$ and $\gamma PJM$ is non- trivial. The transition data can be expressed in terms of reduced matrix elements of the transition operator ${\bf T}^{(1)}$: $\displaystyle\langle\,\Psi(\gamma PJ)\,\|{\bf T}^{(1)}\|\,\Psi(\gamma^{\prime}P^{\prime}J^{\prime})\,\rangle$ $\displaystyle=$ $\displaystyle\sum_{j,k}c_{j}c^{\prime}_{k}\;\langle\,\Phi(\gamma_{j}PJ)\,\|{\bf T}^{(1)}\|\,\Phi(\gamma^{\prime}_{k}P^{\prime}J^{\prime})\,\rangle,$ (2) where $c_{j}$ and $c^{\prime}_{k}$ are, respectively, the expansion coefficients of the CSFs for the lower and upper states, and the summation occurs over all the CSFs for the lower and upper states. The reduced matrix elements are expressed via spin-angular coefficients $d^{(1)}_{ab}$ and operator strengths as: $\displaystyle\langle\,\Phi(\gamma_{j}PJ)\,\|{\bf T}^{(1)}\|\,\Phi(\gamma^{\prime}_{k}P^{\prime}J^{\prime})\,\rangle$ $\displaystyle=$ $\displaystyle\sum_{a,b}d^{(1)}_{ab}\;\langle\,n_{a}l_{a}j_{a}\,\|{\bf T}^{(1)}\|\,n_{b}l_{b}j_{b}\,\rangle.$ (3) Allowing for the fact that we are now using Brink-and-Satchler type reduced matrix elements, we have $\displaystyle\langle\,n_{a}l_{a}j_{a}\,\|{\bf T}^{(1)}\|\,n_{b}l_{b}j_{b}\,\rangle$ $\displaystyle=$ $\displaystyle\left(\frac{(2j_{b}+1)\omega}{\pi c}\right)^{1/2}(-1)^{j_{a}-1/2}\begin{pmatrix}j_{a}~{}~{}~{}~{}1~{}~{}~{}~{}j_{b}\\\ \frac{1}{2}~{}~{}~{}~{}0~{}-\frac{1}{2}\end{pmatrix}\overline{M_{ab}},$ (4) where $\overline{M_{ab}}$ is the radiative transition integral defined by Grant (1974). The factor in front of $\overline{M_{ab}}$ is the Wigner 3-j symbol that gives the angular part of the matrix element. The $\overline{M_{ab}}$ integral can be written $\overline{M_{ab}}=\overline{M^{e}_{ab}}+G\overline{M^{l}_{ab}}$, where $G$ is the gauge parameter. When $G=0$ we get the Coulomb gauge, whereas for $G=\sqrt{2}$ we get the Babushkin gauge. The Babushkin gauge corresponds to the length gauge in the non-relativistic limit and puts weight on the outer part of the wave functions (Grant, 1974; Hibbert, 1974). The Coulomb gauge corresponds to the velocity gauge and puts more weight on the inner part of the wave functions (Papoulia et al., 2019). For E1 transitions, the Babushkin and Coulomb gauges give the same value of the transition moment for exact solutions of the Dirac-equation (Grant, 1974). For approximate solutions, the transition moments differ, and the quantity $dT$, defined as (Froese Fischer, 2009; Ekman et al., 2014) $dT=\frac{|A_{l}-A_{v}|}{\max(A_{l},A_{v})},$ (5) where $A_{l}$ and $A_{v}$ are transition rates in length and velocity form, can be used as an estimation of the uncertainty of the computed rate. ## 3 Computational schemes Calculations were performed in the extended optimal level (EOL) scheme (Dyall et al., 1989) for the weighted average of the even and odd parity states. The CSF expansions were determined using the multireference-single-double (MR-SD) method, allowing single and double (SD) substitutions from a set of important configurations, referred to as the MR, to orbitals in an active set (AS) (Olsen et al., 1988; Sturesson et al., 2007; Fischer et al., 2016). The orbitals in the AS are divided into spectroscopic orbitals, which build the configurations in the MR, and correlation orbitals, which are introduced to correct the initially obtained wave functions. During the different steps of the calculations for C i – iv, the CSF expansions were systematically enlarged by adding layers of correlation orbitals. MCDHF calculations aim to generate an orbital set. The orbital set is then used in RCI calculations based on CSF expansions that can be enlarged to capture additional electron correlation effects. For the same CSF expansion, different orbital sets give different results for both energy levels and transition data. Conventionally, MCDHF calculations are performed for CSF expansions obtained by allowing substitutions not only from the valence subshells, but also from the subshells deeper in the core, accounting for valence-valence (VV), core-valence (CV), and core-core (CC) electron correlation effects. Using orbital sets from such calculations, Pehlivan Rhodin et al. (2017) predicted large $dT$ values for transitions between low- lying states and high Rydberg states, indicating substantial uncertainties in the corresponding transition data. For transitions involving high Rydberg states, it was shown that the velocity gauge gave the more accurate results, which is contradictory to the general belief that the length gauge is the preferred one (Hibbert, 1974). Analyzing the situation more carefully, Papoulia et al. (2019) found that correlation orbitals resulting from MCDHF calculations based on CSF expansions obtained by allowing substitutions from deeper subshells are very contracted in comparison with the outer Rydberg orbitals. As a consequence, the outer parts of the wave functions for the Rydberg states are not accurately described. Thus, the length form that probes the outer part of the wave functions does not produce trustworthy results, while the velocity form that probes the inner part of the wave functions yields more reliable transition rates. In the same work, the authors showed how transition rates that are only weakly sensitive to the choice of gauge can be obtained, by paying close attention to the CSF generation strategies for the MCDHF calculations. In the present work, following the suggestion by Papoulia et al. (2019), the MCDHF calculations were based on CSF expansions for which we impose restrictions on the substitutions from the inner subshells and obtain, as a consequence, correlation orbitals that overlap more with the spectroscopic orbitals of the higher Rydberg states, adding to a better representation of the outer parts of the corresponding wave functions. The MR and orbital sets for each atom and ion are presented in Table 1. The computational scheme, including CSF generation strategies, for each atom and ion is discussed in detail below. The MCDHF calculations were followed by RCI calculations, including the Breit interaction and leading QED effects. Table 1: Summary of the computational schemes for C i – iv. The first column displays the configurations of the targeted states. MR and AS, respectively, denote the multireference sets and the active sets of orbitals used in the MCDHF and RCI calculations, and ${N_{\mathrm{CSFs}}}$ are the numbers of generated CSFs in the final RCI calculations, for the even (e) and the odd (o) parity states. Targeted configurations | MR | AS | $N_{\mathrm{{CSFs}}}$ | | ---|---|---|---|---|--- | C i, $\mathrm{N_{levels}}=100$ | | | $\mathrm{2s2p^{3}}$ | $\mathrm{2s2p^{3}}$ | {11s,10p,10d,9f, | e: 14 941 842 | | $\mathrm{2s^{2}2p}\\{n_{1}\mathrm{s},n_{2}\mathrm{p},n_{3}\mathrm{d,4f}\\}$ | $\mathrm{2s^{2}2p}\\{n_{1}\mathrm{s},n_{2}\mathrm{p},n_{3}\mathrm{d,4f}\\}$ | 7g,6h} | o: 15 572 953 | | ($3\leq n_{1}\leq 6$, $2\leq n_{2}\leq 5$, $3\leq n_{3}\leq 5$) | ($3\leq n_{1}\leq 6$, $2\leq n_{2}\leq 6$, $3\leq n_{3}\leq 5$) | | | | | $\mathrm{2p^{3}}\\{n_{1}\mathrm{s},n_{2}\mathrm{p},n_{3}\mathrm{d}\\}$ | | | | | ($3\leq n_{1}\leq 6$, $3\leq n_{2}\leq 5$, $3\leq n_{3}\leq 6$) | | | | | $\mathrm{2s2p^{2}\\{3s,3p,4p,6p,6d,7s\\}}$ | | | | | $\mathrm{2s2p}\\{n_{1}\mathrm{s},n_{2}\mathrm{p},n_{3}\mathrm{d,4f\\}6d}$ | | | | | ($3\leq n_{1}\leq 6$, $3\leq n_{2}\leq 5$, $3\leq n_{3}\leq 5$) | | | | | C ii, $\mathrm{N_{levels}}=69$ | | | $\mathrm{2s^{2}}nl$($n\leq 6,l\leq 4$) | $\mathrm{2s2p^{2}}$, $\mathrm{2s^{2}}\\{n_{1}\mathrm{s},n_{2}\mathrm{p},n_{3}\mathrm{d},n_{4}\mathrm{f},n_{5}\mathrm{g}\\}$ | {$\mathrm{{14s,14p,12d,12f}},$ | e: 6 415 798 | | $\mathrm{2s^{2}7l}$($l\leq 3$) | $(3\leq n_{1}\leq 9,2\leq n_{2}\leq 9,3\leq n_{3}\leq 7$, | $\mathrm{10g,8h}$} | o: 4 988 973 | | $\mathrm{2s2p^{2}}$, $\mathrm{2p^{3}}$, | $4\leq n_{4}\leq 7,5\leq n_{5}\leq 6)$ | | | | $\mathrm{2s2p3s}$, $\mathrm{2s2p3p}$ | $\mathrm{2p^{3}}$, $\mathrm{2p^{2}}\\{n_{1}\mathrm{s},n_{2}\mathrm{p},n_{3}\mathrm{d},n_{4}\mathrm{f},n_{5}\mathrm{g}\\}$ | | | | | $(3\leq n_{1}\leq 9,4\leq n_{2}\leq 9,3\leq n_{3}\leq 7$, | | | | | $4\leq n_{4}\leq 7,5\leq n_{5}\leq 6)$ | | | | | $\mathrm{2s2p3s}$, $\mathrm{2s2p3p}$ | | | | | C iii, $\mathrm{N_{levels}}=114$ | | | $\mathrm{2s}nl$($n\leq 7,l\leq 4$) | $\mathrm{2s}nl$ ($n\leq 7,l\leq 4$) | {$\mathrm{12s,12p,12d,12f,}$ | e: 1 578 620 | | $\mathrm{2p^{2}}$, $\mathrm{2p\\{3s,3p,3d\\}}$ | $\mathrm{2p^{2}}$, $\mathrm{2p\\{3s,3p,3d\\}}$ | $\mathrm{11g,8h}$} | o: 1 274 147 | | | C iv, $\mathrm{N_{levels}}=53$ | | | $\mathrm{1s^{2}}nl$ ($n\leq 8,l\leq 4$) | $\mathrm{1s^{2}}nl$ ($n\leq 8,l\leq 4$) | {$\mathrm{14s,14p,14d,12f,12g,}$ | e: 1 077 872 | | $\mathrm{1s^{2}6h}$ | $\mathrm{1s^{2}6h}$ | $\mathrm{8h,7i}$} | o: 1 287 706 | | ### 3.1 C i As seen in Table 1, in the computations of neutral carbon, configurations with ${n=7~{}(l=\mathrm{s});6~{}(l=\mathrm{p,d})}$, which are not of direct relevance, were included in the MR set to obtain orbitals that are spatially extended, improving the quality of the outer parts of the wave functions of the higher Rydberg states. The MCDHF calculations were performed using CSF expansions that were produced by SD substitutions from the valence orbitals of the configurations in the MR to the active set of orbitals, with the restriction of allowing maximum one substitution from orbitals with $n=2$. The $\mathrm{1s^{2}}$ core was kept closed and, at this point, the expansions of the atomic states accounted for VV electron correlation. As a final step, an RCI calculation was performed for the largest SD valence expansion augmented by a CV expansion. The CV expansion was obtained by allowing SD substitutions from the valence orbitals and the $\mathrm{1s^{2}}$ core of the configurations in the MR, with the restriction that there should be at most one substitution from $\mathrm{1s^{2}}$. The numbers of CSFs in the final even and odd state expansions are, respectively, 14 941 842 and 15 572 953, distributed over the different $J$ symmetries. ### 3.2 C ii Similarly to the computations in C i, in the computations of the singly- ionized carbon, the configurations $\mathrm{2s^{2}\\{8s,8p,9s,9p\\}}$, which are not our prime targets, were included in the MR set (see also Table 1). In this manner, we generated orbitals that are localized farther from the atomic core. The MCDHF calculations were performed using CSF expansions obtained by allowing SD substitutions from the valence orbitals of the MR configurations. During this stage, the $\mathrm{1s^{2}}$ core remained frozen and the CSF expansions accounted for VV correlation. The final wave functions of the targeted states were determined in an RCI calculation, which included CSF expansions that were formed by allowing SD substitution from all subshells of the MR configurations, with the restriction that there should be at most one substitution from the $\mathrm{1s^{2}}$ core. The numbers of CSFs in the final even and odd state expansions are, respectively, 6 415 798 and 4 988 973, distributed over the different $J$ symmetries. ### 3.3 C iii In the computations of beryllium-like carbon, the MR simply consisted of the targeted configurations (see also Table 1). The CSF expansions used in the MCDHF calculations were obtained by allowing SD substitutions from the valence orbitals, accounting for VV correlation effects. The final wave functions of the targeted states were determined in subsequent RCI calculations, which included CSFs that were formed by allowing single, double, and triple (SDT) substitutions from all orbitals of the MR configurations, with the limitation of leaving no more than one hole in the $\mathrm{1s^{2}}$ atomic core. The final even and odd state expansions, respectively, contained 1 578 620 and 1 274 147 CSFs, distributed over the different $J$ symmetries. ### 3.4 C iv Likewise the computations in C iii, the MR in the computations of lithium-like carbon was solely represented by the targeted configurations (see also Table 1). In the MCDHF calculations, the CSF expansions were acquired by implementing SD electron substitutions from the configurations in the MR, with the restriction of allowing maximum one hole in the $\mathrm{1s^{2}}$ core. In this case, the shape of the correlation orbitals was established by CSFs accounting for valence (V) and CV correlation effects. In the subsequent RCI calculations, the CSF expansions were enlarged by enabling all SDT substitutions from the orbitals in the MR to the active set of orbitals. The final expansions of the atomic states gave rise to 1 077 872 CSFs with even parity and 1 287 706 CSFs with odd parity, respectively, shared among the different $J$ symmetry blocks. ## 4 Results Figure 1: Left panel: Comparison of computed energy levels in the present work with data from the NIST database, for C i – iv. The dashed lines indicate the $-$0.5% and 0.5% relative discrepancies. Right panel: The relative differences between the lifetimes in length and velocity forms, for C i – iv. The dashed and solid lines indicate the 5% and 10% relative differences, respectively. No., as label in the x-axis, corresponds to the No. in Table 4. The energy spectra and wave function composition in $LS$-coupling for the 100, 69, 114, and 53 lowest states, respectively, for C i – iv are given in Table 4. In the tables, the states are given with unique labels (Gaigalas et al., 2017), and the labelling is determined by the CSFs with the largest coefficient in the expansion of Eq. (1). We first summarise the results here, before discussing the individual ions in detail in Sects. 4.1 – 4.4, below. The accuracy of the wave functions from the present calculations was evaluated by comparing the calculated energy levels with experimental data provided via the National Institute of Standards and Technology (NIST) Atomic Spectra Database (Kramida et al., 2019). In the left panel of Fig. 1, energy levels computed in this work are compared with the NIST data. A closer inspection of the figure reveals that the relative discrepancies between the experimental and the computed in this work energies are, in most cases, about $-$0.35%, $-$0.08%, 0.03%, and 0.003%, respectively, for C i – iv. Only for levels of the $\mathrm{2s2p^{3}}$ configuration in C i, the disagreements are larger than 1.0%. The average difference of the computed energy levels relative to the energies from the NIST database is 0.41%, 0.081%, 0.041%, and 0.0044%, respectively, for C i – iv. In Table 4, lifetimes in length and velocity gauges are also presented. The right panel of Fig. 1 presents the relative differences between the lifetimes in length and velocity forms for C i – iv. Except for a few long-lived states that can decay to the ground state only through IC transitions, the relative differences are well below 5%. Table 2: Distribution of the uncertainties $dT$ (in %) of the computed transition rates in C i – iv depending on the magnitude of the rates. The transition rates are arranged in five groups based on the magnitude of the $A$ values (in s-1). The number of transitions, No., the mean $dT$, $\langle dT\rangle$, (in %), and the standard deviations, $\sigma$, are given for each group of transitions, in C i – iv, respectively. The last three rows show the proportions of the transitions with $dT$ less than 20%, 10%, and 5% in all the transitions with $A\geq$ $10^{2}$ s-1 for C i and C ii and $A\geq$ $10^{0}$ s-1 for C iii and C iv, respectively. | C i | | C ii | | C iii | | C iv ---|---|---|---|---|---|---|--- Group | No. | $\langle dT\rangle(\%)$ | $\sigma$ | | No. | $\langle dT\rangle(\%)$ | $\sigma$ | | No. | $\langle dT\rangle(\%)$ | $\sigma$ | | No. | $\langle dT\rangle(\%)$ | $\sigma$ $<10^{0}$ | 62 | 52.6 | 0.34 | | 80 | 29.6 | 0.32 | | 137 | 10.8 | 0.18 | | 20 | 5.92 | 0.061 $10^{0}-10^{2}$ | 156 | 34.0 | 0.25 | | 134 | 17.1 | 0.24 | | 239 | 5.57 | 0.096 | | 10 | 2.38 | 0.017 $10^{2}-10^{4}$ | 451 | 13.2 | 0.15 | | 128 | 14.4 | 0.19 | | 354 | 2.48 | 0.050 | | 6 | 0.667 | 0.0047 $10^{4}-10^{6}$ | 600 | 7.20 | 0.11 | | 167 | 11.8 | 0.15 | | 360 | 1.44 | 0.034 | | 43 | 0.267 | 0.0035 $>10^{6}$ | 284 | 1.68 | 0.020 | | 297 | 1.53 | 0.023 | | 715 | 0.297 | 0.010 | | 307 | 0.205 | 0.0041 $dT$ < 20% | 87.4% | | 89.5% | | 98.4% | | 100% $dT$ < 10% | 77.3% | | 80.7% | | 95.7% | | 100% $dT$ < 5% | 62.0% | | 68.7% | | 91.7% | | 99.4% The accuracy of calculated transition rates can be estimated either by comparisons with other theoretical works and experimental results, when available, or by the quantity $dT$, which is defined in Eq. (5) as the agreement between the values in length and velocity gauges (Froese Fischer, 2009; Ekman et al., 2014). The latter is particularly useful when no experimental measurements are available. Transition data, e.g., wavenumbers; wavelengths; line strengths; weighted oscillator strengths; transition probabilities of E1 transitions; and the accuracy indicators $dT$, are given in Tables 5 – 8, respectively, for C i – iv. Note that the wavenumbers and wavelengths are adjusted to match the level energy values in the NIST database, which are critically evaluated by Haris & Kramida (2017) for C i and Moore & Gallagher (1993) for C ii-iv. When no NIST values are available, the wavenumbers and wavelengths are from the present MCDHF/RCI calculations and marked with * in the tables. To better display the uncertainties $dT$ of the computed transitions rates and their distribution in relation to the magnitude of the transition rate values $A$, the transitions are organized in five groups based on the magnitude of the $A$ values. A statistical analysis of the uncertainties $dT$ of the transitions is performed for the 1553, 806, 1805, and 386 E1 transitions, respectively, for C i – iv. In Table 2, the mean value of the uncertainties $\langle dT\rangle$ and standard deviations $\sigma$ are given for each group of transitions. As seen in Table 2, most of the estimated uncertainties $dT$ are well below 10%. Most of the strong transitions with $A$ > $10^{6}$ s-1 are associated with small uncertainties $dT$, less than 2%, especially for C iii and C iv, for which $\langle dT\rangle$ is 0.297% ($\sigma$ = 0.01) and 0.205% ($\sigma$ = 0.0041), respectively. It is worth noting that, by employing the alternative optimization scheme of the radial orbitals in the present calculations, the uncertainties $dT$ for transitions involving high Rydberg states are significantly reduced. Contrary to the strong transitions, the weaker transitions are associated with relatively large $dT$ values. This is even more pronounced for the first two groups of transitions in C i and C ii, where $A$ is less than $10^{2}$ s-1. These weak E1 transitions are either IC or two-electron one-photon transitions. The rates of the former transitions, in relativistic calculations, are small due to the strong cancellation contributions to the transition moment (Ynnerman & Fischer, 1995), whereas the rates of the latter transitions are identically zero in the simplest approximation of the wave function and only induced by correlation effects (Bogdanovich et al., 2007; Li et al., 2010). These types of transitions are extremely challenging, and therefore interesting from a theoretical point of view, and improved methodology is needed to further decrease the uncertainties of the respective transition data. Fortunately, the weak transitions tend to be of lesser astrophysical importance, either for opacity calculations, or for spectroscopic abundance analyses. Thus, only the transitions with $A\geq$ $10^{2}$ s-1 for C i and C ii, and $A\geq$ $10^{0}$ s-1 for C iii and C iv, are discussed in the paper; although the complete transition data tables, for all computed E1 transitions in C i – iv, are available online. The scatterplots of $dT$ versus $A$ are given in Fig. 2. The mean $dT$ for all presented E1 transitions shown in Fig. 2 is 8.05% ($\sigma$ = 0.12), 7.20% ($\sigma$ = 0.13), 1.77% ($\sigma$ = 0.05), and 0.28% ($\sigma$ = 0.0059), respectively, for C i – iv. A statistical analysis of the proportions of the transitions with $dT$ less than 20%, 10%, and 5% in all the presented E1 transitions is also performed and shown in the last three rows of Table 2. Finally, the present work can be compared with other theoretical calculations. In Fig. 3, $\log gf$ values from the present work are compared with results from MCHF-BP (Fischer, 2006; Tachiev & Fischer, 2000, 1999; Fischer et al., 1998), CIV3 (Hibbert et al., 1993; Corrégé & Hibbert, 2004), and TOPbase data (Cunto & Mendoza, 1992), when available. As shown in the figure, the differences between the $\log gf$ values computed in the present work and respective results from other sources are rather small for most of the transitions. Comparing the MCDHF/RCI results with those from CIV3 calculations by Hibbert et al. (1993), which are frequently used in the abundance analyses, 292(228) out of 378 transitions are in agreement within 20% (10%) for C i, and 78(66) out of 87 transitions are within the same range for C ii. The results from the MCDHF/RCI and MCHF-BP calculations are found to be in very good agreement for C iii–iv, with the relative differences being less than 5% for all the computed transitions. More details about the comparisons with other theoretical calculations, as well as with experimental results, are given in Sects 4.1 – 4.4. Figure 2: Scatterplot of d$T$ values vs. transition rates $A$ of E1 transitions, for C i – iv. The solid lines indicate the 10% relative agreement between the length and velocity gauges. Figure 3: Differences between the calculated $\log gf$ values in this work and results from other theoretical calculations: MCHF-BP (red asterisk), CIV3 (blue plus sign), and TOPbase (black point), for C i – iv. ### 4.1 C i The computed excitation energies, given in Table 4, are compared with results from NIST (Kramida et al., 2019). With the exception of the levels belonging to the $\mathrm{2s2p^{3}}$ configuration, for which the average relative difference between theory and experiment is 1.22%, the mean relative difference for the rest of the states is 0.35%. The complete transition data, for all computed E1 transitions in C i, can be found in Table 5. Based on the statistical analysis of the uncertainties $dT$ shown in table 2, out of the 1335 transitions with $A\geq$ $10^{2}$ s-1, the proportions of the transitions with $dT$ less than 20%, 10%, and 5% are, respectively, 87.4%, 77.3%, and 62.0%. In C i, experimental transition data are available for the $\mathrm{2p3p\rightarrow 2p3s}$, $\mathrm{2p3d\rightarrow 2p3p}$, and $\mathrm{2p4s\rightarrow 2p3p}$ transition arrays using a stabilized arc source (Musielok et al., 1997; Golly et al., 2003; Bacawski et al., 2001). In Table LABEL:tab:com_exp, the experimental relative line strengths, together with their uncertainties, are compared with the present MCDHF/RCI theoretical values and with values from the non-relativistic CIV3 calculations by Hibbert et al. (1993) that included semi-empirical diagonal energy shifts by $LS$ configuration in the interaction matrix in the determination of the wavefunctions. The estimated uncertainties $dT$ of the MCDHF/RCI line strengths are given as percentages in parentheses. In most cases, the theoretical values fall into, or only slightly outside, the range of the estimated uncertainties of the experimental values. Comparing the MCDHF/RCI results with the results from the CIV3 calculations by Hibbert et al. (1993), we see that 41 out of the 50 transitions in common are in good agreement, with the relative differences being less than 10% (see Table LABEL:tab:com_exp). For the $\mathrm{2p4s~{}^{3}P^{o}\rightarrow 2p3p~{}^{3}P}$ transitions and the $\mathrm{2p4s~{}^{3}P_{2}^{o}\rightarrow 2p3p~{}^{3}D_{1}}$ transition, the $S$ values deduced from the present MCDHF/RCI calculations differ substantially from the experimental values, i.e., by more than 20%, while the values from the CIV3 calculations appear to be in better agreement with the corresponding experimental values. Based on the agreement between the length and velocity forms, the estimated uncertainties $dT$ of the present MCDHF/RCI calculations for the above- mentioned transitions are of the order of 8.5% and 1.4%, respectively. For the $\mathrm{2p3d~{}^{3}P_{2}^{o}\rightarrow 2p3p~{}^{3}P_{1}}$, $\mathrm{2p4s~{}^{3}P_{2}^{o}\rightarrow 2p3p~{}^{3}D_{2}}$, and $\mathrm{2p3d~{}^{3}D_{2}^{o}\rightarrow 2p3p~{}^{3}D_{3}}$ transitions, both theoretical results are outside the range of the estimated uncertainties of the experimental values. For the $\mathrm{2p3d~{}^{3}D^{o}\rightarrow 2p3p~{}^{3}P}$ transitions, the evaluated relative line strengths by Golly et al. (2003) slightly differ from the observations by Bacawski et al. (2001). The latter seem to be in better overall agreement with the transition rates predicted by the present calculations. In Table LABEL:tab:CI_com, the computed line strengths and transition rates are compared with values from the spline frozen-cores (FCS) method by Zatsarinny & Fischer (2002) and the MCHF-BP calculations by Fischer (2006). Zatsarinny & Fischer (2002) presented oscillator strengths for transitions from the $\mathrm{2p^{2}~{}^{3}P}$ term to high-lying excited states, while Fischer (2006) considered only transitions from $\mathrm{2p^{2}~{}^{3}P}$, $\mathrm{{}^{1}D}$, and $\mathrm{{}^{1}S}$ to odd levels up to $\mathrm{2p3d~{}^{3}P^{o}}$. As seen in the table, the present MCDHF/RCI results seem to be in better agreement with the values from spline FCS calculations. 76 out of 98 transitions from Zatsarinny & Fischer (2002) agree with present values within 10%, while only 38 out of 78 transitions from Fischer (2006) are within the same range. The relatively large differences with Fischer (2006) may be due to the fact that limited electron correlations were included in their calculations. In the MCHF-BP calculations, two types of correlation, i.e., VV, CV, have been accounted for; however, the CC correlation has not been considered. Additionally, CSF expansions obtained from SD substitutions are not as large as the CSF expansions used in the present calculations. For the majority of the strong transitions with $A$ > $10^{6}$ s-1, there is a very good agreement between the MCDHF/RCI results and the spline FCS values, with the relative difference being less than 5%. On the other hand, for the $\mathrm{2p3d~{}^{3}F\rightarrow 2p^{2}~{}^{3}P}$ and $\mathrm{2p4s~{}^{1}P_{1}\rightarrow 2p^{2}~{}^{3}P}$ transitions, the observed discrepancies between these three methods, i.e., MCDHF/RCI, spline FCS, and MCHF-BP, are quite large. These transitions are all $LS$-forbidden transitions, the former is with $\Delta L$ = 2 and the latter is spin- forbidden transition; these types of transitions are challenging for computations and are always with large uncertainties. For example, for the $\mathrm{2p3d~{}^{3}F_{3}\rightarrow 2p^{2}~{}^{3}P_{2}}$ transition, the $A$ values from MCDHF/RCI, spline FCS, and MCHF-BP calculations are, respectively, 7.92E+06, 6.24E+06, and 1.14E+07 s-1, with the relative difference between each two of them being greater than 20%. Experimental data are, therefore, needed for validating these theoretical results. On the contrary, based on the agreement between the length and velocity forms displayed in the parentheses, the estimated uncertainties of the MCDHF/RCI calculations for the above- mentioned transitions are all less than 0.5%. ### 4.2 C ii The relative differences between theory and experiment for all the energy levels of $\mathrm{2s2p^{2}}$ are 0.16%, while the mean relative difference for the rest of the states is 0.071% (see Table 4). The complete transition data, for all computed E1 transitions in C ii, can be found in Table 6. Out of the presented 592 E1 transitions with $A\geq$ $10^{2}$ s-1, the proportions of the transitions with $dT$ less than 20%, 10%, and 5% are, respectively, 89.5%, 80.7%, and 68.7%. In Table LABEL:tab:com_exp, the lifetimes from the present MCDHF/RCI calculations are compared with available results from the MCHF-BP calculations by Tachiev & Fischer (2000) and observations by Reistad et al. (1986) and Träbert et al. (1999). Träbert et al. (1999) measured lifetimes for the three fine-structure components of the $\mathrm{2s2p^{2}~{}^{4}P}$ term in an ion storage ring. For the measured lifetimes by Reistad et al. (1986) of the doublets terms using the beam-foil technique, a single value for the two fine- structure levels is provided. It can be seen that, in all cases, the MCDHF/RCI computed lifetimes agree with the experimental values by Reistad et al. (1986) within the experimental errors. For the $\mathrm{2s2p^{2}~{}^{4}P_{1/2,3/2,5/2}}$ states, as discussed in Sect. 1, the discrepancies between the measured transition rates by Fang et al. (1993) and by Träbert et al. (1999) are quite large. It is found that the MCDHF/RCI values are in better agreement with the results given by the latter measurements, with a relative difference less than 3%. For these long-lived states, the measured lifetimes are better represented by the MCDHF/RCI results than by the MCHF-BP values. The computed line strengths and transition rates are compared with values from the MCHF-BP calculations by Tachiev & Fischer (2000) and the CIV3 calculations by Corrégé & Hibbert (2004) in Table LABEL:tab:CII_com. We note that the agreement between the present MCDHF/RCI and the MCHF-BP transition rates exhibits a broad variation. In the earlier MCHF-BP and our MCDHF/RCI calculations, the same correlation effects, i.e., VV and CV, have been accounted for. However, the CSF expansions obtained from SD substitutions in the MCHF-BP calculations are not as large as the CSF expansions used in the present calculations, and as a consequence, the $LS$-composition of the configurations might not be predicted as accurately in the former calculations. The MCDHF/RCI results seem to be in better overall agreement with the values from the CIV3 calculations, except for transitions from $\mathrm{2p^{3}~{}^{2}P^{o}}$ to $\mathrm{2s2p^{2}~{}\\{^{4}P,{{}^{2}}S\\}}$ and to $\mathrm{2s^{2}3d~{}^{2}D}$. For these transitions, involving $\mathrm{2p^{3}~{}^{2}P^{o}}$ as the upper level, the transition rates $A$ are of the order of $10^{2}$ – $10^{4}$ s-1. The $dT$ values are relatively large in the present calculations. This is due to the strong cancellation effects caused by, e.g., the strong mixing between the $\mathrm{2p^{3}~{}^{2}P^{o}}$ and $\mathrm{2s2p3s~{}^{2}P^{o}}$ levels for $\mathrm{2p^{3}~{}^{2}P^{o}}\rightarrow\mathrm{2s2p^{2}~{}{{}^{2}}S}$, and the mixing between the $\mathrm{2p^{3}~{}^{2}P^{o}}$ and $\mathrm{2s^{2}4p~{}^{2}P^{o}}$ levels for $\mathrm{2p^{3}~{}^{2}P^{o}}\rightarrow\mathrm{2s^{2}3d~{}{{}^{2}}D}$. Large discrepancies are also observed between the MCDHF/RCI and MCHF-BP results, as well as between the MCHF-BP and CIV3 results for these transitions. Experimental data are, therefore, crucial for validating the aforementioned theoretical results. On the contrary, for the majority of the strong transitions with $A$ > $10^{6}$ s-1, there is a very good agreement between the MCDHF/RCI results and those from the two previous calculations, with the relative differences being less than 5%. ### 4.3 C iii The average relative discrepancy between the computed excitation energies, shown in Table 4, and the NIST recommended values is 0.041%. The complete transition data, for all computed E1 transitions in C iii, can be found in Table 7. Out of the 1668 transitions with $A\geq$ $10^{0}$ s-1, 91.7% (98.4%) of them have $dT$ values less than $5\%$ (20%). Further, the mean $dT$ for all transitions with $A\geq$ $10^{0}$ s-1 is 1.8% with $\sigma$ = 0.05. The lifetimes of the $\mathrm{2s2p~{}^{1}P^{o}_{1}}$, $\mathrm{2p^{2}~{}\\{^{1}S_{0},^{1}D_{2}\\}}$, and $\mathrm{2s3s~{}^{1}S_{0}}$ states were measured by Reistad et al. (1986) using the beam-foil technique, and the oscillator strengths for the $\mathrm{2s2p~{}^{1}P^{o}_{1}\rightarrow 2s^{2}~{}^{1}S_{0}}$ and the $\mathrm{2p^{2}~{}\\{^{1}S_{0},^{1}D_{2}\\}\rightarrow 2s2p~{}^{1}P^{o}_{1}}$ transitions were also provided. Table LABEL:tab:com_exp gives the comparisons between the observed and computed oscillator strengths and lifetimes in C iii. Looking at the table, we see an excellent agreement between the present calculations and those from the MCHF-BP calculations (Tachiev & Fischer, 1999) with the relative difference being less than 0.7%. In all cases, the computed oscillator strengths and lifetimes agree with experiment within the experimental errors. The exceptions are the oscillator strength of the $\mathrm{2p^{2}~{}^{1}S_{0}\rightarrow 2s2p~{}^{1}P^{o}_{1}}$ transition and the lifetime of the $\mathrm{2p^{2}~{}^{1}S_{0}}$ state, for which the computed values slightly differ from the observations. In Table LABEL:tab:CIII_com, the computed line strengths and transition rates are compared with values from the MCHF-BP calculations by Tachiev & Fischer (1999) and the Grasp calculations by Aggarwal & Keenan (2015). For the majority of the transitions, there is an excellent agreement between the MCDHF/RCI and MCHF-BP values with the relative differences being less than 1%. Only 4 out of 60 transitions display discrepancies that are greater than 20%. These large discrepancies are observed for the IC transitions, e.g., $\mathrm{2s3d~{}^{3}D_{2}\rightarrow 2s2p~{}^{1}P^{o}_{1}}$ and $\mathrm{2s3d~{}^{3}D_{2}\rightarrow 2s2p~{}^{1}P^{o}_{1}}$, for which the $dT$ is relatively large. The discrepancies between the MCDHF/RCI and Grasp values are overall large; this is due to the fact that limited electron correlations were included in their calculations. Based on the excellent agreement between the MCDHF/RCI and MCHF-BP results as well as with experiment, we believe that the present transition rates together with the MCHF-BP transition data are more reliable than the ones provided by Aggarwal & Keenan (2015). ### 4.4 C iv The mean relative discrepancy between the computed excitation energies, given in Table 4, and the NIST values is 0.0044%. Out of the presented 366 transitions with $A\geq$ $10^{0}$ s-1 shown in Table 8, only two of them have $dT$ values greater than $5\%$; 94.0% of them with $dT$ being less than 1%. The mean $dT$ for all transitions with $A\geq$ $10^{0}$ s-1 is $0.28\%$ with $\sigma$ = 0.0059. For C iv, there are a number of measurements of transition properties. The transition rates of the $\mathrm{2p~{}^{2}P^{o}_{1/2,3/2}\rightarrow 2s~{}^{2}S_{1/2}}$ transitions were measured by Knystautas et al. (1971) using the beam-foil technique. By using the same technique, the lifetimes for a number of excited states were measured in four different experimental works (Donnelly et al., 1978; Buchet-Poulizac & Buchet, 1973b; Jacques et al., 1980; Peach et al., 1988). In Table LABEL:tab:com_exp, we compare the theoretical results, from present calculations and MCHF-BP calculations, with the NIST recommended values and observed values. The transition rates of the $\mathrm{2p~{}^{2}P^{o}_{1/2,3/2}\rightarrow 2s~{}^{2}S_{1/2}}$ transitions from the present work agree perfectly with the values from the MCHF-BP calculations by Fischer et al. (1998), while they are slightly smaller than the NIST data and the values by Knystautas et al. (1971). A comparison of the lifetimes of the $\mathrm{\\{3s,4s,2p,3p,4p,3d,4d,5d\\}}$ states is made with other theoretical results, i.e., from the MCHF-BP calculations and the Model Potential method. The agreements between these different theoretical results are better than 1% for all these states. Furthermore, the agreement between the computed values and those from observations is also very good except for the $\mathrm{3s~{}^{2}S_{1/2}}$ level, for which the MCDHF/RCI calculations give a slightly smaller lifetime of 0.2350 ns than the observed value of 0.25 $\pm$ 0.01 ns. In Table LABEL:tab:CIV_com, the computed line strengths and transition rates are compared with available values from the MCHF-BP calculations by Fischer et al. (1998). There is an excellent agreement between the two methods with the relative differences being less than 1% for all transitions. ## 5 Reanalysis of the solar carbon abundance Table 3: The $14$ permitted C i lines used as abundance diagnostics in Amarsi et al. (2019). Shown are the upper and lower configurations, oscillator strengths obtained from the present calculations, and oscillator strengths from NIST; the latter being based on the calculations from CIV3 (Hibbert et al., 1993). The estimated uncertainties $dT$ of the oscillator strengths are given as percentages in parentheses. The final two columns show the abundances derived in Amarsi et al. (2019), and the post-corrected values derived here based on the formula $\Delta\log\mathrm{\upvarepsilon_{C}}^{\text{line}}=-\Delta\log gf^{\text{line}}$. | | | $\log gf$ | | | ---|---|---|---|---|---|--- Upper | Lower | $\lambda_{\text{air}}$(nm) | NIST | | MCDHF/RCI($dT$) | | $\log\upvarepsilon_{\mathrm{C}}^{\text{A19}}$ | $\log\upvarepsilon_{\mathrm{C}}^{\text{L20}}$ $\mathrm{2p4p~{}^{1}D_{2}}$ | $\mathrm{2p3s~{}^{1}P_{1}^{o}}$ | 505.217 | -1.30 | | -1.36(0.8%) | | 8.41 | 8.47 $\mathrm{2p4p~{}^{1}P_{1}}$ | $\mathrm{2p3s~{}^{1}P_{1}^{o}}$ | 538.034 | -1.62 | | -1.71(1.4%) | | 8.43 | 8.52 $\mathrm{2p4d~{}^{1}P_{1}^{o}}$ | $\mathrm{2p3p~{}^{1}P_{1}}$ | 658.761 | -1.00 | | -1.05(0.2%) | | 8.33 | 8.38 $\mathrm{2p4d~{}^{3}F_{2}^{o}}$ | $\mathrm{2p3p~{}^{3}D_{1}}$ | 711.148 | -1.08 | | -1.24(0.9%) | | 8.31 | 8.47 $\mathrm{2p4d~{}^{3}F_{4}^{o}}$ | $\mathrm{2p3p~{}^{3}D_{3}}$ | 711.318 | -0.77 | | -0.94(1.5%) | | 8.41 | 8.58 $\mathrm{2p3p~{}^{3}D_{1}}$ | $\mathrm{2p3s~{}^{3}P_{2}^{o}}$ | 1075.40 | -1.61 | | -1.62(1.3%) | | 8.49 | 8.50 $\mathrm{2p3d~{}^{3}F_{2}^{o}}$ | $\mathrm{2p3p~{}^{3}D_{2}}$ | 1177.75 | -0.52 | | -0.46(0.9%) | | 8.46 | 8.40 $\mathrm{2p3d~{}^{3}P_{1}^{o}}$ | $\mathrm{2p3p~{}^{3}P_{0}}$ | 1254.95 | -0.57 | | -0.65(3.3%) | | 8.51 | 8.60 $\mathrm{2p3d~{}^{3}P_{0}^{o}}$ | $\mathrm{2p3p~{}^{3}P_{1}}$ | 1256.21 | -0.52 | | -0.61(3.3%) | | 8.51 | 8.60 $\mathrm{2p3d~{}^{3}P_{1}^{o}}$ | $\mathrm{2p3p~{}^{3}P_{1}}$ | 1256.90 | -0.60 | | -0.70(3.2%) | | 8.46 | 8.56 $\mathrm{2p3d~{}^{3}P_{2}^{o}}$ | $\mathrm{2p3p~{}^{3}P_{1}}$ | 1258.16 | -0.54 | | -0.61(3.4%) | | 8.46 | 8.53 $\mathrm{2p3d~{}^{1}P_{1}^{o}}$ | $\mathrm{2p3p~{}^{1}S_{0}}$ | 2102.31 | -0.40 | | -0.39(0.5%) | | 8.47 | 8.46 $\mathrm{2p4p~{}^{1}D_{2}}$ | $\mathrm{2p3d~{}^{1}F_{3}^{o}}$ | 3085.46 | +0.10 | | +0.07(0.2%) | | 8.41 | 8.44 $\mathrm{2p4d~{}^{1}D_{2}^{o}}$ | $\mathrm{2p4p~{}^{1}P_{1}}$ | 3406.58 | +0.44 | | +0.45(3.1%) | | 8.47 | 8.46 Figure 4: Inferred solar carbon abundances. Black points (A19) are the 3D non- LTE results of Amarsi et al. (2019) for $14$ permitted C i lines. Blue points (L20) are these same results but post-corrected using the new $\log gf$ data. Error bars reflect $\pm 5\%$ uncertainties in the measured equivalent widths as stipulated by those authors. The four lines between $1254\,\mathrm{nm}$ and $1259\,\mathrm{nm}$ discussed in the text have been highlighted in red. The unweighted means $\mu$ (including all $14$ lines) and the standard deviations of the samples $\sigma$ are stated in each panel. One can also attempt to verify the present atomic data empirically, in an astrophysical context. To demonstrate this, a solar carbon abundance analysis was carried out, based on permitted C i lines. Larger errors in the atomic data usually impart a larger dispersion in the line-by-line abundance results, as well as trends in the results with respect to the line parameters. The solar carbon abundance analysis recently presented in Amarsi et al. (2019) was taken as the starting point. Their analysis is based on equivalent widths measured in the solar disk-centre intensity, for $14$ permitted C i lines in the optical and near-infrared, as well as a single forbidden [C i] line at $872.7\,\mathrm{nm}$. Their analysis draws on a three-dimensional (3D) hydrodynamic model solar atmosphere and 3D non-local thermodynamic equilibrium (non-LTE) radiative transfer, that reflects the current state-of-the-art in stellar elemental abundance determinations (e.g. Asplund et al., 2009). For the $14$ permitted C i lines, the authors adopted transition probabilities from NIST, that are based on those of Hibbert et al. (1993) but normalized to a different scale (Haris & Kramida, 2017), corresponding to differences of the order $\pm 0.01\,\mathrm{dex}{}$. Here, we post-correct the solar carbon abundances inferred in Amarsi et al. (2019) from the $14$ permitted C i lines, using the new atomic data derived in the present study (see Table 3). To first-order, for a given spectral line, the change in the inferred abundances are related to the difference in the adopted transition probabilities simply as $\Delta\log\upvarepsilon_{\mathrm{C}}^{\text{line}}=-\Delta\log gf^{\text{line}}$. We briefly note that second-order effects on the inferred abundances, propagated forward from changes to the non-LTE statistical equilibrium when adopting the full set of new $\log gf$ data in the non-LTE model atom, were also tested; these were found to be negligible. The results of this post-correction are illustrated in Fig. 4. We find that the dispersion in the line-by-line abundance results are similar when using the new and the old sets of $\log gf$ data. We also find that the trends in the results with respect to the line parameters are of similar gradients. This is consistent with the finding in Sect. 4.1, that the precision of this new, much larger atomic data set is comparable to that of Hibbert et al. (1993). This new analysis implies a solar carbon abundance of $8.50\,\mathrm{dex}{}$, which is $0.06\,\mathrm{dex}{}$ larger than that inferred in Amarsi et al. (2019) from C i lines, and $0.07\,\mathrm{dex}{}$ larger than the current standard value from Asplund et al. (2009) that is based on C i lines as well as on molecular diagnostics. This increase in the mean abundance is due to $12$ of the $14$ permitted C i lines having lower oscillator strengths in the present calculations, compared to the NIST data set. Six of the lines give results that are larger than the mean ($\log\upvarepsilon\geq 8.51$); included in this set are all four of the lines between $1254\,\mathrm{nm}$ and $1259\,\mathrm{nm}$, which give rise to values of between $8.53$ and $8.60\,\mathrm{dex}$. These four lines have the same upper level configuration, $\mathrm{2p3d\,^{3}P^{o}}$, and a closer inspection of the $LS$-composition reveals that these states are strongly mixed (of the order of $26$%) with $\mathrm{2s2p^{3}\,{}^{3}P^{o}}$ states, which are less accurately described in the present calculations. As a consequence, as shown in Table 3, these transitions appear to be associated with slightly larger uncertainties $dT$ than most of the other lines. Omitting these four lines, or adopting NIST oscillator strengths for them, would reduce the mean abundance from $8.50$ to $8.47\,\mathrm{dex}$. Given that the scatter and trends in the results do not support one set of data over the other, we refrain from advocating a higher solar carbon abundance at this point. Nevertheless, this quite drastic change in the resulting solar carbon abundance highlights the importance of having accurate atomic data for abundance analyses. This is especially relevant in the context of the solar modelling problem, wherein standard models of the solar interior, adopting the solar chemical composition of Asplund et al. (2009), fail to reproduce key empirical constraints, including the depth of the convection zone and interior sound speed that are precisely inferred from helioseismic observations (Basu & Antia, 2008; Zhang et al., 2019). Extra opacity in the solar interior near the boundary of the convection zone would resolve the problem (Bailey et al., 2015). Carbon contributes about $5\%$ of the opacity in this region (Blancard et al., 2012), so a higher carbon abundance would help alleviate the problem, albeit only very slightly. ## 6 Conclusions In the present work, energy levels and transition data of E1 transitions are computed for C i – iv using the MCDHF and RCI methods. Special attention is paid to the computation of transition data involving high Rydberg states by employing an alternative orbital optimization approach. The accuracy of the predicted excitation energies is evaluated by comparing with experimental data provided by the NIST database. The average relative differences of the computed energy levels compared with the NIST data are 0.41%, 0.081%, 0.041%, and 0.0044%, respectively, for C i – iv. The accuracy of the transition data is evaluated based on the relative differences of the computed transition rates in the length and velocity gauges, which is given by the quantity $dT$, and by extensive comparisons with previous theoretical and experimental results. For most of the strong transitions in C i – iv, the $dT$ values are less than 5%. The mean $dT$ for all presented E1 transitions are 8.05% ($\sigma$ = 0.12), 7.20% ($\sigma$ = 0.13), 1.77% ($\sigma$ = 0.050), and 0.28% ($\sigma$ = 0.0059), respectively, for C i – iv. Particularly, for strong transitions with $A>$ $10^{6}$ s-1, the mean $dT$ is 1.68% ($\sigma$ = 0.020), 1.53% ($\sigma$ = 0.023), 0.297% ($\sigma$ = 0.010), and 0.205% ($\sigma$ = 0.0041), respectively, for C i – iv. By employing alternative optimization schemes of the radial orbitals, the uncertainties $dT$ of the computed transition data for transitions involving high Rydberg states are significantly reduced. The agreement between computed transition properties, e.g., line strengths, transition rates, and lifetimes, and experimental values is overall good. The exception is the weak transitions, e.g., the IC transitions, for which the strong cancellation effects are important; however, these effects cannot be properly considered in the present calculations. The present calculations are extended to high Rydberg states that are not covered by previous accurate calculations and this is of special importance in various astrophysical applications. The accurate and extensive sets of atomic data for C i – iv are publicly available for use by the astronomy community. These data should be useful for opacity calculations and for models of stellar structures and interiors. They should also be useful to non-LTE spectroscopic analyses of both early-type and late-type stars. ## Acknowledgements This work is supported by the Swedish research council under contracts 2015-04842, 2016-04185, 2016-03765, and 2020-03940, and by the Knut and Alice Wallenberg Foundation under the project grant KAW 2013.0052. Some of the computations were enabled by resources provided by the Swedish National Infrastructure for Computing (SNIC) at the Multidisciplinary Center for Advanced Computational Science (UPPMAX) and at the High Performance Computing Center North (HPC2N) partially funded by the Swedish Research Council through grant agreement no. 2018-05973. This work was also supported by computational resources provided by the Australian Government through the National Computational Infrastructure (NCI) under the National Computational Merit Allocation Scheme (NCMAS), under project y89. We thank Nicolas Grevesse for insightful comments on an earlier version of this manuscript. We would also like to thank the anonymous referee for her/his useful comments that helped improve the original manuscript. ## Data Availability The full tables of energy levels (Table 4) and transition data (Tables 5 – 8) are available in the online Supplementary Material. ## References * Aggarwal & Keenan (2015) Aggarwal K. M., Keenan F. P., 2015, Monthly Notices of the Royal Astronomical Society, 450, 1151 * Alexeeva et al. (2019) Alexeeva S., Sadakane K., Nishimura M., Du J., Hu S., 2019, ApJ, 884, 150 * Amarsi et al. (2019) Amarsi A. M., Barklem P. S., Collet R., Grevesse N., Asplund M., 2019, A&A, 624, A111 * Asplund et al. (2009) Asplund M., Grevesse N., Sauval A. J., Scott P., 2009, ARA&A, 47, 481 * Bacawski et al. (2001) Bacawski A., Wujec T., Musielok J., 2001, Physica Scripta, 64, 314 * Bailey et al. (2015) Bailey J. E., et al., 2015, Nature, 517, 56 * Basu & Antia (2008) Basu S., Antia H. M., 2008, Phys. Rep., 457, 217 * Berkner et al. (1965) Berkner K., Cooper W., Kaplan S., Pyle R., 1965, Physics Letters, 16, 35 * Blancard et al. (2012) Blancard C., Cossé P., Faussurier G., 2012, ApJ, 745, 10 * Bogdanovich et al. (2007) Bogdanovich P., Karpuškiene R., Rancova O., 2007, Physica Scripta, 75, 669 * Boldt (1963) Boldt G., 1963, Zeitschrift Naturforschung Teil A, 18, 1107 * Buchet-Poulizac & Buchet (1973a) Buchet-Poulizac M. C., Buchet J. P., 1973a, Physica Scripta, 8, 40 * Buchet-Poulizac & Buchet (1973b) Buchet-Poulizac M. C., Buchet J. P., 1973b, Physica Scripta, 8, 40 * Caffau et al. (2010) Caffau E., Ludwig H. G., Bonifacio P., Faraggiana R., Steffen M., Freytag B., Kamp I., Ayres T. R., 2010, A&A, 514, A92 * Chen et al. (2020) Chen X., et al., 2020, ApJ, 889, 157 * Corrégé & Hibbert (2004) Corrégé G., Hibbert A., 2004, Atomic Data and Nuclear Data Tables, 86, 19 * Cunto & Mendoza (1992) Cunto W., Mendoza C., 1992, Rev. Mex. Astron. Astrofis., 23, 107 * Cunto et al. (1993) Cunto W., Mendoza C., Ochsenbein F., Zeippen C. J., 1993, A&A, 275, L5 * Doerfert et al. (1997) Doerfert J., Träbert E., Wolf A., Schwalm D., Uwira O., 1997, Phys. Rev. Lett., 78, 4355 * Donnelly et al. (1978) Donnelly K. E., Kernahan J. A., Pinnington E. H., 1978, J. Opt. Soc. Am., 68, 1000 * Dyall et al. (1989) Dyall K., Grant I., Johnson C., Parpia F., Plummer E., 1989, Comput. Phys. Commun., 55, 425 * Ekman et al. (2014) Ekman J., Godefroid M., Hartman H., 2014, Atoms, 2, 215 * Fang et al. (1993) Fang Z., Kwong V. H. S., Wang J., Parkinson W. H., 1993, Phys. Rev. A, 48, 1114 * Federman & Zsargo (2001) Federman S. R., Zsargo J., 2001, The Astrophysical Journal, 555, 1020 * Fischer (1994) Fischer C. F., 1994, Physica Scripta, 49, 323 * Fischer (2000) Fischer C. F., 2000, Physica Scripta, 62, 458 * Fischer (2006) Fischer C. F., 2006, Journal of Physics B: Atomic, Molecular and Optical Physics, 39, 2159 * Fischer & Tachiev (2004) Fischer C. F., Tachiev G., 2004, Atomic Data and Nuclear Data Tables, 87, 1 * Fischer et al. (1998) Fischer C., Saparov M., Gaigalas G., Godefroid M., 1998, Atomic Data and Nuclear Data Tables, 70, 119 * Fischer et al. (2016) Fischer C. F., Godefroid M., Brage T., Jönsson P., Gaigalas G., 2016, J. Phys. B: At. Mol. Opt. Phys, 49, 182004 * Fischer et al. (2019) Fischer C. F., Gaigalas G., Jönsson P., Bieroń J., 2019, Comput. Phys. Commun., 237, 184 * Fleming et al. (1994) Fleming J., Hibbert A., Stafford R. P., 1994, Physica Scripta, 49, 316 * Foster (1962) Foster E. W., 1962, Proceedings of the Physical Society, 79, 94 * Franchini et al. (2020) Franchini M., et al., 2020, ApJ, 888, 55 * Froese Fischer (2009) Froese Fischer C., 2009, Physica Scripta Volume T, 134, 014019 * Fuhr (2006) Fuhr J. R., 2006, Journal of the American Chemical Society, 128, 5585 * Gaigalas et al. (2017) Gaigalas G., Fischer C. F., Rynkun P., Jönsson P., 2017, Atoms, 5 * Godefroid et al. (2001) Godefroid M., Fischer C. F., Jönsson P., 2001, Journal of Physics B: Atomic, Molecular and Optical Physics, 34, 1079 * Goldbach & Nollez (1987) Goldbach C., Nollez G., 1987, A&A, 181, 203 * Goldbach et al. (1989) Goldbach C., Martin M., Nollez G., 1989, A&A, 221, 155 * Golly et al. (2003) Golly A., Jazgara A., Wujec T., 2003, Physica Scripta, 67, 485 * Goly & Weniger (1982) Goly A., Weniger S., 1982, J. Quant. Spectrosc. Radiative Transfer, 28, 389 * Grant (1974) Grant I. P., 1974, J. Phys. B: At. Mol. Opt. Phys, 7, 1458 * Grant (2007) Grant I. P., 2007, Relativistic Quantum Theory of Atoms and Molecules. Springer, New York * Haris & Kramida (2017) Haris K., Kramida A., 2017, The Astrophysical Journal Supplement Series, 233, 16 * Hibbert (1974) Hibbert A., 1974, Journal of Physics B: Atomic and Molecular Physics, 7, 1417 * Hibbert (1975) Hibbert A., 1975, Computer Physics Communications, 9, 141 * Hibbert et al. (1993) Hibbert A., Biemont E., Godefroid M., Vaeck N., 1993, A&AS, 99, 179 * Jacques et al. (1980) Jacques C., Knystautas E. J., Drouin R., Berry H. G., 1980, Canadian Journal of Physics, 58, 1093 * Jofré et al. (2020) Jofré P., Jackson H., Tucci Maia M., 2020, A&A, 633, L9 * Jones & Wiese (1984) Jones D. W., Wiese W. L., 1984, Phys. Rev. A, 29, 2597 * Jönsson et al. (2013) Jönsson P., Gaigalas G., Bieroń J., Fischer C. F., Grant I., 2013, Comput. Phys. Commun., 184, 2197 * Jönsson et al. (2010) Jönsson P., Li J., Gaigalas G., Dong C., 2010, Atomic Data and Nuclear Data Tables, 96, 271 * Kingston & Hibbert (2000) Kingston A. E., Hibbert A., 2000, Journal of Physics B: Atomic, Molecular and Optical Physics, 33, 693 * Knystautas et al. (1971) Knystautas E., Barrette L., Neveu B., Drouin R., 1971, Journal of Quantitative Spectroscopy and Radiative Transfer, 11, 75 * Kramida et al. (2019) Kramida A., Yu. Ralchenko Reader J., and NIST ASD Team 2019, NIST Atomic Spectra Database (ver. 5.7.1), [Online]. Available: https://physics.nist.gov/asd [2020, February 11]. National Institute of Standards and Technology, Gaithersburg, MD. * Kwong et al. (1993) Kwong V. H. S., Fang Z., Gibbons T. T., Parkinson W. H., Smith P. L., 1993, ApJ, 411, 431 * Li et al. (2010) Li J., Jönsson P., Dong C., Gaigalas G., 2010, Journal of Physics B: Atomic, Molecular and Optical Physics, 43, 035005 * Maecker (1953) Maecker H., 1953, Zeitschrift fur Physik, 135, 13 * Mickey (1970) Mickey D., 1970, Nuclear Instruments and Methods, 90, 77 * Miller et al. (1974) Miller M. H., Wilkerson T. D., Roig R. A., Bengtson R. D., 1974, Phys. Rev. A, 9, 2312 * Moore & Gallagher (1993) Moore C. E., Gallagher J. W., 1993, Tables of spectra of hydrogen, carbon, nitrogen, and oxygen atoms and ions. Boca Raton : CRC Press * Musielok et al. (1997) Musielok J., Veres G., Wiese W., 1997, Journal of Quantitative Spectroscopy and Radiative Transfer, 57, 395 * Nandi et al. (1996) Nandi T., Bhattacharya N., Kurup M. B., Prasad K. G., 1996, Physica Scripta, 54, 179 * Nieva & Przybilla (2006) Nieva M. F., Przybilla N., 2006, ApJ, 639, L39 * Nieva & Przybilla (2008) Nieva M. F., Przybilla N., 2008, A&A, 481, 199 * Nieva & Przybilla (2012) Nieva M. F., Przybilla N., 2012, A&A, 539, A143 * Nussbaumer & Storey (1981) Nussbaumer H., Storey P. J., 1981, A&A, 96, 91 * Nussbaumer & Storey (1984) Nussbaumer H., Storey P. J., 1984, A&A, 140, 383 * Olsen et al. (1988) Olsen J., Roos B. O., Jørgensen P., Jensen H. J. A., 1988, J. Chem. Phys., 89, 2185 * Papoulia et al. (2019) Papoulia A., et al., 2019, Atoms, 7 * Peach et al. (1988) Peach G., Saraph H. E., Seaton M. J., 1988, Journal of Physics B: Atomic, Molecular and Optical Physics, 21, 3669 * Pehlivan Rhodin et al. (2017) Pehlivan Rhodin A., Hartman H., Nilsson H., Jönsson P., 2017, A&A, 598, A102 * Przybilla et al. (2001) Przybilla N., Butler K., Kudritzki R. P., 2001, A&A, 379, 936 * Reistad & Martinson (1986) Reistad N., Martinson I., 1986, Phys. Rev. A, 34, 2632 * Reistad et al. (1986) Reistad N., Hutton R., Nilsson A. E., Martinson I., Mannervik S., 1986, Physica Scripta, 34, 151 * Richter (1958) Richter J., 1958, Zeitschrift fur Physik, 151, 114 * Roberts & Eckerle (1967) Roberts J. R., Eckerle K. L., 1967, Phys. Rev., 153, 87 * Stonkutė et al. (2020) Stonkutė E., et al., 2020, AJ, 159, 90 * Sturesson et al. (2007) Sturesson L., Jönsson P., Froese Fischer C., 2007, CoPhC, 177, 539 * Tachiev & Fischer (1999) Tachiev G., Fischer C. F., 1999, Journal of Physics B: Atomic, Molecular and Optical Physics, 32, 5805 * Tachiev & Fischer (2000) Tachiev G., Fischer C. F., 2000, Journal of Physics B: Atomic, Molecular and Optical Physics, 33, 2419 * Tachiev & Fischer (2001) Tachiev G., Fischer C. F., 2001, Canadian Journal of Physics, 79, 955 * Träbert et al. (1999) Träbert E., Gwinner G., Knystautas E. J., Tordoir X., Wolf A., 1999, Journal of Physics B: Atomic, Molecular and Optical Physics, 32, L491 * VandenBerg et al. (2012) VandenBerg D. A., Bergbusch P. A., Dotter A., Ferguson J. W., Michaud G., Richer J., Proffitt C. R., 2012, ApJ, 755, 15 * Wiese & Fuhr (2007a) Wiese W. L., Fuhr J. R., 2007a, Journal of Physical and Chemical Reference Data, 36, 1287 * Wiese & Fuhr (2007b) Wiese W. L., Fuhr J. R., 2007b, Journal of Physical and Chemical Reference Data, 36, 1737 * Ynnerman & Fischer (1995) Ynnerman A., Fischer C. F., 1995, Phys. Rev. A, 51, 2020 * Zatsarinny & Fischer (2002) Zatsarinny O., Fischer C. F., 2002, Journal of Physics B: Atomic, Molecular and Optical Physics, 35, 4669 * Zhang et al. (2019) Zhang Q.-S., Li Y., Christensen-Dalsgaard J., 2019, ApJ, 881, 103 ## Appendix A Additional tables Table 4: Wave function composition (up to three $LS$ components with a contribution $>$ 0.02 of the total wave function) in $LS$-coupling, energy levels (in cm-1), and lifetimes (in s; given in length ($\tau_{l}$) and velocity ($\tau_{v}$) gauges) for C i – iv. Energy levels are given relative to the ground state and compared with NIST data (Kramida et al., 2019). The full table is available online. Species | No. | State | $LS$-composition | $E_{RCI}$ | $E_{NIST}$ | $\tau_{l}$ | $\tau_{v}$ ---|---|---|---|---|---|---|--- C I | 1 | $\mathrm{2s^{2}2p^{2}(^{3}_{2}P)~{}^{3}P_{0}}$ | 0.88 + 0.03 $\mathrm{2s^{2}2p~{}^{2}P\,7p~{}^{3}P}$ | 0 | 0 | | C I | 2 | $\mathrm{2s^{2}2p^{2}(^{3}_{2}P)~{}^{3}P_{1}}$ | 0.88 + 0.03 $\mathrm{2s^{2}2p~{}^{2}P\,7p~{}^{3}P}$ | 16 | 16 | | C I | 3 | $\mathrm{2s^{2}2p^{2}(^{3}_{2}P)~{}^{3}P_{2}}$ | 0.88 + 0.03 $\mathrm{2s^{2}2p~{}^{2}P\,7p~{}^{3}P}$ | 43 | 43 | | C I | 4 | $\mathrm{2s^{2}2p^{2}(^{1}_{2}D)~{}^{1}D_{2}}$ | 0.85 + 0.05 $\mathrm{2s^{2}2p~{}^{2}P\,7p~{}^{1}D}$ \+ 0.03 $\mathrm{2s^{2}2p~{}^{2}P\,3p~{}^{1}D}$ | 10 275 | 10 193 | | C I | 5 | $\mathrm{2s^{2}2p^{2}(^{1}_{0}S)~{}^{1}S_{0}}$ | 0.78 + 0.06 $\mathrm{2s^{2}2p~{}^{2}P\,7p~{}^{1}S}$ \+ 0.06 $\mathrm{2p^{4}(^{1}_{0}S)~{}^{1}S}$ | 21 775 | 21 648 | | C I | 6 | $\mathrm{2s~{}^{2}S\,2p^{3}(^{4}_{3}S)~{}^{5}S_{2}^{\circ}}$ | 0.93 + 0.04 $\mathrm{2s~{}^{2}S\,2p^{2}(^{3}_{2}P)~{}^{4}P\,7p~{}^{5}S^{\circ}}$ | 33 859 | 33 735 | 3.00E-02 | 1.26E-02 C I | 7 | $\mathrm{2s^{2}2p~{}^{2}P\,3s~{}^{3}P_{0}^{\circ}}$ | 0.91 + 0.04 $\mathrm{2p^{3}(^{2}_{1}P)~{}^{2}P\,3s~{}^{3}P^{\circ}}$ | 60 114 | 60 333 | 3.00E-09 | 3.04E-09 C I | 8 | $\mathrm{2s^{2}2p~{}^{2}P\,3s~{}^{3}P_{1}^{\circ}}$ | 0.91 + 0.04 $\mathrm{2p^{3}(^{2}_{1}P)~{}^{2}P\,3s~{}^{3}P^{\circ}}$ | 60 133 | 60 353 | 3.00E-09 | 3.04E-09 C I | 9 | $\mathrm{2s^{2}2p~{}^{2}P\,3s~{}^{3}P_{2}^{\circ}}$ | 0.91 + 0.04 $\mathrm{2p^{3}(^{2}_{1}P)~{}^{2}P\,3s~{}^{3}P^{\circ}}$ | 60 174 | 60 393 | 3.00E-09 | 3.04E-09 C I | 10 | $\mathrm{2s^{2}2p~{}^{2}P\,3s~{}^{1}P_{1}^{\circ}}$ | 0.92 + 0.04 $\mathrm{2p^{3}(^{2}_{1}P)~{}^{2}P\,3s~{}^{1}P^{\circ}}$ | 61 750 | 61 982 | 2.78E-09 | 2.83E-09 – | – | – | – | – | – | – | – Table 5: Electric dipole transition data for C i from present calculations. Upper and lower states, wavenumber, $\Delta E$, wavelength, $\lambda$, line strength, $S$, weighted oscillator strength, $gf$, transition probability, $A$, together with the relative difference between two gauges of $A$ values, $dT$, provided by the present MCDHF/RCI calculations are shown in the table. Wavelength and wavenumber values are from the NIST database (Kramida et al., 2019) when available. Wavelengths and wavenumbers marked with * are from the present calculations. Only the first ten rows are shown; the full table is available online. Upper | Lower | $\Delta E$(cm-1) | $\lambda$ (Å) | $S$ (a.u. of a${}_{0}^{2}$e2) | $gf$ | $A$ (s-1) | $dT$ ---|---|---|---|---|---|---|--- $\mathrm{2s^{2}2p5d~{}^{3}D_{2}^{o}}$ | $\mathrm{2s^{2}2p^{2}~{}^{3}P_{1}}$ | 86373 | 1157.769 | 1.025E-01 | 2.679E-02 | 2.647E+07 | 0.004 $\mathrm{2s^{2}2p5d~{}^{3}D_{1}^{o}}$ | $\mathrm{2s^{2}2p^{2}~{}^{3}P_{0}}$ | 86362 | 1157.909 | 8.568E-02 | 2.240E-02 | 3.689E+07 | 0.003 $\mathrm{2s^{2}2p5d~{}^{3}D_{3}^{o}}$ | $\mathrm{2s^{2}2p^{2}~{}^{3}P_{2}}$ | 86354 | 1158.018 | 3.197E-01 | 8.358E-02 | 5.897E+07 | 0.003 $\mathrm{2s^{2}2p6s~{}^{3}P_{2}^{o}}$ | $\mathrm{2s^{2}2p^{2}~{}^{3}P_{1}}$ | 86352 | 1158.038 | 1.273E-01 | 3.328E-02 | 3.287E+07 | 0.002 $\mathrm{2s^{2}2p5d~{}^{3}D_{1}^{o}}$ | $\mathrm{2s^{2}2p^{2}~{}^{3}P_{1}}$ | 86346 | 1158.130 | 4.687E-02 | 1.225E-02 | 2.017E+07 | 0.002 $\mathrm{2s^{2}2p5d~{}^{3}D_{2}^{o}}$ | $\mathrm{2s^{2}2p^{2}~{}^{3}P_{2}}$ | 86346 | 1158.131 | 1.049E-01 | 2.742E-02 | 2.708E+07 | 0.000 $\mathrm{2s^{2}2p6s~{}^{3}P_{1}^{o}}$ | $\mathrm{2s^{2}2p^{2}~{}^{3}P_{0}}$ | 86331 | 1158.324 | 2.442E-02 | 6.380E-03 | 1.050E+07 | 0.001 $\mathrm{2s^{2}2p6s~{}^{3}P_{2}^{o}}$ | $\mathrm{2s^{2}2p^{2}~{}^{3}P_{2}}$ | 86325 | 1158.400 | 2.619E-02 | 6.844E-03 | 6.756E+06 | 0.001 $\mathrm{2s^{2}2p5d~{}^{3}D_{1}^{o}}$ | $\mathrm{2s^{2}2p^{2}~{}^{3}P_{2}}$ | 86319 | 1158.492 | 1.367E-03 | 3.571E-04 | 5.875E+05 | 0.002 $\mathrm{2s^{2}2p6s~{}^{3}P_{1}^{o}}$ | $\mathrm{2s^{2}2p^{2}~{}^{3}P_{1}}$ | 86315 | 1158.544 | 6.630E-03 | 1.732E-03 | 2.849E+06 | 0.004 – | – | – | – | – | – | – | – Table 6: Electric dipole transition data for C ii from present calculations. Upper and lower states, wavenumber, $\Delta E$, wavelength, $\lambda$, line strength, $S$, weighted oscillator strength, $gf$, transition probability, $A$, together with the relative difference between two gauges of $A$ values, $dT$, provided by the present MCDHF/RCI calculations are shown in the table. Wavelength and wavenumber values are from the NIST database (Kramida et al., 2019) when available. Only the first ten rows are shown; the full table is available online. Upper | Lower | $\Delta E$(cm-1) | $\lambda$ (Å) | $S$ (a.u. of a${}_{0}^{2}$e2) | $gf$ | $A$ (s-1) | $dT$ ---|---|---|---|---|---|---|--- $\mathrm{2s2p3p~{}^{2}D_{3/2}}$ | $\mathrm{2s^{2}2p~{}^{2}P_{1/2}^{o}}$ | 188581 | 530.275 | 8.159E-02 | 4.673E-02 | 2.771E+08 | 0.015 $\mathrm{2s2p3p~{}^{2}D_{5/2}}$ | $\mathrm{2s^{2}2p~{}^{2}P_{3/2}^{o}}$ | 188551 | 530.359 | 1.515E-01 | 8.678E-02 | 3.430E+08 | 0.015 $\mathrm{2s2p3p~{}^{2}D_{3/2}}$ | $\mathrm{2s^{2}2p~{}^{2}P_{3/2}^{o}}$ | 188517 | 530.454 | 1.661E-02 | 9.511E-03 | 5.636E+07 | 0.015 $\mathrm{2s^{2}7d~{}^{2}D_{3/2}}$ | $\mathrm{2s^{2}2p~{}^{2}P_{1/2}^{o}}$ | 187353 | 533.752 | 1.094E-01 | 6.223E-02 | 3.637E+08 | 0.007 $\mathrm{2s^{2}7d~{}^{2}D_{5/2}}$ | $\mathrm{2s^{2}2p~{}^{2}P_{3/2}^{o}}$ | 187289 | 533.933 | 1.943E-01 | 1.104E-01 | 4.300E+08 | 0.007 $\mathrm{2s^{2}7d~{}^{2}D_{3/2}}$ | $\mathrm{2s^{2}2p~{}^{2}P_{3/2}^{o}}$ | 187289 | 533.933 | 2.205E-02 | 1.254E-02 | 7.321E+07 | 0.007 $\mathrm{2s2p3p~{}^{4}P_{3/2}}$ | $\mathrm{2s^{2}2p~{}^{2}P_{1/2}^{o}}$ | 186443 | 536.355 | 1.779E-07 | 1.007E-07 | 5.830E+02 | 0.017 $\mathrm{2s2p3p~{}^{4}P_{1/2}}$ | $\mathrm{2s^{2}2p~{}^{2}P_{1/2}^{o}}$ | 186427 | 536.402 | 6.007E-07 | 3.400E-07 | 3.936E+03 | 0.039 $\mathrm{2s2p3p~{}^{4}P_{5/2}}$ | $\mathrm{2s^{2}2p~{}^{2}P_{3/2}^{o}}$ | 186402 | 536.473 | 1.227E-05 | 6.942E-06 | 2.678E+04 | 0.020 $\mathrm{2s2p3p~{}^{4}P_{3/2}}$ | $\mathrm{2s^{2}2p~{}^{2}P_{3/2}^{o}}$ | 186380 | 536.537 | 1.327E-06 | 7.506E-07 | 4.343E+03 | 0.027 – | – | – | – | – | – | – | – Table 7: Electric dipole transition data for C iii from present calculations. Upper and lower states, wavenumber, $\Delta E$, wavelength, $\lambda$, line strength, $S$, weighted oscillator strength, $gf$, transition probability, $A$, together with the relative difference between two gauges of $A$ values, $dT$, provided by the present MCDHF/RCI calculations are shown in the table. Wavelength and wavenumber values are from the NIST database (Kramida et al., 2019) when available. Wavelengths and wavenumbers marked with * are from the present calculations. Only the first ten rows are shown; the full table is available online. Upper | Lower | $\Delta{E}$(cm-1) | $\lambda$ (Å) | $S$ (a.u. of a${}_{0}^{2}$e2) | $gf$ | $A$ (s-1) | $dT$ ---|---|---|---|---|---|---|--- $\mathrm{2s7p~{}^{3}P_{1}^{o}}$ | $\mathrm{2s^{2}~{}^{1}S_{0}}$ | 365034* | 273.947* | 7.670E-07 | 8.505E-07 | 2.520E+04 | 0.005 $\mathrm{2s7p~{}^{1}P_{1}^{o}}$ | $\mathrm{2s^{2}~{}^{1}S_{0}}$ | 364896 | 274.051 | 1.043E-02 | 1.156E-02 | 3.423E+08 | 0.010 $\mathrm{2s6p~{}^{1}P_{1}^{o}}$ | $\mathrm{2s^{2}~{}^{1}S_{0}}$ | 357109 | 280.026 | 1.593E-02 | 1.728E-02 | 4.901E+08 | 0.001 $\mathrm{2s6p~{}^{3}P_{1}^{o}}$ | $\mathrm{2s^{2}~{}^{1}S_{0}}$ | 357050 | 280.073 | 5.265E-06 | 5.711E-06 | 1.619E+05 | 0.004 $\mathrm{2p3d~{}^{1}P_{1}^{o}}$ | $\mathrm{2s^{2}~{}^{1}S_{0}}$ | 346712 | 288.423 | 9.317E-04 | 9.818E-04 | 2.627E+07 | 0.003 $\mathrm{2s5p~{}^{3}P_{1}^{o}}$ | $\mathrm{2s^{2}~{}^{1}S_{0}}$ | 344236 | 290.498 | 3.900E-07 | 4.079E-07 | 1.075E+04 | 0.018 $\mathrm{2s5p~{}^{1}P_{1}^{o}}$ | $\mathrm{2s^{2}~{}^{1}S_{0}}$ | 343258 | 291.326 | 4.551E-02 | 4.747E-02 | 1.244E+09 | 0.000 $\mathrm{2p3d~{}^{3}P_{1}^{o}}$ | $\mathrm{2s^{2}~{}^{1}S_{0}}$ | 340127 | 294.007 | 7.645E-07 | 7.903E-07 | 2.035E+04 | 0.005 $\mathrm{2p3d~{}^{3}D_{1}^{o}}$ | $\mathrm{2s^{2}~{}^{1}S_{0}}$ | 337655 | 296.159 | 7.267E-07 | 7.458E-07 | 1.893E+04 | 0.005 $\mathrm{2s4p~{}^{1}P_{1}^{o}}$ | $\mathrm{2s^{2}~{}^{1}S_{0}}$ | 322404 | 310.170 | 3.480E-02 | 3.409E-02 | 7.884E+08 | 0.000 – | – | – | – | – | – | – | – Table 8: Electric dipole transition data for C iv from present calculations. Upper and lower states, wavenumber, $\Delta E$, wavelength, $\lambda$, line strength, $S$, weighted oscillator strength, $gf$, transition probability, $A$, together with the relative difference between two gauges of $A$ values, $dT$, provided by the present MCDHF/RCI calculations are shown in the table. Wavelength and wavenumber values are from the NIST database (Kramida et al., 2019). Only the first ten rows are shown; the full table is available online. Upper | Lower | $\Delta E$(cm-1) | $\lambda$ (Å) | $S$ (a.u. of a${}_{0}^{2}$e2) | $gf$ | $A$ (s-1) | $dT$ ---|---|---|---|---|---|---|--- $\mathrm{8p~{}^{2}P_{3/2}^{o}}$ | $\mathrm{2s~{}^{2}S_{1/2}}$ | 492479 | 203.054 | 5.029E-03 | 7.523E-03 | 3.043E+08 | 0.004 $\mathrm{8p~{}^{2}P_{1/2}^{o}}$ | $\mathrm{2s~{}^{2}S_{1/2}}$ | 492477 | 203.055 | 2.517E-03 | 3.766E-03 | 3.046E+08 | 0.004 $\mathrm{7p~{}^{2}P_{3/2}^{o}}$ | $\mathrm{2s~{}^{2}S_{1/2}}$ | 483950 | 206.633 | 7.885E-03 | 1.159E-02 | 4.527E+08 | 0.001 $\mathrm{7p~{}^{2}P_{1/2}^{o}}$ | $\mathrm{2s~{}^{2}S_{1/2}}$ | 483948 | 206.634 | 3.946E-03 | 5.801E-03 | 4.532E+08 | 0.001 $\mathrm{6p~{}^{2}P_{3/2}^{o}}$ | $\mathrm{2s~{}^{2}S_{1/2}}$ | 470778 | 212.414 | 1.349E-02 | 1.930E-02 | 7.132E+08 | 0.000 $\mathrm{6p~{}^{2}P_{1/2}^{o}}$ | $\mathrm{2s~{}^{2}S_{1/2}}$ | 470775 | 212.416 | 6.753E-03 | 9.657E-03 | 7.139E+08 | 0.000 $\mathrm{5p~{}^{2}P_{3/2}^{o}}$ | $\mathrm{2s~{}^{2}S_{1/2}}$ | 448862 | 222.785 | 2.644E-02 | 3.604E-02 | 1.211E+09 | 0.000 $\mathrm{5p~{}^{2}P_{1/2}^{o}}$ | $\mathrm{2s~{}^{2}S_{1/2}}$ | 448855 | 222.789 | 1.323E-02 | 1.804E-02 | 1.212E+09 | 0.000 $\mathrm{8d~{}^{2}D_{3/2}}$ | $\mathrm{2p~{}^{2}P_{1/2}^{o}}$ | 428244 | 233.511 | 1.264E-02 | 1.644E-02 | 5.027E+08 | 0.002 $\mathrm{8d~{}^{2}D_{5/2}}$ | $\mathrm{2p~{}^{2}P_{3/2}^{o}}$ | 428136 | 233.570 | 2.275E-02 | 2.958E-02 | 6.028E+08 | 0.002 – | – | – | – | – | – | – | – Table 9: Comparison of relative line strengths ($S$), weighted oscillator strengths ($gf$), and lifetimes ($\tau$), or transition probabilities ($A$), with other theoretical work and experimental results for C i – iv. The present values from the MCDHF/RCI calculations are given in the Babushkin(length) gauge. The values in the parentheses are the relative differences between the length and velocity gauges. The references for the experiments are shown in the last column. Note that the sums of the line strengths $S$ have been normalized to 100 for each multiplet in C i. | | C i | | | | | ---|---|---|---|---|---|---|--- Transition array | Mult. | $J_{u}-J_{l}$ | $S$ (a.u. of a${}_{0}^{2}$e2) | | | | MCDHF/RCI | CIV3(a) | Expt. | Expt. | $\mathrm{2s^{2}2p3p-2s^{2}2p3s}$ | $\mathrm{{}^{3}D-~{}^{3}P^{o}}$ | 3 - 2 | 46.67(1.3%) | 46.72 | 46.3 $\pm$ 1.8 (b) | | | | 2 - 1 | 25.29(1.2%) | 25.43 | 25.5 $\pm$ 1.2 (b) | | | | 1 - 0 | 11.25(1.2%) | 11.29 | 11.8 $\pm$ 0.5 (b) | | | | 2 - 2 | 8.047(1.3%) | 7.898 | 7.67 $\pm$ 0.38 (b) | | | | 1 - 1 | 8.213(1.3%) | 8.153 | 8.42 $\pm$ 0.46 (b) | | $\mathrm{2s^{2}2p3p-2s^{2}2p3s}$ | $\mathrm{{}^{3}P-~{}^{3}P^{o}}$ | 2 - 2 | 42.15(0.2%) | 41.92 | 40.6 $\pm$ 0.9 (b) | 41.3(d) | | | 1 - 1 | 7.812(0.3%) | 7.873 | 7.98 $\pm$ 0.23 (b) | 8.1 (d) | | | 1 - 2 | 15.07(0.2%) | 14.79 | 15.1 $\pm$ 0.4 (b) | 15.1(d) | | | 0 - 1 | 11.11(0.2%) | 11.11 | 11.3 $\pm$ 0.3 (b) | 11.6(d) | | | 2 - 1 | 13.41(0.2%) | 13.67 | 14.0 $\pm$ 0.35 (b) | 13.0(d) | | | 1 - 0 | 10.44(0.2%) | 10.64 | 10.9 $\pm$ 0.3 (b) | 10.9(d) | $\mathrm{2s^{2}2p3p-2s^{2}2p3s}$ | $\mathrm{{}^{3}S-~{}^{3}P^{o}}$ | 1 - 2 | 51.96(0.3%) | 51.43 | 52.4 $\pm$ 1.1 (b) | | | | 1 - 1 | 35.49(0.2%) | 35.80 | 34.8 $\pm$ 0.9 (b) | | | | 1 - 0 | 12.54(0.2%) | 12.77 | 12.8 $\pm$ 0.38 (b) | | $\mathrm{2s^{2}2p3d-2s^{2}2p3p}$ | $\mathrm{{}^{3}P^{o}-~{}^{3}S}$ | 2 - 1 | 57.27(3.0%) | 56.85 | 59.0$\pm$3.2 (c) | | | | 1 - 1 | 32.30(2.9%) | 32.56 | 32.1$\pm$4.1 (c) | | | | 0 - 1 | 10.43(2.9%) | 10.59 | 8.9$\pm$2.1 (c) | | $\mathrm{2s^{2}2p3d-2s^{2}2p3p}$ | $\mathrm{{}^{3}P^{o}-~{}^{3}P}$ | 2 - 2 | 42.67(3.3%) | 42.63 | 43.5$\pm$0.4 (c) | | | | 1 - 1 | 9.465(3.2%) | 9.669 | 10.2$\pm$0.6 (c) | | | | 1 - 2 | 13.91(3.3%) | 13.84 | 14.9$\pm$0.8 (c) | | | | 0 - 1 | 11.57(3.3%) | 11.69 | 11.2$\pm$0.5 (c) | | | | 2 - 1 | 11.72(3.4%) | 11.47 | 10.2$\pm$0.5 (c) | | | | 1 - 0 | 10.66(3.3%) | 10.70 | 10.0$\pm$0.4 (c) | | $\mathrm{2s^{2}2p4s-2s^{2}2p3p}$ | $\mathrm{{}^{3}P^{o}-~{}^{3}P}$ | 2 - 2 | 62.05(5.7%) | 50.76 | 51.2$\pm$5.0 (c) | | | | 1 - 1 | 9.312(7.1%) | 8.778 | 8.4$\pm$1.6 (c) | | | | 1 - 2 | 13.79(7.0%) | 14.08 | 14.3$\pm$2.0 (c) | | | | 0 - 1 | 7.335(8.7%) | 9.670 | 10.2$\pm$1.5 (c) | | | | 2 - 1 | 3.113(11.9%) | 8.440 | 10.1$\pm$2.0 (c) | | | | 1 - 0 | 4.397(10.3%) | 8.271 | 5.8$\pm$1.0 (c) | | $\mathrm{2s^{2}2p3d-2s^{2}2p3p}$ | $\mathrm{{}^{3}D^{o}-~{}^{3}P}$ | 3 - 2 | 47.89(1.2%) | 46.50 | 45.5$\pm$2.0 (c) | 45.1(d) | | | 2 - 1 | 26.70(1.2%) | 26.59 | 27.5$\pm$1.2 (c) | 24.2(d) | | | 1 - 0 | 9.708(1.2%) | 11.11 | 11.0$\pm$0.6 (c) | 13.6(d) | | | 2 - 2 | 8.170(0.9%) | 7.559 | 7.5$\pm$0.4 (c) | 8.0 (d) | | | 1 - 1 | 7.100(1.1%) | 7.792 | 8.1$\pm$0.5 (c) | 9.1 (d) | $\mathrm{2s^{2}2p4s-2s^{2}2p3p}$ | $\mathrm{{}^{3}P^{o}-~{}^{3}D}$ | 2 - 3 | 44.42(3.2%) | 44.77 | 44.5$\pm$2.0 (c) | | | | 1 - 2 | 23.65(3.1%) | 24.14 | 24.8$\pm$1.2 (c) | | | | 0 - 1 | 11.10(3.0%) | 11.24 | 11.5$\pm$0.6 (c) | | | | 2 - 2 | 9.145(2.6%) | 9.424 | 8.4$\pm$0.6 (c) | | | | 1 - 1 | 9.301(2.7%) | 9.071 | 9.5$\pm$0.5 (c) | | | | 2 - 1 | 2.373(1.4%) | 1.356 | 1.3$\pm$0.2 (c) | | $\mathrm{2s^{2}2p3d-2s^{2}2p3p}$ | $\mathrm{{}^{3}D^{o}-~{}^{3}D}$ | 3 - 3 | 50.31(1.5%) | 49.19 | 50.0$\pm$5.0 (c) | | | | 2 - 2 | 25.28(1.6%) | 25.72 | 25.4$\pm$2.2 (c) | | | | 1 - 1 | 11.50(1.7%) | 12.68 | 12.3$\pm$0.7 (c) | | | | 2 - 3 | 8.250(0.9%) | 7.291 | 6.6$\pm$0.6 (c) | | | | 1 - 2 | 4.043(1.8%) | 4.621 | 4.7$\pm$0.5 (c) | | $\mathrm{2s^{2}2p3d-2s^{2}2p3p}$ | $\mathrm{{}^{3}F^{o}-~{}^{3}D}$ | 4 - 3 | 43.41(0.4%) | 43.53 | 44.9$\pm$2.0 (c) | | | | 3 - 2 | 30.86(0.4%) | 30.92 | 30.0$\pm$1.5 (c) | | | | 2 - 1 | 20.77(0.3%) | 21.09 | 20.4$\pm$1.0 (c) | | | | 3 - 3 | 1.961(1.2%) | 1.743 | 1.9$\pm$0.2 (c) | | | | 2 - 2 | 2.999(0.9%) | 2.722 | 2.7$\pm$0.2 (c) | | | | C ii | | | | | Configuration | Term | $J$ | $\tau$ (ns) | Ref. | | | | MCDHF/RCI | MCHF-BP(e) | Expt. | | $\mathrm{2s2p^{2}}$ | $\mathrm{~{}^{2}S}$ | 1/2 | 0.4497 (0.7%) | 0.4523 | 0.44$\pm$ 0.02 | (f) | $\mathrm{2s^{2}3s}$ | $\mathrm{~{}^{2}S}$ | 1/2 | 2.292 (0.6%) | 2.266 | 2.4 $\pm$ 0.3 | (f) | $\mathrm{2s^{2}4s}$ | $\mathrm{~{}^{2}S}$ | 1/2 | 2.017 (0.1%) | | 1.9 $\pm$ 0.1 | (f) | $\mathrm{2s^{2}5s}$ | $\mathrm{~{}^{2}S}$ | 1/2 | 3.774 (0.1%) | | 3.7 $\pm$ 0.2 | (f) | $\mathrm{2s2p^{2}}$ | $\mathrm{~{}^{2}P^{o}}$ | 1/2 | 0.2446(0.3%) | 0.2445 | 0.25$\pm$ 0.01 | (f) | | | 3/2 | 0.2445(0.3%) | 0.2449 | 0.25$\pm$ 0.01 | (f) | $\mathrm{2s^{2}3p}$ | $\mathrm{~{}^{2}P^{o}}$ | 1/2 | 9.265(0.7%) | 8.973 | 8.9 $\pm$ 0.4 | (f) | | | 3/2 | 9.255(0.7%) | 8.963 | 8.9 $\pm$ 0.4 | (f) | $\mathrm{2s^{2}4p}$ | $\mathrm{~{}^{2}P^{o}}$ | 1/2 | 3.838(1.3%) | | 3.8 $\pm$ 0.2 | (f) | | | 3/2 | 3.854(1.3%) | | 3.8 $\pm$ 0.2 | (f) | $\mathrm{2p^{3}}$ | $\mathrm{~{}^{2}P^{o}}$ | 1/2 | 0.4998(0.8%) | 0.4966 | 0.48$\pm$ 0.02 | (f) | | | 3/2 | 0.4981(0.8%) | 0.4944 | 0.48$\pm$ 0.02 | (f) | $\mathrm{2s^{2}5p}$ | $\mathrm{~{}^{2}P^{o}}$ | 1/2 | 5.044(0.2%) | | 5.2 $\pm$ 0.3 | (f) | | | 3/2 | 5.099(0.2%) | | 5.2 $\pm$ 0.3 | (f) | $\mathrm{2s^{2}3d}$ | $\mathrm{~{}^{2}D}$ | 3/2 | 0.3490(0.2%) | 0.3493 | 0.34$\pm$ 0.01 | (f) | | | 5/2 | 0.3491(0.2%) | 0.3494 | 0.34$\pm$ 0.01 | (f) | $\mathrm{2s^{2}4d}$ | $\mathrm{~{}^{2}D}$ | 3/2 | 0.7299(0.2%) | | 0.75$\pm$ 0.03 | (f) | | | 5/2 | 0.7304(0.2%) | | 0.75$\pm$ 0.03 | (f) | Configuration | Term | $J$ | $\tau$ (ms) | Ref. | | | | MCDHF/RCI | MCHF-BP(e) | Expt. | | $\mathrm{2s2p^{2}}$ | $\mathrm{~{}^{4}P}$ | 1/2 | 8.151 (47.6%) | 7.654 | 7.95$\pm$0.07 | (g) | | | 3/2 | 106.1 (68.5%) | 96.93 | 104.1$\pm$0.5 | (g) | | | 5/2 | 22.66 (48.0%) | 22.34 | 22.05$\pm$0.07 | (g) | | | C iii | | | | | Transition array | Mult. | $J_{u}-J_{l}$ | $gf$ | Ref. | | | | MCDHF/RCI | MCHF-BP(h) | Expt. | | $\mathrm{2s2p}$ \- $\mathrm{2s^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{1}S}$ | 1-0 | 0.7592(0.1%) | 0.7583 | 0.75 $\pm$ 0.03 | (f) | $\mathrm{2p^{2}}$ \- $\mathrm{2s2p}$ | $\mathrm{{}^{1}S-~{}^{1}P^{o}}$ | 0-1 | 0.1623($<$0.05%) | 0.1622 | 0.152 $\pm$ 0.009 | (f) | $\mathrm{2p^{2}}$ \- $\mathrm{2s2p}$ | $\mathrm{{}^{1}D-~{}^{1}P^{o}}$ | 2-1 | 0.1815(0.5%) | 0.1819 | 0.183 $\pm$ 0.005 | (f) | Configuration | Term | $J$ | $\tau$ (ns) | Ref. | | | | MCDHF/RCI | MCHF-BP(h) | Expt. | | $\mathrm{2s2p}$ | $\mathrm{~{}^{1}P^{o}}$ | 1 | 0.5638(0.1%) ns | 0.5651 | 0.57$\pm$ 0.02 | (f) | $\mathrm{2p^{2}}$ | $\mathrm{~{}^{1}S}$ | 0 | 0.4766($<$0.05%) ns | 0.4764 | 0.51$\pm$ 0.01 | (f) | $\mathrm{2p^{2}}$ | $\mathrm{~{}^{1}D}$ | 2 | 7.240(0.5%) ns | 7.191 | 7.2$\pm$ 0.2 | (f) | $\mathrm{2s3s}$ | $\mathrm{~{}^{1}S}$ | 0 | 1.164($<$0.05%) ns | 1.171 | 1.17$\pm$ 0.05 | (f) | | | C iv | | | | | Transition array | Mult. | $J_{u}-J_{l}$ | $A(10^{8}\mathrm{s^{-1}})$ | Ref. | | | MCDHF/RCI | MCHF-BP(j) | NIST | Expt. | $\mathrm{2p-2s}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}S}$ | 1/2 - 1/2 | 2.632($<$0.05%) | 2.6320 | 2.65 | 2.72$\pm$0.07 | (k) $\mathrm{2p-2s}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}S}$ | 3/2 - 1/2 | 2.646($<$0.05%) | 2.6459 | 2.64 | 2.71$\pm$0.07 | (k) Configuration | Term | $J$ | $\tau$ (ns) | Ref. | | | MCDHF/RCI | MCHF-BP(j) | Model Potential(o) | Expt. | $\mathrm{3s}$ | $\mathrm{~{}^{2}S}$ | 1/2 | 0.2350($<$0.05%) | 0.2350 | 0.236 | 0.25 $\pm$ 0.1 | (l) $\mathrm{4s}$ | $\mathrm{~{}^{2}S}$ | 1/2 | 0.3755($<$0.05%) | 0.3747 | 0.377 | 0.34 $\pm$ 0.035 | (m) $\mathrm{2p}$ | $\mathrm{~{}^{2}P^{o}}$ | 1/2 | 3.799 ($<$0.05%) | 3.799 | 3.79 | 3.7 $\pm$ 0.1 | (k) | | 3/2 | 3.779 ($<$0.05%) | 3.779 | 3.79 | | $\mathrm{3p}$ | $\mathrm{~{}^{2}P^{o}}$ | 1/2 | 0.2146($<$0.05%) | 0.2142 | 0.216 | 0.226 $\pm$ 0.03 | (n) | | 3/2 | 0.2149($<$0.05%) | 0.2145 | 0.216 | | $\mathrm{4p}$ | $\mathrm{~{}^{2}P^{o}}$ | 1/2 | 0.3435($<$0.05%) | | 0.344 | 0.32 $\pm$ 0.03 | (m) | | 3/2 | 0.3440($<$0.05%) | | 0.344 | | $\mathrm{3d}$ | $\mathrm{~{}^{2}D}$ | 3/2 | 0.05717($<$0.05%) | 0.05716 | 0.0572 | 0.0575$\pm$ 0.006 | (n) | | 5/2 | 0.05719($<$0.05%) | 0.05719 | 0.0572 | | $\mathrm{4d}$ | $\mathrm{~{}^{2}D}$ | 3/2 | 0.1312 ($<$0.05%) | | 0.130 | 0.14 $\pm$ 0.015 | (m) | | 5/2 | 0.1313 ($<$0.05%) | | 0.130 | | $\mathrm{5d}$ | $\mathrm{~{}^{2}D}$ | 3/2 | 0.2511 ($<$0.05%) | | 0.251 | 0.23 $\pm$ 0.023 | (m) | | 5/2 | 0.2512 ($<$0.05%) | | 0.251 | | (a)Hibbert et al. (1993); (b)Musielok et al. (1997); (c)Bacawski et al. (2001); (d)Golly et al. (2003); (e)Tachiev & Fischer (2000); (f)Reistad et al. (1986); (g)Träbert et al. (1999); (h)Tachiev & Fischer (1999); (j)Fischer et al. (1998); (k)Knystautas et al. (1971); (l)Donnelly et al. (1978); (m)Buchet- Poulizac & Buchet (1973b); (n)Jacques et al. (1980); (o)Peach et al. (1988). Table 10: Comparison of line strengths ($S$) and transition rates ($A$) with other theoretical results for C i. The present values from the MCDHF/RCI calculations are given in the Babushkin(length) gauge. The wavenumber $\Delta E$ and wavelength $\lambda$ values are taken from the NIST database. The estimated uncertainties $dT$ of the transition rates are given as percentages in parentheses. Transition array | Mult. | $J_{u}-J_{l}$ | $\Delta E$ | $\lambda$ | MCDHF/RCI | | Spline FCS(a) | | MCHF-BP(b) ---|---|---|---|---|---|---|---|---|--- | | | ($\mathrm{cm^{-1}}$) | (Å) | $S$ (a.u. of a${}_{0}^{2}$e2) | $A$ (s-1) | | $S$ (a.u. of a${}_{0}^{2}$e2) | $A$ (s-1) | | $S$ (a.u. of a${}_{0}^{2}$e2) | $A$ (s-1) $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}F^{o}-~{}^{3}P}$ | 3 - 2 | 78172 | 1279.228 | 5.79E-02 | 7.92E+06(0.1%) | | 4.52E-02 | 6.24E+06 | | 8.20E-02 | 1.14E+07 | | 2 - 2 | 78155 | 1279.498 | 1.27E-02 | 2.43E+06(0.3%) | | 9.62E-03 | 1.86E+06 | | 1.16E-02 | 2.25E+06 | | 2 - 1 | 78182 | 1279.056 | 1.08E-02 | 2.07E+06(0.1%) | | 8.95E-03 | 1.73E+06 | | 2.08E-02 | 4.04E+06 $\mathrm{2s^{2}2p4d-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}F^{o}-~{}^{3}P}$ | 3 - 2 | 83717 | 1194.488 | 4.48E-02 | 7.52E+06(0.1%) | | 3.99E-02 | 6.77E+06 | | | | | 2 - 2 | 83709 | 1194.614 | 1.59E-03 | 3.74E+05(0.5%) | | 1.44E-03 | 3.42E+05 | | | | | 2 - 1 | 83736 | 1194.229 | 1.97E-02 | 4.64E+06(0.2%) | | 1.73E-02 | 4.10E+06 | | | $\mathrm{2s^{2}2p5d-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}F^{o}-~{}^{3}P}$ | 3 - 2 | 86283 | 1158.966 | 4.17E-02 | 7.67E+06(0.3%) | | 3.88E-02 | 7.21E+06 | | | | | 2 - 2 | 86274 | 1159.094 | 1.38E-03 | 3.55E+05(0.3%) | | 1.36E-03 | 3.54E+05 | | | | | 2 - 1 | 86301 | 1158.731 | 2.20E-02 | 5.66E+06(0.3%) | | 1.96E-02 | 5.11E+06 | | | $\mathrm{2s^{2}2p3s-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{3}P}$ | 2 - 2 | 60349 | 1657.008 | 2.84E+00 | 2.50E+08(1.2%) | | 2.69E+00 | 2.41E+08 | | 2.93E+00 | 2.61E+08 | | 1 - 2 | 60309 | 1658.121 | 9.46E-01 | 1.39E+08(1.2%) | | 8.95E-01 | 1.33E+08 | | 9.75E-01 | 1.45E+08 | | 2 - 1 | 60376 | 1656.267 | 9.47E-01 | 8.35E+07(1.2%) | | 8.97E-01 | 8.05E+07 | | 9.78E-01 | 8.73E+07 | | 1 - 1 | 60336 | 1657.379 | 5.67E-01 | 8.32E+07(1.2%) | | 5.37E-01 | 8.01E+07 | | 5.84E-01 | 8.67E+07 | | 0 - 1 | 60317 | 1657.907 | 7.57E-01 | 3.33E+08(1.2%) | | 7.17E-01 | 3.20E+08 | | 7.80E-01 | 3.47E+08 | | 1 - 0 | 60352 | 1656.928 | 7.57E-01 | 1.11E+08(1.2%) | | 7.17E-01 | 1.07E+08 | | 7.81E-01 | 1.16E+08 $\mathrm{2s^{2}2p3s-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{3}P}$ | 1 - 2 | 61938 | 1614.507 | 1.80E-04 | 2.86E+04(1.6%) | | 2.72E-04 | 4.40E+04 | | 1.85E-04 | 2.97E+04 | | 1 - 1 | 61965 | 1613.803 | 1.65E-04 | 2.61E+04(1.0%) | | 1.62E-04 | 2.62E+04 | | 1.74E-04 | 2.81E+04 | | 1 - 0 | 61981 | 1613.376 | 2.21E-04 | 3.51E+04(1.1%) | | 2.34E-04 | 3.79E+04 | | 2.26E-04 | 3.65E+04 $\mathrm{2s2p^{3}-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}D^{o}-~{}^{3}P}$ | 3 - 2 | 64043 | 1561.437 | 1.53E+00 | 1.22E+08(2.3%) | | 1.42E+00 | 1.14E+08 | | 1.54E+00 | 1.19E+08 | | 2 - 2 | 64046 | 1561.366 | 2.72E-01 | 3.03E+07(2.1%) | | 2.52E-01 | 2.84E+07 | | 2.75E-01 | 2.96E+07 | | 1 - 2 | 64047 | 1561.339 | 1.81E-02 | 3.35E+06(2.0%) | | 1.67E-02 | 3.14E+06 | | 1.83E-02 | 3.28E+06 | | 2 - 1 | 64073 | 1560.708 | 8.22E-01 | 9.14E+07(2.2%) | | 7.60E-01 | 8.59E+07 | | 8.28E-01 | 8.91E+07 | | 1 - 1 | 64074 | 1560.681 | 2.73E-01 | 5.06E+07(2.1%) | | 2.53E-01 | 4.76E+07 | | 2.76E-01 | 4.94E+07 | | 1 - 0 | 64090 | 1560.282 | 3.65E-01 | 6.78E+07(2.1%) | | 3.38E-01 | 6.37E+07 | | 3.68E-01 | 6.61E+07 $\mathrm{2s2p^{3}-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{3}P}$ | 2 - 2 | 75212 | 1329.562 | 7.98E-01 | 1.42E+08(3.1%) | | 7.44E-01 | 1.35E+08 | | 9.54E-01 | 1.66E+08 | | 1 - 2 | 75210 | 1329.600 | 2.69E-01 | 7.95E+07(3.1%) | | 2.51E-01 | 7.57E+07 | | 3.20E-01 | 9.28E+07 | | 2 - 1 | 75239 | 1329.085 | 2.58E-01 | 4.58E+07(3.1%) | | 2.40E-01 | 4.34E+07 | | 3.12E-01 | 5.44E+07 | | 1 - 1 | 75237 | 1329.123 | 1.64E-01 | 4.85E+07(3.1%) | | 1.53E-01 | 4.63E+07 | | 1.93E-01 | 5.60E+07 | | 0 - 1 | 75238 | 1329.100 | 2.17E-01 | 1.93E+08(3.1%) | | 2.04E-01 | 1.85E+08 | | 2.57E-01 | 2.24E+08 | | 1 - 0 | 75254 | 1328.833 | 2.13E-01 | 6.30E+07(3.1%) | | 1.99E-01 | 6.00E+07 | | 2.54E-01 | 7.37E+07 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}D^{o}-~{}^{3}P}$ | 2 - 2 | 77636 | 1288.055 | 4.63E-04 | 8.69E+04(1.4%) | | 4.27E-04 | 8.10E+04 | | 3.51E-04 | 6.68E+04 | | 2 - 1 | 77663 | 1287.608 | 8.40E-04 | 1.58E+05(1.2%) | | 7.68E-04 | 1.46E+05 | | 7.84E-04 | 1.49E+05 $\mathrm{2s^{2}2p4s-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{3}P}$ | 2 - 2 | 78104 | 1280.333 | 3.23E-01 | 6.17E+07(0.7%) | | 3.21E-01 | 6.21E+07 | | 3.30E-01 | 6.40E+07 | | 1 - 2 | 78073 | 1280.847 | 1.13E-01 | 3.59E+07(0.5%) | | 1.10E-01 | 3.55E+07 | | 1.13E-01 | 3.63E+07 | | 2 - 1 | 78131 | 1279.890 | 1.96E-01 | 3.76E+07(0.5%) | | 1.81E-01 | 3.50E+07 | | 1.77E-01 | 3.44E+07 | | 1 - 1 | 78100 | 1280.404 | 5.55E-02 | 1.77E+07(0.6%) | | 5.57E-02 | 1.79E+07 | | 5.83E-02 | 1.88E+07 | | 0 - 1 | 78088 | 1280.597 | 9.23E-02 | 8.82E+07(0.5%) | | 8.91E-02 | 8.61E+07 | | 9.14E-02 | 8.84E+07 | | 1 - 0 | 78116 | 1280.135 | 1.16E-01 | 3.71E+07(0.5%) | | 1.11E-01 | 3.57E+07 | | 1.10E-01 | 3.56E+07 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}D^{o}-~{}^{3}P}$ | 3 - 2 | 78274 | 1277.550 | 1.68E+00 | 2.31E+08(0.2%) | | 1.61E+00 | 2.23E+08 | | 1.64E+00 | 2.28E+08 | | 2 - 2 | 78264 | 1277.723 | 3.44E-01 | 6.61E+07(0.1%) | | 3.22E-01 | 6.26E+07 | | 3.24E-01 | 6.32E+07 | | 1 - 2 | 78250 | 1277.954 | 1.47E-02 | 4.70E+06($<$0.05%) | | 1.72E-02 | 5.58E+06 | | 1.84E-02 | 5.95E+06 | | 2 - 1 | 78291 | 1277.282 | 8.75E-01 | 1.68E+08(0.2%) | | 8.42E-01 | 1.64E+08 | | 8.73E-01 | 1.70E+08 | | 1 - 1 | 78277 | 1277.513 | 2.43E-01 | 7.80E+07(0.1%) | | 2.66E-01 | 8.61E+07 | | 2.85E-01 | 9.26E+07 | | 1 - 0 | 78293 | 1277.245 | 3.42E-01 | 1.10E+08(0.2%) | | 3.59E-01 | 1.17E+08 | | 3.88E-01 | 1.26E+08 $\mathrm{2s^{2}2p4s-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{3}P}$ | 1 - 2 | 78296 | 1277.190 | 1.14E-02 | 3.64E+06(0.4%) | | 6.78E-03 | 2.20E+06 | | 5.59E-03 | 1.82E+06 | | 1 - 1 | 78323 | 1276.750 | 7.84E-02 | 2.52E+07(0.2%) | | 3.90E-02 | 1.27E+07 | | 2.89E-02 | 9.41E+06 | | 1 - 0 | 78340 | 1276.482 | 5.97E-02 | 1.92E+07(0.1%) | | 2.48E-02 | 8.05E+06 | | 1.55E-02 | 5.05E+06 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}F^{o}-~{}^{3}P}$ | 3 - 2 | 78486 | 1274.109 | 1.00E-02 | 1.39E+06(0.1%) | | 1.18E-02 | 1.65E+06 | | 8.60E-03 | 1.21E+06 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{3}P}$ | 1 - 2 | 78687 | 1270.844 | 1.77E-06 | 5.75E+02(0.4%) | | 3.03E-07 | 9.97E+01 | | 1.28E-05 | 4.23E+03 | | 1 - 1 | 78714 | 1270.408 | 6.02E-04 | 1.96E+05(0.3%) | | 5.66E-04 | 1.87E+05 | | 6.50E-04 | 2.15E+05 | | 1 - 0 | 78731 | 1270.143 | 1.65E-03 | 5.39E+05(0.1%) | | 1.61E-03 | 5.32E+05 | | 1.38E-03 | 4.55E+05 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{3}P}$ | 2 - 2 | 79267 | 1261.552 | 8.17E-01 | 1.66E+08(2.1%) | | 8.13E-01 | 1.67E+08 | | 6.65E-01 | 1.35E+08 | | 1 - 2 | 79275 | 1261.425 | 2.74E-01 | 9.26E+07(2.1%) | | 2.72E-01 | 9.31E+07 | | 2.23E-01 | 7.55E+07 | | 2 - 1 | 79294 | 1261.122 | 2.54E-01 | 5.15E+07(2.1%) | | 2.52E-01 | 5.19E+07 | | 1.98E-01 | 4.02E+07 | | 1 - 1 | 79302 | 1260.996 | 1.69E-01 | 5.72E+07(2.1%) | | 1.68E-01 | 5.75E+07 | | 1.39E-01 | 4.70E+07 | | 0 - 1 | 79306 | 1260.926 | 2.20E-01 | 2.24E+08(2.1%) | | 2.18E-01 | 2.24E+08 | | 1.79E-01 | 1.82E+08 | | 1 - 0 | 79318 | 1260.735 | 2.12E-01 | 7.19E+07(2.1%) | | 2.11E-01 | 7.23E+07 | | 1.69E-01 | 5.73E+07 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}F^{o}-~{}^{1}D}$ | 2 - 2 | 68006 | 1470.449 | 4.71E-04 | 5.91E+04(1.4%) | | | | | 4.22E-04 | 5.36E+04 | | 3 - 2 | 68022 | 1470.094 | 1.52E-02 | 1.36E+06(0.2%) | | | | | 1.55E-02 | 1.41E+06 $\mathrm{2s^{2}2p3s-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{1}D}$ | 1 - 2 | 50160 | 1993.620 | 1.04E-03 | 8.67E+04(1.4%) | | | | | 9.65E-04 | 8.18E+04 | | 2 - 2 | 50200 | 1992.012 | 1.97E-05 | 9.90E+02(3.2%) | | | | | 1.50E-05 | 7.66E+02 $\mathrm{2s^{2}2p3s-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{1}D}$ | 1 - 2 | 51789 | 1930.905 | 3.59E+00 | 3.30E+08(1.5%) | | | | | 3.62E+00 | 3.37E+08 $\mathrm{2s2p^{3}-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}D^{o}-~{}^{1}D}$ | 3 - 2 | 53894 | 1855.483 | 9.91E-06 | 4.71E+02(24.2%) | | | | | 6.60E-06 | 3.01E+02 $\mathrm{2s2p^{3}-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{1}D}$ | 1 - 2 | 65061 | 1537.011 | 8.58E-06 | 1.65E+03(1.7%) | | | | | 8.59E-08 | 1.60E+01 | | 2 - 2 | 65063 | 1536.960 | 2.51E-05 | 2.89E+03(6.9%) | | | | | 7.15E-06 | 8.02E+02 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}D^{o}-~{}^{1}D}$ | 2 - 2 | 67487 | 1481.763 | 3.03E-01 | 3.72E+07(1.2%) | | | | | 2.77E-01 | 3.45E+07 $\mathrm{2s^{2}2p4s-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{1}D}$ | 1 - 2 | 67924 | 1472.231 | 4.55E-03 | 9.48E+05(1.6%) | | | | | 3.68E-03 | 7.77E+05 | | 2 - 2 | 67955 | 1471.552 | 2.02E-04 | 2.53E+04(1.0%) | | | | | 1.77E-04 | 2.25E+04 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}D^{o}-~{}^{1}D}$ | 3 - 2 | 68125 | 1467.877 | 7.46E-03 | 6.72E+05(0.2%) | | | | | 5.35E-03 | 4.88E+05 | | 2 - 2 | 68114 | 1468.106 | 5.36E-05 | 6.76E+03(2.9%) | | | | | 7.36E-05 | 9.40E+03 | | 1 - 2 | 68100 | 1468.410 | 4.59E-02 | 9.64E+06(2.2%) | | | | | 1.13E-02 | 2.40E+06 $\mathrm{2s^{2}2p4s-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{1}D}$ | 1 - 2 | 68147 | 1467.402 | 2.36E-01 | 4.97E+07(2.0%) | | | | | 2.57E-01 | 5.48E+07 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}F^{o}-~{}^{1}D}$ | 3 - 2 | 68336 | 1463.336 | 1.99E+00 | 1.81E+08(0.3%) | | | | | 1.94E+00 | 1.78E+08 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{1}D}$ | 1 - 2 | 68538 | 1459.031 | 2.18E-01 | 4.66E+07(0.1%) | | | | | 2.51E-01 | 5.44E+07 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{1}D}$ | 2 - 2 | 69118 | 1446.797 | 1.83E-05 | 2.46E+03(5.2%) | | | | | 3.09E-05 | 4.13E+03 | | 1 - 2 | 69126 | 1446.630 | 4.50E-06 | 1.01E+03(5.4%) | | | | | 2.65E-06 | 5.92E+02 $\mathrm{2s^{2}2p3s-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{1}S}$ | 1 - 0 | 38704 | 2583.670 | 1.61E-04 | 6.13E+03(5.2%) | | | | | 1.48E-04 | 5.72E+03 $\mathrm{2s^{2}2p3s-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{1}S}$ | 1 - 0 | 40333 | 2479.310 | 6.69E-01 | 2.89E+07(3.7%) | | | | | 6.31E-01 | 2.76E+07 $\mathrm{2s2p^{3}-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{1}S}$ | 1 - 0 | 53605 | 1865.464 | 6.83E-06 | 7.36E+02(23.2%) | | | | | 2.71E-06 | 2.83E+02 $\mathrm{2s^{2}2p4s-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{1}S}$ | 1 - 0 | 56468 | 1770.891 | 9.52E-05 | 1.13E+04(6.6%) | | | | | 8.80E-05 | 1.06E+04 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}D^{o}-~{}^{1}S}$ | 1 - 0 | 56645 | 1765.366 | 1.32E-02 | 1.59E+06(2.9%) | | | | | 6.33E-03 | 7.72E+05 $\mathrm{2s^{2}2p4s-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{1}S}$ | 1 - 0 | 56692 | 1763.909 | 1.72E-02 | 2.07E+06(5.0%) | | | | | 1.99E-02 | 2.43E+06 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{1}S}$ | 1 - 0 | 57083 | 1751.827 | 7.32E-01 | 9.00E+07($<$0.05%) | | | | | 6.67E-01 | 8.33E+07 $\mathrm{2s^{2}2p3d-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{1}S}$ | 1 - 0 | 57670 | 1733.980 | 4.99E-05 | 6.48E+03(8.1%) | | | | | 1.18E-04 | 1.53E+04 $\mathrm{2s^{2}2p4d-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}D^{o}-~{}^{3}P}$ | 2 - 2 | 83454 | 1198.262 | 3.73E-04 | 8.70E+04(0.8%) | | 2.87E-04 | 6.75E+04 | | | | | 2 - 1 | 83481 | 1197.875 | 1.03E-03 | 2.39E+05(0.8%) | | 8.64E-04 | 2.04E+05 | | | $\mathrm{2s^{2}2p5s-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{3}P}$ | 2 - 2 | 83747 | 1194.063 | 1.18E-01 | 2.78E+07(0.9%) | | 1.25E-01 | 2.98E+07 | | | | | 1 - 2 | 83704 | 1194.686 | 4.34E-02 | 1.70E+07(0.8%) | | 4.47E-02 | 1.77E+07 | | | | | 2 - 1 | 83774 | 1193.678 | 1.17E-01 | 2.75E+07(0.6%) | | 1.07E-01 | 2.54E+07 | | | | | 1 - 1 | 83731 | 1194.301 | 1.95E-02 | 7.67E+06(0.9%) | | 2.10E-02 | 8.34E+06 | | | | | 0 - 1 | 83723 | 1194.405 | 3.64E-02 | 4.29E+07(0.7%) | | 3.70E-02 | 4.40E+07 | | | | | 1 - 0 | 83747 | 1194.066 | 4.98E-02 | 1.95E+07(0.7%) | | 4.90E-02 | 1.94E+07 | | | $\mathrm{2s^{2}2p4d-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}D^{o}-~{}^{3}P}$ | 3 - 2 | 83805 | 1193.240 | 6.80E-01 | 1.15E+08(0.1%) | | 6.52E-01 | 1.11E+08 | | | | | 2 - 2 | 83794 | 1193.393 | 1.57E-01 | 3.70E+07(0.1%) | | 1.45E-01 | 3.46E+07 | | | | | 1 - 2 | 83776 | 1193.649 | 5.60E-03 | 2.20E+06(0.1%) | | 5.67E-03 | 2.25E+06 | | | | | 2 - 1 | 83821 | 1193.009 | 3.37E-01 | 7.95E+07($<$0.05%) | | 3.31E-01 | 7.90E+07 | | | | | 1 - 1 | 83803 | 1193.264 | 1.10E-01 | 4.31E+07($<$0.05%) | | 1.08E-01 | 4.29E+07 | | | | | 1 - 0 | 83820 | 1193.030 | 1.65E-01 | 6.50E+07(0.1%) | | 1.61E-01 | 6.39E+07 | | | $\mathrm{2s^{2}2p5s-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{3}P}$ | 1 - 2 | 83833 | 1192.835 | 6.48E-03 | 2.55E+06(0.5%) | | 5.50E-03 | 2.19E+06 | | | | | 1 - 1 | 83860 | 1192.451 | 2.38E-02 | 9.38E+06(0.3%) | | 1.91E-02 | 7.61E+06 | | | | | 1 - 0 | 83877 | 1192.218 | 7.36E-03 | 2.90E+06(0.3%) | | 5.39E-03 | 2.15E+06 | | | $\mathrm{2s^{2}2p4d-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}F^{o}-~{}^{3}P}$ | 3 - 2 | 83903 | 1191.841 | 1.54E-02 | 2.61E+06(0.1%) | | 1.70E-02 | 2.90E+06 | | | $\mathrm{2s^{2}2p4d-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{3}P}$ | 1 - 2 | 83988 | 1190.636 | 3.47E-05 | 1.37E+04(0.3%) | | 3.15E-05 | 1.26E+04 | | | | | 1 - 1 | 84015 | 1190.253 | 1.25E-03 | 4.93E+05(0.1%) | | 1.20E-03 | 4.79E+05 | | | | | 1 - 0 | 84032 | 1190.021 | 2.39E-03 | 9.48E+05($<$0.05%) | | 2.26E-03 | 9.05E+05 | | | $\mathrm{2s^{2}2p4d-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{3}P}$ | 2 - 2 | 84059 | 1189.631 | 2.38E-01 | 5.68E+07(1.0%) | | 2.19E-01 | 5.28E+07 | | | | | 1 - 2 | 84072 | 1189.447 | 8.04E-02 | 3.20E+07(1.0%) | | 7.37E-02 | 2.96E+07 | | | | | 2 - 1 | 84086 | 1189.249 | 5.07E-02 | 1.21E+07(1.1%) | | 4.53E-02 | 1.09E+07 | | | | | 1 - 1 | 84099 | 1189.065 | 5.43E-02 | 2.16E+07(1.0%) | | 5.01E-02 | 2.02E+07 | | | | | 0 - 1 | 84104 | 1188.993 | 6.52E-02 | 7.79E+07(1.0%) | | 5.96E-02 | 7.19E+07 | | | | | 1 - 0 | 84116 | 1188.833 | 5.35E-02 | 2.13E+07(1.1%) | | 4.86E-02 | 1.96E+07 | | | $\mathrm{2s^{2}2p5d-2s^{2}2p^{2}}$ | $\mathrm{{}^{1}D^{o}-~{}^{3}P}$ | 2 - 2 | 86141 | 1160.876 | 6.52E-04 | 1.67E+05($<$0.05%) | | 4.58E-04 | 1.19E+05 | | | | | 2 - 1 | 86168 | 1160.513 | 1.50E-03 | 3.86E+05(0.8%) | | 1.26E-03 | 3.27E+05 | | | $\mathrm{2s^{2}2p6s-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{3}P}$ | 2 - 2 | 86325 | 1158.400 | 2.62E-02 | 6.76E+06(0.1%) | | 3.15E-02 | 8.19E+06 | | | | | 1 - 2 | 86288 | 1158.907 | 1.78E-02 | 7.62E+06(0.2%) | | 1.93E-02 | 8.36E+06 | | | | | 2 - 1 | 86352 | 1158.038 | 1.27E-01 | 3.29E+07(0.2%) | | 1.19E-01 | 3.10E+07 | | | | | 1 - 1 | 86315 | 1158.544 | 6.63E-03 | 2.85E+06(0.4%) | | 7.73E-03 | 3.36E+06 | | | | | 0 - 1 | 86305 | 1158.674 | 1.57E-02 | 2.02E+07(0.3%) | | 1.68E-02 | 2.19E+07 | | | | | 1 - 0 | 86331 | 1158.324 | 2.44E-02 | 1.05E+07(0.1%) | | 2.50E-02 | 1.09E+07 | | | $\mathrm{2s^{2}2p5d-2s^{2}2p^{2}}$ | $\mathrm{{}^{3}D^{o}-~{}^{3}P}$ | 3 - 2 | 86354 | 1158.018 | 3.20E-01 | 5.90E+07(0.3%) | | 2.99E-01 | 5.56E+07 | | | | | 2 - 2 | 86346 | 1158.131 | 1.05E-01 | 2.71E+07($<$0.05%) | | 9.90E-02 | 2.58E+07 | | | | | 1 - 2 | 86319 | 1158.492 | 1.37E-03 | 5.88E+05(0.2%) | | 1.45E-03 | 6.29E+05 | | | | | 2 - 1 | 86373 | 1157.769 | 1.03E-01 | 2.65E+07(0.4%) | | 1.00E-01 | 2.62E+07 | | | | | 1 - 1 | 86346 | 1158.130 | 4.69E-02 | 2.02E+07(0.2%) | | 4.55E-02 | 1.98E+07 | | | | | 1 - 0 | 86362 | 1157.909 | 8.57E-02 | 3.69E+07(0.3%) | | 8.08E-02 | 3.51E+07 | | | Table 10: Continued. (a)Zatsarinny & Fischer (2002); (b)Fischer (2006). Table 11: Comparison of line strengths ($S$) and transition rates ($A$) with other theoretical results for C ii. The present values from the MCDHF/RCI calculations are given in the Babushkin(length) gauge. The wavenumber $\Delta E$ and wavelength $\lambda$ values are taken from the NIST database. The estimated uncertainties $dT$ of the transition rates are given as percentages in parentheses. Transition array | Mult. | $J_{u}-J_{l}$ | $\Delta E$ | $\lambda$ | MCDHF/RCI | | MCHF-BP(a) | | CIV3(b) ---|---|---|---|---|---|---|---|---|--- | | | ($\mathrm{cm^{-1}}$) | (Å) | $S$ (a.u. of a${}_{0}^{2}$e2) | $A$ (s-1) | | $S$ (a.u. of a${}_{0}^{2}$e2) | $A$ (s-1) | | $A$ (s-1) $\mathrm{2s2p^{2}-2s^{2}2p}$ | $\mathrm{{}^{2}D-~{}^{2}P^{o}}$ | 5/2 - 3/2 | 74866 | 1335.708 | 2.03E+00 | 2.89E+08(0.2%) | | 2.03E+00 | 2.90E+08 | | 2.89E+08 | | 3/2 - 3/2 | 74869 | 1335.663 | 2.24E-01 | 4.79E+07($<$0.05%) | | 2.24E-01 | 4.80E+07 | | 4.79E+07 | | 3/2 - 1/2 | 74932 | 1334.532 | 1.13E+00 | 2.42E+08(0.1%) | | 1.13E+00 | 2.43E+08 | | 2.42E+08 $\mathrm{2s2p^{2}-2s^{2}2p}$ | $\mathrm{{}^{2}S-~{}^{2}P^{o}}$ | 1/2 - 3/2 | 96430 | 1037.018 | 1.62E+00 | 1.48E+09(0.7%) | | 1.61E+00 | 1.47E+09 | | 1.47E+09 | | 1/2 - 1/2 | 96493 | 1036.337 | 8.18E-01 | 7.48E+08(0.6%) | | 8.11E-01 | 7.43E+08 | | 7.46E+08 $\mathrm{2s2p^{2}-2s^{2}2p}$ | $\mathrm{{}^{2}P-~{}^{2}P^{o}}$ | 1/2 - 3/2 | 110560 | 904.480 | 9.96E-01 | 1.37E+09(0.3%) | | 9.96E-01 | 1.37E+09 | | 1.36E+09 | | 3/2 - 3/2 | 110602 | 904.142 | 4.95E+00 | 3.41E+09(0.4%) | | 4.94E+00 | 3.41E+09 | | 3.38E+09 | | 1/2 - 1/2 | 110624 | 903.962 | 1.97E+00 | 2.72E+09(0.4%) | | 1.97E+00 | 2.72E+09 | | 2.69E+09 | | 3/2 - 1/2 | 110665 | 903.623 | 9.88E-01 | 6.82E+08(0.4%) | | 9.87E-01 | 6.82E+08 | | 6.76E+08 $\mathrm{2s^{2}3s-2s^{2}2p}$ | $\mathrm{{}^{2}S-~{}^{2}P^{o}}$ | 1/2 - 3/2 | 116474 | 858.559 | 1.81E-01 | 2.89E+08(0.6%) | | 1.83E-01 | 2.93E+08 | | 2.83E+08 | | 1/2 - 1/2 | 116537 | 858.092 | 9.19E-02 | 1.47E+08(0.6%) | | 9.28E-02 | 1.49E+08 | | 1.44E+08 $\mathrm{2s^{2}3d-2s^{2}2p}$ | $\mathrm{{}^{2}D-~{}^{2}P^{o}}$ | 3/2 - 3/2 | 145485 | 687.352 | 3.03E-01 | 4.71E+08(0.2%) | | 3.02E-01 | 4.70E+08 | | 4.60E+08 | | 5/2 - 3/2 | 145487 | 687.345 | 2.72E+00 | 2.82E+09(0.2%) | | 2.71E+00 | 2.82E+09 | | 2.76E+09 | | 3/2 - 1/2 | 145549 | 687.053 | 1.51E+00 | 2.35E+09(0.2%) | | 1.51E+00 | 2.35E+09 | | 2.30E+09 $\mathrm{2s^{2}3p-2s2p^{2}}$ | $\mathrm{{}^{2}P^{o}-~{}^{4}P}$ | 3/2 - 5/2 | 88681 | 1127.626 | 6.13E-07 | 2.16E+02(2.0%) | | 5.04E-07 | 1.79E+02 | | 1.80E+02 $\mathrm{2p^{3}-2s2p^{2}}$ | $\mathrm{{}^{4}S^{o}-~{}^{4}P}$ | 3/2 - 5/2 | 98973 | 1010.371 | 3.45E+00 | 1.70E+09(0.4%) | | 3.44E+00 | 1.70E+09 | | 1.71E+09 | | 3/2 - 3/2 | 99001 | 1010.083 | 2.30E+00 | 1.13E+09(0.4%) | | 2.30E+00 | 1.13E+09 | | 1.14E+09 | | 3/2 - 1/2 | 99023 | 1009.858 | 1.15E+00 | 5.67E+08(0.4%) | | 1.15E+00 | 5.67E+08 | | 5.69E+08 $\mathrm{2p^{3}-2s2p^{2}}$ | $\mathrm{{}^{2}D^{o}-~{}^{4}P}$ | 5/2 - 5/2 | 107407 | 931.030 | 4.41E-06 | 1.86E+03(21.1%) | | 3.52E-06 | 1.48E+03 | | 1.50E+03 | | 3/2 - 3/2 | 107441 | 930.740 | 1.04E-06 | 6.59E+02(22.2%) | | 8.44E-07 | 5.33E+02 | | 5.39E+02 $\mathrm{2s^{2}3s-2s2p^{2}}$ | $\mathrm{{}^{4}P^{o}-~{}^{4}P}$ | 3/2 - 5/2 | 123937 | 806.861 | 6.36E-01 | 6.13E+08(0.4%) | | 3.93E-01 | 3.79E+08 | | | | 1/2 - 3/2 | 123941 | 806.830 | 5.88E-01 | 1.13E+09(0.4%) | | 3.64E-01 | 7.02E+08 | | | | 1/2 - 1/2 | 123963 | 806.687 | 1.18E-01 | 2.27E+08(0.4%) | | 7.27E-02 | 1.40E+08 | | | | 3/2 - 3/2 | 123965 | 806.677 | 1.88E-01 | 1.81E+08(0.4%) | | 1.16E-01 | 1.12E+08 | | | | 5/2 - 5/2 | 123982 | 806.568 | 1.48E+00 | 9.53E+08(0.4%) | | 9.16E-01 | 5.90E+08 | | | | 3/2 - 1/2 | 123987 | 806.533 | 5.88E-01 | 5.67E+08(0.4%) | | 3.63E-01 | 3.51E+08 | | | | 5/2 - 3/2 | 124010 | 806.384 | 6.35E-01 | 4.09E+08(0.4%) | | 3.92E-01 | 2.53E+08 | | $\mathrm{2p^{3}-2s2p^{2}}$ | $\mathrm{{}^{2}P^{o}-~{}^{4}P}$ | 3/2 - 5/2 | 125924 | 794.125 | 2.69E-05 | 2.72E+04(3.2%) | | 7.77E-06 | 7.83E+03 | | 1.72E+02 | | 1/2 - 3/2 | 125953 | 793.947 | 4.39E-06 | 8.88E+03(0.3%) | | 3.80E-06 | 7.65E+03 | | 1.88E+02 | | 3/2 - 3/2 | 125953 | 793.947 | 1.60E-06 | 1.62E+03(8.6%) | | 8.35E-06 | 8.42E+03 | | 1.57E+03 | | 1/2 - 1/2 | 125975 | 793.808 | 1.51E-06 | 3.06E+03(10.2%) | | 1.73E-06 | 3.50E+03 | | 2.04E+01 | | 3/2 - 1/2 | 125975 | 793.808 | 1.44E-05 | 1.46E+04(2.7%) | | 2.48E-05 | 2.50E+04 | | 6.67E+02 $\mathrm{2s^{2}3p-2s2p^{2}}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}D}$ | 1/2 - 3/2 | 56791 | 1760.819 | 2.23E-01 | 4.09E+07(1.6%) | | 2.20E-01 | 4.08E+07 | | 4.37E+07 | | 3/2 - 3/2 | 56802 | 1760.473 | 4.45E-02 | 4.09E+06(1.6%) | | 4.40E-02 | 4.07E+06 | | 4.37E+06 | | 3/2 - 5/2 | 56805 | 1760.395 | 4.01E-01 | 3.68E+07(1.6%) | | 3.96E-01 | 3.67E+07 | | 3.94E+07 $\mathrm{2p^{3}-2s2p^{2}}$ | $\mathrm{{}^{2}D^{o}-~{}^{2}D}$ | 5/2 - 3/2 | 75528 | 1323.995 | 2.27E-01 | 3.33E+07(0.2%) | | 2.25E-01 | 3.28E+07 | | 3.50E+07 | | 5/2 - 5/2 | 75531 | 1323.951 | 3.16E+00 | 4.63E+08(0.1%) | | 3.13E+00 | 4.56E+08 | | 4.88E+08 | | 3/2 - 3/2 | 75534 | 1323.906 | 2.03E+00 | 4.45E+08(0.1%) | | 2.01E+00 | 4.39E+08 | | 4.69E+08 | | 3/2 - 5/2 | 75536 | 1323.862 | 2.30E-01 | 5.05E+07($<$0.05%) | | 2.27E-01 | 4.96E+07 | | 5.31E+07 $\mathrm{2s^{2}3s-2s2p^{2}}$ | $\mathrm{{}^{4}P^{o}-~{}^{2}D}$ | 1/2 - 3/2 | 92034 | 1086.549 | 1.55E-05 | 1.21E+04(1.0%) | | 2.45E-05 | 1.93E+04 | | | | 3/2 - 3/2 | 92058 | 1086.270 | 1.50E-05 | 5.88E+03(1.4%) | | 2.13E-05 | 8.39E+03 | | | | 3/2 - 5/2 | 92060 | 1086.241 | 9.46E-05 | 3.71E+04(0.9%) | | 1.50E-04 | 5.93E+04 | | | | 5/2 - 5/2 | 92105 | 1085.710 | 3.20E-06 | 8.38E+02(3.6%) | | 2.45E-06 | 6.45E+02 | | $\mathrm{2p^{3}-2s2p^{2}}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}D}$ | 1/2 - 3/2 | 94045 | 1063.313 | 1.97E+00 | 1.65E+09(0.9%) | | 1.73E+00 | 1.45E+09 | | 1.63E+09 | | 3/2 - 3/2 | 94045 | 1063.313 | 3.97E-01 | 1.66E+08(0.9%) | | 3.44E-01 | 1.44E+08 | | 1.64E+08 | | 3/2 - 5/2 | 94048 | 1063.284 | 3.55E+00 | 1.49E+09(0.8%) | | 3.11E+00 | 1.30E+09 | | 1.46E+09 $\mathrm{2s^{2}3p-2s2p^{2}}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}S}$ | 1/2 - 1/2 | 35230 | 2838.439 | 7.05E-01 | 3.06E+07(0.1%) | | 7.44E-01 | 3.27E+07 | | 3.32E+07 | | 3/2 - 1/2 | 35241 | 2837.541 | 1.41E+00 | 3.06E+07(0.1%) | | 1.49E+00 | 3.28E+07 | | 3.32E+07 $\mathrm{2p^{3}-2s2p^{2}}$ | $\mathrm{{}^{2}D^{o}-~{}^{2}S}$ | 3/2 - 1/2 | 53972 | 1852.780 | 1.78E-05 | 1.43E+03(1.8%) | | 1.55E-05 | 1.23E+03 | | 1.46E+03 $\mathrm{2p^{3}-2s2p^{2}}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}S}$ | 1/2 - 1/2 | 72484 | 1379.603 | 1.55E-05 | 5.92E+03(59.0%) | | 1.02E-01 | 3.87E+07 | | 2.50E+04 | | 3/2 - 1/2 | 72484 | 1379.603 | 1.78E-04 | 3.39E+04(31.6%) | | 2.13E-01 | 4.05E+07 | | 6.94E+04 $\mathrm{2s^{2}3p-2s2p^{2}}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}P}$ | 1/2 - 3/2 | 21058 | 4748.606 | 2.62E-03 | 2.37E+04(5.9%) | | 2.46E-03 | 2.31E+04 | | 2.52E+04 | | 3/2 - 3/2 | 21069 | 4746.093 | 1.35E-02 | 6.12E+04(5.6%) | | 1.26E-02 | 5.92E+04 | | 6.53E+04 | | 1/2 - 1/2 | 21100 | 4739.292 | 6.33E-03 | 5.76E+04(5.2%) | | 5.75E-03 | 5.42E+04 | | 6.12E+04 | | 3/2 - 1/2 | 21111 | 4736.789 | 1.87E-03 | 8.54E+03(6.5%) | | 1.90E-03 | 8.96E+03 | | 9.23E+03 $\mathrm{2p^{3}-2s2p^{2}}$ | $\mathrm{{}^{2}D^{o}-~{}^{2}P}$ | 5/2 - 3/2 | 39796 | 2512.814 | 2.68E+00 | 5.73E+07(2.7%) | | 2.64E+00 | 5.62E+07 | | 6.20E+07 | | 3/2 - 3/2 | 39801 | 2512.491 | 2.95E-01 | 9.45E+06(2.6%) | | 2.90E-01 | 9.28E+06 | | 1.02E+07 | | 3/2 - 1/2 | 39842 | 2509.881 | 1.49E+00 | 4.81E+07(2.6%) | | 1.47E+00 | 4.72E+07 | | 5.21E+07 $\mathrm{2s^{2}3s-2s2p^{2}}$ | $\mathrm{{}^{4}P^{o}-~{}^{2}P}$ | 1/2 - 3/2 | 56301 | 1776.149 | 4.38E-06 | 7.80E+02(0.4%) | | 9.53E-06 | 1.71E+03 | | | | 3/2 - 3/2 | 56325 | 1775.405 | 5.90E-05 | 5.25E+03(0.4%) | | 1.53E-04 | 1.37E+04 | | | | 1/2 - 1/2 | 56342 | 1774.845 | 4.89E-06 | 8.72E+02(0.5%) | | 1.55E-05 | 2.78E+03 | | | | 3/2 - 1/2 | 56366 | 1774.102 | 1.29E-05 | 1.15E+03(0.3%) | | 3.19E-05 | 2.88E+03 | | $\mathrm{2p^{3}-2s2p^{2}}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}P}$ | 1/2 - 3/2 | 58312 | 1714.890 | 5.79E-01 | 1.14E+08(0.3%) | | 8.67E-01 | 1.71E+08 | | 1.12E+08 | | 3/2 - 3/2 | 58312 | 1714.890 | 2.91E+00 | 2.88E+08(0.3%) | | 4.36E+00 | 4.31E+08 | | 2.83E+08 | | 1/2 - 1/2 | 58354 | 1713.674 | 1.16E+00 | 2.30E+08(0.3%) | | 1.74E+00 | 3.44E+08 | | 2.26E+08 | | 3/2 - 1/2 | 58354 | 1713.674 | 5.79E-01 | 5.74E+07(0.3%) | | 8.70E-01 | 8.62E+07 | | 5.63E+07 $\mathrm{2s^{2}3p-2s^{2}3s}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}S}$ | 1/2 - 1/2 | 15186 | 6584.700 | 1.03E+01 | 3.64E+07(0.2%) | | 1.03E+01 | 3.78E+07 | | 3.70E+07 | | 3/2 - 1/2 | 15197 | 6579.869 | 2.06E+01 | 3.65E+07(0.3%) | | 2.06E+01 | 3.79E+07 | | 3.71E+07 $\mathrm{2p^{3}-2s^{2}3s}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}S}$ | 1/2 - 1/2 | 52440 | 1906.916 | 2.19E-02 | 3.19E+06(0.1%) | | 6.60E-02 | 9.58E+06 | | 3.35E+06 | | 3/2 - 1/2 | 52440 | 1906.916 | 4.31E-02 | 3.15E+06(0.1%) | | 1.30E-01 | 9.45E+06 | | 3.32E+06 $\mathrm{2s^{2}3d-2s^{2}3p}$ | $\mathrm{{}^{2}D-~{}^{2}P^{o}}$ | 3/2 - 3/2 | 13813 | 7239.164 | 5.20E+00 | 6.94E+06(0.6%) | | 5.21E+00 | 6.73E+06 | | 7.09E+06 | | 5/2 - 3/2 | 13815 | 7238.415 | 4.68E+01 | 4.17E+07(0.6%) | | 4.69E+01 | 4.04E+07 | | 4.25E+07 | | 3/2 - 1/2 | 13824 | 7233.325 | 2.60E+01 | 3.48E+07(0.6%) | | 2.60E+01 | 3.37E+07 | | 3.55E+07 $\mathrm{2p^{3}-2s^{2}3d}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}D}$ | 1/2 - 3/2 | 23429 | 4268.202 | 9.99E-05 | 1.30E+03(27.0%) | | 3.11E-01 | 3.99E+06 | | 4.08E+04 | | 3/2 - 5/2 | 23427 | 4268.462 | 1.15E-04 | 7.49E+02(32.7%) | | 5.57E-01 | 3.58E+06 | | 3.33E+04 Table 11: Continued. (a)Tachiev & Fischer (2000); (b)Corrégé & Hibbert (2004). Table 12: Comparison of line strengths ($S$) and transition rates ($A$) with other theoretical results for C iii. The present values from the MCDHF/RCI calculations are given in the Babushkin(length) gauge. The wavenumber $\Delta E$ and wavelength $\lambda$ values are taken from the NIST database. The estimated uncertainties $dT$ of the transition rates are given as percentages in parentheses. Transition array | Mult. | $J_{u}-J_{l}$ | $\Delta E$ | $\lambda$ | MCDHF/RCI | | MCHF-BP(a) | | Grasp(b) ---|---|---|---|---|---|---|---|---|--- | | | ($\mathrm{cm^{-1}}$) | (Å) | $S$ (a.u. of a${}_{0}^{2}$e2) | $A$ (s-1) | | $S$ (a.u. of a${}_{0}^{2}$e2) | $A$ (s-1) | | $S$ (a.u. of a${}_{0}^{2}$e2) | $A$ (s-1) $\mathrm{2s2p-2s^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{1}S}$ | 1 - 0 | 102352 | 977.020 | 2.44E+00 | 1.77E+09(0.1%) | | 2.44E+00 | 1.77E+09 | | 2.38E+00 | 2.15E+09 $\mathrm{2s3p-2s^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{1}S}$ | 1 - 0 | 258931 | 386.203 | 3.06E-01 | 3.59E+09($<$0.05%) | | 3.06E-01 | 3.59E+09 | | 2.70E-01 | 3.16E+09 $\mathrm{2s3p-2s^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{1}S}$ | 1 - 0 | 259711 | 385.043 | 4.29E-05 | 5.08E+05(0.5%) | | 4.37E-05 | 5.18E+05 | | 1.17E-02 | 1.36E+08 $\mathrm{2p^{2}-2s2p}$ | $\mathrm{{}^{3}P-~{}^{3}P^{o}}$ | 1 - 2 | 85007 | 1176.370 | 1.32E+00 | 5.48E+08($<$0.05%) | | 1.32E+00 | 5.50E+08 | | 1.33E+00 | 5.95E+08 | | 0 - 1 | 85034 | 1175.987 | 1.05E+00 | 1.32E+09(0.1%) | | 1.05E+00 | 1.32E+09 | | 1.07E+00 | 1.43E+09 | | 2 - 2 | 85054 | 1175.711 | 3.95E+00 | 9.89E+08(0.1%) | | 3.95E+00 | 9.91E+08 | | 4.00E+00 | 1.07E+09 | | 1 - 1 | 85063 | 1175.590 | 7.90E-01 | 3.30E+08(0.1%) | | 7.89E-01 | 3.31E+08 | | 8.00E-01 | 3.58E+08 | | 1 - 0 | 85087 | 1175.263 | 1.05E+00 | 4.40E+08(0.1%) | | 1.05E+00 | 4.41E+08 | | 1.07E+00 | 4.78E+08 | | 2 - 1 | 85111 | 1174.933 | 1.32E+00 | 3.30E+08(0.1%) | | 1.32E+00 | 3.31E+08 | | 1.33E+00 | 3.59E+08 $\mathrm{2p^{2}-2s2p}$ | $\mathrm{{}^{1}D-~{}^{3}P^{o}}$ | 2 - 2 | 93429 | 1070.331 | 9.42E-05 | 3.13E+04(5.3%) | | 9.51E-05 | 3.17E+04 | | 3.74E-05 | 1.51E+04 | | 2 - 1 | 93485 | 1069.686 | 1.36E-05 | 4.52E+03(7.8%) | | 1.49E-05 | 4.97E+03 | | 3.37E-06 | 1.37E+03 $\mathrm{2p^{2}-2s2p}$ | $\mathrm{{}^{1}S-~{}^{3}P^{o}}$ | 0 - 1 | 130129 | 768.467 | 4.58E-07 | 2.06E+03(17.8%) | | 4.71E-07 | 2.12E+03 | | 2.01E-07 | 1.14E+03 $\mathrm{2s3s-2s2p}$ | $\mathrm{{}^{3}S-~{}^{3}P^{o}}$ | 1 - 2 | 185765 | 538.312 | 4.73E-01 | 2.05E+09(0.1%) | | 4.73E-01 | 2.05E+09 | | 5.03E-01 | 2.11E+09 | | 1 - 1 | 185822 | 538.149 | 2.83E-01 | 1.23E+09(0.1%) | | 2.84E-01 | 1.23E+09 | | 3.01E-01 | 1.27E+09 | | 1 - 0 | 185845 | 538.080 | 9.43E-02 | 4.09E+08(0.1%) | | 9.45E-02 | 4.10E+08 | | 1.00E-01 | 4.22E+08 $\mathrm{2s3s-2s2p}$ | $\mathrm{{}^{1}S-~{}^{3}P^{o}}$ | 0 - 1 | 194779 | 513.401 | 1.45E-08 | 2.17E+02(25.2%) | | 1.76E-08 | 2.64E+02 | | 2.46E-08 | 3.60E+02 $\mathrm{2s3d-2s2p}$ | $\mathrm{{}^{3}D-~{}^{3}P^{o}}$ | 1 - 2 | 217563 | 459.635 | 4.24E-02 | 2.95E+08($<$0.05%) | | 4.24E-02 | 2.95E+08 | | 4.24E-02 | 2.88E+08 | | 2 - 2 | 217564 | 459.633 | 6.36E-01 | 2.65E+09($<$0.05%) | | 6.36E-01 | 2.66E+09 | | 6.36E-01 | 2.60E+09 | | 3 - 2 | 217567 | 459.627 | 3.56E+00 | 1.06E+10($<$0.05%) | | 3.56E+00 | 1.06E+10 | | 3.56E+00 | 1.04E+10 | | 1 - 1 | 217620 | 459.516 | 6.36E-01 | 4.43E+09($<$0.05%) | | 6.36E-01 | 4.43E+09 | | 6.36E-01 | 4.33E+09 | | 2 - 1 | 217621 | 459.514 | 1.91E+00 | 7.96E+09($<$0.05%) | | 1.91E+00 | 7.97E+09 | | 1.91E+00 | 7.79E+09 | | 1 - 0 | 217643 | 459.466 | 8.47E-01 | 5.90E+09($<$0.05%) | | 8.47E-01 | 5.91E+09 | | 8.48E-01 | 5.77E+09 $\mathrm{2s3d-2s2p}$ | $\mathrm{{}^{1}D-~{}^{3}P^{o}}$ | 2 - 1 | 224092 | 446.245 | 4.65E-07 | 2.12E+03(0.3%) | | 2.80E-07 | 1.28E+03 | | 2.48E-07 | 1.14E+03 $\mathrm{2p^{2}-2s2p}$ | $\mathrm{{}^{3}P-~{}^{1}P^{o}}$ | 0 - 1 | 35073 | 2851.142 | 2.69E-06 | 2.36E+02(37.5%) | | 2.99E-06 | 2.65E+02 | | 1.85E-06 | 1.02E+02 | | 2 - 1 | 35149 | 2844.953 | 7.79E-05 | 1.37E+03(8.2%) | | 8.02E-05 | 1.43E+03 | | 2.85E-05 | 3.15E+02 $\mathrm{2p^{2}-2s2p}$ | $\mathrm{{}^{1}D-~{}^{1}P^{o}}$ | 2 - 1 | 43524 | 2297.578 | 4.11E+00 | 1.38E+08(0.5%) | | 4.11E+00 | 1.39E+08 | | 4.18E+00 | 1.34E+08 $\mathrm{2p^{2}-2s2p}$ | $\mathrm{{}^{1}S-~{}^{1}P^{o}}$ | 0 - 1 | 80167 | 1247.383 | 1.99E+00 | 2.10E+09($<$0.05%) | | 1.99E+00 | 2.10E+09 | | 2.24E+00 | 2.69E+09 $\mathrm{2s3s-2s2p}$ | $\mathrm{{}^{3}S-~{}^{1}P^{o}}$ | 1 - 1 | 135860 | 736.047 | 4.69E-07 | 7.92E+02(5.0%) | | 5.42E-07 | 9.17E+02 | | 3.76E-07 | 5.18E+02 $\mathrm{2s3s-2s2p}$ | $\mathrm{{}^{1}S-~{}^{1}P^{o}}$ | 0 - 1 | 144818 | 690.521 | 1.40E-01 | 8.59E+08(0.1%) | | 1.39E-01 | 8.54E+08 | | 1.77E-01 | 9.10E+08 $\mathrm{2s3d-2s2p}$ | $\mathrm{{}^{3}D-~{}^{1}P^{o}}$ | 1 - 1 | 167658 | 596.449 | 5.91E-07 | 1.88E+03(6.5%) | | 7.50E-07 | 2.39E+03 | | 4.14E-07 | 1.12E+03 | | 2 - 1 | 167659 | 596.446 | 9.30E-07 | 1.77E+03(6.2%) | | 4.89E-07 | 9.34E+02 | | 3.79E-07 | 6.15E+02 $\mathrm{2s3d-2s2p}$ | $\mathrm{{}^{1}D-~{}^{1}P^{o}}$ | 2 - 1 | 174130 | 574.281 | 2.93E+00 | 6.25E+09($<$0.05%) | | 2.92E+00 | 6.25E+09 | | 3.37E+00 | 6.40E+09 $\mathrm{2s3p-2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{3}P}$ | 1 - 2 | 121429 | 823.525 | 1.05E-05 | 1.27E+04(0.3%) | | 1.07E-05 | 1.29E+04 | | 7.12E-06 | 7.96E+03 | | 1 - 1 | 121476 | 823.202 | 1.76E-07 | 2.12E+02(1.0%) | | 8.87E-08 | 1.07E+02 | | 1.41E-05 | 1.58E+04 $\mathrm{2s3p-2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{3}P}$ | 1 - 2 | 122209 | 818.269 | 7.20E-04 | 8.84E+05(0.2%) | | 7.05E-04 | 8.65E+05 | | 6.36E-04 | 7.10E+05 | | 2 - 2 | 122222 | 818.181 | 2.17E-03 | 1.60E+06(0.2%) | | 2.15E-03 | 1.58E+06 | | 1.91E-03 | 1.28E+06 | | 0 - 1 | 122251 | 817.988 | 5.76E-04 | 2.13E+06(0.2%) | | 5.65E-04 | 2.08E+06 | | 5.08E-04 | 1.70E+06 | | 1 - 1 | 122256 | 817.950 | 4.29E-04 | 5.27E+05(0.2%) | | 4.26E-04 | 5.23E+05 | | 3.63E-04 | 4.06E+05 | | 2 - 1 | 122269 | 817.863 | 7.30E-04 | 5.39E+05(0.3%) | | 7.20E-04 | 5.31E+05 | | 6.43E-04 | 4.31E+05 | | 1 - 0 | 122285 | 817.758 | 5.78E-04 | 7.12E+05(0.2%) | | 5.69E-04 | 7.00E+05 | | 4.86E-04 | 5.44E+05 $\mathrm{2s3p-2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{1}D}$ | 1 - 2 | 113055 | 884.524 | 3.77E-01 | 3.66E+08($<$0.05%) | | 3.69E-01 | 3.59E+08 | | 7.16E-01 | 5.67E+08 $\mathrm{2s3p-2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{1}D}$ | 1 - 2 | 113835 | 878.464 | 6.05E-05 | 6.00E+04(0.1%) | | 6.32E-05 | 6.27E+04 | | 3.07E-02 | 2.43E+07 $\mathrm{2s3p-2p^{2}}$ | $\mathrm{{}^{1}P^{o}-~{}^{1}S}$ | 1 - 0 | 76411 | 1308.705 | 7.24E-02 | 2.16E+07(0.1%) | | 7.20E-02 | 2.15E+07 | | 1.54E-01 | 2.79E+07 $\mathrm{2s3p-2p^{2}}$ | $\mathrm{{}^{3}P^{o}-~{}^{1}S}$ | 1 - 0 | 77191 | 1295.482 | 1.04E-05 | 3.19E+03(0.2%) | | 1.05E-05 | 3.23E+03 | | 6.65E-03 | 1.20E+06 $\mathrm{2s3p-2s3s}$ | $\mathrm{{}^{1}P^{o}-~{}^{3}S}$ | 1 - 1 | 20718 | 4826.653 | 1.65E-03 | 9.96E+03(1.7%) | | 1.69E-03 | 1.02E+04 | | 4.51E-01 | 3.08E+06 $\mathrm{2s3p-2s3s}$ | $\mathrm{{}^{3}P^{o}-~{}^{3}S}$ | 0 - 1 | 21492 | 4652.775 | 3.57E+00 | 7.20E+07($<$0.05%) | | 3.57E+00 | 7.19E+07 | | 3.65E+00 | 7.44E+07 | | 1 - 1 | 21498 | 4651.548 | 1.07E+01 | 7.20E+07($<$0.05%) | | 1.07E+01 | 7.20E+07 | | 1.05E+01 | 7.14E+07 | | 2 - 1 | 21511 | 4648.720 | 1.78E+01 | 7.22E+07($<$0.05%) | | 1.78E+01 | 7.21E+07 | | 1.83E+01 | 7.46E+07 $\mathrm{2s3p-2s3s}$ | $\mathrm{{}^{1}P^{o}-~{}^{1}S}$ | 1 - 0 | 11761 | 8502.657 | 9.25E+00 | 1.02E+07($<$0.05%) | | 9.25E+00 | 1.02E+07 | | 8.59E+00 | 1.02E+07 $\mathrm{2s3p-2s3s}$ | $\mathrm{{}^{3}P^{o}-~{}^{1}S}$ | 1 - 0 | 12540 | 7973.871 | 1.42E-03 | 1.89E+03(2.6%) | | 1.45E-03 | 1.92E+03 | | 3.69E-01 | 4.33E+05 $\mathrm{2s3d-2s3p}$ | $\mathrm{{}^{3}D-~{}^{1}P^{o}}$ | 1 - 1 | 11079 | 9025.645 | 6.66E-04 | 6.10E+02(1.6%) | | 6.87E-04 | 6.35E+02 | | 1.86E-01 | 1.44E+05 | | 2 - 1 | 11080 | 9024.749 | 2.12E-03 | 1.17E+03(1.4%) | | 2.23E-03 | 1.24E+03 | | 5.55E-01 | 2.58E+05 $\mathrm{2s3d-2s3p}$ | $\mathrm{{}^{1}D-~{}^{1}P^{o}}$ | 2 - 1 | 17551 | 5697.496 | 1.94E+01 | 4.26E+07($<$0.05%) | | 1.94E+01 | 4.28E+07 | | 1.88E+01 | 5.15E+07 $\mathrm{2s3d-2s3p}$ | $\mathrm{{}^{3}D-~{}^{3}P^{o}}$ | 1 - 2 | 10286 | 9721.451 | 2.94E-01 | 2.16E+05($<$0.05%) | | 2.94E-01 | 2.18E+05 | | 3.00E-01 | 2.35E+05 | | 2 - 2 | 10287 | 9720.412 | 4.40E+00 | 1.95E+06($<$0.05%) | | 4.41E+00 | 1.97E+06 | | 4.51E+00 | 2.11E+06 | | 3 - 2 | 10290 | 9717.757 | 2.47E+01 | 7.79E+06($<$0.05%) | | 2.47E+01 | 7.87E+06 | | 2.52E+01 | 8.46E+06 | | 1 - 1 | 10299 | 9709.105 | 4.40E+00 | 3.25E+06($<$0.05%) | | 4.41E+00 | 3.29E+06 | | 4.32E+00 | 3.39E+06 | | 2 - 1 | 10300 | 9708.069 | 1.32E+01 | 5.86E+06($<$0.05%) | | 1.32E+01 | 5.92E+06 | | 1.30E+01 | 6.10E+06 | | 1 - 0 | 10305 | 9703.764 | 5.87E+00 | 4.35E+06($<$0.05%) | | 5.88E+00 | 4.39E+06 | | 6.01E+00 | 4.72E+06 $\mathrm{2s3d-2s3p}$ | $\mathrm{{}^{1}D-~{}^{3}P^{o}}$ | 2 - 1 | 16771 | 5962.446 | 3.11E-03 | 5.97E+03(0.4%) | | 3.23E-03 | 6.24E+03 | | 8.04E-01 | 2.22E+06 Table 12: Continued. (a)Tachiev & Fischer (1999); (b)Aggarwal & Keenan (2015). Table 13: Comparison of line strengths ($S$) and transition rates ($A$) with other theoretical results for C iv. The present values from the MCDHF/RCI calculations are given in the Babushkin(length) gauge. The wavenumber $\Delta E$ and wavelength $\lambda$ values are taken from the NIST database. The estimated uncertainties $dT$ of the transition rates are given as percentages in parentheses. Transition array | Mult. | $J_{u}-J_{l}$ | $\Delta E$ | $\lambda$ | MCDHF/RCI | | MCHF-BP(a) ---|---|---|---|---|---|---|--- | | | ($\mathrm{cm^{-1}}$) | (Å) | $S$ (a.u. of a${}_{0}^{2}$e2) | $A$ (s-1) | | $S$ (a.u. of a${}_{0}^{2}$e2) | $A$ (s-1) $\mathrm{2p-2s}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}S}$ | 1/2 - 1/2 | 64484 | 1550.772 | 9.68E-01 | 2.63E+08($<$0.05%) | | 9.68E-01 | 2.63E+08 | | 3/2 - 1/2 | 64591 | 1548.187 | 1.94E+00 | 2.65E+08($<$0.05%) | | 1.94E+00 | 2.65E+08 $\mathrm{3p-2s}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}S}$ | 1/2 - 1/2 | 320050 | 312.451 | 1.39E-01 | 4.63E+09($<$0.05%) | | 1.40E-01 | 4.64E+09 | | 3/2 - 1/2 | 320081 | 312.420 | 2.78E-01 | 4.62E+09($<$0.05%) | | 2.79E-01 | 4.63E+09 $\mathrm{3s-2p}$ | $\mathrm{{}^{2}S-~{}^{2}P^{o}}$ | 1/2 - 3/2 | 238257 | 419.714 | 2.07E-01 | 2.84E+09($<$0.05%) | | 2.07E-01 | 2.84E+09 | | 1/2 - 1/2 | 238365 | 419.525 | 1.03E-01 | 1.42E+09($<$0.05%) | | 1.03E-01 | 1.42E+09 $\mathrm{3d-2p}$ | $\mathrm{{}^{2}D-~{}^{2}P^{o}}$ | 3/2 - 3/2 | 260288 | 384.190 | 3.26E-01 | 2.91E+09($<$0.05%) | | 3.26E-01 | 2.91E+09 | | 5/2 - 3/2 | 260298 | 384.174 | 2.94E+00 | 1.75E+10($<$0.05%) | | 2.94E+00 | 1.75E+10 | | 3/2 - 1/2 | 260395 | 384.031 | 1.63E+00 | 1.46E+10($<$0.05%) | | 1.63E+00 | 1.46E+10 $\mathrm{4s-2p}$ | $\mathrm{{}^{2}S-~{}^{2}P^{o}}$ | 1/2 - 3/2 | 336756 | 296.951 | 2.74E-02 | 1.06E+09($<$0.05%) | | 2.75E-02 | 1.06E+09 | | 1/2 - 1/2 | 336864 | 296.856 | 1.37E-02 | 5.30E+08($<$0.05%) | | 1.38E-02 | 5.32E+08 $\mathrm{3p-3s}$ | $\mathrm{{}^{2}P^{o}-~{}^{2}S}$ | 1/2 - 1/2 | 17201 | 5813.582 | 6.10E+00 | 3.15E+07($<$0.05%) | | 6.11E+00 | 3.15E+07 | | 3/2 - 1/2 | 17232 | 5802.921 | 1.22E+01 | 3.17E+07($<$0.05%) | | 1.22E+01 | 3.16E+07 $\mathrm{3d-3p}$ | $\mathrm{{}^{2}D-~{}^{2}P^{o}}$ | 3/2 - 3/2 | 4798 | 20841.583 | 1.71E+00 | 9.54E+04($<$0.05%) | | 1.71E+00 | 9.51E+04 | | 5/2 - 3/2 | 4808 | 20796.074 | 1.54E+01 | 5.76E+05($<$0.05%) | | 1.54E+01 | 5.74E+05 | | 3/2 - 1/2 | 4829 | 20705.220 | 8.54E+00 | 4.87E+05($<$0.05%) | | 8.54E+00 | 4.85E+05 $\mathrm{4s-3p}$ | $\mathrm{{}^{2}S-~{}^{2}P^{o}}$ | 1/2 - 3/2 | 81266 | 1230.521 | 1.32E+00 | 7.16E+08($<$0.05%) | | 1.31E+00 | 7.14E+08 | | 1/2 - 1/2 | 81298 | 1230.043 | 6.57E-01 | 3.58E+08($<$0.05%) | | 6.57E-01 | 3.57E+08 Table 13: Continued. * • (a)Fischer et al. (1998).
# Measurements and analysis of response function of cold atoms in optical molasses Subhajit Bhar1<EMAIL_ADDRESS>Maheswar Swar1 Urbashi Satpathi2 <EMAIL_ADDRESS>Supurna Sinha1 Rafael D. Sorkin1,3 Saptarishi Chaudhuri1 Sanjukta Roy1<EMAIL_ADDRESS>1 Raman Research Institute, C. V. Raman Avenue, Sadashivanagar, Bangalore-560080, India. 2International Centre for Theoretical Sciences, Tata Institute of Fundamental Research, Bangalore 560089, India 3Perimeter Institute for Theoretical Physics, 31 Caroline Street North, Waterloo, ON N2L 2Y5, Canada ###### Abstract We report our experimental measurements and theoretical analysis of the position response function of a cloud of cold atoms residing in the viscous medium of an optical molasses and confined by a magneto-optical trap (MOT). We measure the position response function by applying a transient homogeneous magnetic field as a perturbing force. We observe a transition from a damped oscillatory motion to an over-damped relaxation, stemming from a competition between the viscous drag provided by the optical molasses and the restoring force of the MOT. Our observations are in both qualitative and quantitative agreement with the predictions of a theoretical model based on the Langevin equation. As a consistency check, and as a prototype for future experiments, we also study the free diffusive spreading of the atomic cloud in our optical molasses with the confining magnetic field of the MOT turned off. We find that the measured value of the diffusion coefficient agrees with the value predicted by our Langevin model, using the damping coefficient. The damping coefficient was deduced from our measurements of the position response function at the same temperature. ## I Introduction The response of a physical system to an applied force can reveal intrinsic characteristics of the system such as electric polarisability, impedance of an electronic circuit, magnetic susceptibility and optical conductivity Kubo (1966); Mazenko (2006); Balescu (1975); Kumar _et al._ (2020); Pan _et al._ (2020). In a similar context, but without the applied force, the study of diffusive behaviour can provide crucial information regarding transport properties Barkai _et al._ (2014); Beilin _et al._ (2010); Sagi _et al._ (2012). In recent years, the diffusion of a Brownian particle in the presence of quantum zero-point fluctuations was analysed in Sinha and Sorkin (1992); Satpathi _et al._ (2017) starting from the fluctuation-dissipation theorem (FDT) Kubo (1966); Balescu (1975). The key input to the analysis presented in Satpathi _et al._ (2017) is the position response function that describes how the particle reacts to an externally applied force. The specific response function employed in that paper was suggested by the model of a viscous medium. In the present work, we study a concrete experimental realization of such a model, by utilising a three-dimensional configuration of laser beams known as ‘optical molasses’ which enables cooling as well as viscous confinement of the atomic cloud. We find agreement (in a classical regime) with the type of response function that was assumed in Satpathi _et al._ (2017). Aside from the intrinsic interest of a direct measurement of the response function, our experiment lays the groundwork for future experiments that would access the deep quantum regime, where some of the most interesting effects discussed in Satpathi _et al._ (2017) would show up. In this paper, we demonstrate a method to measure the position response function of a cold atomic cloud in a MOT by temporarily subjecting it to a homogeneous magnetic field (transient oscillation method Kim _et al._ (2005)). We observe a transition from a damped oscillatory motion to an over- damped motion of the atomic cloud. This transition stems from a competition between the reactive spring-like force coming from the magneto-optical trap and the viscous drag due to the optical molasses. By turning off the MOT magnetic field, we are also able to study the spatial diffusion of the cold atoms in the viscous medium of an optical molasses, and we verify the Stokes-Einstein-Smoluchowski relation, as described in more detail in Appendix B. The motion of a Brownian particle can be analysed in terms of either the Fluctuation-Dissipation theorem or the Langevin equation. The FDT (which holds both classically and quantum mechanically) relates the spontaneous position and velocity fluctuations of a system in thermal equilibrium to its linear response to an external perturbing force. This allows the spontaneous fluctuations to be determined from the time-dependent response-function and vice versa. The Langevin equationKubo (1966); Mazenko (2006); Ford _et al._ (1988); Balescu (1975); Hohmann _et al._ (2017); Volpe and Volpe (2013); Deng _et al._ (2007); Kessler and Barkai (2012); Graham (2000); Majumdar and Orland (2015); Vulpiani and Baldovin (2020) (in its classical, generalised, and quantum forms) offers a complementary approach which relates the response function directly to the fluctuating forces that drive the position- fluctuations. In this paper, we have adopted the Langevin equation as our starting point, since it enables easy identification of all the forces coming into play. In applying it to the theoretical analysis of the dynamics governing the motion of the cold atoms, we have treated the MOT as an interesting example of an out of equilibrium system, and we have studied it from the point of view of statistical physics, rather than from the viewpoint of cold atom experimenters for whom it serves as a valuable and well-documented source of cold atoms. This type of analysis can be extended to a variety of physical situations where one is interested in the motion of particles in a viscous medium. The paper is organised as follows: In Sec. II, we briefly describe our experimental setup and methods for preparation and detection of the cold atoms. Sec. III is devoted to the position response function of the cold atomic cloud. In Sec. III.1, we set up the theoretical perspective. In Sec. III.2, we describe our method for measuring the response function of the cold atoms, and in Sec. III.3, we compare the analytical results with the experimental observations. In Sec. IV, we present some concluding remarks and future perspective. There are two appendices. The first supplements our treatment of the response function in the body of the paper. The second presents our results on the spatial diffusion of the cold atoms. ## II Preparation and detection of cold atoms Our experiment uses a cold atomic cloud of 87Rb atoms trapped in a MOT inside an ultra-high vacuum (UHV) region ($\sim 10^{-11}$ mbar) in a glass cell. A schematic diagram of the experimental set-up is shown in Fig. 1. The MOT is vapour-loaded from a Rb getter source. An external cavity diode laser (ECDL) serves as the cooling laser, the laser beam being 12 MHz red-detuned from the $5S_{1/2},F=2\rightarrow 5P_{3/2},F^{\prime}=3$ ($D_{2}$) transition of 87Rb. Another ECDL, the repump laser, is tuned to the transition, $5S_{1/2},F=1\rightarrow 5P_{3/2},F^{\prime}=2$, and used to optically pump the atoms back into the cooling cycle. This is a standard procedure in laser cooling experiments. The detuning and intensity of the cooling and repump beams are controlled by acousto-optic modulators (AOM). The restoring force required to confine the cold atoms is provided by a pair of current carrying coils in a near ideal anti-Helmholtz configuration. The fiber coupled laser beams are expanded to have a Gaussian waist diameter of 10 mm and combined in a non-polarizing cube beam splitter. Thereafter, the combined cooling and repump beams are split into three pairs of beams using a combination of half wave plates and polarizing cube beam splitters. Each of the cooling beams is sent through the UHV glass cell and retro-reflected via a quarter waveplate and a mirror. The incoming cooling beams are kept slightly converging so as to account for the losses in the optical elements and to ensure that any radiation-pressure imbalance between the incoming and the retro-reflected beam is eliminated. The cold atoms are detected by a time-of-flight absorption imaging technique, using a short ($\sim$ 100 $\mu$sec) pulse of weak, resonant linearly polarised laser light tuned to the $5S_{1/2},F=2\rightarrow 5P_{3/2},F^{\prime}=3$ transition. The shadow cast by the atoms is imaged onto an ICCD camera with a magnification factor of 0.4. In a typical run of the experiment, we trap and cool about 5 $\times$ 107 atoms at a temperature of around 150 $\mu$K. Figure 1: Schematic diagram of the experimental setup where a cold atomic cloud is produced in a MOT in a glass cell. A magnified view near the cold atomic cloud is shown in the inset. The trajectory of the cold atomic cloud is shown as a series of atomic clouds at successive positions in the XY plane. The cooling beams are retro-reflected using a mirror and a quarter-wave plate. The cylindrical magnetic coils produce the quadrupole magnetic field used for the MOT, and the square coils produce the homogeneous magnetic field used in measuring the response function. The detection of the atomic cloud is done by means of absorption imaging using an ICCD camera. ## III Response Function of the cold atoms ### III.1 The Langevin Equation Our starting point is the Langevin Equation.111In this paper we have used the Langevin equation as the starting point, unlike Ref Sinha and Sorkin (1992); Satpathi _et al._ (2017) where the FDT was used. Note, however, that the crucial input to the Langevin equation is the noise-noise correlation function given in Eq. (2), and this rests entirely on the FDT. In its fully quantum mechanical form, it reads $\small M\ddot{x}+\int_{-\infty}^{t}dt^{\prime}\alpha(t-t^{\prime})\dot{x}(t^{\prime})+kx=\zeta(t)+f(t)$ (1) where $M$ is the mass of the particle, $\alpha(t)$ is the dissipation kernel, and $\zeta(t)$ is the noise related to the dissipation-kernel via the Fluctuation Dissipation Theorem (FDT) Ford _et al._ (1988) as follows: $\begin{split}\small\langle\zeta(t)\rangle=&0\\\ \langle\left\\{\zeta(t),\zeta(t^{\prime})\right\\}\rangle\,=\,&\frac{2}{\pi}\int_{0}^{\infty}d\omega\,\hbar\,\omega\,\text{coth}\left(\frac{\hbar\omega}{2k_{B}T}\right)\\\ &\text{Re}[\tilde{\alpha}(\omega)]\text{cos}(\omega(t-t^{\prime}))\end{split}$ (2) The position-operator, $x(t)$, of the particle at any time $t$ can be obtained by solving these equations. The experiments reported here are at high enough temperatures that the noise can be treated classically. We can therefore take the $\hbar\to 0$ limit of the previous equation, to obtain the noise correlators in their classical form: $\begin{split}\small\langle\zeta(t)\rangle=&0\\\ \langle\zeta(t)\zeta(t^{\prime})\rangle\,=\,&\frac{2k_{B}T}{\pi}\int_{0}^{\infty}d\omega\,\text{Re}[\tilde{\alpha}(\omega)]\text{cos}(\omega(t-t^{\prime}))\end{split}$ (3) (This form of Langevin equation is sometimes termed “generalized” to indicate that the dissipation-kernel is not restricted to being a delta-function.) The last term on the left-hand side of Eq. (1) corresponds to a harmonic force characterized by a spring constant $k$. In our present problem $kx$ corresponds to the restoring force of the MOT. The term $f(t)$ on the right- hand side is a perturbing force, which in this experiment is an additional magneto-optical force induced by the transient homogeneous magnetic field used to measure the position-response function. Taking the expectation value of Eq. (1), and substituting $\langle\zeta(t)\rangle=0$, we obtain a deterministic equation for $\langle x(t)\rangle$, whose Fourier transform is $\small-M\omega^{2}\tilde{x}(\omega)-i\omega\tilde{\alpha}(\omega)\tilde{x}(\omega)+k\tilde{x}(\omega)=\tilde{f}(\omega)$ (4) (Here we have used $x$ to denote the mean position of the particle.) Eq. (4) can be re-expressed as $\small\tilde{x}(\omega)=\tilde{R}(\omega)\tilde{f}(\omega)~{},$ (5) where $\small\tilde{R}(\omega)=\frac{1}{[-M\omega^{2}-i\omega\alpha+k]}$ (6) is the position response function of the particle (in this case the cold atomic cloud) in the frequency domain. Here we have set $\tilde{\alpha}(\omega)=\alpha$, corresponding to the choice of an Ohmic bath to which the system is coupled. This choice is motivated by the optical molasses in the cold-atom experimental setup. In fact, an Ohmic bath is equivalent to a force proportional to the velocity with a fixed coefficient of proportionality or damping coefficient. The present experiment serves as a test of this theoretical model of the molasses. The position response function in the time domain is given by: $\small{R}(t)=\frac{1}{2\pi}\int{\tilde{R}(\omega)e^{-i\omega t}d\omega}$ (7) The position response function obtained thereby from Eq. (6) for the Ohmic bath is $\small R(t)\,=\frac{2}{\alpha_{c}}e^{-\frac{\alpha t}{2M}}\,\text{sinh}\left(\frac{\alpha_{c}t}{2M}\right)$ (8) where $\alpha_{c}=\sqrt{\alpha^{2}-4kM}$. There are three qualitatively distinct cases. For $\alpha^{2}>4kM$, $\alpha_{c}$ is real and one gets an overdamped motion of the cold atomic cloud. For $\alpha^{2}=4kM$, the motion of the cold atomic cloud is critically damped, while for $\alpha^{2}<4kM$, $\alpha_{c}$ is imaginary and the motion of the cold atomic cloud is a damped oscillation. For $k=0$, Eq. (8) reduces to the position response function used in Satpathi _et al._ (2017). (Note that in the MOT, $k$ is always nonzero due to the presence of the non-zero magnetic field gradient.) In Sec. III.3, we will compare the analytically calculated motion of the cold atomic cloud that follows from $R(t)$ via equation (10) below with the experimentally observed oscillatory and damped motions of the cloud. By taking a time derivative of $R(t)$, one can also get the velocity response function, $\small\dot{R}(t)=\frac{1}{\alpha_{c}M}e^{-\frac{\alpha t}{2M}}\left(\alpha_{c}\text{cosh}\left(\frac{\alpha_{c}t}{2M}\right)-\alpha\,\text{sinh}\left(\frac{\alpha_{c}t}{2M}\right)\right)$ (9) It’s an interesting fact that the position response function $R(t)$ can be inferred directly from the mean velocity induced by a homogeneous force whose time-dependence is that of a step-function. By definition, the mean displacement $\langle x(t)\rangle$ is related to $R(t)$ by the equation, $\small\langle x(t)\rangle=\int_{-\infty}^{t}{R(t-t^{\prime})f(t^{\prime})dt^{\prime}}$ (10) (“linear response theory” Sinha and Sorkin (1992); Satpathi _et al._ (2017)). On differentiation, this gives the expectation value $\langle v(t)\rangle$ of the velocity: $\small\langle v(t)\rangle=\int_{-\infty}^{t}{\dot{R}(t-t^{\prime})f(t^{\prime})dt^{\prime}}$ (11) Here $f(t)$ is the external perturbing force, which in our experiment takes the form of a “top-hat function”, $f(t)=f_{0}\,\theta(t+w)\,\theta(-t)$: $f(t)=\begin{cases}f_{0},&\text{for}\;-w<t<0\\\ 0,&\text{for}\;t\leq-w,\;t\geq 0\end{cases}$ (12) (In our experiment $f(t)$ is induced by a bias field. The temporal profile of this field, together with an analysis of how $f(t)$ depends on it, is given in detail in Appendix A.1.) Substituting into Eq. (11), we get: $\displaystyle\langle v(t)\rangle$ $\displaystyle=$ $\displaystyle f_{0}\int_{-w}^{0}{\dot{R}(t-t^{\prime})dt^{\prime}}$ (13) $\displaystyle=$ $\displaystyle-f_{0}\left(R(t)-R(t+w)\right)$ (14) $\small R(t)=-\frac{1}{f_{0}}\langle v(t)\rangle+R(t+w)$ (15) For $w\rightarrow\infty$, $R(t+w)\rightarrow R(\infty)=0$. (This assumes that the MOT is turned on. When it is turned off and only the molasses is present, $R(\infty)$ will be nonzero.) Therefore we get $R(t)=-\frac{1}{f_{0}}\langle v(t)\rangle$ (16) This simple relationship means that one can measure the position response function directly, simply by measuring the expectation value of the velocity. ### III.2 Measurement of the position response function The theoretical expression (8) for the position response function of the cold atoms contains two unknown parameters, the damping-coefficient $\alpha$ and the spring-constant $k$. In order to test the theoretical model that leads to (8), and at the same time determine the values of the parameters $\alpha$ and $k$, one needs to observe how the cloud of cold atoms moves in response to an external force. In our experiment we apply a homogeneous magnetic field (bias field), and then follow the motion of the cloud of cold atoms after the field is switched off. We first prepare the laser-cooled ${}^{87}\text{Rb}$ atoms in a MOT as described in Sec. II. After that, we apply a homogeneous bias field, $B_{b}$. This shifts the trap center to the zero of the new magnetic field. The cold atoms experience a force towards the new center, and equilibrate there within a short interval of time. After 5 sec, we turn the bias field off, and the cold atoms return to the initial trap center, following a trajectory from which the position response function can be inferred. To trace the trajectory, we record the position of the cold atoms at regular intervals of time after turning off the bias field. Fig. 2 is a schematic diagram of the sequence of events in the experiment. We capture and cool the atoms in the MOT from a Rb getter source with a loading time of 15 sec. The cooling beams, having a Gaussian cross-section with a waist size of $10$ mm, are red-detuned by $2.2$ $\Gamma$ from the $5S_{1/2},F=2\rightarrow 5P_{3/2},F^{\prime}=3$ transition, where $\Gamma=38.11(6)\times{10}^{6}s^{-1}$ ($2\pi\times 6.065(9)$ MHz) is the decay rate (natural line-width) of the 87Rb $D_{2}$ transition. Different values of the MOT magnetic field gradient were used in different sets of measurements. For the oscillatory motion shown in Fig. 3, the gradient was $18$ Gauss/cm; for the over-damped motion shown in Fig. 4, it was $3.5$ Gauss/cm. After the preparation stage, we apply the bias field for $5$ seconds (its amplitude being $3$ Gauss for Fig. 3 and $0.6$ Gauss for Fig. 4). Thereafter, we turn the bias field off and wait for a variable time t, after which we switch off the quadrupole magnetic field and the cooling and repumper laser beams simultaneously, and take an absorption image after allowing the cloud to move ballistically for a time $\tau_{tof}=1.2$ ms. The mean position of the cold atomic cloud is inferred by fitting a Gaussian to the column-density profile of the cloud. Figure 2: Timing sequence for measuring the response function of the cold ${}^{87}\text{Rb}$ atoms. We prepare the cold atomic cloud by loading the MOT for 15 sec. Thereafter, we apply a homogeneous bias magnetic field for 5 sec ($w$). After the bias field is switched off, the cooling beam detuning and intensity are changed by a variable amount in order to explore a range of values of $\alpha$. The ensuing motion of the cloud is monitored via time-of- flight absorption imaging. In our experiment, time-of-flight($\tau_{tof}$) is 1.2 ms, the detection gate time ($t_{g}$) is 1 ms, and the pulse width of the imaging beam is 100 $\mu$s. The time separation between successive absorption images ($t_{w}$) is 1 sec. ### III.3 Experimental Results and comparisons with the theory #### III.3.1 Motion of the cold atoms In our experimental runs, we allow the cloud to move ballistically for a time $\tau_{tof}=1.2$ ms after switching off the MOT light beams and the quadrupole field (the bias field having been switched off earlier, of course). This delay lets us acquire the absorption image of the cloud in a magnetic field-free environment. However it introduces a small correction to the mean position of the cloud given by $\langle x_{observed}\rangle\,=\langle x(t)\rangle\,+\tau_{tof}\,\langle v(t)\rangle$ (17) The graphs in Fig. 3 and Fig. 4 show the time variation of $\langle x_{observed}\rangle$ after the bias field is turned off. Each data point shown is the average of three experimental runs, and the error bar is the standard deviation of the mean position, measured as described in Sec. III.2. Figure 3: Position of the cold atoms as a function of time after the homogeneous bias field is switched off, illustrating the underdamped regime. Cooling beam detuning: $-2.2\,\Gamma$, total intensity: $I=16.91\,I_{sat}$; MOT Magnetic field gradient: 18 G/cm; bias magnetic field: 3 Gauss with its direction at an angle to the image plane. The solid line is the best fit between the experimental data and the theoretical prediction from Eq. (26) with $\alpha\,=\,(1.04\pm 0.04)\times 10^{-22}$ kg/sec. Inset: A test fit of the data to Eq. (18) yielded an initial estimate of $\alpha=(1.06\pm 0.24)\times 10^{-22}$ kg/sec. Figure 4: Position of the cold atoms as a function of time after the homogeneous bias field is switched off, illustrating the overdamped regime. Detuning: $-2.2\,\Gamma$, total intensity $I$: 9.73 $I_{sat}$; MOT Magnetic field gradient: 3.5 G/cm; bias magnetic field: 0.6 Gauss with its direction along one of the Cartesian axes in the image plane. The solid line exhibits the best fit between the data and Eq. (26) with $\alpha\,=\,(1.57\pm 0.46)\times 10^{-22}$ kg/sec. Inset: A test fit to Eq. (18) yielded an initial estimate of $\alpha=(1.58\pm 0.24)\times 10^{-22}$ kg/sec. Figure 5: Damping coefficient ($\alpha$) as a function of the light shift. MOT magnetic field gradient: 3.5 G/cm. The data was taken in the overdamped regime exemplified by Fig. 4. --- Figure 6: Position response function deduced directly from velocity for (a) oscillatory motion with $\alpha\,=\,(1.04\pm 0.04)\times 10^{-22}$ kg/sec and (b) damped motion with $\alpha\,=\,(1.57\pm 0.46)\times 10^{-22}$ kg/sec. In both the graphs, solid lines represent the theoretical prediction of $R(t)$ given in Eq. (8) and the shaded region shows the $68\%$ confidence band. The experimental data points correspond to the scaled velocities -$\frac{1}{f_{0}}\langle v(t)\rangle$ of the cold atoms. In Fig. 3, we observe an underdamped oscillatory motion of the cold atomic cloud where the MOT magnetic field gradient is $18$ Gauss/cm and the magnitude of the bias field is $3$ Gauss along the $x$-direction as shown in Fig. 1. In Fig. 4, we observe an over-damped motion of the cold atomic cloud where the MOT magnetic field gradient is $3.5\,$ Gauss/cm and the magnitude of the bias field is $0.6$ Gauss along the $x$-direction. The theoretical curves shown in Fig. 3 and Fig. 4, were obtained by fitting the experimental data to the prediction Eq. (26) (with due regard to Eq. (17)). In these fits, there is only a single fitting parameter: the damping coefficient $\alpha$. In the insets to Fig. 3 and Fig. 4, we have fitted the experimental data to the solution of a damped-harmonic oscillator, $\langle x(t)\rangle\,=A\,e^{\dfrac{(-\alpha+\alpha_{c})t}{2M}}\,+\,B\,e^{\dfrac{-(\alpha+\alpha_{c})t}{2M}}$ (18) without assuming anything further about the form of the position response function. Here $\alpha_{c}$ is defined as earlier, and $M=1.44316060(11)\times 10^{-25}$ kg is the mass of the ${}^{87}Rb$ atom. The fitting parameters in this case were $A$, $B$, and $\alpha$. From these fits, we obtained our initial estimates of $\alpha$. An approach based on a similar 3-parameter fit to the motion of an atom in a MOT was presented in Kim _et al._ (2005). However, while being a correct approximation, it does not capture the details of the external perturbing force in their entirety. In contrast, our approach based on the response function can be used for any form of the perturbing force. Hence it offers a theoretical model which is versatile and widely applicable for this class of experiments. As discussed in Sec. III.1, the cold atomic cloud shows an underdamped oscillatory motion or an over-damped motion in response to the applied bias field depending on whether $\alpha^{2}<4kM$ or $\alpha^{2}>4kM$ respectively, i.e. whether the restoring force due to the magneto-optical trapping overwhelms the viscous force due to the optical molasses or vice versa. As always, $\alpha$ here denotes the damping coefficient and $k$ the spring constant corresponding to the MOT. Both $\alpha$ and $k$ can be calculated from 1D Doppler cooling theory Lett _et al._ (1989); Chang _et al._ (2014) as, $\displaystyle\small\alpha\,$ $\displaystyle=\,4\hbar{\kappa}^{2}\,s_{0}\,\dfrac{2\absolutevalue{\delta}/\Gamma}{\bigg{(}1+2\,s_{0}+\frac{4\delta^{2}}{\Gamma^{2}}\bigg{)}^{2}}$ (19) $\displaystyle k$ $\displaystyle=g\dfrac{\mu_{B}\lambda}{h}\,\alpha\,\dfrac{\partial B_{m}}{\partial x}$ (20) where $\lambda$ is the wavelength and $\kappa=2\pi/\lambda$ is the wavenumber of the cooling beams, $\delta$ is the detuning of the cooling beams from the atomic transition, $\mu_{B}$ is the Bohr magneton, $\dfrac{\partial B_{m}}{\partial x}$ is the MOT magnetic field gradient and $s_{0}$ is the the saturation parameter of the cooling beams defined as $I/I_{sat}$ where $I$ is the total intensity of the cooling beams and $I_{sat}$ is the saturation intensity ($I_{sat}=1.6\,mW/cm^{2}$ for 87Rb $5S_{1/2},F=2\rightarrow 5P_{3/2},F^{\prime}=3$ transition and $\sigma^{\pm}$ polarised light). Hence, the damping coefficient $\alpha$ depends on the detuning and intensity of the cooling beams of the MOT, while the spring constant $k$ has an additional dependence on the magnetic field gradient. Using the simplest possible assumption that the fluorescence from the trapped atoms accurately gives the damping co-efficient in our fitting algorithm described above, we obtain a normalised mean square residual of 8.2% and 5% for the data presented in Fig. 3 and Fig. 4 respectively. However, in the presence of a gradient magnetic field in the MOT and for a Gaussian atom number spatial distribution in the atomic cloud and a Gaussian spatial intensity profile of the cooling laser beams, this simple assumption is likely to be inaccurate. Therefore, we kept $\alpha$ to be a free fitting parameter and obtained a normalised mean square residual to be 2.1% and 2.6% for the data presented in Fig. 3 and Fig. 4 respectively. This indicates that while the fluorescence measurements can give a reasonable estimate of the damping coefficient of cold atoms in the MOT, more accurate values of the damping coefficient can be found using experimental measurements which is modelled well using our theoretical description presented in this paper. #### III.3.2 Estimation of the damping coefficient ($\alpha$) in the MOT In Fig. 5, the damping coefficients ($\alpha$) obtained from fitting the experimental data with the analytical expression given in Eq. (26) and Eq. (30) are plotted against the light shifts, where the light shift ($\Delta$) is given by: $\small\Delta\,=\hbar\,\absolutevalue{\delta}\,\,\dfrac{I/I_{sat}}{1\,+\,4\delta^{2}/\Gamma^{2}}$ (21) where $\delta$ is the detuning of the cooling beam from the atomic transition and $\Gamma$ is the natural linewidth of the atomic transition having transition wavelength $\simeq$ 780 nm. #### III.3.3 Position response function from velocity In Fig. 6a and Fig. 6b, we show comparisons between the theoretically obtained position response functions given in Eq. (8) (solid lines) and the experimentally obtained scaled velocities -$\frac{1}{f_{0}}\langle v(t)\rangle$ (circle with error bars) for the motion of the atomic clouds given in Fig. 3 (oscillatory motion) and Fig. 4 (damped motion). Note that the scaled velocity data agrees very well with the curves for the response functions, confirming Eq. (16), which is indeed a very good approximation to the exact response function (in other words, the top hat function approximates the exact bias field and in turn the perturbing force well). As we vary the molasses parameters and the MOT’s magnetic field-gradient in the experiment, we observe both oscillatory and monotonic motions of the cloud’s centroid $\langle x(t)\rangle$, indicating a transition from an underdamped to an over-damped regime. We did not attempt to explore all the parameters (intensity, detuning, magnetic field gradient) in sufficient detail to pin down the exact transition point between the two regimes. Nevertheless, in the reasonably large parameter space that we have explored, the two regimes appear clearly, as does more generally the systematic variation of the response function with the experimental parameters of the MOT. ## IV Conclusion and Outlook In this work we have measured the position response function of the cold atoms in a MOT by subjecting them to a transient homogeneous magnetic field. We have tested theoretical predictions regarding the nature of the response function, and we have done extensive theoretical analysis and numerical modelling of our experimental observations. One of the significant outcomes of the study has been the verification of the functional form of the position response function which was used as input to a recent theoretical studySatpathi _et al._ (2017) of diffusion, not only in the classical domain dominated by thermal fluctuations, but also in the still- to-be-explored quantum domain where zero point fluctuations are the main driver of the diffusion foo . Our study has led to an interesting experimental observation of a transition from an oscillatory to an over-damped behaviour of the response function as a result of a competition between elastic and dissipative effects. We find a good agreement between our experimental measurements and the theoretical model of a particle moving in a viscous medium and confined by a harmonic-oscillator potential. These measurements can be readily extended to lighter atomic species compared to Rb such as Na and K so as to access a larger range of parameter space to observe a smooth transition of the response function from an under-damped to an over-damped behaviour. We also studied the spatial diffusion of the cold atoms in the optical molasses (Appendix B), observing a behaviour which is consistent with a theoretical model based on the Langevin equation. In particular, the measured value of the diffusion coefficient agreed with the value predicted by the Langevin model, using the damping coefficient deduced from our measurements of the position response function at the same temperature. One novelty of our theoretical analysis is the observation that the position response function can be obtained directly from the velocity (Eq. (16)) if the temporal variation of the perturbing force is a step function. This is confirmed by our experiment. Our theoretical analysis also points out that in the MOT where the magnetic field is linearly proportional to the distance from the centre, the magneto- optical force can be written as the gradient of the square of the local magnitude of the magnetic field as shown in Eq. (23) of the first Appendix. This relationship simplified the theoretical modelling of the perturbing force in our experiment as seen in Eq. (24). Our study provides a general framework to analyse the motion of a particle in optical molasses combined with a restoring force, such as in a MOT, ion-traps in the presence of cooling laser beams Fan _et al._ (2019), or ultra-cold atoms in optical lattices in the presence of additional optical molasses Sherson _et al._ (2010); Bakr _et al._ (2009). These and other similar experimental systems are of current interest in the context of quantum technology devices Amico _et al._ (2017). This study paves the way for exploring spatial diffusion of ultra-cold atoms in the quantum regime where zero point fluctuations dominate over thermal ones Sinha and Sorkin (1992); Satpathi _et al._ (2017); Das _et al._ (2020). The central questions addressed in this paper are rooted in non-equilibrium statistical mechanics, and the fact that we address them using the tools of cold atom physics makes this study inherently interdisciplinary in nature. In future we intend to experimentally measure and analyse the zero point fluctuation driven diffusion in the quantum domain that has been predicted in Sinha and Sorkin (1992); Satpathi _et al._ (2017). In that context we will expand our perspective beyond the classical Langevin Equation to a fully quantum mechanical formulation (quantum Langevin equation). ###### Acknowledgements. This work was partially supported by the Ministry of Electronics and Information Technology (MeitY), Government of India, through the Center for Excellence in Quantum Technology, under Grant4(7)/2020-ITEA. S.R acknowledges funding from the Department of Science and Technology, India, via the WOS-A project grant no. SR/WOS-A/PM-59/2019. This research was supported in part by Perimeter Institute for Theoretical Physics. Research at Perimeter Institute is supported in part by the Government of Canada through the Department of Innovation, Science and Economic Development Canada and by the Province of Ontario through the Ministry of Colleges and Universities. We acknowledge Hema Ramachandran, Meena M. S., Priyanka G. L. and RRI mechanical workshop for the instruments and assistance with the experiments. ## Appendix A Theoretical Modelling: Response function of the cold atoms ### A.1 Perturbing force on cold atoms subjected to a transient homogeneous magnetic field The temporal profile of the bias field used in our experiments is shown in Fig. 7. We fit this profile with the following equation: $\small B_{b}(t)\begin{cases}=0&\text{if $t\leq-w$}\\\\[15.0pt] =B_{0}\Bigg{(}1-e^{-\frac{t+\text{$w$}}{\tau_{1}}}\Bigg{)}&\text{if $-w\leq t\leq 0$}\\\\[15.0pt] =B_{0}\Bigg{(}1-e^{-\frac{\text{$w$}}{\tau_{1}}}\Bigg{)}\,e^{-\frac{t}{\tau_{2}}}&\text{if $t\geq 0$}\\\ \simeq B_{0}\,e^{-\frac{t}{\tau_{2}}}&(\text{for $w>>\tau_{1}$})\end{cases}$ (22) where $B_{0}$ is the magnitude and $w$ is the pulse width of the bias field, and where $\tau_{1}\,\text{ and }\,\tau_{2}$ are the rise time and fall time of the bias field. In our experiment $\tau_{1}\,\text{ and }\,\tau_{2}$ are $912\,\,\mu sec\text{ and }\,\,29.6\,\mu sec$ respectively. The approximation done in the last line of Eq. (22) is due to the fact that the time duration of the bias field ($w$ = 5 sec) is much larger than the $912\,\,\mu sec$ rise time of the bias field. The exact values of $\tau_{1}$ and $\tau_{2}$ depend on the design details of the fast switching circuit for the magnetic field coils in Helmholtz configuration producing the bias field Dedman _et al._ (2001). It is important to have a fast ‘switching off’ of the magnetic field so as to ensure that the measurements taken after switching off the magnetic field are not significantly affected by the time-constant $\tau_{2}$. In any case, we incorporate the effect of $\tau_{1}$ and $\tau_{2}$ on the motion of the atoms in our theoretical model. In a MOT, the $x$-component of the force on the cold atoms, which in Foot (2007) is expressed in terms of $x\,\partial{B_{m}}/\partial{x}$, can be recast as follows to show that the squared $B$-field acts like a potential energy for the atoms: $\begin{split}\small F_{MOT}\,&=-\alpha v-\,g\,\dfrac{\mu_{B}\lambda}{h}\,\alpha\,x\,\dfrac{\partial B_{m}}{\partial x}\\\\[5.0pt] &=-\alpha v-\,g\,\dfrac{\mu_{B}\lambda}{h}\,\alpha\,\dfrac{1}{2C_{m}}\,\dfrac{\partial B_{m}^{2}}{\partial x}\end{split}$ (23) Here, $g=g_{F^{\prime}}m_{F^{\prime}}-g_{F}m_{F}$ for transition between the hyperfine levels $|F,m_{F}\rangle$ and $|F^{\prime},m_{F}^{\prime}\rangle$, $\mu_{B}$ is the Bohr magneton, $\lambda$ is the wavelength of the cooling beams, $h$ is the Planck’s constant, and $\alpha$ is the damping coefficient. In the second line we have used that $B_{m}=C_{m}x$ with $C_{m}$ a constant, which implies that $x\dfrac{\partial B_{m}}{\partial x}\,=\,\dfrac{B_{m}}{C_{m}}\dfrac{\partial B_{m}}{\partial x}=\dfrac{1}{2C_{m}}\dfrac{\partial B_{m}^{2}}{\partial x}$ . In the presence of an additional bias field ($B_{b}$) along the negative $x$-direction, the force on an atom is given by (23) with the bias field added to $B_{m}$: $\begin{split}\small F_{net}\,&=-\alpha v-\,g\,\dfrac{\mu_{B}\lambda}{h}\,\alpha\,\dfrac{1}{2C_{m}}\,\dfrac{\partial(B_{m}-B_{b})^{2}}{\partial x}\\\\[5.0pt] &=-\alpha v-\,g\,\dfrac{\mu_{B}\lambda}{h}\,\alpha\,\dfrac{1}{2C_{m}}\,\bigg{(}\dfrac{\partial B_{m}^{2}}{\partial x}-\,2C_{m}B_{b}\bigg{)}\\\\[5.0pt] &=F_{MOT}\,+\,f(t)\end{split}$ (24) where we used that $\dfrac{\partial B_{b}}{\partial x}\,=\,0$. Therefore $f(t)=g\,\dfrac{\mu_{B}\lambda}{h}\,\alpha\,B_{b}$ (25) Figure 7: Temporal profile of the bias field. The black solid line is the experimental data recorded using a pick-up coil. Insets (a) and (b) show the growth and the decay, respectively, of the bias field as a function of time. After fitting the data using Eq. (22) we obtain $\tau_{1}=(912\pm 0.37)\,\mu sec\,\text{ and }\,\tau_{2}=(29.6\pm 0.056)\,\mu sec$. ### A.2 Mean displacement of the cold atoms Using the expression for the position response function in Eq. (8) and the perturbing force in Eq. (25), we get $\small\langle x(t)\rangle=Ae^{\dfrac{\left(-\alpha+\alpha_{c}\right)t}{2M}}+Be^{\dfrac{-\left(\alpha+\alpha_{c}\right)t}{2M}}+Ce^{\dfrac{-t}{\tau_{2}}}$ (26) where, $\displaystyle A$ $\displaystyle=$ $\displaystyle-\frac{2Mf_{0}}{\alpha_{c}}\left[\frac{\left(e^{\frac{\left(-\alpha+\alpha_{c}\right)w}{2M}}-1\right)}{\alpha-\alpha_{c}}-\frac{\tau_{1}\left(e^{\frac{\left(-\alpha+\alpha_{c}\right)w}{2M}}-e^{\frac{-w}{\tau_{1}}}\right)}{\alpha\tau_{1}-2M-\alpha_{c}\tau_{1}}\right.$ (27) $\displaystyle\left.+\dfrac{\tau_{2}\left(1-e^{\frac{-w}{\tau_{1}}}\right)}{\alpha\tau_{2}-2M-\alpha_{c}\tau_{2}}\right]$ $\displaystyle B$ $\displaystyle=$ $\displaystyle\frac{2Mf_{0}}{\alpha_{c}}\left[\frac{\left(e^{\frac{-\left(\alpha+\alpha_{c}\right)w}{2M}}-1\right)}{\alpha+\alpha_{c}}-\frac{\tau_{1}\left(e^{\frac{-\left(\alpha+\alpha_{c}\right)w}{2M}}-e^{\frac{-w}{\tau_{1}}}\right)}{\alpha\tau_{1}-2M+\alpha_{c}\tau_{1}}\right.$ (28) $\displaystyle\left.+\frac{\tau_{2}\left(1-e^{\frac{-w}{\tau_{1}}}\right)}{\alpha\tau_{2}-2M+\alpha_{c}\tau_{2}}\right]$ $\displaystyle C$ $\displaystyle=$ $\displaystyle\frac{4M\tau_{2}^{2}f_{0}\left(1-e^{\frac{-w}{\tau_{1}}}\right)}{\left(4M^{2}+\tau_{2}^{2}(\alpha^{2}-\alpha_{c}^{2})-4M\alpha\tau_{2}\right)}$ (29) Here, $f_{0}=g\,\dfrac{\mu_{B}\lambda}{h}\,\alpha\,B_{0}$ from Eq. (25). The mean velocity can also be obtained by taking a time derivative of $\langle x(t)\rangle$ in Eq. (26), $\displaystyle\langle v(t)\rangle$ $\displaystyle=$ $\displaystyle\frac{(\alpha_{c}-\alpha)A}{2M}e^{\dfrac{\left(-\alpha+\alpha_{c}\right)t}{2M}}-\frac{(\alpha+\alpha_{c})B}{2M}$ (30) $\displaystyle e^{\dfrac{-\left(\alpha+\alpha_{c}\right)t}{2M}}-\frac{C}{\tau_{2}}e^{\dfrac{-t}{\tau_{2}}}$ Note that for negligible $\tau_{1},\tau_{2}$ and for infinite width, i.e. $w\rightarrow\infty$, using Eq. (27), Eq. (28) and Eq. (29), $\frac{(\alpha_{c}-\alpha)A}{2M}\rightarrow-\frac{f_{0}}{\alpha_{c}},\frac{(\alpha+\alpha_{c})B}{2M}\rightarrow\,-\frac{f_{0}}{\alpha_{c}},\frac{C}{\tau_{2}}\rightarrow 0$ then, using Eq. (30) $\langle v(t)\rangle\rightarrow-f_{0}R(t)$ ($R(t)$ is given by Eq. (8)) and thus Eq. (16) is satisfied. ## Appendix B Spatial diffusion of cold atoms We study the diffusive behaviour of the cold atoms in the viscous medium provided by our optical molasses, exploring different temperatures as the atomic cloud is cooled to lower temperatures via sub-Doppler cooling. When the restoring force produced by the MOT magnetic field is absent, we are in the $k=0$ regime of the Langevin equation which defines our theoretical model. We continue to assume that the dissipation kernel $\alpha(t)$ is simply a delta-function in time, or equivalently that the force exerted on an atom by the optical molasses is $F_{OM}\,=\,-\alpha v$ (31) where v is the velocity of the atom. For consistency, one would hope that essentially the same value of $\alpha$ would explain both the position response function studied above and the diffusive spreading studied here. As is well known, the mean-square distance traveled by the diffusing atom can be determined from the Langevin equation together with the noise-correlator, i.e. from equations (1) and (3). The predicted time-dependence of the spreading depends on how the observation time $\tau$ compares with the “relaxation time” $M/\alpha$. When $\tau\gg M/\alpha$ one finds the familiar Brownian motion, with diffusion coefficient $D$ given by the Stokes-Einstein- Smoluchowski relation: $D=\frac{k_{B}T}{\alpha}$ (32) where $k_{B}$ is the Boltzmann constant and $T$ is the temperature of the cloud. However, for $\tau\ll M/\alpha$ one finds that the mean-square distance travelled grows like $t^{3}$ rather than $t$. In our experiment, the observation time of 20 ms is about 20 times bigger than $M/\alpha\sim 1$ ms. Although this is not enormously greater than unity, it seems sufficiently big for us to ignore the short-time crossover to $t^{3}$ spreading. We have therefore fitted the data under the assumption that we are in the regime of Brownian motion (Wiener process). Figure 8: The plot shows the atomic cloud size expanding in an optical molasses at a temperature of around $120\mu K$ . The solid line is a fit to the experimental data using Eq. (33) . To observe the diffusive spreading of the atoms in the cold atomic cloud, we first loaded the MOT from the background Rb vapour. Thereafter, the MOT magnetic field was switched off, and the cloud was allowed to diffuse in the presence of the cooling laser beams forming the optical molasses, but still in the absence of the MOT magnetic field. In the Brownian motion approximation, an atomic cloud of initial size of $\evaluated{(\Delta r)^{2}}_{t=0}$ expands to a size of $\evaluated{(\Delta r)^{2}}_{t=\tau}$ in time $\tau$ according to the relation: $\small\evaluated{(\Delta r)^{2}}_{t=\tau}\,=\evaluated{(\Delta r)^{2}}_{t=0}\,+\,4\,D\,\tau,$ (33) where, $\Delta r$ is the rms width of the cold atomic cloud. We obtained ${(\Delta r)}^{2}$ directly from the column density profile of the absorption image at time $t$. In other similar experiments Hodapp _et al._ (1995), the density profile was fitted to a Gaussian distribution, whereas the $(\Delta r)^{2}$ shown in Fig. 8 was obtained directly from the absorption images without assuming Gaussianity. This additional generality could become important in the quantum regime of logarithmic spreading, for which the analysis of Satpathi _et al._ (2017) furnishes $(\Delta r)^{2}$ but not the full probability distribution of $\Delta r$. (We know of no proof that the latter will be Gaussian when the diffusion is not classical.) Eq. (32) relates the damping coefficient $\alpha$ to the diffusion coefficient $D$ and thereby allows us to check for consistency between our direct measurement of D (Fig. 8) and the value of the $\alpha$ deduced from our earlier measurements of the position response function. For a temperature of around $120\mu K$ of the cold atomic cloud, the diffusion coefficient obtained from the measurement of the diffusive spreading of the atomic cloud was $(1.01\pm 0.15)\times 10^{-5}$ m2/s yielding a value of $(1.58\pm 0.25)\times 10^{-22}$ kg/s for $\alpha$. For the same temperature, the value of $\alpha$ obtained from the measurement of the position response function was $(1.57\pm 0.46)\times 10^{-22}$ kg/s. The agreement could not be better. ## References * Kubo (1966) R. Kubo, Reports on Progress in Physics 29, 306 (1966). * Mazenko (2006) G. F. Mazenko, _Nonequilibrium Statistical Mechanics_ (Wiley, 2006) p. 478. * Balescu (1975) R. Balescu, _Equilibrium and Non-Equilibrium Statistical Mechanics_, A Wiley interscience publication (Wiley, 1975). * Kumar _et al._ (2020) A. Kumar, M. Rodriguez-Vega, T. Pereg-Barnea, and B. Seradjeh, Phys. Rev. B 101, 174314 (2020). * Pan _et al._ (2020) L. Pan, X. Chen, Y. Chen, and H. Zhai, Nature Physics 16, 767 (2020). * Barkai _et al._ (2014) E. Barkai, E. Aghion, and D. A. Kessler, Phys. Rev. X 4, 021036 (2014). * Beilin _et al._ (2010) L. Beilin, E. Gurevich, and B. Shapiro, Phys. Rev. A 81, 033612 (2010). * Sagi _et al._ (2012) Y. Sagi, M. Brook, I. Almog, and N. Davidson, Phys. Rev. Lett. 108, 093002 (2012). * Sinha and Sorkin (1992) S. Sinha and R. D. Sorkin, Physical Review B 45, 8123 (1992), arXiv:0506196 [cond-mat] . * Satpathi _et al._ (2017) U. Satpathi, S. Sinha, and R. D. Sorkin, Journal of Statistical Mechanics: Theory and Experiment 2017, 123105 (2017), arXiv:1702.06273 . * Kim _et al._ (2005) K. Kim, K.-H. Lee, M. Heo, H.-R. Noh, and W. Jhe, Physical Review A 71, 053406 (2005). * Ford _et al._ (1988) G. W. Ford, J. T. Lewis, and R. F. O’Connell, Phys. Rev. A 37, 4419 (1988). * Hohmann _et al._ (2017) M. Hohmann, F. Kindermann, T. Lausch, D. Mayer, F. Schmidt, E. Lutz, and A. Widera, Physical Review Letters 118, 263401 (2017). * Volpe and Volpe (2013) G. Volpe and G. Volpe, American Journal of Physics 81, 224 (2013), https://doi.org/10.1119/1.4772632 . * Deng _et al._ (2007) Y. Deng, J. Bechhoefer, and N. R. Forde, Journal of Optics A: Pure and Applied Optics 9, S256 (2007). * Kessler and Barkai (2012) D. A. Kessler and E. Barkai, Phys. Rev. Lett. 108, 230602 (2012). * Graham (2000) R. Graham, Journal of Statistical Physics 101, 243 (2000). * Majumdar and Orland (2015) S. N. Majumdar and H. Orland, Journal of Statistical Mechanics: Theory and Experiment 2015, P06039 (2015). * Vulpiani and Baldovin (2020) A. Vulpiani and M. Baldovin, Journal of Statistical Mechanics: Theory and Experiment 2020, 014003 (2020). * Lett _et al._ (1989) P. D. Lett, W. D. Phillips, S. L. Rolston, C. E. Tanner, R. N. Watts, and C. I. Westbrook, Journal of the Optical Society of America B 6, 2084 (1989). * Chang _et al._ (2014) R. Chang, A. L. Hoendervanger, Q. Bouton, Y. Fang, T. Klafka, K. Audo, A. Aspect, C. I. Westbrook, and D. Clément, Phys. Rev. A 90, 063407 (2014). * (22) In the experimental setup, the spring constant $k$ of the MOT is always finite. Therefore one cannot achieve the $k=0$ limit of the position response function in Eq. (8) which corresponds to the position response function used in Satpathi _et al._ (2017). Nevertheless, we get a very good agreement between our theory and experimental data for all finite values of $k$. * Fan _et al._ (2019) M. Fan, C. A. Holliman, A. L. Wang, and A. M. Jayich, Phys. Rev. Lett. 122, 223001 (2019). * Sherson _et al._ (2010) J. F. Sherson, C. Weitenberg, M. Endres, M. Cheneau, I. Bloch, and S. Kuhr, Nature 467, 68 (2010). * Bakr _et al._ (2009) W. S. Bakr, J. I. Gillen, A. Peng, S. Fölling, and M. Greiner, Nature 462, 74 (2009). * Amico _et al._ (2017) L. Amico, G. Birkl, M. Boshier, and L.-C. Kwek, New Journal of Physics 19, 020201 (2017). * Das _et al._ (2020) A. Das, A. Dhar, I. Santra, U. Satpathi, and S. Sinha, Phys. Rev. E 102, 062130 (2020). * Dedman _et al._ (2001) C. J. Dedman, K. G. H. Baldwin, and M. Colla, Review of Scientific Instruments 72, 4055 (2001), https://doi.org/10.1063/1.1408935 . * Foot (2007) C. J. Foot, _Atomic Physics_ , Oxford master series in atomic, optical, and laser physics (Oxford University Press, Oxford, 2007). * Hodapp _et al._ (1995) T. W. Hodapp, C. Gerz, C. Furtlehner, C. I. Westbrook, W. D. Phillips, and J. Dalibard, Applied Physics B 60, 135 (1995).
# Softness, anomalous dynamics, and fractal-like energy landscape in model cell tissues Yan-Wei Li Division of Physics and Applied Physics, School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore 637371, Singapore School of Physics, Beijing Institute of Technology, Beijing 100081, China Leon Loh Yeong Wei Division of Physics and Applied Physics, School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore 637371, Singapore Matteo Paoluzzi Departament de Física de la Matèria Condensada, Universitat de Barcelona, C. Martí Franquès 1, 08028 Barcelona, Spain Massimo Pica Ciamarra<EMAIL_ADDRESS>Division of Physics and Applied Physics, School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore 637371, Singapore CNR–SPIN, Dipartimento di Scienze Fisiche, Università di Napoli Federico II, I-80126, Napoli, Italy ###### Abstract Epithelial cell tissues have a slow relaxation dynamics resembling that of supercooled liquids. Yet, they also have distinguishing features. These include an extended short-time sub-diffusive transient, as observed in some experiments and recent studies of model systems, and a sub-Arrhenius dependence of the relaxation time on temperature, as reported in numerical studies. Here we demonstrate that the anomalous glassy dynamics of epithelial tissues originates from the emergence of a fractal-like energy landscape, particles becoming virtually free to diffuse in specific phase space directions up to a small distance. Furthermore, we clarify that the stiffness of the cells tunes this anomalous behaviour, tissues of stiff cells having conventional glassy relaxation dynamics. ## I Introduction Cells in tissues rearrange in many biological processes, including embryonic development, wound healing, and tumour metastases Manli _et al._ (2012); Poujade _et al._ (2007); Basan _et al._ (2013); Trepat and Sahai (2018). The resulting tissue dynamics is slow and heterogeneous in both space and time, as cells in tissues may spend long transients in the cages formed by their neighbours, before relaxing through cooperative processes Schötz _et al._ (2013); Angelini _et al._ (2011); Bi _et al._ (2016). These observations evidence strong similarities between the dynamics of tissues Angelini _et al._ (2010); Schötz _et al._ (2013); Angelini _et al._ (2011); Bi _et al._ (2016); Sussman _et al._ (2018); Jordan _et al._ (2011); Tanaka and Ogishima (2015); Kalluri and Weinberg (2009) and that of supercooled liquids Debenedetti and Stillinger (2001); Binder and Kob (2011). However, the investigation of the relaxation dynamics of model cell tissue Sussman _et al._ (2018); Sadhukhan and Nandi (2020) where temperature-like stochastic forces drive particles, revealed distinct features not shared by ordinary supercooled liquids. In particular, the relaxation time was observed to grow as $\tau_{\alpha}\propto\exp(-E/T^{x})$ with $x<1$, a sub-Arrhenius behavior markedly distinct from the strong, $x=1$, or a super-Arrhenius, $x>1$, behavior of supercooled liquids Angell (1995). Besides, while in supercooled liquids particles do not diffuse during the transient caging-regime, the mean square displacement scaling as $t^{\alpha}$ with $\alpha\to 0$ at low temperature, cells in tissue may transiently exhibit a sub-diffusive behaviour, during which $1>\alpha>0$ is approximately constant. Experimental findings are compatible with this sub-diffusive behavior Rogers (2007); Schötz _et al._ (2013); Nixon-Abell _et al._ (2016); Armiger _et al._ (2018); Fodor _et al._ (2018), which occurs on a short time scale where the dynamics is likely to be dominated by thermal effects, rather than by self-propulsion of cells. These results are indicative of unusual features of the energy landscape of tissues, which have not yet been rationalized. It has not even been ascertained if the anomalous glassy dynamics ubiquitously occur in epithelial cell tissues, or rather if it depends on the mechanical properties of the cell, recently correlated to their geometrical features Park _et al._ (2015). In this paper, we show that the distinctive sub-diffusive and sub-Arrhenius glassy relaxation only occur in a tissue of highly deformable cells, while conversely a conventional glassy relaxation dynamics occurs. Furthermore, we rationalize that these distinct features signal the existence, in the energy landscape of highly-deformable epithelial tissues, of selected phase space directions along which the system moves almost freely, for short distances. Displacements along these phase space directions trigger cell rearrangement processes, or T1 transitions Weaire and Rivier (1984); Staple _et al._ (2010); Bi _et al._ (2014), that have a negligible energy cost. The physical mechanism leading to sub-diffusion establishes an unexpected connection between the dynamics of cell tissues and that of particles diffusing in random media, and indicates that the energy landscape of tissues is locally fractal- like as that of the Lorentz model close to the percolation threshold van Beijeren (1982); Höfling _et al._ (2006); Zeitz _et al._ (2017). ## II Voronoi model ### II.1 Numerical model We investigate the dynamics of a model of epithelial tissues Farhadifar _et al._ (2007); Staple _et al._ (2010); Bi _et al._ (2015); Manning _et al._ (2010); Fletcher _et al._ (2014); Bi _et al._ (2016), where the configurational degrees of freedom are the centers of mass of the cells, $\\{\mathbf{r}_{i}\\}$, and the shape of cell $i$ is that of the Voronoi cell centered in $\mathbf{r}_{i}$. Biological considerations Farhadifar _et al._ (2007); Staple _et al._ (2010); Bi _et al._ (2015); Manning _et al._ (2010); Fletcher _et al._ (2014); Bi _et al._ (2016); Moshe _et al._ (2018); Giavazzi _et al._ (2018) indicate that the mechanical energy of a cell depends on its area $A_{i}$ and perimeter $P_{i}$, $E_{i}=K_{A}(A_{i}-A_{0}^{i})^{2}+K_{P}(P_{i}-P_{0}^{i})^{2}$, where $A_{0}^{i}$ and $P_{0}^{i}$ are preferred values, while $K_{A}$ and $K_{P}$ are area and perimeter elastic constants. Hence, the dimensionless energy functional is $e=\sum_{i=1}^{N}[(a_{i}-a_{0}^{i})^{2}+r^{-1}(p_{i}-p_{0}^{i})^{2}],$ (1) where the sum runs over all $N=1024$ cells of the system, $a_{i}=A_{i}/l^{2}$ and $p_{i}=P_{i}/l$ with $l$ the unit of length which we have chosen so that $\langle a_{i}\rangle=1$. The preferred area $a_{0}^{i}$ is uniformly distributed in the range $0.8$–$1.2$, to avoid crystallization, while the preferred perimeter is fixed to $p_{0}^{i}=p_{0}\sqrt{a_{0}^{i}}$, with $p_{0}$ the target shape index. The non-dimensional energy then depends on the inverse perimeter modulus, $r=K_{A}l^{2}/K_{P}$, we fix to 1, and on $p_{0}$. This parameter determines the cell deformability, higher values of $p_{0}$ corresponding to more deformable cells Bi _et al._ (2015); Li and Ciamarra (2018). Simulations are performed using periodic boundary conditions. ### II.2 Connection with experiments The use of this model to simulate epithelial tissues poses two challenges. First, one would need to determine the values of the model parameters. While the physical and biological interpretation of the model’s parameters is clear Farhadifar _et al._ (2007); Staple _et al._ (2010), these have never been experimentally determined. Estimating these parameters, and in particular the elastic constants, would require probing the interaction of cells in a tissue, and take into account that the model focuses on a two-dimensional representation. Secondly, one needs to drive the cells via active forces. This is a issue as the features of the active forces inducing the dynamics of cell tissues are still unclear, and indeed different models have been proposed in the literature Barton _et al._ (2017); Bi _et al._ (2016). Specifically, the issue is whereas the biological process leading cell motion also induces aligning interactions between the cells Barton _et al._ (2017). To tackle these issues, we perform simulations in the NVT ensemble, where $T$ should be interpreted as an effective temperature. Specifically, we integrate the equations of motion via the Verlet algorithm, and fix the temperature using a Langevin thermostat Allen (1987). Furthermore, we relate the model parameters to experimental values considering their effect on the dynamics. While the diffusion coefficient of cells in epithelial tissues vary greatly with the control parameters, a typical order of magnitude estimate is $D\simeq 10^{-2}\mu m^{2}/{\rm min}$ (e.g, Armiger _et al._ (2018); Dieterich _et al._ (2008)). We investigate effective temperature values leading to a diffusion coefficients, which we estimate from the mean square displacement of our numerical model, of order $D_{\rm sim}=10^{-3}l^{2}/\tau_{0}$, with $l$ and $\tau$ are length and time units. This is also an order of magnitude estimate, as the diffusion depends on the temperature. Equating the numerical and the experimental diffusion coefficient, and fixing our length unit to the typical cell size, of order $10\mu m$, we estimate our time unit to be approximately $0.06$s. In the following, we investigate the dynamics for $t\leq 10^{5}\tau_{0}\simeq 4h$. This is a short time scale with respect to that of biological processes such as cell reproduction and apoptosis, which affect cell size Puliafito _et al._ (2012); Straetmans and Khain (2019), and with respect to the time scale of cell volume fluctuations Zehnder _et al._ (2015), which we therefore neglect. ## III Conventional and anomalous glassy dynamics: $p_{0}$ dependence Figure 1: Time dependence of the mean square displacement (a), of its log- slope (c) and the self-intermediate scattering function (e), at $p_{0}=3.0$. (b), (d) and (f) show the same quantities at $p_{0}=3.81$. Open symbols in (a)-(d) are results obtained via overdamped simulations, for the indicated low-temperatures values. In all panels, black dots mark the relaxation time. At zero temperature, on increasing the target shape index, this model exhibits a sharp crossover for $p_{0}\simeq 3.81$, which is reminiscent of a rigidity transition Bi _et al._ (2015); Li and Ciamarra (2018); Sussman and Merkel (2018). Here, we compare the relaxation dynamics at $p_{0}=3.0$ and at $p_{0}=3.81$, respectively in the solid phase and close to the crossover. We investigate the MSD $\left\langle\Delta r^{2}(t)\right\rangle=\left\langle\frac{1}{N}\sum_{i=1}^{N}\Delta\mathbf{r}_{i}(t)^{2}\right\rangle$, with $\Delta\mathbf{r}_{i}(t)$ displacement of particle $i$ at time $t$, its log-slope $\Delta(t)=\mathrm{d}\left(\ln\left\langle\Delta r^{2}(t)\right\rangle\right)/\mathrm{d}(\ln(t))$, and the self-intermediate scattering function (ISF) $F_{s}(q,t)=\left\langle\frac{1}{N}\sum_{j=1}^{N}e^{i\mathbf{q}\cdot\Delta\mathbf{r}_{j}(t)}\right\rangle$ with $q=|\mathbf{q}|$ the wavenumber of the first peak of the static structure factor. We define the relaxation time $\tau_{\alpha}$ as $F_{s}(q,\tau_{\alpha})=e^{-1}$. Stiff cells ($p_{0}=3.0$) exhibit a typical glassy behaviour Debenedetti and Stillinger (2001); Binder and Kob (2011); As the temperature decreases the MSD develops an increasingly long plateau during which $\Delta(t)$ attains a small value, and the ISF develops a two-step decay (Figs. 1(a), 1(c) and 1(e)). Conversely, soft cells ($p_{0}=3.81$) relax in a qualitatively different way. Although the dynamics slows down dramatically at low temperatures, the MSD does not exhibit a true plateau, if not at extremely low temperature, and the ISF does not decay in two steps (Figs. 1(b) and 1(f)). More importantly, an extended sub-diffusive behaviour follows the short-time ballistic regime one. Indeed, the log-slope $\Delta(t)$ of the mean square displacement develops an extended plateau, we show in Fig. 1(d). These results clarify that an anomalous glassy dynamics occurs for soft cells, as previously observed Sussman _et al._ (2018), but not for stiff ones. The stiffness of the cells, therefore, does not simply alters the energy scale for particle rearrangement, but rather qualitatively influences the relaxation dynamics. Figure 2: (a) Angell plot representation of the temperature dependence of the relaxation time, for different values of the target shape index $p_{0}$. $T_{g}$ is the glass transition temperature at which the relaxation time is $10^{4}$. (b) The Angell plot with the relaxation time defined as that where the mean square displacement attains a threshold value, as specified in the legend. The relaxation times are scaled so that they equal $\tau_{\alpha}$ at $T_{g}$. The high-$p_{0}$ regime where the anomalous diffusive behaviour occurs, is also that where the relaxation time exhibits a sub-Arrhenius temperature dependence Sussman _et al._ (2018), as we show in Fig. 2(a). Our results obtained in an extended $p_{0}$ range, however, demonstrate that a traditional super-Arrhenius behaviour occurs at low $p_{0}$. A similar crossover is found defining the relaxation time from the decay of correlation function of the area and of the perimeter of the cells, as discussed in Appendix A. A single parameter, $p_{0}$, thus controls the fragility and allows to transit from a sub- to a super-Arrhenius behaviour. We are not aware of other models with a similar crossover. We now clarify that the anomalous sub-diffusive regime and the unusual sub- Arrhenius behaviour are strongly tied. To this end, we define the relaxation time as that at which the mean square displacement reaches a threshold, $\langle\Delta r^{2}(\tau_{\Delta r^{2}_{\rm th}})\rangle=\Delta r^{2}_{\rm th}$. Figures 1(a) and 1(b) show that the mean square displacement at the relaxation time (circles) is $\Delta r^{2}_{\tau_{\alpha}}\simeq 0.1$, regardless of the temperature and of $p_{0}$. Hence, $\tau_{\Delta r^{2}_{\rm th}=0.1}\simeq\tau_{\alpha}$. The relaxation time $\tau_{\Delta r^{2}_{\rm th}}$ decreases with the threshold $\Delta r^{2}_{\rm th}$, becoming increasingly more influenced by the sub-diffusive regime rather than by the subsequent caging regime. When this occurs, the sub-Arrhenius behaviour becomes more apparent, as we illustrate in Fig. 2(b). We then conclude that the sub-diffusive behaviour induces the sub-Arrhenius one, while the caging regime contrasts it. ## IV Anomalous glassy dynamics: physical origin ### IV.1 Particle trajectories The above results demonstrate that cell tissues, for large values of the target shape index $p_{0}$, have a distinctive relaxation dynamics, which is quite different from that of conventional glassy systems. Why is this so? To begin addressing this question, we have repeated the above investigations via overdamped simulations, for selected low-temperature values, and show the results as open circles in Figs. 1(a)-(d). These simulations reproduce the anomalous sub-diffusive regime, demonstrating that this has not an inertial origin. Besides, we have also investigated the relaxation dynamics using cage- relative quantities Shiba _et al._ (2016); Illing _et al._ (2017); Vivek _et al._ (2017); Li _et al._ (2019a), an approach which allows filtering out the effect of long-wavelength fluctuations. These fluctuations might indeed be relevant in two spatial dimensions Mermin and Wagner (1966); Li _et al._ (2019a); Illing _et al._ (2017); Vivek _et al._ (2017); Shiba _et al._ (2016), influencing both the mean square displacement and the relaxation time. We illustrate in Appendix B that there are no considerable differences between the standard and the cage-relative relaxation dynamics; the anomalous sub- diffusive behaviour, therefore, is not the vestige of the vibrational dynamics of the system. Figure 3: Time dependence of the averaged eccentricity $\langle E(t)\rangle$ and of $\langle\rm{cos}\theta\rangle$ (see text) at $p_{0}=3.0$, (a) and (c), and at $p_{0}=3.81$, (b) and (d). The insets in (a) and (b) illustrate particle trajectories at the relaxation time. The horizontal dashed lines in (a) and (b) indicate the Brownian limit, $\langle E(t)\rangle=4/7$. To unveil the microscopic origin of the sub-diffusive behaviour, we then focus on the particle trajectories in the supercooled regime. Example trajectories, evaluated at the relaxation time, are in Figs. 3(a) for $p_{0}=3.0$, and 3(b) for $p_{0}=3.81$. We find that at small $p_{0}$, the trajectories have a round shape reflecting a caging regime, while conversely at large $p_{0}$ they are unusually stretched. To quantify this observation, we describe a trajectory as a sequence of $n_{v}=50$ points equally spaced in time. The eigenvectors of the gyration tensor of this set of points fix the spatial directions along which the fluctuations of the trajectory, as estimated by the squared eigenvalues $\lambda_{1}^{2}\geq\lambda_{2}^{2}$, are maximal and minimal. This allows associating to each trajectory an eccentricity, $E(t)=\left(\lambda_{1}^{2}(t)-\lambda_{2}^{2}(t)\right)^{2}/\left(\lambda_{1}^{2}(t)+\lambda_{2}^{2}(t)\right)^{2}$ Rudnick and Gaspari (1987); Ernst _et al._ (2012). Radially symmetric trajectories have $E=0$, straight lines $E=1$, while Brownian trajectories have $E=4/7$, in two dimensions Rudnick and Gaspari (1987). Consistently, the sample averaged eccentricity attains large values in the ballistic or super- diffusive regimes, reaches the Brownian limit at long times, and it is suppressed in the caging regime, as illustrated in Figs. 3(a) and 3(b). At $p_{0}=3.81$, an intermediate regime occurs between the ballistic and the caging one, where the eccentricity has a plateau. The time dependence of the eccentricity, henceforth, closely resembles that of the log-slope, as apparent comparing Figs. 3(a) and 3(b) with Figs. 1(c) and 1(d). More importantly, the trajectories reveal that the sub-diffusive behaviour results from an anisotropic motion of cells. These anisotropic motion does not correlate with the possible anisotropic shape of the cell, as we show in Appendix C. The stretched trajectories lead to a sub-diffusive dynamics, rather than to a super-diffusive dynamics as one might naively expected, due to the presence of anti-correlations in the motion of the cells. To highlight these correlations, we investigate the angle $\theta$ between consecutive displacements ${\bf r}(t_{0}+t)-{\bf r}(t_{0})$,${\bf r}(t_{0}+2t)-{\bf r}(t_{0}+t)$, over a time $t$. Both at low- and at high-$p_{0}$ values, the time evolution of $\langle\cos\theta\rangle$ resembles that of $\Delta(t)$ and of $E(t)$, as shown in Figs. 3(c) and 3(d). In particular, at large $p_{0}$, we observe an intermediate regime in between the ballistic and the caging ones. In this intermediate regime, $\langle\cos\theta\rangle\approx-0.1$ for a transient. Since this small value of $\langle\cos\theta\rangle$ occurs when the trajectories are elongated, we understand that sub-diffusion emerges as cells are transiently only slightly constrained. Figure 4: (a) Probability distribution of the cell edge length $l$ of energy minima configurations, and (b) dependence of the average energy barrier of T1 transitions on $l$, for $p_{0}=3.0$ (black squares) and $p_{0}=3.81$ (red circles). (a, inset): schematic of a T1 transition in which the edge connecting particles $\rm{a_{1}}$ and $\rm{a_{2}}$ disappears, and a novel edge connecting previously separated cells appears. (c) and (d) illustrate the time dependence of mean square displacement scaled by the temperature, respectively for $p_{0}=3.0$ and $p_{0}=3.81$. Colors indicate different temperature values, as in Fig. 1. ### IV.2 T1 transitions The structural relaxation dynamics is strongly correlated with the topology of the free-energy landscape in glassy systems Berthier and Biroli (2011). Indeed, we now show the existence of phase space directions along which the system is essentially free to diffuse, for short distances, considering the energetic cost of relaxation events involving cell rearrangements, or T1 transitions Weaire and Rivier (1984); Staple _et al._ (2010); Bi _et al._ (2014), one of which is schematically illustrated in the inset of Fig. 4(a). In a T1 transition a cell-edge of length $l$ disappears, as the system overcomes an energy barrier $\Delta e(l)$ we expect to increase with $l$, as observed in the Vertex model Bi _et al._ (2014). We have investigated the edge-length distribution $P(l)$ and the dependence of the average energy barrier on $l$, which are illustrated Figs. 4(a) and 4(b). We detail the procedure used to evaluate these quantities is Appendix D. At small $p_{0}$, $P(l)$ is Gaussian shaped, as observed in the Vertex model Bi _et al._ (2014), and the average energy barrier increases with $l$. At large $p_{0}$, $P(l)$ is broad and has almost a bi-modal shape, which is actually observed at even larger $p_{0}$ values not considered here Li and Ciamarra (2018). In particular, on increasing $p_{0}$ small $l$-values become more probable. The energy cost of T1 transitions involving small edges, e.g. $l\lesssim 0.2$, is sensibly smaller than the energy cost of the other edges, as apparent in Fig. 4(b). Hence, the system is essentially free to diffuse along the specific phase space directions that trigger the T1 transitions involving these small edges. To corroborate this picture we further consider that, since the free diffusion coefficient is proportional to $T$, the mean square displacement should scale as $T$ not only in the ballistic regime but also in the sub-diffusive one. We indeed observe in Figs. 4(c) and 4(d) that, at low $p_{0}$, plots of $\left\langle\Delta r^{2}(t)\right\rangle/T$ only collapse in the ballistic regime, while conversely at high-$p_{0}$ they also collapse in the sub-diffusive one. The emerging scenario reminds the diffusion of a particle in a random media as described by the Lorentz gas models van Beijeren (1982); Höfling _et al._ (2006); Bauer _et al._ (2010); Zeitz _et al._ (2017); Petersen and Franosch (2019), where a particle is free until it hits randomly placed obstacles, and sub-diffusion occurs below the correlation length of the fractal cluster of free space. Figure 5: Time dependence of probability of irreversible consecutive T1 transitions (black) and of the log-slope of mean square displacement (blue) for (a) $p_{0}=3.0$ and $T=0.06$ and for (b) $p_{0}=3.81$ and $T=0.002$. To further support the deep connection between anomalous dynamics and T1 transitions, we consider the probability $P_{\rm irr}$ that two consecutive T1 transitions of the same particle are not one the reverse of the other; this occurs if the two transitions lead to a change in the Voronoi neighbours of the particle. We illustrate in Fig. 11 the dependence of $P_{\rm irr}$ on the time interval $t$ separating the two transitions. To avoid cluttering of data, we consider in (a) $p_{0}=3.0$ and $T=0.06$, and in (b) $p_{0}=3.81$ and $T=0.002$, two state points having close relaxation time, and report results for other parameter values in Appendix E. In the figure, we also superimpose the log-slope $\Delta(t)$ of the mean square displacements. For small $p_{0}$, $P_{\rm irr}$ quickly attains a high, almost constant plateau value, characterizing the caging regime. $P_{\rm irr}$ then approaches $1$ as the system relaxes. For large $p_{0}$, $P_{\rm irr}$ grows essentially as a power-law during the sub-diffusive transient. An inflexion, reminiscent of a plateau in the caging-regime follows the power-law growth and the final approach to $1$. Hence, the sub-diffusive regime is characterized by a scarcity of irreversible transition. ## V Discussion Our study demonstrates that the relaxation dynamics of a model cell tissue qualitatively depends on the stiffness of the cells; while stiff cells exhibit a conventional glass-like relaxation dynamics, soft ones have an extended sub- diffusive transient and a sub-Arrhenius dependence on the relaxation time on the temperature. Consistently, dynamical heterogeneities grow on cooling for stiff cells, while they are almost temperature independence for soft cells, as we demonstrate in Appendix F. The qualitative changes in the relaxation dynamics originate from the emergence of phase space directions along which the system is essentially free to move, in soft cells, and establish an analogy between the energy landscape of cell tissues and the Lorentz model, on short length scales. We do not expect the sub-diffusive behavior we have discussed to be a universal feature of the dynamics of cell tissues. Its occurrence, indeed, might be hidden by the super-diffusive contribution to the mean square displacement of the active forces. To observe our finding one might suppress cell-motility, making the cell tissue dynamics thermal. In order for thermal forces alone to be able to induce the relaxation of the system, it migth be also convenient to consider soft tissues, as those close to the epithelial- mesenchymal transition Jordan _et al._ (2011); Tanaka and Ogishima (2015); Kalluri and Weinberg (2009). We remark, however that a sub-diffusive transient, $r^{2}\propto t^{\beta}$, with $\beta$ constant over an extended period of time, has been observed in some experiments Rogers (2007); Schötz _et al._ (2013); Nixon-Abell _et al._ (2016); Armiger _et al._ (2018); Fodor _et al._ (2018). Our results offer a possible explanation of these experimental findings because at short time the thermal contribution to the mean square displacement ($\propto t$), which is the one we have modeled, dominates over the active contribution ($\propto t^{2}$). ###### Acknowledgements. We acknowledge support from the Singapore Ministry of Education through the Academic Research Fund MOE2017-T2-1-066 (S), and are grateful to the National Supercomputing Centre (NSCC) of Singapore for providing computational resources. MP is supported by the H2020 program under the MSCA grant agreement No. 801370 and by the Secretary of Universities and Research of the Government of Catalonia through Beatriu de Pinós program Grant No. BP 00088 (2018). ## Appendix A Shape correlation functions Figure 6: Time dependence of perimeter (a) and of area (c) correlation functions at $p_{0}=3.0$. (b) and (d) are the time evolution of the same quantities at $p_{0}=3.81$. To prove that the anomalous dynamics are associated to changes in the shapes of the cells, as those one might expect T1 transitions to induce, we investigate the perimeter and area correlation functions. The perimeter correlation function is defined as $C_{P}(t)=\frac{\sum_{i=1}^{N}\left[p_{i}(t)-\langle p_{i}\rangle\right]\left[p_{i}(0)-\langle p_{i}\rangle\right]}{\sum_{i=1}^{N}\left[p_{i}(0)-\langle p_{i}\rangle\right]^{2}},$ (2) where $p_{i}(t)$ is the non-dimensional perimeter of cell $i$ at time $t$ and $\langle p_{i}\rangle$ is the time average value, which is cell dependent due to the polydispersity of our system. The area correlation function $C_{A}(t)$ is similarly defined. We illustrate the time dependence of $C_{P}(t)$ and of $C_{A}(t)$ at $p_{0}=3.0$ and at $p_{0}=3.81$ in Fig. 6. $C_{P}(t)$ and $C_{A}(t)$ demonstrate similar behavior. In particular, at $p_{0}=3.0$, both correlation functions exhibit a two-step decay, which is conversely not apparent at $p_{0}=3.81$. This $p_{0}$ dependence is consistent with that of the ISF (Fig. 1) and CR-ISF (Figs. 7(e) and 7(f)). We further extract from the shape correlation function the perimeter and the area relaxation time, $\tau_{P}$ and $\tau_{A}$, which satisfy $C_{P}(\tau_{P})=C_{A}(\tau_{A})=1/e$. Both relaxation time have a super-Arrhenius temperature dependence at small $p_{0}$, and a sub-Arrhenius one at large $p_{0}$, as we show in Fig. 8(b). The investigation of the relaxation dynamics via the shape-correlation function establishes a coupling between the geometrical properties of the cells and their displacement. Considering that the shape of the cell changes as a consequence of T1 transitions, this result indirectly links anomalous dynamics and T1 transitions. ## Appendix B Cage-relative dynamics Figure 7: Time dependence of the cage-relative mean square displacement (a), of its log-slope (c) and of the cage-relative self-intermediate scattering function (e) at $p_{0}=3.0$. (b), (d) and (f) show the time dependence of the same quantities at $p_{0}=3.81$. The full circles in all panels mark the cage- relative relaxation time $\tau_{\alpha}^{\rm CR}$, which is the time at which the cage-relative self-intermediate scattering function reaches $1/e$. Figure 8: Angell plots, as obtained using different definitions of the relaxation time. In (a), $\tau_{\alpha}^{\rm CR}$ is the cage-relative relaxation time. In (b), $\tau_{\rm P}$ and $\tau_{\rm A}$ are the perimeter and the area relaxation time. We have illustrated in Fig. 1 the mean square displacements (MSD), its log- slope, and the self-intermediate scattering function (ISF). We have additionally investigated the time dependence of these quantities using cage- relative (CR) measures. The CR measures differ from the standard ones in that the CR displacement $\Delta\mathbf{r}_{i}^{\rm CR}(t)=\Delta\mathbf{r}_{i}(t)-1/N_{i}\sum_{j=1}^{N_{i}}\Delta\mathbf{r}_{j}(t)$, where the sum is over the $N_{i}$ neighbors particle $i$ has at time $0$, replaces the displacement $\Delta\mathbf{r}_{i}(t)=\mathbf{r}_{i}(t)-\mathbf{r}_{i}(0)$. Particles moving coherently with their immediate neighbours have a large displacement, but a small CR displacement. Hence, CR measures filter out the effect of coherent displacements, and particularly the effect of long-wavelength fluctuations, which could affect the relaxation dynamics of two-dimensional systems Shiba _et al._ (2016); Vivek _et al._ (2017); Illing _et al._ (2017); Li _et al._ (2019a). Figure 7 shows that, for small $p_{0}$, the CR one reveals a typical glassy behaviour, including an extended plateau in CR-MSD and a two-step decay in CR- ISF at low temperatures, as the standard measure. Similarly, the anomalous sub-diffusive behaviour found at large $p_{0}$ persists when the relaxation dynamics is investigated using CR measures. Indeed, a region of anomalous diffusion is clearly observed in CR-MSD (Fig. 7(b)) and in its log-slope (Fig. 7(d)). This anomalous behaviour, and that observed in the standard quantities in Fig. 1 at the same $p_{0}$ value, occur on the same time scale. We further define the CR relaxation time $\tau_{\alpha}^{\rm CR}$ as the time at which CR-ISF reaches $1/e$, and illustrate the resulting Angell plot in Fig. 8(a). On increasing $p_{0}$, we observe a crossover from a super- to a sub-Arrhenius behaviour, as found in Fig. 2(a) using the standard measure. Overall, the investigation of the relaxation dynamics using CR quantities excludes the possibility that the observed anomalous behaviour occurring at large $p_{0}$ could originate from the emergence of collective particle displacements, like those induced by long-wavelength fluctuations. ## Appendix C Absence of correlation between shape and displacement of a cell Cells in tissue, being deformable objects, may acquire elongated shapes. A cell’s displacement could, therefore, correlate with its shape, e.g. in the anomalous diffusive regime at high $p_{0}$. To investigate this possibility, we first assume the eigenvector associated with the largest eigenvalue of the covariance matrix of the vertices of cell $i$ to identify its principal axis, $\mathbf{v}_{i}$. Next, we consider how the normalized cell displacement $\Delta\mathbf{r}_{i}(t)/|\Delta\mathbf{r}_{i}(t)|$ at time $t$ correlates with the principal axis at time $0$, studying $\cos\alpha_{i}(t)=\Delta\mathbf{r}_{i}(t)/|\Delta\mathbf{r}_{i}(t)|\cdot\mathbf{v_{i}}(0)$. Since the cell’s principal axis is defined up to an angle $\pi$, $\langle\cos\alpha(t)\rangle=0$. We, therefore, focus on $\langle\cos^{2}\alpha(t)\rangle$, which equal $1/2$ in the absence of correlations. In Fig. 9, we show that $\langle\cos^{2}\alpha(t)\rangle$ does equal $1/2$, regardless of the $p_{0}$ value and of the time. Analogous results are obtained at different temperatures. Accordingly, the shape of a cell at a given time does not correlate with its subsequent displacement. This result is consistent with our finding, according to which in the anomalous region cells move along direction inducing T1 transition associated with their short edges. Figure 9: Time dependence of $\langle\cos^{2}\alpha(t)\rangle$ for $p_{0}=3.0$ and $T=0.06$ (black squares) and for $p_{0}=3.81$ and $T=0.002$ (red circles). ## Appendix D T1 energy barrier Figure 10: (a) Illustration of tuning the edge length $l_{{\rm a}}$ between two selected neighboring cells (green) by separating them gradually so as to induce a T1 transition. (b) and (c) the edge length $l_{{\rm a}}$ dependence of the minimised total energy $e(l_{{\rm a}})$ at $p_{0}=3.0$ and $p_{0}=3.81$, respectively. Different colors are for different selected neighboring cell couples. The edge length $l_{{\rm a}}$ and the corresponding energy $e(l_{{\rm a}})$ for the snapshots shown in (a) are indicated in (b). We investigate the energy barrier for T1 transition to occur focusing on systems with $N=100$ cells quenched to their inherent state via the conjugate- gradient algorithm. In these systems, we randomly select two neighbouring cells and indicate with $l_{{\rm a}}$ the length of the Voronoi edge separating them. Then, we gradually increase the separation of the two cells, moving them by small steps along the direction connecting their centres. We fix the step size to $0.1$, $0.01$, and $0.001$ when the distance $dr$ between the cell centers is $dr>0.2$, $0.2>dr>0.1$, and $0.1>dr$, respectively. After each step, we minimize the energy of the tissue using the conjugate-gradient method, keeping fixed the positions of the selected cells. As the distance between the centers of selected cells increases, the length $l_{\rm a}$ of the Voronoi edge separating them decreases, and the energy of the system increases, as visualized in Fig. 10(a). As the length scale increases, the energy of the tissue grows, as illustrated for a few selected cell couples in Fig. 10(b) for $p_{0}=3.0$, and in Fig. 10(c) for $p_{0}=3.81$. The energy suddenly drops as the T1 transition separating the selected particles occurs, as $l_{{\rm a}}$ approaches $0$. The overall change in energy defines the energy barrier $\Delta e(l)=e(l_{{\rm a}}\rightarrow 0)-e(l)$. Figure 4(b) illustrates $\langle\Delta e(l)\rangle$ as a function of the initial edge length $l$. The data are obtained randomly by triggering 200 random T1 transitions, from 24 independent configurations. We note here that in a few instances we have observed drops in the dependence of the energy versus $l_{{\rm a}}$ due to T1 transitions which do not involve the displaced particles. Regardless, we operatively define $\Delta e(l)$ as the difference between the energy of the system as the separating particles undergo a T1 transition and the initial one. ## Appendix E T1 correlations In Fig. 11, we illustrate the time dependence of the probability to find irreversible consecutive T1 transitions, $P_{\rm irr}(t)$, at different temperatures for $p_{0}=3.0$ (panel (a)) and for $p_{0}=3.81$ (panel (b)). Figure 5 shows that data for $p_{0}=3.0$ and $T=0.06$ and for $p_{0}=3.81$ and $T=0.002$ are qualitatively different, and that $P_{\rm irr}(t)$ correlates with the MSD. Here, we notice that the temperature dependence of $P_{\rm irr}(t)$ is qualitatively the same, for different $p_{0}$ values. At higher temperature, it becomes increasingly more probable for two consecutive transitions separated by a small time interval $t$ not to be one the reverse of the other. Furthermore, as the temperature increases the plateau that $P_{\rm irr}(t)$ attains at long-time during the caging regime, reduces in extension and increases in value approaching $1$. Figure 11: Time dependence of the probability that two consecutive T1 transition of a same particle are not one the reverse of the other, for several selected values of temperatures at (a) $p_{0}=3.0$ and at (b) $p_{0}=3.81$. ## Appendix F Dynamical length scales Figure 12: Spatial-temporal correlation functions at time $t_{\rm max}$ for (a) $p_{0}=3.0$ and for (b) $p_{0}=3.81$. $t_{\rm max}$ is the time at which the corresponding four-point susceptibility reaches the maximum. The solid lines are exponential fits. (c) illustrates the dependence of the dynamical correlation length on the relaxation time for different $p_{0}$ values. The existence of a standard and of an anomalous glassy dynamics, respectively at small at a high $p_{0}$ values, suggests that the spatial temporal correlation between the particle displacement may likewise be strongly $p_{0}$ dependent. To investigate this issue, we focus on the decay of the spatial-temporal correlation function Pastore _et al._ (2011); Li _et al._ (2019b): $g_{4}(r_{ij},t)=\langle\omega_{i}(t)\omega_{j}(t)\rangle-\langle\omega_{i}(t)\rangle\langle\omega_{j}(t)\rangle.$ (3) Here $r_{ij}=|\mathbf{r}_{i}(0)-\mathbf{r}_{j}(0)|$ and $\omega_{i}(t)=1(0)$ if $|\mathbf{r}_{i}(t)-\mathbf{r}_{i}(0)|\leq$ ($>$) $l_{*}$. We fix $l_{*}=0.64$, the value at which the peak height of the corresponding four- point susceptibility $\chi_{4}(t)=1/N\sum_{i,j}g_{4}(r_{ij},t)$ is maximal, and fix the time $t=t_{\rm max}$ at which the corresponding $\chi_{4}(t)$ attains the maximum. From the exponential decay of $g_{4}(r,t)$, which is illustrated in Figs. 12(a) and 12(b) for selected $p_{0}$ and temperature values, we then extract the dynamical length scale $\xi$. In Fig. 12(c), we illustrate the dependence of the length scale $\xi$ on the relaxation time, for different values of $p_{0}$. $\xi$ increases as the dynamics slow down. At a given relaxation time, we observe a systematic reduction of $\xi$ on increasing $p_{0}$. This indicates that dynamic heterogeneities decrease as the softness of the particles increases, in line with the absence of a proper glassy behavior for these particles. ## References * Manli _et al._ (2012) C. Manli, H. David, and J. W. Cornelis, Curr. Genomics 13, 267 (2012). * Poujade _et al._ (2007) M. Poujade, E. Grasland-Mongrain, A. Hertzog, J. Jouanneau, P. Chavrier, B. Ladoux, A. Buguin, and P. Silberzan, Proc. Natl. Acad. Sci. U.S.A. 104, 15988 (2007). * Basan _et al._ (2013) M. Basan, J. Elgeti, E. Hannezo, W.-J. Rappel, and H. Levine, Proc. Natl. Acad. Sci. U.S.A. 110, 2452 (2013). * Trepat and Sahai (2018) X. Trepat and E. Sahai, Nature Physics 14, 671 (2018). * Schötz _et al._ (2013) E.-M. Schötz, M. Lanio, J. A. Talbot, and M. L. Manning, J. R. Soc. Interface 10 (2013). * Angelini _et al._ (2011) T. E. Angelini, E. Hannezo, X. Trepat, M. Marquez, J. J. Fredberg, and D. A. Weitz, Proc. Natl. Acad. Sci. U.S.A. 108, 4714 (2011). * Bi _et al._ (2016) D. Bi, X. Yang, M. C. Marchetti, and M. L. Manning, Phys. Rev. X 6, 021011 (2016). * Angelini _et al._ (2010) T. E. Angelini, E. Hannezo, X. Trepat, J. J. Fredberg, and D. A. Weitz, Phys. Rev. Lett. 104, 168104 (2010). * Sussman _et al._ (2018) D. M. Sussman, M. Paoluzzi, M. Cristina Marchetti, and M. Lisa Manning, Europhys. Lett. 121, 36001 (2018). * Jordan _et al._ (2011) N. V. Jordan, G. L. Johnson, and A. N. Abell, Cell Cycle 10, 2865 (2011). * Tanaka and Ogishima (2015) H. Tanaka and S. Ogishima, J. Mol. Cell Biol. 7, 253 (2015). * Kalluri and Weinberg (2009) R. Kalluri and R. A. Weinberg, J. Clin. Invest. 119, 1420 (2009). * Debenedetti and Stillinger (2001) P. Debenedetti and F. Stillinger, Nature 410, 259 (2001). * Binder and Kob (2011) K. Binder and W. Kob, _Glassy Materials and Disordered Solids_ , revised ed. (World Scientific, 2011). * Sadhukhan and Nandi (2020) S. Sadhukhan and S. K. Nandi, arXiv , arXiv:2007.14107 (2020). * Angell (1995) C. A. Angell, Science 267, 1924 (1995). * Rogers (2007) S. S. Rogers, Physical Biology 4, 220 (2007). * Nixon-Abell _et al._ (2016) J. Nixon-Abell, C. J. Obara, A. V. Weigel, D. Li, W. R. Legant, C. S. Xu, H. A. Pasolli, K. Harvey, H. F. Hess, E. Betzig, C. Blackstone, and J. Lippincott-Schwartz, Science 354, aaf3928 (2016). * Armiger _et al._ (2018) T. J. Armiger, M. C. Lampi, C. A. Reinhart-King, and K. N. Dahl, Journal of Cell Science 131, jcs216010 (2018). * Fodor _et al._ (2018) É. Fodor, V. Mehandia, J. Comelles, R. Thiagarajan, N. S. Gov, P. Visco, F. van Wijland, and D. Riveline, Biophysical Journal 114, 939 (2018). * Park _et al._ (2015) J.-A. Park, J. H. Kim, D. Bi, J. A. Mitchel, N. T. Qazvini, K. Tantisira, C. Y. Park, M. McGill, S.-H. Kim, B. Gweon, J. Notbohm, R. Steward Jr, S. Burger, S. H. Randell, A. T. Kho, D. T. Tambe, C. Hardin, S. A. Shore, E. Israel, D. A. Weitz, D. J. Tschumperlin, E. P. Henske, S. T. Weiss, M. L. Manning, J. P. Butler, J. M. Drazen, and J. J. Fredberg, Nat. Mater. 14, 1040 (2015). * Weaire and Rivier (1984) D. Weaire and N. Rivier, Contemp. Phys. 25, 59 (1984). * Staple _et al._ (2010) D. B. Staple, R. Farhadifar, J.-C. Röper, B. Aigouy, S. Eaton, and F. Jülicher, Eur. Phys. J. E 33, 117 (2010). * Bi _et al._ (2014) D. Bi, J. H. Lopez, J. M. Schwarz, and M. L. Manning, Soft Matter 10, 1885 (2014). * van Beijeren (1982) H. van Beijeren, Rev. Mod. Phys 54, 195 (1982). * Höfling _et al._ (2006) F. Höfling, T. Franosch, and E. Frey, Phys. Rev. Lett. 96, 165901 (2006). * Zeitz _et al._ (2017) M. Zeitz, K. Wolff, and H. Stark, European Physical Journal E 40, 1 (2017). * Farhadifar _et al._ (2007) R. Farhadifar, J.-C. Röper, B. Aigouy, S. Eaton, and F. Jülicher, Curr. Biol. 17, 2095 (2007). * Bi _et al._ (2015) D. Bi, J. H. Lopez, J. M. Schwarz, and M. L. Manning, Nat. Phys. 11, 1074 (2015). * Manning _et al._ (2010) M. L. Manning, R. A. Foty, M. S. Steinberg, and E.-M. Schoetz, Proc. Natl. Acad. Sci. U.S.A. 107, 12517 (2010). * Fletcher _et al._ (2014) A. G. Fletcher, M. Osterfield, R. E. Baker, and S. Y. Shvartsman, Biophys. J. 106, 2291 (2014). * Moshe _et al._ (2018) M. Moshe, M. J. Bowick, and M. C. Marchetti, Phys. Rev. Lett. 120, 268105 (2018). * Giavazzi _et al._ (2018) F. Giavazzi, M. Paoluzzi, M. Macchi, D. Bi, G. Scita, M. L. Manning, R. Cerbino, and M. C. Marchetti, Soft matter 14, 3471 (2018). * Li and Ciamarra (2018) Y.-W. Li and M. P. Ciamarra, Phys. Rev. Mater. 2, 045602 (2018). * Barton _et al._ (2017) D. L. Barton, S. Henkes, C. J. Weijer, and R. Sknepnek, PLoS Comput. Biol. 13, 34 (2017). * Allen (1987) M. Allen, _Computer Simulation of Liquids_ (Oxford University Press, Oxford, 1987). * Dieterich _et al._ (2008) P. Dieterich, R. Klages, R. Preuss, and A. Schwab, Proceedings of the National Academy of Sciences of the United States of America 105, 459 (2008). * Puliafito _et al._ (2012) A. Puliafito, L. Hufnagel, P. Neveu, S. Streichan, A. Sigal, D. K. Fygenson, and B. I. Shraiman, Proc. Natl. Acad. Sci. U.S.A. 109, 739 (2012). * Straetmans and Khain (2019) J. Straetmans and E. Khain, J. Stat. Phys. 176, 299 (2019). * Zehnder _et al._ (2015) S. Zehnder, M. Suaris, M. Bellaire, and T. Angelini, Biophys. J. 108, 247 (2015). * Sussman and Merkel (2018) D. M. Sussman and M. Merkel, Soft matter 14, 3397 (2018). * Shiba _et al._ (2016) H. Shiba, Y. Yamada, T. Kawasaki, and K. Kim, Phys. Rev. Lett. 117, 245701 (2016). * Illing _et al._ (2017) B. Illing, S. Fritschi, H. Kaiser, C. L. Klix, G. Maret, and P. Keim, Proc. Natl. Acad. Sci. U. S. A. 114, 1856 (2017). * Vivek _et al._ (2017) S. Vivek, C. P. Kelleher, P. M. Chaikin, and E. R. Weeks, Proc. Natl. Acad. Sci. U. S. A. 114, 1850 (2017). * Li _et al._ (2019a) Y.-W. Li, C. K. Mishra, Z.-Y. Sun, K. Zhao, T. G. Mason, R. Ganapathy, and M. Pica Ciamarra, Proc. Natl. Acad. Sci. U. S. A. 116, 22977 (2019a). * Mermin and Wagner (1966) N. D. Mermin and H. Wagner, Phys. Rev. Lett. 17, 1133 (1966). * Rudnick and Gaspari (1987) J. Rudnick and G. Gaspari, Science 237, 384 (1987). * Ernst _et al._ (2012) D. Ernst, M. Hellmann, J. Köhler, and M. Weiss, Soft Matter 8, 4886 (2012). * Berthier and Biroli (2011) L. Berthier and G. Biroli, Rev. Mod. Phys. 83, 587 (2011). * Bauer _et al._ (2010) T. Bauer, F. Höfling, T. Munk, E. Frey, and T. Franosch, Eur. Phys. J.: Spec. Top. 189, 103 (2010). * Petersen and Franosch (2019) C. F. Petersen and T. Franosch, Soft Matter 15, 3906 (2019). * Pastore _et al._ (2011) R. Pastore, M. P. Ciamarra, A. de Candia, and A. Coniglio, Phys. Rev. Lett. 107, 065703 (2011). * Li _et al._ (2019b) Y.-W. Li, Z.-Q. Li, Z.-L. Hou, T. G. Mason, K. Zhao, Z.-Y. Sun, and M. Pica Ciamarra, Phys. Rev. Mater. 3, 125603 (2019b).
# A Smoking Gun for Planetesimal Formation: Charge Driven Growth into a New Size Range Jens Teiser University of Duisburg-Essen Faculty of Physics Lotharstr. 1-21 D-47057 Duisburg, Germany Maximilian Kruss University of Duisburg-Essen Faculty of Physics Lotharstr. 1-21 D-47057 Duisburg, Germany Felix Jungmann University of Duisburg-Essen Faculty of Physics Lotharstr. 1-21 D-47057 Duisburg, Germany Gerhard Wurm University of Duisburg-Essen Faculty of Physics Lotharstr. 1-21 D-47057 Duisburg, Germany ###### Abstract Collisions electrically charge grains which promotes growth by coagulation. We present aggregation experiments with three large ensembles of basalt beads ($150\,\mu\mathrm{m}-180\,\mu\mathrm{m})$, two of which are charged, while one remains almost neutral as control system. In microgravity experiments, free collisions within these samples are induced with moderate collision velocities ($0-0.2\,\mathrm{m\,s}^{-1}$). In the control system, coagulation stops at (sub-)mm size while the charged grains continue to grow. A maximum agglomerate size of 5 cm is reached, limited only by bead depletion in the free volume. For the first time, charge-driven growth well into the centimeter range is directly proven by experiments. In protoplanetary disks, this agglomerate size is well beyond the critical size needed for hydrodynamic particle concentration as, e.g., by the streaming instabilities. ††journal: ApJL ## 1 Introduction The first stage of planet formation is dominated by hit-and-stick collisions between small dust and ice grains at small collision velocities (Wurm & Blum, 1998; Blum & Wurm, 2008; Johansen et al., 2014; Gundlach & Blum, 2015). Although this first step is fast and efficient, there are several obstacles, which stop this evolution. With increasing agglomerate size, the relative velocities between the colliding aggregates increase (Weidenschilling & Cuzzi, 1993). This leads to restructuring and compaction (Weidling et al., 2009; Meisner et al., 2012). If now two of these compact aggregates collide slowly ($<1\,\mathrm{m\,s}^{-1}$), they rather bounce off than stick to each other. This has been introduced as the bouncing barrier (Güttler et al., 2010; Zsom et al., 2010). Several experiments showed that self-consistent growth indeed comes to a halt at a particle size in the millimeter range (Kruss et al., 2016, 2017; Demirci et al., 2017). Slight shifts in aggregate size are possible depending on temperatures or magnetic fields (Kruss & Wurm, 2018, 2020; Demirci et al., 2019) but the bouncing barrier is a robust finding. Collisions and growth in a protoplanetary disk are governed by the interaction between gas and solids. Hydrodynamic processes therefore have a strong effect on particle evolution. Beyond inducing collisions, they can especially change the local particle concentration (Johansen et al., 2007; Johansen & Youdin, 2007; Chiang & Youdin, 2010; Squire & Hopkins, 2018). If a critical solid-to- gas ratio is reached, the mutual gravity between the solids might lead to the direct formation of a planetesimal (Youdin & Goodman, 2005; Simon et al., 2016; Klahr & Schreiber, 2020). This way, barriers in collisional growth could be prevented. However, while these drag instabilities can be very efficient, they require a minimum size of solids to be present (Dra̧żkowska & Dullemond, 2014; Carrera et al., 2015). Typically, they work best for so-called pebbles with the Stokes-number (ratio between the orbital period and gas-grain-coupling time) $St\sim 1$. Depending on the disk model and the location in the protoplanetary disk, this Stokes number typically translates into particle sizes of the order of a decimeter though somewhat smaller sizes might still work (Yang et al., 2017) under certain conditions. Obviously, to explain planet formation, a severe size gap must be bridged between the millimeter size resulting from the bouncing barrier and the decimeter required for the hydrodynamic processes to work. This bridge might be a charge dominated growth phase. Collisions and friction between particles lead to charge separation upon contact (Lacks & Mohan Sankaran, 2011). For a long time, this was attributed either to different materials in contact (different surface energies) or due to different sizes (Lee et al., 2015). Experiments showed that charge separation also occurs for particles of the same size and material (Jungmann et al., 2018). While the detailed physical processes are poorly understood, the resulting charge distributions in granular samples are well characterized. For a granular sample with particles of the same size and material, a broad charge distribution is the result. By first glimpse, it is similar to a Gaussian distribution, but with heavier tails (Haeberle et al., 2018). The peak position (mean charge) and the full width at half maximum (FWHM) are typically used to characterize the charge distribution of a granular sample. For multiple collisions of particles of the same size and same material the resulting charge distribution peaks at zero charge (Wurm et al., 2019). Within the scope of this paper, the term ”strongly charged” then refers to the FWHM of a corresponding charge distribution. Agitating a granular system for a duration of about 10 min establishes a charge distribution within the system which does not change significantly when the agitation is continued. It only depends on sample and atmospheric parameters as was shown in experiments with monodisperse basalt beads by Wurm et al. (2019). Larger beads are charged more strongly (larger FWHM) than smaller beads. Additionally, the charging of a granular sample strongly depends on the surrounding gas pressure. For a constant granular sample the width of the reached charge distribution follows a curve similar to Paschen’s law of electrical breakthrough in gases. The width (FWHM) of the charge distributions reaches a minimum at a characteristic pressure ($100\,\mathrm{Pa}\,-\,\mathrm{few}\,100\,\mathrm{Pa}$), depending on the sample. At larger pressures, the reached width increases gently, while it increases sharply for pressures smaller than the characteristic value. ## 2 Charge driven aggregation and stability It is obvious that two grains of opposite charge attract each other as a first step in aggregation. This effectively changes the collisional cross section. It is not obvious a priori though why an initial charge on individual grains should be beneficial for the aggregation process later on, once agglomerates have formed. In fact, once two grains of the same absolute charge, but opposite sign, collide and stick to each other this dimer is overall neutral. The long-range Coulomb interaction of net charges is no longer present and the collisional cross section will no longer be strongly enhanced. However, insulating grains do not discharge upon contact. This even holds for metal spheres, if their surface is not extremely cleaned from any contamination (Genc et al., 2019; Kaponig et al., 2020). A dimer or a more complex aggregate still hold charges on their surface. It has to be noted at this point that collisions lead to charge separation (not neutralization) in the first place and grains charged this way have a complex charge pattern with patches of negative and positive charges on their surface (Grosjean et al., 2020; Steinpilz et al., 2020b). This is also valid for grains which are net neutral. The typical configuration are therefore multipoles even on a single grain. Certainly, there are ways to discharge and neutralize grains, which depends on the environment (water content, gas pressure, temperature, radiation, material conductivity). In the experiments here, discharge takes hours, under protoplanetary disk conditions it might be years (Steinpilz et al., 2020a; Jungmann et al., 2018, 2021) (and running experiments by Steinpilz et al., personal communication). So, while these multipole configurations in aggregates might not attract other grains from far away, charges remain highly important during contact. As Coulomb forces decrease with distance of two charges $r_{c}$ as $1/r_{c}^{2}$, two oppositely charged spots on a surface close to the contact point can dominate the sticking force, independent of the net charge budget of the grains. Therefore, collisions lead to sticking at much higher collision velocities for charged grains (Jungmann et al., 2018). It is important to note that in an ensemble of grains net charge is only a proxy that is easily accessible to confirm that grains have a surface charge pattern. Nevertheless, the multipoles determine sticking forces. Aggregates are also more stable, i.e. higher collision velocities are required to destroy them compared to uncharged aggregates (Steinpilz et al., 2020a). Charge patches glue aggregates together. A simple analog for this situation is a salt crystal, which is overall neutral but the alternating charges still provide strong attraction. This way, we expect collisionally charged grains to grow far beyond the bouncing barrier. Therefore, there is a high potential in charge driven growth. It is still unclear though, how large agglomerates can grow this way. In Steinpilz et al. (2020a), cm-size charged agglomerates were observed but their direct formation from individual grains was not traced and took place prior to the free-floating phase in an agitated granular bed. The effects of charging were shown by means of numerical simulations matching the experiments in that case. The key question thus remains: Is charge driven coagulation able to provide the necessary aggregate sizes for drag instabilities to take over? ## 3 Experiment To investigate how large agglomerates can form by collisions of small charged particles, a microgravity experiment is currently being developed for sub- orbital platforms. Here, we report on the first shorter time microgravity experiments during this development, which were conducted at the Bremen Drop Tower (ZARM). Using the catapult mode, microgravity with residual acceleration of $<10^{-6}\,\mathrm{g}$ and a duration of $9.2\,\mathrm{s}$ could be used. ### 3.1 Experimental setup Figure 1: Schematic view of test cells 2 and 3. The upper cell (3) is a vacuum chamber ($p\approx 20\,\mathrm{Pa}$), while cell 2 is at normal pressure. The central part of the experimental setup consists of three test cells, two of which are shown in Fig. 1. All cells have the same geometry with a free volume of $50\,\mathrm{mm}\times 50\,\mathrm{mm}\times 42\,\mathrm{mm}$ and an additional particle reservoir of 14 mm depth and 25 mm length. While cells 2 and 3 are placed in the same unit, cell 1 is placed in a single unit with half the height. Test cell 3 (the upper one in Fig. 1) is designed as a vacuum chamber with a pressure of $20\,\mathrm{Pa}$, while the other two cells are at normal pressure. The side walls of the test cells are copper electrodes which are part of a capacitor at which a DC-voltage of 4 kV can be applied. The experiments are observed with a Raspberry Pi camera (30 frames/s, resolution: $1640\,\times 1230\,\mathrm{Px}$) using bright field illumination in a backlight configuration. The optical resolution is of the order of the particle size. Each sample consists of basalt beads with a size distribution between $150\,\mathrm{\mu m}$ and $180\,\mathrm{\mu m}$ diameter (Whitehouse Scientific). A total amount of 6 g is used in each cell. To reduce collisions between basalt beads and different materials, the top and the bottom of each test cell (including the particle reservoir) are coated with the same basalt beads. ### 3.2 Experiment protocol The test cells can be agitated with a harmonic motion using a voice coil mounted underneath. Prior to the catapult launch this agitation is used to charge the basalt beads for test cells 2 and 3 by collisions and friction due to constant agitation on ground. The duration is 20 min with a frequency of 14 Hz and an amplitude of 4.6 mm (peak to peak), similar to the experiment protocol in Steinpilz et al. (2020a). During this period the sample mostly remains within the sample reservoir and the beads are exposed to numerous collisions and friction among each other. Even in case the particles leave the reservoir, they are only exposed to particle-particle collisions, as the bottom of the test cells is coated with basalt beads and tilted by an angle of $4^{\circ}$, so basically no particles hit the side walls. The sample was left at rest for about 10 min between this agitation period and the catapult launch, which does not change the charge distribution significantly. In contrast to cells 2 and 3, test cell 1 was not agitated on ground, so a sample with minimum charge is used as control experiment. In microgravity, the agitation is used to distribute the sample at the beginning and to keep up a sufficient collision rate to see growth on the short timescales available at the drop tower. The experiment protocol has been changed for the different experiments, so that different aspects of the planned long-duration experiments could be tested on a short timescale. A high voltage at the capacitor plates can be used to estimate the charges in aggregates, but immediately stops further growth processes as the volume is cleared from particles. Agitation can be used to induce a high collision rate, but no charge measurement or particle tracking is possible during this agitation. The different parameters used in the presented experiments are described in section 4. ### 3.3 Collision velocities Due to the limitations of the optical system and the large particle concentration within the test cells, it is not possible to track single particles. However, the collision velocities can be estimated using the parameters of the agitation cycle. As the agitation follows a harmonic, frequency $f$ and maximum amplitude $A_{0}$ directly translate in a maximum velocity of the test cells via $v_{\rm{max}}=2\pi f\cdot A_{0}$. For the parameters used in Fig. 2 ($f=14\,\mathrm{Hz}$, $A_{0}=1.2\,\mathrm{mm}$) this results in a maximum velocity of $v_{\rm{max}}=0.11\,\mathrm{m\,s}^{-1}$. In case of perfectly elastic collisions between the particles and the experiment walls (top and bottom), a resting particle could get a maximum velocity of $v_{\rm{col}}=2v_{\rm{max}}$ or even larger in case of an initial velocity towards the wall. Indeed, non-charged basalt beads collide rather elastically with a coefficient of restitution $\epsilon=v_{\rm{after}}/v_{\rm{before}}$ of the order of $\epsilon=0.9$ (Bogdan et al., 2019). It has to be noted that smaller basalt beads are used and collision velocities are lower in the drop tower experiments compared to the work by Bogdan et al. (2019). Additionally, this situation is slightly more complicated. As the surfaces of the bottom (including the particle compartment) and the top are coated with the same basalt beads as the sample particles, all collisions are collisions between beads at a random impact parameter. Also the charges of the beads themselves influence the collision behavior (Jungmann et al., 2018), as the coefficient of restitution gets smaller for larger charges. Altogether we estimate the maximum collision velocities to be roughly the same as the maximum velocities of the test cells. Additionally, collisions in the free volume of the test cell will damp the original velocity distribution. We therefore assume a broad velocity distribution from almost zero to the maximum velocity of the test cell. ## 4 Results The experiments presented were planned to qualify an experiment hardware for suborbital flights and to test parts of the experiment protocol. Here, we present three different experiment protocols, which were used to show different aspects of the upcoming long-duration experiments. Figure 2: Evolution of the particle ensemble in cell 3 during one experimental run at different times, starting directly after the begin of microgravity and ending when the sample has reached a final state while the test cell is at rest. The width of the chamber of 50 mm can be used as a scale. ### 4.1 First growth The first experiment protocol was used to check if a rather homogeneous particle distribution can be generated by agitation in microgravity. Here, the sample was agitated with a frequency of 14 Hz, an amplitude of 1.2 mm, and a duration of 1 s, starting at the beginning of the microgravity phase. After a short break of 1 s, the agitation was then repeated for a duration of $1\,\mathrm{s}$. Afterwards, the test cells were not moved until the end of the microgravity. Fig. 2 shows the temporal evolution of the particles in the vacuum chamber (cell 3) for this experiment. It starts directly after the last agitation cycle from a well distributed sample, which blocks the illumination almost completely. While the test cell is kept at rest a clustering process can be observed. After around 5 s, the final state is reached as the grown clusters do not collide anymore. Of the three cells, the vacuum chamber shows the most homogeneous sample distribution in the initial state and the largest agglomerate sizes in the final state (see also Fig. 4). According to Wurm et al. (2019) basalt beads charge differently depending on the surrounding pressure. At pressures of about $100\,\mathrm{Pa\,-\,\mathrm{few}\,100\,\mathrm{Pa}}$ the charge distribution is narrowest. For lower pressure the width of the charge distribution rises steeply (Wurm et al., 2019). Additionally, particle motion is not damped significantly by gas drag, as the gas-grain coupling time exceeds the experiment duration. Therefore, this directly leads to larger particle velocities and a higher collision rate. However, the maximum size reached in this experiment run is still restricted to a few millimeters. This can be attributed to the declining collision rate, as the particle velocities are damped by collisions and/or gas drag (depending on the test cell) and therefore also the collision probability goes down. ### 4.2 Charges The charge distribution of single basalt beads cannot be obtained from the data available, as the spatial and temporal resolution of the camera system are not suitable for this. However, the agitation method and therefore the process of particle charging is almost identical to previous studies, either with glass beads (Jungmann et al., 2018, 2021; Steinpilz et al., 2020a) or with basalt beads (Wurm et al., 2019). The width (FWHM) of the corresponding charge distribution scales with the particle size (Wurm et al., 2019), with many studies treating charges on insulators as surface charges only (Lee et al., 2018; Grosjean et al., 2020; Steinpilz et al., 2020b). With a similar charge density on the surface as in Wurm et al. (2019), a charge distribution centered at zero charge with a FWHM of about $3\cdot 10^{-13}\,\mathrm{C}$ can be expected. To roughly estimate the charges on the agglomerates we performed an experiment in which the beads were shaken for 5 s during microgravity (f = 14 Hz, A = 1.2 mm). As shown in Fig. 4 (middle) cm-sized agglomerates form in cell 2. After agitation, a voltage of $\pm 2\,$kV is applied in that cell which accelerates all charged particles towards the electrodes. These single agglomerates are tracked manually and their acceleration is translated to the amount of charge they carry. For this the mass of the agglomerates and its error is estimated via their cross-sections in the images. Fig. 3 shows that the agglomerates are formed from several hundred up to thousands of single particles. Their typical charge is up to $10^{7}$ electron charges. We note that this is only an estimate of the order of magnitude as a detailed analysis is beyond the scope of this paper and not possible with the available data from the short-time experiments. However, this indicates that there are abundant charges on the clusters which might play a role in the agglomeration process. Figure 3: Absolute estimated charges of resulting clusters formed during microgravity. The large error bars result from the high uncertainty of the masses of the agglomerates. ### 4.3 Growing tall Figure 4: Final particle distributions with continuous agitation of 1 Hz under microgravity. Top: sample with minimum charge (no shaking in advance). Middle: 20 min shaking in advance, normal pressure. Bottom: 20 min shaking in advance, vacuum (20 Pa). The evolution presented in Fig. 2 shows that growth by collisions of charged basalt beads is possible in principle. On the other hand it becomes clear that the collision rate is crucial for the outcome and has to be kept on a high level. This was considered in the experiment shown in Fig. 4. Here, the experiment protocol was changed to maintain a certain collision rate, while observation of the grains is still possible. With the onset of microgravity, the test cells were agitated with $f=14\,\mathrm{Hz}$ and $A_{0}=1.2\,\mathrm{mm}$ amplitude for a duration of 3 s. Afterwards, the agitation frequency was reduced to $f=1\,\mathrm{Hz}$ for a duration of 4 s, resulting in an amplitude of 3 mm and a maximum velocity of $v_{\rm{max}}=0.02\,\mathrm{m\,s}^{-1}$. Afterwards, the test cells were at rest for the last 2 s of microgravity. Fig. 4 shows the final particle distributions during this experiment run. It also reveals the systematic differences between the three test cells. The least charged sample (top) only shows minor growth in comparison to the charged sample in the test cell with atmospheric pressure (middle) where larger entities evolve. The vacuum cell (bottom) shows a striking result. Almost all particles are incorporated into one large agglomerate, which therefore has a width of 5 cm (from wall to wall), a thickness of about 2 cm and a total mass of about $6\,\mathrm{g}$. The free volume between the larger agglomerates is also of great interest to interpret the result. In the least charged sample, the particles remain widely distributed in the entire volume. Also in the test cell with atmospheric pressure there is still a significant amount of single beads (or only very small agglomerates) in the free volume between the larger aggregates. This is totally different in cell 3, where the maximum charges can be expected. Single particles are either incorporated into the major agglomerate or stick to the walls of the cell. The free volume is almost completely cleared from small particles. Due to this particle depletion the growth process comes to a halt. Therefore, it can be assumed that the final distribution does not show the maximum sizes achievable by such collisions. ### 4.4 Stability When applying an electric field some large clusters are accelerated towards the electrodes and therefore reach high impact velocities. This can be used to estimate the stability of these clusters. Similar to chapter 4.2 these clusters were tracked manually and their impact velocities determined. An example of a cluster colliding and bouncing off the wall is shown in Fig. 5. Figure 5: Sequence of a 2.5 mm cluster (average diameter) colliding with the electrode. Its impact velocity is about 2 cm/s. The red arrow shows the direction of motion. Collisions between clusters and the (side) walls occur at impact velocities from $10^{-2}\,\mathrm{m\,s}^{-1}$ to $2.1\cdot 10^{-2}\,\mathrm{m\,s}^{-1}$, while the typical cluster sizes (average diameters) range from $2.5\,\mathrm{mm}$ to $5.3\,\mathrm{mm}$. No fragmentation was observed, only sticking or bouncing were observed. Relative velocities in protoplanetary disks depend on the sizes of the collision partners and are lowest for particles of similar size Weidenschilling & Cuzzi (1993). For equally sized particles, the collision velocities are $\leq 10^{-2}\,\mathrm{m\,s}^{-1}$, even for particles of a few cm in size (Weidenschilling & Cuzzi, 1993). As a collision with a solid wall is much more severe than mutual collisions between equally sized clusters, mutual collisions in protoplanetary disks will not destroy the grown agglomerates. The relative velocities are larger for particles of different size, so single particles hitting an agglomerate will be faster. However, the velocities are still in the range of $10^{-1}\,\mathrm{m\,s}^{-1}$, which is exactly in the velocity range of the single basalt beads during the agitation cycles, resulting in growth. ## 5 Conclusion and outlook Charge driven coagulation has already been presented by Steinpilz et al. (2020a) as a possible mechanism to overcome the bouncing barrier in planet formation. Although they could show that the size distribution of agglomerates in an ensemble changes when electrical charges are present, the maximum sizes found were still restricted. Of course, this first step might already be sufficient to help hydrodynamic processes to facilitate further growth (Yang et al., 2017; Schaffer et al., 2018). However, this process is still fragile with respect to the corresponding hydrodynamic models. Here, we present a smoking gun for coagulation to large agglomerate sizes. Although the detailed growth processes are still hidden due to the experimental limitations, there is the proof of concept that small charged grains indeed grow well into the centimeter range and possibly further starting from scratch, i.e. starting with a cloud of individual grains. We observed that growth is only possible if the grains are charged. The size of the largest agglomerates in the experiments presented here is limited by the particle supply (no further collisions) and a non-sufficient experiment duration. Charge driven growth might even continue into the decimeter range. It has to be noted though, that agglomerates in this size range are in danger of destruction by wind erosion, as they already move fast with respect to the surrounding gas. Future experiments will enable us to trace the charge driven coagulation in more detail. As the experimental setup is already dedicated for long duration experiments on suborbital flights, a much better understanding of the detailed growth process can be expected. We thank the anonymous reviewer for the fruitful comments, which helped to improve the manuscript. This work is supported by the German Space Administration (DLR) with funds provided by the Federal Ministry for Economic Affairs and Energy (BMWi) under grant 50WM1762. ## References * Blum & Wurm (2008) Blum, J., & Wurm, G. 2008, ARA&A, 46, 21 * Bogdan et al. (2019) Bogdan, T., Teiser, J., Fischer, N., Kruss, M., & Wurm, G. 2019, Icarus, 319, 133 * Carrera et al. (2015) Carrera, D., Johansen, A., & Davies, M. B. 2015, A&A, 579, A43. https://doi.org/10.1051/0004-6361/201425120 * Chiang & Youdin (2010) Chiang, E., & Youdin, A. N. 2010, Annual Review of Earth and Planetary Sciences, 38, 493 * Demirci et al. (2019) Demirci, T., Krause, C., Teiser, J., & Wurm, G. 2019, A&A, 629, A66 * Demirci et al. (2017) Demirci, T., Teiser, J., Steinpilz, T., et al. 2017, ApJ, 846, 48 * Dra̧żkowska & Dullemond (2014) Dra̧żkowska, J., & Dullemond, C. P. 2014, A&A, 572, A78 * Genc et al. (2019) Genc, E., Mölleken, A., Tarasevitch, D., et al. 2019, Review of Scientific Instruments, 90, 075115. https://doi.org/10.1063/1.5093988 * Grosjean et al. (2020) Grosjean, G., Wald, S., Sobarzo, J. C., & Waitukaitis, S. 2020, Physical Review Materials, 4, 082602 * Gundlach & Blum (2015) Gundlach, B., & Blum, J. 2015, ApJ, 798, 34 * Güttler et al. (2010) Güttler, C., Blum, J., Zsom, A., Ormel, C. W., & Dullemond, C. P. 2010, A&A, 513, A56 * Haeberle et al. (2018) Haeberle, J., Schella, A., Sperl, M., Schröter, M., & Born, P. 2018, Soft Matter, 14, 4987 * Johansen et al. (2014) Johansen, A., Blum, J., Tanaka, H., et al. 2014, in Protostars and Planets VI, ed. H. Beuther, R. S. Klessen, C. P. Dullemond, & T. Henning, 547 * Johansen et al. (2007) Johansen, A., Oishi, J. S., Mac Low, M.-M., et al. 2007, Nature, 448, 1022 * Johansen & Youdin (2007) Johansen, A., & Youdin, A. 2007, ApJ, 662, 627 * Jungmann et al. (2018) Jungmann, F., Steinpilz, T., Teiser, J., & Wurm, G. 2018, Journal of Physics Communications, 2, 095009 * Jungmann et al. (2021) Jungmann, F., Bila, T., Kleinert, L., et al. 2021, Icarus, 355, 114127 * Kaponig et al. (2020) Kaponig, M., Mölleken, A., Tarasevitch, D., et al. 2020, Journal of Electrostatics, 103, 103411 * Klahr & Schreiber (2020) Klahr, H., & Schreiber, A. 2020, ApJ, 901, 54 * Kruss et al. (2016) Kruss, M., Demirci, T., Koester, M., Kelling, T., & Wurm, G. 2016, ApJ, 827, 110 * Kruss et al. (2017) Kruss, M., Teiser, J., & Wurm, G. 2017, A&A, 600, A103 * Kruss & Wurm (2018) Kruss, M., & Wurm, G. 2018, ApJ, 869, 45 * Kruss & Wurm (2020) —. 2020, The Planetary Science Journal, 1, 23 * Lacks & Mohan Sankaran (2011) Lacks, D. J., & Mohan Sankaran, R. 2011, Journal of Physics D Applied Physics, 44, 453001 * Lee et al. (2018) Lee, V., James, N. M., Waitukaitis, S. R., & Jaeger, H. M. 2018, Physical Review Materials, 2, 035602 * Lee et al. (2015) Lee, V., Waitukaitis, S. R., Miskin, M. Z., & Jaeger, H. M. 2015, Nature Physics, 11, 733 * Meisner et al. (2012) Meisner, T., Wurm, G., & Teiser, J. 2012, A&A, 544, A138 * Schaffer et al. (2018) Schaffer, N., Yang, C.-C., & Johansen, A. 2018, A&A, 618, A75 * Simon et al. (2016) Simon, J. B., Armitage, P. J., Li, R., & Youdin, A. N. 2016, The Astrophysical Journal, 822, 55 * Squire & Hopkins (2018) Squire, J., & Hopkins, P. F. 2018, MNRAS, 823 * Steinpilz et al. (2020a) Steinpilz, T., Joeris, K., Jungmann, F., et al. 2020a, Nature Physics, 16, 225 * Steinpilz et al. (2020b) Steinpilz, T., Jungmann, F., Joeris, K., Teiser, J., & Wurm, G. 2020b, New Journal of Physics, 22, 093025 * Weidenschilling & Cuzzi (1993) Weidenschilling, S. J., & Cuzzi, J. N. 1993, in Protostars and Planets III, ed. E. H. Levy & J. I. Lunine, 1031 * Weidling et al. (2009) Weidling, R., Güttler, C., Blum, J., & Brauer, F. 2009, ApJ, 696, 2036 * Wurm & Blum (1998) Wurm, G., & Blum, J. 1998, Icarus, 132, 125 * Wurm et al. (2019) Wurm, G., Schmidt, L., Steinpilz, T., Boden, L., & Teiser, J. 2019, Icarus, 331, 103 * Yang et al. (2017) Yang, C. C., Johansen, A., & Carrera, D. 2017, A&A, 606, A80 * Youdin & Goodman (2005) Youdin, A. N., & Goodman, J. 2005, The Astrophysical Journal, 620, 459 * Zsom et al. (2010) Zsom, A., Ormel, C. W., Güttler, C., Blum, J., & Dullemond, C. P. 2010, A&A, 513, A57
# Parity alternating permutations starting with an odd integer Frether Getachew Kebede111Corresponding author. Department of Mathematics, College of Natural and Computational Sciences, Addis Ababa University, P.O.Box 1176, Addis Ababa, Ethiopia; e-mail<EMAIL_ADDRESS>Fanja Rakotondrajao Département de Mathématiques et Informatique, BP 907 Université d’Antananarivo, 101 Antananarivo, Madagascar; e-mail<EMAIL_ADDRESS> ###### Abstract A Parity Alternating Permutation of the set $[n]=\\{1,2,\ldots,n\\}$ is a permutation with even and odd entries alternatively. We deal with parity alternating permutations having an odd entry in the first position, PAPs. We study the numbers that count the PAPs with even as well as odd parity. We also study a subclass of PAPs being derangements as well, Parity Alternating Derangements (PADs). Moreover, by considering the parity of these PADs we look into their statistical property of excedance. ###### keywords: parity , parity alternating permutation , parity alternating derangement , excedance ###### MSC: [2020] 05A05 , 05A15 , 05A19 ## 1 Introduction and preliminaries A permutation $\pi$ is a bijection from the set $[n]=\\{1,2,\ldots,n\\}$ to itself and we will write it in standard representation as $\pi=\pi(1)\,\pi(2)\,\cdots\,\pi(n)$, or as the product of disjoint cycles. The parity of a permutation $\pi$ is defined as the parity of the number of transpositions (cycles of length two) in any representation of $\pi$ as a product of transpositions. One way of determining the parity of $\pi$ is by obtaining the sign of $(-1)^{n-c}$, where $c$ is the number of cycles in the cycle representation of $\pi$. That is, if the sign of $\pi$ is -1, then $\pi$ is called an odd permutation, and an even permutation otherwise. For example, the permutation $4\,2\,1\,7\,8\,6\,3\,5=(1\,\,4\,\,7\,\,3)(2)(5\,\,8)(6)$, of length 8, is even since it has sign 1. All basic definitions and properties not explained here can be found in for example [8] and [4]. According to [9], a Parity Alternating Permutation over the set $[n]$ is a permutation, in standard form, with even and odd entries alternatively (in this general sense). The set $\mathcal{P}_{n}$ of all parity alternating permutations is a subgroup of the symmetric group $S_{n}$, the group of all permutations over $[n]$. The order of the set set $\mathcal{P}_{n}$ has been studied lately in relations to other number sequences such as Eulerian numbers (see [10, 9]). However, in this paper we will deal only with the parity alternating permutations which in addition have an odd entry in the first position; and we call them PAPs. It can be shown that the set $P_{n}$ containing all PAPs over $[n]$ is a subgroup of the symmetric group $S_{n}$ and also of the group $\mathcal{P}_{n}$. We consider this kind of permutations because, for odd $n$ there are no parity alternating permutations over $[n]$ beginning with an even integer. Avi Peretz determined the number sequence that count the number of PAPs (see A010551). Unfortunately, we could not find any details of his work. In A010551, we can also find the exponential generating function of these numbers due to Paul D. Hanna. Since there is no published proof of this formula we prove it here, as Theorem 2.2. Moreover, the numbers that count the PAPs with even parity and with odd parity (which were not studied before) are determined. By $p_{n}$ we denote the cardinality of the set $P_{n}$ of all PAPs over $[n]$. Let $\phi_{n}$ denote a map from $P_{n}$ to $S_{\lceil\frac{n}{2}\rceil}\times S_{\lfloor\frac{n}{2}\rfloor}$ that relates a PAP $\sigma$ to a pair of permutations $(\sigma_{1},\sigma_{2})$ in the set $S_{\lceil\frac{n}{2}\rceil}\times S_{\lfloor\frac{n}{2}\rfloor}$, in such a way that $\sigma_{1}(i)=\frac{\sigma(2i-1)+1}{2}$ and $\sigma_{2}(i)=\frac{\sigma(2i)}{2}$. It is easy to see that this map is a bijection. For example, the PAPs $5\,2\,1\,4\,3\,6\,7$ and $7\,4\,5\,6\,3\,2\,1$ over [7] are mapped to the pairs $(3\,1\,2\,4,\,1\,2\,3)$ and $(4\,3\,2\,1,\,2\,3\,1)$, respectively. If we consider a PAP $\sigma$ in cycle representations, then each cycle consists of integers of the same parity. Thus, we immediately get cycle representation of $\sigma_{1}$ and $\sigma_{2}$. For instance, the cycle form of the two PAPs above are $(1\,5\,3)(7)(2)(4)(6)$ and $(1\,7)(3\,5)(2\,4\,6)$ which correspond to the pairs $\left((1\,3\,2)(4),\,(1)(2)(3)\right)$ and $((1\,4)(2\,3),\,(1\,2\,3))$, respectively. (Unless stated otherwise we will always use (disjoint) cycle representation of permutations.) Another way of looking at the mapping $\phi_{n}$ is that $\sigma_{1}$ and $\sigma_{2}$ correspond to the parts that contain the odd and even integers in $\sigma$, respectively. Therefore, studying PAPs is similar to studying the two permutations that correspond to the even and the odd integers in the PAP separately and then combining the properties. In Table 1, we give a short summary of properties that permutations and PAPs satisfy (for detailed discussions, see Section 2). | Permutations | PAPs ---|---|--- Seq | $1,1,2,6,24,120,\ldots.$ (A000142) | $1,1,1,2,4,12,\ldots.$ (A010551) EGF | $\frac{1}{1-x}$ | $\frac{2\sqrt{4-x^{2}}+2\cos^{-1}\left(1-x^{2}/2\right)}{(2-x)\sqrt{4-x^{2}}}$ Even (seq) | $1,1,1,3,12,60,\ldots.$ (A001710) | $1,1,1,1,2,6,18,72,\ldots$ Odd (seq) | $1,1,1,3,12,60,\ldots.$ (A001710) | $0,0,0,1,2,6,18,72,\ldots$ Even (EGF) | $\frac{2-x^{2}}{2-2x}$ | $\frac{\sqrt{4-x^{2}}+\cos^{-1}\left(1-\frac{x^{2}}{2}\right)}{(2-x)\sqrt{4-x^{2}}}+\frac{x^{2}}{4}+\frac{x}{2}+\frac{1}{2}$ Odd (EGF) | $\frac{x^{2}}{2-2x}$ | $\frac{\sqrt{4-x^{2}}+\cos^{-1}\left(1-\frac{x^{2}}{2}\right)}{(2-x)\sqrt{4-x^{2}}}-\frac{x^{2}}{4}-\frac{x}{2}-\frac{1}{2}$ Table 1: A comparison table of permutations and PAPs (EGF mean exponential generating function). One interesting subset of $S_{n}$ is the set $D_{n}$ of derangements. For $d_{n}=|D_{n}|$, we have a well known relation $\displaystyle d_{n}=(n-1)[d_{n-1}+d_{n-2}],\,\,d_{0}=1\text{ and }d_{1}=0$ (1) for $n\geq 2$. A proof of this relation may be found in any textbook on combinatorics, but we will have later use of the following bijection due to Mantaci and Rakotondrajao ([6]). They define $\psi_{n}$ to be the bijection between $D_{n}$ and $[n-1]\times(D_{n-1}\cup D_{n-2})$ as follows: let $D_{n}^{(1)}$ denote the set of derangements over $[n]$ having the integer $n$ in a cycle of length greater than 2, and $D_{n}^{(2)}$ be the set of derangements over $[n]$ having $n$ in a transposition. These two sets are disjoint and their union is $D_{n}$. Then for $\delta\in D_{n}$ define $\psi_{n}(\delta)=(i,\delta^{\prime})$, where $i=\delta^{-1}(n)$ and $\delta^{\prime}$ is the derangement obtained from 1. $\bullet$ $\delta\in D_{n}^{(1)}$ by removing $n$ or 2. $\bullet$ $\delta\in D_{n}^{(2)}$ by removing the transposition $(i\,\,n)$ and then decreasing all integers greater than $i$ by 1. For instance, the pairs $(2,(1\,5\,2)(3\,4))$ and $(2,(1\,2)(3\,4))$ correspond to the derangements $(1\,5\,2\,6)(3\,4)$ and $(1\,3)(4\,5)(2\,6)$, respectively, for $n=6$. We denote the restricted bijections $\psi_{n}|_{D_{n}^{(1)}}$ and $\psi_{n}|_{D_{n}^{(2)}}$ by $\psi_{n}^{(1)}$ and $\psi_{n}^{(2)}$, respectively. Another important, and more difficult to prove, recurrence relation that the numbers $d_{n}$ satisfy is $\displaystyle d_{n}=n\,d_{n-1}+(-1)^{n},\,\,d_{0}=0$ (2) for $n\geq 1$. We will later make a use of the bijection $\tau_{n}:([n]\times D_{n-1})\backslash F_{n}\longrightarrow D_{n}\backslash E_{n}$ given by the second author ([7]) proving the recurrence. Where $E_{n}$ is the set containing the derangement $\Delta_{n}=(1\,2)(3\,4)\cdots(n-1\,\,\,n)$ for even $n$, and is empty for odd $n$. $F_{n}$ is the set containing the pair $(n,\,\Delta_{n-1})$ when $n$ is odd, and is empty when $n$ is even. Thus, the inverse $\zeta_{n}$ of $\tau_{n}$ relates an element of $[n-1]\times D_{n-1}$ with every derangement over $[n]$ that has the integer $n$ in a cycle of length greater than 2, and an element of $\\{n\\}\times D_{n-1}\backslash F_{n}$ with every derangement over $[n]$ in which $n$ lies in a transposition. Classifying derangements by their parity, we denote the number of even and odd derangements over $[n]$ by $d^{e}_{n}$ and $d^{o}_{n}$, respectively. Clearly $d_{n}=d^{e}_{n}+d^{o}_{n}$. Moreover, the numbers $d^{e}_{n}$ and $d^{o}_{n}$ satisfy the relations $\displaystyle d^{e}_{n}=(n-1)[d^{o}_{n-1}+d^{o}_{n-2}]\quad\text{ and }\quad d^{o}_{n}=(n-1)[d^{e}_{n-1}+d^{e}_{n-2}],$ (3) for $n\geq 2$ with initial conditions $d^{e}_{0}=1$, $d^{e}_{1}=0$, $d^{o}_{0}=0$, and $d^{o}_{1}=0$ ([6], Proposition 4.1). We will put a major interest on Parity Alternating Derangements (PADs) which are the derangements which also are parity alternating permutations starting with odd integers. Let $\mathfrak{d}_{n}$ denote cardinality of the set of PADs $\mathfrak{D}_{n}=D_{n}\cap P_{n}$. The restricted bijection $\Phi_{n}=\phi_{n}|_{\mathfrak{D}_{n}}:\mathfrak{D}_{n}\longrightarrow D_{\lceil\frac{n}{2}\rceil}\times D_{\lfloor\frac{n}{2}\rfloor}$ will let us consider the odd parts and the even parts of any given PAD regarded as ordinary derangements with smaller length than the length of the PAD. The mapping $\Phi_{n}$ plays the central role in our investigations. In Table 2, we display the connection of ordinary derangements and PADs (for detailed discussions, see Section 3). Finding explicit expressions for some of the generating functions are still open questions. On the other hand the EGF for the PADs for example is the solution to an eighth order differential equation with polynomial coefficients, and also is expressible in terms of Hadamard products of some known generating functions. | Derangements | PADs ---|---|--- Seq | $1,0,1,2,9,44,\ldots$ (A000166) | $1,0,0,0,1,2,4,18,81,396,\ldots$ EGF | $\frac{e^{-x}}{1-x}$ | open RR | $d_{n}=(n-1)[d_{n-1}+d_{n-2}]$ | relation (4) RR | $d_{n}=nd_{n-1}+(-1)^{n}$ | relation (5) Even (seq) | $1,0,0,2,3,24,130,\ldots$ (A003221) | $1,0,0,0,1,0,4,6,45,192,976\ldots$ Odd (seq) | $0,0,1,0,6,20,135,\ldots$ (A000387) | $0,0,0,0,0,2,0,12,36,204,960,\ldots$ Even (EGF) | $\frac{(2-x^{2})e^{-x}}{2(1-x)}$ | open Odd (EGF) | $\frac{x^{2}e^{-x}}{2(1-x)}$ | open Even (RR) | $d^{e}_{n}=(n-1)[d^{o}_{n-1}+d^{o}_{n-2}]$ | relation (6) Odd (RR) | $d^{o}_{n}=(n-1)[d^{e}_{n-1}+d^{e}_{n-2}]$ | relation (7) Even - Odd | $(-1)^{n-1}(n-1)$ | $(-1)^{n-2}\Big{\lceil}\frac{n-2}{2}\Big{\rceil}\Big{\lfloor}\frac{n-2}{2}\Big{\rfloor}$ Table 2: A comparison table of derangements and PAPs, RR represents recurrence relation. In section 4, we study excedance distribution over PADs by means of the corresponding distributions for the two derangements obtained by $\Phi_{n}$. ## 2 Parity Alternating Permutations (PAPs) As we stated in the introduction, we use splitting method by the mapping $\phi_{n}$ in the study of PAPs. One application of this is that the number of PAPs of length $n$ is $\displaystyle p_{n}=|S_{\lceil\frac{n}{2}\rceil}||S_{\lfloor\frac{n}{2}\rfloor}|=\lceil n/2\rceil!\lfloor n/2\rfloor!.$ $n$ | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 ---|---|---|---|---|---|---|---|---|---|---|--- $p_{n}$ | 1 | 1 | 1 | 2 | 4 | 12 | 36 | 144 | 576 | 2880 | 14400 Table 3: First few terms of the sequence $\\{p_{n}\\}_{0}^{\infty}$. ###### Proposition 2.1. The numbers $p_{n}$ satisfy the recurrence relation $\displaystyle p_{n}=\lceil n/2\rceil p_{n-1},$ for $n\geq 1$ and $p_{0}=1$. ###### Proof. First let us define a mapping $\omega_{n}:S_{n}\longrightarrow[n]\times S_{n-1}$ by $\displaystyle\omega_{n}(\pi)=(i,\pi^{\prime}),$ where $\pi^{\prime}$ is obtained from $\pi\in S_{n}$ by removing the integer $n$, and $i=\pi^{-1}(n)$. One can easily see that $\omega_{n}$ is a bijection. Now let us take a PAP $\sigma$ over $[n]$. Then $\phi_{n}$ maps $\sigma$ to a pair $(\sigma_{1},\sigma_{2})$. Define then a mapping $\Omega:P_{n}\longrightarrow\Bigl{[}\big{\lceil}\frac{n}{2}\big{\rceil}\Bigr{]}\times P_{n-1}$ as follows: for $n=2m$ $\displaystyle\Omega(\sigma)=\left(i,\phi_{2m}^{-1}(\sigma_{1},\,\sigma_{2}^{\prime})\right),$ where $(i,\sigma_{2}^{\prime})=\omega_{m}(\sigma_{2})$, and for $n=2m+1$ $\displaystyle\Omega(\sigma)=\left(i,\phi_{2m+1}^{-1}(\sigma_{1}^{\prime},\,\sigma_{2})\right),$ where $(i,\sigma_{1}^{\prime})=\omega_{m+1}(\sigma_{1})$. The mapping $\Omega$ is a bijection since $\omega_{n}$ is a bijection for every $n\geq 1$. In any case, there are $\big{\lceil}\frac{n}{2}\big{\rceil}$ possibilities for $i$. ∎ As a consequence, we get the following theorem. ###### Theorem 2.2. The exponential generating function $P(x)=\sum_{n\geq 0}p_{n}\frac{x^{n}}{n!}$ of the sequence $\\{p_{n}\\}_{n=0}^{\infty}$ has the closed formula $P(x)=\displaystyle\frac{2}{2-x}+\displaystyle\frac{\cos^{-1}(1-\displaystyle\frac{x^{2}}{2})}{(2-x)\sqrt{1-\displaystyle\frac{x^{2}}{4}}}.$ ###### Proof. Based on the recurrence relation in Proposition 2.1, we obtain the following relations $\displaystyle P_{0}(x)=\frac{x}{2}P_{1}(x)+1\quad\text{ and }\quad P_{1}(x)=\frac{x}{2}P_{0}(x)+\frac{1}{2}\int_{0}^{x}P_{0}(t)\,dt,$ where $P_{0}(x)=\sum_{n\geq 0}p_{2n}\frac{x^{2n}}{(2n)!}$ and $P_{2}(x)=\sum_{n\geq 0}p_{2n+1}\frac{x^{2n+1}}{(2n+1)!}$. Clearly, $P(x)=P_{0}(x)+P_{1}(x)$. Additionally, $P_{0}(x)$ satisfies the differential equation $\displaystyle\left(1-\frac{x^{2}}{4}\right)P^{\prime}_{0}(x)=\frac{x^{2}+2}{2x}P_{0}(x)-\frac{1}{x}.$ Thus, we obtain the formulas $\displaystyle P_{0}(x)=\frac{4}{4-x^{2}}+\frac{4x\sin^{-1}\left(\frac{x}{2}\right)}{(4-x^{2})^{3/2}}\quad\text{ and }\quad P_{1}(x)=\frac{8}{4x-x^{3}}+\frac{8x\sin^{-1}\left(\frac{x}{2}\right)}{x(4-x^{2})^{3/2}}-\frac{2}{x}.$ Therefore, $P(x)=\displaystyle\frac{2}{2-x}+\displaystyle\frac{\cos^{-1}(1-\displaystyle\frac{x^{2}}{2})}{(2-x)\sqrt{1-\displaystyle\frac{x^{2}}{4}}}.\qed$ For classification of PAPs in terms of their parity, we use $P^{e}_{n}$ and $P^{o}_{n}$ to denote the set of even PAPs and odd PAPs, respectively, and $p_{n}^{e}$ and $p_{n}^{o}$ as their cardinality, respectively. Thus, $\displaystyle p_{n}=p^{e}_{n}+p^{o}_{n}.$ $n$ | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 ---|---|---|---|---|---|---|---|---|---|---|--- $p_{n}^{e}$ | 1 | 1 | 1 | 1 | 2 | 6 | 18 | 72 | 288 | 1440 | 7200 $p_{n}^{o}$ | 0 | 0 | 0 | 1 | 2 | 6 | 18 | 72 | 288 | 1440 | 7200 Table 4: First few terms of the sequences $\\{p_{n}^{e}\\}_{0}^{\infty}$ and $\\{p_{n}^{o}\\}_{0}^{\infty}$. Our goal is now to study the relationships between these two sequences. ###### Theorem 2.3. The numbers $p_{n}^{e}$ and $p_{n}^{o}$ satisfy the recurrence relations $\displaystyle p_{n}^{e}=\lfloor(n-1)/2\rfloor p_{n-1}^{o}+p_{n-1}^{e}$ $\displaystyle p_{n}^{o}=\lfloor(n-1)/2\rfloor p_{n-1}^{e}+p_{n-1}^{o},$ for $n\geq 1$, with initial conditions $p^{e}_{0}=1$ and $p^{o}_{0}=0$. ###### Proof. Let $S_{n}^{e}$ and $S_{n}^{o}$ be the set of even and odd permutations, respectively. Define two mappings $\omega_{n}^{e}:S_{n}^{e}\longrightarrow[n-1]\times S_{n-1}^{o}\cup S_{n-1}^{e}$ and $\omega_{n}^{o}:S_{n}^{o}\longrightarrow[n-1]\times S_{n-1}^{e}\cup S_{n-1}^{o}$ by $\displaystyle\omega_{n}^{e}(\pi)=\begin{cases}(i,\pi^{\prime}),\text{ if }i\neq n\\\ \pi^{\prime\prime},\text{ otherwise}\end{cases}\quad\text{ and }\quad\omega_{n}^{o}(\pi)=\begin{cases}(j,\pi^{\prime}),\text{ if }j\neq n\\\ \pi^{\prime\prime},\text{ otherwise}\end{cases},$ respectively, where $i=\pi^{-1}(n)$, $\pi^{\prime}$ is obtained from $\pi$ by removing the integer $n$, and $\pi^{\prime\prime}$ is obtained from $\pi$ by removing the cycle $(n)$, for $\pi\in S_{n}^{e}$. Similarly for $\omega_{n}^{o}$. It is easy to see that both mappings $\omega_{n}^{e}$ and $\omega_{n}^{o}$ are bijections. The mapping $\omega_{n}^{e}$ changes the parity of $\pi$ when it results in $\pi^{\prime}$ and preserves when it results in $\pi^{\prime\prime}$. This is because the signs of $\pi$, $\pi^{\prime}$ and $\pi^{\prime\prime}$ are $(-1)^{n-c}$, $(-1)^{n-1-c}$, and $(-1)^{n-1-c+1}$, respectively, where $c$ is the number of cycles in $\pi$. For the mapping $\omega_{n}^{o}$ we apply similar argument. Now consider a PAP $\sigma$ in $P_{n}$. Then $\phi_{n}$ maps $\sigma$ in to a pair $(\sigma_{1},\sigma_{2})$. Following the notation in the proof of Proposition 2.1, let us define two mappings $\Omega^{e}:P_{n}^{e}\longrightarrow\Bigl{[}\big{\lfloor}\frac{n-1}{2}\big{\rfloor}\Bigr{]}\times P_{n-1}^{o}\cup P_{n-1}^{e}$ and $\Omega^{o}:P_{n}^{o}\longrightarrow\Bigl{[}\big{\lfloor}\frac{n-1}{2}\big{\rfloor}\Bigr{]}\times P_{n-1}^{e}\cup P_{n-1}^{o}$ as follows: 1. 1. when $n$ is even $\displaystyle\Omega^{e}(\sigma)=\begin{cases}\left(i,\,\phi_{n}^{-1}(\sigma_{1},\sigma_{2}^{\prime})\right),\text{ if }i\neq\frac{n}{2}\\\ \phi_{n}^{-1}(\sigma_{1},\sigma_{2}^{\prime\prime}),\text{ otherwise}\end{cases}\text{and }$ $\displaystyle\Omega^{o}(\sigma)=\begin{cases}\left(i,\,\phi_{n}^{-1}(\sigma_{1},\sigma_{2}^{\prime})\right),\text{ if }i\neq\frac{n}{2}\\\ \phi_{n}^{-1}(\sigma_{1},\sigma_{2}^{\prime\prime}),\text{ otherwise}\end{cases},$ where $i=\sigma_{2}^{-1}(\frac{n}{2})$, and both $\sigma^{\prime}_{2}$, $\sigma^{\prime\prime}_{2}$ are obtained from $\sigma_{2}$ by the mapping $\omega_{\frac{n}{2}}^{e}$ when $\sigma\in P_{n}^{e}$ and by the mapping $\omega_{\frac{n}{2}}^{o}$ when $\sigma\in P_{n}^{o}$, 2. 2. when $n$ is odd $\displaystyle\Omega^{e}(\sigma)=\begin{cases}\left(j,\,\phi_{n}^{-1}(\sigma_{1}^{\prime},\sigma_{2})\right),\text{ if }j\neq\frac{n+1}{2}\\\ \phi_{n}^{-1}(\sigma_{1}^{\prime\prime},\sigma_{2}),\text{ otherwise}\end{cases}\text{and }$ $\displaystyle\Omega^{o}(\sigma)=\begin{cases}\left(j,\,\phi_{n}^{-1}(\sigma_{1}^{\prime},\sigma_{2})\right),\text{ if }j\neq\frac{n+1}{2}\\\ \phi_{n}^{-1}(\sigma_{1}^{\prime\prime},\sigma_{2}),\text{ otherwise}\end{cases},$ where $j=\sigma_{1}^{-1}(\frac{n+1}{2})$, and both $\sigma^{\prime}_{1}$, $\sigma^{\prime\prime}_{1}$ are obtained from $\sigma_{1}$ by the mapping $\omega_{\frac{n+1}{2}}^{e}$ when $\sigma\in P_{n}^{e}$ and by the mapping $\omega_{\frac{n+1}{2}}^{o}$ when $\sigma\in P_{n}^{o}$. Since $\omega_{n}$ is bijection for $n\geq 2$, both $\Omega^{e}$ and $\Omega^{o}$ are bijections too. Note that in both mappings there are $\big{\lfloor}\frac{n-1}{2}\big{\rfloor}$ possibilities for $i$ ($i\neq\frac{n}{2}$) and similarly for $j$ ($j\neq\frac{n+1}{2}$). ∎ ###### Proposition 2.4. For any positive integer $n\geq 3$, we have $\displaystyle p_{n}^{e}=p_{n}^{o}.$ ###### Proof. Multiplying a PAP by a transposition $(1,n)$ if $n$ is odd, or by $(1,n-1)$ if $n$ is even, we obtain a PAP having opposite parity. This multiplication means swapping the first and the last odd integer of a PAP in standard representation. It creates a bijection between $P^{e}_{n}$ and $P^{o}_{n}$. ∎ By applying Proposition 2.4 and considering $p(x)$, we get: ###### Corollary 2.5. The exponential generating functions of the sequences $\\{p_{n}^{e}\\}_{n\geq 0}$ and $\\{p_{n}^{o}\\}_{n\geq 0}$ have the closed forms $\displaystyle P^{e}(x)=\displaystyle\frac{1}{2}\left(P(x)+\frac{x^{2}}{2}+x+1\right)\quad\text{ and }\quad P^{o}(x)=\displaystyle\frac{1}{2}\left(P(x)-\frac{x^{2}}{2}-x-1\right).$ $\square$ ## 3 Parity Alternating Derangements (PADs) As a result of the bijection $\Phi_{n}$ in the introduction above, we can determine the number $\mathfrak{d}_{n}$ of PADs over $[n]$ as follows: $\displaystyle\mathfrak{d}_{n}=d_{\lceil n/2\rceil}d_{\lfloor n/2\rfloor}=\sum_{j=0}^{\lceil n/2\rceil}\sum_{i=0}^{\lfloor n/2\rfloor}\lceil n/2\rceil!\lfloor n/2\rfloor!\frac{(-1)^{i+j}}{j!\,i!}.$ $n$ | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 ---|---|---|---|---|---|---|---|---|---|--- $d_{n}$ | 1 | 0 | 1 | 2 | 9 | 44 | 265 | 1854 | 14833 | 133496 $\mathfrak{d}_{n}$ | 1 | 0 | 0 | 0 | 1 | 2 | 4 | 18 | 81 | 396 Table 5: First few values of $d_{n}$ and $\mathfrak{d}_{n}$. In the next theorem we give a formula for the number of PADs, connected to the relation (1). ###### Theorem 3.6. The number of PADs over $[n]$ satisfy the recurrence relation $\displaystyle\mathfrak{d}_{n}=s\big{(}\mathfrak{d}_{n-1}+(n-2-s)\left(\mathfrak{d}_{n-3}+\mathfrak{d}_{n-4}\right)\big{)},$ (4) where $s=\big{\lfloor}\frac{n-1}{2}\big{\rfloor}=\frac{2n-3-(-1)^{n}}{4}$, for $n\geq 4$, with initial conditions $\mathfrak{d}_{0}=1$, $\mathfrak{d}_{1}=0$, $\mathfrak{d}_{2}=0$ and $\mathfrak{d}_{3}=0$. ###### Proof. The proof is splitted into two cases, for PADs over a set of odd and even sizes. Let $(\delta_{1},\delta_{2})$ be the corresponding pair of a PAD $\delta\in\mathfrak{D}_{n}$ under the mapping $\Phi_{n}$. Define a mapping $\Psi:\mathfrak{D}_{n}\longrightarrow[s]\times\left(\mathfrak{D}_{n-1}\cup[n-2-s]\times(\mathfrak{D}_{n-3}\cup\mathfrak{D}_{n-4})\right)$ as follows: Case I: for odd $n$, depending on the following two elective properties of $\delta$, the mapping $\Psi$ will be defined as: 1. 1. in the event of the largest entry $\frac{n+1}{2}$ of $\delta_{1}$ being in a cycle of length greater than 2, we let $\displaystyle\Psi(\delta)=\left(i,\,\Phi_{n}^{-1}(\delta^{\prime}_{1},\,\delta_{2})\right),$ where $(i,\delta^{\prime}_{1})=\psi_{\frac{n+1}{2}}^{(1)}(\delta_{1})$; 2. 2. in the event when $\frac{n+1}{2}$ lies in a transposition in $\delta_{1}$, we distinguish two cases: * (a) if the largest entry $\frac{n-1}{2}$ of $\delta_{2}$ is contained in a cycle of length greater than 2, then $\displaystyle\Psi(\delta)=\left(i,\,j,\,\Phi_{n}^{-1}(\delta^{\prime}_{1},\,\delta^{\prime}_{2})\right),$ where $(i,\,\delta^{\prime}_{1})=\psi_{\frac{n+1}{2}}^{(2)}(\delta_{1})$ and $(j,\,\delta^{\prime}_{2})=\psi_{\frac{n-1}{2}}^{(1)}(\delta_{2})$; * (b) if $\frac{n-1}{2}$ is contained in a cycle of length 2 in $\delta_{2}$, then $\displaystyle\Psi(\delta)=\left(i,\,j,\,\Phi_{n}^{-1}(\delta^{\prime}_{1},\,\delta^{\prime}_{2})\right),$ where $(i,\,\delta^{\prime}_{1})=\psi_{\frac{n+1}{2}}^{(2)}(\delta_{1})$ and $(j,\,\delta^{\prime}_{2})=\psi_{\frac{n-1}{2}}^{(2)}(\delta_{2})$. Case II: for even $n$, 1. 1. in the event of the largest entry $\frac{n}{2}$ of $\delta_{2}$ lies in a cycle of length greater than 2, we let $\displaystyle\Psi(\delta)=\left(i,\,\Phi_{n}^{-1}(\delta_{1},\,\delta^{\prime}_{2})\right),$ where $(i,\,\delta^{\prime}_{2})=\psi_{\frac{n}{2}}^{(1)}(\delta_{2})$; 2. 2. in the event when $\frac{n}{2}$ being in a transposition in $\delta_{2}$, we distinguish two cases: * (a) if the largest entry $\frac{n}{2}$ in $\delta_{1}$ contained in a cycle of length greater than 2, then $\displaystyle\Psi(\delta)=\left(i,\,j,\,\Phi_{n}^{-1}(\delta^{\prime}_{1},\,\delta^{\prime}_{2})\right),$ where $(i,\,\delta^{\prime}_{1})=\psi_{\frac{n}{2}}^{(1)}(\delta_{1})$ and $(j,\,\delta^{\prime}_{2})=\psi_{\frac{n}{2}}^{(2)}(\delta_{2})$; * (b) if $\frac{n}{2}$ contained in a cycle of length 2 in $\delta_{1}$, then $\displaystyle\Psi(\delta)=\left(i,\,j,\,\Phi_{2n}^{-1}(\delta^{\prime}_{1},\,\delta^{\prime}_{2})\right),$ where $(i,\,\delta^{\prime}_{1})=\psi_{\frac{n}{2}}^{(2)}(\delta_{1})$ and $(j,\,\delta^{\prime}_{2})=\psi_{\frac{n}{2}}^{(2)}(\delta_{2})$. Since $\psi_{n}$ is a bijection for any $n\geq 2$, one can easily conclude that $\Psi$ is a bijection too. Note that in both cases there are $\big{\lfloor}\frac{n-1}{2}\big{\rfloor}=s$ possibilities for $i$ and $\big{\lfloor}\frac{n-2}{2}\big{\rfloor}=n-2-s$ possibilities for $j$. Thus, the formula in the Theorem follows. ∎ The next theorem is connected to the relation (2). ###### Theorem 3.7. The number $\mathfrak{d}_{n}$ of PADs also satisfies the relation $\displaystyle\mathfrak{d}_{n}=s\mathfrak{d}_{n-1}+(-1)^{s}d_{n-s},$ (5) where $s=\big{\lceil}\frac{n}{2}\big{\rceil}=\frac{2n+1-(-1)^{n}}{4}$, for $n\geq 1$ with $d_{1}=0$, $d_{0}=1$ and $\mathfrak{d}_{0}=1$. ###### Proof. Distinguishing by means of the parity of $n$, we can write the relation (5) as: $\displaystyle\mathfrak{d}_{n}=d_{s}\big{(}s\,d_{s-1}+(-1)^{s}\big{)}\,\,\text{ when }n\text{ is even, and }$ $\displaystyle\mathfrak{d}_{n}=\big{(}s\,d_{s-1}+(-1)^{s}\big{)}\,d_{s-1}\,\,\text{ when }n\text{ is odd}.$ Now, take a PAD $\delta$ in $\mathfrak{D}_{n}$ and introduce two mappings as: * 1. $Z_{0}:\big{(}D_{s}\backslash E_{s}\big{)}\times D_{s}\longrightarrow[s]\times\big{(}D_{s-1}\backslash F_{s}\big{)}\times D_{s}$ by $\displaystyle\delta\xrightleftharpoons[\Phi^{-1}_{n}]{\Phi_{n}}(\delta_{1},\delta_{2})\xrightleftharpoons[id_{s}\times\tau_{s}]{id_{s}\times\zeta_{s}}\big{(}\delta_{1},(i,\delta^{\prime}_{2})\big{)}\xrightleftharpoons[h]{h}(i,(\delta_{1},\delta^{\prime}_{2}))\xrightleftharpoons[id\times\Phi_{n}]{id\times\Phi_{n}^{-1}}\big{(}i,\Phi_{n}^{-1}(\delta_{1},\delta^{\prime}_{2})\big{)},$ and * 2. $Z_{1}:\big{(}D_{s}\backslash E_{s}\big{)}\times D_{s-1}\longrightarrow[s]\times\big{(}D_{s-1}\backslash F_{s}\big{)}\times D_{s-1}$ by $\displaystyle\delta\xrightleftharpoons[\Phi^{-1}_{n}]{\Phi_{n}}(\delta_{1},\delta_{2})\xrightleftharpoons[\tau_{s}\times id_{s-1}]{\zeta_{s}\times id_{s-1}}\big{(}(i,\delta^{\prime}_{1}),\delta_{2}\big{)}\xrightleftharpoons[h_{2}]{h_{1}}\big{(}i,(\delta^{\prime}_{1},\delta_{2})\big{)}\xrightleftharpoons[id\times\Phi_{n}^{-1}]{id\times\Phi_{n}}\big{(}i,\Phi_{n}^{-1}(\delta^{\prime}_{1},\delta_{2})\big{)}.$ Note that $h$, $h_{1}$, and $h_{2}$ are the obvious recombination maps. Since all the functions we used to define the two mappings $Z_{0}$ and $Z_{1}$ are injective, both $Z_{0}$ and $Z_{1}$ are bijection mappings. ∎ In order to classify PADs with respect to their parity, we let $\mathfrak{D}_{n}^{e}$ and $\mathfrak{D}_{n}^{o}$ denote the set of even and odd PADs over $[n]$, respectively. Moreover, $\mathfrak{d}^{e}_{n}=|\mathfrak{D}_{n}^{e}|$ and $\mathfrak{d}^{o}_{n}=|\mathfrak{D}_{n}^{o}|$. Obviously, $\mathfrak{d}_{n}=\mathfrak{d}^{e}_{n}+\mathfrak{d}^{o}_{n}$. $n$ | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 ---|---|---|---|---|---|---|---|---|---|---|--- $\mathfrak{d}_{n}^{e}$ | 1 | 0 | 0 | 0 | 1 | 0 | 4 | 6 | 45 | 192 | 976 $\mathfrak{d}_{n}^{o}$ | 0 | 0 | 0 | 0 | 0 | 2 | 0 | 12 | 36 | 204 | 960 Table 6: First few values of the number of even and odd PADs. ###### Proposition 3.8. The numbers of even and odd PADs satisfy the relations $\displaystyle\mathfrak{d}^{e}_{n}=d^{e}_{\lfloor\frac{n}{2}\rfloor}d^{e}_{\lceil\frac{n}{2}\rceil}+d^{o}_{\lfloor\frac{n}{2}\rfloor}d^{o}_{\lceil\frac{n}{2}\rceil},$ $\displaystyle\mathfrak{d}^{o}_{n}=d^{e}_{\lfloor\frac{n}{2}\rfloor}d^{o}_{\lceil\frac{n}{2}\rceil}+d^{e}_{\lceil\frac{n}{2}\rceil}d^{o}_{\lfloor\frac{n}{2}\rfloor},$ for $n\geq 0$, with initial conditions $d^{e}_{0}=1$, $d^{e}_{1}=0$, $d^{o}_{0}=0$, and $d^{o}_{1}=0$. ###### Proof. Let $\delta$ be a PAD over $[n]$. Then, there exist $\delta_{1}\in D_{\lceil\frac{n}{2}\rceil}$ and $\delta_{2}\in D_{\lfloor\frac{n}{2}\rfloor}$ such that $\Phi_{n}(\delta)=(\delta_{1},\delta_{2})$. If $\delta\in\mathfrak{D}_{n}^{e}$, then $\delta_{1}$ and $\delta_{2}$ must have the same parity. Thus, $\mathfrak{d}^{e}_{n}=d^{e}_{\lfloor\frac{n}{2}\rfloor}d^{e}_{\lceil\frac{n}{2}\rceil}+d^{o}_{\lfloor\frac{n}{2}\rfloor}d^{o}_{\lceil\frac{n}{2}\rceil}$. If $\delta\in\mathfrak{D}_{n}^{e}$, then $\delta_{1}$ and $\delta_{2}$ must have opposite parities. Hence, $\mathfrak{d}^{o}_{n}=d^{e}_{\lfloor\frac{n}{2}\rfloor}d^{o}_{\lceil\frac{n}{2}\rceil}+d^{e}_{\lceil\frac{n}{2}\rceil}d^{o}_{\lfloor\frac{n}{2}\rfloor}$. ∎ ###### Corollary 3.9. The number of PADs with even parity and with odd parity satisfy the recurrence relations $\displaystyle\mathfrak{d}_{n}^{e}=s\left(\mathfrak{d}_{n-1}^{o}+(n-2-s)(\mathfrak{d}_{n-3}^{e}+\mathfrak{d}_{n-4}^{e})\right)$ (6) $\displaystyle\mathfrak{d}_{n}^{o}=s\left(\mathfrak{d}_{n-1}^{e}+(n-2-s)(\mathfrak{d}_{n-3}^{o}+\mathfrak{d}_{n-4}^{o})\right),$ (7) where $s=\big{\lfloor}\frac{n-1}{2}\big{\rfloor}=\frac{2n-3-(-1)^{n}}{4}$, for $n\geq 4$ with initial conditions $\mathfrak{d}_{0}^{e}=1$, $\mathfrak{d}_{0}^{o}=0$, and $\mathfrak{d}_{i}^{e}=\mathfrak{d}_{i}^{o}=0$ for $i=1,2,3$. ###### Proof. It is enough to clarify the effect of the bijection $\psi_{n}$ on the parity of a derangement over $[n]$, the rest is just applying the bijection $\Psi$ from the proof of Theorem 3.6. Letting $\delta$ be in $D_{n}$, $\delta^{\prime}$ has sign either $(-1)^{n-1-c}$ if $\delta\in D_{n}^{(1)}$, or $(-1)^{n-2-(c-1)}=(-1)^{n-1-c}$ if $\delta\in D_{n}^{(2)}$. Here $\delta^{\prime}$ is the derangement obtained from $\delta$ by applying $\psi_{n}$, and $c$ is the number of cycles in the cycle representation of $\delta$. This means, the bijection $\psi_{n}$ changes the parity of a derangement. ∎ ###### Definition 3.10. Let $\delta$ be a derangement over $[n]$ in standard cycle representation and let $C_{1}=(1\,\,a_{2}\,\,\cdots\,\,a_{m})$ be the first cycle. Following [3], an extraction point of $\delta$ is an entry $e\geq 2$ if $e$ is the smallest number in the set $\\{2,\ldots,n\\}\backslash\\{a_{2}\\}$ for which $C_{1}$ does not end with the numbers of $\\{2,\ldots,e\\}\backslash\\{a_{2}\\}$ written in decreasing order. The $(n-1)$ derangements, $\delta_{n,i}=(1\,\,i\,\,n\,\,n-1\,\,\cdots\,\,i+2\,\,i+1\,\,i-1\,\,i-2\,\,\cdots\,\,3\,\,2)$ for $i\in[2,n]$, that do not have extraction points are called the exceptional derangements and the set of exceptional derangements is denoted by $X_{n}$. Note that the extraction point (if it exists) must belong to the first or the second cycle. Following this approach we may introduce: ###### Definition 3.11. We call the PAD $\displaystyle\Phi_{n}^{-1}(\delta_{\lceil\frac{n}{2}\rceil,i},\,\delta_{\lfloor\frac{n}{2}\rfloor,j}),\text{ for }i\in\big{[}2,\lceil n/2\rceil\big{]}\text{ and }j\in\big{[}2,\lfloor n/2\rfloor\big{]}$ an exceptional PAD and we let $\mathcal{X}_{n}$ denote the set containing them. ###### Example 3.12. If $n=8$, then $\displaystyle\mathcal{X}_{8}$ $\displaystyle=\\{\Phi_{8}^{-1}(\delta_{4,2},\,\delta_{4,2}),\,\,\Phi_{8}^{-1}(\delta_{4,2},\,\delta_{4,3}),\,\,\Phi_{8}^{-1}(\delta_{4,2},\,\delta_{4,4}),\,\,\Phi_{8}^{-1}(\delta_{4,3},\,\delta_{4,2}),\,\,\Phi_{8}^{-1}(\delta_{4,3},\,\delta_{4,3}),$ $\displaystyle\quad\quad\Phi_{8}^{-1}(\delta_{4,3},\,\delta_{4,4}),\,\,\Phi_{8}^{-1}(\delta_{4,4},\,\delta_{4,2}),\,\,\Phi_{8}^{-1}(\delta_{4,4},\,\delta_{4,3}),\,\,\Phi_{8}^{-1}(\delta_{4,4},\,\delta_{4,4})\\}$ $\displaystyle=\\{(1\,3\,7\,5)(2\,4\,8\,6),\,\,(1\,3\,7\,5)(2\,6\,8\,4),\,\,(1\,3\,7\,5)(2\,8\,6\,4),\,\,(1\,5\,7\,3)(2\,4\,8\,6),\,\,(1\,5\,7\,3)$ $\displaystyle\quad\quad(2\,6\,8\,4),\,\,(1\,5\,7\,3)(2\,8\,6\,4),\,\,(1\,7\,5\,3)(2\,4\,8\,6),\,\,(1\,7\,5\,3)(2\,6\,8\,4),\,\,(1\,7\,5\,3)(2\,8\,6\,4)\\}.$ ###### Remark 3.13. Since the exceptional derangements over $[n]$ have sign $(-1)^{n-1}$, all the exceptional PADs in $\mathcal{X}_{n}$ have the same parity, with sign $(-1)^{n-2}=(-1)^{n}$. ###### Remark 3.14. As it was proved in [3], the number of the exceptional derangements in $X_{n}$ is the difference of the number of even and odd derangements, i.e., $d_{n}^{e}-d_{n}^{o}=(-1)^{n-1}(n-1)$. Chapman ([5]) also provide a bijective proof for the same formula. Below we give the idea of the proof due to Benjamin, Bennett, and Newberger ([3]). Let $f_{n}$ be the involution on $D_{n}\backslash X_{n}$ defined by $\displaystyle f_{n}(\pi)=f_{n}\big{(}(1\,\,a_{2}\,\,X\,\,e\,\,Y\,\,Z)\,\pi^{\prime}\big{)}=(1\,\,a_{2}\,\,X\,\,Z)(e\,\,Y)\,\pi^{\prime}$ for $\pi$ in $D_{n}\backslash X_{n}$ with the extraction point $e$ in the first cycle; and vice versa for the other $\pi$ in $D_{n}\backslash X_{n}$ with the extraction point $e$ in the second cycle. $a_{2}$ is the second element in the first cycle of $\pi$; $X$, $Y$, and $Z$ are ordered subsets of $[n]$, $Y\neq\emptyset$ and $Z$ consist the elements of $\\{2,3,\ldots,e-1\\}\backslash\\{a_{2}\\}$ written in decreasing order, and $\pi^{\prime}$ is the rest of the derangement in $\pi$. Since the number of the cycles in $\pi$ and $f_{n}(\pi)$ differ by one, they must have opposite parity. Labeling $\mathfrak{d}_{n}^{e}{-}\mathfrak{d}_{n}^{o}$ as $\mathfrak{f}_{n}$, we have the following result. ###### Proposition 3.15. The difference $\mathfrak{f}_{n}$ counts the number of exceptional PADs over $[n]$ and its closed formula is given by $\displaystyle\mathfrak{f}_{n}=(-1)^{n-2}\Big{\lceil}\frac{n-2}{2}\Big{\rceil}\Big{\lfloor}\frac{n-2}{2}\Big{\rfloor}.$ (8) ###### Proof. Let $\delta$ be in $\mathfrak{D}_{n}\backslash\mathcal{X}_{n}$. Then $\Phi_{n}$ map $\delta$ with the pair $(\delta_{1},\delta_{2})$. Define a mapping $F$ from $\mathfrak{D}_{n}\backslash\mathcal{X}_{n}$ to itself as $\displaystyle F(\delta)=\begin{cases}\Phi_{n}^{-1}\left(f_{\lceil\frac{n}{2}\rceil}(\delta_{1}),\,\delta_{2}\right)\text{ if }n\text{ is odd}\\\ \Phi_{n}^{-1}\left(\delta_{1},\,f_{\lfloor\frac{n}{2}\rfloor}(\delta_{2})\right)\text{ otherwise}\end{cases}$ Since $f_{n}$ is a bijection and changes parity, $F$ is a bijection and also $\delta$ and $F(\delta)$ have opposite parity. The leftovers, which are the PADs in $\mathcal{X}_{n}$ with sign $(-1)^{n-2}$, are counted by $\lceil\frac{n-2}{2}\rceil\lfloor\frac{n-2}{2}\rfloor$. ∎ $n$ | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 ---|---|---|---|---|---|---|---|---|---|---|---|---|--- $\mathfrak{f}_{n}$ | 1 | 0 | 0 | 0 | 1 | -2 | 4 | -6 | 9 | -12 | 16 | -20 | 25 Table 7: First few values the difference $\mathfrak{f}_{n}=\mathfrak{d}_{n}^{e}-\mathfrak{d}_{n}^{o}$. This sequence looks like an alternating version of A002620, to which Paul Barry constructed an EGF. From [2], we learned that he used Mathematica to generate the formulas by taking the Inverse Laplace Transform of the ordinary generating function described in [1]. However, we propose the following more direct, constructive proof. ###### Theorem 3.16. The exponential generating function of the difference $\mathfrak{f}_{n}$ has the closed form $\displaystyle\frac{e^{x}}{8}+\frac{e^{-x}}{8}(2x^{2}+6x+7).$ ###### Proof. From the closed formula of $\mathfrak{f}_{n}$ in Proposition 3.15, we have $\displaystyle(n-1)^{2}=\mathfrak{f}_{2n}$ $\displaystyle=\mathfrak{d}_{2n}^{e}-\mathfrak{d}_{2n}^{o}$ $\displaystyle n(n-1)=\mathfrak{f}_{2n+1}$ $\displaystyle=-(\mathfrak{d}_{2n+1}^{e}-\mathfrak{d}_{2n+1}^{o}).$ Hence, $\displaystyle\sum_{n\geq 0}\mathfrak{f}_{2n}\frac{x^{n}}{(2n)!}=\sum_{n\geq 0}(n-1)^{2}\frac{x^{2n}}{(2n)!}=\frac{x^{2}-3x+4}{8}e^{x}+\frac{x^{2}+3x+4}{8}e^{-x},\quad\text{ and }$ $\displaystyle\sum_{n\geq 0}\mathfrak{f}_{2n+1}\frac{x^{2n+1}}{(2n+1)!}=-\sum_{n\geq 0}n(n-1)\frac{x^{n}}{n!}=-\frac{x^{2}-3x+3}{8}e^{x}+\frac{x^{2}+3x+3}{8}e^{-x}$ Thus, $\displaystyle\sum_{n\geq 0}\mathfrak{f}_{2n}\frac{x^{2n}}{(2n)!}+\sum_{n\geq 0}\mathfrak{f}_{2n+1}\frac{x^{2n+1}}{(2n+1)!}$ $\displaystyle=\frac{e^{x}}{8}+\frac{e^{-x}}{8}(2x^{2}+6x+7)$ is the desired formula. ∎ ## 4 Excedance distribution over PADs In this section, we focus on excedance distribution in PADs. ###### Definition 4.17. We say that a permutation $\sigma$ has an excedance on $i\in[n]$ if $\sigma(i)>i$. In this case, $i$ is said to be an excedant. We give the notation below in the study of this property: $\displaystyle\mathfrak{d}_{n,k}$ $\displaystyle=|\\{\delta\in\mathfrak{D}_{n}:\,\,\delta\text{ has }k\text{ excedances}\\}|,$ $\displaystyle\mathfrak{d}_{n,k}^{o}$ $\displaystyle=|\\{\delta\in\mathfrak{D}^{o}(n):\,\,\delta\text{ has }k\text{ excedances}\\}|,$ $\displaystyle\mathfrak{d}_{n,k}^{e}$ $\displaystyle=|\\{\delta\in\mathfrak{D}^{e}(n):\,\,\delta\text{ has }k\text{ excedances}\\}|.$ Mantaci and Rakotondrajao ([6]) studied the excedance distribution in derangements, i.e., the numbers $\displaystyle d_{n,k}$ $\displaystyle=|\\{\delta\in D_{n}:\,\,\delta\text{ has }k\text{ excedances}\\}|,$ $\displaystyle d^{e}_{n,k}$ $\displaystyle=|\\{\delta\in D_{n}:\,\,\delta\text{ is an even derangemnt and }\text{ has }k\text{ excedances}\\}|,$ $\displaystyle d^{o}_{n,k}$ $\displaystyle=|\\{\delta\in D_{n}:\,\,\delta\text{ is an odd derangemnt and }\text{ has }k\text{ excedances}\\}|.$ ###### Remark 4.18. Since the number of excedances in a derangement over $[n]$ is in the range $[1,n-1]$, the number of excedances of a PAD over $[n]$ is at least 2 and at most $n-2$. ###### Proposition 4.19. The numbers $\mathfrak{d}_{n,k}$, $\mathfrak{d}_{n,k}^{e}$, and $\mathfrak{d}_{n,k}^{o}$ are symmetric, that is $\displaystyle\mathfrak{d}_{n,k}=\mathfrak{d}_{n,n-k},\,\,\mathfrak{d}^{e}_{n,k}=\mathfrak{d}^{e}_{n,n-k},\text{ and }\,\,\mathfrak{d}^{o}_{n,k}=\mathfrak{d}^{o}_{n,n-k}.$ ###### Proof. The bijection from $\mathfrak{D}_{n}$ to it self, defined as $\delta\mapsto\delta^{-1}$ for $\delta\in\mathfrak{D}_{n}$, associates a PAD having $k$ excedances with a PAD having $n-k$ exeedances and also preserves parity. ∎ | | | $\mathfrak{d}_{n,k}$ | | | | ---|---|---|---|---|---|---|--- $n\setminus k$ | 2 | 3 | 4 | 5 | 6 | 7 | 4 | 1 | | | | | | 5 | 1 | 1 | | | | | 6 | 1 | 2 | 1 | | | | 7 | 1 | 8 | 8 | 1 | | | 8 | 1 | 14 | 51 | 14 | 1 | | 9 | 1 | 28 | 169 | 169 | 28 | 1 | 10 | 1 | 42 | 483 | 884 | 483 | 42 | 1 Table 8: First few terms of the number of PADs in terms of number of excedances | | | $\mathfrak{d}^{e}_{n,k}$ | | | | ---|---|---|---|---|---|---|--- $n\setminus k$ | 2 | 3 | 4 | 5 | 6 | 7 | 4 | 1 | | | | | | 5 | 0 | 0 | | | | | 6 | 1 | 2 | 1 | | | | 7 | 0 | 3 | 3 | 0 | | | 8 | 1 | 8 | 27 | 8 | 1 | | 9 | 0 | 13 | 83 | 83 | 13 | 0 | 10 | 1 | 22 | 243 | 444 | 243 | 22 | 1 | | | $\mathfrak{d}^{o}_{n,k}$ | | | | ---|---|---|---|---|---|---|--- $n\setminus k$ | 2 | 3 | 4 | 5 | 6 | 7 | 4 | 0 | | | | | | 5 | 1 | 1 | | | | | 6 | 0 | 0 | 0 | | | | 7 | 1 | 5 | 5 | 1 | | | 8 | 0 | 6 | 24 | 6 | 0 | | 9 | 1 | 15 | 86 | 86 | 15 | 1 | 10 | 0 | 20 | 240 | 440 | 240 | 20 | 0 Table 9: First few values of $\mathfrak{d}_{n,k}$ in terms of their parity ###### Proposition 4.20. The excedance distribution of a PAD is given by $\displaystyle\mathfrak{d}_{n,k}=\begin{cases}\sum_{i=1}^{k-1}d_{\lceil\frac{n}{2}\rceil,i}\,d_{\lfloor\frac{n}{2}\rfloor,k-i},\,\,\text{ if }\,2\leq k\leq\lfloor\frac{n}{2}\rfloor\\\ \sum_{i=1}^{n-k-1}d_{\lceil\frac{n}{2}\rceil,i}\,d_{\lfloor\frac{n}{2}\rfloor,n-k+i},\,\,\text{ if }\,\lfloor\frac{n}{2}\rfloor<k\leq n{-}2\end{cases}.$ ###### Proof. To find the number of excedances of a PAD $\delta$ over $[n]$, we sum up the number of exeedances in $\delta_{1}$ and in $\delta_{2}$, where $(\delta_{1},\,\delta_{2})$ is the image of $\delta$ defined in $\Phi_{n}$. Since there are $d_{m,i}$ derangements in $D_{m}$ having $i$ excedances, for $i\in[1,\,\,m-1]$, the products $d_{m,i}\,d_{l,k-i}$, for $m=\lceil\frac{n}{2}\rceil$ and $l=\lfloor\frac{n}{2}\rfloor$, determine the number of PADs over $[n]$ having $k$ excedances. Summing up the products over the range $i=1,2,\ldots,k{-}1$ will give the first formula. The second formula follows from Proposition 4.19. ∎ ###### Corollary 4.21. The excedance distribution of PADs in terms of their parity is given by: $\displaystyle\mathfrak{d}_{n,k}^{e}=\begin{cases}\sum_{i=1}^{k-1}(d^{e}_{\lceil\frac{n}{2}\rceil,i}\,d^{e}_{\lfloor\frac{n}{2}\rfloor,k-i}+d^{o}_{\lceil\frac{n}{2}\rceil,i}\,d^{o}_{\lfloor\frac{n}{2}\rfloor,k-i}),\,\,\text{ if }\,2\leq k\leq\lfloor n/2\rfloor\\\ \sum_{i=1}^{n-k-1}(d^{e}_{\lceil\frac{n}{2}\rceil,i}\,d^{e}_{\lfloor\frac{n}{2}\rfloor,n-k+i}+d^{o}_{\lceil\frac{n}{2}\rceil,i}\,d^{o}_{\lfloor\frac{n}{2}\rfloor,n-k+i}),\,\,\text{ if }\,\lfloor n/2\rfloor<k\leq n{-}2\end{cases},$ $\displaystyle\mathfrak{d}_{n,k}^{o}=\begin{cases}\sum_{i=1}^{k-1}(d^{e}_{\lceil\frac{n}{2}\rceil,i}\,d^{o}_{\lfloor\frac{n}{2}\rfloor,k-i}+d^{o}_{\lceil\frac{n}{2}\rceil,i}\,d^{e}_{\lfloor\frac{n}{2}\rfloor,k-i}),\,\,\text{ if }\,2\leq k\leq\lfloor n/2\rfloor\\\ \sum_{i=1}^{n-k-1}(d^{e}_{\lceil\frac{n}{2}\rceil,i}\,d^{o}_{\lfloor\frac{n}{2}\rfloor,n-k+i}+d^{o}_{\lceil\frac{n}{2}\rceil,i}\,d^{e}_{\lfloor\frac{n}{2}\rfloor,n-k+i}),\,\,\text{ if }\,\lfloor n/2\rfloor<k\leq n{-}2\end{cases}.$ $\square$ An immediate consequence of this Corollary is ###### Proposition 4.22. We have $\displaystyle\mathfrak{f}_{n,k}=\mathfrak{d}_{n,k}^{e}-\mathfrak{d}_{n,k}^{o}=(-1)^{n}\,\max\\{k-1,\,n{-}(k+1)\\}$ for $n\geq 4$ and $2\leq k\leq n{-}2$. ###### Proof. Mantaci and Rakotondrajao (see [6]) have proved the identity $d^{o}_{n,k}-d^{e}_{n,k}=(-1)^{n}$ using recursive argument. Applying this with the Corollary 4.21, we get the desired formula. ∎ | | | $\mathfrak{f}_{n,k}$ | | | | ---|---|---|---|---|---|---|--- $n\setminus k$ | 2 | 3 | 4 | 5 | 6 | 7 | 4 | 1 | | | | | | 5 | -1 | -1 | | | | | 6 | 1 | 2 | 1 | | | | 7 | -1 | -2 | -2 | -1 | | | 8 | 1 | 2 | 3 | 2 | 1 | | 9 | -1 | -2 | -3 | -3 | -2 | -1 | 10 | 1 | 2 | 3 | 4 | 3 | 2 | 1 Table 10: The first few values of the difference $\mathfrak{d}^{e}_{n,k}-\mathfrak{d}^{o}_{n,k}$. ###### Theorem 4.23. The exponential generating function for the sequence $\\{\mathfrak{f}_{n,k}\\}$ has the closed form $\displaystyle\frac{1}{(1-u)^{2}}\left(u^{2}e^{-x}+e^{-ux}-2u\cosh{\sqrt{u}x}+\frac{u+u^{2}}{\sqrt{u}}\sinh{\sqrt{u}x}-(1-u)^{2}\right).$ ###### Proof. Let $\mathfrak{f}_{n}(u)=\sum_{k=2}^{n-2}\mathfrak{f}_{n,k}u^{k}$ and $\mathfrak{f}(x,u)=\sum_{n\geq 4}\mathfrak{f}_{n}(u)\frac{x^{n}}{n!}$. From Proposition 4.22, we have $\displaystyle\mathfrak{f}_{2m,k}$ $\displaystyle=\begin{cases}k-1,\quad\text{if }2\leq k\leq m\\\ 2m-k-1,\quad\text{if }m<k\leq 2m{-}2\end{cases},$ $\displaystyle\mathfrak{f}_{2m+1,k}$ $\displaystyle=\begin{cases}-(k-1),\quad\text{if }2\leq k\leq m\\\ -(2m-k),\quad\text{if }m<k\leq 2m{-}1\end{cases},$ for $m\geq 2$. So, $\displaystyle\mathfrak{f}_{2m}(u)$ $\displaystyle=\sum_{k=2}^{m}(k-1)u^{k}+\sum_{k=m+1}^{2m-2}(2m-k-1)u^{k}=\frac{u^{2}-2u^{m+1}+u^{2m}}{(1-u)^{2}},$ $\displaystyle\mathfrak{f}_{2m+1}(u)$ $\displaystyle=\sum_{k=2}^{m}-(k-1)u^{k}+\sum_{k=m+1}^{2m-1}-(2m-k)u^{k}=\frac{u^{m+2}+u^{m+1}-u^{2}-u^{2m+1}}{(1-u)^{2}},$ $\displaystyle\mathfrak{f}(x,u)$ $\displaystyle=\sum_{m\geq 2}\mathfrak{f}_{2m}(u)\frac{x^{2m}}{(2m)!}+\sum_{m\geq 2}\mathfrak{f}_{2m+1}(u)\frac{x^{2m+1}}{(2m+1)!}$ $\displaystyle=\frac{1}{(1-u)^{2}}\left(u^{2}e^{-x}+e^{-ux}-2u\cosh{\sqrt{u}x}+\frac{u+u^{2}}{\sqrt{u}}\sinh{\sqrt{u}x}-(1-u)^{2}\right).$ ∎ ## Methodological remarks In this paper, most of our results are obtained in a way of splitting the permutations into two subwords. However, this method is not always applicable. One example is the number of PADs avoiding the pattern $p=1\,2$. The only derangement that avoid $p$ is $(1\,\,n\,)(2\,\,\,n{-}1\,)\cdots(\frac{n}{2}\,\,\,\frac{n+2}{2})$, for even $n$, that is, the derangement over $[n]$ with entries in decreasing order when written in linear representation. However, it does not exist if $n$ is odd, since $\frac{n+1}{2}$ is a fixed point. The PAD $\delta$ created from a pair $(\delta_{1},\delta_{2})$, by the mapping $\Phi_{n}^{-1}$, of two even length derangements that both avoids the pattern $p$ is $\delta=(1\,\,n{-}1\,)(3\,\,\,n{-}3\,)\cdots\big{(}\frac{n-2}{2}\,\,\,\frac{n+2}{2}\big{)}(2\,\,n\,)(4\,\,\,n{-}2\,)\cdots\big{(}\frac{n}{2}\,\,\,\frac{n+4}{2}\big{)}$, which is $n{-}1\,\,n\,\,n{-}3\,\,n{-}2\cdots 3\,\,4\,\,1\,\,2$ in linear form, has length $n\equiv 0\pmod{4}$. However, each pair $i\,\,i{+}1$, where $i$ is an entry in odd position, is a subword with the occurrence of the pattern $p$ in $\delta$. This indicates that $\delta_{1}$ and $\delta_{2}$ avoid $p$ but $\delta$ doesn’t. Things get even more complicated with patterns of length greater than 2. Final Remarks: As for now, we have not been successful in finding the recurrence relations and generating functions for the sequences $\\{\mathfrak{d}_{n,k}\\}_{n=0}^{\infty}$, $\\{\mathfrak{d}_{n,k}^{e}\\}_{n=0}^{\infty}$, and $\\{\mathfrak{d}_{n,k}^{o}\\}_{n=0}^{\infty}$. ## Acknowledgements The first author acknowledges the financial support extended by the cooperation agreement between International Science Program at Uppsala University and Addis Ababa University. Special thanks go to Prof. Jörgen Backelin, Prof. Paul Vaderlind and Dr. Per Alexandersson of Stockholm University - Dept. of Mathematics, for all their valuable inputs and suggestions. Many thanks to our colleagues from CoRS - Combinatorial Research Studio, for lively discussions and comments. ## References * [1] Paul Barry. On a central transform of integer sequences. arXiv preprint arXiv:2004.04577, 2020. * [2] Paul Barry. Private communication to the first author, $14^{th}$, December 2020. * [3] Arthur T Benjamin, Curtis T Bennett, and Florence Newberger. Recounting the odds of an even derangement. Mathematics Magazine, 78(5):387–390, 2005. * [4] Miklós Bóna. Combinatorics of permutations. CRC Press, 2012. * [5] Robin Chapman. An involution on derangements. Discrete Mathematics, 231(1):121–122, 2001. * [6] Roberto Mantaci and Fanja Rakotondrajao. Exceedingly deranging! Advances in Applied Mathematics, 30(1-2):177–188, 2003\. * [7] Fanja Rakotondrajao. k-fixed-points-permutations. Integer: Electronic Jornal of Combinatorial Number Theory, 7(A36):A36, 2007. * [8] Richard P Stanley. Enumerative combinatorics volume 1 second edition. Cambridge studies in advanced mathematics, 2011. * [9] Shinji Tanimoto. Combinatorics of the group of parity alternating permutations. Advances in Applied Mathematics, 44(3):225–230, 2010. * [10] Shinji Tanimoto. Parity alternating permutations and signed eulerian numbers. Annals of Combinatorics, 14(3):355–366, 2010.
# Will Artificial Intelligence supersede Earth System and Climate Models? Christopher Irrgang Helmholtz Centre Potsdam, German Research Centre for Geosciences GFZ, Potsdam, Germany Niklas Boers Department of Mathematics and Computer Science, Free University of Berlin, Germany Potsdam Institute for Climate Impact Research, Potsdam, Germany Department of Mathematics and Global Systems Institute, University of Exeter, UK Maike Sonnewald Program in Atmospheric and Oceanic Sciences, Princeton University, Princeton, NJ 08540, USA NOAA/OAR Geophysical Fluid Dynamics Laboratory, Ocean and Cryosphere Division, Princeton, NJ 08540, USA University of Washington, School of Oceanography, Seattle, WA, USA Elizabeth A. Barnes Colorado State University, Fort Collins, USA Christopher Kadow German Climate Computing Center DKRZ, Hamburg, Germany Joanna Staneva Helmholtz-Zentrum Geesthacht, Center for Material and Coastal Research HZG, Geesthacht, Germany Jan Saynisch-Wagner Helmholtz Centre Potsdam, German Research Centre for Geosciences GFZ, Potsdam, Germany ###### Abstract We outline a perspective of an entirely new research branch in Earth and climate sciences, where deep neural networks and Earth system models are dismantled as individual methodological approaches and reassembled as learning, self-validating, and interpretable Earth system model-network hybrids. Following this path, we coin the term ”Neural Earth System Modelling” (NESYM) and highlight the necessity of a transdisciplinary discussion platform, bringing together Earth and climate scientists, big data analysts, and AI experts. We examine the concurrent potential and pitfalls of Neural Earth System Modelling and discuss the open question whether artificial intelligence will not only infuse Earth system modelling, but ultimately render them obsolete. For decades, scientists have utilized mathematical equations to describe geophysical and climate processes and to construct deterministic computer simulations that allow for the analysis of such processes. Until recently, process-based models had been considered irreplaceable tools that helped to understand the complex interactions in the coupled Earth system and that provided the only tools to predict the Earth system’s response to anthropogenic climate change. The provocative thought that Earth system models (ESMs) might lose their fundamental importance in the advent of novel artificial intelligence (AI) tools has sparked both a gold-rush feeling and contempt in the scientific communities. On the one hand, deep neural networks have been developed that complement and have started to outperform the skill of process-based models in various applications, ranging from numerical weather prediction to climate research. On the other hand, most neural networks are trained for isolated applications and lack true process knowledge. Regardless, the daily increasing data streams from Earth system observation (ESO), increasing computational resources, and the availability and accessibility of powerful AI tools, particularly in machine learning (ML), have led to numerous innovative frontier applications in Earth and climate sciences. Based on the current state, recent achievements, and recognised limitations of both process-based modelling and AI in Earth and climate research, we draw a perspective on an imminent and profound methodological transformation, hereafter named Neural Earth System Modelling (NESYM). To solve the emerging challenges, we highlight the necessity of new transdisciplinary collaborations between the involved communities. ## Overview on Earth System Modelling and Earth System Observations Earth system models (ESMs)1 combine process-based models of the different sub- systems of the Earth system into an integrated numerical model that yields for a given state of the coupled system at time $t$ the tendencies associated with that state, i.e., a prediction of the system state for time $t+1$. The individual model components, or modules, describe sub-systems including the atmosphere, the oceans, the carbon and other biogeochemical cycles, radiation processes, as well as land surface and vegetation processes and marine ecosystems. These modules are then combined by a dynamic coupler to obtain a consistent state of the full system for each time step. For some parts of the Earth system, the primitive physical equations of motion are known explicitly, such as the Navier-Stokes equations that describe the fluid dynamics of the atmosphere and oceans (Fig. 1). In practise, it is impossible to numerically resolve all relevant scales of the dynamics and approximations have to be made. For example, the fluid dynamical equations for the atmosphere and oceans are integrated on discrete spatial grids, and all processes that operate below the grid resolution have to be parameterised to assure a closed description of the system. Since the multi-scale nature of the dynamics of geophysical fluids implies that the subgrid-scale processes interact with the larger scales that are resolved by the model, (stochastic) parameterization of subgrid-scale processes is a highly non-trivial, yet unavoidable, part of climate modelling 2, 3, 4. Figure 1: Symbolic representation of Earth system components and exemplary deterministic or stochastic coupling mechanisms on long and short spatio- temporal scales. For other parts of the Earth system, primitive equations of motion, such as the Navier-Stokes equations for atmospheric motion, do not exist. Essentially, this is due to the complexity of the Earth system, where many phenomena that emerge at a macroscopic level are not easily deducible from microscopic-scales that may or may not be well-understood. A typical example is given by ecosystems and the physiological processes governing the vegetation that covers vast parts of the land surface, as well as their interactions with the atmosphere, the carbon and other geochemical cycles. Also for these cases, approximations in terms of parameterizations of potentially crucial processes have to be made. Regardless of the specific process, such parameterizations induce free parameters in ESMs, for which suitable values have to be found empirically. The size of state-of-the-art ESMs mostly prohibits systematic calibration methods such as, e.g., the ones based on Bayesian inference, and the models are therefore often tuned manually. The quality of the calibration as well as the overall accuracy of the model can only be assessed with respect to relatively sparse observations of the last 170 years, at most, and there is no way to assess the models’ skill in predicting future climate conditions5. The inclusion of free parameters possibly causes biases or structural model errors and the example of the discretized spatial grid suggests that the higher the spatial resolution of an ESM, the smaller the potential errors. Likewise, it is expected that the models’ representation of the Earth system will become more accurate the more processes are resolved explicitly. The inclusion of a vastly increasing number of processes, together with continuously rising spatial resolution, have indeed led to the development of comprehensive ESMs that have become irreplaceable tools to analyse and predict the state of the Earth system. From the first assessment report of the Intergovernmental Panel on Climate Change (IPCC) in 1990 to the fifth phase of the Climate Model Intercomparison Project (CMIP5)6 and the associated fifth IPCC assessment report in 2014, the spatial resolution has increased from around 500km to up to 70km. In accordance, the CMIP results show that the models have, over the course of two decades, greatly improved in their accuracy to reproduce crucial characteristics of the Earth system, such as the evolution of the global mean temperatures (GMT) since the beginning of instrumental data in the second half of the 19th century, or the average present-day spatial distribution of temperature or precipitation 7, 8. Despite the tremendous success of ESMs, persistent problems and uncertainties remain: (1) A crucial quantity for the evaluation of ESMs is the equilibrium climate sensitivity (ECS), defined as the amount of equilibrium GMT increase that results from an instantaneous doubling of atmospheric carbon dioxide 9. There remains a large ECS range in current ESM projections and reducing these uncertainties, and hence the uncertainties of future climate projections, is one of the key challenges in the development of ESMs. Nevertheless, from CMIP5 to CMIP6, the likely range of ECS has widened from $2.$1–$4.7^{\circ}$C to $1.8$–$5.6^{\circ}$C10, 11. A highly promising line of research in this regard focuses on the identification of emergent constraints, which in principle allow to narrow down the projected range for a model variable of interest, given that the variable has a concise relationship with another model variable that can be validated against past observations 12, 13. The development of suitable data-driven techniques for this purpose is still in its infancy. (2) Both theoretical considerations and paleoclimate data suggest that several sub-systems of the Earth system can abruptly change their state in response to gradual changes in forcing14, 15. There is concern that current ESMs will not be capable of predicting future abrupt climate changes, because the instrumental era of less than two centuries has not experienced comparable transitions, and model validation against paleoclimate data evidencing such events remains impossible due to the length of the relevant time scales16. In an extensive search, many relatively abrupt transitions have been identified in future projections of CMIP5 models17, but due to the nature of these rare, high-risk events, the accuracy of ESM in predicting them remains untested. (3) Current ESMs are not yet suitable for assessing the efficacy or the environmental impact of carbon dioxide removal techniques, which are considered key mitigation options in pathways realizing the Paris Agreement 18. Further, ESMs still insufficiently represent key environmental processes such as the carbon cycle, water and nutrient availability, or interactions between land use and climate. This can impact the usefulness of land-based mitigation options that rely on actions such as biomass energy with carbon capture and storage or nature-based climate solutions 19, 20. (4) The distributions of time series encoding Earth system dynamics typically exhibit heavy tails. Extreme events such as heat waves and droughts, but also extreme precipitation events and associated floods, have always caused tremendous socio-economic damages. With ongoing anthropogenic climate change, such events are projected to become even more severe, and the attribution of extremes poses another outstanding challenge of Earth system science21. While current ESMs are very skilful in predicting average values of climatic quantities, there remains room for improvement in representing extremes. In addition to the possible solutions to these fundamental challenges, improvements of the overall accuracy of ESMs can be expected from more extensive and more systematic integration of the process-based numerical models with observational data. Earth system observations (ESOs) are central to ESMs, serving a multitude of purposes. ESOs are used to evaluate and compare process-based model performance, to generate model parameters and initial model states, or as boundary forcing of ESMs 22, 23. ESOs are also used to directly influence the model output by either tuning or nudging parameters that describe unmodeled processes, or by the more sophisticated methods of data assimilation that alter the model’s state variables to bring the model output in better agreement with the observations24. To incorporate uncertainty into model predictions, variational interfaces have been used25. Existing techniques for assimilating data into ESMs fall into two main categories, each with their own limitations. Gradient-based optimization, as in four-dimensional variational (4DVar) schemes, is the current state of the art for efficiency and accuracy, but currently requires time consuming design and implementation of adjoint calculation routines tailored to each model. Ensemble-based Kalman filter (EnKF) schemes are gradient-free but produce unphysical outputs and rely on strong statistical assumptions that are often unsatisfied, leading to biases and overconfident predictions26. The main problems of contemporary ESM data assimilation are 1) nonlinear dynamics and non-Gaussian error budgets in combination with the high dimensionality of many ESM components 27, 28, 29, and 2) constraining the governing processes over the different spatio-temporal scales found in coupled systems 30, 31. ML approaches can be used to combine the accuracy of 4DVar with the flexibility of EnKF, essentially allowing optimization-based assimilation in cases where gradients are currently unavailable. Furthermore, these traditional approaches of model and observation fusion have slowly been expanded or replaced by ML methods in recent years 32, 33, 34. ESOs cover a wide range of spatio-temporal scales and types, ranging from a couple of centimeters to tens of thousands of kilometers, and from seconds and decades to millennia. The types of observations range from in-situ measurements of irregular times and spaces (ship cruises, buoy arrays, etc.), over single time series (ice and sediment cores, tide gauges, etc.) to satellite-based global 2D or 3D data fields (altimetry, gravimetry, radio occultation, etc.). The amount of available observations is rapidly increasing and has reached a threshold where automated analysis becomes crucial. Yet, the available observational data pool still contains large gaps in time and space that prevent building a holistic observation-driven picture of the coupled Earth system, which result from insufficient spatio-temporal data resolution, too short observation time periods, and largely unobserved compartments of Earth systems like, for instance, abyssal oceans. The combination of these complex characteristics render Earth system observations both challenging and particularly interesting for AI applications. ## From Machine Learning-based Data Exploration Towards Learning Physics ML and other AI techniques have achieved stunning results in computer vision 35, speech and language models 36, medical science 37, 38, economical and societal analytics 39, and other disciplines 40, 41. Due to this wide-spread integration into both fundamental research and end-user products, and despite shortcomings and inherent limitations 42, 43, 44, 45, ML is already praised as a key disruptive technology of the 21st century 46. In contrast, the usage of ML in Earth and climate sciences is still in its infancy. A key observation is that ML concepts from computer vision and automated image analysis can be isomorphically transferred to ESO imagery. Pioneering studies demonstrated the feasibility of ML for remote sensing data analysis, classification tasks, and parameter inversion already in the 1990s 47, 48, 49, 50, and climate-model emulation in the early 2000s 51. The figurative Cambrian explosion of AI techniques in Earth and climate sciences, however, only began over the last five years and will rapidly continue throughout the coming decades. Under the overarching topic of ESO data exploration, ML has been applied for a huge variety of statistical and visual use cases. Classical prominent examples are pattern recognition in geo-spatial observations, climate data clustering, automated remote-sensing data analysis, and time series prediction 52, 32. In this context, ML has been applied across various spatial and temporal scales, ranging from short-term regional weather prediction to Earth-spanning climate phenomena. Significant progress has been made in developing purely data-driven weather prediction networks, which start to compete with process-based model forecasts 53, 54, 55. ML contributed to the pressing need to improve the predictability of natural hazards, for instance, by uncovering global extreme- rainfall teleconnections 56, or by improving long-term forecasts of the El Niño Southern Oscillation (ENSO)57, 58. ML-based image filling techniques were utilized to reconstruct missing climate information, allowing to correct previous global temperature records 59. Furthermore, ML was applied to analyze climate data sets, e.g., to extract specific forced signals from natural climate variability 60, 61 or to predict clustered weather patterns 62. In these applications, the ML tools function as highly specialized agents that help to uncover and categorize patterns in an automated way. A key methodological advantage of ML in comparison to covariance-based spatial analysis lies in the possibility to map nonlinear processes 63, 64. At the same time, such trained neural networks lack actual physical process knowledge, as they solely function through identifying and generalizing statistical relations by minimizing pre-defined loss measures for a specific task 65. Consequently, research on ML in Earth and climate science differs fundamentally from the previously described efforts of advancing ESMs in terms of methodological development and applicability. Concepts of utilizing ML not only for physics-blind data analyses, but also as surrogates and methodological extensions for ESMs have only recently started to shape 66. Scientists started pursuing the aim that ML methods learn aspects of Earth and climate physics, or at least plausibly relate cause and effect. The combination of ML with process-based modelling is the essential distinction from the previous ESO data exploration. Lifting ML from purely diagnosis-driven usage towards the prediction of geophysical processes will also be crucial for aiding climate change research and the development of mitigation strategies 67. Following this reasoning, ML methods can be trained with process-based model data to inherit a specific geophysical causation or even emulate and accelerate entire forward simulations. For instance, ML has been used in combination with ESMs and ESOs to invert space-borne oceanic magnetic field observations to determine the global ocean heat content 33. Similarly, a neural network has been trained with a continental hydrology model to recover high-resolution terrestrial water storage from satellite gravimetry 34. ML plays an important role for upscaling unevenly distributed carbon flux measurements to improve global carbon monitoring systems68. As such, the eddy covariance technique was combined with ML to measure the net ecosystem exchange of $\text{CO}_{2}$ between ecosystems and the atmosphere, offering a unique opportunity to study ecosystem responses to climate change 69. ML has shown remarkable success in representing subgrid-scale processes and other parameterizations of ESMs, given that sufficient training data were available. As such, neural networks were applied to approximate turbulent processes in ocean models 70 and atmospheric subgrid processes in climate models 71. Several studies highlight the potential for ML-based parameterization schemes 72, 73, 74, 75, 76, helping step-by-step to gradually remove numerically and human-induced simplifications and other biases of ESMs 77. While some well-trained ML tools and simple hybrids have shown higher predictive power than traditional state-of-the-art process-based models, only the surface of new possibilities, but also of new scientific challenges, has been scratched. So far, ML, ESMs, and ESO have largely been independent tools. Yet, we have reached the understanding that applications of physics-aware ML and model-network hybrids pose huge benefits by filling up niches where purely process-based models persistently lack reliability. ## The Fusion of Process-based Models and Artificial Intelligence Figure 2: Successive stages of the fusion process of Earth system models and artificial intelligence. The idea of hybrids of process-based and ML models is not new 78. So far, hybrids have almost exclusively been thought of as numerical models that are enhanced by ML to either improve the models’ performance in the sense of a useful metric, or to accelerate the forward simulation time in exchange for a decrease in simulation accuracy. Along with the general advance regarding the individual capabilities and limitations of ESMs and ML methods, respectively, also the understanding of how ML can enhance process-based modelling has evolved. This progress allows ML to take over more and more components of ESMs, gradually blending the so far strict distinction between process-based modelling and data-driven ML approaches. Even more so, entirely new methodological concepts are dawning that justify acknowledging Neural Earth System Modelling as a distinct research branch (Fig. 2). The long-term goal will be to consistently integrate the recently discovered advantages of ML into the already decade-long source of process knowledge in Earth system science. However, this evolution does not come without methodological caveats, which need to be investigated carefully. For the sake of comparability, we distinguish between weakly coupled NESYM hybrids, i.e., an ESM or AI technique benefits from information from the respective other, and strongly coupled NESYM hybrids, i.e., fully coupled model-network combinations that dynamically exchange information between each other. The emergent development of weak hybrids is predominantly driven by the aim to resolve the previously described ESM limitations, particularly unresolved and especially sub-grid scale processes. Neural networks can emulate such processes after careful training with simulation data from a high-resolution model that resolves the processes of interest, or with relevant ESO data. The next methodological milestone will be the integration of such trained neural networks into ESMs for operational usage. First tests have indicated that the choice of the AI technique, e.g., neural networks versus random forests, seems to be crucial for the implementation of learning parameterization schemes, as they can significantly deteriorate the ESM’s numerical stability 79. Thus, it is not only important to identify how neural networks can be trained to resolve ESM limitations, but also how such ML-based schemes can be stabilized in the model physics context and how their effect on the process-based simulation can be evaluated and interpreted 80. The limitations of ML-based parameterization approaches can vary widely for different problems or utilized models and, consequently, should be considered for each learning task individually 81. Nevertheless, several ideas have been proposed to stabilize ML parameterizations, e.g., by enforcing physical consistency through customized loss functions in neural networks and specific network architectures 82, 75, or by optimizing the considered high-resolution model training data 76. In addition, an ESM blueprint has been proposed, in which learning parameterizations can be targeted through searching an optimal fit of statistical measures between ESMs, observations, and high-resolution simulations 83. In this context, further efforts have been made to enhance an ESM not with ML directly, but in combination with a data assimilation system 24. For instance, emulating a Kalman filter scheme with ML has been investigated 84, 85, an ML-based estimation of atmospheric forcing uncertainties used as error covariance information in data assimilation has been proposed 86, as well as further types of Kalman-network hybrids 87, 88. In the second class of weak hybrids, the model and AI tasks are transposed, such that the information flow is directed from the model towards the AI tool. Here, neural networks are trained directly with model state variables, their trajectories, or with more abstract information like seasonal signals, interannual cycles, or coupling mechanisms. The goal of the ML application might not only be model emulation, but also inverting non-linear geophysical processes 33, learning geophysical causation 89, or predicting extreme events 90, 91. In addition to these inference and generalization tasks, a key question in this sub-discipline is whether a neural network can learn to outperform the utilized process-based trainer model in terms of physical consistency or predictive power. ESOs play a vital role in this context, as they can serve as additional training constraints for a neural network training, allowing it to build independent self-evaluation measures 34. The given examples generally work well for validation and prediction scenarios within the given training distribution. Out-of-distribution samples, in contrast, pose a huge challenge for supervised learning, which renders the “learning from the past” principle particularly ill-posed for prediction tasks in NESYM. As a consequence, purely data-driven AI methods will not be able to perform accurate climate projections on their own, because of the both naturally and anthropogenically induced non-stationarity of the climate and Earth system. Overcoming these limitations requires a deeper holistic integration in terms of strongly coupled hybrids and the consideration of further, less constrained training techniques like unsupervised training 92 and generative AI methods 93, 73, 94. For example, problems of pure AI methods with non-stationary training data can be attenuated by combining them with physical equations describing the changing energy-balance of the Earth system due to anthropogenic greenhouse-gas emissions 95. In addition, first steps towards physics-informed AI have been made by ML-based and data-driven discovery of physical equations 96 and by the implementation of neural partial differential equations 97, 98 into the context of climate modelling 99. Continuous maturing of the methodological fusion process will allow building hybrids of neural networks, ESMs and ESOs that dynamically exchange information. ESMs will soon utilize output from supervised and unsupervised neural networks to optimize their physical consistency and, in turn, feed back improved information content to the ML component. ESOs form another core element and function as constraining ground truth of the AI-infused process prediction. Similar to the adversarial game of generative networks 100, or coupling mechanisms in an ESM 101, also strongly coupled NESYM hybrids will require innovative interfaces that control the exchange of information that are, so far, not available. In addition, we formulate key characteristics and goals of this next stage: (1) Hybrids can better reproduce and predict out-of-distribution samples and extreme events, (2) hybrids perform constrained and consistent simulations that obey physical conservation laws despite potential shortcomings of the hybrids’ individual components, (3) hybrids include integrated adaptive measures for self-validation and self- correction, and (4) NESYM allows replicability and interpretability. We believe that cross-discipline collaborations between Earth system and AI scientists will become more important than ever to achieve these milestones. Frontier applications of Neural Earth system models are manifold. Yet, ultimately, NESYM hybrids need to drastically improve the current forecast limits of geophysical processes and contribute towards understanding the Earth’s susceptible state in a changing climate. Consequently, not only the fusion of ESM and AI will be in the research focus, but also AI interpretability and resolving the common notion of a black box (Fig. 3). ## Peering into the Black Box Figure 3: Qualitative comparison of isolated (AI - Artificial Intelligence, ESM - Earth System Model, ESO - Earth System Observation) and hybrid methodological approaches. The respective approaches are represented as trajectories in a meta space of hybrid coupling degree, interpretability, and prediction skill. The goal of Neural Earth System Modelling (NESYM) is to integrate the interpretability and to exceed the prediction skill of the respective isolated approaches. In this meta space, the necessary research increment to achieve this goal can be described through an increase in the degree of hybrid coupling (Fusion of AI and ESM) and an increase in interpretability (XAI - explainable AI, IAI - interpretable AI). ML has emerged as a set of methods based on the combination of statistics, applied mathematics and computer science, but it comes with a unique set of hurdles. Peering into the black box and understanding the decision making process of the ML method, termed explainable AI (XAI), is critical to the use of these tools. Especially in the physical sciences, adaptation of ML suffers from a lack of interpretability, particularly supervised ML. In contrast and in addition to XAI stands the call for interpretable AI (IAI), i.e., building specifically interpretable ML models from the beginning on, instead of explaining ML predictions through post-process diagnostics 102. Ensuring that what is ‘learned’ by the machine is physically tractable or causal, and not due to trivial coincidences 103, is important before ML tools are used, e.g., in an ESM setting targeted at decision making. Thus, explainability provides the user with trust in the ML output, improving its transparency. This is critical for ML use in the policy-relevant area of climate science as society is making it increasingly clear that understanding the source of AI predictive skill is of crucial importance104, 105. Ensuring the ML method is getting the right answers for the right reasons is essential given the transient nature of the climate system. As the climate continues to respond to anthropogenic climate change, NESYM will be required to make predictions of continually evolving underlying distributions and XAI/IAI will be critical to ensuring that the skill of the ML method can be explained, and inspire trust in its extrapolation to future climate regimes. There are many ML tools at our disposal, and XAI can assist researchers in choosing the optimal ML architecture, inputs, outputs, etc. By analyzing the decision making process, climate scientists will be able to better incorporate their own physical knowledge into the ML method, ultimately leading to improved results. Perhaps least appreciated in geoscientific applications thus far is the use of XAI to discover new science 106. When the ML method is capable of making a prediction, XAI allows us to ask “what did it learn?”. In this way, ML becomes more than just a prediction and allows scientists to ask “why?” as they normally would, but now with the power of ML. Explaining the source of an ML applications skill can be done retrospectively102. The power of XAI for climate and weather applications has very recently been demonstrated 107, 106, 108. For example, neural networks coupled with the XAI attribution method known as layerwise relevance propagation (LRP)109, 110 have revealed modes of variability within the climate system, sources of predictability across a range of timescales, and indicator patterns of climate change 106, 61. There is also evidence that XAI methods can be used to evaluate climate models against observations, identifying the most important climate model biases for the specific prediction task 111. However, these methods are in their infancy and there is vast room for advancements in their application, making it explicitly appropriate to employ them within the physical sciences. Unsupervised ML can be intuitively IAI through the design of experiments. For example, applying clustering on closed model budgets of momentum ensures all relevant physics are represented, and can be interpreted in terms of the statistically dominant balances. In this manner, different regimes can be discovered 112, 92. Adversarial learning has been an effective tool for generating super-resolution fields of atmospheric variables in climate models 94. Furthermore, unsupervised ML approaches have been proposed for discovering and quantifying causal interdependencies and dynamical links inside a system, such as the Earth’s climate 89, 113. The development of ESMs is increasingly turning to process-oriented diagnostics (PODs)114, where a certain process is targeted and used as a benchmark for model improvement. A revolution of analysis tools has been called for, and ML is poised to be part of this change 115, 116, 66. For instance, the POD approach has been applied to evaluate the ability of ESM projections to simulate atmospheric interactions and to constrain climate projection uncertainties 117. Given the importance of both explainability and interpretability for improving ML generalization and scientific discovery, we need climate scientists working together with AI scientists to develop methods that are tailored to the field’s needs. This is not just an interesting exercise - it is essential for the proper use of AI for NESYM development and use (Fig. 3). Earth and climate scientists can aid the development of consistent benchmarks that allow evaluating both stand-alone ML and NESYM hybrids in terms of geophysical consistency 118. However, help of the AI community is needed to resolve other recently highlighted ML pitfalls, for instance, translating the concepts of adversarial examples and deep learning artifacts 119 into the ESM context or finding new measures to identify and avoid shortcut learning 45 in NESYM hybrids. In summary, only combined efforts and continuous development of both ESM and AI will evolve Neural Earth System Modelling. Our perspective should not only be seen as the outline of a promising scientific pathway to achieve a better understanding of the Earth’s present and future state, but also as an answer to the recent support call from the AI community 120. Based on the recent advances in applying AI to Earth system and climate sciences, it seems to be a logical succession that AI will take over more and more tasks of traditional statistical and numerical ESM methods. Yet, in its current stage, it also seems unthinkable that AI alone can solve the climate prediction problem. In the forthcoming years, AI will necessarily need to rely on the geophysical determinism of process-based modelling and on careful human evaluation. However, once we find solutions to the foreseeable limitations described above and can build interpretable and geophysically consistent AI tools, this next evolutionary step will seem much more likely. ## Acknowledgements This study was funded by the Helmholtz Association and by the Initiative and Networking Fund of the Helmholtz Association through the project Advanced Earth System Modelling Capacity (ESM). NB acknowledges funding by the Volskwagen foundation and the European Union’s Horizon 2020 research and innovation program under grant agreement No 820970 (TiPES). ## Authors’ contributions CI conceived the paper and organized the collaboration. All authors contributed to writing and revising all chapters of this manuscript. In particular, NB and CI drafted the ESM overview, JSW and JS drafted the ESO and DA overview, CI and CK drafted the chapter ’From Machine Learning-based Data Exploration Towards Learning Physics’, CI and JSW and NB drafted the chapter ’The Fusion of Process-based Models and Artificial Intelligence’, MS and EB and CI drafted the chapter ’Peering into the Black Box’. ## Competing Interest The authors declare no competing interest. ## References * 1 Prinn, R. G. Development and application of earth system models. _Proceedings of the National Academy of Sciences_ 110, 3673–3680 (2013). URL https://www.pnas.org/content/110/Supplement_1/3673. https://www.pnas.org/content/110/Supplement_1/3673.full.pdf. * 2 Lin, J. W.-B. & Neelin, J. D. Considerations for stochastic convective parameterization. _Journal of the Atmospheric Sciences_ 59, 959 – 975 (2002). URL https://journals.ametsoc.org/view/journals/atsc/59/5/1520-0469_2002_059_0959_cfscp_2.0.co_2.xml. * 3 Klein, R. Scale-dependent models for atmospheric flows. _Annual Review of Fluid Mechanics_ 42, 249–274 (2010). * 4 Berner, J. _et al._ Stochastic parameterization: Toward a new view of weather and climate models. _Bulletin of the American Meteorological Society_ 98, 565 – 588 (2017). URL https://journals.ametsoc.org/view/journals/bams/98/3/bams-d-15-00268.1.xml. * 5 Knutti, R. Should we believe model predictions of future climate change? _Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences_ 366, 4647–4664 (2008). URL https://royalsocietypublishing.org/doi/abs/10.1098/rsta.2008.0169. https://royalsocietypublishing.org/doi/pdf/10.1098/rsta.2008.0169. * 6 Taylor, K. E., Stouffer, R. J. & Meehl, G. A. An overview of cmip5 and the experiment design. _Bulletin of the American Meteorological Society_ 93, 485 – 498 (2012). URL https://journals.ametsoc.org/view/journals/bams/93/4/bams-d-11-00094.1.xml. * 7 Stocker, T. _et al._ (eds.) _Climate Change 2013: The Physical Science Basis. Contribution of Working Group I to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change_ (Cambridge University Press, Cambridge, United Kingdom and New York, NY, USA, 2013). URL www.climatechange2013.org. * 8 Eyring, V. _et al._ Overview of the coupled model intercomparison project phase 6 (cmip6) experimental design and organization. _Geoscientific Model Development_ 9, 1937–1958 (2016). URL https://gmd.copernicus.org/articles/9/1937/2016/. * 9 Knutti, R., Rugenstein, M. A. & Hegerl, G. C. Beyond equilibrium climate sensitivity. _Nature Geoscience_ 10, 727–736 (2017). * 10 Meehl, G. A. _et al._ Context for interpreting equilibrium climate sensitivity and transient climate response from the cmip6 earth system models. _Science Advances_ 6 (2020). URL https://advances.sciencemag.org/content/6/26/eaba1981. https://advances.sciencemag.org/content/6/26/eaba1981.full.pdf. * 11 Zelinka, M. D. _et al._ Causes of higher climate sensitivity in cmip6 models. _Geophysical Research Letters_ 47, e2019GL085782 (2020). URL https://agupubs.onlinelibrary.wiley.com/doi/abs/10.1029/2019GL085782. E2019GL085782 10.1029/2019GL085782, https://agupubs.onlinelibrary.wiley.com/doi/pdf/10.1029/2019GL085782. * 12 Cox, P. M., Huntingford, C. & Williamson, M. S. Emergent constraint on equilibrium climate sensitivity from global temperature variability. _Nature_ 553, 319–322 (2018). URL http://www.nature.com/doifinder/10.1038/nature25450. * 13 Hall, A., Cox, P., Huntingford, C. & Klein, S. Progressing emergent constraints on future climate change. _Nature Climate Change_ 9, 269–278 (2019). URL http://dx.doi.org/10.1038/s41558-019-0436-6. * 14 Lenton, T. M. _et al._ Tipping elements in the earth’s climate system. _Proceedings of the National Academy of Sciences_ 105, 1786–1793 (2008). URL https://www.pnas.org/content/105/6/1786. https://www.pnas.org/content/105/6/1786.full.pdf. * 15 Boers, N., Ghil, M. & Rousseau, D.-D. Ocean circulation, ice shelf, and sea ice interactions explain dansgaard–oeschger cycles. _Proceedings of the National Academy of Sciences_ 115, E11005–E11014 (2018). URL https://www.pnas.org/content/115/47/E11005. https://www.pnas.org/content/115/47/E11005.full.pdf. * 16 Valdes, P. Built for stability. _Nature Geoscience_ 4, 414–416 (2011). * 17 Drijfhout, S. _et al._ Catalogue of abrupt shifts in intergovernmental panel on climate change climate models. _Proceedings of the National Academy of Sciences_ 112, E5777–E5786 (2015). URL https://www.pnas.org/content/112/43/E5777. https://www.pnas.org/content/112/43/E5777.full.pdf. * 18 Masson-Delmotte, V. _et al._ (eds.) _IPCC: Global Warming of 1.5 Degree Celsius: IPCC Special Report on the Impacts of Global Warming of 1.5 Degree Celsius_ (Intergovernmental Panel on Climate Change (IPCC), 2018). URL https://www.ipcc.ch/sr15. * 19 Shukla, P. _et al._ (eds.) _Climate Change and Land: an IPCC special report on climate change, desertification, land degradation, sustainable landmanagement, food security, and greenhouse gas fluxes in terrestrial ecosystems_ (Intergovernmental Panel on Climate Change (IPCC), 2019). URL https://www.ipcc.ch/srccl/. * 20 Pörtner, H. _et al._ (eds.) _IPCC Special Report on the Ocean and Cryosphere in a Changing Climate_ (Intergovernmental Panel on Climate Change (IPCC), 2019). URL https://www.ipcc.ch/srocc/. * 21 Otto, F. E. _et al._ Attribution of extreme weather events in Africa: a preliminary exploration of the science and policy implications. _Climatic Change_ 132, 531–543 (2015). * 22 Balsamo, G. _et al._ Satellite and in situ observations for advancing global earth surface modelling: A review. _Remote Sensing_ 10 (2018). URL https://www.mdpi.com/2072-4292/10/12/2038. * 23 Hersbach, H. _et al._ The era5 global reanalysis. _Quarterly Journal of the Royal Meteorological Society_ 146, 1999–2049 (2020). URL https://rmets.onlinelibrary.wiley.com/doi/abs/10.1002/qj.3803. https://rmets.onlinelibrary.wiley.com/doi/pdf/10.1002/qj.3803. * 24 Evensen, G. _Data assimilation: the ensemble Kalman filter_ (Springer Science & Business Media, 2009). * 25 Blei, D. M., Kucukelbir, A. & McAuliffe, J. D. Variational inference: A review for statisticians. _Journal of the American Statistical Association_ 112, 859–877 (2017). URL https://doi.org/10.1080/01621459.2017.1285773. https://doi.org/10.1080/01621459.2017.1285773. * 26 Houtekamer, P. L. & Zhang, F. Review of the Ensemble Kalman Filter for Atmospheric Data Assimilation. _Monthly Weather Review_ 144, 4489–4532 (2016). URL http://journals.ametsoc.org/doi/10.1175/MWR-D-15-0440.1. * 27 van Leeuwen, P. J. Nonlinear data assimilation in geosciences: an extremely efficient particle filter. _Quarterly Journal of the Royal Meteorological Society_ 136, 1991–1999 (2010). URL https://rmets.onlinelibrary.wiley.com/doi/abs/10.1002/qj.699. https://rmets.onlinelibrary.wiley.com/doi/pdf/10.1002/qj.699. * 28 van Leeuwen, P. J., Künsch, H. R., Nerger, L., Potthast, R. & Reich, S. Particle filters for high-dimensional geoscience applications: A review. _Quarterly Journal of the Royal Meteorological Society_ 145, 2335–2365 (2019). URL https://rmets.onlinelibrary.wiley.com/doi/abs/10.1002/qj.3551. https://rmets.onlinelibrary.wiley.com/doi/pdf/10.1002/qj.3551. * 29 Vetra-Carvalho, S. _et al._ State-of-the-art stochastic data assimilation methods for high-dimensional non-gaussian problems. _Tellus A: Dynamic Meteorology and Oceanography_ 70, 1–43 (2018). URL https://doi.org/10.1080/16000870.2018.1445364. https://doi.org/10.1080/16000870.2018.1445364. * 30 Penny, S. G. _et al._ Strongly coupled data assimilation in multiscale media: Experiments using a quasi-geostrophic coupled model. _Journal of Advances in Modeling Earth Systems_ 11, 1803–1829 (2019). URL https://agupubs.onlinelibrary.wiley.com/doi/abs/10.1029/2019MS001652. https://agupubs.onlinelibrary.wiley.com/doi/pdf/10.1029/2019MS001652. * 31 Browne, P. A., de Rosnay, P., Zuo, H., Bennett, A. & Dawson, A. Weakly coupled ocean–atmosphere data assimilation in the ecmwf nwp system. _Remote Sensing_ 11 (2019). URL https://www.mdpi.com/2072-4292/11/3/234. * 32 Salcedo-Sanz, S. _et al._ Machine learning information fusion in Earth observation: A comprehensive review of methods, applications and data sources. _Information Fusion_ 63, 256–272 (2020). * 33 Irrgang, C., Saynisch, J. & Thomas, M. Estimating global ocean heat content from tidal magnetic satellite observations. _Scientific Reports_ 9, 7893 (2019). URL http://dx.doi.org/10.1038/s41598-019-44397-8. * 34 Irrgang, C., Saynisch‐Wagner, J., Dill, R., Boergens, E. & Thomas, M. Self‐Validating Deep Learning for Recovering Terrestrial Water Storage From Gravity and Altimetry Measurements. _Geophysical Research Letters_ 47 (2020). URL https://onlinelibrary.wiley.com/doi/abs/10.1029/2020GL089258https://onlinelibrary.wiley.com/doi/10.1029/2020GL089258. * 35 Voulodimos, A., Doulamis, N., Doulamis, A. & Protopapadakis, E. Deep Learning for Computer Vision: A Brief Review. _Computational Intelligence and Neuroscience_ 2018 (2018). * 36 Brown, T. B. _et al._ Language models are few-shot learners. _arXiv_ (2020). 2005.14165. * 37 Loh, E. Medicine and the rise of the robots: A qualitative review of recent advances of artificial intelligence in health. _BMJ Leader_ 2, 59–63 (2018). * 38 Topol, E. J. High-performance medicine: the convergence of human and artificial intelligence. _Nature Medicine_ 25, 44–56 (2019). URL http://dx.doi.org/10.1038/s41591-018-0300-7. * 39 Nosratabadi, S. _et al._ Data science in economics: Comprehensive review of advanced machine learning and deep learning methods. _Mathematics_ 8, 1–25 (2020). * 40 Hagerty, A. & Rubinov, I. Global AI Ethics: A Review of the Social Impacts and Ethical Implications of Artificial Intelligence. _arXiv_ (2019). 1907.07892. * 41 Perc, M., Ozer, M. & Hojnik, J. Social and juristic challenges of artificial intelligence. _Palgrave Communications_ 5, 1–7 (2019). * 42 Papernot, N. _et al._ The limitations of deep learning in adversarial settings. _Proceedings - 2016 IEEE European Symposium on Security and Privacy, EURO S and P 2016_ 372–387 (2016). 1511.07528. * 43 Adadi, A. & Berrada, M. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI). _IEEE Access_ 6, 52138–52160 (2018). * 44 Walton, P. Artificial intelligence and the limitations of information. _Information (Switzerland)_ 9 (2018). * 45 Geirhos, R. _et al._ Shortcut learning in deep neural networks. _Nature Machine Intelligence_ 2, 665–673 (2020). URL http://dx.doi.org/10.1038/s42256-020-00257-zhttp://www.nature.com/articles/s42256-020-00257-z. 2004.07780. * 46 Girasa, R. Ai as a disruptive technology. In _Artificial Intelligence as a Disruptive Technology_ , 3–21 (Springer, 2020). * 47 Dawson, M., Olvera, J., Fung, A. & Manry, M. Inversion of Surface Parameters Using Fast Learning Neural Networks. In _[Proceedings] IGARSS ’92 International Geoscience and Remote Sensing Symposium_ , vol. 2, 910–912 (IEEE, 1992). URL http://ieeexplore.ieee.org/document/578294/. arXiv:1011.1669v3. * 48 Miller, D. M., Kaminsky, E. J. & Rana, S. Neural network classification of remote-sensing data. _Computers and Geosciences_ 21, 377–386 (1995). * 49 Serpico, S. B., Bruzzone, L. & Roli, F. An experimental comparison of neural and statistical non-parametric algorithms for supervised classification of remote-sensing images. _Pattern Recognition Letters_ 17, 1331–1341 (1996). URL https://doi.org/10.1016/S0167-8655(96)00090-6. * 50 Hsieh, W. W. & Tang, B. Applying Neural Network Models to Prediction and Data Analysis in Meteorology and Oceanography. _Bulletin of the American Meteorological Society_ 79, 1855–1870 (1998). URL http://journals.ametsoc.org/doi/abs/10.1175/1520-0477{%}281998{%}29079{%}3C1855{%}3AANNMTP{%}3E2.0.CO{%}3B2. z0022. * 51 Knutti, R., Stocker, T. F., Joos, F. & Plattner, G. K. Probabilistic climate change projections using neural networks. _Climate Dynamics_ 21, 257–272 (2003). * 52 Lary, D. J., Alavi, A. H., Gandomi, A. H. & Walker, A. L. Machine learning in geosciences and remote sensing. _Geoscience Frontiers_ 7, 3–10 (2016). URL http://dx.doi.org/10.1016/j.gsf.2015.07.003http://linkinghub.elsevier.com/retrieve/pii/S1674987115000821. * 53 Arcomano, T. _et al._ A Machine Learning-Based Global Atmospheric Forecast Model. _Geophysical Research Letters_ 47, 1–9 (2020). * 54 Weyn, J. A., Durran, D. R. & Caruana, R. Can machines learn to predict weather? Using deep learning to predict gridded 500‐hPa geopotential height from historical weather data. _Journal of Advances in Modeling Earth Systems_ (2019). * 55 Weyn, J. A., Durran, D. R. & Caruana, R. Improving Data‐Driven Global Weather Prediction Using Deep Convolutional Neural Networks on a Cubed Sphere. _Journal of Advances in Modeling Earth Systems_ 12 (2020). * 56 Boers, N. _et al._ Complex networks reveal global pattern of extreme-rainfall teleconnections. _Nature_ 566, 373–377 (2019). URL http://dx.doi.org/10.1038/s41586-018-0872-x. * 57 Ham, Y.-g., Kim, J.-h. & Luo, J.-j. Deep learning for multi-year ENSO forecasts. _Nature_ 573, 568–572 (2019). URL http://dx.doi.org/10.1038/s41586-019-1559-7http://www.nature.com/articles/s41586-019-1559-7. * 58 Yan, J., Mu, L., Wang, L., Ranjan, R. & Zomaya, A. Y. Temporal Convolutional Networks for the Advance Prediction of ENSO. _Scientific Reports_ 10, 1–15 (2020). * 59 Kadow, C., Hall, D. M. & Ulbrich, U. Artificial intelligence reconstructs missing climate information. _Nature Geoscience_ 13, 408–413 (2020). URL http://dx.doi.org/10.1038/s41561-020-0582-5. * 60 Barnes, E. A., Hurrell, J. W., Ebert‐Uphoff, I., Anderson, C. & Anderson, D. Viewing Forced Climate Patterns Through an AI Lens. _Geophysical Research Letters_ 46, 13389–13398 (2019). URL https://doi.org/10.1029/2019GL084944https://onlinelibrary.wiley.com/doi/10.1029/2019GL084944. * 61 Barnes, E. A. _et al._ Indicator Patterns of Forced Change Learned by an Artificial Neural Network. _Journal of Advances in Modeling Earth Systems_ 12 (2020). 2005.12322. * 62 Chattopadhyay, A., Hassanzadeh, P. & Pasha, S. Predicting clustered weather patterns: A test case for applications of convolutional neural networks to spatio-temporal climate data. _Scientific Reports_ 10, 1317 (2020). URL http://dx.doi.org/10.1038/s41598-020-57897-9http://arxiv.org/abs/1811.04817http://www.nature.com/articles/s41598-020-57897-9. 1811.04817. * 63 Ramachandran, P., Zoph, B. & Le, Q. V. Searching for activation functions. _arXiv_ 1–13 (2017). 1710.05941. * 64 Lu, Z., Hunt, B. R. & Ott, E. Attractor reconstruction by machine learning. _Chaos_ 28 (2018). URL http://dx.doi.org/10.1063/1.5039508. 1805.03362. * 65 Rumelhart, D. E., Hinton, G. E. & Williams, R. J. Learning representations by back-propagating errors. _Nature_ 323, 533–536 (1986). arXiv:1011.1669v3. * 66 Reichstein, M. _et al._ Deep learning and process understanding for data-driven Earth system science. _Nature_ 566, 195–204 (2019). URL http://dx.doi.org/10.1038/s41586-019-0912-1. * 67 Huntingford, C. _et al._ Machine learning and artificial intelligence to aid climate change research and preparedness. _Environmental Research Letters_ 14 (2019). * 68 Jung, M. _et al._ Scaling carbon fluxes from eddy covariance sites to globe: synthesis and evaluation of the fluxcom approach. _Biogeosciences_ 17, 1343–1365 (2020). URL https://bg.copernicus.org/articles/17/1343/2020/. * 69 Tramontana, G. _et al._ Partitioning net carbon dioxide fluxes into photosynthesis and respiration using neural networks. _Global Change Biology_ 26, 5235–5253 (2020). URL https://onlinelibrary.wiley.com/doi/abs/10.1111/gcb.15203. https://onlinelibrary.wiley.com/doi/pdf/10.1111/gcb.15203. * 70 Bolton, T. & Zanna, L. Applications of Deep Learning to Ocean Data Inference and Subgrid Parameterization. _Journal of Advances in Modeling Earth Systems_ 11, 376–399 (2019). URL http://doi.wiley.com/10.1029/2018MS001472. * 71 Rasp, S., Pritchard, M. S. & Gentine, P. Deep learning to represent subgrid processes in climate models. _Proceedings of the National Academy of Sciences_ 115, 9684–9689 (2018). URL http://arxiv.org/abs/1806.04731http://www.pnas.org/lookup/doi/10.1073/pnas.1810286115. 1806.04731. * 72 O’Gorman, P. A. & Dwyer, J. G. Using Machine Learning to Parameterize Moist Convection: Potential for Modeling of Climate, Climate Change, and Extreme Events. _Journal of Advances in Modeling Earth Systems_ (2018). 1806.11037. * 73 Gagne, D. J., Christensen, H. M., Subramanian, A. C. & Monahan, A. H. Machine Learning for Stochastic Parameterization: Generative Adversarial Networks in the Lorenz ’96 Model. _Journal of Advances in Modeling Earth Systems_ 12 (2020). 1909.04711. * 74 Han, Y., Zhang, G. J., Huang, X. & Wang, Y. A Moist Physics Parameterization Based on Deep Learning. _Journal of Advances in Modeling Earth Systems_ 0–2 (2020). * 75 Beucler, T., Pritchard, M., Gentine, P. & Rasp, S. Towards Physically-consistent, Data-driven Models of Convection. _arXiv_ 2–6 (2020). URL http://arxiv.org/abs/2002.08525. 2002.08525. * 76 Yuval, J. & O’Gorman, P. A. Stable machine-learning parameterization of subgrid processes for climate modeling at a range of resolutions. _Nature Communications_ 11, 1–10 (2020). URL http://dx.doi.org/10.1038/s41467-020-17142-3. 2001.03151. * 77 Brenowitz, N. D. & Bretherton, C. S. Prognostic Validation of a Neural Network Unified Physics Parameterization. _Geophysical Research Letters_ 1–10 (2018). URL https://agupubs-onlinelibrary-wiley-com.emedien.ub.uni-muenchen.de/doi/pdf/10.1029/2018GL078510http://doi.wiley.com/10.1029/2018GL078510. * 78 Krasnopolsky, V. M. & Fox-Rabinovitz, M. S. Complex hybrid models combining deterministic and machine learning components for numerical climate modeling and weather prediction. _Neural Networks_ 19, 122–134 (2006). * 79 Brenowitz, N. D. _et al._ Machine Learning Climate Model Dynamics: Offline versus Online Performance. _ArXiv_ 1–6 (2020). URL http://arxiv.org/abs/2011.03081. 2011.03081. * 80 Brenowitz, N. D., Beucler, T., Pritchard, M. & Bretherton, C. S. Interpreting and Stabilizing Machine-Learning Parametrizations of Convection. _Journal of the Atmospheric Sciences_ 77, 4357–4375 (2020). URL https://journals.ametsoc.org/view/journals/atsc/77/12/jas-d-20-0082.1.xml. 2003.06549. * 81 Seifert, A. & Rasp, S. Potential and limitations of machine learning for modeling warm‐rain cloud microphysical processes. _Journal of Advances in Modeling Earth Systems_ (2020). URL https://onlinelibrary.wiley.com/doi/10.1029/2020MS002301. * 82 Beucler, T., Rasp, S., Pritchard, M. & Gentine, P. Achieving conservation of energy in neural network emulators for climate modeling. _arXiv_ 2–6 (2019). 1906.06622. * 83 Schneider, T., Lan, S., Stuart, A. & Teixeira, J. Earth System Modeling 2.0: A Blueprint for Models That Learn From Observations and Targeted High-Resolution Simulations. _Geophysical Research Letters_ 44, 12,396–12,417 (2017). 1709.00037. * 84 Cintra, R. S. & Velho, H. F. d. C. Data Assimilation by Artificial Neural Networks for an Atmospheric General Circulation Model: Conventional Observation. _Bulletin of the American meteorological Society_ 77, 437–471 (2014). URL http://arxiv.org/abs/1407.4360http://www.intechopen.com/books/advanced-applications-for-artificial-neural-networks/data-assimilation-by-artificial-neural-networks-for-an-atmospheric-general-circulation-model. 1407.4360. * 85 Wahle, K., Staneva, J. & Guenther, H. Data assimilation of ocean wind waves using Neural Networks. A case study for the German Bight. _Ocean Modelling_ 96, 117–125 (2015). URL http://linkinghub.elsevier.com/retrieve/pii/S146350031500116Xhttp://dx.doi.org/10.1016/j.ocemod.2015.07.007. * 86 Irrgang, C., Saynisch-Wagner, J. & Thomas, M. Machine Learning-Based Prediction of Spatiotemporal Uncertainties in Global Wind Velocity Reanalyses. _Journal of Advances in Modeling Earth Systems_ 12 (2020). * 87 Brajard, J., Carrassi, A., Bocquet, M. & Bertino, L. Combining data assimilation and machine learning to emulate a dynamical model from sparse and noisy observations: A case study with the lorenz 96 model. _Journal of Computational Science_ 44, 101171 (2020). URL http://www.sciencedirect.com/science/article/pii/S1877750320304725. * 88 Ruckstuhl, Y., Janjić, T. & Rasp, S. Training a convolutional neural network to conserve mass in data assimilation. _Nonlinear Processes in Geophysics Discussions_ 2020, 1–15 (2020). URL https://npg.copernicus.org/preprints/npg-2020-38/. * 89 Runge, J. _et al._ Inferring causation from time series in Earth system sciences. _Nature Communications_ 10, 1–13 (2019). URL http://dx.doi.org/10.1038/s41467-019-10105-3. * 90 Boers, N. _et al._ Prediction of extreme floods in the eastern Central Andes based on a complex networks approach. _Nature Communications_ 5, 1–7 (2014). * 91 Qi, D. & Majda, A. J. Using machine learning to predict extreme events in complex systems. _Proceedings of the National Academy of Sciences of the United States of America_ 117, 52–59 (2020). * 92 Sonnewald, M., Dutkiewicz, S., Hill, C. & Forget, G. Elucidating ecological complexity: Unsupervised learning determines global marine eco-provinces. _Science Advances_ 6, 1–12 (2020). * 93 Leinonen, J., Guillaume, A. & Yuan, T. Reconstruction of Cloud Vertical Structure With a Generative Adversarial Network. _Geophysical Research Letters_ 46, 7035–7044 (2019). URL https://onlinelibrary.wiley.com/doi/abs/10.1029/2019GL082532. * 94 Stengel, K., Glaws, A., Hettinger, D. & King, R. N. Adversarial super-resolution of climatological wind and solar data. _Proceedings of the National Academy of Sciences of the United States of America_ 117, 16805–16815 (2020). * 95 Huber, M. & Knutti, R. Anthropogenic and natural warming inferred from changes in Earth’s energy balance. _Nature Geoscience_ 5, 31–36 (2012). * 96 Zanna, L. & Bolton, T. Data‐driven Equation Discovery of Ocean Mesoscale Closures. _Geophysical Research Letters_ 1–13 (2020). * 97 Lagaris, I. E., Likas, A. & Fotiadis, D. I. Artificial neural networks for solving ordinary and partial differential equations. _IEEE Transactions on Neural Networks_ 9, 987–1000 (1998). 9705023. * 98 Raissi, M., Perdikaris, P. & Karniadakis, G. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. _Journal of Computational Physics_ 378, 686–707 (2019). URL https://linkinghub.elsevier.com/retrieve/pii/S0021999118307125. * 99 Ramadhan, A. _et al._ Capturing missing physics in climate model parameterizations using neural differential equations. _arXiv_ (2020). URL http://arxiv.org/abs/2010.12559. 2010.12559. * 100 Goodfellow, I. J. _et al._ Generative adversarial networks. _arXiv_ (2014). 1406.2661. * 101 Hurrell, J. W. _et al._ The community earth system model: A framework for collaborative research. _Bulletin of the American Meteorological Society_ 94, 1339 – 1360 (01 Sep. 2013). URL https://journals.ametsoc.org/view/journals/bams/94/9/bams-d-12-00121.1.xml. * 102 Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. _Nature Machine Intelligence_ 1, 206–215 (2019). URL http://dx.doi.org/10.1038/s42256-019-0048-x. 1811.10154. * 103 Balaji, V. Climbing down Charney’s ladder: Machine Learning and the post-Dennard era of computational climate science. _arXiv_ (2020). URL http://arxiv.org/abs/2005.11862. 2005.11862. * 104 Ethics guidelines for trustworthy ai. https://ec.europa.eu/digital-single-market/en/news/ethics-guidelines-trustworthy-ai (2019). Accessed: 2021-01-06. * 105 Executive order on promoting the use of trustworthy artificial intelligence in the federal government. https://www.whitehouse.gov/presidential-actions/executive-order-promoting-use-trustworthy-artificial-intelligence-federal-government/ (2020). Accessed: 2021-01-06. * 106 Toms, B. A., Barnes, E. A. & Ebert‐Uphoff, I. Physically Interpretable Neural Networks for the Geosciences: Applications to Earth System Variability. _Journal of Advances in Modeling Earth Systems_ 12, 1–20 (2020). URL http://arxiv.org/abs/1912.01752https://onlinelibrary.wiley.com/doi/abs/10.1029/2019MS002002. 1912.01752. * 107 McGovern, A. _et al._ Making the black box more transparent: Understanding the physical implications of machine learning. _Bulletin of the American Meteorological Society_ 100, 2175 – 2199 (01 Nov. 2019). URL https://journals.ametsoc.org/view/journals/bams/100/11/bams-d-18-0195.1.xml. * 108 Ebert-Uphoff, I. & Hilburn, K. Evaluation, tuning and interpretation of neural networks for working with images in meteorological applications. _Bulletin of the American Meteorological Society_ 1 – 47 (31 Aug. 2020). URL https://journals.ametsoc.org/view/journals/bams/aop/BAMS-D-20-0097.1/BAMS-D-20-0097.1.xml. * 109 Olden, J. D., Joy, M. K. & Death, R. G. An accurate comparison of methods for quantifying variable importance in artificial neural networks using simulated data. _Ecological Modelling_ 178, 389 – 397 (2004). URL http://www.sciencedirect.com/science/article/pii/S0304380004001565. * 110 Bach, S. _et al._ On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. _PLOS ONE_ 10, 1–46 (2015). URL https://doi.org/10.1371/journal.pone.0130140. * 111 Barnes, E. A., Mayer, K., Toms, B., Martin, Z. & Gordon, E. Identifying opportunities for skillful weather prediction with interpretable neural networks. _arXiv_ (2020). 2012.07830. * 112 Sonnewald, M., Wunsch, C. & Heimbach, P. Unsupervised Learning Reveals Geography of Global Ocean Dynamical Regions. _Earth and Space Science_ 6, 784–794 (2019). * 113 Runge, J., Nowack, P., Kretschmer, M., Flaxman, S. & Sejdinovic, D. Detecting and quantifying causal associations in large nonlinear time series datasets. _Science Advances_ 5 (2019). URL https://advances.sciencemag.org/content/5/11/eaau4996. https://advances.sciencemag.org/content/5/11/eaau4996.full.pdf. * 114 Maloney, E. D. _et al._ Process-oriented evaluation of climate and weather forecasting models. _Bulletin of the American Meteorological Society_ 100, 1665 – 1686 (01 Sep. 2019). URL https://journals.ametsoc.org/view/journals/bams/100/9/bams-d-18-0042.1.xml. * 115 Eyring, V. _et al._ Taking climate model evaluation to the next level. _Nature Climate Change_ 9, 102–110 (2019). URL http://dx.doi.org/10.1038/s41558-018-0355-y. * 116 Schlund, M. _et al._ Constraining Uncertainty in Projected Gross Primary Production With Machine Learning. _Journal of Geophysical Research: Biogeosciences_ 125 (2020). * 117 Nowack, P., Runge, J., Eyring, V. & Haigh, J. D. Causal networks for climate model evaluation and constrained projections. _Nature Communications_ 11, 1–11 (2020). URL http://dx.doi.org/10.1038/s41467-020-15195-y. * 118 Rasp, S. _et al._ WeatherBench: A benchmark dataset for data-driven weather forecasting. _Journal of Advances in Modeling Earth Systems_ (2020). URL http://arxiv.org/abs/2002.00469. 2002.00469. * 119 Buckner, C. Understanding adversarial examples requires a theory of artefacts for deep learning. _Nature Machine Intelligence_ (2020). URL http://dx.doi.org/10.1038/s42256-020-00266-y. * 120 Rolnick, D. _et al._ Tackling Climate Change with Machine Learning. _arXiv_ (2019). URL http://arxiv.org/abs/1906.05433. 1906.05433.
msam10 # Bridging the gap between collisional and collisionless shock waves Antoine Bret1,2 and Asaf Pe’er3 Email address for correspondence: <EMAIL_ADDRESS>1ETSI Industriales, Universidad de Castilla-La Mancha, 13071 Ciudad Real, Spain 2Instituto de Investigaciones Energéticas y Aplicaciones Industriales, Campus Universitario de Ciudad Real, 13071 Ciudad Real, Spain 3Department of Physics, Bar-Ilan University, Ramat-Gan, 52900, Israel (?; revised ?; accepted ?. - To be entered by editorial office) ###### Abstract While the front of a fluid shock is a few mean-free-paths thick, the front of a collisionless shock can be orders of magnitude thinner. By bridging between a collisional and a collisionless formalism, we assess the transition between these two regimes. We consider non-relativistic, un-magnetized, planar shocks in electron/ion plasmas. In addition, our treatment of the collisionless regime is restricted to high Mach number electrostatic shocks. We find that the transition can be parameterized by the upstream plasma parameter $\Lambda$ which measures the coupling of the upstream medium. For $\Lambda\lesssim 1.12$, the upstream is collisional, i.e. strongly coupled, and the strong shock front is about $\mathcal{M}_{1}\lambda_{\mathrm{mfp},1}$ thick, where $\lambda_{\mathrm{mfp},1}$ and $\mathcal{M}_{1}$ are the upstream mean-free- path and Mach number respectively. A transition occurs for $\Lambda\sim 1.12$ beyond which the front is $\sim\mathcal{M}_{1}\lambda_{\mathrm{mfp},1}\ln\Lambda/\Lambda$ thick for $\Lambda\gtrsim 1.12$. Considering $\Lambda$ can reach billions in astrophysical settings, this allows to understand how the front of a collisionless shock can be orders of magnitude smaller than the mean-free- path, and how physics transitions continuously between these 2 extremes. ## 1 Introduction Shock waves are very common in systems that involve fluid flows. Such systems occur on very different scales - from the microphysical scale to astronomical scales. As such, the properties of the shocks can vary considerably, depending on the environment. From the microphysical point of view, it is useful to discriminate between collisional and collisionless shocks. Collisional shock waves, first discovered in the $19^{th}$ century (Salas, 2007), can occur in any fluid as the result of a steepening of a large amplitude sound wave, or collision of two media (Zel’dovich & Raizer, 2002). The front of a collisional shock is necessarily at least a few mean-free-paths thick, as the dissipation from the downstream to the upstream occurs via binary collisions. Collisionless shock waves were discovered later and can only form in plasma (Petschek, 1958; Buneman, 1964; Sagdeev, 1966). The dissipation is provided by collective plasma phenomena instead of binary collisions. As a result, the front of such shocks can be orders of magnitude thinner than the mean-free- path. For example, the front of the bow shock of the earth magnetosphere in the solar wind is some 100 km thick (Bale et al., 2003; Schwartz et al., 2011). Yet, the proton mean-free-path at this location is about the Sun-Earth distance, nearly 7 orders of magnitude longer. Hence, if the earth bow shock were collisional, its front would be about 1 a.u. thick (see also Balogh & Treumann (2013) §2.1.3 and references therein). Is it possible to bridge between these two regimes? How does a shock switches from a regime where its front is a few mean-free-paths thick, to another regime where its front is million times smaller? Exploring the intermediate case, bridging between collisional and collisionless shocks, is the aim of this article. On the collisionless side, shock-accelerated particles which can enhance the density jump, or external magnetization which can reduce it, will be ignored (Berezhko & Ellison, 1999; Bret & Narayan, 2018, 2019; Bret, 2020). The method implemented is explained in Section 2. The big picture is as follows: we first present an evaluation of the shock front thickness in the collisional regime, then in the collisionless regime. The first task is achieved in Section 3 using the Mott-Smith _ansatz_ (Mott-Smith, 1951), which writes the distribution function at any place along the shock profile, as a linear combination of the upstream and downstream Maxwellians. We then follow Tidman (1967) in Section 4 for the collisionless case before we bridge between the 2 expressions of the front thickness in Section 5, to propose an expression of the front thickness valid from the collisional to the collisionless regime. Figure 1: Setup and notations. ## 2 Method As previously stated, a fluid shock is mediated by collisions while a collisionless shock is mediated by collective effects. For a plasma where only electrostatic fields are active (such is the case for an electrostatic shock, the kinetic equation accounting for both kinds of effects would formally read (Kulsrud (2005), p. 9), $\frac{\partial F}{\partial t}+\mathbf{v}\cdot\frac{\partial F}{\partial\mathbf{r}}+\frac{q\mathbf{E}}{m}\cdot\frac{\partial F}{\partial\mathbf{v}}=\left(\frac{\partial F}{\partial t}\right)_{c}+\left(\frac{\partial F}{\partial t}\right)_{w},$ (1) where $q$ and $m$ are the charge and the mass of the species considered. The first term of the right-hand-side, namely $(\partial F/\partial t)_{c}$, stands for the rate of change of the distribution $F$ due to collisions. It is typically given by the Fokker-Planck operator. The second term, $(\partial F/\partial t)_{w}$, accounts for the effects of the waves and is given, for example, by the quasi-linear operator. In principle, accounting at once for these two collision terms with appropriate collision operators, should allow to describe a shock wave from the collisional to the fully collisionless regimes. Resolving the shock front requires a formalism capable of resolving the entire shock profile. This is a notoriously difficult problem which has been greatly aided by the introduction of the so-called Mott-Smith _ansatz_ (Mott-Smith, 1951). Initially introduced for a neutral fluid, this _ansatz_ consisted in approximating the molecular distribution function $F$ along the shock profile by a linear combination of the upstream and downstream drifting Maxwellians, $\displaystyle F(\mathbf{v})=$ $\displaystyle n_{1}(x)\left(\frac{m}{2\pi k_{B}T_{1}}\right)^{3/2}\exp\left(-\frac{m}{2k_{B}T_{1}}(\mathbf{v}-\mathbf{U}_{1})^{2}\right)$ (2) $\displaystyle+~{}n_{2}(x)\left(\frac{m}{2\pi k_{B}T_{2}}\right)^{3/2}\exp\left(-\frac{m}{2k_{B}T_{2}}(\mathbf{v}-\mathbf{U}_{2})^{2}\right),$ where $T_{1,2}$ and $\mathbf{U}_{1,2}$ are the upstream (subscript 1) and downstream (subscript 2) temperatures and velocities respectively, determined by the Rankine-Hugoniot (RH) jump conditions (see figure 1)111All temperatures are not always considered constant in the main articles cited here (Mott- Smith, 1951; Tidman, 1958, 1967). Yet, they are considered so when it comes to computing the shock profile.. The boundary conditions for the functions $n_{1,2}(x)$ are, $\displaystyle n_{1}(+\infty)$ $\displaystyle=N_{1,0}~{}~{}~{}~{}$ $\displaystyle n_{1}(-\infty)=0,$ $\displaystyle n_{2}(+\infty)$ $\displaystyle=0~{}~{}~{}~{}$ $\displaystyle n_{2}(-\infty)=N_{2,0},$ (3) where again $N_{1,0}$ and $N_{2,0}$ fulfill the RH jump conditions. Taking then the appropriate moments of the dispersion equation gives a differential equation which allows to determine the respective weights of the 2 Maxwellians in terms of $x$, hence the shock profile together with its front thickness (Mott-Smith, 1951). The method implemented here consists in dealing with the collisional and the collisionless regimes separately. * • We study the _collisional regime_ in Section 3. There we apply the Mott-Smith _ansatz_ using the BGK collision term (Bhatnagar et al., 1954) as a collision operator for $(\partial F/\partial t)_{c}$ in Eq. (1), with $(\partial F/\partial t)_{w}=0$. Notably, Bhatnagar et al. (1954) presented 4 different collision operators through their Eqs. (3, 4, 5-6, 15-19). Those given by Eqs. (3, 4, 5-6), like $\nu(f-f_{0})$222Here $\nu$ is a collision frequency, $f$ the distribution function and $f_{0}$ the equilibrium distribution function., have been widely used although they do not conserve all 3 quantities: particle number and/or momentum and/or energy. In Bhatnagar et al. (1954), only the operator of Eqs. (15-19) does conserve all 3, hence this is the one used here. Tidman (1958) used the Fokker-Planck operator to deal with the problem, considering Eq. (1) with $(\partial F/\partial t)_{w}=0$ and, $\left(\frac{\partial F}{\partial t}\right)_{c}=\frac{4\pi e^{4}}{m_{i}^{2}}\ln\Lambda\left(\frac{\partial F}{\partial t}\right)_{c,FP},$ (4) where $(\partial F/\partial t)_{c,FP}$ is the Fokker-Planck collision operator, $m_{i}$ the ion mass, and $\Lambda$ the number of particles in the Debye sphere, that is, the co-called “plasma parameter” which measures the coupling of the plasma. As we shall see in Section 6, the present treatment provides a more adequate bridging to the collisionless regime than Tidman (1958)’s Fokker-Planck result. * • For the _collisionless regime_ we follow in Section 4 the collisionless result of Tidman (1967) who also used the Mott-Smith _ansatz_. In recent years, the correctness of this approximation, namely that the distribution function is well approximated by superimposed drifting Maxwellians, was validated numerically using Particle-In-Cell simulations (Spitkovsky, 2008). Tidman (1967) considered Eq. (1) with $(\partial F/\partial t)_{c}=0$, describing $(\partial F/\partial t)_{w}=0$ by the quasi-linear operator. Having assessed the width of the shock front in the collisional and the collisionless regimes, we then bridge between the 2 expressions of the shock thickness in Section 5. ## 3 Collisional regime: applying Bhatnagar et al. (1954) collision term to the Mott-Smith _ansatz_ We switch to the reference frame of the shock and assume steady state in this frame. As in Tidman (1958), we consider the distributions are functions of $(v_{x},v_{y},v_{z})$ and assume quantities only vary with the $x$ coordinate. We therefore set $\partial_{y,z,t}=0$ and $E_{y,z}=0$ so that equation (1), with $(\partial F/\partial t)_{w}=0$, reads for the ion distribution $F$, $v_{x}\frac{\partial F}{\partial x}+\frac{eE_{x}}{m_{i}}\frac{\partial F}{\partial v_{x}}=\left(\frac{\partial F}{\partial t}\right)_{c,BGK},$ (5) where $m_{i}$ is the ion mass. The BGK collision term now reads (Bhatnagar et al., 1954; Gross & Krook, 1956), $\left(\frac{\partial F}{\partial t}\right)_{c,BGK}=\frac{1}{\sigma}\left(N^{2}\Phi-NF\right),$ (6) which vanishes for a Maxwellian distribution. According to Bhatnagar et al. (1954), $``\sigma^{-1}$ $\times$ a density” is a collision frequency $\nu$. In the present setting we define, $\frac{N_{1,0}}{\sigma}=\nu\sim\frac{v_{\mathrm{thi},1}}{\lambda_{\mathrm{mfp},1}}~{}~{}\Rightarrow~{}~{}\sigma=\frac{N_{1,0}\lambda_{\mathrm{mfp},1}}{v_{\mathrm{thi},1}},$ (7) where $v_{\mathrm{thi},1}$ and $\lambda_{\mathrm{mfp},1}$ are the upstream thermal velocity and mean-free-path respectively. Then $N$ and $\Phi$ are given by Eqs. (15-19) of Bhatnagar et al. (1954), $\displaystyle N$ $\displaystyle=$ $\displaystyle\int Fd^{3}v=n_{1}(x)+n_{2}(x),$ (8) $\displaystyle\Phi$ $\displaystyle=$ $\displaystyle\left(\frac{m_{i}}{2\pi k_{B}T}\right)^{3/2}\exp\left(-\frac{m_{i}}{2k_{B}T}(\mathbf{v}-\mathbf{q})^{2}\right),$ (9) $\displaystyle\mathbf{q}$ $\displaystyle=$ $\displaystyle\frac{1}{N}\int\mathbf{v}Fd^{3}v,$ (10) $\displaystyle\frac{3k_{B}T}{m_{i}}$ $\displaystyle=$ $\displaystyle\frac{1}{N}\int(\mathbf{v}-\mathbf{q})^{2}Fd^{3}v,$ (11) where $F$ has been considered of the form (2). Multiplying equation (5) by $v_{y}^{2}$ and integrating over $d^{3}v$ (see detailed calculation reported in Appendix A) gives an exact, simple result, $U_{1}\frac{k_{B}T_{1}}{m_{i}}~{}\frac{\partial n_{1}}{\partial x}+U_{2}\frac{k_{B}T_{2}}{m_{i}}~{}\frac{\partial n_{2}}{\partial x}=\frac{1}{3\sigma}(U_{1}-U_{2})^{2}~{}n_{1}n_{2}.$ (12) This differential equation is structurally identical to the ones found in Mott-Smith (1951); Tidman (1958). We show in Appendix B how it yields density profiles like the ones pictured in Figure 1, of the form, $\displaystyle n_{1}(x)$ $\displaystyle=$ $\displaystyle N_{1,0}\frac{1}{1+e^{-x/\ell}},$ $\displaystyle n_{2}(x)$ $\displaystyle=$ $\displaystyle N_{2,0}\frac{e^{-x/\ell}}{1+e^{-x/\ell}},$ (13) implicitly defining the shock width $\ell$. From Eq. (39) we find the thickness of the shock according to the present formalism, $\ell=3\sigma\frac{k_{B}(T_{1}-T_{2})U_{2}}{N_{1,0}m_{i}(U_{1}-U_{2})^{2}}=3\frac{N_{1,0}\lambda_{\mathrm{mfp},1}}{v_{\mathrm{thi},1}}\frac{k_{B}(T_{1}-T_{2})U_{2}}{N_{1,0}m_{i}(U_{1}-U_{2})^{2}}.$ (14) It is now convenient to use the RH jump conditions to express $\ell$ in terms of the upstream quantities, like the upstream Mach number and mean-free-path. The calculations reported in Appendix C give, $\displaystyle\ell$ $\displaystyle=$ $\displaystyle\lambda_{\mathrm{mfp},1}\frac{U_{1}}{v_{\mathrm{thi},1}}\frac{(\mathcal{M}_{1}^{2}+3)(5\mathcal{M}_{1}^{2}(\mathcal{M}_{1}^{2}+2)-3)}{5\mathcal{M}_{1}^{2}(\mathcal{M}_{1}^{2}-1)^{2}},$ (15) $\displaystyle=$ $\displaystyle\lambda_{\mathrm{mfp},1}\frac{(\mathcal{M}_{1}^{2}+3)(5\mathcal{M}_{1}^{2}(\mathcal{M}_{1}^{2}+2)-3)}{5\mathcal{M}_{1}(\mathcal{M}_{1}^{2}-1)^{2}},$ where $\mathcal{M}_{1}$ is the upstream Mach number333Here we set $v_{\mathrm{thi},1}\sim c_{s1}$ in order to write $U_{1}/v_{\mathrm{thi},1}\sim\mathcal{M}_{1}$, where $c_{s1}$ is the upstream sound speed. An exact calculation only changes the end result by a factor of order unity. Moreover, the same factor also modifies the collisionless shock width (18). Therefore, the critical plasma parameter $\Lambda_{c}$ defined by Eq. (20) for the collisional/collisionless transition, remains unchanged when considering $v_{\mathrm{thi},1}\sim c_{s1}$.. We eventually obtain the following limits for the shock width $\ell$, $\ell=\lambda_{\mathrm{mfp},1}\times\left\\{\begin{array}[]{r}\frac{12}{5}(\mathcal{M}_{1}-1)^{-2}~{}~{}\mathrm{for}~{}~{}\mathcal{M}_{1}\sim 1,\\\ \mathcal{M}_{1}~{}~{}\mathrm{for}~{}~{}\mathcal{M}_{1}\rightarrow\infty.\end{array}\right.$ (16) Figure 2: Shock front thickness in units of the mean-free-path, $\ell/\lambda_{\mathrm{mfp},1}$, in terms of the upstream Mach number $\mathcal{M}_{1}$. The function $\ell/\lambda_{\mathrm{mfp},1}$ is plotted in Figure 2 in terms of the Mach number. It reaches a minimum for $\mathcal{M}_{1}=3.53$ with $\ell/\lambda_{\mathrm{mfp},1}=6$. Such a “U” shape has also been found in Tidman (1958). We shall further comment on Tidman (1958) in Section 6. ## 4 Collisionless regime Our expression (16) of the shock width cannot be used to bridge all the way to collisionless shocks since it has been derived from the kinetic equation (1) without the $(\partial F/\partial t)_{w}$ collision term. Yet, collisionless shocks are sustained by the mechanism described by this very term. Tidman (1967) treated the problem of a collisionless shock by setting $(\partial F/\partial t)_{c}=0$ in Eq. (1) and considering the quasi-linear operator for $(\partial F/\partial t)_{w}$. The Mott-Smith _ansatz_ was also implemented in this study. Tidman (1967) could not derive an equation of the form (12) allowing to extract an analytical shock profile. Further analysis in Biskamp & Pfirsch (1969) and Tidman & Krall (1969) concluded that the quasi- linear formalism is not non-linear enough to fully render a shock. Yet, Tidman (1967) could derive the following estimate of the width of the front, $\displaystyle\ell$ $\displaystyle=$ $\displaystyle A\frac{U_{1}}{\omega_{pi,1}}=A\frac{U_{1}}{v_{\mathrm{thi},1}}\frac{v_{\mathrm{thi},1}}{\omega_{pi,1}}$ (17) $\displaystyle=$ $\displaystyle A\mathcal{M}_{1}\lambda_{\mathrm{Di},1},$ where $\lambda_{\mathrm{Di},1}$ is the upstream ionic Debye length and $A$ is a parameter expected to be of order $\mathcal{O}(10)$. We can eventually cast this result under the form, $\ell=A\mathcal{M}_{1}\frac{\ln\Lambda}{\Lambda}\lambda_{\mathrm{mfp,1}},$ (18) where $\Lambda$ is the plasma parameter already introduced in Eq. (4) and we have used (Fitzpatrick (2014), p. 10), $\lambda_{\mathrm{mfp,1}}=\frac{\Lambda}{\ln\Lambda}\lambda_{\mathrm{Di},1}.$ (19) Notably, Tidman (1967) only addressed high Mach numbers turbulent shocks triggered by electrostatic instabilities. The forthcoming bridging between the 2 regimes is therefore only valid for such shocks. Weibel shocks sustained by electromagnetic instabilities are therefore excluded (Stockem et al., 2014; Ruyer et al., 2017). Figure 3: Plot of $\ell(\mathcal{M}_{1}\lambda_{\mathrm{mfp},1})^{-1}$ in terms of $\Lambda$. For a collisional plasma with small $\Lambda$, the front thickness $\ell$ is given by Eq. (16). The collisionless thickness is given by Eq. (18). The width of the front for any plasma parameter $\Lambda$ is given by the red curve. Only valid for strong shock (see end of Section 4). Figure 4: Value of $\Lambda_{c}$ for which Eqs. (16 & 18) intersect, in terms of $A$ defined through Eq. (17). ## 5 Bridging between the 2 regimes Figure 3 shows the collisional and collisionless expressions of $\ell(\mathcal{M}_{1}\lambda_{\mathrm{mfp},1})^{-1}$ from Eqs. (16, 18). For upstream Mach number $\mathcal{M}_{1}>$ a few (4-5), these 2 expressions intersect for a critical plasma parameter $\Lambda_{c}$ defined by, $\lambda_{\mathrm{mfp},1}\mathcal{M}_{1}=A\mathcal{M}_{1}\frac{\ln\Lambda_{c}}{\Lambda_{c}}\lambda_{\mathrm{mfp,1}}~{}~{}\Rightarrow~{}~{}1=A\frac{\ln\Lambda_{c}}{\Lambda_{c}},$ (20) fulfilled for $\Lambda_{c}\sim 1.12$ and then for $\Lambda_{c}\sim 35$ (for $A=10$). For $\Lambda<1.12$, the upstream is strongly coupled, that is, collisional, and the width of the front will be given by the collisional result (16). For $\Lambda>1.12$, the upstream is weakly coupled, that is, collisionless, and the relevant front width is therefore the collisionless result (18). Hence, the larger value of $\Lambda_{c}\sim 35$ where the 2 expressions intersect again is not physically meaningful. For such values of $\Lambda$, the upstream is collisionless so that the collisionless result applies. The transition between the 2 regimes occurs therefore for a critical plasma parameter $\Lambda_{c}=1.12$, coinciding with the transition of the upstream from the strongly coupled/collisional regime, to the weakly coupled/collisionless regime. Although this value of $\Lambda_{c}$ has been computed for $A=10$, Figure 4 shows it is poorly sensitive to $A$ as long as $A=\mathcal{O}(10)$. Note that this value of $\Lambda_{c}=1.12$ is only indicative. For example, Lee & More (1984) developed an electron conductivity model for _dense_ plasmas requiring $\ln\Lambda\geq 2$, i.e, $\Lambda\geq e^{2}=7.39$. Therefore, while $\Lambda=35$ probably pertains to weakly collisional plasmas, the value $\Lambda_{c}=1.12$ only gives a general idea of where the transition occurs. The width of the front for any plasma parameter $\Lambda$ is eventually given by the red curve in Figure 3. Simply put, the nature of the shock is the same as the nature of the upstream. Both are collisional or collisionless together. The non-monotonic behavior in the collisionless regime is just the consequence of the non-monotonic variation of the mean-free-path in terms of the plasma parameter. The function $g(x)=A\ln x/x$ reaches a max for $x=e$ with $g(e)=3.67$, still for $A=10$. ## 6 Comparison with Tidman (1958) A calculation parallel to the one performed in Section 3 for the collisional regime was achieved in Tidman (1958). However, as we show here, the bridging it provides to the collisionless regime is inadequate. For the ion distribution function $F$, Tidman (1958) used the Fokker-Planck operator for $(\partial F/\partial t)_{c}$ in Eq. (1), set $(\partial F/\partial t)_{w}=0$, and found for strong shocks444See Eq. (6.6) of Tidman (1958) where $V$ is the sound speed and $K$ is the Mach number., $\ell_{T}=\alpha\frac{c_{s}^{4}}{N_{1,0}\Gamma}\mathcal{M}_{1}^{4},$ (21) where $\alpha=29.1/512\pi$ and $\Gamma=\frac{4\pi e^{4}}{m_{i}^{2}}\ln\Lambda$. We can recast this result under the form, $\ell_{T}=4\pi\alpha~{}\lambda_{\mathrm{mfp,1}}\mathcal{M}_{1}^{4}\sim 0.23~{}\lambda_{\mathrm{mfp,1}}\mathcal{M}_{1}^{4},$ (22) where we have used Eq. (19). As a consequence, bridging the collisional result of Tidman (1958) with the collisionless result of Tidman (1967), that is, bridging Eq. (22) with Eq. (18), implicitly defines a critical plasma parameter $\Lambda_{c}$ through, $\frac{4\pi\alpha}{A}\mathcal{M}_{1}^{3}=\frac{\ln\Lambda_{c}}{\Lambda_{c}},$ (23) yielding a Mach number-dependent value of $\Lambda_{c}$ and having no solution if the left-hand-side is larger than the maximum of the right hand-side, that is, for $\mathcal{M}_{1}>2.51$ (considering $A=10$). As opposed to that, the $\propto\mathcal{M}_{1}$ scaling of the collisional $\ell$ given by BGK-derived Eq. (16) is essential to give a value of $\Lambda_{c}$ independent of the upstream Mach number $\mathcal{M}_{1}$, with a switch from the collisional to the collisionless regime when the upstream becomes collisionless. We therefore find that BGK provides a better bridging to the collisionless regime than Fokker-Planck. Hazeltine (1998) already noted the capacity of the BGK operator to behave adequately in the collisionless limit. Computing the moments of the kinetic equation with the BGK operator, he could derive a _non- local_ expression of the heat flux in the collisionless regime, as expected when the mean-free-path becomes large (Hammett & Perkins, 1990; Hazeltine, 1998). Indeed, the BGK operator was specifically designed to provide an operator capable of giving an adequate description of low-density plasmas (Bhatnagar et al., 1954). The physical reason for the better behavior of the BGK operator when the mean free path becomes large could be that regardless of the mean free path, BGK assumes the equilibrium distribution function is a Maxwellian, since the collision term (6) vanishes for $F=N\phi$, where $\phi$ is a Maxwellian (see Eq. 9). In contrast, the Fokker-Planck operator does not assume any a priori form of the equilibrium distribution function. It can even be used to prove that such a function is a Maxwellian. Yet, the collision rate is implicitly assumed large compared to the dynamic terms $v/L$ in the Fokker-Planck equation (Kulsrud (2005), p. 213) since the derivation of the Fokker-Planck operator involves a Taylor expansion in time, implicitly assuming collisions are frequent enough (Kulsrud (2005), Eq. 29-30, p. 204 or Chandrasekhar (1943), §II.4). Therefore, when collisions become scarce, the BGK formalism keeps forcing, by design, a Maxwellian equilibrium, while Fokker-Planck progressively loses validity. ## 7 Conclusion We propose a bridging between collisional and collisionless shocks. The collisional “leg” is worked out using the Moot-Smith _ansatz_ (Mott-Smith, 1951) with the “full” BGK collision term which behaves correctly in the large mean-free-path limit (Bhatnagar et al., 1954; Gross & Krook, 1956; Hazeltine, 1998). The collisionless part is from Tidman (1967), valid for strong turbulent electrostatic shocks. The result makes perfect physical sense. As long as the upstream is strongly coupled, that is, collisional with $\Lambda\lesssim 1.12$, the strong shock is collisional with a front thickness $\sim\mathcal{M}_{1}\lambda_{\mathrm{mfp},1}$ given by Eq. (16). From $\Lambda\gtrsim 1.12$, the shock switches to the collisionless regime, with a front thickness $\ell\sim\mathcal{M}_{1}\lambda_{\mathrm{mfp},1}\ln\Lambda/\Lambda$, given by Eq. (18). We show that the BGK treatment of the collisional regime provides a better bridge to the collisionless regime than the Fokker-Planck model. Nevertheless, a confusing feature remains: in the collisional limit, one would expect the BGK and the Fokker-Planck treatments to merge. Yet, they don’t, as evidenced by their different $\mathcal{M}_{1}$ scaling for the strong shock width ($\propto\mathcal{M}_{1}$ for BGK vs. $\propto\mathcal{M}_{1}^{4}$ for Fokker- Planck). The reason for this could be that the collision frequency used in BGK (Eq. 7) does not depend on the particle velocity. However, this is still unclear to us. A smoother transition between the 2 regimes could be assessed from Eq. (1) considering both $(\partial F/\partial t)_{c}$ and $(\partial F/\partial t)_{w}$ at once, whereas we here switched them on and off according to the regime considered. The Mott-Smith _ansatz_ could still be applied, while using BGK for $(\partial F/\partial t)_{c}$ and the operator proposed by Dupree (1966) (as suggested in Tidman (1967)) or Baalrud et al. (2008), for $(\partial F/\partial t)_{w}$. Although the present theory is formally restricted to high Mach number, un- magnetized, electrostatic shocks, it may help understand how the value of $\Lambda\sim 10^{10}$ observed in the solar wind (see for example Fitzpatrick (2014), p. 8) yields an earth bow shock thickness orders of magnitude shorter than the mean-free-path. ## 8 Acknowledgments A.B. acknowledges support by grants ENE2016-75703-R from the Spanish Ministerio de Economía y Competitividad and SBPLY/17/180501/000264 from the Junta de Comunidades de Castilla-La Mancha. A. P. acknowledges support from the European Research Council via ERC consolidating grant #773062 (acronym O.M.J.). Thanks are due to Anatoly Spitkovsky, Bill Dorland, Ian Hutchinson, Ellen Zweibel and Richard Halzeltine for valuable inputs. ## Appendix A Proof of Eq. (12) Equation (12) is the $v_{y}^{2}$ moment of Eq. (5). The left-hand-side is calculated in Tidman (1958). Note that the term proportional to $E_{x}$ vanishes in this moment. We only detail here the calculation proper to the present work, that is, that of the right-hand-side. For this we need $\Phi$, hence $\mathbf{q}$ and $T$ defined by Eqs. (8-11). According to Eq. (10), $\mathbf{q}$ is given by, $\mathbf{q}=\frac{1}{N}\int\mathbf{v}Fd^{3}v=\frac{1}{n_{1}+n_{2}}\int\mathbf{v}Fd^{3}v.$ (24) Since $F$ is the sum of 2 drifting Maxwellians given by Eq. (2), we find for $\mathbf{q}$, $\mathbf{q}=\frac{n_{1}\mathbf{U}_{1}+n_{2}\mathbf{U}_{2}}{n_{1}+n_{2}}\equiv q~{}\mathbf{e}_{x},$ (25) where $\mathbf{e}_{x}$ is the unit vector of the $x$ axis. For $T$ we then get from (11)555The factor 2 in the second term of Eq. (A) comes from $\int v_{z}^{2}F=\int v_{y}^{2}F$., $\displaystyle\frac{3k_{B}T}{m_{i}}$ $\displaystyle=$ $\displaystyle\frac{1}{n_{1}+n_{2}}\int(\mathbf{v}-\mathbf{q})^{2}Fd^{3}v$ $\displaystyle=$ $\displaystyle\frac{1}{n_{1}+n_{2}}\int[(v_{x}-q)^{2}+v_{y}^{2}+v_{z}^{2}]Fd^{3}v$ $\displaystyle=$ $\displaystyle\frac{1}{n_{1}+n_{2}}\int(v_{x}-q)^{2}Fd^{3}v+\frac{2}{n_{1}+n_{2}}\int v_{y}^{2}Fd^{3}v$ $\displaystyle=$ $\displaystyle\frac{1}{n_{1}+n_{2}}\int(v_{x}-q)^{2}Fd^{3}v+\frac{2(k_{B}T_{1}n_{1}+k_{B}T_{2}n_{2})}{m_{i}(n_{1}+n_{2})}$ $\displaystyle\Rightarrow k_{B}T$ $\displaystyle=$ $\displaystyle\frac{n_{1}k_{B}T_{1}+n_{2}k_{B}T_{2}}{n_{1}+n_{2}}+\frac{n_{1}n_{2}}{3(n_{1}+n_{2})^{2}}m_{i}(U_{1}-U_{2})^{2}.$ (27) Let us now write explicitly the $v_{y}^{2}$ moment of the right-hand-side ($rhs$) of Eq. (5), $rhs=\frac{1}{\sigma}\int v_{y}^{2}(-NF+N^{2}\Phi)d^{3}v=\underbrace{\frac{N^{2}}{\sigma}\int v_{y}^{2}\Phi d^{3}v}_{\mathbf{1}}-\underbrace{\frac{N}{\sigma}\int v_{y}^{2}Fd^{3}v}_{\mathbf{2}}.$ (28) From (8) we see $N$ does not depend on $\mathbf{v}$. It can therefore be taken out of the integrals. Computing 2 we find, $\displaystyle\frac{N}{\sigma}\int v_{y}^{2}Fd^{3}v$ $\displaystyle=$ $\displaystyle\frac{n_{1}+n_{2}}{\sigma}\left(n_{1}\frac{k_{B}T_{1}}{m_{i}}+n_{2}\frac{k_{B}T_{2}}{m_{i}}\right),$ (29) $\displaystyle=$ $\displaystyle\frac{k_{B}}{m_{i}\sigma}(n_{1}+n_{2})(n_{1}T_{1}+n_{2}T_{2}).$ Then we compute 1. $\displaystyle\frac{N^{2}}{\sigma}\int v_{y}^{2}\Phi d^{3}v$ $\displaystyle=$ $\displaystyle\frac{(n_{1}+n_{2})^{2}}{\sigma}\int v_{y}^{2}\left(\frac{m_{i}}{2\pi k_{B}T}\right)^{3/2}\exp\left(-\frac{m_{i}}{2k_{B}T}(\mathbf{v}-\mathbf{q})^{2}\right)d^{3}v,$ $\displaystyle=$ $\displaystyle\frac{(n_{1}+n_{2})^{2}}{\sigma}\left(\frac{m_{i}}{2\pi k_{B}T}\right)^{3/2}\underbrace{\int v_{y}^{2}\exp\left(-\frac{m_{i}}{2k_{B}T}(\mathbf{v}-\mathbf{q})^{2}\right)d^{3}v}_{\mathbf{3}}.$ For 3 we get, $\mathbf{3}=(2\pi)^{3/2}\left(\frac{k_{B}T}{m_{i}}\right)^{5/2},$ (30) so that 1 gives, $\displaystyle\mathbf{1}$ $\displaystyle=$ $\displaystyle\frac{(n_{1}+n_{2})^{2}}{\sigma}\left(\frac{m_{i}}{2\pi k_{B}T}\right)^{3/2}(2\pi)^{3/2}\left(\frac{k_{B}T}{m_{i}}\right)^{5/2}$ (31) $\displaystyle=$ $\displaystyle\frac{(n_{1}+n_{2})^{2}}{\sigma}\frac{k_{B}T}{m_{i}}.$ Finally, Eq. (28) simplifies nicely and reads, $\displaystyle rhs$ $\displaystyle=$ $\displaystyle\frac{(n_{1}+n_{2})^{2}}{\sigma}\frac{k_{B}T}{m_{i}}-\frac{k_{B}}{m_{i}\sigma}(n_{1}+n_{2})(n_{1}T_{1}+n_{2}T_{2}),$ (32) $\displaystyle=$ $\displaystyle\frac{1}{3\sigma}n_{1}n_{2}(U_{1}-U_{2})^{2},$ in agreement with the right-hand-side of Eq. (12). ## Appendix B Derivation of the density profiles (3) from Eq. (12) Let us define $\alpha,\beta,\gamma$ from Eq. (12) by, $\underbrace{U_{1}\frac{k_{B}T_{1}}{m_{i}}}_{\alpha}\frac{\partial n_{1}}{\partial x}+\underbrace{U_{2}\frac{k_{B}T_{2}}{m_{i}}}_{\beta}\frac{\partial n_{2}}{\partial x}=\underbrace{\frac{1}{3\sigma}(U_{1}-U_{2})^{2}}_{\gamma}n_{1}n_{2}.$ (33) Consider now the matter conservation equation obtained equating the $v_{x}$ moments of (2) between any $x$ and $x=+\infty$, $n_{1}(x)U_{1}+n_{2}(x)U_{2}=N_{1,0}U_{1}.$ (34) Differentiate with respect to $x$ gives, $\frac{\partial n_{1}(x)}{\partial x}U_{1}+\frac{\partial n_{2}(x)}{\partial x}U_{2}=0~{}~{}\Rightarrow~{}~{}\frac{\partial n_{2}(x)}{\partial x}=-\frac{\partial n_{1}(x)}{\partial x}\frac{U_{1}}{U_{2}},$ (35) and use the result to eliminate $\partial n_{2}/\partial x$ in (33), $\frac{\partial n_{1}}{\partial x}\left(\alpha-\beta\frac{U_{1}}{U_{2}}\right)=\gamma n_{1}n_{2}~{}~{}\Rightarrow~{}~{}\frac{\partial n_{1}}{\partial x}\frac{1}{n_{1}n_{2}}=\frac{\gamma}{\alpha-\beta\frac{U_{1}}{U_{2}}}.$ (36) Making now use again of the conservation equation (34) to write, $n_{2}=(N_{1,0}-n_{1})\frac{U_{1}}{U_{2}},$ (37) one gets, $\frac{U_{2}}{U_{1}}\frac{\partial n_{1}}{\partial x}\frac{1}{n_{1}(N_{1,0}-n_{1})}=\frac{U_{2}}{U_{1}}\frac{\partial n_{1}}{\partial x}\frac{1}{N_{1,0}}\left(\frac{1}{n_{1}}+\frac{1}{N_{1,0}-n_{1}}\right)=\frac{\gamma}{\alpha-\beta\frac{U_{1}}{U_{2}}}.$ (38) We eventually obtain, $\frac{\partial n_{1}}{\partial x}\left(\frac{1}{n_{1}}+\frac{1}{N_{1,0}-n_{1}}\right)=N_{1,0}\frac{\gamma}{\alpha-\beta\frac{U_{1}}{U_{2}}}\frac{U_{1}}{U_{2}}\equiv-\ell^{-1},$ (39) where $\ell$ is the shock thickness since the solution accounting for the boundary conditions (2) is, $n_{1}(x)=N_{1,0}\frac{1}{1+e^{-x/\ell}}.$ (40) From (34) one then obtains for $n_{2}(x)$, $n_{2}(x)=N_{2,0}\frac{e^{-x/\ell}}{1+e^{-x/\ell}}.$ (41) ## Appendix C Derivation of Eq. (15) from Eq. (14) We first cast Eq. (14) under the form, $\ell=3\sigma\frac{k_{B}T_{1}(1-T_{2}/T_{1})(U_{2}/U_{1})U_{1}}{N_{1,0}m_{i}U_{1}^{2}(1-U_{2}/U_{1})^{2}}.$ (42) We then use the RH jump conditions (see for example Fitzpatrick (2014) p. 216, or Thorne & Blandford (2017) p. 905), $\left(\frac{U_{2}}{U_{1}}\right)^{-1}=\frac{N_{2,0}}{N_{1,0}}=\frac{\gamma+1}{\gamma-1+2\mathcal{M}_{1}^{-2}},$ (43) and, $\frac{T_{2}}{T_{1}}=\frac{P_{2}}{P_{1}}\frac{N_{1,0}}{N_{2,0}}$ (44) with, $\frac{P_{2}}{P_{1}}=\frac{2\gamma\mathcal{M}_{1}^{2}-\gamma+1}{\gamma+1}.$ (45) Substituting these ratios and setting, $\mathcal{M}_{1}^{2}=\frac{U_{1}^{2}}{\gamma P_{1}/N_{1,0}}$ (46) we get to Eq. (15) with $\gamma=5/3$. ## References * Baalrud et al. (2008) Baalrud, S. D., Callen, J. D. & Hegna, C. C. 2008 A kinetic equation for unstable plasmas in a finite space-time domain. Physics of Plasmas 15 (9), 092111. * Bale et al. (2003) Bale, S. D., Mozer, F. S. & Horbury, T. S. 2003 Density-transition scale at quasiperpendicular collisionless shocks. Phys. Rev. Lett. 91, 265004. * Balogh & Treumann (2013) Balogh, A. & Treumann, R.A. 2013 Physics of Collisionless Shocks: Space Plasma Shock Waves. Springer New York. * Berezhko & Ellison (1999) Berezhko, E. G. & Ellison, Donald C. 1999 A Simple Model of Nonlinear Diffusive Shock Acceleration. ApJ 526 (1), 385–399. * Bhatnagar et al. (1954) Bhatnagar, P. L., Gross, E. P. & Krook, M. 1954 A model for collision processes in gases. i. small amplitude processes in charged and neutral one-component systems. Phys. Rev. 94, 511. * Biskamp & Pfirsch (1969) Biskamp, D. & Pfirsch, D. 1969 Comments on “Turbulent Shock Waves in Plasmas”. Physics of Fluids 12 (3), 732–733. * Bret (2020) Bret, Antoine 2020 Can We Trust MHD Jump Conditions for Collisionless Shocks? ApJ 900 (2), 111. * Bret & Narayan (2018) Bret, Antoine & Narayan, Ramesh 2018 Density jump as a function of magnetic field strength for parallel collisionless shocks in pair plasmas. Journal of Plasma Physics 84 (6), 905840604. * Bret & Narayan (2019) Bret, A. & Narayan, R. 2019 Density jump as a function of magnetic field for collisionless shocks in pair plasmas: The perpendicular case. Physics of Plasmas 26 (6), 062108. * Buneman (1964) Buneman, O. 1964 Models of Collisionless Shock Fronts. Physics of Fluids 7 (11), S3–S8. * Chandrasekhar (1943) Chandrasekhar, S. 1943 Stochastic problems in physics and astronomy. Rev. Mod. Phys. 15, 1–89. * Dupree (1966) Dupree, T. H. 1966 A Perturbation Theory for Strong Plasma Turbulence. Physics of Fluids 9 (9), 1773–1782. * Fitzpatrick (2014) Fitzpatrick, R. 2014 Plasma Physics: An Introduction. Taylor & Francis. * Gross & Krook (1956) Gross, E. P. & Krook, M. 1956 Model for collision processes in gases: Small-amplitude oscillations of charged two-component systems. Phys. Rev. 102, 511. * Hammett & Perkins (1990) Hammett, Gregory W. & Perkins, Francis W. 1990 Fluid moment models for landau damping with application to the ion-temperature-gradient instability. Phys. Rev. Lett. 64, 3019–3022. * Hazeltine (1998) Hazeltine, R. D. 1998 Transport theory in the collisionless limit. Physics of Plasmas 5 (9), 3282–3286. * Kulsrud (2005) Kulsrud, Russell M 2005 Plasma physics for astrophysics. Princeton, NJ: Princeton Univ. Press. * Lee & More (1984) Lee, Y. T. & More, R. M. 1984 An electron conductivity model for dense plasmas. Physics of Fluids 27 (5), 1273–1286. * Mott-Smith (1951) Mott-Smith, H. M. 1951 The Solution of the Boltzmann Equation for a Shock Wave. Physical Review 82 (6), 885–892. * Petschek (1958) Petschek, H. E. 1958 Aerodynamic Dissipation. Reviews of Modern Physics 30, 966–974. * Ruyer et al. (2017) Ruyer, C., Gremillet, L., Bonnaud, G. & Riconda, C. 2017 A self-consistent analytical model for the upstream magnetic-field and ion-beam properties in Weibel-mediated collisionless shocks. Physics of Plasmas 24 (4), 041409. * Sagdeev (1966) Sagdeev, R. Z. 1966 Cooperative Phenomena and Shock Waves in Collisionless Plasmas. Reviews of Plasma Physics 4, 23. * Salas (2007) Salas, Manuel D. 2007 The curious events leading to the theory of shock waves. Shock Waves 16 (6), 477–487. * Schwartz et al. (2011) Schwartz, Steven J., Henley, Edmund, Mitchell, Jeremy & Krasnoselskikh, Vladimir 2011 Electron temperature gradient scale at collisionless shocks. Phys. Rev. Lett. 107, 215002. * Spitkovsky (2008) Spitkovsky, Anatoly 2008 Particle acceleration in relativistic collisionless shocks: Fermi process at last? Astrophys. J. Lett. 682, L5–L8. * Stockem et al. (2014) Stockem, A., Fiuza, F., Bret, A., Fonseca, R. A. & Silva, L. O. 2014 Exploring the nature of collisionless shocks under laboratory conditions. Scientific Reports 4, 3934. * Thorne & Blandford (2017) Thorne, K.S. & Blandford, R.D. 2017 Modern Classical Physics: Optics, Fluids, Plasmas, Elasticity, Relativity, and Statistical Physics. Princeton University Press. * Tidman (1958) Tidman, D. A. 1958 Structure of a Shock Wave in Fully Ionized Hydrogen. Physical Review 111 (6), 1439–1446. * Tidman (1967) Tidman, D. A. 1967 Turbulent Shock Waves in Plasmas. Physics of Fluids 10 (3), 547–564. * Tidman & Krall (1969) Tidman, Derek A. & Krall, Nicholas A. 1969 Reply to Comments by D. Biskamp and D. Pfirsch. Physics of Fluids 12 (3), 733–735. * Zel’dovich & Raizer (2002) Zel’dovich, Ya B & Raizer, Yu P 2002 Physics of shock waves and high-temperature hydrodynamic phenomena. Dover Publications.
# Traffic Flow Estimation using LTE Radio Frequency Counters and Machine Learning Forough Yaghoubi Ericsson ABStockholmSweden<EMAIL_ADDRESS>, Armin Catovic Schibsted Media GroupStockholmSweden <EMAIL_ADDRESS>, Arthur Gusmao Ericsson ABStockholmSweden <EMAIL_ADDRESS>, Jan Pieczkowski Ericsson ABStockholmSweden <EMAIL_ADDRESS>and Peter Boros Ericsson ABStockholmSweden <EMAIL_ADDRESS> ###### Abstract. As the demand for vehicles continues to outpace construction of new roads, it becomes imperative we implement strategies that improve utilization of existing transport infrastructure. Traffic sensors form a crucial part of many such strategies, giving us valuable insights into road utilization. However, due to cost and lead time associated with installation and maintenance of traffic sensors, municipalities and traffic authorities look toward cheaper and more scalable alternatives. Due to their ubiquitous nature and wide global deployment, cellular networks offer one such alternative. In this paper we present a novel method for traffic flow estimation using standardized LTE/4G radio frequency performance measurement counters. The problem is cast as a supervised regression task using both classical and deep learning methods. We further apply transfer learning to compensate that many locations lack traffic sensor data that could be used for training. We show that our approach benefits from applying transfer learning to generalize the solution not only in time but also in space (i.e., various parts of the city). The results are very promising and, unlike competing solutions, our approach utilizes aggregate LTE radio frequency counter data that is inherently privacy- preserving, readily available, and scales globally without any additional network impact. Intelligent Transportation Systems, Traffic Flow, LTE, Radio Frequency, Machine Learning, Transfer Learning ††copyright: none††doi: ††isbn: ††conference: ; 2021††journalyear: ;††price: ## 1\. Introduction The increasing number of vehicles in the public roadway network, relative to the limited construction of new roads, has caused recurring congestion in the U.S. and throughout the industrialized world (Lawrence A. Klein, 2006). In the U.S. alone, the total cost of lost productivity caused by traffic congestion was estimated at $87 billion in 2018 (Fleming, [n.d.]). While one solution is to build new and expand existing roads, this is costly and takes time. A complementary approach is to implement strategies that improve the utilization of existing transport infrastructure. These strategies are found in Intelligent Transportation Systems (ITS) roadway and transit programs that have among their goals reducing travel time, easing delay and congestion, improving safety, and reducing pollutant emissions (Lawrence A. Klein, 2006). Traffic flow sensor technology forms a key component of ITS. Traffic sensors can be categorized as in-roadway (e.g. inductive loop sensors and magnetometers), or over-roadway (e.g. traffic cameras, radar, infrared and laser sensors). More recently there has been a surge of ad-hoc over-roadway sensor technology, including road-side cellular network masts, Bluetooth and Wi-Fi sensors, as well as telemetry collected from connected vehicles, smartphones and GPS devices. Cellular network masts are particularly appealing, combining ubiquity of cellular network technology (e.g. LTE or more specifically E-UTRA), with strict high availability requirements. While there have been previous approaches in utilizing cellular networks for understanding traffic flow, they’ve either been intrusive due to using user data, or not practical from a network operations perspective. In this paper we present a novel method for traffic flow estimation that leverages standard LTE/E-UTRA performance management (PM) counters, as defined by the 3rd Generation Partnership Project (3GPP) (3GPP, 2019). Namely, we utilize two radio frequency (RF) measurements - path loss distribution, and timing advance distribution counters, aggregated over 15 minute intervals. These counters are inherently privacy-preserving, and are continuously collected by nearly all LTE networks around the world, independent of network vendor. Thus our solution is non-invasive, and highly practical as it can be scaled across vast geographic regions with no live network impact. Our contributions in this paper are threefold: 1. (1) We present a novel method for estimating traffic flow using classical and deep learning regression models trained on E-UTRA RF counters (features) and vehicle counts from actual traffic sensors (targets). 2. (2) We evaluate the performance of our models by applying the learned model to different time samples, referred to as temporal generalization in this paper. 3. (3) We evaluate the performance of our models by applying the learned model to different road segments lacking ground truth data, referred to as spatial generalization in this paper; it is shown that due to difference in traffic distribution, the performance of our models is sub-optimal, hence we improve the accuracy using two transfer learning approaches. The rest of the paper is organized as follows: in section 2 we summarize the related works in the area of traffic flow estimation; our overall solution including feature selection/transformation and learning algorithms is explained in section 3; section 4 introduces two different transfer learning approaches; in section 5 we evaluate the performance of our models in terms of temporal and spatial generalization; ethical aspects are considered in section 6; finally the key takeaways are summarized in section 7. ## 2\. Related Works Traffic flow estimation has and continues to be a popular research topic. In (Zewei et al., 2015; George et al., 2013; Ma et al., 2013; Haferkamp et al., 2017; Nam et al., 2020) traffic flow estimation approaches are presented using data gathered from different sources such as cameras, acoustic sensors, magnetometers and spatially separated magnetic sensors. These solutions are not efficient due to coverage limitations and effort required in terms of installation and maintenance. To cope with these problems (Hansapalangkul et al., 2007; Pattara-Atikom and Peachavanish, 2007; Hongsakham et al., 2008; Caceres et al., 2012; Xing et al., 2019; Ji and Hong, 2019; Wang et al., 2020) propose the use of mobile subscriber data in traffic flow estimation. In (Hansapalangkul et al., 2007; Pattara-Atikom and Peachavanish, 2007; Hongsakham et al., 2008) the cell dwelling time and global positioning system (GPS) coordinates of a mobile subscriber are used to estimate the traffic congestion. These methods however have the disadvantage of high power consumption on a mobile device due to constant use of GPS, and are inherently intrusive. Another type of cellular data is considered in (Caceres et al., 2012) where authors propose a traffic flow estimation algorithm based on the number of subscribers in cars making a voice call. With today’s heavy usage of streaming and social media services however, voice calls are hardly representative of the traffic density, which limits the accuracy of such an approach. The authors in (Xing et al., 2019) use the travel trajectory of different mobile subscribers to detect in-vehicle users and henceforth compute the number of vehicles on a specific road. Tracking individual mobile users however is highly contentious and in most countries any user-identifiable or user-sensitive information limits the real-time usage of such data. In more recent work (Ji and Hong, 2019), the authors propose a method to predict the traffic speed and direction using wireless communication access logs including S1 application protocol (S1AP) collected from multiple radio base stations (RBS) located within a predetermined distance from the road. Due to high- intensity nature of S1AP signalling, tracing on the S1 interface in every single RBS leads to increase in network load, which is undesirable as it could lead to network overload with potentially catastrophic consequences. Furthermore, S1AP exposes potentially sensitive subscriber information allowing for the user to be fingerprinted or tracked. Another approach as presented in (Wang et al., 2020), describes a ”data fusion” approach, i.e. combining taxi GPS data with vehicle counts from license plate recognition (LPR) devices. The scalability of such approach is constrained due to limited availability of LPR and taxi GPS data. Therefore in this paper, we propose a new solution to traffic flow estimation problem based on aggregate LTE/E-UTRA radio frequency counter data that is inherently privacy preserving, readily available, and does not impose any extra load on the network. ## 3\. Method In this paper we describe two different approaches to using E-UTRA RF counters for traffic flow estimation. One approach involves using uplink path loss distribution as a feature vector in our model. Here we reason that different number of vehicles, i.e. obstacles on the road, can be represented by different path loss distributions. By optimizing for the number of vehicles the model should be able to discriminate between vehicles and all other users in the vicinity. In the second approach we use radio propagation delay, or more specifically timing advance (TA), as our feature vector. LTE radio base stations (eNBs) estimate the propagation delay on every random access (RA) initiated by a user. These propagation delays are aggregated from all RAs and represented as a distribution over discretized distances, where every bin represents a certain distance range from the eNB. By selecting only the bins corresponding to the known distances between the eNB and the relevant road segments, we can directly capture the road users, i.e. vehicles. Figure 1. Traffic flow estimation system presented in this paper. Path loss and timing advance features are described in detail in section 3.1. In both cases we use supervised regression techniques to train and evaluate our models. Fig. 1 shows the high level view of our system. In this paper we work in two domains: the source domain consists of training and validation data - the feature and target variables; the second domain, referred to as target domain, is where we perform inference using only the feature variables - however we use ground truth data for evaluation purposes. Our solution depends on the following assumptions: * • The eNB, or more specifically the sector antenna, is located within line-of- sight (LOS) of the relevant road segment, * • The distance between the road segment and the sector antenna is known, * • The relevant road segment consists of predominantly vehicular traffic, * • Traffic sensors used to supply ground truth data completely capture the traffic flow along the relevant road segment. In the following subsections we describe our feature and target variables, and learning algorithms. ### 3.1. Feature and Target Variables Figure 2. An example of a laser-based traffic sensor used in this paper; photograph by Holger Ellgaard, distributed under a CC BY-SA 3.0 license. #### 3.1.1. Traffic Sensor Data Target variables, i.e. ground truth data, consist of total vehicle counts aggregated over 15-minute intervals. The data is collected from a number of laser-based traffic sensors around inner Stockholm. Fig. 2 shows an example of such a sensor. Each sensor scans one lane of the road. For each road segment we sum the values from each lane to obtain total vehicle counts per 15-minute interval. We remove any samples where one of the sensor’s values are missing (e.g. due to a malfunction). #### 3.1.2. Path Loss Features Path loss (PL) is the attenuation of electromagnetic wave caused by free-space losses, absorption (e.g. by atmospheric particles), and scattering off various obstacles and surfaces. Radio propagation models attempt to account for this attenuation, and are a pivotal component in cellular network planning. Hata- Okamura (Hata, 1980) models are one such family of radio propagation models used to approximate cellular network coverage in different environments. In LTE, eNBs estimate PL values for all the scheduled users on every transmit time interval (TTI), which is typically 1ms. These estimates, represented as decibel (dB) values are then placed into discretized PL bins; in our case we have 21 bins, where each bin covers a range of 5dB, starting from $<$50dB and going up to $>$140dB. These estimates are done per-frequency band; in our case we have three bands, 800MHz, 1800MHz (two separate antennas working in this band) and 2600MHz, so we concatenate PL bins for all bands, resulting in total of 4 x 21 = 48 PL features. We don’t apply any filtering or transformation to PL features, and we treat all PL bins equally. Even though PL estimates are done every 1ms, the actual data available to us is aggregated in 15-minute intervals. The motivation behind our use of PL features is that different traffic conditions will result in different radio wave scattering characteristics, leading to different path loss distributions. A condition where there are no vehicles on the road will be represented by path loss distribution $PL_{a}$, which would be representative of radio wave losses due to predominantly indoor users and pedestrians. On the other hand a condition where the traffic flow is greater than zero would result in path loss distribution $PL_{b}$ where $PL_{b}\neq PL_{a}$, since radio wave scattering off vehicle surfaces would yield a different path loss ”signature”. Our trained algorithms should be able to discriminate between such conditions. Figure 3. Spatial granularity of an eNB. #### 3.1.3. Timing Advance Features Timing advance (TA) is estimated for every user connection request, or more specifically on every random access. TA estimation is dependent on successful completion of an RRC Connection Request procedure, and the 11-bit TA command. The name is a slight misnomer, since TA features are actually represented as discretized distance bins/ranges, representing the distance between the user and the sector antenna. In our case we have 35 bins, starting from $<$80m up to 100km; typically only the first few bins are incremented, as users are normally within 500m of the antenna (otherwise they will be handed over to another sector, or another eNB). Fig. 3 shows the spatial granularity of an eNB and how the TA features may be represented. Unlike PL features, we do actually apply a distance selection filter to TA features. Our aim is to consider only road users (vehicles), which means selecting TA bins/features that represent the known distances between the relevant road segment and the sector antenna. Lets assume that TA value ranges are indicated by $M$ bins where bin $ta_{i}$ corresponds to a distance interval shown by $[d_{i};d_{i+1})$ given a known distance $d$ between the road segment and the sector antenna, we choose the TA bin index $ta_{i}$ where $d_{i}\leq d\leq d_{i+1}$. Just like traffic sensor data and PL features, TA features are also aggregated in 15-minute intervals. #### 3.1.4. Cyclic Time Features As traffic exhibits strong seasonality, it is beneficial to give our models temporal information. To encode this information, a common method is to transform the date-time representation into cyclic time features using a $\sin$ and $\cos$ transformation as follows: $\displaystyle x_{\sin}=$ $\displaystyle\sin(\frac{2\pi x}{\max(x)})$ $\displaystyle x_{\cos}=$ $\displaystyle\cos(\frac{2\pi x}{\max(x)})$ where $x$ can be hour, day and month. By using the above equation, we convert time-of-day, day-of-week and week-of-month to the corresponding cyclic time features. As the time granularity for our data is in minutes, we set the $\max(x)$ to $24*60$, $7*24*60$, and $4*7*24*60$ respectively. #### 3.1.5. Road-dependent Features As traffic flow depends on the road characteristics, we also apply different road-dependent features. These features are easily extracted from e.g. OpenStreetMap services. In this paper we use the following road-dependent features: number of lanes, maximum speed limit, and road category, i.e. highway, large city road and small city road. ### 3.2. Learning Algorithms In this paper we compare two supervised learning approaches for traffic flow estimation. In the first approach we evaluate a number of different classical regression algorithms. In the second approach we take into account the history of time samples using gradient based Long Short-term Memory (LSTM). #### 3.2.1. Classical Regression Models Classical regression assumes independence between time samples. Since we’re working with fairly coarse 15-min aggregate intervals, it is reasonable to assume this independence. Regression then amounts to estimating a function $f(\bm{x};\bm{\theta})$, which transforms a feature vector $\bm{x}$ to a target variable $y$. Function parameters $\bm{\theta}=(\theta_{0},\theta_{1},...,\theta_{k})$ are found by minimizing expected loss, typically a mean squared error (MSE) of the form $MSE(\bm{\theta})=\frac{1}{N}\sum_{n=1}^{N}(y_{n}-f(\bm{x_{n}};\bm{\theta}))^{2}$, where $N$ corresponds to total number of 15-min aggregate samples. We evaluate a number of different regression algorithms including Support Vector Regressor (SVR), Kernel Ridge (KR), Decision Tree (DT), and Random Forest (RF). Each algorithm also requires setting its internal parameters, or hyperparameters. Since the total number of hyperparameters is small, we use grid search method to exhaustively search through the hyperparameter space and pick the combination of parameters that yield the best performance. We apply a time- dependent train/test split, e.g. by selecting the first 6 weeks for training, and the following 2 weeks for testing; compared to a random assignment of train/test data, our approach is more in line with how the algorithm would be used in practice, and is more representative of the generalization capability in the real-world setting. #### 3.2.2. LSTM LSTM is a specific kind of recurrent neural network (RNN) that has the ability to capture long-term time dependencies and bridge time intervals in excess of 1000 steps even in case of noisy, in-compressible input sequences (Du et al., 2017). Similar to other types of RNNs, LSTM has a chain structure with modified repeating modules. In each module, instead of having a single neural network layer, there are four layers that interact with each other. More detailed information about LSTM architecture can be found in (Smagulova and James, 2019). The architecture of our LSTM based traffic flow estimator consists of two LSTM layers, followed by a dropout regularization layer, and then finally the two fully-connected (FC) layers. The two LSTM, as well as the two FC layers, use the rectified linear unit (ReLU) activation function, while the output layer activates with the linear function. ## 4\. Transfer Learning Approaches The learning approaches mentioned above optimize the model for temporal generalization where we use all available locations in our training set, but withhold a contiguous period of time (e.g. two weeks) for test purposes. However, we would like our models to generalize well across all possible locations, even never-before-seen locations, which may potentially have completely different traffic patterns/distributions. We refer to this problem as spatial generalization. To cope with this problem we use transfer learning (TL) approaches. TL focuses on transferring the knowledge between different domains and can be a promising solution to overcome the spatial generalization problem. Recently, there has been lot of work focusing on transfer learning and proposing efficient solutions (Zhuang et al., 2019; Pan and Yang, 2009; Tan et al., 2018). These studies categorize TL into three subcategories based on different situations involving source and target domain data and the tasks, including inductive, transductive, and unsupervised transfer learning. Our work can be fitted into transductive transfer learning where the source label data are available while no label data for target domain is provided. Here the assumption is that the task between target and source domain is the same, but the domain marginal or conditional distributions are different. Among the proposed transductive TL algorithms, we evaluate two approaches - one based on instant weighting and the second one based on deep domain adaptation. We explain each of the algorithms in detail in the following sections. ### 4.1. Instant Weighting The data-based TL approaches, such as instant weighting, focus on transferring the knowledge by adjustment of the source data. Assuming that the source and target domain only differ in marginal distribution, a simple idea for transformation is to assign weights to source domain data equal to the ratio of source and target domain marginal distribution. Therefore the general loss function of the learning algorithm is given by: (1) $\min_{\theta}\frac{1}{N_{s}}\sum_{1}^{N_{s}}\alpha_{i}J(\theta(x_{i}^{s}),y_{i}^{s})+\lambda\gamma(\theta)$ where $J$ represents the loss of source data and $\alpha_{i}$ is the weighting parameter and is equal to: (2) $\alpha_{i}=\frac{P^{T}(x)}{P^{S}(x)}.$ In the literature, there exist many ways to compute $\alpha_{i}$; in (Huang et al., 2006) the authors used Kernel Mean Matching (KMM) to estimate the ratio by matching the means of target and source domain data in the reproducing- kernel Hilbert space where the problem of finding weights can be written as follows: (3) $\displaystyle\min_{\alpha}\frac{1}{2}\alpha^{T}K\alpha-\kappa\alpha$ $\displaystyle s.t\sum_{i}^{N_{s}}\alpha_{i}-N_{s}\leq N_{s}\epsilon$ $\displaystyle\alpha\in[0,B]$ where $N_{s}$ shows the number of sample in source domain data and $K$ is kernel matrix and is defined as: (4) $K=\begin{bmatrix}k_{ss}&k_{st}\\\ k_{ts}&k_{tt}\end{bmatrix},$ while $k_{ss}=k(x_{s},x_{s})$ and $\kappa_{i}=\frac{N_{s}}{N_{T}}\sum_{i}^{N_{T}}k(x_{i},x_{Tj})$. ### 4.2. Domain Adaptation Deep learning algorithms have received lot of attention from researchers having successfully outperformed many traditional machine learning methods in tasks such as computer vision and natural language processing (NLP). Therefore in the TL area many researchers also utilize deep learning techniques. In this paper, we use discrepancy-based domain adaptation, where a deep neural network is used to learn the domain-independent feature representations. In deep neural networks, the early layers tends to learn more generic transferable features, while domain-dependent features are extracted in the terminal layers. Therefore, to decrease the gap between the distribution in the last layers, we add multiple adaptation layers with discrepancy loss as regularizer. The deep learning model used for feature extraction is the LSTM model explained in previous section. The pretrained LSTM model will be used to extract the features for both source and target domains. After that the primary goal is to reduce the difference between target and source domain distribution. The term maximum mean discrepancy (MMD) is widely used in TL literature as a metric to compute the distance between two distribution (Wang et al., 2017; Gretton et al., 2012). Fig. 4 shows the architecture of our domain adaptation network based on LSTM. Figure 4. The domain adaptation network based on LSTM. Let $f$ denote the function for feature representation of our pretrained model, then the distance between the feature distribution of source and target domain is given by: (5) $d(p,q)=\sup_{f\in F}{E_{p}\\{f(x)\\}-E_{q}\\{f(y)\\}}$ where $\sup$ defines the supremum, $E$ denotes the expectation and $x$ and $y$ are independently and identically distributed (i.i.d) samples from $p$ and $q$, respectively. The above equation can be easily computed using the kernel trick where it can be expressed by expectation of kernel functions. Therefore, the square of equation (5) can be reformulated as follows: (6) $d^{2}_{k}(p,q)=E_{x_{p}^{s}x_{p}^{s}k(x_{p}^{s},x_{p}^{s})}+E_{x_{q}^{t}x_{q}^{t}k(x_{q}^{t},x_{q}^{t})}-2E_{x_{p}^{s}x_{q}^{t}k(x_{p}^{s},x_{q}^{t})},$ where $x_{p}^{s}$ and $x_{q}^{t}$ are the samples from source and target domain respectively, and $k$ is the kernel defined as $\exp(\frac{-\left\lVert x_{i}-x_{j}\right\rVert^{2}}{\gamma})$. To adapt the pretrained model for the target data samples, the objective function of our TL algorithm is given by (Long et al., 2015): (7) $\min_{\theta}\frac{1}{N_{s}}\sum_{1}^{N_{s}}J(\theta(x_{i}^{s}),y_{i}^{s})+\lambda\sum_{l=l_{1}}^{l_{2}}d_{k}^{2}(D^{s}_{l},D^{t}_{l}),$ where $J$ is the loss for source domain in LSTM network, $l_{1}$ and $l_{2}$ indicate the layer indices between which the regularization is effective, and $D^{s}_{l}$ and $D^{t}_{l}$ are $l$ layer representation of the source and target samples, respectively. The parameter $\lambda$ is a trade off term so that the objective function can benefit both from TL and deep learning. ## 5\. Results We use approximately 8 weeks worth of data, where every data sample corresponds to a 15-min interval, so we have $\scriptstyle\sim$ 96 * 7 * 8 = 5376 data samples. The data are collected from six different locations around inner Stockholm; each location corresponds to a road segment with a traffic sensor and a nearby LTE eNB. We evaluate models using PL and TA features independently and across a range of regression algorithms. When evaluating temporal generalization we use all locations during training and split the data into 80/20 train/test sets, which corresponds to approximately 6 weeks of contiguous training data, and 2 weeks of test data. When evaluating spatial generalization we use all time samples for training but we randomly assign road segments into source and target domains. For evaluation purposes we use coefficient of determination $R^{2}$ defined as follows: (8) $R^{2}=1-\frac{SS_{res}}{SS_{tot}}$ where $\displaystyle SS_{tot}=\sum_{i}\left(y_{i}-\bar{y}\right)^{2}$ (9) $\displaystyle SS_{res}=\sum_{i}\left(y_{i}-\hat{y}_{i}\right)^{2}$ $SS_{tot}$ represents total sum of squares, and $SS_{res}$ represents residual sum of squares, while $\bar{y}$ and $\hat{y_{i}}$ are the observed data mean and the predicted traffic flow respectively. A model that always predicts observed data mean will have $R^{2}=0$; models with observations worse than the observed data mean will have negative values; the most optimal value is $R^{2}=1$, so we want our models to be as close to 1 as possible. The set of classical regression algorithms used for training are Support Vector Regression (SVR), Kernel Ridge (KR), Decision Trees (DT) and Random Forest (RF). We also train a deep learning model with two LSTM layers followed by a dropout layer and two fully-connected layers activated with the ReLU function. The hyperparameters providing the best $R^{2}$ score on the test set for our models are found using grid search and presented in Table 1. The corresponding results for both temporal and spatial generalization performance are shown in Table 2. Table 1. Hyperparameters used in this paper. Models | Parameters ---|--- | Kernel = rbf SVR | $C$ = 10 | $\gamma$ = 0.001 | Kernel = rbf KR | $\alpha$ = 1 | $\gamma$ = 0.01 DT | Maximum depth = 10 RF | Maximum depth = 30 | Learning rate = 0.0009 LSTM | Hidden size = 100 | Epochs = 300 | Dropout rate = 0.2 | Window = 5 Table 2. $R^{2}$ score for temporal and spatial generalization performance using TA and PL features. Higher scores are better. Models | Temporal Generalization | Spatial Generalization ---|---|--- TA | PL | TA | PL SVR | 0.754 | 0.786 | 0.12 | -0.62 KR | 0.862 | 0.888 | -0.79 | -0.63 DT | 0.938 | 0.946 | -3.22 | -0.37 RF | 0.946 | 0.959 | -0.96 | 0.017 LSTM | 0.845 | 0.901 | 0.087 | -1.67 The results in Table 2 indicate that all regression algorithms perform reasonably well in terms of temporal generalization, using either TA or PL features. The Random Forest (RF) model outperforms all the others, including the LSTM model, with an average $R^{2}$ score of 0.95. These results validate our initial assumption that due to a fairly coarse 15-min aggregate interval, it is safe to assume independence between time steps, hence deep learning based LSTM does not add any additional value. A more visual representation of the RF algorithm performance is shown in Fig. 5 where we compare traffic flow estimates from our model against the actual values across three different locations. The algorithm does not always capture the peaks - our hypothesis is that more training samples with varied traffic flow distributions are needed for the model to generalize even better. (a) (b) (c) Figure 5. Traffic flow estimates in terms of number of vehicles per 15-min interval using our Random Forest (RF) model compared to the ground truth, for 5(a) Road 1, 5(b) Road 2 and 5(c) Road 3. Despite good temporal generalization performance, the average $R^{2}$ score for spatial generalization is very low for all regression models. This poor performance is due to inherent difference between the source and target domain distributions. In order to improve spatial generalization we use two types of transfer learning (TL) algorithms, namely instant weighting and deep domain adaptation. In the first approach we implement the instant weighting for classical regression. For each test location, we compute the weights solving the quadratic optimization problem, and then retrain the model using these weights. Since the RF model yields the highest $R^{2}$ score on temporal generalization we apply instant weighting to RF only. Table 3 presents the $R^{2}$ scores of RF model for both TA and PL features with and without applying the instant weighting. The results indicate that instant weighting can only improve the performance when TA features are used. Since the TA features represent the road users more explicitly we expect there to be some minimum similarity between all domain distributions. On the other hand PL represents all users, including indoor users, and therefore PL features are highly sensitive to physical layout of the environment, i.e. number of buildings, thickness of walls, heights of buildings etc. Table 3. $R^{2}$ score for spatial generalization using the Random Forest (RF) model with and without transfer learning (TL). Higher scores are better. Test Road | PL Features | TA Features ---|---|--- No TL | TL | No TL | TL 1 | 0.02 | -0.47 | -0.96 | 0.71 2 | 0.02 | -0.75 | -0.96 | 0.72 3 | 0.02 | -0.86 | -0.96 | 0.42 Mean | 0.02 | -0.69 | -0.96 | 0.62 Table 4. $R^{2}$ score for spatial generalization using the LSTM model with and without transfer learning (TL). Higher scores are better. Test Road | PL Features | TA Features ---|---|--- No TL | TL | No TL | TL 1 | -0.20 | 0.24 | 0.24 | 0.66 2 | -1.73 | -1.02 | -0.62 | 0.61 3 | -3.09 | -0.52 | -0.13 | 0.61 Mean | -1.67 | -0.43 | -0.17 | 0.63 In the second approach, we implement the deep domain adaptation algorithm as shown in Fig. 4. We freeze the two LSTM layers and the two fully-connected layers using the pre-trained weights, while we train the final two fully- connected layers using the MMD regularizer. As there is no target domain label data available only the source output is considered in the loss function. Table 4 shows the performance of spatial generalization using the LSTM and deep domain adaptation. The LSTM model performs reasonably well using TA features, with average $R^{2}$ score very similar to what we saw using RF and instant weighting. ## 6\. Ethical Considerations One of the main motivations for the work presented in this paper concerns user privacy and integrity. Traffic cameras and automated license plate recognition devices allow for unprecedented levels of identification and tracking. This is all the more true for user data obtained from cellular networks and mobile devices. Our approach as presented in this paper uses data that is inherently privacy-preserving - we use readily available radio frequency counters that are aggregated on cell level and per definition do not contain any information about individual users, nor could this information be reconstructed. It is therefore impossible to identify or track any individual user based on this data. With that in mind we can state that the work presented in this paper does not raise any ethical issues. ## 7\. Conclusion Traffic flow estimation has traditionally involved forecasting methods based on observations from dedicated traffic sensors. Firstly these methods don’t scale well since we require large number of sensors. Secondly we would need a separate forecasting model for every road, since roads don’t exhibit homogeneous behaviour. Finally our traffic estimation performance would be susceptible to drastic changes in driver behaviour or road conditions, such as traffic accidents and road works. To overcome these limitations alternative approaches have been proposed, including using various forms of cellular network data to estimate traffic flow. However existing approaches are either user invasive, or can potentially result in adverse operational impacts to cellular networks. In this paper we propose a traffic flow estimation method using inherently anonymous and widely available LTE/E-UTRA radio frequency counters, namely path loss and timing advance counters, effectively turning LTE eNBs into traffic sensors. We cast traffic flow estimation as a supervised regression problem, where path loss and timing advance counters are used as primary features, and vehicle counts from actual traffic sensors as target or ground truth variables. We demonstrated excellent performance using both Random Forest and LSTM regression models. Since we have limited amount of ground truth data, i.e. we only had access to six different locations, we also evaluated the performance of two different transfer learning approaches, namely instant weighting, and deep domain adaptation. With transfer learning we demonstrated reasonable performance using either Random Forests or LSTMs, but using only timing advance features. Our hypothesis is that with more data and more locations the performance will improve further still. While our models are not perfect estimators, they are still extremely useful - they capture the shape of the traffic very well, and for most purposes provide a good-enough estimate of the traffic flow. The output of these models can be used for anomaly detection, for example for detecting traffic congestion or accidents. All this can be achieved without having to install any additional sensors - we simply re-use LTE radio base stations that are permanently fixed in their locations with near 100% uptime. ## Acknowledgments We would like to thank the following people for their support throughout the project, and for facilitating the network and traffic sensor data without which none of this would be possible: Elin Allison, Madeleine Körling and Jyrki Lehtinen from Telia Company AB; Anders Broberg and Tobias Johansson from City of Stockholm; Annika Engström from KTH Royal Institute of Technology and Digital Demo Stockholm; Chris Deakin and Chris Holmes from WM5G Limited; Mo Elhabiby and Mike Grogan from Vodafone UK. We would also like to extend our gratitude to Leif Jonsson, Jesper Derehag, Carolyn Cartwright and Simone Ferlin, for reviewing our paper and providing valuable feedback. ## References * (1) * 3GPP (2019) 3GPP. 2019. _Performance measurements Evolved Universal Terrestrial Radio Access Network (E-UTRAN)_. Technical Specification (TS) 32.425. 3rd Generation Partnership Project (3GPP). V16.5.0. * Caceres et al. (2012) Noelia Caceres, Luis M Romero, Francisco G Benitez, and Jose M del Castillo. 2012. Traffic flow estimation models using cellular phone data. _IEEE Transactions on Intelligent Transportation Systems_ 13, 3 (2012), 1430–1441. * Du et al. (2017) Shengdong Du, Tianrui Li, Xun Gong, Yan Yang, and Shi Jinn Horng. 2017. Traffic flow forecasting based on hybrid deep learning framework. In _2017 12th International Conference on Intelligent Systems and Knowledge Engineering (ISKE)_. IEEE, 1–6. * Fleming ([n.d.]) Sean Fleming. [n.d.]. Traffic congestion cost the US economy nearly $87 billion in 2018. _World Economic Forum_ ([n. d.]). https://www.weforum.org/agenda/2019/03/traffic-congestion-cost-the-us-economy-nearly-87-billion-in-2018/ * George et al. (2013) Jobin George, Leena Mary, and KS Riyas. 2013. Vehicle detection and classification from acoustic signal using ANN and KNN. In _2013 international conference on control communication and computing (ICCC)_. IEEE, 436–439. * Gretton et al. (2012) Arthur Gretton, Dino Sejdinovic, Heiko Strathmann, Sivaraman Balakrishnan, Massimiliano Pontil, Kenji Fukumizu, and Bharath K Sriperumbudur. 2012\. Optimal kernel choice for large-scale two-sample tests. In _Advances in neural information processing systems_. 1205–1213. * Haferkamp et al. (2017) Marcus Haferkamp, Manar Al-Askary, Dennis Dorn, Benjamin Sliwa, Lars Habel, Michael Schreckenberg, and Christian Wietfeld. 2017. Radio-based traffic flow detection and vehicle classification for future smart cities. In _2017 IEEE 85th Vehicular Technology Conference (VTC Spring)_. IEEE, 1–5. * Hansapalangkul et al. (2007) T Hansapalangkul, P Keeratiwintakorn, and W Pattara-Atikom. 2007\. Detection and estimation of road congestion using cellular phones. In _2007 7th International Conference on ITS Telecommunications_. IEEE, 1–4. * Hata (1980) M. Hata. 1980. Empirical formula for propagation loss in land mobile radio service. _IEEE Transacations on Vehicular and Technology VT-29_ 3 (1980), 317–325. * Hongsakham et al. (2008) W Hongsakham, W Pattara-Atikom, and R Peachavanish. 2008\. Estimating road traffic congestion from cellular handoff information using cell-based neural networks and K-means clustering. In _2008 5th International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology_ , Vol. 1. IEEE, 13–16. * Huang et al. (2006) Jiayuan Huang, Arthur Gretton, Karsten Borgwardt, Bernhard Schölkopf, and Alex Smola. 2006\. Correcting sample selection bias by unlabeled data. _Advances in neural information processing systems_ 19 (2006), 601–608. * Ji and Hong (2019) Byoungsuk Ji and Ellen J Hong. 2019. Deep-learning-based real-time road traffic prediction using long-term evolution access data. _Sensors_ 19, 23 (2019), 5327. * Lawrence A. Klein (2006) David R. P. Gibson Lawrence A. Klein, Milton K. Mills. 2006. Chapter 1 - Introduction. In _Traffic Detector Handbook: Third Edition \- Volume I_. Federal Highway Administration, 1–2. * Long et al. (2015) Mingsheng Long, Yue Cao, Jianmin Wang, and Michael Jordan. 2015\. Learning transferable features with deep adaptation networks. In _International conference on machine learning_. PMLR, 97–105. * Ma et al. (2013) Wenteng Ma, Daniel Xing, Adam McKee, Ravneet Bajwa, Christopher Flores, Brian Fuller, and Pravin Varaiya. 2013. A wireless accelerometer-based automatic vehicle classification prototype system. _IEEE Transactions on Intelligent Transportation Systems_ 15, 1 (2013), 104–111. * Nam et al. (2020) Daisik Nam, Riju Lavanya, R Jayakrishnan, Inchul Yang, and Woo Hoon Jeon. 2020. A Deep Learning Approach for Estimating Traffic Density Using Data Obtained from Connected and Autonomous Probes. _Sensors_ 20, 17 (2020), 4824. * Pan and Yang (2009) Sinno Jialin Pan and Qiang Yang. 2009. A survey on transfer learning. _IEEE Transactions on knowledge and data engineering_ 22, 10 (2009), 1345–1359. * Pattara-Atikom and Peachavanish (2007) Wasan Pattara-Atikom and Ratchata Peachavanish. 2007. Estimating road traffic congestion from cell dwell time using neural network. In _2007 7th International Conference on ITS Telecommunications_. IEEE, 1–6. * Smagulova and James (2019) Kamilya Smagulova and Alex Pappachen James. 2019. A survey on LSTM memristive neural network architectures and applications. _The European Physical Journal Special Topics_ 228, 10 (2019), 2313–2324. * Tan et al. (2018) Chuanqi Tan, Fuchun Sun, Tao Kong, Wenchang Zhang, Chao Yang, and Chunfang Liu. 2018\. A survey on deep transfer learning. In _International conference on artificial neural networks_. Springer, 270–279. * Wang et al. (2017) Jitian Wang, Han Zheng, Yue Huang, and Xinghao Ding. 2017\. Vehicle type recognition in surveillance images from labeled web-nature data using deep transfer learning. _IEEE Transactions on Intelligent Transportation Systems_ 19, 9 (2017), 2913–2922. * Wang et al. (2020) Pu Wang, Jiyu Lai, Zhiren Huang, Qian Tan, and Tao Lin. 2020. Estimating Traffic Flow in Large Road Networks Based on Multi-Source Traffic Data. _IEEE Transactions on Intelligent Transportation Systems_ (2020). * Xing et al. (2019) Jiping Xing, Zhiyuan Liu, Chunliang Wu, and Shuyan Chen. 2019\. Traffic Volume Estimation in Multimodal Urban Networks Using Cell Phone Location Data. _IEEE Intelligent Transportation Systems Magazine_ 11, 3 (2019), 93–104. * Zewei et al. (2015) Xu Zewei, Wei Jie, and Chen Xianqiao. 2015. Vehicle recognition and classification method based on laser scanning point cloud data. In _2015 International Conference on Transportation Information and Safety (ICTIS)_. IEEE, 44–49. * Zhuang et al. (2019) Fuzhen Zhuang, Zhiyuan Qi, Keyu Duan, Dongbo Xi, Yongchun Zhu, Hengshu Zhu, Hui Xiong, and Qing He. 2019\. A comprehensive survey on transfer learning. _arXiv preprint arXiv:1911.02685_ (2019).
# Streaming Models for Joint Speech Recognition and Translation Orion Weller1*, Matthias Sperber2, Christian Gollan2, Joris Kluivers2 1Brigham Young University 2Apple oweller<EMAIL_ADDRESS> ###### Abstract Using end-to-end models for speech translation (ST) has increasingly been the focus of the ST community. These models condense the previously cascaded systems by directly converting sound waves into translated text. However, cascaded models have the advantage of including automatic speech recognition output, useful for a variety of practical ST systems that often display transcripts to the user alongside the translations. To bridge this gap, recent work has shown initial progress into the feasibility for end-to-end models to produce both of these outputs. However, all previous work has only looked at this problem from the consecutive perspective, leaving uncertainty on whether these approaches are effective in the more challenging streaming setting. We develop an end-to-end streaming ST model based on a re-translation approach and compare against standard cascading approaches. We also introduce a novel inference method for the joint case, interleaving both transcript and translation in generation and removing the need to use separate decoders. Our evaluation across a range of metrics capturing accuracy, latency, and consistency shows that our end-to-end models are statistically similar to cascading models, while having half the number of parameters. We also find that both systems provide strong translation quality at low latency, keeping 99% of consecutive quality at a lag of just under a second. ††*Work done during an internship with Apple ## 1 Introduction Speech translation (ST) is the process of translating acoustic sound waves into text in a different language than was originally spoken in. This paper focuses on ST in a particular setting, as described by two characteristics: (1) We desire models that translate in a streaming fashion, where users desire the translation before the speaker has finished. This setting poses additional difficulties compared to consecutive translation, forcing systems to translate without knowing what the speaker will say in the future. (2) Furthermore, the speaker may want to verify that their speech is being processed correctly, intuitively seeing a streaming transcript while they speak Fügen (2008); Hsiao et al. (2006). For this reason, we consider models that produce both transcripts and translation jointly.111This corresponds to the mandatory transcript case in the proposed categorization by Sperber and Paulik (2020). Previous approaches to streaming ST have typically utilized a cascaded system that pipelines the output of an automatic speech recognition (ASR) system through a machine translation (MT) model for the final result. These systems have been the preeminent strategy, taking the top place in recent streaming ST competitions Pham et al. (2019); Jan et al. (2019); Elbayad et al. (2020); Ansari et al. (2020). Despite the strong performance of these cascaded systems, there are also some problems: error propagation from ASR output to MT input Ruiz and Federico (2014); ASR/MT training data mismatch and loss of access to prosodic/paralinguistic speech information at the translation stage Sperber and Paulik (2020); and potentially sub-optimal latencies in the streaming context. End-to-end (E2E) models for ST have been proposed to remedy these problems, leveraging the simplicity of a single model to sidestep these issues. E2E models are also appealing from computational and engineering standpoints, reducing model complexity and decreasing parameter count. Although initial research has explored E2E models for joint speech recognition and translation, no previous works have examined them in the streaming case, a crucial step in using them for many real-world applications. To understand this area more fully, we develop an E2E model to compare with its cascading counterpart in this simultaneous joint task. We build off the models proposed by Sperber et al. (2020) in the consecutive case, extending them for use in the streaming setting. We also use the re-translation technique introduced by Niehues et al. (2018) to maintain simplicity while streaming. To reduce model size, we introduce a new method for E2E inference, producing both transcript and translation in an interleaved fashion with one decoder. As this task requires a multi-faceted evaluation along several axes, we provide a suite of evaluations to highlight the differences of these major design decisions. This suite includes assessing translation quality, transcription quality, lag of the streaming process, output flicker, and consistency between the transcription and translation. We find that our E2E model performs similarly to the cascaded model, indicating that E2E networks are a feasible and promising direction for streaming ST. ## 2 Proposed Method ### Network Architecture In the ST survey provided by Sperber et al. (2020), they introduce several E2E models that could be used for the joint setting. As our work focuses on providing a simple but effective approach to streaming ST, we focus on the CONCAT model, which generates both the transcript and translation in a concatenated fashion. We compare this E2E model against the standard cascading approach, following the architecture and hyperparameter choices used in Sperber et al. (2020). All audio input models use the same multi-layer bidirectional LSTM architecture, stacking and downsampling the audio by a factor of three before processing. We note that although bidirectional encoders are unusual with standard ASR architectures, re-translation makes them possible. The cascaded model’s textual encoder follows the architecture described in Vaswani et al. (2017) but replaces self-attention blocks with LSTMs. Decoder networks are similar, but use unidirectional LSTMs. More implementation details can be found in Appendix A. In order to reduce model size and inference time for E2E networks, we introduce a novel method for interleaving both transcript and translation in generation, removing the need to use separate decoders. This method extends the CONCAT model proposed by Sperber et al. (2020) to jointly decode according to the ratio given by the parameter $\gamma$ (Figure 1). When $\gamma=0.0$, we generate the transcript tokens until completion, followed by the translation tokens (vice versa for $\gamma=1.0$). At $\gamma=0.0$, our model is equivalent to the previously proposed model. Defining $\mathrm{count_{i}}$ as the count of $i$ tokens previously generated, transcription tokens as st and translation tokens as tt, we generate the next token as a transcription token if: $\displaystyle(1.0-\gamma)*(1+\mathrm{count_{tt}})>\gamma*(1+\mathrm{count_{st}})$ This approach enables us to produce tokens in an interleaving fashion, given the hyperparameter $\gamma$. Figure 1: Example token representations (En→De) for three different interleaving parameters (Section 2). Language tokens indicate whether the data corresponds to the source transcript or the target translation and are used with a learned embedding that is summed with the word embeddings, as described in Sperber et al. (2020). Figure 2: Left: average lag in seconds vs BLEU score. Right: average lag in seconds vs WER score. All points are the mean of each configuration’s score across the eight target languages. Configurations are the cross product of the values for $K$ and $F$, see Section 2: Inference. Note that points near 1.0 AL have appx. 99% of the unconstrained BLEU score. Results for the E2E model use $\gamma=0.5$. . Metric | Params | Model | De | Es | Fr | It | Nl | Pt | Ro | Ru | Average ---|---|---|---|---|---|---|---|---|---|---|--- BLEU $\uparrow$ | 217M | Cascade | 18.8 | 22.7 | 27.0 | 18.9 | 22.5 | 21.9 | 17.9 | 13.0 | 20.3 | 107M | E2E $\gamma$=0.0 | 18.1 | 23.1 | 27.0 | 18.7 | 22.3 | 22.2 | 17.6 | 12.2 | 20.2 | 107M | E2E $\gamma$=0.3 | 17.7 | 22.6 | 26.3 | 18.0 | 21.5 | 21.5 | 17.0 | 12.1 | 19.6 | 107M | E2E $\gamma$=0.5 | 18.2 | 22.8 | 27.0 | 18.6 | 21.9 | 21.9 | 17.1 | 12.0 | 19.9 | 107M | E2E $\gamma$=1.0 | 18.2 | 22.8 | 27.1 | 18.9 | 22.2 | 22.3 | 17.6 | 12.7 | 20.2 WER $\downarrow$ | 217M | Cascade | 25.9 | 24.0 | 23.1 | 25.6 | 28.5 | 26.4 | 24.4 | 23.1 | 25.1 | 107M | E2E $\gamma$=0.0 | 24.2 | 23.5 | 23.3 | 23.0 | 23.4 | 25.3 | 24.1 | 23.6 | 23.8 | 107M | E2E $\gamma$=0.3 | 24.1 | 23.6 | 22.9 | 23.8 | 23.4 | 25.7 | 24.1 | 24.1 | 24.0 | 107M | E2E $\gamma$=0.5 | 24.5 | 23.9 | 22.9 | 23.8 | 23.4 | 25.7 | 24.3 | 23.6 | 24.0 | 107M | E2E $\gamma$=1.0 | 23.6 | 22.9 | 22.3 | 23.0 | 22.4 | 24.7 | 23.4 | 22.7 | 23.1 Table 1: BLEU and WER scores for models trained on different target languages. Bold scores indicate results that are statistically similar to the best score using a bootstrap permutation test with $\alpha=0.05$. ### Re-translation We use the re-translation method Niehues et al. (2018); Arivazhagan et al. (2020a, b) as it provides a simple way to handle the streaming case. This method works by simply re-translating the utterance as new data arrives, updating its former prediction. As we are generating both transcript and translation, this avoids the challenging issue of combining the requirements for both components: streaming speech models need to manage the audio signal variability across time while streaming translation models need to overcome issues with reordering and lack of future context. Alternative strategies to the re-translation approach include the chunk-based strategy explored by Liu et al. (2020), which commits to all previous output chunks and Ren et al. (2020) who utilize an additional segmenter model trained via CTC Graves et al. (2006) to create segments that are translated via wait-k Ma et al. (2019). Although these approaches show effective results, they add additional complexity without addressing issues particular to streaming transcription. ### Inference In order to generate quality-latency curves, we use several techniques to reduce latency and flicker at the cost of quality. The first is the mask-k method proposed by Arivazhagan et al. (2020b), masking the last $K$ output tokens. The second method is a form of constrained decoding: we define a hyperparameter $F$ that sets the number of free tokens allowed to change in the next re-translation. Thus, we constrain future output to match the first $\textit{len}(\textit{tokens})-F$ tokens of the current output. All models use values $\\{0,1,2,3,4,5,7,10,100\\}$ for $K$ and $\\{0,1,2,3,4,5,7,10,15,20,25,100\\}$ for $F$. For interleaving models, we set $K$ and $F$ on both transcript and translation tokens. Model | En→De Incr. | En→De Full | En→Es Incr. | En→Es Full | Mean Incr. | Mean Full ---|---|---|---|---|---|--- Cascade | 13.8 | 13.2 | 12.2 | 11.6 | 14.1 | 13.4 Concat $\gamma$=0.0 | 17.6 | 16.7 | 14.9 | 13.8 | 17.0 | 16.0 Concat $\gamma$=0.3 | 17.2 | 16.6 | 14.3 | 13.7 | 16.6 | 15.8 Concat $\gamma$=0.5 | 17.8 | 16.5 | 14.8 | 13.3 | 17.3 | 15.7 Concat $\gamma$=1.0 | 17.3 | 16.8 | 14.9 | 13.7 | 16.9 | 15.8 Table 2: Consistency scores for En→De, En→Es, and average results over all languages; lower is better (see Sperber et al. (2020)). _Incr._ stands for the incremental consistency score, or the average consistency throughout re- translation. Bold scores indicate results that are statistically similar to the best score using a bootstrap permutation test with $\alpha=0.05$. ## 3 Experimental Settings ### Data We use the MuST-C corpus di Gangi et al. (2019) since it is the largest publicly available ST corpus, consisting of TED talks with their English transcripts and translations into eight other language pairs. The dataset consists of at least 385 hours of audio for each target language. We utilize the log Mel filterbank speech features provided with the corpus as input for the ASR and E2E models. To prepare the textual data, we remove non- speech artifacts (e.g. “(laughter)” and speaker identification) and perform subword tokenization using SentencePiece Kudo and Richardson (2018) on the unigram setting. Following previous work for E2E ST models, we use a relatively small vocabulary and share transcription and translation vocabularies. We use MuST-C dev for validation and report results on tst- COMMON, utilizing the segments provided (Appendix D). ### Prefix Sampling We implement techniques developed by Niehues et al. (2018); Arivazhagan et al. (2020b) for improving streaming ST, sampling a random proportion of each training instance as additional data to teach our models to work with partial input. See Appendix C for implementation details. ### Metrics We evaluate these models on a comprehensive suite of metrics: sacrebleu (_BLEU_ , Post (2019)) for translation quality, word error rate (_WER_ , Fiscus (1997)) for transcription quality, average lag (_AL_ , Ma et al. (2019)) for the lag between model input and output, and normalized erasure (_NE_ , Arivazhagan et al. (2020a)) for output flicker. Measuring consistency is a nascent area of research; we use the robust and simple lexical consistency metric defined by Sperber et al. (2020), which uses word-level translation probabilities. To show how consistent these results are while streaming, we compute an incremental consistency score, averaging the consistency of each re-translation. ## 4 Results Results for the quality-latency curves created by the use of constrained decoding and mask-k (Section 3) are shown in Figure 2. Unconstrained settings are used for all results in table form. For convenience, bold scores indicate the highest performing models in each metric according to a bootstrap permutation test. ### Translation Quality We see in Table 1 that the cascaded model slightly outperforms some E2E models, while achieving statistically similar performance to the $\gamma=1.0$ model. We note however, that the cascaded model has nearly twice as many parameters as the E2E models (217M vs 107M). When we examine these models under a variety of different inference conditions (using constrained decoding and mask-k as in Arivazhagan et al. (2020a)), we further see this trend illustrated through the quality vs latency trade-off (left of Figure 2), with both models retaining 99% of their BLEU at less than 1.0 AL. ### Transcription Quality Conversely, Table 1 and the right of Figure 2 show that the $\gamma=1.0$ E2E model performs similarly or slightly better than the cascaded model across all inference parameters and all target languages. With an AL of 1.5, the E2E model loses only 3% of its performance. ### Consistency The E2E models perform worse than the cascaded on consistency, with the best models being approximately 18% less consistent (Table 2). The cascaded model also maintains better scores through each re-translation (_Incr._).222Initial experiments indicate that the triangle E2E architecture Sperber et al. (2020) model may perform better on consistency in our streaming setting, but due to time constraints we were not able to explore this further. Future work exploring alternative architectures or decoding techniques Le et al. (2020) may provide fruitful avenues of research. ### Flicker We note that the flicker scores for cascade and E2E models are similar, with both having normalized erasure scores of less than 1 and the majority of inference settings having less than the “few-revision” threshold of 0.2 (proposed by Arivazhagan et al. (2020a)). More NE details are found in Appendix B. ### Interleaving Rate Table 1 also shows us the overall results for different interleaving rates. We see that interleaving at a rate of 1.0 has the best quality scores (0.7 less WER than the next best rate, the base $\gamma=0.0$ model) but the worst consistency (Table 2). Conversely, $\gamma=0.3$ has the worst quality scores but the best consistency. ## 5 Conclusion We focus on the task of streaming speech translation, producing both a target translation and a source transcript from an audio source. We develop an end- to-end model to avoid problems that arise from the use of cascaded models for streaming ST. We further introduce a new method for joint inference for end- to-end models, generating both translation and transcription tokens concurrently. We show that our novel end-to-end model, with only half the number of parameters, is comparable to standard cascaded models across a variety of evaluation categories: transcript and translation quality, lag of streaming, consistency between transcript and translation, and re-translation flicker. We hope that this will spur increased interest in using end-to-end models for practical applications of streaming speech translation. ## References * Ansari et al. (2020) Ebrahim Ansari, Amittai Axelrod, Nguyen Bach, Ondřej Bojar, Roldano Cattoni, Fahim Dalvi, Nadir Durrani, Marcello Federico, Christian Federmann, Jiatao Gu, Fei Huang, Kevin Knight, Xutai Ma, Ajay Nagesh, Matteo Negri, Jan Niehues, Juan Pino, Elizabeth Salesky, Xing Shi, Sebastian Stüker, Marco Turchi, Alexander Waibel, and Changhan Wang. 2020. FINDINGS OF THE IWSLT 2020 EVALUATION CAMPAIGN. In _Proceedings of the 17th International Conference on Spoken Language Translation_ , pages 1–34, Online. Association for Computational Linguistics. * Arivazhagan et al. (2020a) Naveen Arivazhagan, Colin Cherry, Wolfgang Macherey, and George Foster. 2020a. Re-translation versus streaming for simultaneous translation. _arXiv preprint arXiv:2004.03643_. * Arivazhagan et al. (2020b) Naveen Arivazhagan, Colin Cherry, Isabelle Te, Wolfgang Macherey, Pallavi Baljekar, and George Foster. 2020b. Re-translation strategies for long form, simultaneous, spoken language translation. In _ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_ , pages 7919–7923. IEEE. * Elbayad et al. (2020) Maha Elbayad, Ha Nguyen, Fethi Bougares, Natalia Tomashenko, Antoine Caubrière, Benjamin Lecouteux, Yannick Estève, and Laurent Besacier. 2020\. On-trac consortium for end-to-end and simultaneous speech translation challenge tasks at iwslt 2020. _arXiv preprint arXiv:2005.11861_. * Fiscus (1997) Jonathan G Fiscus. 1997. A post-processing system to yield reduced word error rates: Recognizer output voting error reduction (rover). In _1997 IEEE Workshop on Automatic Speech Recognition and Understanding Proceedings_ , pages 347–354. IEEE. * Fügen (2008) Christian Fügen. 2008. _A System for Simultaneous Translation of Lectures and Speeches_. Ph.D. thesis, University of Karlsruhe. * Gal and Ghahramani (2016) Yarin Gal and Zoubin Ghahramani. 2016. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning. In _International Conference on Machine Learning (ICML)_. * di Gangi et al. (2019) Antonino Mattia di Gangi, Roldano Cattoni, Luisa Bentivogli, Matteo Negri, and Marco Turchi. 2019. MuST-C : a Multilingual Speech Translation Corpus. In _North American Chapter of the Association for Computational Linguistics (NAACL)_ , Minneapolis, USA. * Glorot and Bengio (2010) Xavier Glorot and Yoshua Bengio. 2010. Understanding the difficulty of training deep feedforward neural networks. In _International Conference on Artificial Intelligence and Statistics (AISTATS)_ , Sardinia, Italy. * Graves et al. (2006) Alex Graves, Santiago Fernández, Faustino Gomez, and Jürgen Schmidhuber. 2006. Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks. In _Proceedings of the 23rd international conference on Machine learning_ , pages 369–376. * Hsiao et al. (2006) Roger Hsiao, Ashish Venugopal, Thilo Köhler, Ting Zhang, Paisarn Charoenpornsawat, Andreas Zollmann, Stephan Vogel, Alan W. Black, Tanja Schultz, and Alex Waibel. 2006. Optimizing components for handheld two-way speech translation for an English-Iraqi Arabic system. In _Annual Conference of the International Speech Communication Association (InterSpeech)_ , pages 765–768, Pittsburgh, USA. * Jan et al. (2019) Niehues Jan, Roldano Cattoni, Stuker Sebastian, Matteo Negri, Marco Turchi, Salesky Elizabeth, Sanabria Ramon, Barrault Loic, Specia Lucia, and Marcello Federico. 2019. The iwslt 2019 evaluation campaign. In _16th International Workshop on Spoken Language Translation 2019_. * Kingma and Ba (2014) Diederik P. Kingma and Jimmy L. Ba. 2014. Adam: A Method for Stochastic Optimization. In _International Conference on Learning Representations (ICLR)_ , Banff, Canada. * Kudo and Richardson (2018) Taku Kudo and John Richardson. 2018. SentencePiece: A simple and language independent subword tokenizer and detokenizer for neural text processing. _Empirical Methods in Natural Language Processing (EMNLP)_ , pages 66–71. * Le et al. (2020) Hang Le, Juan Pino, Changhan Wang, Jiatao Gu, Didier Schwab, and Laurent Besacier. 2020. Dual-decoder transformer for joint automatic speech recognition and multilingual speech translation. In _Proceedings of the 28th International Conference on Computational Linguistics_ , pages 3520–3533, Barcelona, Spain (Online). International Committee on Computational Linguistics. * Liu et al. (2020) Danni Liu, Gerasimos Spanakis, and Jan Niehues. 2020. Low-latency sequence-to-sequence speech recognition and translation by partial hypothesis selection. _arXiv preprint arXiv:2005.11185_. * Ma et al. (2019) Mingbo Ma, Liang Huang, Hao Xiong, Renjie Zheng, Kaibo Liu, Baigong Zheng, Chuanqiang Zhang, Zhongjun He, Hairong Liu, Xing Li, et al. 2019. Stacl: Simultaneous translation with implicit anticipation and controllable latency using prefix-to-prefix framework. In _Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics_ , pages 3025–3036. * Neubig et al. (2018) Graham Neubig, Matthias Sperber, Xinyi Wang, Matthieu Felix, Austin Matthews, Sarguna Padmanabhan, Ye Qi, Devendra Singh Sachan, Philip Arthur, Pierre Godard, John Hewitt, Rachid Riad, and Liming Wang. 2018. XNMT: The eXtensible Neural Machine Translation Toolkit. In _Conference of the Association for Machine Translation in the Americas (AMTA) Open Source Software Showcase_ , Boston, USA. * Niehues et al. (2018) Jan Niehues, Ngoc-Quan Pham, Thanh-Le Ha, Matthias Sperber, and Alex Waibel. 2018\. Low-latency neural speech translation. _arXiv preprint arXiv:1808.00491_. * Paszke et al. (2019) Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Köpf, Edward Yang, Zach DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In _Advances in Neural Information Processing Systems (NeuIPS)_ , Vancouver, Canada. * Pham et al. (2019) Ngoc-Quan Pham, Thai-Son Nguyen, Thanh-Le Ha, Juan Hussain, Felix Schneider, Jan Niehues, Sebastian Stüker, and Alexander Waibel. 2019. The iwslt 2019 kit speech translation system. In _Proceedings of the 16th International Workshop on Spoken Language Translation_. * Post (2019) Matt Post. 2019. A Call for Clarity in Reporting BLEU Scores. In _Conference on Machine Translation (WMT)_ , pages 186–191, Brussels, Belgium. * Ren et al. (2020) Yi Ren, Jinglin Liu, Xu Tan, Chen Zhang, QIN Tao, Zhou Zhao, and Tie-Yan Liu. 2020\. Simulspeech: End-to-end simultaneous speech to text translation. In _Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics_ , pages 3787–3796. * Ruiz and Federico (2014) Nicholas Ruiz and Marcello Federico. 2014. Assessing the impact of speech recognition errors on machine translation quality. In _11th Conference of the Association for Machine Translation in the Americas (AMTA), Vancouver, BC, Canada_. * Sperber and Paulik (2020) Matthias Sperber and Matthias Paulik. 2020. Speech Translation and the End-to-End Promise: Taking Stock of Where We Are. In _Association for Computational Linguistic (ACL)_ , Seattle, USA. * Sperber et al. (2020) Matthias Sperber, Hendra Setiawan, Christian Gollan, Udhyakumar Nallasamy, and Matthias Paulik. 2020. Consistent Transcription and Translation of Speech. _Transactions of the Association for Computational Linguistics (TACL)_. * Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention Is All You Need. In _Neural Information Processing Systems Conference (NIPS)_ , pages 5998–6008, Long Beach, USA. ## Appendix A Model Details In this section we will describe implementation details of the model architectures (shown in Figure 3) and training processes. ### Model Architectures Unless otherwise noted, the same hyperparameters are used for all models. Weights for the speech encoder are initialized based on a pre-trained attentional ASR task that is identical to the ASR part of the direct multitask model. Other weights are initialized according to Glorot and Bengio (2010). The speech encoder is a 5-layer bidirectional LSTM with 700 dimensions per direction. Attentional decoders consist of 2 Transformer blocks Vaswani et al. (2017) but use 1024-dimensional unidirectional LSTMs instead of self- attention, except for the CONCAT model, which uses 3 layers. For the cascade’s MT model, encoder/decoder both contain 6 layers with 1024-dimensional LSTMs. Subword embeddings are size 1024. We regularize using LSTM dropout with $p=0.3$, decoder input word-type dropout Gal and Ghahramani (2016), and attention dropout, both $p=0.1$. We apply label smoothing with strength $\epsilon=0.1$. Figure 3: Architectures of the cascade and concatenated model ### Training We optimize using Adam Kingma and Ba (2014) with $\alpha=0.0005$, $\beta_{1}=0.9$, $\beta_{2}=0.98$, 4000 warm-up steps, and learning rate decay by using the inverse square root of the iteration of each instance. We set the batch size dynamically based on the sentence length, such that the average batch size is 128 utterances. The training is stopped when the validation score has not improved over 10 epochs, where the validation score is corpus- level translation BLEU score (for the E2E and MT models) and corpus-level WER for the cascade’s ASR model. Figure 4: Left: average lag in seconds vs NE score. Right: NE vs WER score. All values are the mean of the results from the eight target languages. For decoding and generating n-best lists, we use beam size 5 and polynomial length normalization with exponent 1.5. Our implementation is based on PyTorch Paszke et al. (2019) and XNMT Neubig et al. (2018), and all models are trained in single-GPU environments, employing Tesla V100 GPUs with 32 GB memory. Most E2E and ASR models converged after approximately 30 epochs or 5 days of training. MT models converged after approximately 50 epochs or 2 days of training. ## Appendix B Normalized Erasure (Output Flicker) We see similar curves for both the cascaded model and the E2E model when comparing normalized erasure in Figure 4. We see that most settings have an NE score of less than 0.2, while virtually all settings are less than 1. We note that a proportion of 0.2 for NE means that, on average, 1/5 of the tokens change once before they settle to their final state. ## Appendix C Prefix Training We used prefix training to increase stability and reduce flickering in the streaming setting. We conducted this by utilizing each training instance twice in each epoch: one as normal and the other with only the prefix. The length of the prefixes were randomly sampled from [0, 1]. We found that this additional data augmentation was particularly helpful; without it, the models would hallucinate the rest of a partial sentence. We further found that starting the prefix sampling data augmentation too late in training was also negative. After testing initial models on the dev set, we found that starting this additional augmentation 15 epochs after training was best. ## Appendix D Utterance Segmentation We follow the audio segments provided in the MuST-C corpus, created through a use of human alignment and XNMT Neubig et al. (2018). We note that there exist a variety of methods for creating segments for such models, however, we leave additional exploration of E2E alignment methods as future work.
# Virtual laser scanning with HELIOS++: A novel take on ray tracing-based simulation of topographic 3D laser scanning Lukas Winiwarter<EMAIL_ADDRESS>Alberto Manuel Esmorís Pena Hannah Weiser Katharina Anders Jorge Martínez Sanchez Mark Searle Bernhard Höfle<EMAIL_ADDRESS>3DGeo Research Group, Institute of Geography, Heidelberg University, Germany Centro Singular de Investigación en Tecnoloxías Intelixentes, CiTIUS, USC, Spain Interdisciplinary Center for Scientific Computing (IWR), Heidelberg University, Germany ###### Abstract Topographic laser scanning is a remote sensing method to create detailed 3D point cloud representations of the Earth’s surface. Since data acquisition is expensive, simulations can complement real data given certain premises are available: i) a model of 3D scene and scanner, ii) a model of the beam-scene interaction, simplified to a computationally feasible while physically realistic level, and iii) an application for which simulated data is fit for use. A number of laser scanning simulators for different purposes exist, which we enrich by presenting HELIOS++. HELIOS++ is an open-source simulation framework for terrestrial static, mobile, UAV-based and airborne laser scanning implemented in C++. The HELIOS++ concept provides a flexible solution for the trade-off between physical accuracy (realism) and computational complexity (runtime, memory footprint), as well as ease of use and of configuration. Unique features of HELIOS++ include the availability of Python bindings (`pyhelios`) for controlling simulations, and a range of model types for 3D scene representation. Such model types include meshes, digital terrain models, point clouds and partially transmissive voxels, which are especially useful in laser scanning simulations of vegetation. In a scene, object models of different types can be combined, so that representations spanning multiple spatial scales in different resolutions and levels of detail are possible. We aim for a modular design, where the core components of platform, scene, and scanner can be individually interchanged, and easily configured by working with XML files and Python bindings. Virtually scanned point clouds may be used for a broad range of applications. Our literature review of publications employing virtual laser scanning revealed the four categories of use cases prevailing at present: data acquisition planning, method evaluation, method training and sensing experimentation. To enable direct interaction with 3D point cloud processing and GIS software, we support standard data formats for input models (Wavefront Objects, GeoTIFFs, ASCII xyz point clouds) and output point clouds (LAS/LAZ and ASCII). HELIOS++ further allows the simulation of beam divergence using a subsampling strategy, and is able to create full- waveform outputs as a basis for detailed analysis. As generation and analysis of waveforms can strongly impact runtimes, the user may set the level of detail for the subsampling, or optionally disable full-waveform output altogether. A detailed assessment of computational considerations and a comparison of HELIOS++ to its predecessor, HELIOS, reveal reduced runtimes by up to 83 %. At the same time, memory requirements are reduced by up to 94 %, allowing for much larger (i.e. more complex) 3D scenes to be loaded into memory and hence to be virtually acquired by laser scanning simulation. ###### keywords: software , LiDAR simulation , point cloud , data generation , voxel , vegetation modelling , diffuse media ††journal: Remote Sensing of Environment ## 1 Introduction Simulation of physical processes is often carried out when experiments are not feasible or simply impossible, or to find parameters that produce a certain outcome if inversion is non-trivial. In virtual laser scanning (VLS), simulations of LiDAR (Light Detection and Ranging) create 3D point clouds from models of scenes, platforms, and scanners (Figure 1), that aim to recreate real-world scenarios of laser scanning acquisitions. Such simulated point clouds may, for certain use cases, replace real data, and may even allow for analyses where real data capture is not feasible, e.g. due to technical, economical or logistic constraints, or when simulating hardware which is not yet existing. However, there are use cases where VLS is not appropriate, for example, when analysing effects only partially modelled in the simulation such as penetration of the laser into opaque objects. In a similar argument, (passive) photogrammetry is inadequate for the reconstruction of a non- textured flat area, but still a useful method for many other tasks. Therefore, VLS can be seen as a tool to acquire 3D geospatial data under certain _premises_. These include: 1. 1. an adequate model of the 3D scene and the scanner, as well as the platform behaviour, 2. 2. a simplification of the real-world beam-scene interactions to a computationally feasible and physically realistic level, and finally, 3. 3. an application for which VLS data is fit for use. VLS can easily and cheaply produce large amounts of data with very well defined properties and known ground truth. Parameters (e.g. tree attributes such as crown base height) can be extracted from the scene model (e.g. a mesh object that is being scanned) automatically and without errors, and these parameters can in turn be used for training or validation of algorithms that attempt to extract them from point cloud data. Due to the low cost compared to real data acquisitions, VLS can be combined with Monte-Carlo simulations to solve non-continuous optimisation problems on scan settings and acquisition strategies. In a research workflow, VLS experiments may be employed to identify promising candidate settings before carrying out a selected number of real experiments that are used to answer the respective research questions. Figure 1: Schematic concept of HELIOS++, showcasing platforms (boxed labels) and object models composing a scene (non-boxed labels). A variety of model types to represent 3D scenes are supported: terrain models, voxel models (custom .vox format or XYZ point clouds) and mesh models. For platforms, four options are currently supported: airplane, multicopter, ground vehicle and static tripod. A schematic diverging laser beam and its corresponding waveform (magenta) is shown being emitted from the airplane and interacting with a mesh model tree and the rasterised ground surface. In this paper, we present a novel take on ray tracing-based VLS covering the given premises (1-3). This is implemented as the open source software package HELIOS++ (Heidelberg LiDAR Operations Simulator ++)111HELIOS++ is available on GitHub (https://github.com/3dgeo-heidelberg/helios) and is licensed under both GNU GPL and GNU LGPL. HELIOS++ is also indexed with Zenodo (Winiwarter et al., 2021).. HELIOS++ is the successor of HELIOS (Bechtold and Höfle, 2016), with a completely new code base, implemented in C++ (whereas the former version was implemented in Java). HELIOS++ improves over HELIOS in terms of memory footprint and runtime, and also in functionality, correctness of implemented algorithms, and usability, making it versatile and highly performant to end users. We first motivate the need for a novel VLS framework by a survey of previous methodologies to laser scanning simulation and their applications, and point out the unique features of HELIOS++ (Section 2). In Section 3, we present the architecture and design considerations of HELIOS++. We then show different types of applications and conduct a systematic literature survey of uses of HELIOS in Section 4. Technical considerations concerning the handling of big 3D scenes and ray tracing are dealt with in Section 5 and conclusions are drawn in Section 6. ## 2 Existing implementations and state of the art virtual laser scanning Simulating a process within a system always involves a simplified substitute of reality. The complexity of this substitute depends on the process understanding, on computational considerations and on the specific problem that is to be solved. Approaches with different levels of simulation complexity exist regarding i) the type of input scene model (e.g. 2.5D digital elevation models (DEMs) vs. 3D meshes) and ii) how the interaction of beam and object is modeled (e.g. single ray/echo vs. full-waveform). An overview of publications of these approaches is listed in Table 1. Publication | Platforms | | Beam --- div. FWF | Scene | Comments North (1996), North et al. (2010) | satellite | ✓ | ✓ | 3D | FLIGHT model Lewis (1999) | ALS | ✓ | ✓ | 3D | | used by --- Calders et al. (2013) & Disney et al. (2010) Tulldahl and Steinvall (1999) | ALS | ✓ | ✓ | 3D | for bathymetry Ranson and Sun (2000) | | ALS --- (nadir) ✓ | ✓ | 3D | Holmgren et al. (2003) | ALS | | | 3D | Goodwin et al. (2007) | ALS | | | 3D | LITE model Lohani and Mishra (2007) | ALS | | | 2.5D | Morsdorf et al. (2007) | ALS | ✓ | ✓ | 3D | using POVray Kim et al. (2009) | ALS | | | 3D | Kukko and Hyyppä (2009) | ALS, MLS | ✓ | ✓ | 2.5D | Hodge (2010) | TLS | ✓ | ✓ | 2.5D | Kim et al. (2012) | ALS | ✓ | ✓ | 3D | Wang et al. (2013) | TLS | | | 3D | Gastellu-Etchegorry et al. (2015) | | satellite, --- ALS, TLS ✓ | ✓ | 3D | DART model Bechtold and Höfle (2016) | | ALS, TLS, --- MLS, ULS ✓ | ✓ | 3D | HELIOS Table 1: Overview of virtual laser scanning simulators and associated publications. For each simulator, a check mark (✓) is added if they support simulation of finite (non-zero) beam divergence (”Beam div.”) and full waveforms (”FWF”). Scene representation may be in full 3D or 2.5D, i.e. raster-based. A simple airborne laser scanning (ALS) simulator is presented by Lohani and Mishra (2007) for use in research and education. Their tool comes with a user- friendly graphical user interface (GUI) and allows selecting different scanner and trajectory configurations. This simulator models the laser ray as an infinitesimal beam with zero divergence to simplify the ray tracing procedure. The scene is represented by a 2.5D elevation raster, which allows only a simplified representation of the Earth’s surface. In a different application, the interactions between laser beams and forest canopies are investigated by Goodwin et al. (2007), Holmgren et al. (2003) and Lovell et al. (2005). They present approaches of combining 3D forest modelling and ALS simulation using ray tracing. As in Lohani and Mishra (2007), their simulators all model the laser beam as an infinite straight line, which intersects the scene in one distinct point. Full-waveform laser scanners can record the full waveform of the backscattered signal, providing information about the objects in a scene that are illuminated by the conic beam. This waveform from a finite footprint was specifically simulated by Ranson and Sun (2000) in the forestry context and by Tulldahl and Steinvall (1999) for airborne LiDAR bathymetry. Morsdorf et al. (2007) use the ray tracing software POV-Ray222http://www.povray.org/ to model the waveform of laser scans of a 3D tree model from a combination of intensity and depth images. Kukko and Hyyppä (2009) aim for a more complete and universal LiDAR simulator that considers platform and beam orientation, pulse transmission depending on power distribution and laser beam divergence, beam interaction with the scene, and full-waveform recording. Their approach models the physics involved in LiDAR measurements in high detail, and is demonstrated for a use case in forestry. Similar to the work by Lohani and Mishra (2007), they use 2.5D elevation maps to represent the scene. This makes the simulator useful for airborne simulations over terrain or building models, where the scene is scanned from above and scene elements are assumed to be solid and opaque. However, it is less suited for penetrable 3D objects such as vegetation or overhanging geometries, especially for the simulation of ground-based acquisitions which are less in a bird’s-eye view perspective. Kim et al. (2012) present a similarly detailed simulator, which includes radiometric simulation and recording of the waveform and includes explicit 3D object representations. It is unclear if it also supports static platforms such as TLS. TLS simulations are the sole focus of some studies for specific applications, such as leaf area index inversion (Wang et al., 2013) or TLS measurement error quantification (Hodge, 2010). While Wang et al. (2013) use a more simple model assuming no beam divergence, the simulation described by Hodge (2010) includes both the modelling of beam divergence and recording of the waveform. Their simulation again uses 2.5D elevation models to represent the scene, which is appropriate for their particular objective of error quantification in high- resolution, short-range TLS of natural surfaces, specifically fluvial sediment deposits, of small scenes (area of $1\text{\times}1\text{\,}\mathrm{m}$). Established Monte-Carlo ray tracing simulator are used and being extended for airborne and satellite laser scanning simulations. The librat model (Calders et al., 2013; Disney et al., 2009, 2010), a modular development of ARARAT (Lewis and Muller, 1993) is such a Monte-Carlo simulator. Similarly, North et al. (2010) extend the 3D radiative transfer model FLIGHT (North, 1996) to model satellite LiDAR waveforms. Monte-Carlo methods represent a simple, robust and versatile set of techniques to solve multi-dimensional problems by repeatedly sampling from a probability density function describing the system that is investigated (Disney et al., 2000). These stochastic methods are useful for simulating multi-scattering processes, e.g. for modeling canopy reflectance. The main drawback of Monte-Carlo ray tracing methods are high computation times to simulate sufficient photons for the scattering model to converge to an accurate solution (Disney et al., 2000; Gastellu-Etchegorry et al., 2016). The LiDAR extension of the Discrete Anisotropic Radiative Transfer (DART) model attempts to alleviate these restrictions by quickly selecting scattering directions of simulated photons using the so-called Box method and modeling their propagation and interaction using a Ray Carlo method, which combines classical Monte-Carlo and ray tracing methods (Gastellu-Etchegorry et al., 2016). A comprehensive review of simulators for the generation of point cloud data, including LiDAR simulators, is presented in Schlager et al. (2020) with focus on their applicability to generate data in the context of driver assistance systems and autonomous driving vehicles. In this context, they analyse algorithms with respect to their fidelity, operating principles, considered effects and possible improvements. In contrast to most of the previously mentioned approaches, HELIOS++ provides a framework for full 3D laser scanning simulation with multiple platforms (terrestrial (TLS), mobile (MLS), UAV-borne (ULS) and airborne (ALS)), and a flexible system to represent scenes, which allows combination of input data from multiple sources and data formats (Figure 1). The simulation of beam divergence and full waveform recording are supported. While HELIOS++ may not be as realistic in terms of physical accuracy regarding the energy budget of a single laser shot as, e.g., DART, it provides a sensible trade-off between computational efforts and resulting point cloud quality. Users can simulate VLS over a large range of scales, and even combine different scales in one scene. For example, a highly detailed tree model with individual leaves might be placed in a forest scene represented by (transmissive) voxels (Weiser et al., 2021), while using a rasterised digital terrain model as ground surface. This allows to model the influence of the surrounding of a particular object of interest on the derived VLS point clouds. Furthermore, HELIOS++ aims for high usability by providing a comprehensible set of parameters, while not overwhelming the users with options. These parameters represent the state of the art and are supported by peer-reviewed literature. Since HELIOS++ can be used from the command line and from within Python, workflows integrating HELIOS++ can be easily scripted and automated, as well as linked to external software (e.g. GIS, 3D point cloud processing software, Jupyter Notebooks, and others). HELIOS++ comes with an extensive documentation of all algorithms that are used in the simulations, including examples and references to the relevant literature describing the implemented methods. Furthermore, the open source implementation allows any user to i) inspect and ii) alter/adapt the source code of the program. For ease of use, we provide pre-compiled versions that are ready to use for major operating systems (Microsoft Windows 10 and Debian Linux 10.7 Buster), and the option to use Python as a scripting language to create, manipulate and simulate VLS data acquisition with HELIOS++. ## 3 Implementation of HELIOS++ This section introduces the concepts and interfaces of HELIOS++, for which the important components are the overall architecture and modules (Section 3.1), platforms (Section 3.2), scanners and laser beam deflectors (Section 3.3), the waveform simulation (Section 3.4), input formats (Section 3.5) and output formats (Section 3.6), and the aspect of randomness and repeatability of results (Section 3.7). ### 3.1 Architecture and modules The central element in HELIOS++ simulations is a _survey_. A survey contains links to the _scene_ , which defines the objects that are scanned, the _platform_ , on which the virtual scanner is mounted and moved through the scene, and to the _scanner_ itself. Furthermore, a survey contains a number of _legs_ , which represent waypoints for the platform. The scene consists of a number of _parts_. Each part represents one input source, for example, a 3D mesh file, or a voxel file. Multiple parts may be combined in a scene, and no limitation to the combination of different data source types is imposed. HELIOS++ uses internationally accepted standard file formats for input and output. These elements are defined through Extensible Markup Language (XML) files, that are referenced using (relative) file paths. XML is a text-based format, which can easily be manipulated using a text editor or an XML editor. Figure 2 presents the different files along with a subset of the parameters that can be set in the respective files. Figure 2: File structure of HELIOS++ survey, scene, platform, and scanner. A survey consists of one or more legs, and a single scanner, platform, and scene, respectively. A scene is built up from one or more parts, which can be of different data type. A survey may then be run through either i) the command line, where an executable is provided or ii) the Python bindings `pyhelios`. The Python bindings allow access to the simulation parameters as parsed from the XML files, including changing the parameters programmatically for each simulation run. For example, different scanners can be exchanged automatically and simulated in sequence in a single Python script without changing any other input and settings of the simulation. `pyhelios` furthermore allows access to the simulation result, i.e. the point cloud and the platform trajectory, and converts them into a NumPy array either at the end of the simulation run or through a callback function that can be executed every $n$-th cast laser ray. In this way, a live preview of the point cloud acquisition can be implemented in Python to give a visual impression of the ongoing simulation. A Python script distributed with HELIOS++ acts as such a visualiser, and is called with the same commands as the standalone executable. A user may thus quickly switch between using the pure C++ implementation without visualisation or the Python bindings with visualisation, as presented in Listing 1. Listing 1: Comparison between running HELIOS++ as an executable and through the Python wrapper providing an interactive viewer ⬇ 1run\helios.exe data\surveys\arbaro_demo.xml \--lasOutput \--writeWaveform 2python pyhelios\helios.py data\surveys\arbaro_demo.xml \--lasOutput \--writeWaveform With `pyhelios`, it is also possible to combine HELIOS++ with tools like _Jupyter Notebooks_ , allowing for explanations along-side code and figures. We include sample notebooks in the documentation of HELIOS++. One of these samples is shown in Figure 3. Figure 3: Screenshot of a Jupyter Notebook showcasing the Python bindings of HELIOS++ by plotting the trajectory of a simulation over flat terrain. ### 3.2 Supported laser scanning platforms Both static and dynamic platforms are supported by HELIOS++, which can resemble an airplane (`LinearPath` platform), a multicopter (`Multicopter` platform), a ground-based vehicle (`GroundVehicle` platform) or a static tripod (`Static` platform). In the case of the `LinearPath` platform, the vehicle is moved with a constant speed from one waypoint (leg) to the next one. The orientation of the platform is always towards the next waypoint. The `Multicopter` platform additionally simulates acceleration and deceleration of the platform. In the turn mode _smooth_ , the platform banks to make more smooth turns at the waypoints instead of stopping and turning on the spot. This mode simulates the _banked angle turns_ available in the flight protocols of major drone companies (e.g. DJI). Custom yaw angles for the beginning and the end of each leg can be provided. The `GroundVehicle` platform is bound to elements of the scene defined as ground in the respective material file (cf. Section 3.5). Furthermore, it considers maximum turn radii and implements three-point-turns to resemble a car or tractor on which a scanner is mounted (i.e. MLS). The different platforms and scene types (Section 3.5) are shown in Figure 1. As expected, choice of platform influences the resulting pointcloud, which is shown in Figure 4. Figure 4: Point clouds resulting from virtual laser scanning of the scene shown in Figure 1, using (a) an airplane, (b) a multicopter, (c) a ground vehicle and (d) a static tripod as platform. Since HELIOS++ records which objects are generating which return, the points can be perfectly assigned to the objects, here illustrated by distinct colouring. ### 3.3 Supported laser scanners and scan deflectors The core component of the simulation is the laser scanner, which includes a model for the scan deflector. The choice of the deflector influences the resulting scan pattern (Fig. 5). Figure 5: Different scan patterns depending on the deflector used in the simulation: a) rotating mirror, b) fibre-optic line scanner, c) oscillating mirror and d) slanted rotating mirror (Palmer scanner). The patterns shown here result from simulations with HELIOS++ using the respective deflectors, albeit with unrealistic settings of pulse repetition rate and scanning frequency, in order to show the patterns more clearly. HELIOS++ supports polygonal mirror deflection, resulting in parallel scan lines with even point density (Fig. 5a), fibre-optics, where the beam is fed through fibreoptic cables to point in different directions, resulting in a similar scan pattern to (a), but without the need of mechanically moving parts (Fig. 5b), a swinging mirror, which results in a zig-zag-pattern with increased point densities at the extrema (Fig. 5c) and rotating slanted mirrors, resulting in a conical point pattern at a constant scan angle off- nadir, also referred to as Palmer scanner (Fig. 5d). ### 3.4 Waveform simulation During real LiDAR acquisitions, a laser beam sent out from the laser scanner has a finite footprint, i.e. a non-zero area that intersects with the scene. For every infinitesimal point in this area, energy is transmitted back to the detector where it is recorded as the integral of intensities over the area. By considering this received intensity over time, it is possible to extract multiple echoes from one laser pulse, corresponding to multiple targets that were hit by parts of the intersection area, respectively. In HELIOS++, the non-zero beam divergence is simulated by subrays, that are sampled in a regular pattern around the central ray (Fig. 6). Every subray has its own base intensity, which is calculated according to Equation 1, representing a 2D Gaussian power distribution (Carlsson et al., 2001). $I=I_{0}\exp\left(-2r^{2}/w^{2}\right)$ (1) where $I_{0}$ [$\mathrm{W}$] is the peak power, $w$ [$\mathrm{m}$] the local beam divergence and $r$ [$\mathrm{m}$] the radial distance from the power maximum, i.e. the ray centre. $w$ is calculated using Equation 2 from the beam waist radius $w_{0}$ [$\mathrm{m}$], and the helper values $\omega=\frac{\lambda R}{\pi w_{0}^{2}}$ and $\omega_{0}=\frac{\lambda R_{0}}{\pi w_{0}^{2}}$ with $\lambda$ [$\mathrm{nm}$] as the wavelength, $R$ [$\mathrm{m}$] the range of the target and $R_{0}$ [$\mathrm{m}$] the focusing length of the laser. $w=w_{0}\sqrt{\omega_{0}^{2}+\omega^{2}}$ (2) Every subray is individually cast into the scene and intersected with objects. If it hits an object, a return is generated and recorded by the detector. The respective subray does not continue in the scene through transmission or reflection. In the case of the transmissive voxel model (cf. Section 3.5), a subray may either fully traverse a voxel or produce a return, but never both. Therefore, the property of transmissivity of a scene part is coupled to the use of multiple subrays. The power returned from the object is further dependent on the material specified in the scene definition (Section 3.5). The number of subrays generated can be set by the user by providing the `beamSampleQuality` parameter in the XML file of the scanner or the survey. The `beamSampleQuality` corresponds to the number of concentric circles where subrays are sampled. For each circle, the number of subrays is defined as $\lfloor 2\pi i\rfloor$ where $i=1,\dots,$`beamSampleQuality` is the circle index. This ensures that the angular distance between adjacent subrays is approximately constant, i.e. each subray represents a solid angle of equal size. A central subray is always added. Figure 6 shows the subray distribution for three different beam sample qualities. Figure 6: Subray configurations for beam sample qualities of (a) 2 (7 subrays), (b) 5 (93 subrays), and (c) 9 (279 subrays). The colour of the subrays corresponds to the normalised intensity at this location within the beam cone: low (purple) to high (yellow). The black circle represents the single beam divergence (at the $1/e^{2}$ points, here: $0.3\text{\,}\mathrm{mrad}$), the gray circle twice the beam divergence. The pulse shape in time is approximated using bins of a regular, user-defined size via the parameter `binWidth_ns`. Each bin’s power is calculated according to Equation 3, where $t$ [$\mathrm{ns}$] is the time and $\tau$ [$\mathrm{ns}$] is the pulse length of the scanner divided by 1.75 (Carlsson et al., 2001). $I$ [$\mathrm{W}$] is calculated for each subray according to Equation 1. $P(t)=I\left(\frac{t}{\tau}\right)^{2}\exp\left(-\frac{t}{\tau}\right)$ (3) For every pulse, the subrays are collected as a representation of the full returned waveform, and the recorded power for each bin is taken as the sum of the subray’s waveforms, shifted according to the different ranges of the subrays. Range difference is converted to time difference by using the speed of light as a constant ($c=\;$$299,792,458\text{\,}\mathrm{m}\text{\,}{\mathrm{s}}^{-1}$). A local maximum filter is then used on the summed waveform to detect peaks, which are regarded as echoes and exported as points. Optionally, a Gaussian may be fitted to the resulting waveform, to get a measure of echo width (i.e. standard deviation of the Gaussian). The position of the point along the range (time) axis, however, is taken from the local maximum, not from the fitted Gaussian, as the outgoing waveform (Eq. 3) is not Gaussian. ### 3.5 Input scene models HELIOS++ supports data formats that are (de-facto) standards in the field, and can be created and manipulated with free software such as QGIS333https://qgis.org/en/site/, CloudCompare444https://www.danielgm.net/cc/, Blender555https://www.blender.org/ or AmapVOX666Vincent et al. (2017), http://amap-dev.cirad.fr/projects/amapvox. #### 3.5.1 Wavefront objects In wavefront object files (file extension .obj), meshes are represented as lists of triangles (triples of point IDs) and points (triples of coordinates). This allows opaque 3D models of arbitrary complexity. In addition, material properties can be assigned to the objects in so-called material files (file extension .mtl; Wavefront Technologies, 1992). #### 3.5.2 GeoTIFF models Raster models created with common GIS tools, such as a digital elevation or a digital surface model, can be included. The raster is converted to a triangular mesh on import, where the pixel centres are interpreted as points. Invalid pixels (i.e. no-data values) are ignored in the triangulation, resulting in holes in the mesh. This data type allows an simple representation of the Earth’s surface. As prescribed by the file format, only 2.5D-data is supported. #### 3.5.3 Point clouds In ASCII xyz files, point clouds can be used as a raw input to HELIOS++. On import, the point clouds are voxelised using a voxel size defined by the user. In addition, a normal vector used for intensity calculation can be assigned to the voxels (by nearest neighbour or mean reducing) or calculated from the point cloud on the spot (cf. Section 5). Alternatively, all laser rays can be set to have an incidence angle of 0°, or the rays are intersected with the actual faces of the cube representing the voxel. Material properties can be defined in the respective scene part definition of the XML file. #### 3.5.4 Transmissive voxels Especially for modelling vegetation, we include support for voxel data containing plant area density information (e.g. created using the AmapVOX software (Vincent et al., 2017), file extension .vox). Multiple modes are supported in this case: * 1. opaque voxels, where the voxels are represented by solid cubes with a fixed side length equal to voxel resolution, * 2. adaptive scaling of opaque voxel cubes (with optional random but reproducible shift to avoid regular patterns in the result), where the scaling (side length of a voxel $a$) is dependent on the plant area density $PAD$, the user-defined scaling factor $\alpha$, the base voxel size $a_{0}$ and the maximum $PAD$, related by Equation 4, and * 3. transmissive voxels (explained in more detail in the following). Point clouds resulting from a tree simulation using these different modes are shown in Figure 7. The opaque voxel models also support the definition of material properties through the scene part’s definition. A comprehensive study on the effects of different levels of detail in modelling forests using the first two modes on the extraction of forestry and point cloud parameters is performed by Weiser et al. (2021). $a=a_{0}\left(\frac{PAD}{PAD_{max}}\right)^{\alpha}$ (4) Here, $a$ is the side length of the resulting cube, $a_{0}$ the base voxel size, $PAD$ the plant area density of each individual voxel, $PAD_{max}$ a maximum value for the plant area density and $\alpha$ a user-defined scaling factor. Figure 7: VLS point cloud of a tree based on different voxel modes, using a UAV as a platform. Point clouds are simulated using input voxel models with different fixed sizes (a-c), PAD-dependent scaled voxels (d and e) and transmissive voxels (f). The transmissive voxels use an extinction approach as presented by North (1996). The extinction coefficient $\sigma$ is calculated following Equation 5: $\sigma=\frac{\mu_{L}}{2\pi}\int_{0}^{2\pi}g_{L}\left|\Omega^{\prime}\cdot\Omega_{L}\right|d\Omega_{L}$ (5) Here, $\mu_{L}$ is the plant area density (PAD) as defined in the .vox file, $g_{L}$ is the probability taken from the leaf angle distribution at direction L, and $\left|\Omega^{\prime}\cdot\Omega_{L}\right|$ is the cosine of the angle between the incident ray $\Omega^{\prime}$ and the direction $\Omega_{L}$. This leaf angle distribution is supplied via a look-up-table as shown in Table 2, and examples for planophile, erectophile, plagiophile, extremophile, spherical and uniform distributions are provided. Subsequently, a random number $R$ is drawn from a uniform distribution $\in[0,1)$. The expectation of the intersection of a subray with a voxel given the extinction coefficient $\sigma$ is then given as in Equation 6, where $s$ is the distance of an echo after entering the voxel (i.e. traversed distance). $s=\frac{-\ln(R)}{\sigma}$ (6) If the actual length of the subray traversing the voxel is larger than $s$, no echo is recorded for this ray in the voxel, and it is assumed to have transmitted through the voxel, potentially intersecting the next voxel or another object. Otherwise, an echo is created at distance $s$ from where the subray first entered the respective voxel, and the subray is not continued through the scene (North, 1996). Irrespective of the input type, each scene part can be individually scaled, rotated and translated within the scene, allowing the re-use of models to create multiple instances of similar objects. These transformations can also be manipulated using the Python bindings. For rotations, both extrinsic (”global”, fixed coordinate axes) and intrinsic (”local”, rotating coordinate axes) modes are supported. HELIOS++ allows arbitrary combinations of all possible object models in a scene. For example, a GeoTIFF can be used as digital terrain model on which a vehicle is moving, combined with transmissive voxel models of trees, mesh models for tree stems and a point cloud of buildings, which is voxelised by HELIOS++ (Figure 1). Such combination allows highly versatile use of the scene definition, while considering shadowing and overlapping effects between the different types of object models, and is one of the key advantages over other VLS software. ### 3.6 Output format and options The simulated point clouds can be written either as LAS file (Version 1.0, point format 1), according to the format definition by the American Society of Photogrammetry and Remote Sensing (ASPRS, 2011), as LAZ file, which is a lossless compression of LAS provided by the LASzip library777https://laszip.org/, Isenburg (2013), or as ASCII file, where the point coordinates and attributes are written in columns. The LAS/LAZ formats allow for smaller file sizes and faster read/write access (cf. Section 5.4), while the ASCII format can be parsed by any program working with text files. In addition to creating compressed LAS files, HELIOS++ can also compress output ASCII files, if smaller output sizes are required. Uncompressing LAZ or compressed ASCII files is possible by invoking HELIOS++ using the `--unzip` flag. In the simulation of full waveforms (see Section 3.4), the echo width is optionally estimated for every recorded return. Since this estimation uses a non-linear least squares method and is hence expensive to calculate, the estimation has to be switched on explicitly by the user. In typical use cases, this increases the runtime of the simulation by a factor of 1.3. If the full waveform is to be exported, a separate ASCII file will be written. This file contains information about every beam that generated at least one hit consisting in the beam origin and the beam direction, as well as the sampled returned waveform. The bin size as well as the maximum length of this sampled waveform can be defined by the user (Section 3.4). To connect the point cloud output with the full waveform output, the point cloud has an attribute `fullwaveIndex`, which corresponds to the `fullwaveIndex` in the ASCII waveform output. Multiple points, if coming from the same beam, may have the same `fullwaveIndex`. The output created by HELIOS++ can be easily converted to standard waveform formats such as PulseWaves and LAS 1.4 WDP. Furthermore, HELIOS++ comes with an option to export the trajectory of the platform. The time interval between successive platform positions can be defined by the user in the survey XML file. In addition to the time and the position of the platform, the attitude angles of roll, pitch and yaw are exported. ### 3.7 Randomness and repeatability The trajectory of the platform, the scanner definitions and the ray-scene interactions with the exception of transmissive voxels are deterministic. To allow for more realistic point clouds, random noise sources may be introduced at various points of the simulation. A single distance measurement is attributed with a random ranging error, drawn from a normal distribution with parameters defined for each scanner. Additionally, random platform noise can be added to simulate trajectory estimation or scan position localisation errors. By default, the system time is used to generate randomness, which results in different outcomes for every simulation run. However, the ability to produce the identical results in repeated simulation runs is a major advantage of VLS over real data acquisitions and may be aspired for certain use cases. To allow for repeatable survey results while using randomness, there are additional options to define a custom seed for pseudo-randomness generation. Still, when using multithreading, each thread accesses the randomness generator in a non-deterministic order, resulting in different outputs for every simulation run. This can be avoided by running HELIOS++ in single-threaded mode and with a fixed seed. This will generate the exact same result in repeated runs, even for transmissive voxels, as the random number drawn from the uniform distribution is also created based on the seed. ## 4 Applications of laser scanning simulation To investigate the usability of the HELIOS++ concept of providing a sensible trade-off between physical reality and computational complexity, along with the support of generic input files to create the 3D scene, we present applications of laser scanning simulation. This literature survey is based on studies that use HELIOS++’s predecessor, HELIOS (Bechtold and Höfle, 2016) or pre-release versions of HELIOS++. Since the functionality of HELIOS, apart from the visualisation module, is fully integrated in HELIOS++, the publications presented in the following provide an adequate review of simulation applications. The analysis is conducted on all publications that cite the original publication of Bechtold and Höfle (2016), and that actively use HELIOS in their research according to the published article. It consists of journal papers, conference contributions and posters. The following questions are posed to analyse the contents of each publication. * 1. What is the scientific target of the simulation? * 2. How many simulations are carried out on which platform (ALS, TLS, ULS, MLS)? * 3. Which type of model is employed to compose the scene for the simulation? * 4. Which parameters are extracted from the point cloud, if applicable? * 5. What is the resulting VLS point cloud compared to? The synthesis of the literature review allows a grouping of the publications and of the purposes of laser scanning simulation into four main categories: (1) optimising or analysing different scan settings and acquisition modes, (2) comparing parameters extracted from simulated data to error-free ground truth, (3) generating training data for supervised machine learning, and (4) testing and development of novel or future algorithms and sensors. ### 4.1 Data acquisition planning and scan setting effects Data acquisition planning at its core is an optimisation problem. The goal is to acquire data that is fit for the purpose of the analysis by minimal effort. This can be a minimum number of scan positions, flight lines, etc., that is sufficient for the specific requirement to the data, such as coverage of the scene regarding occlusion or a certain point density. While flight planning tools and visibility analysis-based methods can be used to create potential acquisition plans, they usually lack methods to verify the fitness for use of obtained data. With virtual laser scanning, assuming that at least a coarse model of the area of interest is available, a 3D point cloud can be generated and tested for its fitness directly in the planned application. There is no need to define proxy metrics like target point density, required accuracy and overlap as required by simple survey planning tools. This way, users can ensure that the acquired data will meet all requirements such as coverage, adequate representation of geometry, and resolution before actually going to the field to collect the data, by running their analyses on the simulated point clouds and interpreting the results. Similarly, the effects of different scan settings on parameters extracted from the simulated point clouds can be studied with HELIOS++, as single variables can easily be manipulated in a way that is isolated from all other influences, including environmental influences, which are very difficult to control in repeated real-world acquisitions. For example, the effects of flying height and maximum scan angle on the resulting ground point density and resolution (i.e. illuminated area per beam) may be analysed. Similarly, TLS scans of different resolution can be simulated and the effect of resolution on the results of data analysis (e.g. the quality of extraction of tree stems) can be quantified. While the quality of the simulation in terms of being physically realistic highly depends on the input models, even a coarse model can be useful to estimate occlusion and resulting point densities. Such analyses, carried out prior to real data acquisitions, can save valuable time in the field. Using a Monte-Carlo approach, multiple simulations using different parameters are carried out to create the optimal (in terms of number of positions, time, etc.) acquisition plan. Existing publications in this category include Backes et al. (2020), who use a digital surface model created from photogrammetry to simulate acquisition of an alpine valley by TLS and ULS. They estimate the minimal detectable change, to optimise scan positions and trajectories for change analysis. Similar analyses, but with focus on resulting point density and completeness of data acquisition, are carried out based on a DEM provided by public agencies by Lin and Wang (2019). They are able to show that a downsampled scene model can provide accurate measures for simulated point densities when also reducing the pulse frequency. This validates the simulation’s representation of spatial scales. A validation of scan position planning based on viewshed analysis is carried out with respect to achieved accuracy, point density and completeness by Previtali et al. (2019). Their use cases are acquisitions of complex archaeological sites, where complete coverage is often difficult to achieve due to occlusions. In two examples, they optimise the positions of 97 and 16 scan positions, respectively, to scan a basilica and part of an ancient food storage in Italy. The effect of scan settings on point cloud-based parameters is analysed for forestry settings by Hämmerle et al. (2017), who extract understory tree heights from TLS and ULS data. They use 3D models created with the Arbaro tree generator888http://arbaro.sourceforge.net/ (Weber and Penn, 1995), on which they densely sample points to create a reference point cloud as ground truth. They find a favourable trade-off between acquisition effort and accuracy of results for a number of three TLS scan positions around a tree object of interest (Hämmerle et al., 2017). Similarly, Li et al. (2020) create 3D tree models using ONYXTREE999http://www.onyxtree.com/ and evaluate the influence of scan parameters on extracted values of diameter at breast height (DBH), tree height, stem curve and crown volume. They succeed in reducing the root-mean- squared error on these values by iteratively adapting scan parameters. Weiser et al. (2021) analyse the different opaque voxel models at different scales and their effect on common tree metrics, showing that a scaled voxel model requires much less complexity (i.e., allows for a larger voxel size) than binary voxels at a fixed scale. Schäfer et al. (2019) use the novel transmissive voxel-support of HELIOS++ for similar analyses with ALS and ULS data. ### 4.2 Algorithm and method evaluation: validation and calibration In numerous point cloud analysis methods, the objective is to extract certain parameters describing the objects of interest. For example, in forestry applications, ALS and TLS point clouds are commonly used to derive the diameter at breast height of trees, crown radii, tree heights, and tree species (Giannetti et al., 2018). To calibrate the extraction algorithms as well as to validate the results, in-situ measurements are required, which are laborious and costly to acquire, and not free of error. An alternative to this in-situ data acquisition can be provided by HELIOS++, if the parameters to be extracted from the point cloud can be derived from the objects within the scene, or the scene is created dynamically according to the parameters. For example, a method may be designed to measure DBH from a point cloud. In VLS, the true values of DBH can be derived from the stem models, or the virtual tree models themselves may be generated according to given values of DBH. In addition to being error-free, the domain of the parameters (e.g. the range of DBH values) used in the simulation can be defined by the user. A simulated example can therefore be picked to have exactly the properties needed in a specific application (say, use case-specific DBH values between $20\text{\,}\mathrm{cm}25\text{\,}\mathrm{cm}$), whereas real objects with the required properties as ground truth may be hard to find. In addition to parameter extraction, perfect ground truth can be used to evaluate the performance of classification algorithms. Currently, classification methods are difficult to compare, as authors use different evaluation approaches, mainly due to the lack of semantic reference data. Since in VLS the interaction between the laser beam and the scene can be attributed to the exact mesh face that is hit during ray casting, and therefore to a specific object, the ground truth data is free of error even in highly complex and multi-echo scenarios. This has compelling advantages: First, reference data can be created automatically, drastically cutting costs and providing the possibility of generating massive amounts. And second, it removes labelling errors, as manual labelling will always include some degree of human error. Considerable research using HELIOS to validate or calibrate extracted parameters has been undertaken in multiple forestry applications. Liu et al. (2019a) and Liu et al. (2019b) estimated leaf angles on trees, where ground truth is practically impossible to obtain, because in real settings, leaves are continuously moving due to wind. VLS allows to validate their approach of leaf angle estimation and subsequently apply it to real data. Wang et al. (2020) validate their calculation of photon recollision probability over spatial locations using VLS data. Zhu et al. (2020) use two different methods to assess Leaf Area Index (LAI) and compare these methods with ground truth obtained from the object models, allowing them a comparison with the true LAI of the input objects. For tree segmentation, ground truth is also difficult to obtain, as tree canopies often intersect each other. By using HELIOS, Wang (2020) and Xiao et al. (2019) obtain perfect ground truth for training and validation of their segmentation methods. Wang (2020) achieve 2.9 % and 19.8 % RMSE for tree height and crown diameter estimates. In a non-forestry application, road curves are reconstructed and the reconstruction is compared with the model parameters (Zhang et al., 2019), achieving a relative accuracy of 0.6 % in circle radii estimation using VLS data. Requirements for Building Information Modelling (BIM) are evaluated by (Rebolj et al., 2017), who generate around 100 point clouds to ascertain the influence of parameter values. They define accuracy criteria for the successful identification of building elements in the scans. Bechtold et al. (2016) test a segmentation tool for rock outcrops by using HELIOS as a simulator. They show the value of simulated test data for method development by easily generating point clouds with different occlusions and scan settings, resulting in a multitude of point densities and point patterns. ### 4.3 Method training With the advent of neural networks as supervised machine learning method in geospatial domains, the need for training data has grown almost indefinitely. While raster-based approaches can make use of existing pre-trained networks by domain transfer (Pires de Lima and Marfurt, 2019), no such networks exist for point-based deep learning such as PointNet/PointNet++ (Qi et al., 2017). Simulated data, though only replicating parts of reality, can be used by neural networks to learn basic descriptors which describe point cloud neighbourhoods, e.g. planes, corners or edges. From these descriptors, higher- level features are derived, which are subsequently used in classification or regression (Winiwarter et al., 2019). Once a network has learned to represent data in form of these features, it can be adjusted to real data by adding relatively small amounts of training data, in approaches shown to work for the image domain (Danielczuk et al., 2019). Further research is needed on this approach especially for point cloud data, but HELIOS++ allows easy and fast generation of labelled training data, which does not suffer from ground truth errors. As an example application of this purpose, Martínez Sánchez et al. (2019) use HELIOS-simulated data to train and evaluate a semantic classification of an urban scene created from OpenStreetMap101010https://www.openstreetmap.org/ models. The use of VLS allowed a quantification of their classifier’s total error, which amounts to 0.5 % on simulated point clouds. ### 4.4 Sensing experimentation The fourth category summarises publications concerned with the development of novel sensors and methods, such as Park et al. (2020), who present a new Time- of-Flight sensor and compare its results to a HELIOS simulation. In general, the parameters of the virtual sensors can be tuned to resemble a non-existent sensor, the performance of which can then be simulated without the need of actually building a prototype. Especially when looking for potential improvements of current sensors, this enables the identification of weak links or bottlenecks for certain use cases. More in-depth experimentation is also conceivable, where e.g. a novel deflector model shall be simulated. Due to the open-source license, a developer may take the HELIOS++ framework and would only have to implement a new deflection method, whereas the scene and all other components can be used as is to run a simulation, testing the usability of the novel deflector model. Especially when considering the short lifecycle of current hardware, simulation may be the only way to ensure fitness for use of the sensor. Equivalent simulators are widely used in remote sensing, as especially tools working with data acquired by satellites can be developed using the simulated data, and are then ready to use as soon as the first real data is delivered. ## 5 Computational considerations Since modern laser scanners can measure millions of points per second, it is crucial to have an efficient implementation for the ray-scene intersection. In this section, we first present theoretical considerations on the ray tracing implemented in HELIOS++, the modelling of vegetation and options of generating large and complex scenes. In Section 5.4, a comparison between HELIOS++ and its predecessor HELIOS is carried out. ### 5.1 Ray tracing implementation using a kD-Tree A scene $S$ in HELIOS++ context can be mathematically described as a set of primitives, so $S=\\{P_{1},\ldots,P_{n}\\}$. For each primitive $P$ its boundaries are defined considering an axis-aligned bounding box, its centroid, and the set of vertices $V=\\{v_{1},\ldots,v_{m}\\}$ composing the primitive. Each primitive supports rotation, scaling and translation together with a material specification defining its reflectance and specularity. Certain primitives, such as transmissive voxels, also support a look-up table which can be used for vegetation modelling, as explained in Section 5.2. The scene building process consists in generating the set of primitives composing the scene. Multiple input sources are supported (cf. Section 3.5), so it is possible to build a scene considering Wavefront Object files, point clouds as ASCII files specifying $(X,Y,Z)$ coordinates for each point, GeoTIFF files, or a custom voxel file format based on AMAPVox (Vincent et al., 2017). When building a scene, different sources can be considered, as each one is associated to its own scene part. It is also possible to apply aforementioned affine transformations to an entire scene part. This can be used, for instance, to load the same object multiple times in one scene and placing it in different locations through translations. Saving and loading already built scenes is relying on boost serialisation technology (Ramey, 2004). Ray intersections are computed through a recursive search performed over a kD- Tree containing all primitives (Bentley, 1975). Let $O$ be the ray origin and $\hat{v}$ the normalised ray direction vector. When recursively searching through the kD-Tree starting at $O$, $\hat{v}$ is used to consider which node must be visited until a leaf node is reached. For each recursive search operation $s$, coordinates are analysed, so $s\equiv 0\mod 3$ means the $X$ coordinate splits the space, $s\equiv 1\mod 3$ means the Y coordinate splits the space, and $s\equiv 2\mod 3$ means the Z coordinate splits the space. Once inside a leaf node, ray intersections with respect to each primitive are computed. The minimum distance intersection $t_{0}$ is the time the ray needed to enter the primitive while $t_{1}$ is the time the ray needed to leave it. It is possible to have only $t_{0}$ determined, as is the case for triangles, since they are only intersected once per ray. For the special case of primitives which support multiple ray intersections, such as transmissive voxels, the process is repeated considering consecutive origins. Suppose we have a transmissive voxel intersected by a ray $\\{O_{1},\hat{v}\\}$: If this voxel lets the ray pass through it, then the next ray intersection will be found considering the ray $\\{O_{2},\hat{v}\\}$, with $O_{2}=O_{1}+(t_{1}+\varepsilon)\hat{v}$, where $\varepsilon$ is a small decimal number to assure getting out of the previously intersected primitive. ### 5.2 Vegetation modelling: Transmissive voxels Beam horizontal component | Beam vertical component | Hit probability $g_{L}$ ---|---|--- 1.000000 | 0.000000 | 0.424413 0.999683 | 0.025180 | 0.424682 0.998732 | 0.050345 | 0.425489 $\vdots$ | $\vdots$ | $\vdots$ 0.009444 | 0.999955 | 0.848789 0.000000 | 1.000000 | 0.848822 Table 2: Values from a look-up table (LUT) for the hit probability $g_{L}$ at a given beam direction $L$, represented by horizontal and vertical component. The numbers here correspond to an erectophile distribution and are obtained using a numerical integration method on the formulae from North (1996). The probability $g_{L}$ is normalised over $2\pi$. One of the new functionalities of HELIOS++ is vegetation modelling through transmissive voxel primitives. For this purpose, a look-up table for leaf angle distribution is used to compute ray intensity with respect to a $\sigma$ (cross-section) value obtained from it. The values in this look-up table represent a hit probability $g_{L}$ given the horizontal and vertical component of the beam vector (Tab. 2). The intensity $I$ is calculated according to Equation 7 where $P$ is the emitted power from Equation 3 (Carlsson et al., 2001), $d$ is the distance between ray origin and intersection point, $\alpha^{2}$ is the square of the scanner receiver diameter, $\beta^{2}$ is the square of the scanner beam divergence and $\lambda$ is the product between atmospheric factor and scanner efficiency. The $\sigma$ value can be seen as function of ray direction vector $\sigma(\hat{v})$, so it will have a different value depending on ray incidence. $I\propto\lambda\sigma\frac{P\alpha^{2}}{4{\pi}d^{4}\beta^{2}}$ (7) Voxels can operate in transmissive mode. This implies the voxel will let the ray continue and not return a signal after intersection if $\sigma=0$. If $\sigma>0$, then the return of the ray is randomly determined by sampling a value $u$ from a uniform distribution $\in[0,1)$ and computing $s=\frac{-\log(u)}{\sigma}$. If the value of $s$ is greater than the distance between both intersection points at the voxel, the ray will continue. Otherwise, the ray will stop at the voxel. For the non-transmissive mode, only voxels with transmittance $1.0$ will allow rays to continue. Thus, by using detailed voxels operating in transmissive mode together with an appropriate look-up table specification, HELIOS++ is capable of simulating laser scanning of vegetation from a precomputed leaf angle distribution following North (1996). To achieve high-quality output, it is recommended to apply individual leaf angle distributions for each vegetation type within a scene. ### 5.3 Strategy to handle large and complex scenes When loading point clouds from ASCII xyz files, big files might not be entirely containable in memory. For this purpose, a two-stage algorithm is used to digest point clouds of arbitrary size. The first stage simply finds the minimum and maximum values for each coordinate and counts the total number of points. The second stage builds all necessary voxels to represent the point cloud inside HELIOS++. For this purpose, a voxel grid is allocated. Then, voxels which have points inside them are built as HELIOS++ primitives. For each voxel its spatial coordinates $(X,Y,Z)$, normal vector components $(N_{x},N_{y},N_{z})$ and colour components $(R,G,B)$ are considered, if available, as may be the case for photogrammetric point clouds. The colour for the voxel is computed as the average of each colour component for all points inside the voxel. We approximate the sRGB color space by averaging the squares of the red, green, and blue values, and taking the square root of this average as the resulting value for the voxel. If normals are provided, the voxel normal can be determined either as the normal of the point which is closest to the voxel centre or the average of each normal component for all points. In case no normal vectors are provided, HELIOS++ can estimate them using singular value decomposition (SVD, Golub and Kahan (1965)). All points inside a voxel are considered to build a matrix of coordinates, for which singular values and singular vectors are obtained. Then, the singular vector of the smallest singular value is the orthonormal vector defining the plane which best fits the point set in terms of the smallest sum of squared orthogonal residuals. It can hence be understood as the voxel normal vector. The normal estimation method can be applied to the entire input point cloud in a single stage for small point clouds. If the size of input point cloud is too big with respect to RAM, the normal estimation is performed in batch mode, dividing the workload into smaller parts. Each batch extracts points inside its voxels while ignoring points outside its scope. ### 5.4 Performance comparison As the direct successor of HELIOS (Bechtold and Höfle, 2016), it is interesting to compare the performance of HELIOS++ with the original implementation in Java. However, HELIOS++ was not just a port of the code, but also comes with multiple computational improvements over its predecessor. One of these improvements concerns the ability for the user to better leverage accuracy at the cost of runtime, or vice versa, by setting parameters accordingly. For some tasks, a rough, thereby faster simulation may suffice, whereas for other tasks very detailed simulation is required, but smaller sample sizes can be used or long processing times are acceptable. Another improvement concerns the binning mechanism for the full-waveform simulation, which is used for maxima detection even if the waveform is not written to an output file. In HELIOS++, we use the parameters `binSize_ns` and `maxFullwaveRange_ns` for this binning, where the bin size is used for both the outgoing pulse and the returned waveform. To limit the impact on performance of very low-incidence rays with a high sampling quality, the user can provide a maximum length of the recorded waveform, beyond which any further echoes are discarded. To compare the performances of HELIOS++ (Version 1.0.0) and HELIOS (Version 2018-09-24), we carry out a number of simulations using different parameter settings, and record the runtime as well as peak memory usage. We present three different scenes, at three different complexity levels, and make use of the option in HELIOS++ to write different file types and to skip the echo width determination, if not required. The first example scenario represents an ALS survey over terrain, which is created by loading a digital terrain model in GeoTIFF format as described in Section 3.5. Since the GeoTIFF loader in the Java version was faulty and this version is no longer maintained, we created a mesh in Wavefront Object format as an input for the Java version. We simulate two flight lines at an altitude of $1,500\text{\,}\mathrm{m}\,\mathrm{a}\mathrm{s}\mathrm{l}$ over terrain with an extent of $26\text{\times}17.8\text{\,}\mathrm{km}$. As scanner we use a _Leica ALS50-II_ , and a Cessna SR-22 as platform. The scene is shown in Figure 8. Figure 8: Visualisation of example survey for airborne laser scanning (ALS), using a digital terrain model as object and a standard ALS instrument as scanner. The 3”-SRTM model by the USGS is used as input raster. The second scenario represents a mobile laser scan of an urban area, where a car is driving through downtown buildings modelled as prisms, cylinders and pyramids (Figure 9). The trajectory of the car is $208\text{\,}\mathrm{m}$ long and its average speed is $20\text{\,}\mathrm{m}\text{\,}{\mathrm{s}}^{-1}$. The scanner is an oblique- mounted _RIEGL VUX-1UAV_. Figure 9: Visualisation of example survey for mobile laser scanning (MLS), with a car driving between objects of simple geometry. The yellow points show the acquired point cloud of scanned parts. In the third and final scenario, we present a TLS survey of a vegetation scene with a large potential for multi-echoes, as it represents two trees generated with the Arbaro Tree Simulator (Weber and Penn, 1995), scanned using a _RIEGL VZ-400_ from two static positions. This simulation has a large number of geometric primitives. A visual is depicted in Figure 10. Figure 10: Visualisation of example survey for terrestrial laser scanning (TLS), scanning two highly complex tree models created using the Arbaro Tree Simulator. The yellow points show the acquired point cloud of scanned parts. Each simulation scenario was carried out three times on an Intel-i9-7900 @ $3.3\text{\,}\mathrm{GHz}$ with $64\text{\,}\mathrm{GB}$ of RAM, and I/O on an SSD connected via SATA. The average runtime and memory consumption values are given in Table 3. | | HELIOS (Java) --- Version 2018-09-24 | HELIOS++ --- Version 1.0.0 Echo width | ✓ | ✓ | | Waveform output | ✓ | ✓ | ✓ | File format | XYZ | XYZ | XYZ | LAS | Scene 1 --- (ALS, GeoTIFF) | $4,712.1\pm 613.8\text{\,}\mathrm{s}$ --- $6,570\text{\,}\mathrm{MB}$ | $2,773.0\pm 53.8\text{\,}\mathrm{s}$ --- $2,246\text{\,}\mathrm{MB}$ | $2,403.8\pm 18.0\text{\,}\mathrm{s}$ --- $2,246\text{\,}\mathrm{MB}$ | $1,912.1\pm 20.22\text{\,}\mathrm{s}$ --- $2,246\text{\,}\mathrm{MB}$ | Scene 2 --- (MLS, geometric primitives) | $68.0\pm 2.3\text{\,}\mathrm{s}$ --- $560\text{\,}\mathrm{MB}$ | $23.9\pm 0.2\text{\,}\mathrm{s}$ --- $24\text{\,}\mathrm{MB}$ | $23.0\pm 0.3\text{\,}\mathrm{s}$ --- $24\text{\,}\mathrm{MB}$ | $16.1\pm 0.6\text{\,}\mathrm{s}$ --- $24\text{\,}\mathrm{MB}$ | Scene 3 --- (TLS, tree models) | $362.6\pm 6.7\text{\,}\mathrm{s}$ --- $5,360\text{\,}\mathrm{MB}$ | $97.2\pm 1.3\text{\,}\mathrm{s}$ --- $314\text{\,}\mathrm{MB}$ | $74.4\pm 1.0\text{\,}\mathrm{s}$ --- $314\text{\,}\mathrm{MB}$ | $62.4\pm 0.3\text{\,}\mathrm{s}$ --- $314\text{\,}\mathrm{MB}$ Table 3: Performance comparison of HELIOS and HELIOS++ with different options. We use default parameters for waveform modelling (``beamSampleQuality=3— and ``binSize_ns=0.25—; ``numBins=100— and ``numFullwaveBins=200— for HELIOS++ and HELIOS, respectively). Runtimes are average of three runs ($\pm$ standard deviation), memory footprint is the highest value (maximum) during the full run. In the case with largest improvement over HELIOS, runtimes of HELIOS++ are lower by 83 % (TLS) while using only 6 % of the previously required memory. Especially for large scenes, the issue of memory footprint has been a limiting factor for the usability of HELIOS. In the case of the ALS scene, the memory consumption can be reduced by 66 %, from more than $6\text{\,}\mathrm{GB}$ to just above $2\text{\,}\mathrm{GB}$, with a runtime reduction of 51 %. For the MLS scene using the simple geometric shapes, the reduced processing overhead leads to a reduction of 96 % in memory usage and 76 % in runtime. From these results we deduce a significant improvement both in runtime and memory footprint when comparing any configuration of HELIOS++ to the previous Java version. ## 6 Conclusions With HELIOS++, we present an open-source laser scanning simulation framework that enables highly performant virtual laser scanning (VLS). In its C++ implementation and with the possibility to use the `pyhelios` package in Python to manipulate simulation parameters, we opt for an efficient and easy- to-use software. While physical accuracy and realism may be superseded by complementing simulation software, HELIOS++ provides a flexible solution to balance computational requirements (runtime, memory footprint) and quality of results (physical realism), while being easy to use and configure for users. Different studies using the HELIOS concept demonstrate the usefulness of virtually acquired laser scanning data for analyses in different categories of purpose. The main categories are (a) planning of flight patterns (ULS, ALS) or scan positions (TLS); (b) generation of ground-truth data for validation of algorithms that extract parameters (i.a. for forestry) from point clouds, (c) generation of training data for supervised machine learning, and (d) testing and development of novel sensors and algorithms. The novel object model of HELIOS++, the transmissive voxel, allows a stochastic simulation of penetrable objects, such as vegetation canopy, without the need for a highly detailed 3D mesh model. Generally, HELIOS++ performs up to of $5.8\,\times$ faster than its predecessor HELIOS on common scenes, and allows the user to omit intensive calculations if not required (e.g. the calculation of the echo width for returned pulses). To summarise, HELIOS++ is a versatile, easy-to-use, well documented scientific software for virtual laser scanning simulations that provides a tool to generate VLS data, thereby complementing data obtained by real-world laser scanning. The framework invites users to experiment and develop new ideas, while being able to rely on established algorithms from the literature. #### Acknowledgements The authors wish to acknowledge the contribution of Patrick Herbers (Ruhr- Universität Bochum), for the improvement the triangle-ray intersection in C++, which helped to significantly reduce the runtime of HELIOS++. Figures 1, 4, 5, and 8 show an airplane model CC-BY Emmanuel Beranger. Figures 1 and 4 show a house model by free3d.com user `gerhald3d`, and a drone model by cgtrader.com user `CGaxr`. ## References * ASPRS (2011) ASPRS, 2011. LAS specification. https://www.asprs.org/wp-content/uploads/2010/12/LAS_1_4_r13.pdf. * Backes et al. (2020) Backes, D., Smigaj, M., Schimka, M., Zahs, V., Grznárová, A., Scaioni, M., 2020\. River Morphology Monitoring of a Small-Scale Alpine Riverbed using Drone Photogrammetry and LiDAR. ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences XLIII-B2-2020, 1017–1024. doi:10.5194/isprs-archives-XLIII-B2-2020-1017-2020. * Bechtold et al. (2016) Bechtold, S., Hämmerle, M., Höfle, B., 2016. Simulated full-waveform laser scanning of outcrops for development of point cloud analysis algorithms and survey planning: An application for the HELIOS LiDAR simulation framework, in: Proceedings of the 2nd Virtual Geoscience Conference, Bergen, Norway, pp. 57–58. URL: http://lvisa.geog.uni-heidelberg.de/papers/2016/Bechtold_et_al_2016.pdf. * Bechtold and Höfle (2016) Bechtold, S., Höfle, B., 2016\. HELIOS: A Multi-Purpose LiDAR Simulation Framework for Research, Planning and Training of Laser Scanning Operations with Airborne, Ground-Based Mobile and Stationary Platforms. ISPRS Annals of Photogrammetry, Remote Sensing and Spatial Information Sciences III–3, 161–168. doi:10.5194/isprs-annals-III-3-161-2016. * Bentley (1975) Bentley, J.L., 1975. Multidimensional binary search trees used for associative searching. Communications of the ACM 18, 509–517. doi:https://doi.org/10.1145/361002.361007. * Calders et al. (2013) Calders, K., Lewis, P., Disney, M., Verbesselt, J., Herold, M., 2013. Investigating assumptions of crown archetypes for modelling LiDAR returns. Remote Sensing of Environment 134, 39–49. doi:10.1016/j.rse.2013.02.018. * Carlsson et al. (2001) Carlsson, T., Steinvall, O., Letalick, D., 2001. Signature simulation and signal analysis for 3-D laser radar. * Danielczuk et al. (2019) Danielczuk, M., Matl, M., Gupta, S., Li, A., Lee, A., Mahler, J., Goldberg, K., 2019. Segmenting Unknown 3D Objects from Real Depth Images using Mask R-CNN Trained on Synthetic Data, in: 2019 International Conference on Robotics and Automation (ICRA), pp. 7283–7290. doi:10.1109/ICRA.2019.8793744. * Disney et al. (2000) Disney, M., Lewis, P., North, P., 2000. Monte Carlo ray tracing in optical canopy reflectance modelling. Remote Sensing Reviews 18, 163–196. doi:10.1080/02757250009532389. * Disney et al. (2010) Disney, M.I., Kalogirou, V., Lewis, P., Prieto-Blanco, A., Hancock, S., Pfeifer, M., 2010\. Simulating the impact of discrete-return lidar system and survey characteristics over young conifer and broadleaf forests. Remote Sensing of Environment 114, 1546–1560. doi:10.1016/j.rse.2010.02.009. * Disney et al. (2009) Disney, M.I., Lewis, P.E., Bouvet, M., Prieto-Blanco, A., Hancock, S., 2009. Quantifying Surface Reflectivity for Spaceborne LiDAR via Two Independent Methods. IEEE Transactions on Geoscience and Remote Sensing 47, 3262–3271. doi:10.1109/TGRS.2009.2019268. * Gastellu-Etchegorry et al. (2015) Gastellu-Etchegorry, J.P., Yin, T., Lauret, N., Cajgfinger, T., Gregoire, T., Grau, E., Feret, J.B., Lopes, M., Guilleux, J., Dedieu, G., Malenovský, Z., Cook, B., Morton, D., Rubio, J., Durrieu, S., Cazanave, G., Martin, E., Ristorcelli, T., 2015\. Discrete Anisotropic Radiative Transfer (DART 5) for Modeling Airborne and Satellite Spectroradiometer and LIDAR Acquisitions of Natural and Urban Landscapes. Remote Sensing 7, 1667–1701. doi:10.3390/rs70201667. * Gastellu-Etchegorry et al. (2016) Gastellu-Etchegorry, J.P., Yin, T., Lauret, N., Grau, E., Rubio, J., Cook, B.D., Morton, D.C., Sun, G., 2016\. Simulation of satellite, airborne and terrestrial LiDAR with DART (I): Waveform simulation with quasi-Monte Carlo ray tracing. Remote Sensing of Environment 184, 418–435. doi:10.1016/j.rse.2016.07.010. * Giannetti et al. (2018) Giannetti, F., Puletti, N., Quatrini, V., Travaglini, D., Bottalico, F., Corona, P., Chirici, G., 2018. Integrating terrestrial and airborne laser scanning for the assessment of single-tree attributes in Mediterranean forest stands. European Journal of Remote Sensing 51, 795–807. doi:10.1080/22797254.2018.1482733. * Golub and Kahan (1965) Golub, G., Kahan, W., 1965. Calculating the Singular Values and Pseudo-Inverse of a Matrix. Journal of the Society for Industrial and Applied Mathematics Series B Numerical Analysis 2, 205–224. doi:https://doi.org/10.1137/0702016. * Goodwin et al. (2007) Goodwin, N.R., Coops, N.C., Culvenor, D.S., 2007. Development of a simulation model to predict LiDAR interception in forested environments. Remote Sensing of Environment 111, 481–492. doi:10.1016/j.rse.2007.04.001. * Hodge (2010) Hodge, R.A., 2010. Using simulated Terrestrial Laser Scanning to analyse errors in high-resolution scan data of irregular surfaces. ISPRS Journal of Photogrammetry and Remote Sensing 65, 227–240. doi:10.1016/j.isprsjprs.2010.01.001. * Holmgren et al. (2003) Holmgren, J., Nilsson, M., Olsson, H., 2003. Simulating the effects of LiDAR scanning angle for estimation of mean tree height and canopy closure. Canadian Journal of Remote Sensing 29, 623–632. doi:10.5589/m03-030. * Hämmerle et al. (2017) Hämmerle, M., Lukač, N., Chen, K.C., Koma, Z., Wang, C.K., Anders, K., Höfle, B., 2017. Simulating Various Terrestrial and UAV LiDAR Scanning Configurations for Understory Forest Structure Modelling. ISPRS Annals of Photogrammetry, Remote Sensing and Spatial Information Sciences IV-2/W4, 59–65. doi:10.5194/isprs-annals-IV-2-W4-59-2017. * Isenburg (2013) Isenburg, M., 2013. LASzip: lossless compression of LiDAR data. Photogrammetric Engineering and Remote Sensing 79, 209–217. URL: https://www.cs.unc.edu/~isenburg/lastools/download/laszip.pdf. * Kim et al. (2012) Kim, S., Lee, I., Lee, M., 2012. LiDAR Waveform Simulation over Complex Targets. ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences XXXIX-B7, 517–522. doi:10.5194/isprsarchives-XXXIX-B7-517-2012. * Kim et al. (2009) Kim, S., Min, S., Kim, G., Lee, I., Jun, C., 2009\. Data simulation of an airborne lidar system, in: Turner, M.D., Kamerman, G.W. (Eds.), Laser Radar Technology and Applications XIV, SPIE. p. 73230C. doi:10.1117/12.818545. * Kukko and Hyyppä (2009) Kukko, A., Hyyppä, J., 2009\. Small-footprint Laser Scanning Simulator for System Validation, Error Assessment, and Algorithm Development. Photogrammetric Engineering & Remote Sensing 75, 1177–1189. doi:10.14358/PERS.75.10.1177. * Lewis (1999) Lewis, P., 1999. Three-dimensional plant modelling for remote sensing simulation studies using the Botanical Plant Modelling System. Agronomie 19, 185–210. doi:10.1051/agro:19990302. * Lewis and Muller (1993) Lewis, P., Muller, J.P., 1993\. The advanced radiometric ray tracer: Ararat for plant canopy reflectance simulation. Proc. 29th Conf. Int. Soc. Photogramm. Remote Sens. 29. URL: http://www.isprs.org/proceedings/XXIX/congress/part7/26_XXIX-part7.pdf. * Li et al. (2020) Li, L., Mu, X., Soma, M., Wan, P., Qi, J., Hu, R., Zhang, W., Tong, Y., Yan, G., 2020. An Iterative-Mode Scan Design of Terrestrial Laser Scanning in Forests for Minimizing Occlusion Effects. IEEE Transactions on Geoscience and Remote Sensing , 1–20doi:10.1109/TGRS.2020.3018643. * Pires de Lima and Marfurt (2019) Pires de Lima, R., Marfurt, K., 2019\. Convolutional Neural Network for Remote-Sensing Scene Classification: Transfer Learning Analysis. Remote Sensing 12, 86\. URL: http://dx.doi.org/10.3390/rs12010086, doi:10.3390/rs12010086. * Lin and Wang (2019) Lin, C.H., Wang, C.K., 2019\. Point Density Simulation for ALS Survey, in: Proceedings of the 11th International Conference on Mobile Mapping Technology (MMT2019), pp. 157–160. URL: https://www.geog.uni-heidelberg.de/md/chemgeo/geog/gis/mmt2019-lin_and_wang_compr.pdf. * Liu et al. (2019a) Liu, J., Skidmore, A.K., Wang, T., Zhu, X., Premier, J., Heurich, M., Beudert, B., Jones, S., 2019a. Variation of leaf angle distribution quantified by terrestrial LiDAR in natural European beech forest. ISPRS Journal of Photogrammetry and Remote Sensing 148, 208–220. doi:10.1016/j.isprsjprs.2019.01.005. * Liu et al. (2019b) Liu, J., Wang, T., Skidmore, A.K., Jones, S., Heurich, M., Beudert, B., Premier, J., 2019b. Comparison of terrestrial LiDAR and digital hemispherical photography for estimating leaf angle distribution in European broadleaf beech forests. ISPRS Journal of Photogrammetry and Remote Sensing 158, 76–89. doi:10.1016/j.isprsjprs.2019.09.015. * Lohani and Mishra (2007) Lohani, B., Mishra, R.K., 2007\. Generating LiDAR Data in Laboratory: LiDAR Simulator, in: ISPRS Workshop on Laser Scanning 2007 and SilviLaser 2007, pp. 264–269. URL: https://www.isprs.org/proceedings/XXXVI/3-W52/final_papers/Lohani_2007.pdf. * Lovell et al. (2005) Lovell, J.L., Jupp, D., Newnham, G.J., Coops, N.C., Culvenor, D.S., 2005. Simulation study for finding optimal LiDAR acquisition parameters for forest height retrieval. Forest Ecology and Management 214, 398–412. doi:10.1016/j.foreco.2004.07.077. * Martínez Sánchez et al. (2019) Martínez Sánchez, J., Váquez Alvarez, A., López Vilarino, D., Fernández Rivera, F., Cabaleiro Domínguez, J.C., Fernández Pena, T., 2019. Fast Ground Filtering of Airborne LiDAR Data Based on Iterative Scan-Line Spline Interpolation. Remote Sensing 11, 2256\. doi:10.3390/rs11192256. * Morsdorf et al. (2007) Morsdorf, F., Frey, O., Koetz, B., Meier, E., 2007\. Ray Tracing for Modeling of Small Footprint Airborne Laser Scanning Returns. doi:10.3929/ETHZ-B-000107380. * North (1996) North, P., 1996. Three-dimensional forest light interaction model using a Monte Carlo method. IEEE Transactions on Geoscience and Remote Sensing 34, 946–956. doi:10.1109/36.508411. * North et al. (2010) North, P.R.J., Rosette, J.A.B., Suárez, J.C., Los, S.O., 2010\. A Monte Carlo radiative transfer model of satellite waveform LiDAR. International Journal of Remote Sensing 31, 1343–1358. doi:10.1080/01431160903380664. * Park et al. (2020) Park, M., Baek, Y., Dinare, M., Lee, D., Park, K.H., Ahn, J., Kim, D., Medina, J., Choi, W.J., Kim, S., et al., 2020. Hetero-integration enables fast switching time-of-flight sensors for light detection and ranging. Scientific Reports 10, 2764\. doi:10.1038/s41598-020-59677-x. * Previtali et al. (2019) Previtali, M., Díaz-Vilariño, L., Scaioni, M., Frías Nores, E., 2019\. Evaluation of the Expected Data Quality in Laser Scanning Surveying of Archaeological Sites, in: 4th International Conference on Metrology for Archaeology and Cultural Heritage, Florence, Italy, pp. 19–24. URL: https://re.public.polimi.it/retrieve/handle/11311/1124569/476842/Previtali%20et%20al%202019%20MetroArchaeo.pdf. * Qi et al. (2017) Qi, C.R., Yi, L., Su, H., Guibas, L.J., 2017. PointNet++: Deep hierarchical feature learning on point sets in a metric space, in: Advances in Neural Information Processing Systems, pp. 5099–5108. * Ramey (2004) Ramey, R., 2004. C++ BOOST Serialization. URL: https://www.boost.org/doc/libs/1_72_0/libs/serialization/doc/index.html. * Ranson and Sun (2000) Ranson, K.J., Sun, G., 2000\. Modeling LiDAR Returns from Forest Canopies. IEEE Transactions on Geoscience and Remote Sensing 38, 2617–2626. doi:10.1109/36.885208. * Rebolj et al. (2017) Rebolj, D., Pučko, Z., Babič, N.C., Bizjak, M., Mongus, D., 2017. Point cloud quality requirements for Scan-vs-BIM based automated construction progress monitoring. Automation in Construction 84, 323–334. doi:10.1016/j.autcon.2017.09.021. * Schlager et al. (2020) Schlager, B., Muckenhuber, S., Schmidt, S., Holzer, H., Rott, R., Maier, F.M., Saad, K., Kirchengast, M., Stettinger, G., Watzenig, D., Ruebsam, J., 2020. State-of-the-Art Sensor Models for Virtual Testing of Advanced Driver Assistance Systems/Autonomous Driving Functions. SAE International Journal of Connected and Automated Vehicles 3, 233–261. doi:10.4271/12-03-03-0018. * Schäfer et al. (2019) Schäfer, J., Faßnacht, F., Höfle, B., Weiser, H., 2019\. Das SYSSIFOSS-Projekt: Synthetische 3D-Fernerkundungsdaten für verbesserte Waldinventurmodelle, in: 2\. Symposium zur angewandten Satellitenerdbeoachtung, Cologne, Germany. URL: https://www.dialogplattform-erdbeobachtung.de/downloads/cms/Stele3/EO-Symposium_Jannika.Schaefer.pdf. * Tulldahl and Steinvall (1999) Tulldahl, H.M., Steinvall, K.O., 1999\. Analytical waveform generation from small objects in LiDAR bathymetry. Applied optics 38, 1021–1039. doi:10.1364/ao.38.001021. * Vincent et al. (2017) Vincent, G., Antin, C., Laurans, M., Heurtebize, J., Durrieu, S., Lavalley, C., Dauzat, J., 2017. Mapping plant area index of tropical evergreen forest by airborne laser scanning. A cross-validation study using LAI2200 optical sensor. Remote Sensing of Environment 198, 254 – 266. URL: http://www.sciencedirect.com/science/article/pii/S003442571730233X, doi:https://doi.org/10.1016/j.rse.2017.05.034. * Wang (2020) Wang, D., 2020. Unsupervised semantic and instance segmentation of forest point clouds. ISPRS Journal of Photogrammetry and Remote Sensing 165, 86–97. doi:10.1016/j.isprsjprs.2020.04.020. * Wang et al. (2020) Wang, D., Schraik, D., Hovi, A., Rautiainen, M., 2020\. Direct estimation of photon recollision probability using terrestrial laser scanning. Remote Sensing of Environment 247, 111932. doi:10.1016/j.rse.2020.111932. * Wang et al. (2013) Wang, Y., Xie, D., Yan, G., Zhang, W., Mu, X., 2013\. Analysis on the inversion accuracy of LAI based on simulated point clouds of terrestrial LiDAR of tree by ray tracing algorithm, in: IGARSS 2013-2013 IEEE International Geoscience and Remote Sensing Symposium, IEEE, Piscataway. pp. 532–535. doi:10.1109/IGARSS.2013.6721210. * Wavefront Technologies (1992) Wavefront Technologies, 1992. Appendix B1. Object Files (.obj), Advanced Visualizer Manual. URL: http://fegemo.github.io/cefet-cg/attachments/obj-spec.pdf. * Weber and Penn (1995) Weber, J., Penn, J., 1995. Creation and rendering of realistic trees, in: Mair, S.G., Cook, R. (Eds.), Computer graphics proceedings, Association for Computing Machinery, New York, NY. pp. 119–128. doi:10.1145/218380.218427. * Weiser et al. (2021) Weiser, H., Winiwarter, L., Fassnacht, F., Höfle, B., 2021\. Opaque Voxel-based Tree Models for Virtual Laser Scanning in Forestry Applications. In preparation. * Winiwarter et al. (2019) Winiwarter, L., Mandlburger, G., Schmohl, S., Pfeifer, N., 2019\. Classification of ALS Point Clouds Using End-to-End Deep Learning. PFG – Journal of Photogrammetry, Remote Sensing and Geoinformation Science URL: https://doi.org/10.1007/s41064-019-00073-0, doi:10.1007/s41064-019-00073-0. * Winiwarter et al. (2021) Winiwarter, L., Pena, A.M.E., Weiser, H., Anders, K., Saches, J.M., Searle, M., Höfle, B., 2021. 3dgeo-heidelberg/helios: Version 1.0.0. URL: https://doi.org/10.5281/zenodo.4452871, doi:10.5281/zenodo.4452871. * Xiao et al. (2019) Xiao, W., Zaforemska, A., Smigaj, M., Wang, Y., Gaulton, R., 2019. Mean shift segmentation assessment for individual forest tree delineation from airborne lidar data. Remote Sensing 11, 1263\. doi:10.3390/rs11111263. * Zhang et al. (2019) Zhang, Z., Li, J., Guo, Y., Yang, C., Wang, C., 2019\. 3D Highway Curve Reconstruction From Mobile Laser Scanning Point Clouds. IEEE Transactions on Intelligent Transportation Systems , 1–11doi:10.1109/TITS.2019.2946259. * Zhu et al. (2020) Zhu, X., Liu, J., Skidmore, A.K., Premier, J., Heurich, M., 2020. A voxel matching method for effective leaf area index estimation in temperate deciduous forests from leaf-on and leaf-off airborne LiDAR data. Remote Sensing of Environment 240\. doi:10.1016/j.rse.2020.111696.
# Thermodynamics of light emission Antoine Rignon-Bret<EMAIL_ADDRESS>École Normale Supérieure, 45 rue d’Ulm, F-75230 Paris, France ###### Abstract Some interactions between classical or quantum fields and matter are known to be irreversible processes. Here we associate an entropy to the electromagnetic field from well-known notions of statistical quantum mechanics, in particular the notion of diagonal entropy. We base our work on the study of spontaneous emission and light diffusion. We obtain a quantity which allows to quantify irreversibility for a quantum and classical description of the electromagnetic field, that we can study and interpret from a thermodynamical point of view. ## I Introduction Quantification of irreversibility has been a main issue for a long time in physics, and becomes even more important nowadays with the development of new branches as quantum information or quantum thermodynamics. Among all the physical phenomena involving irreversible processes, we will focus here on some well known process appearing in field theory. The equations governing classical or quantum field theories are known to be reversible, however some processes described by these theories exhibit an irreversible behavior. For instance, we can think about radiation damping in classical electrodynamics [1], emission of gravitational waves in relativistic gravitation [2] or spontaneous emission in quantum electrodynamics [3]. The way of this irreversible behavior arises from reversible field equations is very well understood and belongs to the established knowledge of physics. However, in this paper, we will try to quantify explicitly the degree of irreversibility of such matter-radiation interactions, using the framework of thermodynamics and statistical mechanics. We will focus here on the electromagnetic field and we will show that we can associate an entropy variation to such processes, which vanishes when there is no asymmetry between emission of light and absorption, but which is strictly positive in the opposite case. The situation is analog to the fall of a ball in a gravity field. The ball falls because of gravity, but if there were no dissipation, it could bounce back and reach the initial height. However, it does not, because by hitting the ground, some part of the mechanical energy of the ball go to the atoms in the ground. They get excited and so the final state of the ball is the state of maximal entropy or minimal energy. If we had a precise enough thermometer, we could observe a temperature variation because of this excitation, and this is the signature of an increasing entropy variation. We aim to think some matter-radiations interactions, as the spontaneous emission process, in the same way, which means that we have to find the corresponding vibrating degrees of freedom which makes the ground state stable unlike the excited states, and get an entropy from them. By doing this, we will associate some entropy to the electromagnetic field. The idea of associating an entropy to the electromagnetic field is old, as old as thermodynamics and electromagnetism are. Actually it is by confronting these two theories together that Einstein understood that radiation was made of indivisible quanta of energy [4], which lead to quantum mechanics, with the success we know. However, if Einstein worked with a thermostatted box where radiation was trapped and reached equilibrium with matter, the aim in the following work is different. We focus on quantum or classical electromagnetic processes, possibly involving a single atom and a single photon, so there is no associated thermal equilibrium and we are not in the thermodynamic limit. However, as entropy is an extensive quantity, unlike temperature, the notion still makes sense as long as we can count states. Finding the ”good” microscopic definition of entropy is still a challenge. Aside from the well known Von Neumann entropy $S=-k_{B}Tr\big{(}\hat{\rho}ln\hat{\rho}\big{)}$, more general frameworks have been developed [5]. For this work, we will be particularly interested in the diagonal entropy [6], which has been introduced to quantify the degree of irreversibility of some transformations in an isolated system. Indeed, the classical Von Neumann entropy associated to quantum systems always vanishes for unitary evolutions. In the following, after reminding the main features of spontaneous emission in section II, we will give in section III an appropriate definition of the entropy of the electromagnetic field, that we will relate to diagonal entropy. In particular, we will give a thermodynamical interpretation to the formula and we will check that it verifies the basic requirements for an entropy. In section IV, we will talk a bit about classical electrodynamics and we will show that the electromagnetic field entropy formula obtained in the previous sections is still valid and makes sense. ## II The framework of spontaneous emission To fully understand our purpose, it may be useful to return to the general framework of quantum electrodynamics and spontaneous emission. We will only remind some basic results and make some comments, further details or proofs are given in Scully’s book [3]. The general hamiltonian is given by : $\hat{H}=\hat{H}_{atom}+\hat{H}_{em}+\hat{H}_{int}$ (1) $\hat{H}=\frac{\hbar\omega_{0}}{2}(Id+\hat{\sigma}_{z})+\sum_{(\textbf{k},s)}\hbar\omega_{\textbf{k}}\hat{a}_{\textbf{k},s}^{\dagger}\hat{a}_{\textbf{k},s}+\sum_{(\textbf{k},s)}\hbar\lambda_{\textbf{k},s}(\hat{\sigma}_{+}\hat{a}_{\textbf{k},s}+\hat{\sigma}_{-}\hat{a}_{\textbf{k},s}^{\dagger})$ (2) where : $\lambda_{\textbf{k},s}=\textbf{d}\cdot\textbf{u}_{\textbf{k},s}\sqrt{\frac{\hbar\omega_{\textbf{k}}}{2\epsilon_{0}V}}$ (3) is a result obtained from quantum field theory. Here, d is the dipole moment operator matrix element between the atom ground state and excited state, $\textbf{u}_{\textbf{k},s}$ is a unit vector in Fourier space in the direction k and polarization $s$, $\hat{a}_{\textbf{k},s}$($\hat{a}^{\dagger}_{\textbf{k},s}$) being the electromagnetic field state $(\textbf{k},s)$ annihilation (creation) operator, $\hat{\sigma}_{+}$($\hat{\sigma}_{-}$) the atom quantum energy creation (annihilation) operator, and finally $V=L^{3}$ the volume of the cavity which encloses the atom. As we study the desexcitation of an atom, there is at most one photon in the cavity, and the state of the system atom + field is the following : $\ket{\Psi(t)}=c_{0}(t)\ket{e}\ket{0}+\sum_{(\textbf{k},s)}c_{(\textbf{k},s)}(t)\ket{g}\ket{\textbf{k},s}$ (4) where $\ket{e}$($\ket{g}$) in the atom excited (ground) state. By using Schrodinger equation with hamiltonian (2) and projecting it onto proper states we get the following equation system : $\displaystyle\dot{c}_{0}(t)=-i\omega_{0}c_{0}(t)-i\sum_{(\textbf{k},s)}\lambda_{\textbf{k},s}c_{\textbf{k},s}(t)$ (5) $\displaystyle\dot{c}_{\textbf{k},s}(t)=-i\omega_{\textbf{k}}c_{\textbf{k},s}(t)-i\lambda_{\textbf{k},s}c_{0}(t)$ (6) Here is the point. Until now, we considered that the atom was enclosed in a box of finite volume $V=L^{3}$. The size of the box is relevant because it constrains the electromagnetic modes that can propagate in the box by the quantification of the wave vector k. To lead the calculation further, we will assume that $L$ is very big such that the sums on k can be replaced by integrals. Then, after using the Weiskoppf-Wigner approximation, we can show that : ${c}_{0}(t)=e^{-\frac{\Gamma}{2}t-i\omega_{0}t}$ (7) With : $\Gamma=\frac{d^{2}\omega_{0}^{3}}{3\pi\hbar\epsilon_{0}c^{3}}$ (8) From (7), the probability to find the atom in its excited state after time $t$ is simply : $P_{e}(t)=e^{-\Gamma t}$ (9) The process is here clearly irreversible, the atom will stay forever in its ground state after de-excitation. The reason why we get an irreversible behavior starting from the Schrodinger equation which is clearly reversible is that we considered that the box was very big. If the box is relatively small, after emission, the photon can reflect on the cavity mirrors and come back to the atom in order to be reabsorbed, so with finite $L$, we do not get a decaying exponential but a periodic function, the photon being emitted, reflected, reabsorbed, then re-emitted, etc. If $L$ goes to infinity, the photon is never reflected by the walls of the cavity and the process is irreversible. Mathematically, a discrete sum of periodic function is still a periodic function, while a continuous sum (integral) of periodic function is not a periodic function anymore, so by doing the approximation of replacing sums by integrals we exhibited the irreversibility in the calculations. Now, we are interested in the electromagnetic wave packet components, so in the values of the $c_{k,s}$. We can inject (7) into (6) to find (when $t>>\Gamma$) : $c_{\textbf{k},s}=\frac{\lambda_{\textbf{k},s}}{(\omega_{k}-\omega_{0})+i\frac{\Gamma}{2}}e^{-i\omega_{\textbf{k}}t}$ (10) By taking the norm squared and by summing on the k of constant modulus, we get the $\omega$ probability distribution : $P(\omega)=\sum_{(\textbf{k},s)}\lvert c_{\textbf{k},s}\rvert^{2}\delta(\omega_{k}-\omega)=\frac{1}{\pi}\frac{\frac{\Gamma}{2}}{(\omega-\omega_{0})^{2}+\frac{\Gamma^{2}}{4}}$ (11) It is a lorentzian distribution, with standard deviation $\Delta\omega=\frac{\Gamma}{2}$ which is also its frequency width. Its time length is $\frac{1}{\Gamma}$, exactly as is the atom excited state lifetime. ## III Electromagnetic field entropy ### III.1 Statistical point of view As Eq.(4) shows, the system atom + photon is always in a pure state. The matrix density of the system reads : $\hat{\rho}=\ket{\Psi(t)}\bra{\Psi(t)}=\hat{U}_{\hat{H}}(t,0)\ket{\Psi(0)}\bra{\Psi(0)}\hat{U}^{\dagger}_{\hat{H}}(t,0)$ (12) where $\hat{U}_{\hat{H}}(t,0)$ is the unitary evolution operator between times $0$ and $t$. The Von Neumann entropy $S=-k_{B}Tr\big{(}\hat{\rho}ln\hat{\rho}\big{)}$ always vanishes at all time because $\hat{\rho}$ is always in a pure state and it seems that we can’t describe the irreversibility of the spontaneous emission process by a corresponding amount of entropy. It is true as long as we stay with the pure Von Neumann entropy. However, we can introduce a little trick. The system atom + electromagnetic field is not completely isolated, because there is a reflecting cavity which surrounds it. This cavity itself is placed in a dynamic universe and interact with it. Indeed, because of long-range interaction as gravity that we can never really remove, such tiny effects will play a role if we wait a long time enough. For instance, as the photon wave vector $k$ is related to its impulsion, and when the wave packet reflects on the cavity walls, it transfers some momentum to the box, that we can schematically write as : $\displaystyle\ket{\Psi(t)}=\bigg{(}c_{0}(t)\ket{e}\ket{0}+\sum_{(\textbf{k},s)}c_{(\textbf{k},s)}(t)\ket{g}\ket{\textbf{k},s}\bigg{)}\otimes\ket{box}\longrightarrow$ (13) $\displaystyle c_{0}(t)\ket{e}\ket{0}\otimes\ket{box}+\sum_{(\textbf{k},s)}c_{(\textbf{k},s)}(t)\ket{g}\ket{\textbf{-k},s}\otimes\ket{2\textbf{k}}_{box}$ (14) By momentum conservation. Afterwards, the box can entangle with, for instance, dust present in the environment which starts a Von Neumann infinite regress [7]. If we are interested only in the system atom + electromagnetic field, we trace out the environment comprised of the box, dust, and any exterior system in the universe which interacts with our box. We get from (4): $\displaystyle\hat{\rho}_{sys}=Tr_{env}\hat{\rho}_{sys+env}$ (15) $\displaystyle=\lvert c_{0}(t)\rvert^{2}\ket{e}\ket{0}\bra{0}\bra{e}+\sum_{(\textbf{k},s)}\lvert c_{\textbf{k},s}\rvert^{2}\ket{g}\ket{\textbf{k},s}\bra{\textbf{k},s}\bra{g}$ (16) Which is a diagonal density matrix representing a mixed state, and in consequence has non-vanishing Von Neumann entropy. We used here the framework of quantum decoherence [7, 8], which ensures that after a long enough time, all the off diagonal entries of the density matrix vanish, because of entanglement with an always existing environment. Likewise, diagonal entropy has been introduced by A.Polkovnikov [6] in order to study formally the irreversibility processes in quantum systems which are meant to be closed. The definition of this diagonal entropy is : $S_{d}=-\sum_{n}\rho_{nn}ln\rho_{nn}$ (17) where $\rho_{nn}$ is the $nn$-entry of the density matrix. Polkovnikov showed that it had all the good properties that we can expect from a suitable definition of entropy. This diagonal entropy has already been investigated by many authors [9, 10, 11, 12, 13]. We can directly show that the diagonal entropy of the pure state (4) is the von Neumann entropy of the density matrix (16) obtained from a decoherent process. It reads : $S=-\lvert c_{0}\rvert^{2}ln\lvert c_{0}\rvert^{2}-\sum_{(\textbf{k},s)}\lvert c_{\textbf{k},s}\rvert^{2}ln\lvert c_{\textbf{k},s}\rvert^{2}$ (18) As from (7) $\lvert c_{0}(t)\rvert$ goes from $1$ to $0$ when $t$ goes from $0$ to infinity, we can define the spontaneous emission entropy variation as $t\longrightarrow\infty$ as: $\Delta S=-\sum_{(\textbf{k},s)}\lvert c_{\textbf{k},s}\rvert^{2}ln\lvert c_{\textbf{k},s}\rvert^{2}$ (19) which is a strictly positive quantity. ### III.2 Explicit calculation of entropy We can calculate (18) and (19) directly from the values of (10). However, we will make an approximation here to get a simpler result. Indeed, first, from (3) we know that $\lambda_{k,s}$ is proportional to $sin\theta$, where $\theta$ is the angle between the dipole moment vector d and the wave vector k. This gives us that, typically, k has non-zero probability distribution in the following solid angle (for a given norm of the wave vector $k_{0}$) : $\Omega_{k_{0}}=\int_{0}^{\pi}\int_{0}^{2\pi}k_{0}^{2}sin^{3}\theta\mathrm{d}\theta\mathrm{d}\phi=\frac{8\pi\omega_{0}^{2}}{3c^{2}}$ (20) The first approximation that we will make is setting that the distribution of the wave vector k is uniform in this angular distribution $\Omega_{k_{0}}$ in Fourier space, and vanishes elsewhere Secondly, from (11), the frequency distribution is lorentzian. As we saw, the characteristic size of the frequency width is $\frac{\Gamma}{2}<<\omega_{0}$ which is the lorentzian mean for standard values of the two quantities. So we will replace the lorentzian distribution by a uniform distribution centered on $\omega_{0}$ and of width $\delta\omega=\frac{\Gamma}{2}$. Thus, we can replace $\sum_{(\textbf{k},s)}\longrightarrow\int\frac{\mathrm{d^{3}}k}{(\frac{2\pi}{L})^{3}}$ in (19) and get from it : $\Delta S\simeq\int_{-\infty}^{+\infty}\frac{1}{\pi}\frac{\frac{\Gamma}{2}}{(\omega-\omega_{0})^{2}+\frac{\Gamma^{2}}{4}}ln\bigg{(}\frac{8\pi\omega_{0}^{2}}{3c^{3}}\bigg{(}\frac{L}{2\pi}\bigg{)}^{3}\pi\frac{(\omega-\omega_{0})^{2}+\frac{\Gamma^{2}}{4}}{\frac{\Gamma}{2}}\bigg{)}{\mathrm{d}\omega}$ (21) $\Delta S\simeq ln\frac{V\omega_{0}^{2}\delta\omega}{3\pi c^{3}}$ (22) The volume $K_{0}=\frac{\omega_{0}^{2}\delta\omega}{3\pi c^{3}}$ is the typical volume in Fourier space of the available Fourier modes of the photon emitted. The more they are, the more ”irreversible” is the desexcitation. We see that it corresponds to a volume $V_{0}$ in the real space such as : $V_{0}=\frac{3\pi c^{3}}{\omega_{0}^{2}\delta\omega}$ (23) And from now we will write : $\Delta S=ln\frac{V}{V_{0}}$ (24) The interaction between the electromagnetic field and the atom broadens the frequency range in Fourier space available for the photon. Irreversibility comes from the fact that there is not only one mode $\omega_{0}$ of the electromagnetic field that can be excited but many of them. However, the formula (24) gives us another interpretation. To make it clearer, let consider first the 1D case. In the same way as we derived (22), we can show that for a one-dimensional cavity we can associate to the electromagnetic field the entropy : $\Delta S^{(1D)}=ln\frac{L\delta\omega}{2\pi c}$ (25) Where $\delta\omega$ is the standard deviation of the one dimensional lorentzian frequency distribution. The wave packet typical length $\delta x$ can be obtained from the Heisenberg relation $\delta x\delta k\simeq\frac{1}{2}$, so we can write (25) as : $\Delta S^{(1D)}\simeq ln\frac{L}{\delta x}$ (26) Up to a irrelevant $4\pi$ factor. Actually, it is true as long as we can set $c_{0}=0$ (remember that it is rigorously true only for $L\longrightarrow\infty$). In order to understand why, let consider the the typical time of de-excitation $\tau_{em}$, which is the inverse of the spontaneous emission rate. The typical time taken by the photon to explore the whole box is $\frac{L}{c}$. Furthermore, if we define $\tau_{ps}$ as the time taken by the energy quantum to explore the whole phase space, we should consider the time $\frac{L}{c}$ when the energy quantum is propagating freely in the box and the time $\tau_{em}$, when the energy quantum is inside the atom (when the atom is in the excited state). So : $\tau_{ps}=\frac{L}{c}+\tau_{em}$ (27) as long as $L>>\delta x$, we can neglect $\tau_{em}$ before $\frac{L}{c}$ and the entropy is indeed given by (26). But in general we expect that the entropy is given by : $\Delta S^{(1D)}=ln\frac{\tau_{ps}}{\tau_{em}}$ (28) because the entropy counts the number of states accessible to the energy quantum, and the time that the atom spends in the excited state $\tau_{em}$ is the same time as the wave packet spends in the mesh of size $\delta x\simeq c\tau_{em}$. The formula (28) enhances the fact that irreversibility is just a matter of time scale. Now, if we decide to decrease the size of the cavity up to it becomes on the same order as the typical length of the wave packet, the coefficient $c_{0}$ cannot be neglected anymore. In this case we get from (27) : $\tau_{ps}\simeq\frac{L}{c}+\tau_{em}\simeq 2\tau_{em}$ (29) Therefore, when $L\simeq\delta x$ : $\Delta S^{(1D)}=ln\frac{2\tau_{em}}{\tau_{em}}=ln2$ (30) In that case, the atom will be half of the time in the excited state and half of the time in the ground state, and the process will almost seem to be reversible. We can make an analogy with a one particule Joule Gay-Lussac expansion, initially contained in a box of length $\frac{L}{2}$ and which can propagate in the whole box of size $L$ when the constraint is released. In the three-dimensional case, the volume $V_{0}$ given by (23) is not the typical volume of the wave packet, which is approximatly, for $r=ct>>\frac{c}{\delta\omega}>>\frac{c}{\omega_{0}}$ : $V_{\textbf{r}}\simeq\frac{8\pi}{3}r^{2}\frac{1}{2\delta k}$ (31) However, our theory tells us that if we wait a long time compared to the decoherence time, everything happens as if the wavepacket occupies a volume $V_{0}$ during a time $\tau_{em}=\frac{1}{\Gamma}=\frac{1}{2\delta\omega}$. In other words, our phase space, which is the total volume $V$ is discretized in a mesh of volumes $V_{0}$, and the wave packet explores it. Actually, in this framework, the wave packet looks more like a particule than a wave packet, and the formula (24) is exactly the ideal gas entropy for one particule. If we add more excited atoms in the box (with the same energy gap), we should get : $S=Nln\frac{V}{V_{0}}$ (32) as the photons do not interact. The philosophy here is we can see the spontaneous emission here as a purely thermodynamic process. Let suppose that at the beginning we ”turned-off” the interaction between the atom and the electromagnetic field. The atom is forced to stay in the excited state and there is no entropy. At $t=0$, we turn on the interaction and the energy quantum can now move freely in the whole box. Here the matter-radiation interaction plays the role of the partition, and when this constraint is released, the system can reach a new equilibrium with entropy given by (32). Seen in this way, spontaneous emission is analog to a classical Joule Gay- Lussac expansion and the state where the energy quantum is inside the atom (the state when the atom is in its excites state) is just a particular state among others. ## IV Classical electrodynamics Let consider an oscillating electron hooked to a spring, with pulsation $\omega_{0}$. If we ignore the electrodynamic laws, the electron movement is just : $\textbf{r}(t)=\textbf{r}_{0}e^{i\omega_{0}t}$ (33) However, from the Maxwell equations, we can recover the Larmor formula giving the radiated power at time $t$ and at distance $r$ from the oscillator [1] : $P_{ray}(r,t)=\frac{1}{6\pi\epsilon_{0}c^{3}}\ddot{d}^{2}(t-\frac{r}{c})$ (34) This involves that the electron mechanical energy decays. We get : $\displaystyle\frac{d\langle E\rangle_{T}}{dt}_{ray}=\frac{-e^{2}}{6\pi\epsilon_{0}c^{3}}\frac{1}{T}\int_{t}^{t+T}a(t^{\prime})\frac{dv(t^{\prime})}{dt^{\prime}}\mathrm{dt^{\prime}}$ (35) $\displaystyle=\frac{+e^{2}}{6\pi\epsilon_{0}c^{3}}\frac{1}{T}\bigg{(}\int_{t}^{t+T}v(t^{\prime})\frac{da(t^{\prime})}{dt^{\prime}}-[v(t^{\prime})a(t^{\prime})]_{t}^{t+T}\bigg{)}$ (36) where $T=\frac{2\pi}{\omega_{0}}$. But $[v(t^{\prime})a(t^{\prime})]_{t}^{t+T}\simeq 0$ because the electron trajectory is almost periodic. Therefore the electron is submitted to a radiative force : $F_{ray}(r,t)=\frac{e^{2}}{6\pi\epsilon_{0}c^{3}}\frac{da(t)}{dt}$ (37) which in near harmonic regime with pseudo pulsation $\omega_{0}$ becomes : $F_{ray}(r,t)=-m\frac{v(t)}{\tau}$ (38) where : $\frac{1}{\tau}=\frac{e^{2}\omega_{0}^{2}}{6\pi m\epsilon_{0}c^{3}}$ (39) Therefore, at time $t>0$, the electron trajectory is : $\textbf{r}(t>0)=\textbf{r}_{0}e^{-\frac{t}{2\tau}+i\omega_{0}t}$ (40) Thus, by comparing Eq (33) and (40), we understand that the interaction matter-radius allows new ways of vibrating, because the frequency range broadens. Of course, it is very similar to the spontaneous emission process we studied in the previous sections. But it is also similar to the ball which hits the ground and transforms its mechanical energy into heat. As their total entropy increases because they have new ways of vibrating, the entropy of the electromagnetic field rises because the electron can excite new Fourier modes of the electromagnetic field. Therefore the electromagnetic field entropy increasing should be equal to : $\Delta S=-\sum_{(\textbf{k},s)}p(\textbf{k},s)lnp(\textbf{k},s)$ (41) where $p(\textbf{k},s)$ is the amount of energy of the emitted signal going into the Fourier mode $(\textbf{k},s)$ (divided by the total energy). Of course, the Fourier transform of (40) is easy to calculate, and its amplitude squared gives the energy contained in the Fourier modes. The calculations lead again to a lorentzian distribution with mean value $\omega_{0}$ and standard deviation $\frac{1}{2\tau}$. Thus, if we enclose our oscillator in a cubic box of volume $V=L^{3}$, the density of states is $(\frac{L}{2\pi})^{3}$. Following the same steps and the same approximations as in the previous sections, we get at the end : $\Delta S=ln\frac{V}{V_{0}}$ (42) With : $V_{0}=\frac{6\pi\tau c^{3}}{\omega_{0}^{2}}$ (43) Of course, it is totally similar to what we found previously. It’s not surprising, the entropy formula (22) we found previously does not involve any purely quantum quantity, and it should also apply to the classical case. The oscillator energy splits into many components, because there are many available modes of the electromagnetic field, and this splitting is responsible of the irreversibility. Thanks to (42), we can interpret the radiation force (37) as an entropic force. As for the spontaneous emission, the entropy associated to this splitting can be interpreted as the perfect gaz entropy, which enhances a description of the Fourier modes in terms of particules. Therefore, a corpuscular description of the Fourier modes seems to be relevant for matter-radiation interaction. ## V Conclusion We found a quantity for the classical and quantum electromagnetic field which measures the degree of irreversibility of the matter-radiation process. We claimed that we could associate an entropy production to all irreversible processes, and in particular to matter-radiation processes. If we focused on the electromagnetic field here, we can think about irreversibility processes in other field theories, as gravity. Indeed, the gravitational wave emission is analog to the electromagnetic wave emission of the dipole, except that the gravitational wave dipole moment vanishes, and is replaced by the quadrupole moment. However, gravity is a more complex topic than electromagnetism. First, gravitation is not linear. Second, the volume $V$ of the ”box” containing the gravitationnal system is itself solution of the field equations. ## VI Acknowledgements I want to thank my friends Martin Caelen, Helmy Chekir, Léonard Ferdinand and Pierre Vallet for their feedback on this work and for having encouraged me to write it down. I want espacially to thank Mark T. Mitchison for his careful reading and his invaluable advice. ## References * Feynman _et al._ [1965] R. P. Feynman, R. B. Leighton, and M. Sands, American Journal of Physics 33, 750 (1965). * Einstein and Rosen [1937] A. Einstein and N. Rosen, Journal of the Franklin Institute 223, 43 (1937). * Scully and Zubairy [1999] M. O. Scully and M. S. Zubairy, “Quantum optics,” (1999). * Einstein [1905] A. Einstein, Annalen der Physik , 1 (1905). * Šafránek _et al._ [2020] D. Šafránek, A. Aguirre, J. Schindler, and J. Deutsch, arXiv preprint arXiv:2008.04409 (2020). * Polkovnikov [2011] A. Polkovnikov, Annals of Physics 326, 486 (2011). * Laloë [2001] F. Laloë, American Journal of Physics 69, 655 (2001). * Zurek [2003] W. H. Zurek, Reviews of modern physics 75, 715 (2003). * Santos _et al._ [2011] L. F. Santos, A. Polkovnikov, and M. Rigol, Physical review letters 107, 040601 (2011). * Giraud and García-Mata [2016] O. Giraud and I. García-Mata, Physical Review E 94, 012122 (2016). * Piroli _et al._ [2017] L. Piroli, E. Vernier, P. Calabrese, and M. Rigol, Physical Review B 95, 054308 (2017). * Wang _et al._ [2020] Z. Wang, Z.-H. Sun, Y. Zeng, H. Lang, Q. Hong, J. Cui, and H. Fan, Physics Letters A , 126333 (2020). * Sun _et al._ [2020] Z.-H. Sun, J. Cui, and H. Fan, Physical Review Research 2, 013163 (2020).
# SUTRA: An Approach to Modelling Pandemics with Undetected Patients, and Applications to COVID-19 Manindra Agrawal, Madhuri Kanitkar, Deepu Phillip, Tanima Hajra, Arti Singh, Avaneesh Singh, Prabal Pratap Singh and Mathukumalli Vidyasagar MA, DP, TH, ArS, AvS, PPS are at Indian Institute of Technology Kanpur, Kanpur, UP 208016; MK is the Vice Chancellor of Maharashtra University of Health Sciences, Nashik, MH 422004; MV is with the Department of Artificial Intelligence, Indian Institute of Technology Hyderabad, Kandi, TS 502284; MA is the corresponding author. Email: <EMAIL_ADDRESS> ###### Abstract The Covid-19 pandemic has two key properties: (i) asymptomatic cases (both detected and undetected) that can result in new infections, and (ii) time- varying characteristics due to new variants, Non-Pharmaceutical Interventions etc. We develop a model called SUTRA (Susceptible, Undetected though infected, Tested positive, and Removed Analysis) that takes into account both of these two key properties. While applying the model to a region, two parameters of the model can be learnt from the number of daily new cases found in the region. Using the learnt values of the parameters the model can predict the number of daily new cases so long as the learnt parameters do not change substantially. Whenever any of the two parameters changes due to the key property (ii) above, the SUTRA model can detect that the values of one or both of the parameters have changed. Further, the model has the capability to relearn the changed parameter values, and then use these to carry out the prediction of the trajectory of the pandemic for the region of concern. The SUTRA approach can be applied at various levels of granularity, from an entire country to a district, more specifically, to any large enough region for which the data of daily new cases are available. We have applied the SUTRA model to thirty-two countries, covering more than half of the world’s population. Our conclusions are: (i) The model is able to capture the past trajectories very well. Moreover, the parameter values, which we can estimate robustly, help quantify the impact of changes in the pandemic characteristics. (ii) Unless the pandemic characteristics change significantly, the model has good predictive capability. (iii) Natural immunity provides significantly better protection against infection than the currently available vaccines. These properties of the model make it useful for policy makers to plan logistics and interventions. ## 1 Introduction The COVID-19 pandemic caused by the SARS-CoV-2 virus has by now led to more than 600 million reported cases and more than six million deaths worldwide, as of October 1, 2022 [1]. By way of comparison, the infuenza epidemic of 1957 led to 20,000 deaths in the UK and 80,000 deaths in the USA, while the 1968 influenza pandemic led to 30,000 deaths in the UK and 100,000 deaths in the USA [2]. In contrast, the COVID-19 pandemic has already led to more than one million deaths in the USA and more than 170,000 deaths in the UK [2]. Therefore the COVID-19 pandemic is the most deadly since the Spanish Flu pandemic which started in 1918. In the USA, 675,000 people, or 0.64% of the population, died in that pandemic [3], compared to 0.33% of the population in the current pandemic. In order to cope with a health crisis of this magnitude, governments everywhere require accurate projections of the progress of the pandemic, both in space and over time, and at various levels of granularity. In addition, decision-makers also require assessments of the relative effectiveness of different non-pharmaceutical interventions (NPIs) such as lockdowns. Over the past century or so, various mathematical models have been developed to predict the trajectory of a pandemic. These can be classified in three broad categories: statistical models, state-space models, and empirical models [4]. The most popular among these are compartment models, a type of state- space models. These models divide population into disjoint compartments representing different stages of infection, and are based on the premise that the disease spreads when an infected person comes into contact with a susceptible person. In the initial SIR model [5], the population was divided into three compartments: S (Susceptible), I (Infected), and R (Removed / Recovered). Subsequently an intermediate compartment of E (Exposed) was introduced between S and I [6]. In SEIR model, interactions between S and E do not lead to fresh infections. For pandemics like COVID-19, that have significant number of asymptomatic patients, instead of E, compartment A (Asymptomatic) with interactions between S and A also leading to new infections, is more suited [7]. For COVID-19 pandemic, a number of models have been introduced to capture its trajectory. Two significant features of this pandemic are the presence of a large number of undetected cases and time-varying parameter values due to the emergence of various mutants, lockdowns etc. For any model to capture the trajectory well, it must take into account these two features. Some of the proposed models have large number of compartments in order to make the model biologically more realistic (for example [8]). These models have a large number of parameters, and estimating their values reliably is not possible from reported data due to the well-known phenomenon of the bias-variance tradeoff in statistics. This is even more true when the parameter values change with time. Besides the ones mentioned above, there are other models work with four compartments (and consequently a small number of parameters). In these models, parameter estimation is easier (for example [9, 10, 11]). However, even these models have to make certain assumptions about how the parameter values change. For example, [9] assumes the parameter values are derived from other considerations and do not change with time, while both [10] and [11] assume that parameters change following a specific equation. ## 2 Our Contributions Against this background, in this paper we propose a four compartment model called SUTRA (Susceptible, Undetected, Tested positive, and Removed, Approach).111In Sanskrit, the word Sutra also means an aphorism. Sutras are a genre of ancient and medieval Hindu texts, and depict a code strung together by a genre. The components are same as in [10], however, there are differences in the dynamics (with better epidemiological justification, see 3). This has following consequences. ### 2.1 Fundamental Equation The model admits a time invariant relationship (called fundamental equation) between detected new cases (${\mathcal{N}}_{T}$), detected active cases (${\mathcal{T}}$), and total detected cases (${\mathcal{C}}_{T}$): ${\mathcal{T}}(t)=\frac{1}{\tilde{\beta}}{\mathcal{N}}_{T}(t+1)+\frac{1}{\tilde{\rho}P_{0}}{\mathcal{C}}_{T}(t){\mathcal{T}}(t)$ where $\tilde{\beta}$ and $\tilde{\rho}$ are parameters of the model and $P_{0}$ the population of region under study (see section 3). ### 2.2 Estimation of $\tilde{\beta}$ and $\tilde{\rho}$ We can efficiently estimate values of the two parameters $\tilde{\beta}$ and $\tilde{\rho}$ for the entire duration of the pandemic using the fundamental equation and simple linear regression (see section 5). These parameters have the following interpretation: * • Parameter $\tilde{\beta}\approx\beta$, a standard parameter denoting contact rate (also called transmission rate). * • We may interpret $\tilde{\rho}\approx\epsilon$ where $\epsilon$ is detection ratio, the ratio of detected new cases to total number of new cases (as done in an earlier version of our model [12]), however, it is not satisfactory since in every region, value of $\tilde{\rho}$ is observed to increase by a very large factor ($1000$ or more) in the initial couple of months of pandemic before becoming much less volatile (see section 9). A better interpretation is $\tilde{\rho}\approx\epsilon\rho$ where $\rho P_{0}$ is effective population under the pandemic influence (see section 3). The large initial increase then makes sense since the effective population under pandemic at the beginning is a very small fraction of $P_{0}$ and increases very rapidly. ### 2.3 Phases of Pandemic The value of parameter $\tilde{\beta}$ reduces when restrictive measures like lockdowns are imposed, and increases when these measures are lifted or a more infectious mutant arrives. Change in the value of parameter $\tilde{\rho}$ happens for many reasons. For example, when pandemic spreads to newer regions, or is completely eliminated from a region, or a part of susceptible population gets vaccine-induced immunity, or a part of population with immunity (acquired through vaccination or prior infection) loses it (see section 6). When the value of $\tilde{\beta}$ or $\tilde{\rho}$ changes significantly, the trajectory of the pandemic changes. The model captures it as a phase change and recomputes the new values. As explained in section 6, the model can detect when the values of $\tilde{\beta}$ or $\tilde{\rho}$ are changing, and when do they stabilize. ### 2.4 Future Projection With the knowledge of $\tilde{\rho}$ and $\tilde{\beta}$, model can efficiently compute trajectory of the pandemic for the entire duration. The computed trajectory is a good estimate for future also as long as parameters do not change significantly (see section 7). ### 2.5 Estimation of $\rho$ and $\epsilon$ To understand the impact of pandemic better, it is desirable to estimate values of $\rho$ and $\epsilon$ separately instead of their product. We show (Theorem 1) that given values of $\tilde{\beta}$ and $\tilde{\rho}$ along with number of total and active infections on starting date of the simulation, there are only finitely many possible values for $\rho$ and $\epsilon$. Further, there is a unique canonical value for both at each time instant. Since infections count at the start of simulation is not known, one requires a more realistic condition to be able to compute $\rho$ and $\epsilon$ values. Towards this, we show that the condition can be replaced by knowledge of value of $\epsilon$ or $\rho$ at any one time instant (section 8). Former requirement is met with a good serosurvey of the region at any point in time. Latter requirement is achievable if we can identify the time when pandemic has spread all over the region making $\rho$ close to $1$. One also requires that there is little vaccine-induced immunity at the time, since vaccine immunity reduces $\rho$ (see Lemma 2). For COVID-19, as the Omicron mutant arrived nearly eighteen months after pandemic started and is supposed to have bypassed vaccine-immunity nearly completely (as we also show in section 10), we can assume $\rho\approx 1$ sometime after Omicron reached a region that did not implement strict control measures at the time. For such regions (that cover almost the entire world barring exceptions like China), we can estimate values of $\rho$ and $\epsilon$. ### 2.6 Analysis of the Past The computed parameter values provide a quantification of impact of various events during the course of the pandemic. This includes impact of lockdowns and other restriction measures and arrival of new mutants. Section 9 does this analysis for four countries. ### 2.7 Analysis of Immunity Loss The Omicron mutant caused widespread loss of immunity. Applying the model on thirty-two countries covering all continents and more than half the world’s population, we deduce that loss of vaccine-immunity was significantly more than natural immunity conferred by prior exposure to any variant (see section 10). This, coupled with the fact that vaccines continue to protect against severe infection, strongly suggests that the best strategy to manage the pandemic is to allow it to spread after vaccinating the population. ## 3 Model Formulation Perhaps the earliest paper to propose a pandemic model incorporating asymptomatic patients is [7]. In this paper, the population is divided into four compartments: $S$, $A$ (for Asymptomatic), $I$ (for Infected) and $R$. Interactions between members of $S$ and $A$, as well as between members of $S$ and $I$, can lead to fresh infections. In that paper, it is assumed that almost all persons in $A$ escape detection, while almost all persons in $I$ are detected by the health authorities. While the SAIR model of [7] is a good starting point for modeling diseases with asymptomatic patients, and has been used in a few models for COVID-19 ([9] for example), it is not a good fit for COVID-19 for the following reasons: (i) due to contact tracing, some fraction of $A$ does get detected and is often of similar order as detected symptomatic ones, (ii) many symptomatic cases are not detected. Therefore, the size of $I$ cannot be estimated well. In the present paper we propose a different grouping, namely: $S$ = Susceptible Population, $U$ = Undetected cases in the population, $T$ = Tested Positive, either asymptomatic or symptomatic, and $R$ = Removed, either through recovery or death. This leads to the SUTRA model, where the last A in SUTRA stands for “approach.” Same division is used for models in [10, 11]. As is standard, we use symbols $S$, $U$, $T$, $R$ to also represent (time varying) fractional size of the four compartments. The category $R$ of removed can be further subdivided into $R_{U}$ denoting those who are removed from $U$, and $R_{T}$ denoting those who are removed from $T$. As in the conventional SAIR model [7], interactions between members of $S$ on the and members of $U$ or $T$, can lead to the person in S getting infected with a certain likelihood. $S$$U$$T$$R$$\beta SU$$\epsilon\beta SU$$\gamma T$$\gamma U$ Figure 1: Flowchart of the SUTRA model A compartmental diagram of the SUTRA model is shown in Figure 1. Typically, to handle undetected cases, models assume that the size of $T$ is $\epsilon$ fraction of size of $U+T$ (see for example, [10, 11]). This is essentially equivalent to the assumption that detected new cases are $\epsilon$ fraction of new infections, as assumed in our model. Epidemiologically, all the new cases (but for rare exceptions) will remain undetected for a few days (until the symptoms appear). Therefore, one needs to justify the choice of $\epsilon\beta SU$ for detected new cases. We argue as follows: * • Recently infected persons have higher chances of getting detected for two reasons. For symptomatic cases, the symptoms appears within a few days. For asymptomatic cases, they are detected through contact tracing which mostly starts with a symptomatic case and the asymptomatic cases detected would all be infected after the initiating symptomatic case. * • Number of new cases do not change dramatically over a few days and so number of detected cases over past few days can be taken to be proportional to $\beta SU$, number of most recent cases. A few additional reasonable assumptions have been made to simplify parameter estimation. Specifically, * • It is assumed that the removal rate for both compartments $T$ and $U$ is the same. This can be justified because, due to contact tracing, a significant fraction of patients in $T$ are asymptomatic, and those people recover at the same rate as the asymptomatic people in $U$. Even for the small fraction in $T$ who develop complications and pass away, the time duration is very close to that of those who recover. * • There is no interaction shown between the $T$ and $S$ compartments. In most countries, those who test positive (whether symptomatic or not) are either kept in institutional quarantine, or told to self-quarantine. In reality, there might still be a small amount of contact between $T$ and $S$. However, neglecting this does not significantly change the dynamics of the model, and greatly simplifies the parameter estimation. With these considerations, the governing equations for the SUTRA model are: $\dot{S}=-\beta SU,$ (1) $\dot{U}=\beta SU-\epsilon\beta SU-\gamma U,\dot{T}=\epsilon\beta SU-\gamma T,$ (2) $\dot{R}_{U}=\gamma U,\dot{R}_{T}=\gamma T.$ (3) Since these quantities denote the fraction of the population within each compartment, we have $S+U+T+R_{U}+R_{T}=1.$ There are three parameters in above equations, namely $\beta$, $\gamma$, and $\epsilon$. The interpretation of these parameters is as follows: * • $\beta$ = The expected number of susceptible persons infected by an infected person in one day; it is called the contact rate or transmission rate. * • $\gamma$ = Removal rate, the rate at which infected people are removed including both recoveries and deaths. * • $\epsilon$ = Rate at which infected patients in $U$ move over to $T$. As shown later, it also equals the ratio $T/(U+T)$ most of the time, and is thus called the detection rate. Later, we introduce two more parameters $\rho$ and $c$, and derive expressions for $\tilde{\beta}$ and $\tilde{\rho}$ in terms of $\beta$, $\epsilon$, $\rho$, and $c$. ### 3.1 Analyzing Model Equations Defining $M=U+T$, $R=R_{U}+R_{T}$, we get from equations (2) and (3) that $\dot{M}+\dot{R}=\beta SU=\frac{1}{\epsilon}(\dot{T}+\dot{R}_{T}),$ (4) resulting in $M+R=\frac{1}{\epsilon}(T+R_{T})+c$ (5) for an appropriate constant of integration $c$. Adding equations (2) gives $\dot{M}=\beta SU-\gamma M=\frac{1}{\epsilon}(\dot{T}+\gamma T)-\gamma M,$ or $\frac{d(Me^{\gamma t})}{dt}=\frac{1}{\epsilon}\frac{d(Te^{\gamma t})}{dt},$ (6) resulting in $M=\frac{1}{\epsilon}T+de^{-\gamma t}$ (7) for some constant $d$. Since $e^{-\gamma t}$ is a decaying exponential, it follows that, except for an initial transient period, the relationship $M=\frac{1}{\epsilon}I$ holds. This in turn implies that $U=M-T=\frac{1-\epsilon}{\epsilon}T$. How long is the transient period? Observe that the constant $d$ equals $M(0)-\frac{1}{\epsilon}T(0)$ which is close to zero since fraction of infected cases at the start of pandemic is very small. Therefore the transient period will not last more than a few days. As we will see later, such transient periods will recur at various stages of pandemic and all of them remain small. Define $N_{T}=\dot{T}+\dot{R}_{T}=\epsilon\beta SU$, the fraction of population detected to be positive at time $t$, and $C_{T}=T+R_{T}$, the fraction of population detected to be infected up to time $t$. The above simplifications allow us to rewrite equation (2) as: $\begin{split}N_{T}&=\epsilon\beta SU=\beta(1-\epsilon)ST\\\ &=\beta(1-\epsilon)(1-(M+R))T\\\ &=\beta(1-\epsilon)(1-\frac{1}{\epsilon}(T+R_{T})-c)T\\\ &=\beta(1-\epsilon)(1-c)T-\frac{\beta(1-\epsilon)}{\epsilon}C_{T}T\end{split}$ (8) Rearrange (8) as $T=\frac{1}{\tilde{\beta}}N_{T}+\frac{1}{\epsilon(1-c)}C_{T}T,$ (9) where $\tilde{\beta}=\beta(1-\epsilon)(1-c).$ ### 3.2 Discretization of the Model Relationships The progression of a pandemic is typically reported via two daily statistics: The number of people who test positive, and the number of people who are removed (including both recoveries and deaths). The second statistics has a problem though: there is no agreement on when to classify an infected person as removed. Some do it when RTPCR test is negative, some do it when symptoms are gone for a certain period, and some others do it after a fixed period of time. For the purpose of modeling, this classification needs to be done at the time when an infected person is no longer capable of infecting others. This is hard to decide, and so is almost never done. Further, some countries do not report second statistics at all (UK for example). In such a situation, we cannot rely on reported data, and instead compute $R_{T}$ by fixing $\gamma$ to an appropriate value as discussed in section 4. Let ${\mathcal{T}}(t)$ denote the number of active detected cases on day $t$, ${\mathcal{R}}_{T}(t)$ denote the number of detected cases that are removed on or before day $t$, and ${\mathcal{N}}_{T}(t)$ denote the number of cases detected on day $t$. Note that all three are integers, and $t$ is also a discrete counter. In contrast, in the SUTRA model, $T$, $R_{T}$ and $N_{T}$ are fractions in $[0,1]$, while $t$ is a continuum. Therefore, ${\mathcal{T}}(t)=P\int_{t-1}^{t}T(s)ds,{\mathcal{R}}_{T}=P\int_{t-1}^{t}R_{T}(s)ds,{\mathcal{N}}_{T}=P\int_{t-1}^{t}N_{T}(s)ds$ where $P$ is the effective population that is potentially affected by the pandemic. Now we introduce the parameter measuring the spread of the pandemic. Define number $\rho$, called the reach, which equals $P/P_{0}$, where $P$ is the effective population and $P_{0}$ is the total population of the group under study, e.g., the entire country, or an individual state, or a district (this parameter is also introduced and studied in [11]). The reach parameter $\rho$ is usually nondecreasing, starts at $0$, and increases towards $1$ over time (situations where it decreases are discussed later). While the underlying population $P_{0}$ is known, the reach $\rho$ is not known and must be inferred from the data. Substituting $P=\rho P_{0}$, and integrating equation (9) over a day gives a relationship that involves only measurable and computable quantities ${\mathcal{T}}$, ${\mathcal{C}}_{T}={\mathcal{T}}+{\mathcal{R}}_{T}$, and ${\mathcal{N}}_{T}$, and the parameters of the model, namely ${\mathcal{T}}(t)=\frac{1}{\tilde{\beta}}{\mathcal{N}}_{T}(t+1)+\frac{1}{\tilde{\rho}P_{0}}{\mathcal{C}}_{T}(t){\mathcal{T}}(t),$ (10) where $\tilde{\rho}=\epsilon\rho(1-c).$ Note that ${\mathcal{N}}_{T}$ is shifted forward by one day since new infections reported on day $t+1$ are determined by active infections and susceptible population on day $t$. Eq. (10) is the fundamental equation governing the pandemic. It establishes a linear relationship between ${\mathcal{N}}_{T}$, ${\mathcal{T}}$, and ${\mathcal{C}}_{T}{\mathcal{T}}$, that can be computed using the fundamental equation and the first three equations below (after fixing $\gamma$). In addition to the fundamental equation, we will need discrete forms of other equations of the model to compute all quantities. We group them in two—first the quantities that can be computed from ${\mathcal{N}}_{T}$: $\begin{split}{\mathcal{T}}(t)&={\mathcal{N}}_{T}(t)+(1-\gamma){\mathcal{T}}(t-1)\\\ {\mathcal{R}}_{T}(t)&={\mathcal{R}}_{T}(t-1)+\gamma{\mathcal{T}}(t-1)\\\ {\mathcal{C}}_{T}(t)&={\mathcal{T}}(t)+{\mathcal{R}}_{T}(t)={\mathcal{N}}_{T}(t)+{\mathcal{C}}_{T}(t-1)\end{split}$ (11) The second group is of equations that involve numbers that cannot be computed from reported data: $\begin{split}{\mathcal{N}}(t)&=\beta(1-\epsilon)S(t-1){\mathcal{M}}(t-1)\\\ {\mathcal{M}}(t)&={\mathcal{N}}(t)+(1-\gamma){\mathcal{M}}(t-1)\\\ {\mathcal{R}}(t)&={\mathcal{R}}(t-1)+\gamma{\mathcal{M}}(t-1)\\\ {\mathcal{C}}(t)&={\mathcal{M}}(t)+{\mathcal{R}}(t)={\mathcal{N}}(t)+{\mathcal{C}}(t-1)\\\ {\mathcal{U}}(t)&={\mathcal{M}}(t)-{\mathcal{T}}(t)\\\ {\mathcal{R}}_{U}(t)&={\mathcal{R}}(t)-{\mathcal{R}}_{T}(t)\\\ S(t)&=1-\frac{{\mathcal{C}}(t)}{\rho P_{0}}\end{split}$ (12) It is easy to see that all the quantities can be computed using above equations in addition to the fundamental equation once the parameter values used in the equations are available. ## 4 Fixing $\gamma$ As discussed in the previous section, reported removal data does not provide a good estimate for $\gamma$. In [13], median duration of infection for asymptomatic cases was estimated in the range $[6.5,9.5]$ and mean duration for symptomatic cases in the range $[10.9,15.8]$ days with a caveat that the duration reduces when children are included. In [14], infection duration for symptomatic cases was observed to be less than $10$ days. Since our groups $U$ and $T$ consist of a mix of asymptomatic and symptomatic cases, and it is likely that an infected person stops infecting others before becoming RTPCR negative, we take the mean duration of infection for both groups to be $10$ days, implying $\gamma=0.1$. All our simulations are done using the above value of $\gamma$ and show a good fit with the actual trajectories. ## 5 Estimation of $\tilde{\beta}$ and $\tilde{\rho}$ One of the distinctive features of our approach is a methodology for estimating the values of all the parameters in the pandemic model from reported raw data on the number of daily new cases. The model has five parameters $\gamma$, $\beta$, $\epsilon$, $\rho$, and $c$. At a first glance, these appear all independent, however, we show in section 8 that last four are essentially determined by $\tilde{\beta}$ and $\tilde{\rho}$. In this section, we show how to estimate $\tilde{\beta}$ and $\tilde{\rho}$ from reported data using the fundamental equation. Let $\widehat{{\mathcal{N}}}_{T}(t)$ be the reported new infections on day $t$. Note that $\widehat{{\mathcal{N}}}_{T}(t)$ may not be the same as detected new infections on day $t$ since there may be delays in reporting detected cases. Moreover, weekends often see fewer tests being done, causing unexpected variations in $\widehat{{\mathcal{N}}}_{T}$. To remove latter, we average $\widehat{{\mathcal{N}}}_{T}(t)$ over a week, and let $\widetilde{{\mathcal{N}}}_{T}(t)=\frac{1}{7}\sum_{j=0}^{6}\widehat{{\mathcal{N}}}_{T}(t-j).$ Let $\widetilde{{\mathcal{C}}}_{T}(t)=\sum_{s=0}^{t}\widetilde{{\mathcal{N}}}_{T}(s)$, the total number of reported cases until day $t$, and $\widetilde{{\mathcal{T}}}(t)$ be the number of reported active cases on day $t$ computed inductively using equation $\widetilde{{\mathcal{T}}}(t)=\widetilde{{\mathcal{N}}}_{T}(t)+(1-\gamma)\widetilde{{\mathcal{T}}}(t-1)$. Fix a time interval $[t_{0},t_{1}]$. Define ($t_{1}-t_{0}$)-dimensional vectors ${\mathbf{u}}$, ${\mathbf{v}}$, ${\mathbf{w}}$ as follows: ${\mathbf{u}}(t-t_{0})=\widetilde{{\mathcal{T}}}(t),t_{0}\leq t<t_{1},$ ${\mathbf{v}}(t-t_{0})=\widetilde{{\mathcal{N}}}_{T}(t+1),t_{0}\leq t<t_{1},$ ${\mathbf{w}}(t-t_{0})=\frac{1}{P_{0}}\widetilde{{\mathcal{C}}}_{T}(t)\widetilde{{\mathcal{T}}}(t),t_{0}\leq t<t_{1}.$ Then the following linear regression problem is solved: $\min_{\tilde{\beta},\tilde{\rho}}||{\mathbf{u}}-\frac{1}{\tilde{\beta}}{\mathbf{v}}-\frac{1}{\tilde{\rho}}{\mathbf{w}}||^{2}.$ The quality of the fit parameter, usually denoted by $R^{2}$, is computed as follows: $R^{2}=1-\frac{||{\mathbf{u}}-\frac{1}{\tilde{\beta}}{\mathbf{v}}-\frac{1}{\tilde{\rho}}{\mathbf{w}}||^{2}}{||{\mathbf{u}}||^{2}},$ with the optimal parameter choices. The closer $R^{2}$ is to one, the better is the quality of the fit. At times, when there are relatively few data points ($t_{1}-t_{0}$ is small), or the data has significant errors, above linear regression method fails to work (e.g., estimated parameter value becomes negative). In such situations we use a different method for estimation that is more tolerant to errors as described below. Let $\displaystyle R^{2}_{\beta}$ $\displaystyle=$ $\displaystyle 1-\frac{|{\mathbf{u}}-\frac{1}{\tilde{\beta}}{\mathbf{v}}-\frac{1}{\tilde{\rho}}{\mathbf{w}}|^{2}}{|{\mathbf{u}}-\frac{1}{\tilde{\rho}}{\mathbf{w}}|^{2}}$ $\displaystyle R^{2}_{\rho}$ $\displaystyle=$ $\displaystyle 1-\frac{|{\mathbf{u}}-\frac{1}{\tilde{\beta}}{\mathbf{v}}-\frac{1}{\tilde{\rho}}{\mathbf{w}}|^{2}}{|{\mathbf{u}}-\frac{1}{\tilde{\beta}}{\mathbf{v}}|^{2}}$ Find values of $\tilde{\beta}>0$ and $\tilde{\rho}>0$ that maximize the product $R^{2}=R^{2}_{\beta}\cdot R^{2}_{\rho}$. This choice ensures that both $\tilde{\beta}$ and $\tilde{\rho}$ play almost equally significant roles in minimizing the error. Further, the desired maximum of $R^{2}_{\beta}R^{2}_{\rho}$ is guaranteed to exist: ###### Lemma 1. When ${\mathbf{u}}$ is independent of ${\mathbf{v}}$ as well as ${\mathbf{w}}$, there is a maxima of $R^{2}$ with $R^{2}_{\beta},R^{2}_{\rho},\tilde{\beta},\tilde{\rho}>0$. The only situation when the above method will not yield the desired maxima of $R^{2}$ is when ${\mathbf{u}}$ is dependent on either ${\mathbf{v}}$ or ${\mathbf{w}}$. Former implies that ${\mathcal{T}}$ is proportional to ${\mathcal{N}}_{T}$ over the time period, or equivalently, $S$ does not change over the period. This implies ${\mathcal{N}}=0={\mathcal{N}}_{T}={\mathcal{T}}$ for the period. Similarly, latter implies that ${\mathcal{T}}$ is proportional to ${\mathcal{C}}_{T}{\mathcal{T}}$ for the duration, or equivalently, ${\mathcal{C}}_{T}$ does not change over the period. This also implies that ${\mathcal{N}}_{T}=0={\mathcal{N}}$. Either case occurs when the pandemic has effectively ended and there are no new cases for an extended period. The uncertainty in the parameter estimation is computed using the standard mean-square error formula for linear regression. We use it to compute $95$% confidence interval ranges for $\tilde{\beta}$ and $\tilde{\rho}$ values. ## 6 Phases of the Pandemic The parameters $\rho$, $\beta$ and $\epsilon$ are not constant, and vary over time. This causes changes in $\tilde{\beta}$ and $\tilde{\rho}$ as well. The contact rate $\beta$ changes for following reasons: * • Emergence of new and more infectious variants of the virus, which would spread faster than its predecessor. It takes time for the new variant to overtake whatever existed previously, which is why this factor would cause $\beta$ to increase over a period. * • Non-compliance with COVID guidelines. The $\beta$ parameter measures the likelihood of infection when an infected person (from either $U$ or $T$) meets a susceptible person from $S$. Thus $\beta$ increases if people do not wear masks, or fail to maintain social distancing, and the like. * • The parameter can also decrease suddenly, with almost a step change, due to non-pharmaceutical interventions such as lockdowns. The reach $\rho$ changes for following reasons: * • Spread of the pandemic to parts of the region that were previously untouched by it causes $\rho$ to increase. The parts may even be physically co-located with parts already touched by the pandemic comprising of those people who had completely isolated themselves. * • Elimination of the pandemic from parts of the region that were under its influence causes $\rho$ to decrease by the fraction of still susceptible population of the parts. * • Vaccination of susceptible people causes $\rho$ to decrease, as these people moving out of susceptible compartment can be viewed as effective population under the pandemic reducing. Similarly, loss of immunity among immune population causes $\rho$ to increase as this can be viewed as effective population under the pandemic increasing. This is formalized by the following lemma. ###### Lemma 2. Suppose $\rho_{\text{g}ain}$ is the fraction of susceptible population that became immune via vaccination, and $\rho_{\text{l}oss}$ is the fraction of immune population that lost immunity over a specified period of time. Then the new trajectory of the pandemic is obtained by multiplying both $\beta$ and $\rho$ (equivalently both $\tilde{\beta}$ and $\tilde{\rho}$) by $1+\frac{\rho_{\text{l}oss}-\rho_{\text{g}ain}}{\rho}$. Finally, the detection rate $\epsilon$ may increase due to more comprehensive testing, and may decrease due to reduction in testing. The changes in parameter values occur either as a slow drift over an extended period of time, or as sudden rise and fall. We divide the entire timeline of the pandemic into phases, such that within each phase, the parameters are (nearly) constant. A phase change occurs when one or more parameter values change significantly. It could be due to a quick change for reasons listed above, or accumulated slow change over an extended period. By convention, we include the duration of change in a parameter as part of new phase and call it drift period of the phase. The remaining duration of a phase is called stable period of the phase. When the value of $\epsilon$ changes, then the relationship $T=\epsilon M$ breaks down. The following lemma shows that $T$ converges to $\epsilon M$ as soon as $\epsilon$ stabilizes to its new value. ###### Lemma 3. Suppose a new phase begins at time $t_{0}$ with a drift period of $d$ days. Further, suppose that the value of parameter $\epsilon$ changes from $\epsilon_{0}$ to $\epsilon_{1}$ during the drift period. Then, ${\mathcal{M}}(t_{0}+d)=\frac{1}{\epsilon_{1}}{\mathcal{T}}(t_{0}+d)$. The above analysis leads to the following methodology of phase identification and parameter estimation for phases: 1. 1. Suppose first phase starts at $t=0$. Consider a small initial drift period $d$ (we start with $d=10$) and a small time interval $[0,t_{1}]$, and compute the values of $\tilde{\beta}$ and $\tilde{\rho}$ for this interval. 2. 2. Increase the value of $t_{1}$ and adjust the value of $d$ until value of $R^{2}$ stabilizes. Freeze the computed values of $\tilde{\beta}$ and $\tilde{\rho}$ for the phase. 3. 3. Increase the value of $t_{1}$ further until the fundamental equation has significant errors. This indicates that a new phase has started. 4. 4. Repeat the same with every subsequent phase. We demonstrate the above methodology for one phase (phase $\\#9$) in India: when the delta-variant started spreading rapidly in the country during April 2021. In Appendix A, we have plotted points $(\widetilde{{\mathcal{T}}}-\frac{1}{\tilde{\beta}}\widetilde{{\mathcal{N}}}_{T},\frac{1}{P_{0}}\widetilde{{\mathcal{C}}}_{T}*\widetilde{{\mathcal{T}}})$ for different values of $t_{1}$. During the drift period of a phase, when the parameter values are changing, the points continuously drift away from a line passing through the origin (Figures 14, 14, 16) indicating that equation 10 is not satisfied. When the phase stabilizes, the points corresponding to the period line up nicely (Figures 16, 18, 18, 20, 20) indicating that the equation 10 is now satisfied. The plots also show that values of $\tilde{\beta}$ and $\tilde{\rho}$ are changing quickly during the drift period, and do not change much during stable period. This leads to easy identification of phases and stable period within. ### 6.1 Parameter values during drift period We have so far seen how to estimate values of $\tilde{\beta}$ and $\tilde{\rho}$ during stable period of every phase. However, in order to simulate the course of the pandemic, it is necessary to have the values of the parameters during the drift period as well. Suppose $d$ is the number of days in drift period, and $b_{0}$ and $b_{1}$ are the computed values of a parameter in the previous and the current phases. Then its value will move from $b_{0}$ to $b_{1}$ during the drift period. A natural way of fixing its value during the period is to use either arithmetic or geometric progression. That is, on $i$th day in the drift period the value is set to $b_{0}+\frac{i}{d}\cdot(b_{1}-b_{0})$ or $b_{0}\cdot(\frac{b_{1}}{b_{0}})^{i/d}$ respectively. Among these, geometric progression captures the way parameters change better: * • When a new, more infectious, mutant spreads in a population, its infections grow exponentially initially. This corresponds to a multiplicative increase in $\beta$. * • Similarly, a new virus spreads in a region exponentially at the beginning. This corresponds to a multiplicative increase in $\rho$. * • A lockdown typically restricts movement sharply causing a multiplicative decrease in $\beta$. * • A change in testing strategy typically gets implement fast in a region, causing a multiplicative change in $\epsilon$. For these reasons, we assume that changes in parameters $\beta(1-\epsilon)$ (this is the effective contact rate due to quarantining of detected cases), $\rho$, and $\epsilon$ are multiplicative. Further, changes in parameter $c$ are additive as it is constant of integration ensuring continuity between two phases. Therefore, we may assume that changes in $1-c$ are multiplicative. This leads to the conclusion that changes in $\tilde{\beta}$ and $\tilde{\rho}$ are also multiplicative. Having defined how the parameters change during drift periods, we assume that the equations (10), (11), and (12) hold on all days. When in drift period, even the parameter values in the equations change daily as defined above. Subsequent sections show that our model with these assumptions is able to capture the trajectory of the pandemic very well. ## 7 Future Projections Once the quantities $\tilde{\beta},\tilde{\rho}$ are estimated as above for current phase, equations (10) and (11) can be used to compute values of ${\mathcal{N}}_{T}$, ${\mathcal{C}}_{T}$, and ${\mathcal{T}}$ for the entire phase duration. If the model captures the dynamics well, the predictions for daily new cases ${\mathcal{N}}_{T}$ should match closely with averaged reported numbers $\widetilde{{\mathcal{N}}}_{T}$ after the phase enters stable period as long as parameters do not change significantly. Indeed, this is confirmed by our simulations of trajectories in multiple countries. For example, for the phase $\\#9$ of India discussed in the previous section, predicted trajectory changed rapidly when the phase was in drift period, and stabilized when it transitioned to stable period (see Figure 2). Figure 2: Predicted Trajectories for India during April-June, 2021 This property allows one to accurately predict the future course of the pandemic once the present phase stabilizes. We used this to make several successful predictions in the past. Some notable ones were predicting the timing and height of the peak of second wave of India ten days in advance [15], predicting timing of the peak of third wave in India as well as many states of the country [16], predicting timing and height of the peak of Delta- wave in UK ten days in advance [17], and predicting timing and height of the peak of Delta-wave in US more than a month in advance [18]. The predictions for India and its states were useful to the policy-makers in planning the required capacity for providing health care, and scheduling nonpharmaceutical interventions such as school reopenings. ## 8 Estimation of $\rho$ and $\epsilon$ After fixing values of $\gamma$, the model has four parameters left: $\beta$, $\rho$, $\epsilon$ and $c$. We have seen how to estimate values of composite parameters $\tilde{\beta}$ and $\tilde{\rho}$ at all times, which allows us to compute the trajectory of daily new detected cases for a region. We can obtain more information about the pandemic if values of $\rho$ and $\epsilon$ can be estimated separately. For example, $\epsilon$ will enable us to estimate trajectory of total daily new cases, including undetected ones. We provide more applications in the next two sections. Without any additional information, besides the daily new detected infections time series, it is not possible to estimate value of $\epsilon$: ###### Lemma 4. Given detected new cases trajectory, ${\mathcal{N}}_{T}(t)$, $0\leq t\leq t_{F}$, there exist infinitely many total new cases trajectories and corresponding values of $\epsilon$ consistent with ${\mathcal{N}}_{T}$. In this section, we show that with just one additional data point—${\mathcal{C}}(0)$ and ${\mathcal{M}}(0)$, the total number of cases up to time $t=0$ and total active cases at $t=0$—-the number of possible trajectories for ${\mathcal{N}}(t)$, consistent with the given data, becomes finite: ###### Theorem 1. Given detected new cases trajectory, ${\mathcal{N}}_{T}(t)$, $0\leq t\leq t_{F}$ and ${\mathcal{C}}(0)$, there exist only finitely many trajectories for ${\mathcal{N}}(t)$ consistent with ${\mathcal{N}}_{T}$. Further, a good estimate for all the trajectories can be obtained efficiently. As is shown in the proof of above theorem (see Appendix C), the trajectory for ${\mathcal{N}}(t)$ is unique for the first phase, but there may be multiple ones for subsequent phases, identified by a unique value for the pair $(\epsilon,c)$ for each. Using the observation that the value of $\epsilon$ from one phase to next will not change significantly, we can identify a unique canonical trajectory for total new cases: for each phase, given the possible values of $\epsilon$ that give rise to consistent trajectories for ${\mathcal{N}}(t)$, choose the canonical value of $\epsilon$ to be the one closest to the canonical value of $\epsilon$ of previous phase (for first phase, there is anyway a unique value of $\epsilon$). The corresponding trajectory for ${\mathcal{N}}(t)$ is called canonical trajectory. As the theorem also states, the canonical value of $\epsilon$ and corresponding value of $c$ can be efficiently estimated. This, in turn, provides values of $\beta=\frac{\tilde{\beta}}{(1-\epsilon)(1-c)}$ and $\rho=\frac{\tilde{\rho}}{\epsilon(1-c)}$ for the phase. In this way, we get the values of all parameters at all times. ### 8.1 Calibrating the Model Above shows how to estimate parameter values for all phases, provided we know the values of ${\mathcal{C}}(0)$. This is equivalent to finding out the values of parameters $\epsilon$ for the first phase, say $\epsilon_{1}$, since ${\mathcal{C}}(0)=\frac{1}{\epsilon_{1}}{\mathcal{C}}_{T}(0)$ (since $c_{1}=0$ as shown in the proof of theorem 1). While this is good in theory, we do not know $\epsilon_{1}$ in practice. Moreover, time $t=0$ when the data becomes available for the first time is unlikely to be the time when the pandemic begins, and hence we may not have $c_{1}=0$. However, $c_{1}=0$ will still be a good estimate since $R(0)$ is likely to be very small. We estimate value of $\epsilon_{1}$ from other information available about the pandemic. This is called calibrating the model. We can calibrate the model in two ways: * • A sero-survey at time $t_{0}$ provides a good estimate of ${\mathcal{C}}(t_{0}-\delta)$, where $\delta$ equals the time taken for antibodies to develop. Once we accurately estimate $\epsilon_{1}$, the model can compute ${\mathcal{C}}(t)$ at all times $t$. We choose a suitable value of $\epsilon_{1}$ ensuring that model computation matches with the sero-survey result at time $t-\delta$. * • When the pandemic has been active long enough in a region without major, long- term restrictions, we may assume that it has reached all sections of society, making $\rho$ close to $1$. Again, we can choose $\epsilon_{1}$ that ensures that the reach of the pandemic is close to $1$ at suitable time. While using the above two methods for calibrating the model, following points need to be kept in mind: Using serosurveys. Many serosurveys suffer from significant sampling biases. For example, if a survey is done using residual sera from a period of high infection numbers, it is likely to significantly overestimate the seroprevalence because a large fraction of uninfected persons would not venture to give blood sample in such a period. In order to minimize sampling biases, therefore, one should use serosurveys done during a period of low infection numbers. Even then, some uninfected people may not participate making the estimates higher than actual. To further reduce bias, one should ideally be able to use multiple serosurveys as well as use the fact that reach is close to $1$ by a given time. Using reach. As observed earlier (Lemma 2), parameter $\rho$ is impacted by several factors, including gain and loss of immunity. Therefore, $\rho$ may not be close to $1$ even when the pandemic has spread over entire population. To capture this, we define $\rho_{\text{a}ctual}$ to denote the actual reach of pandemic, so $\rho=\rho_{\text{a}ctual}\cdot(1+\frac{\rho_{\text{l}oss}-\rho_{\text{g}ain}}{\rho_{\text{a}ctual}})=\rho_{\text{a}ctual}+\rho_{\text{l}oss}-\rho_{\text{g}ain}$. To calibrate the model using $\rho_{\text{a}ctual}\approx 1$ at certain time, we need an estimate of $\rho_{\text{l}oss}-\rho_{\text{g}ain}$ too, which introduces more errors in calibration. There are regions where neither an accurate sero-survey is available, and it is evident that reach is nowhere close to $1$. For such regions, calibration cannot be done with any confidence, and so estimation of all parameter values is not possible. ## 9 Analysis of the Past The parameter table of a country enables us to quantify the impacts of various events like the arrival of a new mutant, or a lockdown. Moreover, through the reach parameter, we can also explain the somewhat mysterious phenomenon of multiple peaks occurring in rapid succession that was observed in many countries. Below, we discuss in detail the progression of the pandemic in four countries. The time series data for India was sourced from [19], and for rest of the countries from [1]. ### 9.1 India Model computed trajectory for detected cases shows an excellent match with the reported trajectory: Figure 3: Predicted and Actual Trajectories for India For estimating all parameters, the calibration was done using sero survey done in December 2020 [20], a period of low infection. Estimates of seropositivity computed by the model were matched with two other serosurveys [21, 22], and very good agreement was found. Further, our model showed that the reach was close to maximum by December 2021, a very likely scenario. Interestingly, the detection rate $\epsilon$ stayed almost unchanged at $1/32$ throughout the course of the pandemic. Comparing the timeline of the pandemic [23] with parameter table below, we observe the following. Table 1: Parameter Table for India Ph No | Start | Drift | $\beta$ | $1/\epsilon$ | $\rho$ ---|---|---|---|---|--- 1 | 03-03-2020 | 4 | $0.29\pm 0.04$ | $32$ | $0\pm 0$ 2 | 19-03-2020 | 1 | $0.33\pm 0.02$ | $32\pm 0$ | $0\pm 0$ 3 | 12-04-2020 | 4 | $0.16\pm 0$ | $32\pm 0$ | $0.033\pm 0.003$ 4 | 17-06-2020 | 30 | $0.16\pm 0.01$ | $32\pm 0$ | $0.204\pm 0.02$ 5 | 20-08-2020 | 17 | $0.16\pm 0$ | $32\pm 0$ | $0.364\pm 0.012$ 6 | 29-10-2020 | 10 | $0.18\pm 0$ | $32\pm 0$ | $0.423\pm 0.008$ 7 | 18-12-2020 | 20 | $0.19\pm 0.01$ | $32\pm 0$ | $0.448\pm 0.005$ 8 | 11-02-2021 | 35 | $0.38\pm 0.01$ | $32\pm 0$ | $0.462\pm 0.008$ 9 | 30-03-2021 | 25 | $0.28\pm 0.01$ | $32\pm 0$ | $0.828\pm 0.009$ 10 | 25-05-2021 | 0 | $0.27\pm 0$ | $32\pm 0$ | $0.859\pm 0.026$ 11 | 20-06-2021 | 38 | $0.51\pm 0.01$ | $32\pm 0$ | $0.92\pm 0.001$ 12 | 20-08-2021 | 2 | $0.58\pm 0.02$ | $32\pm 0$ | $0.92\pm 0.004$ 13 | 01-11-2021 | 35 | $0.61\pm 0.01$ | $32\pm 0$ | $0.949\pm 0.001$ 14 | 26-12-2021 | 9 | $1.56\pm 0.21$ | $32\pm 0$ | $1.03\pm 0.031$ 15 | 10-01-2022 | 7 | $1.18\pm 0.02$ | $32.1\pm 0.1$ | $1.035\pm 0.022$ 16 | 06-02-2022 | 1 | $1.56\pm 0.01$ | $32.1\pm 0$ | $1.02\pm 0.011$ 17 | 24-03-2022 | 20 | $3.45\pm 0.26$ | $32.1\pm 0$ | $1.044\pm 0.003$ 18 | 02-06-2022 | 5 | $2.89\pm 0.06$ | $32.4\pm 2.1$ | $1.078\pm 0.113$ First wave (March to October 2020): The strict lockdown imposed at the end of March 2020 brought down the contact rate $\beta$ by a factor of two. The reach was very small until May ($\approx 0.03$) but increased to $0.36$ between the end of June and the end of August. This was caused by reverse migration of workers and a partial lifting of lockdown that happened during this period. Second wave (February to July 2021): The arrival of the Delta variant caused the value of $\beta$ to rise to $0.38$ in February 2021. As the variant began to spread in different parts of the country, most states imposed restrictions, which reduced the nationwide $\beta$ to $0.28$ by April. In the same month, $\rho$ increased sharply to $0.83$. Note that while increase in $\rho$ was clearly due to Delta variant, the increase happened more than a month after increase in $\beta$. We have observed this delayed increase phenomenon in $\rho$ repeatedly. The removal of all restrictions by August caused $\beta$ to increase to $0.58$. This suggests that the Delta variant was more infectious by a factor of $\approx 2$ compared to original variant. Third wave (December 2021 to March 2022): The arrival of the Omicron variant caused $\beta$ to increase sharply to $1.56$ and $\rho$ to increase to $1.03$ (from $0.95$) by the end of December. In January, mild restrictions were imposed across the country, causing $\beta$ to drop to $1.18$. These were lifted in February, and $\beta$ went back up to $1.56$ in February. A ripple (April 2022 to September 2022): The value of $\beta$ increased to around $3$ by June. This, coupled with an increase in $\rho$ from $1.02$ to $1.08$ (indicating around $6\%$ population losing natural immunity) caused a ripple that peaked in July. At present, around $98\%$ of population is estimated to have natural immunity. ### 9.2 UK Model computed trajectory for detected cases shows a good match with the reported trajectory: Figure 4: Predicted and Actual Trajectories for UK For estimating all parameters, the calibration was done using two of the three serosurveys reported in [24]. During the period of serosurveys, the infection numbers were going up and down, which made calibration a little tricky. We used the numbers from the last two surveys as well as the observation that reach has remained stationary from November 2021 (suggesting that $\rho_{\text{a}ctual}$ has been around $1$ since then) for calibration. The detection rate started at $1/9.3$ and over time increased to almost one in three cases. Comparing the timeline of the pandemic [25] with parameter table below, we observe the following. Table 2: Parameter Table for UK Ph No | Start | Drift | $\beta$ | $1/\epsilon$ | $\rho$ ---|---|---|---|---|--- 1 | 14-03-2020 | 10 | $0.26\pm 0.01$ | $9.3$ | $0.02\pm 0.001$ 2 | 17-04-2020 | 0 | $0.15\pm 0$ | $9.3\pm 0$ | $0.063\pm 0.003$ 3 | 07-07-2020 | 10 | $0.24\pm 0.01$ | $9.3\pm 0$ | $0.08\pm 0.002$ 4 | 01-09-2020 | 7 | $0.27\pm 0.01$ | $9.3\pm 0$ | $0.122\pm 0.004$ 5 | 30-09-2020 | 0 | $0.21\pm 0.01$ | $9.3\pm 0$ | $0.296\pm 0.032$ 6 | 09-11-2020 | 5 | $0.26\pm 0.01$ | $9.3\pm 0$ | $0.3\pm 0.021$ 7 | 03-12-2020 | 25 | $0.32\pm 0.01$ | $8.7\pm 1.2$ | $0.594\pm 0.087$ 8 | 29-01-2021 | 40 | $0.68\pm 0.04$ | $8.7\pm 0.3$ | $0.616\pm 0.053$ 9 | 15-05-2021 | 25 | $0.61\pm 0.03$ | $7.7\pm 1.9$ | $0.763\pm 0.116$ 10 | 03-08-2021 | 28 | $0.37\pm 0.01$ | $6.5\pm 0.1$ | $0.921\pm 0.043$ 11 | 18-09-2021 | 27 | $0.5\pm 0.02$ | $5.9\pm 0.2$ | $0.956\pm 0.032$ 12 | 06-11-2021 | 27 | $0.54\pm 0.01$ | $5.4\pm 0.1$ | $1.046\pm 0.034$ 13 | 13-12-2021 | 20 | $0.74\pm 0$ | $4.2\pm 0$ | $1.014\pm 0.001$ 14 | 18-01-2022 | 20 | $0.87\pm 0.19$ | $3.5\pm 0.2$ | $1.013\pm 0.035$ 15 | 28-02-2022 | 15 | $0.93\pm 0.03$ | $3.1\pm 0.1$ | $1.048\pm 0.007$ 16 | 14-04-2022 | 10 | $0.73\pm 0.09$ | $2.9\pm 1.6$ | $1.019\pm 0.343$ 17 | 01-06-2022 | 3 | $1.71\pm 0.09$ | $2.8\pm 0.4$ | $1.02\pm 0.106$ First wave (March to July 2020): The strict lockdown imposed in March 2020 brought down the contact rate $\beta$ from $0.26$ to $0.15$ in mid-April. However, almost simultaneously, $\rho$ increased three-fold causing another peak. By July, $\beta$ was back up to $0.24$ after removal of restrictions. Second wave (September 2020 to January 2021): This wave was primarily caused by increase in value of $\rho$ from $0.08$ to $0.6$. This increase in $\rho$ was a natural consequence of very small effective population until August (less than one percent) and easing of lockdown from July (as noted above, increase in $\rho$ happens with a lag). As the numbers started increasing, fresh restrictions were put in place bringing $\beta$ down by $20\%$ by September-end. This caused the cases to peak by October-end (by that time $\rho$ increased to $0.3$). As the lockdown was eased before Christmas, both $\beta$ and $\rho$ started increasing causing and second bigger peak in January 2021. The rise in case number caused another lockdown, but this time $\beta$ did not decrease. Note that a new variant, called Alpha, started spreading in UK rapidly in December 2020. It was believed to be significantly more infectious than earlier one. This appears to be the reason why the value of $\beta$ did not decrease in January, and went up slightly instead. Third wave (February 2021 to October 2021): Lockdown was eased during February-March which resulted in a significant rise in $\beta$ to $0.68$ by second half of March. This jump, however, caused only a slight change in trajectory because reach stayed around $0.6$ and more than $85\%$ of population within reach had natural immunity by then. Numbers started rising from mid-June due to increase in $\rho$ (again a delayed increase). There were three peaks in quick succession: The first caused by increase in $\rho$ to $0.76$ in July, the second caused by further increase in $\rho$ to $0.92$ in August (when $\beta$ came down to $0.37$ during this period, likely caused by precautions taken by people due to high numbers), and the third caused by increase in $\beta$ to $0.5$ in addition to a slight increase in $\rho$. This increase in $\beta$ was likely due to the Delta variant now active in the country. Fourth wave (November 2021 to August 2022): In November the Omicron variant arrived causing $\beta$ to increase further. The wave had four peaks (although the second one got a bit messed up due to reporting of very large numbers on 31st January of backlog cases). These peaks were all caused by increase in $\beta$ – to $0.74$ in December, to $0.87$ in January-end, to $0.93$ in March, and finally to $1.71$ in June. The stepwise increase is connected to levels of restrictions imposed. At present, around $92\%$ of population is estimated to have natural immunity. ### 9.3 US Model computed trajectory for detected cases shows an excellent match with the reported trajectory: Figure 5: Predicted and Actual Trajectories for USA For estimating all parameters, the calibration was done using the serosurvey [26]. The samples were taken from life insurance applications. The calibration was further supported by the fact that $\rho$ has not changed since December, suggesting that $\rho_{\text{a}ctual}$ was close to $1$ at the time. The detection rate has slowly decreased from $1/3.5$ to $1/4$ during the course of the pandemic. Comparing the timeline of the pandemic [27] with parameter table below, we observe the following. Table 3: Parameter Table for US Ph No | Start | Drift | $\beta$ | $1/\epsilon$ | $\rho$ ---|---|---|---|---|--- 1 | 15-03-2020 | 3 | $0.31\pm 0.02$ | $3.5$ | $0.007\pm 0.001$ 2 | 13-04-2020 | 40 | $0.18\pm 0.01$ | $3.5\pm 0$ | $0.038\pm 0.003$ 3 | 11-06-2020 | 12 | $0.18\pm 0.01$ | $3.5\pm 0$ | $0.113\pm 0.006$ 4 | 03-09-2020 | 65 | $0.24\pm 0.01$ | $3.5\pm 0$ | $0.255\pm 0.019$ 5 | 01-12-2020 | 10 | $0.24\pm 0$ | $3.5\pm 0$ | $0.325\pm 0.007$ 6 | 30-12-2020 | 5 | $0.26\pm 0.01$ | $3.5\pm 0$ | $0.391\pm 0.022$ 7 | 19-02-2021 | 7 | $0.23\pm 0.01$ | $3.5\pm 0$ | $0.462\pm 0.009$ 8 | 08-03-2021 | 16 | $0.45\pm 0.02$ | $3.6\pm 0.6$ | $0.434\pm 0.12$ 9 | 06-06-2021 | 10 | $0.38\pm 0.01$ | $3.6\pm 0$ | $0.459\pm 0.002$ 10 | 26-06-2021 | 21 | $0.65\pm 0.01$ | $3.7\pm 0.1$ | $0.512\pm 0.044$ 11 | 11-08-2021 | 3 | $0.46\pm 0.01$ | $3.8\pm 0.1$ | $0.583\pm 0.036$ 12 | 11-09-2021 | 0 | $0.31\pm 0.01$ | $3.8\pm 0.1$ | $0.697\pm 0.09$ 13 | 17-10-2021 | 28 | $0.42\pm 0.01$ | $3.9\pm 0$ | $0.756\pm 0.007$ 14 | 28-11-2021 | 4 | $0.6\pm 0.01$ | $3.9\pm 0$ | $0.734\pm 0.003$ 15 | 22-12-2021 | 6 | $0.53\pm 0.01$ | $4.2\pm 0.2$ | $1.088\pm 0.038$ 16 | 24-02-2022 | 39 | $1.87\pm 0.02$ | $4.3\pm 0$ | $1.083\pm 0.026$ 17 | 06-04-2022 | 35 | $1.02\pm 0.02$ | $4.2\pm 0.2$ | $1.195\pm 0.032$ 18 | 24-06-2022 | 5 | $1.03\pm 0.68$ | $4.1\pm 0$ | $1.201\pm 0.056$ 19 | 08-07-2022 | 5 | $0.71\pm 0.07$ | $4\pm 0$ | $1.279\pm 0.018$ 19 | 08-07-2022 | 5 | $0.58\pm 0.06$ | $4\pm 0$ | $1.317\pm 0.024$ First wave (March to August 2020): Restrictions imposed in April 2020 brought down the contact rate $\beta$ from $0.31$ to $0.18$ by mid-May. However, almost simultaneously, $\rho$ increased to $0.04$ causing a flat trajectory. In June, most restrictions were lifted. This increased $\rho$ further to $0.11$ causing a peak in July-end. The value of $\beta$, however, did not increase. This could be due to precautions taken by a large number of people. Second wave (September 2020 to February 2021): By October, $\beta$ went up to $0.24$ and stayed around this value until the end of the wave. The value of $\rho$ increased in three steps: to $0.26$ during September-October period, to $0.33$ in December, and to $0.39$ in January. This causes three successive peaks in November, December, and January. Third wave (March 2021 to November 2021): There were two peaks separated by more than four months in this period. The Delta variant appeared to have arrived in March causing $\beta$ to increase to $0.45$. However, it caused only a small peak since $\rho$ stayed around $0.5$ until July, and more than $75\%$ of population under reach had natural immunity. The reach started increasing in August to eventually become $0.7$ by mid-September causing another peak (yet another case of delayed increase in $\rho$). Fourth wave (December 2021 to March 2022): The Omicron variant started spreading in December causing $\beta$ to increase to $0.6$, but the numbers did not increase much by December-end, since $\rho$ did not change by much. Then the reach increased substantially to $1.08$ in a short time leading to a very sharp and high peak. By February, the wave subsided, and even though $\beta$ jumped to $1.87$ in March, it did not cause cases to increase as more than $90\%$ of population was immune by then. Fifth wave (April 2022 to September 2022): The primary cause of this wave appears to be a loss of natural immunity. By May, $\rho$ increased to $1.2$ and is close to $1.3$ at present. This implies that more than $20\%$ of population has lost natural immunity in past six months. Immunity loss at such a scale has not been observed in the other three countries discussed here. Reasons for this are not clear. At present, around $85\%$ of population is estimated to have natural immunity. ### 9.4 South Africa Model computed trajectory for detected cases shows an excellent match with the reported trajectory: Figure 6: Predicted and Actual Trajectories for South Africa For estimating all parameters, the calibration was done using the serosurvey [28]. The calibration was further supported by the fact that $\rho$ has not changed since November suggesting that $\rho_{\text{a}ctual}$ has been close to $1$ since then. The detection rate has remained almost unchanged at $1/17$ during the course of the pandemic. Comparing the timeline of the pandemic [29] with parameter table below, we observe the following. Table 4: Parameter Table for South Africa Ph No | Start | Drift | $\beta$ | $1/\epsilon$ | $\rho$ ---|---|---|---|---|--- 1 | 16-04-2020 | 15 | $0.18\pm 0.01$ | $17$ | $0.036\pm 0.009$ 2 | 03-06-2020 | 25 | $0.19\pm 0.01$ | $17.1\pm 0$ | $0.241\pm 0.005$ 3 | 21-08-2020 | 10 | $0.16\pm 0.01$ | $17.1\pm 0$ | $0.288\pm 0.007$ 4 | 11-09-2020 | 15 | $0.25\pm 0.01$ | $17.1\pm 0$ | $0.329\pm 0.007$ 5 | 08-11-2020 | 5 | $0.27\pm 0.01$ | $17.1\pm 0$ | $0.4\pm 0.007$ 6 | 03-12-2020 | 5 | $0.31\pm 0$ | $17.1\pm 0$ | $0.5\pm 0.037$ 7 | 28-12-2020 | 12 | $0.46\pm 0.01$ | $17.9\pm 0.1$ | $0.485\pm 0.009$ 8 | 11-02-2021 | 40 | $0.65\pm 0.02$ | $18\pm 0$ | $0.536\pm 0.002$ 9 | 11-04-2021 | 36 | $0.4\pm 0.01$ | $18.1\pm 0$ | $0.752\pm 0.004$ 10 | 06-06-2021 | 10 | $0.43\pm 0.01$ | $19.6\pm 0.4$ | $0.938\pm 0.037$ 11 | 24-07-2021 | 30 | $0.99\pm 0.02$ | $18.2\pm 2.1$ | $0.923\pm 0.108$ 12 | 01-11-2021 | 22 | $1.58\pm 0.05$ | $17.5\pm 1.1$ | $1.033\pm 0.026$ 13 | 31-12-2021 | 7 | $1.27\pm 0.01$ | $16.4\pm 0.1$ | $1.01\pm 0.01$ 14 | 20-01-2022 | 25 | $1.56\pm 0.01$ | $16.2\pm 0$ | $1.044\pm 0.001$ 15 | 11-03-2022 | 30 | $4.35\pm 0.04$ | $16.1\pm 0$ | $1.027\pm 0$ 16 | 16-04-2022 | 15 | $2.99\pm 0.02$ | $15.8\pm 0.2$ | $1.06\pm 0.004$ 17 | 12-06-2022 | 38 | $2.56\pm 0.15$ | $15.6\pm 0$ | $1.076\pm 0.002$ First wave (April to August 2020): Restrictions imposed in March and April 2020 brought down the contact rate $\beta$ to around $0.2$. A significant increase in $\rho$ to $0.24$ by June- end caused the first wave that peaked in July-end. Second wave (September 2020 to February 2021): Restrictions were lowered in September causing increase in $\beta$ value to $0.25$. Reach also continued to increase slowly to $0.5$. This caused only a slow rise since more than $40\%$ of population under the reach was already immune. The Beta variant arrived in December causing an immediate jump in $\beta$ value to $0.46$. This caused the second peak in January. Restrictions were reimposed in December to control the rise in the numbers due to Beta variant. The fact that $\beta$ still increased substantially shows that infectiousness of this variant was quite high. Third wave (March 2021 to October 2021): With removal of restrictions measures by March, value of $\beta$ further increased to $0.65$. However, since increase in $\rho$ is typically delayed and more than $90\%$ of population within reach was already immune, the rise in $\beta$ did not cause increase in numbers. The numbers started rising when $\rho$ started increasing in April to become $0.75$ by mid-May. Restrictions were partly brought back causing $\beta$ to come down to around $0.4$. The value of $\rho$ further went up to $0.94$ causing a peak in July. As the numbers started coming down from the peak in August, an unusual phenomenon occurred. Cases started increasing once again, there was a short peak in second half of August, and then the numbers came down once again but with a slightly less steep slope than before. Our model shows that this happened due to a sharp increase in value of $\beta$ to nearly $1$ from $0.43$. Note that restrictions were being increased during June-July and were relaxed only from September, so the increase in $\beta$ was not due to relaxations. Was this caused by Omicron variant that was detected later in South Africa? The sharp increase in $\beta$ which then stayed high certainly suggests so. Fourth wave (November 2021 to March 2022): In November, with more and more relaxations, Omicron caused $\beta$ to further increase to $1.58$ and $\rho$ to $1.03$ resulting in a high peak. No restrictions were imposed this time, and so the numbers rose and fell sharply. Fifth wave (March 2022 to June 2022): A further increase in $\beta$ to around $3$ by April resulted in another peak in mid-May. This peak was, however, a small one since reach was stationary around $1.05$ and more than $95\%$ of population had natural immunity. At present, around $98\%$ of population is estimated to have natural immunity. ## 10 Analysis of the Immunity Loss After the South African authorities announced the emergence of a new variant of concern (VOC), later named Omicron, the epidemiology community started analysing the ability of the Omicron variant to bypass immunity provided by vaccination, or prior exposure, or both. Our objective in this section is to provide a quantitative analysis using the SUTRA model. But before that, we give a brief summary of the vast literature based on laboratory (as opposed to population-level) studies. Everywhere in the world where it was discovered, the Omicron VOC soon replaced all other variants and was responsible for a massive increase in cases. This was due to high transmissibility conferred by the mutation, ensuring a tight binding to the ACE 2 receptor facilitating immune escape [30]. The immune escape phenomenon was reported by many groups studying the neutralization activity of sera from both infected and vaccinated individuals; see [31, 32, 33, 34, 35]. The immunity conferred by complete vaccination decreased from 80% for the Delta variant to about 30% for the Omicron variant. People infected with the Delta were better off than those infected with the initial Beta variant. There was a complete loss of neutralizing antibodies in over 50% of the vaccinated individuals and the decrease in titres varied from 43-122 fold between vaccines [36]. A booster Pfizer dose could generate an anti-Omicron neutralizing response, but titres were 6-23 fold lower than those for Delta variant. Sera from vaccinated individual of the Pfizer or Astra Zeneca vaccine barely inhibited the Omicron variant five months after complete vaccination [37]. In addition, Omicron was completely or partially resistant to neutralization by all monoclonal antibodies tested [30]. Overall, most studies confirmed that sera from convalescent as well as fully vaccinated individuals irrespective of the vaccine (BNT162b2, mRNA-1273, Ad26.COV2.5 or ChAdOx1-nCoV19, Sputnik V or BBIBP-CorV) contained very low to undetectable levels of nAbs against Omicron. A booster with a third dose of mRNA vaccine appeared to restore neutralizing activity but the duration over which this effect may last has not been confirmed. Double vaccination followed by Delta breakthrough infection, or prior infection followed by mRNA vaccine double vaccination, appear to generate increased protective levels of neutralizing antibodies [38]. Viral escape from neutralising antibodies can facilitate breakthrough infections in vaccinated and convalescent individuals; however, pre-existing cellular and innate immunity could protect from severe disease [38, 39]. Mutations in Omicron can knock out or substantially reduce neutralization by most of the large panel of potent monoclonal antibodies and antibodies under commercial development. Studies also showed that neutralizing antibody titers against BA.2 were similar to those against the BA.1 variant. A third dose of the vaccine was needed for induction of consistent neutralizing antibody titers against either the BA.1 or BA.2.3,4 variants, suggesting a substantial degree of cross-reactive natural immunity [40]. The studies above indicate that vaccine immunity was lost substantially against Omicron, and natural immunity provided better protection. All the studies were done in laboratories or in a small section of population, and our analysis in this section complements them as it is based on population-wise data. ### 10.1 Vaccine Immunity before Omicron We first analyze the gain in immunity due to vaccination before the arrival of Omicron. For this propose, we downloaded an extensive list of serosurveys, carried out in various countries and maintained by the site [41], eliminated surveys that were not done at national level, or had small sample sizes, or had high risk of bias. Nineteen countries remained after this pruning. These sero-surveys were used together with the SUTRA model to capture the pandemic trajectories and estimate parameter values in these countries. We identify two values for each country: 1. 1. Value $\frac{1}{\rho}-1$ before arrival of Omicron. All the nineteen countries had restrictions removed well before Omicron and therefore, it is reasonable to expect that $\rho_{\text{a}ctual}$ was close to $1$ by the arrival of Omicron in the country. As shown in Lemma 2, $\rho=\rho_{\text{a}ctual}+\rho_{\text{l}oss}-\rho_{\text{g}ain}\approx 1+\rho_{\text{l}oss}-\rho_{\text{g}ain}$ at the time. In other words, $\rho_{\text{g}ain}-\rho_{\text{l}oss}\approx 1-\rho$. Since calibration of the model provides only an approximate value of $\rho$, we use fractional gain in immunity $\frac{\rho_{\text{g}ain}-\rho_{\text{l}oss}}{\rho}\approx\frac{1}{\rho}-1$ which is likely to be more robust. 2. 2. Fraction of uninfected population in the country that has received at least one dose of vaccination at the onset of Omicron wave. To estimate this number, we assume that the two types of immunity, vaccine and natural, are independent random variables, implying that the fraction with hybrid immunity is the product of vaccine immunity and natural immunity fractions. With this assumption, and using vaccination data from [42], we can estimate the required value. Figure 7 plots the above two numbers. It shows a very strong correlation between the two numbers implying that vaccination provided excellent immunity before Omicron. Figure 7: $\frac{1}{\rho}-1$ vs. Percentage of Vaccinated & Uninfected Population ### 10.2 Loss of Vaccine Immunity after Omicron We can measure immunity loss due to Omicron by comparing the value of $\rho$ after Omicron arrives in a country with the value before its arrival. This change will be almost entirely due to immunity loss since $\rho_{\text{a}ctual}\approx 1$ before Omicron as discussed above. We first compare it with vaccine-only immunity present in the population to get an estimate of how much of it was lost. Figure 8 plots these two numbers. Again, a very strong correlation is observed between the numbers. This, and the fact that the slope of best-fit line is close to $1$, suggests that almost all of immunity loss was due to loss of vaccination immunity. Figure 8: Immunity Loss after Omicron vs. Vaccinated & Uninfected Population The above conclusion is further strengthened by the next plot where we compare immunity loss due to Omicron with the natural immunity present in the population before the arrival of mutation. Figure 9 plots these two numbers. It shows a very strong negative correlation implying that natural immunity provided excellent protection against Omicron. Figure 9: Immunity Loss after Omicron vs. Naturally Immune Population ### 10.3 Incorporating More Countries To make our conclusions more broad-based, we include seventeen more countries based on following criteria: 1. 1. All continents are represented well (five from Africa, two from North America, four from South America, thirteen from Asia, eleven from Europe, and one from Australia) 2. 2. Populous countries are simulated (except China for which it is not possible to calibrate the model). More than half the world’s population lives in these countries. 3. 3. It is likely that $\rho_{\text{a}ctual}$ was close to maximum in these countries at the time of Omicron’s arrival, allowing us to calibrate the model. Adding these countries to the plots, we find little change in the correlations (see Figures 10, 11, 12), further strengthening the conclusions. Figure 10: $\frac{1}{\rho}-1$ vs. Percentage of Vaccinated & Uninfected Population Figure 11: Immunity Loss after Omicron vs. Vaccinated & Uninfected Population Figure 12: Immunity Loss after Omicron vs. Naturally Immune Population Taken together, these plots show that Omicron bypassed vaccination immunity almost completely, but natural immunity provided excellent protection. A clear conclusion is that countries that followed zero-COVID strategy – strictly control the spread and vaccinate entire population – suffered maximum during the Omicron wave. Indeed, a perusal of Table 5 shows that in countries where the reach was very low before the arrival of the Omicron variant saw very large percentage increases in reach thereafter. It also suggests that the best strategy for managing the pandemic for a country is to allow the virus to freely spread after vaccinating entire population. Trying to control its spread even after vaccination will not build natural immunity in the population and there will always be a chance of fresh outbreaks. We can see it happening in China at present. ## Acknowledgments The work of MA, DP, TH, Arti S, Avaneesh S, & Prabal S was supported by grants from CII and Infosys Foundation, and MV was supported by the Science and Engineering Research Board, India. ## References * [1] Worldometers. 2022 COVID-19 Coronavirus Pandemic. https://www.worldometers.info/coronavirus/. * [2] Honigsbaum M. 2020 Revisiting the 1957 and 1968 influenza pandemics. The Lancet 395, 1824–1826. * [3] for Disease Control C, Prevention. 2019 1918 Pandemic (H1N1 virus). https://www.cdc.gov/flu/pandemic-resources/1918-pandemic-h1n1.html. * [4] Siettos CI, Russo L. 2013 Mathematical modeling of infectious disease dynamics. Virulence 4, 295–306. * [5] Kermack WO, McKendrick AG. 1927 A contribution to the mathematical theory of epidemics. Proceedings of The Royal Society A 117, 700–721. * [6] Hethcote HW. 1976 Qualitative analyses of communicable disease models. Mathematical Biosciences 28, 335–356. * [7] Robinson M, Stilianakis NI. 2013 A model for the emergence of drug resistance in the presence of asymptomatic infections. Mathematical Biosciences 243, 163–177. * [8] Giordano, G., Blanchini, F., Bruno, R. et al.. 2020 Modelling the COVID-19 epidemic and implementation of population-wide interventions in Italy. Nat Med 26, 855–860. * [9] Martínez-Guerra R, Flores-Flores JP. 2021 An algorithm for the robust estimation of the COVID-19 pandemic’s population by considering undetected individuals. Applied Mathematics and Computation 405, 126273. * [10] Deo V, Grover G. 2021 A new extension of state-space SIR model to account for Underreporting – An application to the COVID-19 transmission in California and Florida. Results in Physics 24, 104182. * [11] Calafiore G C and Novara C and Possieri C. 2020 A time-varying SIRD model for the COVID-19 contagion in Italy. Annual Rev Control 50, 361–372. * [12] Agrawal M, Kanitkar M, Vidyasagar M. 2020 Modelling the spread of SARS-CoV-2 pandemic - Impact of lockdowns & interventions. Indian Journal of Medical Research. * [13] Byrne AW, McEvoy D, Collins AB, et al. 2020 Inferred duration of infectious period of SARS-CoV-2: rapid scoping review and analysis of available evidence for asymptomatic and symptomatic COVID-19 cases. BMJ Open 10. * [14] Daniel Owusu, Mary A Pomeroy, Nathaniel M Lewis et al. 2021 Persistent SARS-CoV-2 RNA Shedding Without Evidence of Infectiousness: A Cohort Study of Individuals With COVID-19. The Journal of Infectious Diseases. * [15] Manindra Agrawal. 2021 Tweet on 29th April. https://twitter.com/agrawalmanindra/status/1387734516807073792?s=20. * [16] Manindra Agrawal. 2022 Tweet on 13th February. https://twitter.com/agrawalmanindra/status/1492569073783570432?s=20&t=ULH1TMDIArosjgjZLvvRXw. * [17] Manindra Agrawal. 2021a Tweet on 10th July. https://twitter.com/agrawalmanindra/status/1413762687213789185?s=20&t=ULH1TMDIArosjgjZLvvRXw. * [18] Manindra Agrawal. 2021b Tweet on 25th July. https://twitter.com/agrawalmanindra/status/1419176814954500096?s=20. * [19] COVID-19 India. 2021 COVID-19 India. https://www.covid19india.org/. * [20] Manoj V Murhekar, Jeromie Wesley Vivian Thangaraj T et al.. 2021a SARS-CoV-2 seroprevalence among the general population and healthcare workers in India. Int J Infect Dis. * [21] Manoj V Murhekar, Tarun Bhatnagar, Sriram Selvaraju, V Saravanakumar, Jeromie Wesley Vivian Thangaraj et al.. 2021b SARS-CoV-2 antibody seroprevalence in India, August–September, 2020: findings from the second nationwide household serosurvey. LANCET Global Health 9, E257–E266. * [22] Manoj V Murhekar, Tarun Bhatnagar, Jeromie Wesley Vivian Thangaraj et al.. 2021c Seroprevalence of IgG antibodies against SARS-CoV-2 among the general population and healthcare workers in India, June–July 2021: A population-based cross-sectional study. PLOS Medicine. * [23] Wikipedia. 2022 Timeline of the COVID-19 pandemic in India. https://en.wikipedia.org/wiki/Timeline_of_the_COVID-19_pandemic_in_India. * [24] Ward H, Cooke G, Atchison C, Whitaker M et al.. 2021 Prevalence of antibody positivity to SARS-CoV-2 following the first peak of infection in England: Serial cross-sectional studies of 365,000 adults. LANCET Regional Health. * [25] Wikipedia. 2022 Timeline of the COVID-19 pandemic in the United Kingdom. https://en.wikipedia.org/wiki/Timeline_of_the_COVID-19_pandemic_in_the_United_Kingdom. * [26] Robert Stout and Steven Rigatti. 2021 Seroprevalence of SARS-CoV-2 Antibodies in the US Adult Asymptomatic Population as of September 30, 2020. JAMA Newwork Open. * [27] Wikipedia. 2022 Timeline of the COVID-19 pandemic in the United States. https://en.wikipedia.org/wiki/Timeline_of_the_COVID-19_pandemic_in_the_United_States. * [28] Nicole Wolter, Stefano Tempia, Anne von Gottberg, Jinal Bhiman et al.. 2022 Seroprevalence of SARS-CoV-2 after the second wave in South Africa in HIV-infected and uninfected persons: a cross-sectional household survey. Clinical Infectious Diseases. * [29] Wikipedia. 2022 Timeline of the COVID-19 pandemic in South Africa. https://en.wikipedia.org/wiki/Timeline_of_the_COVID-19_pandemic_in_South_Africa. * [30] Dejnirattisai W, Huo J, Zhou D et al.. 2022 SARS-CoV-2 Omicron-B.1.1.529 leads to widespread escape from neutralizing antibody responses. Cell 185, 467–484.e15. * [31] Zhang L, Li Q, Liang Z et al.. 2022 The significant immune escape of pseudotyped SARS-CoV-2 variant Omicron. Emerging Microbes & Infections 11, 1–5. * [32] Ren SY, Wang WB, Gao RD, Zhou AM. 2022 Omicron variant (B.1.1.529) of SARS-CoV-2: Mutation, infectivity, transmission, and vaccine resistance. World Journal of Clinical Cases 10, 1–11. * [33] Cele S, Jackson L, Khoury DS et al.. 2021 SARS-CoV-2 Omicron has extensive but incomplete escape of Pfizer BNT162b2 elicited neutralization and requires ACE2 for infection. Nature. * [34] Zeng C, Evans JP, Qu P et al.. 2021 Neutralization and Stability of SARS-CoV-2 Omicron Variant. bioRxiv. * [35] Ai J, Zhang H, Zhang Y et al.. 2022 Omicron variant showed lower neutralizing sensitivity than other SARS-CoV-2 variants to immune sera elicited by vaccines after boost. Emerging Microbes & Infections 11, 337–343. * [36] St. Denis WFGBKJ, Hoelzemer A et al.. 2022 mRNA-based COVID-19 vaccine boosters induce neutralizing immunity against SARS-CoV-2 Omicron variant. Cell 185, 457–466.e4. * [37] Planas D, Saunders N, Maes P et al.. 2022 Considerable escape of SARS-CoV-2 Omicron to antibody neutralization. Nature 602, 671–675. * [38] Flemming A. 2022 Omicron, the great escape artist. Nature Reviews Immunology 22, 75. * [39] no JMC, Alshammary H, Tcheou J et al.. 2022 Activity of convalescent and vaccine serum against SARS-CoV-2 Omicron. Nature 602, 682–688. * [40] Yu J, Collier AY, Rowe M et al.. 2022 Neutralization of the SARS-CoV-2 Omicron BA.1 and BA.2 Variants. New England Journal of Medicine 386, 1579–1580. * [41] Arora RK, Joseph A, Wyk JV et al.. 2022 Serotracker. https://serotracker.com/en. * [42] Our World in Data. 2022 Vaccination Statistics. https://ourworldindata.org/covid-vaccinations. ## Appendix A India Phase 9 Plots All the plots in this section have $\frac{1}{P_{0}}\widetilde{{\mathcal{C}}}_{T}\widetilde{{\mathcal{T}}}$ on $x$-axis and $\widetilde{{\mathcal{T}}}-\frac{1}{\tilde{\beta}}\widetilde{{\mathcal{N}}}_{T}$ on $y$-axis. Figure 13: Phase Plot on April 21, 2021 Figure 14: Phase Plot on April 23, 2021 Figure 15: Phase Plot on April 25, 2021 Figure 16: Phase Plot on April 27, 2021 Figure 17: Phase Plot on May 07, 2021 Figure 18: Phase Plot on May 17, 2021 Figure 19: Phase Plot on May 27, 2021 Figure 20: Phase Plot on June 07, 2021 ## Appendix B Pandemic Status in Countries at the Time of Omicron Arrival In this section, we provide the data used for analysis in section 10. The table below lists, for $32$ countries, the percentage of population that had vaccine immunity (given at least one shot of vaccine fifteen days before) when Omicron arrived in the country. It also lists percentage of population with natural immunity (estimated by the model), with hybrid immunity (assuming two types of immunity are independent), with vaccine-only immunity (difference vaccine immunity and hybrid immunity), and within reach of the pandemic at the time. Table 5: Status of Pandemic at the Time of Omicron Arrival | Serosurvey | Vaccination | Natural | Hybrid | Only | Increase ---|---|---|---|---|---|--- Country | Available | Immunity % | Immunity % | Immunity % | Vaccination | in Reach % | | | | | Immunity % | Australia | N | 78.8 | 2.7 | 2.1 | 76.7 | 96.9 Bangladesh | N | 60.0 | 79.0 | 47.4 | 12.6 | 5.7 Brazil | N | 80.4 | 63.7 | 51.2 | 29.2 | 12.8 Canada | Y | 84.8 | 46.9 | 39.8 | 45.0 | 31.2 Chile | N | 90.1 | 58.3 | 52.5 | 37.6 | 31.8 Croatia | Y | 52.7 | 39.6 | 20.9 | 31.8 | 31.8 Ecuador | N | 79.9 | 62.7 | 50.1 | 29.8 | 21.8 Ethiopia | N | 7.7 | 79.2 | 6.1 | 1.6 | 17.2 France | Y | 75.6 | 34.0 | 25.7 | 49.9 | 62.2 Greece | Y | 71.2 | 29.6 | 21.1 | 50.1 | 54.7 India | Y | 60.7 | 81.1 | 49.2 | 11.5 | 11.3 Indonesia | Y | 55.5 | 80.3 | 44.6 | 10.9 | 1.7 Iran | N | 73.5 | 60.2 | 44.2 | 29.3 | 24.2 Israel | Y | 69.3 | 35.6 | 24.7 | 44.6 | 63.4 Italy | Y | 77.7 | 45.4 | 35.3 | 42.4 | 43.3 Japan | N | 80.5 | 4.5 | 3.6 | 76.9 | 84.3 Jordan | Y | 43.5 | 74.3 | 32.3 | 11.2 | 15.2 Kenya | Y | 8.7 | 82.9 | 7.2 | 1.5 | 7.8 Lithuania | Y | 60.0 | 81.6 | 49.0 | 11.0 | 24.8 Mexico | Y | 63.5 | 61.0 | 38.7 | 24.8 | 15.7 Nigeria | N | 2.8 | 87.3 | 2.4 | 0.4 | 0.3 Norway | Y | 78.2 | 4.6 | 3.6 | 74.6 | 93.3 Oman | Y | 63.1 | 73.3 | 46.3 | 16.8 | 11.9 Pakistan | N | 53.0 | 76.7 | 40.7 | 12.3 | 4.6 Philippines | N | 57.2 | 79.4 | 45.4 | 11.8 | 14.1 Portugal | Y | 89.3 | 18.5 | 16.5 | 72.8 | 62.5 Singapore | N | 86.9 | 16.1 | 14.0 | 72.9 | 73.3 South Africa | Y | 26.0 | 73.8 | 19.2 | 6.8 | 10.7 Spain | Y | 81.2 | 46.6 | 37.8 | 43.4 | 42.4 UK | Y | 78.3 | 83.3 | 65.2 | 13.1 | 3.4 US | Y | 73.4 | 55.9 | 41.0 | 32.4 | 32.5 Vietnam | N | 81.3 | 18.8 | 15.3 | 66.0 | 56.7 ## Appendix C Proofs In this Appendix, we provide proofs of all lemmas and theorems stated in the paper. * Lemma 1 When ${\mathbf{u}}$ is independent of ${\mathbf{v}}$ as well as ${\mathbf{w}}$, there is a maxima of $R^{2}$ with $R^{2}_{\beta},R^{2}_{\rho},\tilde{\beta},\tilde{\rho}>0$. ###### Proof. Let $x=\frac{1}{\tilde{\beta}}$ and $y=\frac{1}{\tilde{\rho}}$. Then we have: $\begin{split}R^{2}&=\frac{xy(2{\mathbf{v}}^{T}{\mathbf{u}}-y{\mathbf{v}}^{T}{\mathbf{w}}-x{\mathbf{v}}^{T}{\mathbf{v}})(2{\mathbf{w}}^{T}{\mathbf{u}}-x{\mathbf{w}}^{T}{\mathbf{v}}-y{\mathbf{w}}^{T}{\mathbf{w}})}{|{\mathbf{u}}-y{\mathbf{w}}|^{2}|{\mathbf{u}}-x{\mathbf{v}}|^{2}}\end{split}$ (13) with $R^{2}_{\beta}>0$ iff $2{\mathbf{w}}^{T}{\mathbf{u}}-x{\mathbf{w}}^{T}{\mathbf{v}}-y{\mathbf{w}}^{T}{\mathbf{w}}>0$ and $R^{2}_{\epsilon}>0$ iff $2{\mathbf{v}}^{T}{\mathbf{u}}-y{\mathbf{v}}^{T}{\mathbf{w}}-x{\mathbf{v}}^{T}{\mathbf{v}}>0$. The denominator of equation (13) is always positive since ${\mathbf{u}}$ is independent of ${\mathbf{v}}$ as well as ${\mathbf{w}}$. The numerator is a product of four linear terms in the unknowns $x$ and $y$. Therefore the value of $R^{2}$ is positive inside the polygon defined by: $\displaystyle x$ $\displaystyle\geq$ $\displaystyle 0$ $\displaystyle y$ $\displaystyle\geq$ $\displaystyle 0$ $\displaystyle 2{\mathbf{v}}^{T}{\mathbf{u}}-y{\mathbf{v}}^{T}{\mathbf{w}}-x{\mathbf{v}}^{T}{\mathbf{v}}$ $\displaystyle\geq$ $\displaystyle 0$ $\displaystyle 2{\mathbf{w}}^{T}{\mathbf{u}}-x{\mathbf{w}}^{T}{\mathbf{v}}-y{\mathbf{w}}^{T}{\mathbf{w}}$ $\displaystyle\geq$ $\displaystyle 0$ and is zero on the boundaries. This guarantees that there exists at least one maxima inside the polygon. ∎ * Lemma 2 Suppose $\rho_{\text{g}ain}$ is the fraction of susceptible population that became immune via vaccination, and $\rho_{\text{l}oss}$ is the fraction of immune population that lost immunity over a specified period of time. Then the new trajectory of the pandemic is obtained by multiplying both $\beta$ and $\rho$ (equivalently both $\tilde{\beta}$ and $\tilde{\rho}$) by $1+\frac{\rho_{\text{l}oss}-\rho_{\text{g}ain}}{\rho}$. ###### Proof. During the course of the pandemic, first few phases did not have any vaccination or immunity loss. Consider the first phase with either immunity loss or gain through vaccination or both. Suppose fraction of population that gains immunity through vaccination in this phase is $\rho_{\text{g}ain}$ and the fraction of population that loses immunity is $\rho_{\text{l}oss}$. Consider a time instant $t$ in the stable period of the phase when the changes in immunity have already taken place. Then the fraction of removed population would be ${\mathcal{R}}(t)+\rho_{\text{g}ain}P_{0}-\rho_{\text{l}oss}P_{0}$ where ${\mathcal{R}}(t)$ is the fraction of removed population if there was no change in immunity levels. Therefore, the fundamental equation (10) changes to: $\displaystyle{\mathcal{N}}_{T}(t+1)$ $\displaystyle=$ $\displaystyle\beta(1-\epsilon)S{\mathcal{T}}$ $\displaystyle=$ $\displaystyle\beta(1-\epsilon)(1-\frac{1}{\rho P_{0}}({\mathcal{M}}+{\mathcal{R}}+\rho_{\text{g}ain}P_{0}-\rho_{\text{l}oss}P_{0})){\mathcal{T}}$ $\displaystyle=$ $\displaystyle\beta(1-\epsilon)(1-\frac{\rho_{\text{g}ain}-\rho_{\text{l}oss}}{\rho}-\frac{1}{\rho P_{0}}({\mathcal{M}}+{\mathcal{R}})){\mathcal{T}}$ $\displaystyle=$ $\displaystyle\beta(1-\epsilon)f(1-\frac{1}{\rho fP_{0}}({\mathcal{M}}+{\mathcal{R}})){\mathcal{T}}$ $\displaystyle=$ $\displaystyle\beta(1-\epsilon)f(1-c-\frac{1}{\epsilon\rho fP_{0}}({\mathcal{T}}+{\mathcal{R}}_{T})){\mathcal{T}}$ $\displaystyle=$ $\displaystyle\tilde{\beta}f(1-\frac{1}{\tilde{\rho}fP_{0}}({\mathcal{T}}+{\mathcal{R}}_{T})){\mathcal{T}}$ where $f=1-\frac{\rho_{\text{g}ain}-\rho_{\text{l}oss}}{\rho}$. Therefore, fundamental equation now holds with values of $\beta$ and $\rho$ multiplied by $f$. The other equations of the model (11 and 12) can easily be seen to hold with the same change in values of $\beta$ and $\rho$. ∎ * Lemma 3 Suppose a new phase begins at time $t_{0}$ with a drift period of $d$ days. Further, suppose that the value of parameter $\epsilon$ changes from $\epsilon_{0}$ to $\epsilon_{1}$ during the new phase. Then, ${\mathcal{M}}(t_{0}+d)=\frac{1}{\epsilon_{1}}{\mathcal{T}}(t_{0}+d)$. ###### Proof. The model parameters $\beta(1-\epsilon)$, $\epsilon$, $\rho$ and $1-c$ change multiplicatively until they stabilize to new values. Suppose that in one day, $\beta(1-\epsilon)$ changes by a factor of $f_{b}$, $\epsilon$ by a factor of $f_{e}$, $\rho$ by a factor of $f_{r}$ and $1-c$ by a factor of $f_{c}$. Then, composite parameter $\tilde{\beta}=\beta(1-\epsilon)(1-c)$ will change by a factor of $f_{b}f_{c}$ and $\tilde{\rho}=\epsilon\rho(1-c)$ will change by a factor of $f_{e}f_{r}f_{c}$. We know that after the drift period, $\tilde{\beta}$ and $\tilde{\rho}$ stabilize. Suppose that $\epsilon$ continues to change even after the drift period of $d$ days and stabilizes after $D$ days in the phase. We now consider following cases: * • $\tilde{\beta}$ changes during the drift period and $\beta(1-\epsilon)$ changes on day $D-1$. In that case, change in $\tilde{\beta}$ on day $D-1$ equals $f_{b}f_{c}$ which is not equal to $1$ since $\tilde{\beta}$ changes during drift period. This is not possible. * • $\tilde{\beta}$ changes during the drift period and $\beta(1-\epsilon)$ does not change on day $D-1$. Then change in $\tilde{\beta}$ on day $D-1$ equals $f_{c}$, and since $\tilde{\beta}$ does not change on day $D-1$, $f_{c}=1$. Computation leading up to derivation of equation (15), as in the proof of Theorem 1, can be carried out for day $D$ (since all parameters stabilize by then), and substituting $y=\frac{1}{f_{c}}=1$ in the equation (15) we get $f_{e}=x=1$ (note that $\tilde{\rho}_{D}=\tilde{\rho}_{D-1}$). This contradictions the assumption that $\epsilon$ changes during the phase. * • $\tilde{\rho}$ changes during drift period and $\rho$ changes on day $D-1$. Then change in $\tilde{\rho}$ on day $D-1$ equals $f_{r}f_{e}f_{c}$ which is not equal to $1$ since $\tilde{\rho}$ changes during drift period. This is not possible. * • $\tilde{\rho}$ changes during drift period and $\rho$ does not change on day $D-1$. Then change in $\tilde{\rho}$ on day $D-1$ equals $f_{e}f_{c}$ which must be equal to $1$ since $\tilde{\rho}$ does not change after $d$ days. Therefore, $f_{e}=1/f_{c}$. Going back to equation (15) and substituting $x=f_{e}=1/f_{c}=y$, we again get $f_{e}=1=f_{c}$, contradicting the assumption that $\epsilon$ changes during the phase. Together, the cases above cover all possibilities and hence we conclude that $D=d$ and therefore, ${\mathcal{M}}(t_{0}+d)=\frac{1}{\epsilon_{1}}{\mathcal{T}}(t_{0}+d)$. ∎ * Lemma 4 Given detected new cases trajectory, ${\mathcal{N}}_{T}(t)$, $0\leq t\leq t_{F}$, there exist infinitely many total new cases trajectories and corresponding values of $\epsilon$ consistent with ${\mathcal{N}}_{T}$. ###### Proof. Given ${\mathcal{N}}_{T}(t)$, $0\leq t\leq t_{F}$, we can compute phases of the trajectory and values of $\tilde{\beta}$ and $\tilde{\rho}$ for all phases as shown in section 5, as well as ${\mathcal{C}}_{T}(t)$, ${\mathcal{T}}(t)$ and ${\mathcal{R}}_{T}(t)$ for the entire duration. Choose any value $\epsilon_{0}$ in the range $[0.9,1.0]$. Fix $\epsilon=\epsilon_{0}$ and $c=0$. This allows us to compute the values of $\beta$ and $\rho$ for all phases (the value of $\rho$ will be at most $\frac{1}{0.9}$ times the value of $\tilde{\rho}$ at any time). Let ${\mathcal{N}}(t)=\frac{1}{\epsilon}{\mathcal{N}}_{T}(t)$, and ${\mathcal{M}}(t)=\frac{1}{\epsilon}{\mathcal{T}}(t)$ for $0\leq t\leq t_{F}$. Setting ${\mathcal{R}}(0)=0$, and using the equation (12) for $R$, we get that ${\mathcal{R}}(t)=\gamma\sum_{s=0}^{t-1}{\mathcal{M}}(s)=\frac{\gamma}{\epsilon}\sum_{s=0}^{t-1}{\mathcal{T}}(s)=\frac{1}{\epsilon}{\mathcal{R}}_{T}(t)$. Then, for all $t$, $0\leq t\leq t_{F}$: $\displaystyle{\mathcal{N}}(t)$ $\displaystyle=$ $\displaystyle\frac{1}{\epsilon}{\mathcal{N}}_{T}(t)$ $\displaystyle=$ $\displaystyle\frac{1}{\epsilon}\tilde{\beta}(1-\frac{1}{\tilde{\rho}P_{0}}({\mathcal{T}}(t-1)+{\mathcal{R}}_{T}(t-1))){\mathcal{T}}(t-1)$ $\displaystyle=$ $\displaystyle\tilde{\beta}(1-\frac{1}{\rho P_{0}}({\mathcal{M}}(t-1)+{\mathcal{R}}(t-1))){\mathcal{M}}(t-1)$ $\displaystyle=$ $\displaystyle\beta(1-\epsilon)S(t-1){\mathcal{M}}(t-1).$ It is straightforward to see that the remaining model equations (12) are also satisfied. As there are infinitely many values in the range $[0.9,1.0]$, the proof is complete. ∎ * Theorem 1 Given detected new cases trajectory, ${\mathcal{N}}_{T}(t)$, $0\leq t\leq t_{F}$ and ${\mathcal{C}}(0)$, there exist only finitely many trajectories for ${\mathcal{N}}(t)$ consistent with ${\mathcal{N}}_{T}$. Further, a good estimate for all the trajectories can be obtained efficiently. ###### Proof. Proof is by induction on the number of phases. In the base case we have only one phase. For this phase, there is no drift period since there are no previous values of parameters. Therefore, parameter values stay the same throughout the phase duration. Let $\beta_{1}$, $\rho_{1}$, $\epsilon_{1}$ and $c_{1}$ be the parameter values governing the actual trajectory for this phase. Therefore, ${\mathcal{M}}(t)=\frac{1}{\epsilon_{1}}{\mathcal{T}}(t)$ and ${\mathcal{R}}(t)=\frac{1}{\epsilon_{1}}{\mathcal{R}}_{T}(t)+c_{1}\rho_{1}P_{0}$ for the entire phase. Note that ${\mathcal{R}}(0)=0$ since at time $t=0$, when the pandemic starts, there are no recoveries. Hence, $c_{1}=0$. Further, $\epsilon_{1}={\mathcal{T}}(0)/{\mathcal{M}}(0)={\mathcal{T}}(0)/{\mathcal{C}}(0)$. From this, we can compute $\displaystyle\rho_{1}$ $\displaystyle=$ $\displaystyle\frac{1}{\epsilon_{1}}\tilde{\rho}_{1}$ $\displaystyle\beta_{1}$ $\displaystyle=$ $\displaystyle\tilde{\beta}_{1}/(1-\epsilon_{1})$ giving values of all parameters for first phase, using which the trajectory can be computed for the first phase uniquely. Suppose there are finitely trajectories up to phase $i-1$. Fix any one trajectory with values of four parameters in phase $i-1$ being $\beta_{0}$, $\rho_{0}$, $\epsilon_{0}$ and $c_{0}$. Let $t_{0}$ be the time when phase $i$ starts. We have: $\begin{bmatrix}{\mathcal{M}}(t_{0})\\\ {\mathcal{C}}(t_{0})\end{bmatrix}=\frac{1}{\epsilon_{0}}\begin{bmatrix}{\mathcal{T}}(t_{0})\\\ {\mathcal{C}}_{T}(t_{0})\end{bmatrix}+c_{0}\rho_{0}\begin{bmatrix}0\\\ P_{0}\end{bmatrix}.$ Suppose phase $i$ has a drift period of $d$ days. In the model, the parameter values change multiplicatively during the drift period. Let $\epsilon_{j}=\epsilon_{0}x^{j}$, and $1-c_{j}=(1-c_{0})/y^{j}$ for $1\leq j\leq d$, where $x$ and $y$ are unknown multipliers by which the two parameters change every day. The final value of the parameters will be $\epsilon_{d}=\epsilon_{0}x^{d}$ and $1-c_{d}=(1-c_{0})/y^{d}$. Let $\tilde{\beta}_{j}=\tilde{\beta}_{0}(\frac{\tilde{\beta}_{d}}{\tilde{\beta}_{0}})^{j/d}$ and $\tilde{\rho}_{j}=\tilde{\rho}_{0}(\frac{\tilde{\rho}_{d}}{\tilde{\rho}_{0}})^{j/d}$, for $1\leq j\leq d$. These numbers can be computed since $\tilde{\beta}_{0}$, $\tilde{\beta}_{d}$, $\tilde{\rho}_{0}$, and $\tilde{\rho}_{d}$ are known. Let $\beta_{j}=\frac{\tilde{\beta}_{j}}{(1-\epsilon_{j})(1-c_{j})}$ and $\rho_{j}=\frac{\tilde{\rho}_{j}}{\epsilon_{j}(1-c_{j})}$ for $1\leq j\leq d$. Then we can write: $\begin{bmatrix}{\mathcal{M}}(t_{0}+j)\\\ {\mathcal{R}}(t_{0}+j)\end{bmatrix}=\begin{bmatrix}g_{j}&0\\\ \gamma&1\end{bmatrix}\cdot\begin{bmatrix}{\mathcal{M}}(t_{0}+j-1)\\\ {\mathcal{R}}(t_{0}+j-1)\end{bmatrix}$ where $\displaystyle g_{j}$ $\displaystyle=$ $\displaystyle\beta_{j-1}(1-\epsilon_{j-1})(1-\frac{{\mathcal{M}}(t_{0}+j-1)+{\mathcal{R}}(t_{0}+j-1)}{\rho_{j-1}P_{0}})-\gamma+1$ $\displaystyle=$ $\displaystyle\tilde{\beta}_{j-1}(\frac{1}{1-c_{j-1}}-\frac{\epsilon_{j-1}}{\tilde{\rho}_{j-1}}\frac{{\mathcal{M}}(t_{0}+j-1)+{\mathcal{R}}(t_{0}+j-1)}{P_{0}})-\gamma+1$ $\displaystyle=$ $\displaystyle\tilde{\beta}_{j-1}(\frac{y^{j-1}}{1-c_{0}}-\frac{\epsilon_{0}x^{j-1}}{\tilde{\rho}_{j-1}}\frac{{\mathcal{M}}(t_{0}+j-1)+{\mathcal{R}}(t_{0}+j-1)}{P_{0}})-\gamma+1$ Therefore, both ${\mathcal{M}}(t_{0}+j)$ and ${\mathcal{R}}(t_{0}+j)$ are polynomials in $x$ and $y$. It is straightforward to show that the degrees of ${\mathcal{M}}(t_{0}+j)$ and ${\mathcal{R}}(t_{0}+j)$ equal $2^{j}-j-1$ and $2^{j-1}-j-2$ respectively. At the end of drift period, we have: $\displaystyle\begin{bmatrix}{\mathcal{M}}(t_{0}+d)\\\ {\mathcal{R}}(t_{0}+d)\end{bmatrix}$ $\displaystyle=$ $\displaystyle\frac{1}{\epsilon_{d}}\begin{bmatrix}{\mathcal{T}}(t_{0}+d)\\\ {\mathcal{R}}_{T}(t_{0}+d)\end{bmatrix}+c_{d}\rho_{d}\begin{bmatrix}0\\\ P_{0}\end{bmatrix}$ (14) $\displaystyle=$ $\displaystyle\frac{1}{\epsilon_{d}}\begin{bmatrix}{\mathcal{T}}(t_{0}+d)\\\ {\mathcal{R}}_{T}(t_{0}+d)\end{bmatrix}+c_{d}\frac{\tilde{\rho}_{d}}{\epsilon_{d}(1-c_{d})}\begin{bmatrix}0\\\ P_{0}\end{bmatrix}$ $\displaystyle=$ $\displaystyle\frac{1}{\epsilon_{0}x^{d}}\begin{bmatrix}{\mathcal{T}}(t_{0}+d)\\\ {\mathcal{R}}_{T}(t_{0}+d)\end{bmatrix}+\frac{\tilde{\rho}_{d}}{\epsilon_{0}x^{d}}(\frac{y^{d}}{1-c_{0}}-1)\begin{bmatrix}0\\\ P_{0}\end{bmatrix}\mbox{~{}~{}~{}~{}~{}~{}}$ For what values of unknown multipliers $x$ and $y$ are the relationships in equation (14) satisfied? To see this, we analyze the quantities ${\mathcal{N}}(t_{0}+d)$, ${\mathcal{M}}(t_{0}+d)$, and ${\mathcal{C}}(t_{0}+d)$. From (12) we have: $\displaystyle{\mathcal{C}}(t_{0}+d)$ $\displaystyle=$ $\displaystyle{\mathcal{C}}(t_{0}+d-1)+{\mathcal{N}}(t_{0}+d)$ $\displaystyle{\mathcal{M}}(t_{0}+d)$ $\displaystyle=$ $\displaystyle{\mathcal{N}}(t_{0}+d)+(1-\gamma){\mathcal{M}}(t_{0}+d-1)$ $\displaystyle{\mathcal{N}}(t_{0}+d)$ $\displaystyle=$ $\displaystyle\beta_{d-1}(1-\epsilon_{d-1})S(t_{0}+d-1){\mathcal{M}}(t_{0}+d-1)$ $\displaystyle=$ $\displaystyle\beta_{d-1}(1-\epsilon_{d-1})\left(1-\frac{1}{\rho_{d-1}P_{0}}{\mathcal{C}}(t_{0}+d-1)\right){\mathcal{M}}(t_{0}+d-1)$ Therefore, $\displaystyle{\mathcal{C}}(t_{0}+d-1)$ $\displaystyle=$ $\displaystyle\frac{1}{\epsilon_{d}}{\mathcal{C}}_{T}(t_{0}+d)+c_{d}\rho_{d}P_{0}-\frac{1}{\epsilon_{d}}{\mathcal{N}}_{T}(t_{0}+d)$ $\displaystyle=$ $\displaystyle\frac{1}{\epsilon_{d}}{\mathcal{C}}_{T}(t_{0}+d-1)+c_{d}\rho_{d}P_{0}$ $\displaystyle{\mathcal{M}}(t_{0}+d-1)$ $\displaystyle=$ $\displaystyle\frac{1}{\epsilon(1-\gamma)}({\mathcal{T}}(t_{0}+d)-{\mathcal{N}}(t_{0}+d))$ $\displaystyle=$ $\displaystyle\frac{1}{\epsilon}{\mathcal{T}}(t_{0}+d-1)$ $\displaystyle{\mathcal{N}}(t_{0}+d)$ $\displaystyle=$ $\displaystyle\beta_{d-1}(1-\epsilon_{d-1})\left(1-\frac{1}{\epsilon_{d}\rho_{d-1}P_{0}}{\mathcal{C}}_{T}(t_{0}+d-1)-\frac{c_{d}\rho_{d}}{\rho_{d-1}}\right)\frac{1}{\epsilon_{d}}{\mathcal{T}}(t_{0}+d-1)$ $\displaystyle=$ $\displaystyle\frac{\tilde{\beta}_{d-1}}{\epsilon_{d}(1-c_{d-1})}\left(1-\frac{1}{\epsilon_{d}\rho_{d-1}P_{0}}{\mathcal{C}}_{T}(t_{0}+d-1)-\frac{c_{d}\rho_{d}}{\rho_{d-1}}\right){\mathcal{T}}(t_{0}+d-1).$ Since ${\mathcal{N}}(t_{0}+d)=\frac{1}{\epsilon_{d}}{\mathcal{N}}_{T}(t_{0}+d)=\frac{\tilde{\beta}_{d-1}}{\epsilon_{d}}\left(1-\frac{1}{\tilde{\rho}_{d-1}P_{0}}{\mathcal{C}}_{T}(t_{0}+d-1)\right){\mathcal{T}}(t_{0}+d-1)$ where the second equality is from fundamental equation (10), we have: $\displaystyle 1-\frac{1}{\tilde{\rho}_{d-1}P_{0}}{\mathcal{C}}_{T}(t_{0}+d-1)$ $\displaystyle=$ $\displaystyle\frac{1}{1-c_{d-1}}\left(1-\frac{1}{\epsilon_{d}\rho_{d-1}P_{0}}{\mathcal{C}}_{T}(t_{0}+d-1)-\frac{c_{d}\rho_{d}}{\rho_{d-1}}\right)$ $\displaystyle=$ $\displaystyle\frac{1}{1-c_{d-1}}-\frac{c_{d}\rho_{d}}{(1-c_{d-1})\rho_{d-1}}-\frac{1}{\epsilon_{d}\rho_{d-1}(1-c_{d-1})P_{0}}{\mathcal{C}}_{T}(t_{0}+d-1)$ $\displaystyle=$ $\displaystyle\frac{y^{d-1}}{1-c_{0}}-\frac{\epsilon_{d-1}c_{d}\rho_{d}}{\tilde{\rho}_{d-1}}-\frac{1}{x\tilde{\rho}_{d-1}P_{0}}{\mathcal{C}}_{T}(t_{0}+d-1)$ $\displaystyle=$ $\displaystyle\frac{y^{d-1}}{1-c_{0}}-\frac{c_{d}\tilde{\rho}_{d}}{x(1-c_{d})\tilde{\rho}_{d-1}}-\frac{1}{x\tilde{\rho}_{d-1}P_{0}}{\mathcal{C}}_{T}(t_{0}+d-1)$ $\displaystyle=$ $\displaystyle\frac{y^{d-1}}{1-c_{0}}+\frac{\tilde{\rho}_{d}}{x\tilde{\rho}_{d-1}}-\frac{y^{d}\tilde{\rho}_{d}}{x(1-c_{0})\tilde{\rho}_{d-1}}-\frac{1}{x\tilde{\rho}_{d-1}P_{0}}{\mathcal{C}}_{T}(t_{0}+d-1).$ Therefore, $\displaystyle x$ $\displaystyle=$ $\displaystyle\frac{\left(\frac{\tilde{\rho}_{d}}{\tilde{\rho}_{d-1}}-\frac{y^{d}\tilde{\rho}_{d}}{(1-c_{0})\tilde{\rho}_{d-1}}-\frac{1}{\tilde{\rho}_{d-1}P_{0}}{\mathcal{C}}_{T}(t_{0}+d-1)\right)}{\left(1-\frac{y^{d-1}}{1-c_{0}}-\frac{1}{\tilde{\rho}_{d-1}P_{0}}{\mathcal{C}}_{T}(t_{0}+d-1)\right)}.$ (15) Equation (15) expresses $x$ as a degree $d$ rational function of $y$. Substituting this in the polynomial ${\mathcal{M}}(t_{0}+d)$, we obtain a degree $d(2^{d}-d-1)$ rational function in $y$. Equating it to $\frac{1}{\epsilon_{d}}{\mathcal{T}}(t_{0}+d)=\frac{1}{\epsilon_{0}x^{d}}{\mathcal{T}}(t_{0}+d)$ results in a polynomial in $y$ of degree bounded by $d(2^{d}-d)$, say $P(y)$, whose roots are the possible values of $y$. Therefore, there are at most $d(2^{d}-d)$ many values of $(x,y)$ that satisfy all the required equations. While the potential number of solutions is very large for even moderate values of $d$ (say $d>20$), most of the solutions are likely to be infeasible since a feasible solution needs to satisfy the conditions that $\epsilon_{d}$ must be in the range $[0,1]$ and $c_{d}$ in the range $[-1,1]$. Further, there is a simple and efficient way to list out good estimates for all the solutions: since $-1<c_{d}<1$, one can step through possible values of $c_{d}$ in the range using small discrete steps, compute the value of $y$ for the chosen value of $c_{d}$, compute value of $x$ using equation (15), and then check if $P(y)$ is close to zero. This will list out good estimates of all feasible solutions. ∎
# Topology and complexity of the hydrogen bond network in classical models of water Fausto Martelli IBM Research Europe, Hartree Centre, Daresbury, WA4 4AD, United Kingdom<EMAIL_ADDRESS>Department of Physics and CNR Institute of Complex Systems, Sapienza University of Rome, P.le Aldo Moro 5, 00185 Roma, Italy<EMAIL_ADDRESS> ###### Abstract Over the years, plenty of classical interaction potentials for water have been developed and tested against structural, dynamical and thermodynamic properties. On the other hands, it has been recently observed (F. Martelli et. al, ACS Nano, 14, 8616–8623, 2020) that the topology of the hydrogen bond network (HBN) is a very sensitive measure that should be considered when developing new interaction potentials. Here we report a thorough comparison of 11 popular non polarizable classical water models against their HBN, which is at the root of water properties. We probe the topology of the HBN using the ring statistics and we evaluate the quality of the network inspecting the percentage of broken and intact HBs. For each water model, we assess the tendency to develop hexagonal rings (that promote crystallization at low temperatures) and pentagonal rings (known to frustrate against crystallization at low temperatures). We then introduce the _network complexity index_ , a general descriptor to quantify how much the topology of a given network deviates from that of the ground state, namely of hexagonal or cubic ice. Remarkably, we find that the network complexity index allows us to relate, for the first time, the dynamical properties of different water models with their underlying topology of the HBN. Our study provides a benchmark against which the performances of new models should be tested against, and introduces a general way to quantify the complexity of a network which can be transferred to other materials and that links the topology of the HBN with dynamical properties. Finally, our study introduces a new perspective that can help in rationalizing the transformations among the different phases of water and of other materials. Water models, Classical potentials, Network topology, Network complexity, Hydrogen bonds ††preprint: AIP/123-QED ## I Introduction On our planet, water is the only substance that can be found co-existing in the solid, liquid and vapour phases outside research laboratories. Its molecular simplicity hides a remarkably wide list of anomalous behaviors that stretch over the the most complex phase diagram of any pure substance Salzmann (2019), and whose origin lies in a critical point located at low temperatures and low pressures Palmer _et al._ (2014); Sellberg _et al._ (2014); Debenedetti, Sciortino, and Zerze (2020); Kringle _et al._ (2020); Kim _et al._ (2020). Nonetheless, water plays a primary role in many industrial, biological and geological processes. Therefore, there is a great interest in developing classical interaction potentials able to embrace water’s complex nature. This intent is aggravated by the wide span of thermodynamic conditions at which water exists (from low temperatures and low pressures of the interstellar medium to high pressures and high temperatures in the core of planets), and by the broad range of timescales required for several processes to occur (from heterogeneous nucleation to protein folding to geological processes). After a long-lasting debate Poole _et al._ (1992); Liu _et al._ (2010); Limmer and Chandler (2011); Wikfeldt, Nilsson, and Pettersson (2011); Palmer, Car, and Debenedetti (2013); Limmer and Chandler (2013); Palmer _et al._ (2014); Limmer and Chandler (2014); Chandler (2016); Palmer _et al._ (2016a, 2018, b), it is now accepted that liquid water is a mixture Palmer _et al._ (2014); Sellberg _et al._ (2014); Palmer _et al._ (2018); Debenedetti, Sciortino, and Zerze (2020); Kim _et al._ (2020); Kringle _et al._ (2020) of molecules whose local neighborhood constantly change between an ordered tetrahedral state with local lower density, and a more distorted tetrahedral state with local higher density Martelli _et al._ (2020); Shi and Tanaka (2020a, b); Shi, Russo, and Tanaka (2018a); Russo and Tanaka (2014a); Akahane and Tanaka (2018); Santra _et al._ (2015); Huang _et al._ (2009); Nilsson and Pettersson (2015); Wikfeldt, Nilsson, and Pettersson (2011); De Marzio _et al._ (2017); Martelli (2019). These two local structures continuously interchange with each other, giving rise to a complex network of bonds that actively determines the properties of water at the macro scale Martelli (2019). The percentage of such local environments depends on the thermodynamic conditions Martelli (2019); Wikfeldt, Nilsson, and Pettersson (2011), and eventually liquid water becomes a 1:1 mixture at low temperatures and pressures Palmer _et al._ (2014); Debenedetti, Sciortino, and Zerze (2020). Classical interaction potentials have been developed over the years to reproduce experimental structural, dynamical and/or thermodynamic properties of water (over a relatively reduced set of thermodynamic points), and several studies have compared them against a plethora of observables Palmer _et al._ (2016b); Pekka and Lennart (2001); Mao and Zhang (2012); Lee and Kim (2019); Jorgensen _et al._ (1983); Harrach and Drossel (2014); González _et al._ (2010); Zielkiewicz (2005); Steinczinger, Jóvári, and Pusztai (2017); Dix, Lue, and Carbone (2018). On the other hand, even though Bernal and Fowler first recognized almost 100 years ago that water molecules build a complex network of bonds that is at the heart of water’s anomalous behavior Bernal and Fowler (1933), the pivotal role of the hydrogen bond network (HBN) has so far been showcased only in a handful of important cases involving transformations between complex phases of water Tse _et al._ (1999); Marton̆ák, Donadio, and Parrinello (2004, 2005); Palmer _et al._ (2014); Shephard _et al._ (2017); Martelli _et al._ (2018); Martelli (2019), the mutual interactions between water and biological membranes Martelli, Crain, and Franzese (2020), and between water and graphene sheets for technological purposes Chiricotto _et al._ . In particular, the inspection of the topology of the HBN in biological environments is opening new avenues in enhancing the efficacy of new drugs/vaccines Martelli, Calero, and Franzese (2021). The HBN has never been used as a metric to test water models. Two reasons for this lack of comparison are: (i) it is very hard to have a direct experimental description of the HBN in liquid phases, and (ii) it is hard to probe the HBN from a computational perspective. Here we fill the gap of the point (ii) above by comparing the HBN of 11 popular and widely adopted classical water models at ambient conditions. We study the TIP3P Jorgensen _et al._ (1983) the SPC Berendsen _et al._ (1981), the SPC/E Berendsen, Grigera, and Straatsma (1987) and the flexible SPC Toukan and Rahman (1985); Amira, Spángberg, and Hermansson (2004) as 3-points models; the TIP4P Jorgensen _et al._ (1983), the TIP4P-Ice Abascal _et al._ (2005), the TIP4P/2005 Abascal and Vega (2005), the flexible TIP4P/2005 González and Abascala (2011) and the TIP4P-Ew Horn _et al._ (2004) as 4-points models; the TIP5P Mahoney and Jorgensen (2000) and the TIP5P-Ew Rick (2004) as 5-points models. We probe the topology of the HBN using the ring statistics, a theoretical tool that has seen increasingly high relevance in determining the properties of bulk water Martelli _et al._ (2016, 2018); Martelli (2019); Formanek and Martelli (2020); Palmer _et al._ (2014); Santra _et al._ (2015); Leoni _et al._ (2019); Camisasca _et al._ (2019); Marton̆ák, Donadio, and Parrinello (2004, 2005); Russo and Tanaka (2014b); Palmer _et al._ (2014); Fitzner _et al._ (2019); Shi and Tanaka (2018), of aqueous solutions Bakó _et al._ (2017); Pothoczki, Pusztai, and Bakó (2018, 2019); Li _et al._ (2020a) and of water under confinement Martelli, Crain, and Franzese (2020); Chiricotto _et al._ . We then measure the quality of the HBN for each water model in terms of broken and intact HBs, a measure intimately linked to the fluidity and tetrahedrality of water DiStasio Jr. _et al._ (2014); Martelli, Crain, and Franzese (2020). We also introduce the concept of _network complexity index_ $\xi$ that allows to quantify how much a given HBN deviates from the HBN of water in the crystalline phase at low temperatures and ambient pressure, i.e., hexagonal or cubic ice. As a showcase, we apply this index $\xi$ to one of the rings definition and counting scheme here adopted. Remarkably, the index $\xi$ allows us to link dynamical properties of water with the topology of the underlying HBN and the corresponding structural properties. Finally, we show that the HBN topology is more sensitive to the size of the simulation box with respect to, e.g., structural properties measured _via_ the two-bodies pair correlation function. Therefore, the inspections of the network topology should always be considered when simulating network-forming materials. This issue becomes particularly relevant when dealing with _ab initio_ molecular dynamics simulations which are restricted to small simulation cells. The article is organized as follows. In Section II we report the details of the numerical simulations, the ring definitions and counting schemes, and we introduce the network complexity index. In Section III we report our main findings for all 11 water models here inspected. Conclusions and final remarks are reported in Section IV. ## II Computational details In this section we describe the numerical setup, the protocols implemented to count rings and we introduce the definition of the network complexity index. ### II.1 Numerical simulations Our study is based on classical molecular dynamics simulations of systems composed of $N=1100$ water molecules described by different 11 interaction potentials in the isobaric ($N$$p$$T$) ensemble. We have employed Nosé-Hoover thermostat Nosé (1984); Hoover (1985) with 0.2 ps relaxation time to maintain constant temperature at $T=300$ K, and Parrinello-Rahman barostat Parrinello and Rahman (1981) with 2 ps relaxation time to maintain constant pressure at 1 bar. We have performed simulations with the GROMACS 18.0.1 package Abraham _et al._ (2015). All simulations have been equilibrated for 1 ns. The production runs achieved 3 ns. For each water model we have averaged over 10 independent trajectories. Our analysis investigates the properties of liquid water in the presence of thermal noise. ### II.2 Ring statistics and network complexity index In order to compute the ring statistics it is necessary to follow two steps. First of all, it is necessary to define the link between atoms/molecules. Possible definitions can be based on the formation of bonds, interaction energies, geometric distances, etc.. The second step is the definition of ring and the corresponding counting scheme. This task is of particular relevance in directional networks, like water or silica, were the donor/acceptor nature of the bonds breaks the symmetry in the linker search path. Several definitions of rings and counting schemes have been reported in the literature King (1967); Rahman and Stillinger (1973); Guttman (1990); Franzblau (1991); Wooten (2002); Yuan and Cormack (2002); Roux and Jund (2010). Such different definitions have yielded to different interpretations of numerical results even for the simplest crystalline structures, not to mention more complicated networks such as amorphous silicate structures Jin _et al._ (1994); Hobbs _et al._ (1998); Guttman (1990); Marians and Hobbs (1988, 1990); Yuan and Cormack (2002) and water Fitzner _et al._ (2019); Camisasca _et al._ (2019); Martelli (2019); Santra _et al._ (2015). In the case of water, these inconsistencies have been recently reconciled by Formanek and Martelli showing that they were caused by different counting schemes Formanek and Martelli (2020). Here, we report a thorough analysis of the HBN for the 11 classical non- polarizable models of water. This benchmark study adopts the three ring definitions and counting schemes reported in Ref. Formanek and Martelli (2020). The definition of HB follows the geometric construction described in Ref. Luzar and Chandler (1996). In this regard, any quantitative measure of HBs in liquid water is somewhat ambiguous, since the notion of an HB itself is not uniquely defined. However, qualitative agreement between the definition here adopted and several other proposed definitions have been deemed satisfactory over a wide range of thermodynamic conditions, including the one here investigated Prada-Gracia, Shevchuk, and Rao (2013); Shi, Russo, and Tanaka (2018b). We construct rings by starting from a tagged water molecule (marked as 1 in fig. 1) and recursively traversing the HBN until the starting point is reached again or the path exceeds the maximal ring size considered (12 water molecules in our case). In the first scheme sketched in fig. 1 a), molecule 1 _donates_ a HB emphasized by the red arrow, and we restrict our counting only to the shortest rings King (1967), i.e., to rings that can not be further decomposed into smaller ones (the 6-folded ring in this case). We will refer to this scheme to as d1. This scheme emphasizes the directional nature of the HBs resulting in an enhanced hexagonal character of the HBN Formanek and Martelli (2020). It is therefore well suited for investigating phenomena such as nucleation, where hexagonal rings are the elemental building blocks of the final, crystal network. In the second scheme sketched in fig. 1 b), we consider only rings formed when molecule 1 _accepts_ a HB and we restrict our counting only to the shortest ring King (1967) (the 5-folded ring in this case). We will refer to this scheme to as d2. This scheme emphasizes the formation of pentagonal rings in the network Formanek and Martelli (2020), which are known to play an important role at supercooled conditions frustrating against crystallization Russo and Tanaka (2014a); Shi, Russo, and Tanaka (2018a); Martelli (2019), as well as in promoting the crystallization of clathrate structures Li _et al._ (2020b). These two counting schemes d1 and d2 provide drastically different distributions Formanek and Martelli (2020) because of the different energies involved in accepting and donating a HB. As we will show in the forthcoming discussion, such difference translates into distinct percentages of coordination defects of the kind $\textit{A}_{2}\textit{D}_{1}$ and $\textit{A}_{1}\textit{D}_{2}$, where $\textit{A}_{x}\textit{D}_{y}$ indicates that a water molecules accepts $x$ and donates $y$ HBs. In the third scheme sketched in fig. 1 c) we loosen the previous restrictions. Molecule 1 can now either accept or donate a HB, hence counting both the 6-folded and the 5-folded ring. In the following, we will refer to this counting scheme to as d3. As shown in Ref. Formanek and Martelli (2020), these three different counting schemes carry different, but complementary physical information and, therefore, allow us to make a proper comparison on the topology of the HBN generated by different classical models of water. Figure 1: Schematic representation of the three counting schemes. Red filled circles represent oxygen atoms, while red empty circles represent hydrogen atoms. The network of HBs is represented by the arrows Water molecule labeled as 1 is the starting molecule from which rings are counted. a): Molecule donates a HB (red arrow) and the shortest ring is the hexagonal one. b): Molecule 1 can accept a HB (one of the two red arrows) and the shortest ring is the pentagonal one c): Molecule 1 can either donate one HB or accept a HB, generating both the short hexagonal and pentagonal ring. When studying the topology of a network via the ring statistics, the probability distribution P(n) of having a n-folded ring is a normalized quantity that does not reflect the overall number of each n-folded ring. On the other hand, the actual number of rings is important to understand the degree of complexity in a network (for a given number of molecules). The stable crystalline phase of water at ambient pressure is the hexagonal(cubic) ice, Ih(c), characterized by an hexagonal network of bonds. In a sample of liquid water at higher temperatures, the thermal noise allows water molecules to explore a larger configurational space and, hence, the underlying HBN hosts also shorter and longer rings. Intuitively, such network is more ”complex” with respect to the HBN of the ground state, and its fluctuations are related to the dynamical properties (translational and rotational diffusion) of water molecules. Knowing the symmetry of the HBN at the ground state, we can measure the deviation from it. We introduce the _network complexity index_ $\xi$ defined as the ratio between the number of 6-folded rings and the total number of rings: $\xi=\frac{n_{6}}{\sum_{i=3}^{12}n_{i}}$ (1) where $n_{i}$ is the number of the $i-$th ring. The sum on the denominator of eq. 1 runs from n=3, the shortest ring length possible, to n=12, the longest ring length here considered. The network complexity index $\xi$ encodes the number of rings in a given network. For the ground state Ih(c), the network complexity index is trivially $\xi=1$. We can define the case of maximal disorder in the network when each ring length occur with the same frequency in the network. This case corresponds to a flat distribution in P(n) and, since in our case we consider 10 possible ring lengths (from n=3 to n=12), $\xi=0.1$. Here, we compute $\xi$ for the counting scheme d3, scheme c) in fig. 1 which, compared to the other schemes, accounts for the presence in the network of longer rings. The definition of network complexity index can be extended to other networks with different ground states (i.e., non-hexagonal networks), in which case the nominator in eq. 1 will be $n_{j}$, $j$ being the length of the characteristic ring length at the ground state. However, when comparing the index $\xi$ in different simulations, care must be taken in order to ensure that (i) the same definition of ring and counting schemes are used, (ii) the same maximum search path (or maximal ring size) is implemented. It is worthy to remark, at this point, that our analysis occurs in the presence of thermal noise. As recently shown by Montes de Oca et al., the percentage of broken and intact bonds drastically changes upon removal of the thermal noise de Oca _et al._ (2020), i.e., in correspondence with the inherent potential energy surfaces (IPES). While we do expect quantitative differences in correspondence with the IPES with respect to the results that we will show in Section III, we are confident that the overall trend should be preserved. ## III Results In this section, we report and discuss our results in terms of topology and complexity of the HBNs, as well as their quality in terms of broken and intact HBs for the water 11 models. We then compare the network complexity indices and we relate them to the dynamical properties of each water model. Finally, we show and discuss the effects of the simulation box size on the topology of the HBN. ### III.1 3 points models we start our analysis by comparing, in fig. 2, the oxygen-oxygen two-bodies pair correlation functions g2(r) for 3-points models with the g2(r) obtained from various scattering experiments Skinner _et al._ (2013); Soper and Benmore (2008) (open circles and squares) and _ab initio_ molecular dynamics (AIMD) simulations DiStasio Jr. _et al._ (2014) (open triangles) at the PBE0 level of theory accounting for non-local van der Waals/dispersion interactions DiStasio Jr. _et al._ (2014). It is worth to mention, at this point, that the PBE0+vdw level of theory gives an accurate g2(r) compared to the experimental one, but does not capture the correct density difference between liquid water at ambient conditions and Ih. We can observe that the TIP3P model (black line) has the first peak at $\sim 2.8$ nm with intensity comparable with that of scattering experiments and _ab initio_ simulations. On the other hand, after the first minimum the distribution is almost flat with no peaks at larger distances. This is indicative of a lack of structurization at larger distances. The SPC model (red line) is able to perform better than the TIP3 model, with a g2(r) showing hints of a second and a third peak, though shifted compared to the experimental and _ab initio_ g2(r). On the other hand, the intensity of the first peak overcomes the experimental and AIMD first peak. The SPC/E model (green line) results in a better g2(r) in terms of intensity and position of the second and third peak with respect to the experimental and the _ab initio_ g2(r), reflecting the better performances in density and diffusion constant than the SPC model Berendsen, Grigera, and Straatsma (1987). On the other hand, the intensity of the first peak is further enhanced. The flexible SPC water model (blue line) is a re-parametrization of the SPC water model in which the O–H stretching is made anharmonic, and thus the dynamical behavior is well described and bulk density and permettivity are correct Praprotnik, Janežič, and Mavri (2004). The g2(r) of the flexible SPC model is characterized by a high intensity first peak and a deeper first minimum with respect to the rigid SPC model, and correctly captures the profile of the g2(r) at larger distances. Overall, the sequence TIP3, SPC, SPC/E and flexible SPC is characterized by an increment in the height of the first peak that also shifts slightly towards lower distances, a corresponding deepening of the first minimum and a gradual appearance of a second and third peak that tend to overlap with the experimental and with AIMD ones for the SPC/E and the flexible SPC models. Figure 2: The oxygen-oxygen two-bodies pair correlation, g2(r), of liquid water for the TIP3P (black), SPC (red), SPC/E (green) and the flexible SPC model (blue). The g2(r) obtained from various scattering experiments Skinner _et al._ (2013); Soper and Benmore (2008) and _ab initio_ molecular dynamics simulations DiStasio Jr. _et al._ (2014) are reported for comparison with open symbols. In fig. 3 we report the probability distribution P(n) of having a n-folded ring, with n$\in[3,12]$ for the classical 3 points models. The distribution for the TIP3P model is reported as open circles, SPC as open squares, SPC/E as open diamonds and flexible SPC as open triangles. It is worth to remark, at this point, that the P(n) is a normalized distribution and, therefore, it does not reflect the actual number of each ring. Panel a) reports the P(n) according to the definition d1 sketched in fig. 1 a). According to this counting scheme that emphasizes the directionality of the HBs, all networks have a dominating hexagonal character. Such character is milder in the TIP3P model (black open circles), which shows a very broad P(n) and whose network accommodates also 10- and 11-folded rings. The hexagonal character of the network grows moving to the SPC model (red open squares), with a corresponding reduction of longer (n$>$8) rings. The reduction of longer rings develops along with an enhancement of 5- and 7-folded rings, which are comparable to 6-folded rings in terms of energy Camisasca _et al._ (2019). In particular, the HBN of the SPC model described with this counting scheme is almost completely deprived of rings with n$>$10\. The hexagonal character of the network further grows in the SPC/E (green open diamonds) and in the flexible SPC (open blue triangles). These models show almost identical P(n)s. In particular, besides the enhanced hexagonal character, the HBNs of these models have an enhanced pentagonal and heptagonal character. Consequently, the contribution from longer rings, namely rings with n=8 and n=9, is less marked. Therefore, loosening the holonomic constraints and allowing the (re-parametrized) SPC model to vibrate has a major effect not just on the structural properties (as shown from the g2(r), fig. 2) but also on the topology of the HBN. Panel b) reports the P(n) according to the definition d2 sketched in fig. 1 b). As shown in Ref. Formanek and Martelli (2020), this counting scheme emphasizes the pentagonal character of the HBN, as the starting water molecule must accept one HB instead of donating. Interestingly, the TIP3P model shows, according to this counting scheme, an almost equal pentagonal and hexagonal character with a tail populating configurations up to n=10. The HBN of the SPC model, on the other hand, shows an improved pentagonal character with a slight increase in the hexagonal character and a reduction of longer rings which mostly disappear at n$>$9\. The pentagonal character of the HBN over the hexagonal one become particularly dominant in the SPC/E and in the flexible SPC models, for which rings characterized by n$>$8 are mostly absent. Panel c) reports the P(n) according to the definition d3 sketched in fig. 1 c). Since this counting scheme does not discern among the donor/acceptor character of the HBs and does not implement the shortest path criterion King (1967), the resulting topology is more complex with respect to the previous counting schemes. Therefore, for this scheme we also compute the network topology index $\xi$ (eq. 1) and report them in table 1. According to this counting scheme, the HBN for all the inspected models show a similar character in terms of hexagonal and heptagonal rings. The TIP3P model shows a quite broad distribution with a considerable contribution of longer (n$>$7) rings. The presence of rings with n=12 reflects the more complex topology of the network with respect to the previous counting schemes. Interestingly, such distribution is comparable with that of the TIP3P model optimized MacKerell Jr. _et al._ (1998) for biological simulations Martelli, Crain, and Franzese (2020). Because of the very broad character of P(n), the TIP3P model has a low value of network complexity index $\xi=0.1496$, not too far from the value $\xi=0.1$ that characterizes (as described in Section II.2) a flat distribution P(n) with equal populations. Moving to the SPC model, we observe a slight enhancement of n=5, and a more pronounced enhancement of n=6 and n=7 which, as for the TIP3P model, equally characterize the HBN. Contrarily, the network is deprived of longer rings, namely of n=10, n=11 and n=12 rings. The slight increase in n=6 causes a small increment in the network complexity index to $\xi=0.1665$. Moving to the SPC/E model, we observe a mild increment in n=5, with a relevant increment in n=6 and n=7, as well as a marked depletion of longer rings. The complexity index for the SPC/E model further increases to $\xi=0.1863$. A similar distribution characterizes the HBN of the flexible SPC model, which almost overlaps with the P(n) of the SPC/E model. The distribution of the flexible SPC model suggests that, as mentioned above, the introduction of flexibility plays a very important role in shaping the network. The complexity index for the flexible SPC model rises to $\xi=0.1922$. Figure 3: Probability distributions of the hydrogen-bonded n-folded rings, P(n), for liquid water at ambient conditions described by the TIP3P (black open circles), the SPC (red open squares), the SPC/E (green open diamonds) and the flexible SPC (blue open triangles) models. Panel a) reports the P(n) according to the definition sketched in fig. 1 a); panel b) reports the p(n) according to the definition sketched in fig. 1 b); panel c) reports the P(n) according to the definition sketched in fig. 1 c). | TIP3P | SPC | SPC/E | SPC-Flex ---|---|---|---|--- $\xi$ | 0.1496 | 0.1665 | 0.1863 | 0.1922 Table 1: Values of the network complexity index $\xi$ computed for the rings counting scheme d3 for the 3 points models of water. In fig. 4, we report the percentage of broken and intact HBs for the 3 point models of water and, for comparison, the _ab initio_ liquid water at the PBE0 level of theory with vdW long range interactions at a temperature of 330K to account for nuclear quantum effects DiStasio Jr. _et al._ (2014). We adopt the following syntax: $\textit{A}_{x}\textit{D}_{y}$ indicates the number of acceptors ($\textit{A}_{x}$) and donors ($\textit{D}_{y}$) HBs. The network in _ab initio_ liquid water (black open circles) is dominated by a intact HBs ($\textit{A}_{2}\textit{D}_{2}$), which account for $\sim 48\%$. The second highest configuration is the $\textit{A}_{1}\textit{D}_{2}$, with $\sim 20\%$, followed by $\textit{A}_{2}\textit{D}_{1}$ ($12\%$), $\textit{A}_{2}\textit{D}_{2}$ ($\sim 10\%$) and $\textit{A}_{3}\textit{D}_{2}$ ($\sim 5\%$). The percentage of this last configuration, i.e., the configuration $\textit{A}_{3}\textit{D}_{2}$, is shared with all 3 points models. The percentage of broken and intact HBs in the TIP3P model (red open squares) quantitatively differs from the distribution of _ab initio_ liquid water, but qualitatively shows a similar behaviour. In particular, the distribution for the TIP3P model is dominated by a markedly reduced amount of intact HBs with a percentage $\textit{A}_{2}\textit{D}_{2}\sim 33\%$, followed by $\textit{A}_{1}\textit{D}_{2}$ with $\sim 23\%$, $\textit{A}_{2}\textit{D}_{1}$ with $\sim 12\%$ and an almost equal percentage of $\textit{A}_{1}\textit{D}_{1}$. The low percentage of intact HBs explains the very broad distribution of rings (fig. 3) as well as the absence of structurization beyond the first hydration shell in the g2(r) (fig. 2). Moving from the TIP3 model to the SPC model (open green diamonds) we observe a slight decrease in the configuration $\textit{A}_{1}\textit{D}_{1}$ to $\sim 10\%$ and an increment in the percentage of intact HBs to $\sim 37\%$. Therefore, the SPC model is characterized by a slightly enhanced percentage of fully coordinated configurations which cause an enhancement in 5-, 6-, and 7-folded rings, as reported in fig. 3, and are responsible for the appearance of the second peak in the g2(r) (fig. 2). The modified SPC models, namely the SPC/E (open blue triangles) and the flexible SPC model (open orange left triangles), are both characterized by a distribution that mostly overlap with the _ab initio_ liquid water potential. The high percentage of intact HBs reaches values in the order of $\sim 45-48\%$ for both SPC/E and the flexible SPC models, and this enhanced tetra-coordinated character explains the enhanced number of 5-, 6-, and 7-folded rings and the corresponding decrease in longer rings, as reported in fig. 3, as well as an enhancement in the intensity of the first peak in the g2(r) and a corresponding deepening of the following minimum. Figure 4: Percentage-wise decomposition of the intact HBs per water molecule into acceptor-(A) and donor-(D) for _ab initio_ liquid water at T=330 K as black open circles, and for the 3 points models. The TIP3P model is reported as red open squares, the SPC as green open diamonds, the SPC/E as blue open triangles and the flexible SPC as orange open left triangles. The $x$-axis labels $\textit{A}_{x}\textit{D}_{y}$ indicate the number of acceptor ($\textit{A}_{x}$) and donor ($\textit{D}_{y}$) HBs. For clarity we omit combinations with minor contributions, e.g., $\textit{A}_{3}\textit{D}_{1}$, $\textit{A}_{0}\textit{D}_{y}$, $\textit{A}_{x}\textit{D}_{0}$, etc. ### III.2 4 points models In fig. 5 we report the g2(r) for the 4 points models, namely the TIP4 (black line), the TIP4P-Ew (red line), the TIP4P/2005 (green line), the flexible TIP4P/2005 (blue line) and the TIP4P-ice (orange line) models. As for the 3 points models, the open symbols represent the g2(r) from experimental data Skinner _et al._ (2013); Soper and Benmore (2008) and from _ab initio_ molecular dynamics simulations DiStasio Jr. _et al._ (2014), and serve as a comparison. The TIP4P model is characterized by the intensity of the first peak as high as $\sim 3$, the closest to the experimental/_ab initio_ ones. The depth of the following minimum is slightly more pronounced than the experimental/AIMD. At larger distances, on the other hand, the g2(r) almost overlap with the experimental/AIMD ones. The TIP4P-Ew (red line) has been optimized for simulating water in biological environments Horn _et al._ (2004). With respect to the TIP4P model, the TIP4P-Ew shows a higher intensity in the first peak of the g2(r), reaching $\sim 3.2$. Other features of the g2(r) are qualitatively overlapping with the experimental/AIMD distribution functions. A similar trend characterizes the g2(r) for the TIP4P/2005 (green line) and for the flexible TIP4P/2005 (blue line) which are mostly indistinguishable from the distribution function of the TIP4P-Ew function. The TIP4P-Ice, on the other hand, shows a g2(r) more structured with respect to the previous ones. The intensity of the first peak reaches values as close as $\sim 3.5$, and the following minimum shows a deeper depth. Finally, the intensity of the second peak is slightly more pronounced with respect to the experimental/AIMD ones. Such ”over”-structurization is not surprising considering that this water model has been developed to study crystalline and non crystalline solid forms of ice Abascal _et al._ (2005). Overall, moving from the TIP4P model to the TIP4P-Ew, TIP4P/2005, flexible TIP4P/2005 and the TIP4P-Ice we observe a systematic enhancement of the first peak which also slightly shifts towards larger values. All models well capture the second and third peaks. Figure 5: The oxygen-oxygen two-bodies pair correlation, g2(r), of liquid water for the TIP4P (black), TIP4P-Ew (red), TIP4P/2005 (green), the flexible TIP4P/2005 (blue), and the TIP4P-Ice (orange) models. The g2(r) obtained from various scattering experiments Skinner _et al._ (2013); Soper and Benmore (2008) and _ab initio_ molecular dynamics simulations DiStasio Jr. _et al._ (2014) are reported for comparison with open symbols. In fig. 6 we report the probability distribution P(n) of having a n-folded ring, with n$\in[3,12]$ for the TIP4P (black open circles), the TIP4P-Ew (red open squares), the TIP4P/2005 (green open diamonds), the flexible TIP4P/2005 (blue open triangles) and the TIP4P-Ice (orange open left triangles) models. In panel a) we report the P(n) according to the definition d1 sketched in fig. 1 a). All classical models of water show a P(n) maximized at n=6. In particular, the TIP4P model shows, among all the 4 points models here inspected, the broader distribution, with an almost equal contribution of 5-folded and 7-folded rings. Interestingly, the distribution for the TIP4P model is less broad compared with the distribution for the TIP3P model reported in fig. 3 a), with almost no rings longer than n=9. The HBN of the TIP4P-Ew model shows a slight increase in the pentagonal and in the hexagonal character and a corresponding lower character in rings longer than n=7. Very similar distributions occur for the TIP4P/2005 and for the flexible TIP4P/2005 models. The hexagonal character of the HBN is further emphasized in the TIP4P-Ice model for which we can observe also a slight increment in the pentagonal character and a depletion of rings longer than n=7. This enhanced hexagonal character of the HBN is a consequence of the parametrization of this model which has been optimized to reproduce crystalline and non-crystalline solid forms of ice. In panel b) of fig. 6 we report the P(n) computed using the rings and counting scheme definition d2 sketched in fig. 1 b). According to such counting scheme, all distributions are maximized over n=5. In particular, the P(n) for the TIP4P model shows the broader distribution, with some contribution from n=8, while the contributions to the HBN from longer rings are mostly negligible. The TIP4P-Ew model shows an enhanced pentagonal character and a slightly increment of the hexagonal character as well, with a reduction of the contribution coming from rings longer than n=6. Such tendency becomes more pronounced moving to the TIP4P/2005 and to the flexible TIP4P/2005 models, while differences among these three models are minimal. On the other hand, the pentagonal character of the HBN is particularly enhanced in the TIP4P-Ice model whose HBN, according to this counting scheme, dose not account for rings longer than n=7. In panel c) we report the P(n) computed using the counting scheme d3 sketched in fig. 1 c). For this scheme, we also compute the complexity indices $\xi$, reported in table 2. We can observe that the distribution P(n) for the TIP4P model is fairly broad and mostly dominated by an almost equal amount of 6- and 7-folded rings followed by 5-folded rings, while the amount of longer rings decreases with increasing the rings lengths. The complexity index for the TIP4P model is $\xi=0.1721$, which is larger compared to the TIP3 and the SPC models, but lower compared to the other 3 points models here inspected. The network of TIP4P-Ew model is characterized by a consistent increment in the hexagonal character, followed by the heptagonal character and a smaller increment in the pentagonal character. Longer rings weight less with respect to the TIP4P model. The complexity index for the TIP4P-Ew is $\xi=0.1887$, larger compared to the TIP4P model. The distribution P(n) for the TIP4P/2005 and the flexible TIP4P/2005 models mostly overlap with the P(n) for the TIP4P-Ew, and have a similar complexity index, namely $\xi=0.1915$ for the TIP4P/2005 model and $\xi=0.1984$ for the flexible TIP4P/2005 model. Interestingly, the P(n) for the TIP4P-Ew, for the TIP4P/2005 and for the flexible TIP4P/2005 models show a breaking in the equal hexagonal and heptagonal character of the network, with a slight predominance of the hexagonal character which justifies the higher values of the index $\xi$. The network of the TIP4P-Ice model is characterized by a further enhancement of the hexagonal character, followed by the heptagonal and hexagonal characters, with a depletion of longer rings. Such enhancement in the hexagonal character is reflected by the complexity index $\xi=0.2155$, the closest to the value of ice among all models inspected in this work. The similar distributions of rings for the TIP4P-Ew, TIP4P/2005 and flexible TIP4P/2005 in all counting schemes suggest that such models perform equally in terms of HBN, and this reflects the very similar g2(r) (fig. 5). On the other hand, the TIP4P-Ice, which shows a more structured g2(r) (fig. 5), tends to favour a network with a shorter connectivity. Figure 6: Probability distributions of the hydrogen-bonded n-folded rings, P(n), for liquid water at ambient conditions described by the TIP4P (black open circles), the TIP4P-EW (red open squares), the TIP4P/2005 (green open diamonds), the flexible TIP4P/2005 (blue open triangles), and the TIP4P-Ice (open orange triangles) models. | TIP4P | TIP4P-Ew | TIP4P/2005 | TIP4P/2005-Flex | TIP4P-Ice ---|---|---|---|---|--- $\xi$ | 0.1721 | 0.1887 | 0.1915 | 0.1984 | 0.2155 Table 2: Values of the network complexity index $\xi$ computed for the rings counting scheme d3 for the 4 points models of water. In fig. 7 we report the percentage of broken and intact HBs for the 4 points classical models and for _ab initio_ liquid water. We can observe that the percentage of broken and intact HBs for all 4 points models qualitatively resembles that of _ab initio_ liquid water (open black circles). With respect to AIMD water, the TIP4P model (open red squares) underestimates the percentage of $\textit{A}_{2}\textit{D}_{2}$ configurations to $\sim 43\%$ and a slightly higher percentage of $\textit{A}_{1}\textit{D}_{2}$ configurations ($\sim 22\%$). All other configurations mostly overlap with the configurations of AIMD water. The percentage of broken and intact HBs for the TIP4P-Ew (open green diamonds), for the TIP4P/2005 (open blue triangles) and the flexible TIP4P/2005 (open orange left triangles) mostly overlap with each other, reflecting the almost overlapping distributions of rings and the similar values of the complexity index $\xi$ (fig. 6). In particular, we observe almost no differences between the distribution for the TIP4P/2005 and the flexible TIP4P/2005 models. With respect to the TIP4P model, they recover the percentage of $\textit{A}_{2}\textit{D}_{2}$ for the AIMD liquid water, while showing a small reduction of $\textit{A}_{1}\textit{D}_{1}$ defects ($\sim 7\%$). Such increment in the percentage of $\textit{A}_{2}\textit{D}_{2}$ to $\sim 48\%$ reflects the stronger 5-, 6- and 7-folded character of the HBN described in fig. 6. The TIP4P-Ice model, on the other hand, overestimate the percentage of intact HBs with respect to AIMD water, reaching $\sim 55\%$ of the total configurations. This enhanced 4-folded coordination explains the distribution of rings (fig.6) describing an HBN particularly enriched in 6- and 7-folded rings, as well as the higher intensity in the first peak of the g2(r) (fig 5). Overall, we can state that, overall, the 4 points models are characterized by similar HBNs, with the exception of the TIP4P model which shows a broad distribution of rings caused by a lower percentage of intact HBs, and the TIP4P-Ice model, whose high percentage of intact HBs causes generates a network characterized by a strong hexagonal and heptagonal character with a marked reduction of longer rings. Figure 7: Percentage-wise decomposition of the intact HBs per water molecule into acceptor-(A) and donor-(D) for _ab initio_ liquid water at T=330 K as black open circles, and for the 4 points models. The TIP4P model is reported as red open squares, the TIP4P-Ew model as green open diamonds, the TIP4P/2005 model as blue open triangles, the flexible TIP4P/2005 as orange open left triangles, and the TIP4P-Ice model as brown open lower triangles. The distribution for the TIP4P/2005 model almost perfectly overlaps with the distribution for the flexible TIP4P/2005 model. ### III.3 5 points models In fig. 8 we compare the g2(r) of two 5 points models, namely the TIP5P (black line) and the TIP5P-E (red line) models with the g2(r) obtained from various scattering experiments Skinner _et al._ (2013); Soper and Benmore (2008) and _ab initio_ molecular dynamics simulations DiStasio Jr. _et al._ (2014) reported as open symbols. We can observe that the g2(r) for both 5 points models mostly overlap. With respect to the 3 points and to the 4 points models, the g2(r) of both 5 points models are more closer to the experimental and to the AIMD g2(r). The intensity of the first peak for the TIP5P and the TIP5P-E models is below 3, namely $\sim$2.9. The depth of the first minimum is slightly more pronounced with respect to the experimental and the AIMD g2(r), while experimental and AIMD peaks at longer distances are well captured. Figure 8: The oxygen-oxygen two-bodies pair correlation, g2(r), of liquid water for the TIP5P (black) and the TIP5P-EW (red) models. The g2(r) obtained from various scattering experiments Skinner _et al._ (2013); Soper and Benmore (2008) and _ab initio_ molecular dynamics simulations DiStasio Jr. _et al._ (2014) are reported for comparison with open symbols. In fig. 9 we report the probability distribution P(n) of having a n-folded ring, with n$\in[3,12]$ for the TIP5P model (open black circles) and for the TIP5P-E model (open red squares). In panel a) we report the P(n) computed using the counting scheme d1 sketched in fig. 1 a) and which emphasizes the directionality of the HBN. Both the TIP5P and the TIP5P-E models provide similar distributions which are slightly maximized at n=6, and contributions of rings up to n=11. In panel b) we report the P(n) computed using, as a counting scheme, the definition d2 sketched in fig. 1 b). As for the previous case, also in this case the two distributions are mostly indistinguishable and with an almost equal pentagonal and hexagonal character. Overall, the pentagonal character of both HBNs is less pronounced with respect to the four points models, and qualitatively resemble the P(n) for the three points model TIP3P (fig. 3 b)). In panel c) we report the P(n) computed using the definition d3 and sketched in fig. 1 c). For this counting scheme, we also compute the complexity indices reported in table 3. As for the previous counting schemes, the differences between the two P(n)’s are minimal, indicating that the two model provide similar HBNs. Such similarity can be quantified observing that both models have almost the same complexity index $\xi$, i.e., $\xi=0.1774$ for the TIP5P model and $\xi=0.1754$ for the TIP4P-Ew model. The values of $\xi$ for both networks are comparable with that of the TIP4P model but, with respect to the TIP4P model, the hexagonal character of the HBN is more pronounced for both the TIP5P and the TIP5P-E models. Besides the TIP4P-Ice model, among all other interaction potentials here inspected, the TIP5P and the TIP5P-E models are the only one for which the counting scheme d3 is (slightly) maximized towards n=6. Figure 9: Probability distributions of the hydrogen-bonded n-folded rings, P(n), for liquid water at ambient conditions described by the TIP5P (black open circles) and the TIP5P-EW (red open squares). | TIP5P | TIP5P-E ---|---|--- $\xi$ | 0.1774 | 0.1754 Table 3: Values of the network complexity index $\xi$ computed for the rings counting scheme d3 for the 5 points models of water. In fig. 10 we report the percentage of broken and intact HBs for the 5 points classical models and for _ab initio_ liquid water. With respect to the AIMD liquid water (open black circles), both the TIP5P (open red squares) and the TIP5P-E (open green diamonds) models are characterized by a markedly lower percentage of intact HBs ($\sim 36\%$), while low-coordinated defects occur in higher percentages. The $\textit{A}_{1}\textit{D}_{2}$ configuration account for the $\sim 22\%$ of the total configurations, followed by $\textit{A}_{2}\textit{D}_{1}$ configurations with a percentage of $\sim 16\%$ and $\textit{A}_{1}\textit{D}_{1}$ configurations with $\sim 12\%$ of the total configurations. The low amount of intact HBs (compared with the AIMD water), reflects the broad distribution of rings and the corresponding low value of the network complexity indices $\xi$ (fig. 9). Figure 10: Percentage-wise decomposition of the intact HBs per water molecule into acceptor-(A) and donor-(D) for _ab initio_ liquid water at T=330 K as black open circles, and for the 5 points models. The TIP5P model is reported as red open squares and the TIP5P-EW model as green open diamonds. ### III.4 Relation between network complexity and dynamical properties We here show how the network complexity is linked to dynamical properties, namely the translational and rotational diffusion. This link comes from the observation that bonding can be viewed as a competition between the energy gained from the formation of a bond, and the entropy loss due to the reduction in configurational volume that occurs when two particles are constrained to stay close relative to each other. The establishment of an extended network of bonds occurs when the energy gain (that controls the lifetime of bonds) balances the entropy loss. For each water model we have computed the diffusion coefficient and the rotational relaxation time ($\tau_{rot}$), and we have reported them against the network complexity index $\xi$. We have computed the diffusion coefficient from the mean squared displacement, and the rotational relaxation time $\tau_{rot}$ from the integral of the rotational autocorrelation function as reported in Refs Calero, Stanley, and Franzese (2016); Martelli, Calero, and Franzese (2021) $C_{rot}(t)=\left<\textit{OH}(t)\cdot\textit{OH}(0)\right>$, i.e., $\tau_{rot}=\int_{0}^{+\infty}C_{rot}(t)dt$. Fig. 11 show the values of these three observables in a three dimensional plot, with projections on the corresponding two dimensional spaces. The values for 3-points models are reported in red, the values for 4-points models are reported in green, and the values of 5-points models are reported in blue. We can observe a clear correlation between the complexity of the HBN and the dynamical properties of water molecules. The models with the highest diffusion coefficients and the fastest rotational relaxation times are the TIP3P and the SPC models, which are also characterized by the lowest values of the index $\xi$. Contrarily, the TIP4P-Ice model is the model with the lowest diffusion coefficient and the slowest rotational relaxation time, and the highest index $\xi$. Overall, we can observe that for all models of water the higher the value of $\xi$ (reported in ascending order in the tables 1, 2, 3), the slower the diffusion coefficient and the rotational relaxation time. Therefore, we can assert that there is a clear correlation between the complexity of the HBN and dynamical properties, a relation never observed before. Such correlation suggests that faster diffusion and rotations allow water molecules to increase the possible connections between each other, hence increasing the configurational space that the network can explore resulting in a more complex topology able to host a larger amount of longer rings. Figure 11: Three dimensional plot reporting the projection on the corresponding two dimensions of the values acquired by the 11 classical models of water. We report the values for the network complexity index $\xi$, the diffusion coefficient and the rotational relaxation time $\tau_{rot}$. Data for 3-points models are reported in red, while data for the 4-points models are reported in green and data for the 5-points models are reported in blue. ### III.5 System size dependence We now turn our attention to the study of finite size effects, commonly inspected when computing physical quantities to check whether a system suffers from periodicity artifacts. In fig. 12 a) we report the oxygen-oxygen g2(r) for a simulation box containing 500 water molecules (black continuous line), 1000 water molecules (red dashed line) and 1500 water molecules (green dotted- dashed line) interacting with the TIP4P classical interaction potential. We can observe that the three g2(r) perfectly overlap. In fig. 12 b) we report the probability distribution P(n) of having a n-folded ring for the three cases inspected above and computed according to the ring definition and counting scheme d3. We can observe that the distribution computed for the smaller simulation box with N=500 molecules is remarkably different from the distributions with N=1000 and N=1500 molecules. In particular, we observe a strong enhancement of n=12 rings which causes a reduced contribution of shorter rings to the P(n). This result indicates that a search path of n=12 water molecules is too long for a small simulation box with only N=500 water molecules, and the increment in n=12 is caused by periodicity in the simulation box. It is worth to mention that the definitions d1 and d2 do not show such behavior (data not reported), as the network investigated with these definitions does not host rings as long as n=12 (see fig. 6 upper and middle panels). Therefore, although the g2(r) for N=500 is the same as the g2(r) for larger simulation boxes, care must be taken when inspecting the network topology and in how such inspection is performed. Figure 12: Panel a): system size dependence on the oxygen-oxygen two bodies pair correlation function for the TIP4P model using a box containing 500 water molecules (black continuous line), 1000 water molecules (red dashed line) and 1500 water molecules (green dotted-dashed line). Panel b): ring distribution for a system described by the TIP4P model in a box containing 500 water molecules (black pluses), 1000 water molecules (red crosses) and 1500 water molecules (green stars). ## IV Conclusions In this article we have tested 11 popular non polarizable classical interaction potentials for water against their hydrogen bond networks (HBNs). We have probed the topology of the HBN using three schemes that emphasize different physical features. We have evaluated the quality of the HBNs in terms of broken and intact HBs, and we have linked our results to structural properties measured _via_ the two bodies pair correlation function g2(r). We have then introduced the network complexity index $\xi$ that measures how much the topology of a HBN deviates from that of the ground state, and we have tested it to one of the three rings counting schemes. We have shown that the index $\xi$ is directly related to dynamical properties, hence establishing a clear cause-effect relationship between molecular motions and network connectivity. Finally, we have inspected how periodicity artifacts can influence the topology of the HBN. We have performed all studies at ambient conditions, i.e., T=300 K and p=1 bar. Although different water models have (very) different melting points, their network topology –and hence their network complexity– remain roughly unchanged away from the limit of supercooling Formanek and Martelli (2020). On the other hand, when water is under confinement, water molecules in the proximity of the surfaces undergo a drastic change in the dynamics Samatas _et al._ (2018); Martelli, Crain, and Franzese (2020); Chiricotto _et al._ ; Gallo, Rovere, and Chen (2010); Camisasca, Marzio, and Gallo (2020); Iorio, Camisasca, and Gallo (2019); Iorio _et al._ (2020); Tenuzzo, Camisasca, and Gallo (2020); Calero and Franzese (2020) and in network topology Martelli, Crain, and Franzese (2020); Chiricotto _et al._ . Therefore, the choice of a given interaction potential becomes of particular relevance. In the class of 3 points models, we have tested the TIP3P, the SPC, the SPC/E and the flexible SPC models. We have found that the TIP3P model is characterized by the less structured network, with broad distributions or rings and a network rich in coordination defects which allows the network to arrange in long rings. In particular, the counting scheme d3 gives a low value of complexity index $\xi$ which reflects the large distance from the HBN of crystalline ice for which $\xi=1.0$. The broad rings distribution and the low percentage of intact HBs explain the absence of a second hydration peak in the g2(r). The percentage of intact HBs increases in the SPC model which is, therefore, characterized by an HBN with fewer longer rings and by a g2(r) with signatures of a second hydration peak. The SPC/E and the flexible SPC models are characterized by a further increase in the percentage of intact HBs, comparable with that of _ab initio_ liquid water. The resulting HBNs accommodate an even lower percentage of longer rings and, therefore, the network complexity index is higher compared to the previous models, indicating a closer (but still very far) HBN to the HBN of Ih(c), in agreement with the more structured g2(r). It is of particular interest to observe how the introduction of flexibility in the SPC model drastically affects the topology of the HBN. In the class of 4 points models, we have tested the TIP4P, the TIP4P-Ew, the TIP4P/2005, the flexible TIP4P/2005 and the TIP4P-Ice models. The TIP4P model is the only model whose network accommodates a lower percentage of intact HBs with respect to _ab initio_ liquid water. The topology of the corresponding HBN is hence the most complex, i.e., with a low complexity index. The TIP4P-Ew, the TIP4P/2005 and the flexible TIP4P/2005 models are characterized by a similar percentage of intact HBs, comparable with that of _ab initio_ liquid water. The corresponding HBNs have comparable topologies and values of complexity indices. The TIP4P-Ice model, finally, shows a higher percentage of intact HBs with respect to _ab initio_ liquid water. The topology of the corresponding HBN is the less complex among all models here studied, with small contributions of longer rings and the highest value of $\xi$ index. Such results explain the over structured g2(r) with respect to both _ab initio_ water and experimental results. In the class of 5 points models, we have tested the TIP5P and the TIP5P-E models. Both models have similar HBNs, characterized by a lower percentage of broken HBs with respect to _ab initio_ liquid water. Both networks are fairly complex, accommodating longer rings causing low values of the network complexity index. Overall, the balance between intact HBs and network topology allows the 5 points models to be the better models in reproducing the _ab initio_ and experimental water g2(r). Overall, we have shown that water models endowed with the fastest dynamics are able to establish more complex networks, while models with the slowest dynamics establishes networks more closely related to that of cubic or hexagonal ice. In particular, among the 11 models here inspected the TIP3P and the SPC models are the ones with the fastest dynamics and with the network deviating the most from that of ice. On the other hand, the TIP4P-Ice model is the one with the slowest dynamics and, hence, with a network of HB the closer to that of ice. Finally, we have shown that the topology of the HBN might be affected by finite size effects when other observables such as, e.g., the two body pair correlation function, do not show such sensitivity. In conclusion, the topology of the HBN and its quality in terms of broken and intact HBs are more sensitive quantities than other physical observables Martelli, Crain, and Franzese (2020); Chiricotto _et al._ . Therefore, the properties of the HBN should be inspected along with all other properties such as, e.g., structural, dynamical and thermodynamic properties when developing new interaction potentials. Our study provides a benchmark evaluated following three different ring definitions and counting schemes. New interaction potentials should be tested against such results. Nonetheless, the network complexity index provides a direct quantitative measure of how much a HBN is complex and far from the HBN at the ground state, and a direct link to the dynamical properties. Such quantity can be transferred to other materials. The effect of polarization on the topology of the HBN and on its quality should be investigated. ###### Acknowledgements. We acknowledge support from the STFC Hartree Centre’s Innovation Return on Research programme, funded by the Department for Business, Energy and Industrial Strategy. ## References * Salzmann (2019) C. G. Salzmann, J. Chem. Phys. 150, 060901 (2019). * Palmer _et al._ (2014) J. C. Palmer, F. Martelli, Y. Liu, R. Car, A. Z. Panagiotopoulos, and P. G. Debenedetti, Nature 510, 385 (2014). * Sellberg _et al._ (2014) J. A. Sellberg, C. Huang, T. A. McQueen, N. D. Loh, H. Laksmono, D. Sclesinger, R. G. Sierra, D. Nordlund, C. Y. Hampton, D. Starodub, D. P. DePonte, M. Beye, C. Chen, A. V. Martin, A. Barty, K. T. Wikfeldt, T. M. Weiss, C. Caronna, J. Feldkamp, L. B. Skinner, M. M. Seibert, M. Messerschmidt, G. J. Williams, S. Boutet, L. G. M. Pettersson, M. J. Bogan, and A. Nilsson, Nature 510, 381 (2014). * Debenedetti, Sciortino, and Zerze (2020) P. G. Debenedetti, F. Sciortino, and G. H. Zerze, Science 369, 289 (2020). * Kringle _et al._ (2020) L. Kringle, W. A. Thornley, B. D. Kay, and G. A. Kimmel, Science 369, 1490 (2020). * Kim _et al._ (2020) K. H. Kim, K. Amann-Winkel, N. Giovambattista, A. Spah, F. Perakis, H. Pathak, M. L. Parada, C. Yang, D. Mariedahl, T. Eklund, T. J. Lane, S. You, S. Jeong, M. Weston, J. H. Lee, I. Eom, M. Kim, J. Park, P. P. S.H. Chun, and A. Nilsson, Science 370, 978 (2020). * Poole _et al._ (1992) P. H. Poole, F. Sciortino, U. Essmann, and H. E. Stanley, Nature 360, 324 (1992). * Liu _et al._ (2010) Y. Liu, J. C. Palmer, A. Z. Panagiotopoulos, and P. G. Debenedetti, J. Chem. Phys. 137, 214505 (2010). * Limmer and Chandler (2011) D. T. Limmer and D. Chandler, J. Chem. Phys. 135, 134503 (2011). * Wikfeldt, Nilsson, and Pettersson (2011) K. T. Wikfeldt, A. Nilsson, and L. G. M. Pettersson, Phys. Chem. Chem. Phys. 13, 19918 (2011). * Palmer, Car, and Debenedetti (2013) J. C. Palmer, R. Car, and P. G. Debenedetti, Faraday Discuss. 167, 77 (2013). * Limmer and Chandler (2013) D. T. Limmer and D. Chandler, J. Chem. Phys. 138, 214504 (2013). * Limmer and Chandler (2014) D. T. Limmer and D. Chandler, Proc. Natl. Acad. Sci. USA 111, 9413 (2014). * Chandler (2016) D. Chandler, Nature 531, E1 (2016). * Palmer _et al._ (2016a) J. C. Palmer, F. Martelli, Y. Liu, R. Car, A. Z. Panagiotopoulos, and G. D. P, Nature 531, E2 (2016a). * Palmer _et al._ (2018) J. C. Palmer, A. Haji-Akbari, R. S. Singh, F. Martelli, R. Car, A. Z. Panagiotopoulos, and P. G. Debenedetti, J. Chem. Phys. 148, 137101 (2018). * Palmer _et al._ (2016b) J. C. Palmer, R. S. Singh, R. Chen, F. Martelli, and P. G. Debenedetti, Mol. Phys. 114, 2580 (2016b). * Martelli _et al._ (2020) F. Martelli, F. Leoni, F. Sciortino, and J. Russo, J. Chem. Phys. 153, 104503 (2020). * Shi and Tanaka (2020a) R. Shi and H. Tanaka, Proc. Natl. Ac. Sci. 117, 26591 (2020a). * Shi and Tanaka (2020b) R. Shi and H. Tanaka, J. Am. Chem. Soc. 142, 2868 (2020b). * Shi, Russo, and Tanaka (2018a) R. Shi, J. Russo, and H. Tanaka, J. Chem. Phys. 149, 224502 (2018a). * Russo and Tanaka (2014a) J. Russo and H. Tanaka, Nat. Commun. 5, 3556 (2014a). * Akahane and Tanaka (2018) J. R. K. Akahane and H. Tanaka, Proc. Natl. Ac. Sci. 115, E3333 (2018). * Santra _et al._ (2015) B. Santra, R. A. DiStasio Jr., F. Martelli, and R. Car, Mol. Phys. 113, 2829 (2015). * Huang _et al._ (2009) C. Huang, K. T. Wikfeldt, T. Tokushima, D. Nordlund, Y. Harada, U. Bergmann, M. Niebuhr, T. M. Weiss, Y. Horikawa, M. Leetmaa, M. P. Ljungberg, O. Takahashi, A. Lenz, L. Ojamäe, A. P. Lyubartsev, S. Shin, L. G. M. Pettersson, and A. Nilsson, Proc. Natl. Acad. Sci. USA 106, 15214 (2009). * Nilsson and Pettersson (2015) A. Nilsson and L. G. M. Pettersson, Nat. Commun. 6, 8998 (2015). * De Marzio _et al._ (2017) M. De Marzio, G. Camicasca, M. Rovere, and P. Gallo, J. Chem. Phys. 146, 084502 (2017). * Martelli (2019) F. Martelli, J. Chem. Phys. 150, 094506 (2019). * Pekka and Lennart (2001) M. Pekka and N. Lennart, J. Phys. Chem. A 105, 9954 (2001). * Mao and Zhang (2012) Y. Mao and Y. Zhang, Chem. Phys. Lett. 542, 37 (2012). * Lee and Kim (2019) S. H. Lee and J. Kim, Mol. Phys. 117, 1926 (2019). * Jorgensen _et al._ (1983) W. L. Jorgensen, J. Chandrasekhar, J. D. Madura, R. W. Impey, and M. L. Klein, J. Chem. Phys. 79, 926 (1983). * Harrach and Drossel (2014) M. F. Harrach and B. Drossel, J. Chem. Phys. 140, 174501 (2014). * González _et al._ (2010) B. S. González, E. G. Noya, C. Vega, and L. M. Sesé, J. Phys. Chem. B 114, 2484 (2010). * Zielkiewicz (2005) J. Zielkiewicz, J. Chem. Phys. 123, 104501 (2005). * Steinczinger, Jóvári, and Pusztai (2017) Z. Steinczinger, P. Jóvári, and L. Pusztai, J. Mol. Phys. 228, 19 (2017). * Dix, Lue, and Carbone (2018) J. Dix, L. Lue, and P. Carbone, J. Comp. Chem. 39, 2051 (2018). * Bernal and Fowler (1933) J. D. Bernal and R. H. Fowler, J. Chem. Phys. 1, 515 (1933). * Tse _et al._ (1999) J. S. Tse, D. D. Klug, C. A. Tulk, I. Swainson, E. C. Svensson, C.-K. Loong, V. Shpakov, V. R. Belosludov, R. V. Belosludov, and Y. Kawazoe, Nature 400, 647 (1999). * Marton̆ák, Donadio, and Parrinello (2004) R. Marton̆ák, D. Donadio, and M. Parrinello, Phys. Rev. Lett. 92, 225702 (2004). * Marton̆ák, Donadio, and Parrinello (2005) R. Marton̆ák, D. Donadio, and M. Parrinello, J. Chem. Phys. 122, 134501 (2005). * Shephard _et al._ (2017) J. J. Shephard, S. Ling, G. Sosso, A. Michaelides, B. Slater, and C. G. Salzmann, J. Phys. Chem. Lett. 8, 1645 (2017). * Martelli _et al._ (2018) F. Martelli, N. Giovambattista, S. Torquato, and R. Car, Phys. Rev. Materials 2, 075601 (2018). * Martelli, Crain, and Franzese (2020) F. Martelli, J. Crain, and G. Franzese, ACS Nano 14, 8616 (2020). * (45) M. Chiricotto, F. Martelli, G. Giunta, and P. Carbone, J. Phys. Chem. C Under review. * Martelli, Calero, and Franzese (2021) F. Martelli, C. Calero, and G. Franzese, “Re-defining the concept of hydration water near soft interfaces,” (2021), arXiv:2101.06136 [cond-mat.soft] . * Berendsen _et al._ (1981) H. J. C. Berendsen, J. P. M. Postma, W. F. van Gunsteren, and J. Hermans, in _Intermolecular Forces_ , edited by B. Pullman (Springer, Dordrecht, 1981) pp. 331–342. * Berendsen, Grigera, and Straatsma (1987) H. J. C. Berendsen, J. R. Grigera, and T. P. Straatsma, J. Phys. Chem. 91, 6269 (1987). * Toukan and Rahman (1985) K. Toukan and A. Rahman, Phys. Rev. B 31, 2643 (1985). * Amira, Spángberg, and Hermansson (2004) S. Amira, D. Spángberg, and K. Hermansson, Chem. Phys. 303, 327 (2004). * Abascal _et al._ (2005) J. L. F. Abascal, E. Sanz, R. G. Fernández, and C. Vega, J. Chem. Phys. 122, 234511 (2005). * Abascal and Vega (2005) J. L. F. Abascal and C. Vega, J. Chem. Phys. 123, 234505 (2005). * González and Abascala (2011) M. A. González and J. L. F. Abascala, J. Chem. Phys. 135, 224516 (2011). * Horn _et al._ (2004) H. W. Horn, W. C. Swope, J. W. Pitera, J. D. Madura, T. J. Dick, G. L. Hura, and T. Head-Gordon, J. Chem. Phys. 120, 9665 (2004). * Mahoney and Jorgensen (2000) M. W. Mahoney and W. L. Jorgensen, J. Chem. Phys. 122, 8910 (2000). * Rick (2004) S. W. Rick, J. Chem. Phys. 120, 6085 (2004). * Martelli _et al._ (2016) F. Martelli, H.-Y. Ko, E. C. Oğuz, and R. Car, Phys. Rev. B 97, 064105 (2016). * Formanek and Martelli (2020) M. Formanek and F. Martelli, AIP Adv. 10, 055205 (2020). * Leoni _et al._ (2019) F. Leoni, R. Shi, H. Tanaka, and J. Russo, J. Chem. Phys. 151, 044505 (2019). * Camisasca _et al._ (2019) G. Camisasca, D. Schlesinger, I. Zhovtobriukh, G. Pitsevich, and L. G. M. Pettersson, J. Chem. Phys. 151, 034508 (2019). * Russo and Tanaka (2014b) J. Russo and H. Tanaka, Nat. Commun. 5, 3556 (2014b). * Fitzner _et al._ (2019) M. Fitzner, G. C. Sosso, S. J. Cox, and A. Michaelides, Proc. Natl. Acad. Sci. USA 116, 2009 (2019). * Shi and Tanaka (2018) R. Shi and H. Tanaka, Proc. Natl. Acad. Sci. USA 115, 1980 (2018). * Bakó _et al._ (2017) I. Bakó, J. Oláh, A. Lábas, S. Bálint, L. Pusztai, and M. C. B. Funel, J. Mol. Liq. 228, 25 (2017). * Pothoczki, Pusztai, and Bakó (2018) S. Pothoczki, L. Pusztai, and I. Bakó, J. Phys. Chem. B 122, 6790 (2018). * Pothoczki, Pusztai, and Bakó (2019) S. Pothoczki, L. Pusztai, and I. Bakó, J. Phys. Chem. B 123, 7599 (2019). * Li _et al._ (2020a) L. Li, J. Zhong, Y. Yan, J. Zhang, J. Xu, J. S. Francisco, and X. C. Zeng, Proc. Natl. Acad. Sci. USA 117, 24701 (2020a). * DiStasio Jr. _et al._ (2014) R. A. DiStasio Jr., B. Santra, Z. Li, X. Wu, and R. Car, J. Chem. Phys. 141, 084502 (2014). * Nosé (1984) S. Nosé, Mol. Phys. 52, 255 (1984). * Hoover (1985) W. G. Hoover, Phys. Rev. A 31, 1695 (1985). * Parrinello and Rahman (1981) M. Parrinello and A. Rahman, J. Appl. Phys. 52, 7182 (1981). * Abraham _et al._ (2015) M. J. Abraham, T. Murtola, R. Schulz, S. Páll, J. C. Smith, B. Hess, and E. Lindahl, SoftwareX 1, 19 (2015). * King (1967) S. V. King, Nature 213, 1112 (1967). * Rahman and Stillinger (1973) A. Rahman and F. H. Stillinger, J. Am. Chem. Soc. 95, 7943 (1973). * Guttman (1990) L. Guttman, J. Non-Cryst. Solids 116, 145 (1990). * Franzblau (1991) D. S. Franzblau, Phys. Rev. B 44, 4925 (1991). * Wooten (2002) F. Wooten, Acta Cryst. A 58, 346 (2002). * Yuan and Cormack (2002) X. Yuan and A. N. Cormack, Comp. Mater. Sci. 24, 343 (2002). * Roux and Jund (2010) S. L. Roux and P. Jund, Comp. Mater. Sci. 49, 70 (2010). * Jin _et al._ (1994) W. Jin, R. K. Kalia, P. Vashishta, and J. P. Rino, Phys. Rev. B 118, 50 (1994). * Hobbs _et al._ (1998) L. W. Hobbs, C. E. Jesurum, V. Pulim, and B. Berger, Philos. Mag. A 68, 679 (1998). * Marians and Hobbs (1988) C. S. Marians and L. W. Hobbs, J. Non-Cryst. Solids 160, 317 (1988). * Marians and Hobbs (1990) C. S. Marians and L. W. Hobbs, J. Non-Cryst. Solids 119, 269 (1990). * Luzar and Chandler (1996) A. Luzar and D. Chandler, Nature 379, 55 (1996). * Prada-Gracia, Shevchuk, and Rao (2013) D. Prada-Gracia, R. Shevchuk, and F. Rao, J. Chem. Phys. 139, 084501 (2013). * Shi, Russo, and Tanaka (2018b) R. Shi, J. Russo, and H. Tanaka, J. Chem. Phys. 149, 224502 (2018b). * Li _et al._ (2020b) L. Li, J. Zhong, Y. Yan, J. Zhang, J. Xu, J. S. Francisco, and X. C. Zeng, Proc. Natl. Acad. Sci. USA 117, 24701 (2020b). * de Oca _et al._ (2020) J. M. M. de Oca, F. Sciortino, , and G. A. Appignanesi, J. Chem. Phys. 152, 244503 (2020). * Skinner _et al._ (2013) L. B. Skinner, C. Huang, D. Schlesinger, L. G. M. Pettersson, A. Nilsson, and C. J. Benmore, J. Chem. Phys. 138, 074506 (2013). * Soper and Benmore (2008) A. K. Soper and C. J. Benmore, Phys. Rev. Lett. 101, 065502 (2008). * Praprotnik, Janežič, and Mavri (2004) M. Praprotnik, D. Janežič, and J. Mavri, J. Phys. Chem. A 108, 11056 (2004). * MacKerell Jr. _et al._ (1998) A. D. MacKerell Jr., D. Bashford, M. Bellott, R. L. D. Jr, J. D. Evanseck, M. J. Field, S. Fischer, J. Gao, H. Guo, S. Ha, D. Joseph-McCarthy, L. Kuchnir, K. Kuczera, F. T. K. Lau, C. Mattos, S. Michnick, T. Ngo, D. T. Nguyen, B. Prodhom, W. E. Reiher, B. Roux, M. Schlenkrich, J. C. Smith, R. Stote, J. Straub, M. Watanabe, J. Wiórkiewicz-Kuczera, D. Yin, and M. Karplus, J. Phys. Chem. B 102, 3586 (1998). * Calero, Stanley, and Franzese (2016) C. Calero, E. H. Stanley, and G. Franzese, Materials 9, 319 (2016). * Samatas _et al._ (2018) S. Samatas, C. Calero, F. Martelli, and G. Franzese, arXiv:1811.01911 [cond-mat.soft] (2018). * Gallo, Rovere, and Chen (2010) P. Gallo, M. Rovere, and S.-H. Chen, J. Phys. Chem. Lett. 1, 729 (2010). * Camisasca, Marzio, and Gallo (2020) G. Camisasca, M. D. Marzio, and P. Gallo, J. Chem. Phys. 153, 224503 (2020). * Iorio, Camisasca, and Gallo (2019) A. Iorio, G. Camisasca, and P. Gallo, Sci. China Phys. Mech. Astron. 62, 107011 (2019). * Iorio _et al._ (2020) A. Iorio, M. Minozzi, G. Camisasca, M. Rovere, and P. Gallo, Philosophical Magazine 100, 2582 (2020). * Tenuzzo, Camisasca, and Gallo (2020) L. Tenuzzo, G. Camisasca, and P. Gallo, Molecules 25, 4570 (2020). * Calero and Franzese (2020) C. Calero and G. Franzese, J. Mol. Liq. 317, 114027 (2020).
top=1in, bottom=1in, left=1in, right=1in universitá degli studi di pavia dipartimento di fisica corso di laurea magistrale in scienze fisiche Bit Commitment in Operational Probabilistic Theories Tesi per la Laurea Magistrale di Lorenzo Giannelli Chiar.mo Prof. Giacomo Mauro D'Ariano Dott. Alessandro Tosini Anno Accademico 2019-2020 A Piero The aim of this thesis is to investigate the bit commitment protocol in the framework of operational probabilistic theories. In particular a careful study is carried on the feasibility of bit commitment in the non-local boxes theory and in order to do this new aspects of the theory are presented. Lo scopo di questa tesi è di investigare il protocollo di bit commitment all'interno delle teorie probabilistiche operazionali. In particolare si è analizzato attentamente la fattibilità del protocollo all'interno della teoria dei non-local boxes e i nuovi aspetti della teoria emersi in questa analisi sono presentati. CHAPTER: INTRODUCTION The study of quantum foundations is a discipline of science that seeks to understand the most characterizing aspects of quantum theory, to reformulate it and even propose new generalizations. An active area of research in quantum foundations is therefore to find alternative formulations of quantum theory which rely on physically compelling principles in attempt to find a re-derivation of the quantum formalism in terms of operational axioms. One of the most interesting effort in this direction is made investigating the relations between the current operational axioms and the main results of quantum information theory. As a recent example it has been proved in Ref. [1] that the no information without disturbance (NIWD) theorem, i.e. the impossibility in quantum theory to extract information without disturbing the state of the system or its correlations with other systems, is independent of both local discriminability and purification, two of the defining axioms of quantum theory. Especially the latter, as we will see thoroughly in the first Chapter, is considered as a characteristic and distinctive quantum trait but now NIWD can be exhibited in absence of it and also of most of the principles of quantum theory. The NIWD property spawns other no-go theorems, which represent some of the most famous and classic results in quantum information theory. Among these results one can certainly list the no-cloning theorem, the no-programming and the impossibility of perfectly secure bit commitment. In our thesis our efforts are focused on the latter. A bit commitment (BC) protocol is meant to allow one party, Alice, to send a bit to a second party, Bob, in such a way that Bob cannot read the bit until Alice allows for its disclosure, while Alice cannot change the value of the bit after she encoded it. The bit commitment protocol is a very important primitive in cryptography, and perfectly secure protocols are known to be impossible in classical information theory. This is the case in quantum theory as well. The proof involves a very important characterization theorem for general theories with purification [2]. Numerous bit commitment protocols have been proposed in literature and this cryptography primitive has been deeply studied, especially the possibility of unconditionally secure bit commitment, both in quantum and classical information theory, due to its importance in practical applications. We would like to investigate the relations between the theorem of no-bit commitment and the operational axioms that characterize quantum theory. A general strategy to apprehend the nature of these links is to test the validity of the theorem in a theory that lacks one or more principles; naturally the first try has to be made excluding the purification, the most quantum feature. The answer of how to start in the analysis comes directly from the literature on bit commitment protocol itself. In fact, after it was proved to be impossible in quantum theory the protocol has been tested in more general scenarios, in particular in more non-local scenarios. To understand what it means we need to take a step back to the two parties of the protocol, Alice e Bob. Assume that they are not able to communicate but have access to physical states that they can use to generate joint correlations. In this experiment the outcome of the measurements on the state of their local systems are given by random variables. Obviously, causality constrains the correlations to be non signalling, and on the other side quantum theory prevents the strength of the non-local correlation to violate Bell's inequalities [3], where the maximal value is known as Cirel'son's bound [4]. A well-known variant of a Bell inequality is the Clauser, Horne, Shimony & Holt (CHSH) inequality [5], which can be expressed as [6] \begin{equation*} \sum_{x,y\in\{0,1\}}\mathsf{Pr}(a_x\oplus b_y=x\cdot y)\le2\sqrt{2}\,. \end{equation*} Where $x$ and $y$ denote the choice of Alice's and Bob's measurement, respectively, $a_x\in\{0,1\}$, $b_y\in\{0,1\}$ the respective binary outcomes, and $\oplus$ addition modulo 2. However, if we only care about the causality constrain, Cirel'son's bound can be violated up to the maximal value of 4. Popescu and Rohrlich, who first noticed [7], raised the question of why nature is not more non-local and why does quantum mechanics not allow a stronger violation of the CHSH inequality. Following this lead, bit commitment has been studied in Popescu Rohrlich (PR) non-local boxes theory. It has been claimed, as in Ref. [8], that bit commitment was admissible in PR-box theory but a counter-proof reached from Short, Gisin, and Popescu [9]. It was formulated on the behalf that, since non-locality is the main reason to prevent bit commitment in quantum theory, it would not be possible that in a theory more non-local than the quantum one BC would work. However, thereafter a new protocol was proposed by Buhrman et al. [10] and the argument of the counter-proof of Short, Gisin and Popescu is not able to deny it. The first aim of our work is to bring some clarity, pointing out the limitations of the framework of validity about the results that have been claimed until now. In our path new aspects of PR-box theory emerged and have been studied. Even if the theory is still far to be considered complete, important characterizations have been made. We will show that all the BC protocols proposed so far respect the limitations of PR-box theory under which we are able to prove a no-bit commitment theorem. Furthermore we are able to create, with very similar arguments, a counter-proof of the protocol proposed in Ref. [10]. Finally, a surprising result is observed. Relaxing the constraint of PR-box theory and including some recent developments, a scheme of bit commitment that is perfectly secure seems possible. Even if there are actually no "operational" evidences within the theory to deny it, by the same fundamental reason expressed by Short, Gisin, and Popescu, we think that future advancements of the theory would reconsider the protocol as cheatable. However, as a matter of fact, now the answer to the existence of a theory with entanglement that admits bit commitment, seems to be “Yes!”. A synopsis of the thesis is the following: In the first Chapter the framework of operational probabilistic theories (OPTs) is presented by first introducing the operational language that expresses the possible connections between events, and then by dressing the elements of the language with a probabilistic structure. After that, the principles for the OPT of quantum mechanics are stated (given the framework, the rule of connectivity among events are given). In the second Chapter, we will discuss the bit commitment protocol. We will start with an historical perspective and then we will rigorously define the protocol in the language of OPT. We will mainly deal only with perfectly secure bit commitment since our analysis is carried in the OP framework and there are not yet the technical tools to analyse unconditionally secure BC in OPTs. In the third section of the Chapter, we will study the proof of the impossibility of perfectly secure bit commitment in quantum theory done in Ref. [2]. This is a very elegant and solid demonstration and our purpose is to adapt the proof in order to comprehend other theories than the quantum one. As we will see, non-locality and entanglement are the key reasons that plays in favor of the impossibility of BC and so it seems reasonable to try to extend the impossibility proof also to other non-local theories. In the third Chapter, we will analyse the probabilistic theory corresponding to the popular PR-box model. A comprehensive study has never been made and an organic theory is not available. We propose to bring clarity to the actual model considering only bipartite correlated boxes, highlighting its limitations, and some progress by analysing new aspects such as perfect discriminability and states purification. We will also add some considerations and prevision on $N$-partite correlated boxes. Finally, in the fourth Chapter we propose a proof of impossibility of perfectly secure bit commitment in PR-boxes. Even if the PR-box model does not constitute a proper theory it has been often used in numerous application in literature, such as in protocol of perfectly or unconditionally secure BC. However this led to neglect important elements of the theory and in return numerous results published can be proved false when contextualized in the PR-box OPT (also if it is still incomplete). As the last remark, we conjecture that including tripartite correlated boxes in the theory, secure bit commitment would be possible. The reason is the following. We will see that what prevents BC in quantum theory is that every state can be purified and its purification is unique up to local reversible transformations. Exactly these local operations allow the cheating in the protocol. When we deal with PR-box theory limited to bipartite correlated boxes, only one internal state can be purified, but its purification is unique and this is enough to allow cheating. On the contrary, when we admit tripartite correlated boxes, the uniqueness of purification is lost and exactly this uncommon feature could open the door to the possibility of secure bit commitment. CHAPTER: OPERATIONAL PROBABILISTIC THEORY The purpose of this Chapter is to introduce the framework of operational probabilistic theories (OPTs) and to express quantum theory as an OPT. In this Chapter we will follow Ref. [11]. The framework of operational probabilistic theories consists of two distinct conceptual ingredients: an operational structure, describing circuits that produce outcomes, and a probabilistic structure, which assigns probabilities to the outcomes in a consistent way. The operational structure summarizes all the possible circuits that can be constructed in a given physical theory, in this setting a rigorous formulation of the elements of the circuits: systems, transformations, and their composition is given, which constitutes the grammar for the probabilistic description of an experiment. However, it is only the probabilistic structure that promotes the operational language from a merely descriptive tool to a framework for predictions, the predictive power being the crucial requirement for any scientific theory and for its testability - the essence of science itself. Different OPTs will have different rules for assigning the joint probabilities of events. Working in this framework allows us to deal with a wide range of probabilistic theories, including not only quantum and classical theories, but also the theory of Popescu-Rohrlich (PR), or non-local, boxes. In the first part of the Chapter the framework is provided by first introducing the operational language that expresses the possible connections between events, and then by dressing the elements of the language with a probabilistic structure. Then, in the second part of the Chapter we formulate the principles for the OPT of quantum theory. In fact, once the framework is defined, the rule of connectivity among events are given. § THE FRAMEWORK A theory for making predictions about joint events depending on their reciprocal connections is what we call an operational probabilistic theory. We see that OPT is a non-trivial extension of probability theory. To the joint events we associate not only their joint probability, but also a circuit that connects them. When the events in the circuit have a well-defined order, the circuit is mathematically described by a directed acyclic graph (a graph with directed edges and without loops). The basic element of an OPT - the notion of event - gets dressed with wires that allow us to connect it with other events. Such wires are the systems of the theory. In agreement with the directed nature of the graph, there are input and output systems. The events are the transformations, whereas the transformations with no input system are the states (corresponding to preparations of systems), and those with no output system are the effects (corresponding to observations of systems). Since the purpose of a single event is to describe a process connecting an input with an output, the full circuit associated to a probability is a closed one, namely a circuit with no input and no output. The circuit framework is mathematically formalized in the language of category theory. In this language, an OPT is a category, whose systems and events are objects and arrows, respectively. Every arrow has an input and an output object, and arrows can be sequentially composed. The associativity, existence of a trivial system, and commutativity of the parallel composition of systems of quantum theory technically correspond to having a strict symmetric monoidal category. §.§ Primitive notions and notation The primitive notions of any operational theory are those of test, event, and system. A test $\{\tA_i\}_{i\in \rI}$ is the collection of events $\tA_i$, where $i$ labels the element of the outcome space I. In addition to comprising a collection of events, the notion of test carries also the event connectivity of the theory that is achieved by the systems. These can represent the input and the output of the test. The resulting representation of a test is the following diagram: \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em { \poloFantasmaCn{\rA}\qw& \gate{\{\tE_x\}_{x\in \rX}}& \poloFantasmaCn{\rB}\qw& \qw \end{aligned} \end{equation*} The wire on the left labeled as A represents the input system, whereas the wire on the right labeled as B represents the output system. The same diagrammatic representation is also used for any of the events, namely for $x\in \rX$ \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em { \poloFantasmaCn{\rA}\qw& \gate{\tE_x}& \poloFantasmaCn{\rB}\qw& \qw \end{aligned} \end{equation*} In the following, the systems will be denoted by capital Roman letters A,B,...,Z, whereas the events by capital calligraphic letters $\mathcal{A,B,}...,\mathcal{Z}$. Different tests can be combined in a circuit, which is a directed acyclic graph where the links are the systems (oriented from left to right, namely from input to output) and the nodes are the boxes of the tests. The same graph can be built up for a single test instance, namely with the network nodes being events instead of tests, corresponding to a joint outcome for all tests. The circuit graph is obtained precisely by using the following rules. Sequential Composition of Test When the output system of test $\{\mathcal{C}_x\}_{x\in \rX}$ and the input system of test $\{\tD_y\}_{y\in \rY}$ coincide, the two tests can be composed in sequence as follows: \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em { \poloFantasmaCn{\rA}\qw& \gate{\{\tC_x\}_{x\in \rX}}& \poloFantasmaCn{\rB}\qw& \gate{\{\tD_y\}_{y\in \rY}}& \poloFantasmaCn{\rC}\qw& \qw \, \eqqcolon \, \Qcircuit @C=1em @R=.7em { \poloFantasmaCn{\rA}\qw& \gate{\{\tE_{(x,y)}\}_{(x,y)\in \rX\times \rY}}& \poloFantasmaCn{\rC}\qw& \qw \end{aligned} \end{equation*} resulting in the test $\{\tE_{(x,y)}\}_{(x,y)\in \rX\times \rY}$ called sequential composition of $\{\tC_x\}_{x\in \rX}$ and $\{\tD_y\}_{y\in \rY}$. In formulas we will also write $\tE_{(x,y)}:=\tD_y\tC_x$. Identity Test For every system A, one can perform the identity test (shortly identity) that "leaves the system alone". Formally, this is the deterministic test $\mathcal{I}_\textnormal{A}$ with the property \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { \poloFantasmaCn{\rA}\qw& \gate{\mathcal{I}_\rA}& \poloFantasmaCn{\rA}\qw& \gate{\mathcal{C}}& \poloFantasmaCn{\rB}\qw& \qw \end{aligned} \, =\, \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { \poloFantasmaCn{\rA}\qw& \gate{\mathcal{C}}& \poloFantasmaCn{\rB}\qw& \qw \end{aligned}\,, \end{equation*} \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { \poloFantasmaCn{\rB}\qw& \gate{\mathcal{D}}& \poloFantasmaCn{\rA}\qw& \gate{\mathcal{I}_\rA}& \poloFantasmaCn{\rA}\qw& \qw \end{aligned} \, =\, \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { \poloFantasmaCn{\rB}\qw& \gate{\mathcal{D}}& \poloFantasmaCn{\rA}\qw& \qw \end{aligned}\,, \end{equation*} where the above identities must hold for any event $ \Qcircuit @C=.5em @R=.5em { & \poloFantasmaCn{\rA} \qw & \gate{\mathcal{C}} & \poloFantasmaCn{\rB} \qw &\qw} $ and $ \Qcircuit @C=.5em @R=.5em { & \poloFantasmaCn{\rB} \qw & \gate{\mathcal{D}} & \poloFantasmaCn{\rA} \qw &\qw} $,respectively. The sub-index A will be dropped from $\mathcal{I}_\text{A}$ where there is no ambiguity. Operationally Equivalent Systems We say that two systems A and $\text{A}^\prime$ are operationally equivalent - denoted as $\text{A}^\prime\simeq \text{A}$ or just $\text{A}^\prime=\text{A}$ - if there exist two deterministic events $ \Qcircuit @C=.5em @R=.5em { & \poloFantasmaCn{\rA} \qw & \gate{\mathcal{U}} & \poloFantasmaCn{\rA^\prime} \qw &\qw} $ and $ \Qcircuit @C=.5em @R=.5em { & \poloFantasmaCn{\rA^\prime} \qw & \gate{\mathcal{V}} & \poloFantasmaCn{\rA} \qw &\qw} $ such that \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R {& \poloFantasmaCn{\rA} \qw & \gate{\mathcal{U}} & \poloFantasmaCn{\rA^\prime} \qw & \gate{\mathcal{V}} & \poloFantasmaCn{\rA} \qw & \qw} \end{aligned} \ =\ \begin{aligned} \Qcircuit @C=1em @R=.7em @! R {& \poloFantasmaCn{\rA} \qw & \gate{\mathcal{I}} & \poloFantasmaCn{\rA} \qw & \qw} \end{aligned}\,, \end{equation*} \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R {& \poloFantasmaCn{\rA^\prime} \qw & \gate{\mathcal{V}} & \poloFantasmaCn{\rA} \qw & \gate{\mathcal{U}} & \poloFantasmaCn{\rA^\prime} \qw & \qw} \end{aligned} \ =\ \begin{aligned} \Qcircuit @C=1em @R=.7em @! R {& \poloFantasmaCn{\rA^\prime} \qw & \gate{\mathcal{I}} & \poloFantasmaCn{\rA^\prime} \qw & \qw} \end{aligned}\,. \end{equation*} Accordingly, if $\{\tC_x\}_{x\in \rX}$ is any test for system A, performing an equivalent test on system $\text{A}^\prime$ means performing the test $\{\tC^\prime_x\}_{x\in \rX}$ defined as \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R {& \poloFantasmaCn{\rA^\prime} \qw & \gate{\mathcal{C}^\prime_x} & \poloFantasmaCn{\rA^\prime} \qw & \qw} \end{aligned} \ =\ \begin{aligned} \Qcircuit @C=1em @R=.7em @! R {& \poloFantasmaCn{\rA^\prime} \qw & \gate{\mathcal{V}} & \poloFantasmaCn{\rA} \qw & \gate{\mathcal{C}_x} & \poloFantasmaCn{\rA} \qw & \gate{\mathcal{U}} & \poloFantasmaCn{\rA^\prime} \qw & \qw} \end{aligned}\,. \end{equation*} Composite System Given two systems A and B, one can join them into the single composite system AB. As a rule, the system AB is operationally equivalent to the system BA, and we will identify them in the following. This means that the system composition is commutative, \begin{equation*} \text{AB}=\text{BA}. \end{equation*} We will call a system trivial system, reserving for it the letter I, if it corresponds to the identity in the system composition, namely \begin{equation*} \text{AI}=\text{IA}=\text{A}. \end{equation*} The trivial system corresponds to having no system, namely I carries no information. Finally we require the composition of systems to be associative, namely \begin{equation*} \text{A(BC)}=\text{(AB)C}. \end{equation*} In other words, if we iterate composition on many systems we always end up with a composite system that only depends on the components, and not on the particular composition sequence according to which they have been composed. Systems then make an Abelian monoid. A test with input system AB and output system CD represents an interaction process (see the parallel composition of tests in the following). Parallel Composition of Tests Any two tests $ \Qcircuit @C=.5em @R=.5em { & \poloFantasmaCn{\rA} \qw & \gate{\{\mathcal{C}_x\}_{x\in \rX}} & \poloFantasmaCn{\rB} \qw &\qw} $ $ \Qcircuit @C=.5em @R=.5em { & \poloFantasmaCn{\rC} \qw & \gate{\{\mathcal{D}_y\}_{y\in \rY}} & \poloFantasmaCn{\rD} \qw &\qw} $ can be composed in parallel as follows: \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R {& \poloFantasmaCn{\rA} \qw & \gate{\{\mathcal{C}_x\}_{x\in \rX}} & \poloFantasmaCn{\rB} \qw &\qw\\ & \poloFantasmaCn{\rC} \qw & \gate{\{\mathcal{D}_y\}_{y\in \rY}} & \poloFantasmaCn{\rD} \qw &\qw} \end{aligned} \ =:\ \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { & \poloFantasmaCn{\rA\rC} \qw & \gate{\{\tF_{(x,y)}\}_{(x,y)\in \rX\times \rY}} & \poloFantasmaCn{\rB\rD} \qw &\qw } \end{aligned}\,. \end{equation*} The test $ \Qcircuit @C=.8em @R=.5em { & \poloFantasmaCn{\rA\rC} \qw & \gate{\{\tF_{(x,y)}\}_{(x,y)\in \rX\times \rY}} & \poloFantasmaCn{\rB\rD} \qw &\qw } $ is the parallel composition of tests $ \Qcircuit @C=.5em @R=.5em { & \poloFantasmaCn{\rA} \qw & \gate{\{\mathcal{C}_x\}_{x\in \rX}} & \poloFantasmaCn{\rB} \qw &\qw} $ and $ \Qcircuit @C=.5em @R=.5em { & \poloFantasmaCn{\rC} \qw & \gate{\{\mathcal{D}_y\}_{y\in \rY}} & \poloFantasmaCn{\rD} \qw &\qw} $, where $\{\tF_{(x,y)}\}_{(x,y)\in \rX\times \rY}\equiv\{\tC_x\otimes\tD_y\}_{(x,y)\in \rX\times \rY}$. Parallel and sequential composition of tests commute, namely one has \begin{equation} \label{c:parallel composition} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { & \poloFantasmaCn{\rA} \qw & \gate{\tC_z} & \poloFantasmaCn{\rB} \qw & \gate{\tA_x} & \poloFantasmaCn{\rC} \qw & \qw \\ & \poloFantasmaCn{\rD} \qw & \gate{\tB_y} & \poloFantasmaCn{\rE} \qw & \gate{\tD_w} & \poloFantasmaCn{\rF} \qw & \qw \gategroup{1}{3}{2}{3}{.7em}{--} \gategroup{1}{5}{2}{5}{.7em}{--}} \end{aligned} \ =\ \begin{aligned} \Qcircuit @C=1em @R=1.2em @! R { & \poloFantasmaCn{\rA} \qw & \gate{\tC_z} & \poloFantasmaCn{\rB} \qw & \gate{\tA_x} & \poloFantasmaCn{\rC} \qw & \qw \\ & \poloFantasmaCn{\rD} \qw & \gate{\tB_y} & \poloFantasmaCn{\rE} \qw & \gate{\tD_w} & \poloFantasmaCn{\rF} \qw & \qw \gategroup{1}{3}{1}{5}{.7em}{--} \gategroup{2}{3}{2}{5}{.7em}{--}} \end{aligned}\,. \end{equation} When one of the two operation is the identity, we wull omit the identity box and drawn only a straight line: \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { & \poloFantasmaCn{\rA} \qw & \gate{\tC_x} & \poloFantasmaCn{\rB} \qw & \qw \\ & \qw & \poloFantasmaCn{\rC} \qw & \qw & \qw \end{aligned} \end{equation*} Therefore, as a consequence of commutation between sequetial and parallel composition, we have the following identity: \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { & \poloFantasmaCn{\rA} \qw & \gate{\tC_x} & \qw & \poloFantasmaCn{\rB} \qw & \qw \\ & \poloFantasmaCn{\rC} \qw & \qw & \gate{\tD_y} & \poloFantasmaCn{\rD} \qw & \qw \end{aligned} \ =\ \begin{aligned} \Qcircuit @C=1em @R=1.2em @! R { & \poloFantasmaCn{\rA} \qw & \qw & \gate{\tC_x} & \poloFantasmaCn{\rB} \qw & \qw \\ & \poloFantasmaCn{\rC} \qw & \gate{\tD_y} & \qw & \poloFantasmaCn{\rD} \qw & \qw \end{aligned}\,. \end{equation*} Preparation Tests and Observation Tests Tests with a trivial input system are called preparation tests, and tests with a trivial output system are called observation tests. They will be represented as follows: \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R {& \prepareC {\{\rho_x\}_{x\in \rX}} & \poloFantasmaCn{\rB} \qw & \qw \end{aligned} \ :=\ \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { & \poloFantasmaCn{\rI} \qw & \gate{\{\rho_x\}_{x\in \rX}} & \poloFantasmaCn{\rB} \qw & \qw \end{aligned}\,, \end{equation*} \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { & & \poloFantasmaCn{\rA} \qw & \measureD{\{a_y\}_{y\in \rY}} \end{aligned} \ :=\ \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { & \poloFantasmaCn{\rA} \qw & \gate{\{a_y\}_{y\in \rY}} & \poloFantasmaCn{\rI} \qw & \qw \end{aligned}\,. \end{equation*} The corresponding events will be called preparation events and observation events. In formulas we will also write $|\rho_i)_\text{A}$ to denote a preparation event and $(a_j|_\text{A}$ to denote an observation event. Closed Circuits Using the above rules we can build up closed circuits, i.e. circuits with no input and no output system. An example is given by the following circuit: \begin{equation} \label{c:closed1} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { \multiprepareC{3}{\{\Psi_{i}\}}& \qw\poloFantasmaCn{\rA}& \multigate{1}{\{\tA_{j}\}}& \qw\poloFantasmaCn{\rB}& \gate{\{\tC_{l}\}}& \qw\poloFantasmaCn{\rC}& \multigate{1}{\{\tE_{n}\}}& \qw\poloFantasmaCn{\rD}& \multimeasureD{2}{\{\tG_{q}\}} \\ \pureghost{\{\Psi_{i}\}}& \qw\poloFantasmaCn{\rE}& \ghost{\{\tA_{j}\}}&\qw\poloFantasmaCn{\rF}& \multigate{1}{\{\tD_{m}\}}& \qw\poloFantasmaCn{\rG}& \ghost{\{\tE_{n}\}} \\ \pureghost{\{\Psi_{i}\}}& \qw\poloFantasmaCn{\rH}& \multigate{1}{\{\tB_{k}\}}& \qw\poloFantasmaCn{\rL}& \ghost{\{\tD_{m}\}}&\qw\poloFantasmaCn{\rM}& \multigate{1}{\{\tF_{p}\}}& \qw\poloFantasmaCn{\rN}&\pureghost{\{\tG_{q}\}}\qw \\ \pureghost{\{\Psi_{i}\}}& \qw\poloFantasmaCn{\rO}& \ghost{\{\tB_{k}\}}& \qw& \qw\poloFantasmaCn{\rP}& \qw& \ghost{\{\tF_{p}\}} \\ \end{aligned} \end{equation} where we omitted the probability spaces of each test. Independent Systems For any (generally open) circuit constructed according to the above rules we call a set of systems independent if for each couple of systems in the set the two are not connected by a unidirected path (i.e. following the arrow from the input to the output). For example, in Eq. (<ref>) the sets {A,E}, {H,O}, {A,E,H,O}, {A,L}, {A,E,L,P} are independent, whereas e.g. the sets {A,M}, {A,B}, {A,E,N} are not. A maximal set of independent systems is called a slice. We are now in position to move towards the general purpose of an operational probabilistic theory: predicting and accounting for the joint probability of events corresponding to a particular circuit of connections. Given a closed circuit, as in Eq. (<ref>), we are left with just a joint probability distribution. Therefore, to a closed circuit of event as the following: \begin{equation} \label{c:closed2} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { \multiprepareC{3}{\Psi_{i}}& \qw\poloFantasmaCn{\rA}& \multigate{1}{\tA_{j}}& \qw\poloFantasmaCn{\rB}& \gate{\tC_{l}}& \qw\poloFantasmaCn{\rC}& \multigate{1}{\tE_{n}}& \qw\poloFantasmaCn{\rD}& \multimeasureD{2}{\tG_{q}} \\ \pureghost{\Psi_{i}}& \qw\poloFantasmaCn{\rE}& \ghost{\tA_{j}}&\qw\poloFantasmaCn{\rF}& \multigate{1}{\tD_{m}}& \qw\poloFantasmaCn{\rG}& \ghost{\tE_{n}} \\ \pureghost{\Psi_{i}}& \qw\poloFantasmaCn{\rH}& \multigate{1}{\tB_{k}}& \qw\poloFantasmaCn{\rL}& \ghost{\tD_{m}}&\qw\poloFantasmaCn{\rM}& \multigate{1}{\tF_{p}}& \qw\poloFantasmaCn{\rN}&\pureghost{\tG_{q}}\qw \\ \pureghost{\Psi_{i}}& \qw\poloFantasmaCn{\rO}& \ghost{\tB_{k}}& \qw& \qw\poloFantasmaCn{\rP}& \qw& \ghost{\tF_{p}} \\ \end{aligned} \end{equation} we will associate a joint probability $p(i,j,k,l,m,n,p,q)$ which we will consider as parametrically dependent on the circuit, namely, for a different choice of events and/or different connections we will have a different joint probability. Since we are interested only in the joint probabilities and their corresponding circuits, we will build up probabilistic equivalence classes, and define: Two events from system A to system B are equivalent if they occur with the same joint probability with the other events within any circuit. We will call transformation from A to B - denoted as $\tA\in\mathsf{Transf}(\rA\rightarrow\rB)$ - the equivalence class of events from A to B that are equivalent in the above sense. Likewise we will call instrument an equivalence class of tests, state an equivalence class of preparation events, and effect an equivalence class of observation events. We will denote the set of states of system A as $\mathsf{St}(\rA)$, and the set of its effects as $\mathsf{Eff}(\rA)$. Clearly, the input systems belonging to two different elements of an equivalence class will be operationally equivalent, and likewise for output systems. We now can define an operational probabilistic theory as follows: An operational probabilistic theory (OPT) is a collection of systems and transformations, along with rules for composition of systems and parallel and sequential composition of transformations. The OPT assigns a joint probability to each closed circuit. Therefore, in an OPT every test from the trivial system I to itself is a probability distribution $\{p_i\}_{i\in \rX}$ for the set of joint outcomes X, with $p(i):=p_i\in\left[0,1\right]$ and $\sum_{i\in\rX}p(i)=1$. Compound events from the trivial system to itself are independent, namely their joint probability is given by the product of the respective probabilities for both the parallel and the sequential composition, namely \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R {& \prepareC{\rho_{i_1}} & \poloFantasmaCn{\rA} \qw & \measureD{a_{i_2}}\\ & \prepareC{\sigma_{j_1}} & \poloFantasmaCn{\rB} \qw & \measureD{b_{j_2}} \end{aligned} \ =\ \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { \prepareC{\rho_{i_1}} & \poloFantasmaCn{\rA} \qw & \measureD{a_{i_2}} & \prepareC{\sigma_{j_1}} & \poloFantasmaCn{\rB} \qw & \measureD{b_{j_2}} \end{aligned} \,. \end{equation*} A special case of OPT is the deterministic OPT, where all probabilities are 0 or 1. §.§ States and effects Using the parallel and sequential composition of transformation it follows that any closed circuit can be regarded as the composition of a preparation event and an observation event, for example the circuit in Eq. (<ref>) can be cut along a slice as follows: \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { \multiprepareC{3}{\Psi_{i}}& \qw\poloFantasmaCn{\rA}& \multigate{1}{\tA_{j}}& \qw\poloFantasmaCn{\rB}& \qw \\ \pureghost{\Psi_{i}}& \qw\poloFantasmaCn{\rE}& \ghost{\tA_{j}}&\qw\poloFantasmaCn{\rF}& \multigate{1}{\tD_{m}}& \qw\poloFantasmaCn{\rG}& \\ \pureghost{\Psi_{i}}& \qw\poloFantasmaCn{\rH}& \multigate{1}{\tB_{k}}& \qw\poloFantasmaCn{\rL}& \ghost{\tD_{m}}&\qw\poloFantasmaCn{\rM}& \multigate{1}{\tF_{p}}& \qw \\ \pureghost{\Psi_{i}}& \qw\poloFantasmaCn{\rO}& \ghost{\tB_{k}}& \qw& \qw\poloFantasmaCn{\rP}& \qw& \ghost{\tF_{p}} \\ \end{aligned} \ + \ \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { & \qw\poloFantasmaCn{\rB}& \gate{\tC_{l}}& \qw\poloFantasmaCn{\rC}& \multigate{1}{\tE_{n}}& \qw\poloFantasmaCn{\rD}& \multimeasureD{2}{\tG_{q}} \\ & & & \qw\poloFantasmaCn{\rG}& \ghost{\tE_{n}} \\ & & & & & \qw\poloFantasmaCn{\rN}& \pureghost{\tG_{q}}\qw \\ \\ \end{aligned} \end{equation*} and thus is equivalent to the following state-effect circuit: \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R {& \prepareC{(\Psi_i,\tA_j,\tB_k,\tD_m,\tF_p)} & \poloFantasmaCn{\rB\rG\rN} \qw & \measureD{(\tC_l,\tE_n,\tG_q)} \end{aligned} \end{equation*} Therefore, a state $\rho\in\st{A}$ is a functional over effects $\eff{A}$, the functional being denoted with the pairing $(a|\rho)$ with $a\in\eff{A}$ and analogously an effect $a\in\eff{A}$ is a functional over states $\st{A}$. By taking linear combinations of functionals we see that $\st[R]{A}:=\mathsf{Span}_\mathbb{R}\left[\st{A}\right]$ and $\eff[R]{A}:=\mathsf{Span}_\mathbb{R}\left[\eff{A}\right]$ are dual spaces, and states are positive linear functionals over effects, and effects are positive linear functional over states ($\st[R]{A}$ and $\eff[R]{A}$ are assume finite dimensional and we denote as $D_\rA:=\text{dim}\st[R]{A}\equiv\text{dim}\eff[R]{A}$ also called size of system A). In the following we also denote by $\mathsf{St}_1(\rA)$ and $\mathsf{Eff}_1(\rA)$ the sets of deterministic states and effects, respectively. According to the above definition, two states are different if and only if there exists an effect which occurrs on them with different joint probabilities. We also have that two effects are different if and only if there exists a state on which they have different probabilities. In particular, given two states $\rho_0\ne\rho_1\in\st{A}$ we will say that an effect $a\in\eff{A}$ separates the state $\rho_0$ and $\rho_1$ when $(\rho_1|a)\ne(\rho_0|a)$, namely when the effect occurs with different joint probabilities over the two states (the analogous relation holds for separable states respect to effects). Therefore we conclude that: States are separating for effects and effects are separating for states. It is possible to demonstrate that in any convex OPT if two states (effects) $\rho_0$,$\rho_1\in\st{A}$ ($a_0,a_1\in\eff{A}$) are distinct, then one can discriminate them with error probability strictly smaller than $\frac{1}{2}$. §.§ Transformations From what we said before, the following circuit is a state of system BFHO: \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { \multiprepareC{3}{\Psi}& \qw\poloFantasmaCn{\rA}& \multigate{1}{\tA}& \qw\poloFantasmaCn{\rB}& \qw \\ \pureghost{\Psi}& \qw\poloFantasmaCn{\rE}& \ghost{\tA}&\qw\poloFantasmaCn{\rF}& \qw \\ \pureghost{\Psi}& \qw\poloFantasmaCn{\rH}& \qw \\ \pureghost{\Psi}& \qw\poloFantasmaCn{\rO}& \qw& \\ \end{aligned} \end{equation*} This means that any transformation connected to some output systems of a state maps the state into another state of generally different systems. Thus, while states and effects are linear functionals over each other, we can always regard a transformation as a map between states. In particular, a transformation $\tT\in\transf{A}{B}$ is always associated to a map $\hat{\tT}$ from $\st{A}$ to $\st{B}$, uniquely defined as \begin{equation*} \hat{\tT}:|\rho)\in\st{A}\mapsto\hat{\tT}|\rho)=|\tT\rho)\in\st{B}. \end{equation*} Similarly the transformation can be associated to a map from $\eff{A}$ to $\eff{B}$. The map $\hat{\tT}$ can be linearly extended to a map from $\st[R]{A}$ to $\st[R]{B}$. Notice that the linear extension of $\tT$ (which we will denote by the same symbol) is well defined. In fact, a linear combination of states of A is null - in formula $\sum_{i}c_i|rho_i)=0$ - if and only if $\sum_{i}c_i(a|rho_i)=0$ for every $a\in\eff{A}$, and since for every $b\in\eff{B}$ we have $(b|\tT\in\eff{A}$, then $(b|\tT(\sum_{i}c_i|\rho_i))=\sum_{i}c_i(b|\tT|\rho_i)=(b|\sum_{i}c_i\tT|\rho_i)=0$, and finally We want to stress that if two transformations $\tT,\tT^\prime\in\transf{A}{B}$ correspond to the same map $\hat{\tT}$ from $\st{A}$ to $\st{B}$, this does not mean that the two transformations are the same, since as an equivalence class, they must occur with the same joint probability in all possible circuits. In terms of state mappings, the same definition of the transformation as equivalence class corresponds to say that $\tT,\tT^\prime\in\transf{A}{B}$ as maps from states of AR to states of BR are the same for all possible systems R of the theory, namely $\tT=\tT^\prime\in\transf{A}{B}$ if and only if \begin{equation} \label{eq:equal maps} \forall\rR,\;\forall\Psi\in\st{AR}\quad \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { \multiprepareC{1}{\Psi}& \qw\poloFantasmaCn{\rA}& \gate{\tT}& \qw\poloFantasmaCn{\rB}& \qw \\ \pureghost{\Psi}& \qw\poloFantasmaCn{\rR}& \qw&\qw&\qw \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { \multiprepareC{1}{\Psi}& \qw\poloFantasmaCn{\rA}& \gate{\tT^\prime}& \qw\poloFantasmaCn{\rB}& \qw \\ \pureghost{\Psi}& \qw\poloFantasmaCn{\rR}& \qw&\qw&\qw \end{aligned}\,. \end{equation} Indeed, there exist cases of OPT where there are transformations $\tT=\tT^\prime\in\transf{A}{B}$ corresponding to the same map when applied to $\st{A}$ and not when applied to $\st{AR}$ for some system R, a relevant example is fermionic theory [12, 13]. Since we can take linear combinations of linear transformations, $\transf{A}{B}$ can be embedded in the vector space $\transf{A}{B}$. The deterministic transformations, whose set will be denoted as $\mathsf{Transf}_1(\rA\rightarrow\rB)$, will be also called channels. Finally, a transformation $\tU\in\transf{A}{B}$ is reversible if there exists another transformation $\tU^{-1}\in\transf{B}{A}$ such that $\tU^{-1}\tU = \tI_\rA$ and $\tU\tU^{-1}=\tI_\rB$. The set of reversible transformations from A to B will be denoted by $\textsf{RevTransf}(\rA\rightarrow\rB)$. §.§ Coarse-graining and refinement When dealing with probabilistic events, a natural notion is that of coarse-graining, corresponding to merging events into a single event. According to probability theory, the probability of a coarse-grained event $\rS\subseteq\rX$ subset of the outcome space X is the sum of probabilities of the elements of S, namely $p(\rS)=\sum_{i\in\rS}p(i)$. We then correspondingly have that the coarse-grained event $\tT_\rS$ of a test $\{\tT_i\}_{i\in\rX}$ will be given by \begin{equation} \label{eq:coarse-grained event} \tT_\rS=\sum_{i\in\rS}\tT_i. \end{equation} We stress that the equal sign in Eq. (<ref>) is to be meant in the sense of Eq. (<ref>). In addition to the notion of coarse-grained event we have also that of coarse-grained test, corresponding to the collection of a coarse-grained events $\{\tT_{\rX_l}\}_{l\in\rZ}$ from a partition $\rX=\cup_{l\in\rZ}\rX_l$ of the outcome space $\rX$, with $\rX_i\cap\rX_j=\emptyset$ for $i\neq j$. The converse procedure of coarse-graining is what we call refinement. If $\tT_\rS$ the coarse-graining in Eq. (<ref>), we call any sum $\sum_{i\in\rS^\prime}\tT_i$ with $\rS^\prime\subseteq\rS$ a refinement of $\tT_\rS$. The same notion can be analogously considered for a test. Intuitively, a test that refines another is a test that extracts more detailed information, namely it is a test with better "resolving power". The notion of refinement is translated to transformations (hence also to states, and effects), as equivalence classes of events. Refinement and coarse-graining define a partial ordering in the set of transformations $\transf{A}{B}$, writing $\tD\prec\tC$ if $\tD$ is a refinement of $\tC$. A transformation $\tC$ is atomic if it has only trivial refinement, namely $\tC_i$ refines $\tC$ implies that $\tC_i=p\tC$ for some probability $p\ge0$. A test that consists of atomic transformations is a test whose "resolving power" cannot be further improved. It is often useful to refer to the set of all possible refinements of a given event $\tC$. This set is called refinement set of the event $\tC\in\transf{A}{B}$, and is denoted by $\textsf{RefSet}(\tC)$. In formula, $\textsf{RefSet}(\tC):=\{\tD\in\transf{A}{B}|\tD\prec\tC\}.$ In the special case of states, we will use the word pure as a synonym of atomic. A pure state describes an event providing maximal knowledge about the system's preparation, namely a knowledge that cannot be further refined (we will denote with $\mathsf{PurSt}(\rA)$ the set of pure states of system $\rA$). As usual, a state that is not pure will be called mixed. An important notion is that of internal state. A state is called internal when any other state can refine it: precisely, $\omega\in\st{A}$ is internal if for every $\rho\in\st{A}$ there is a non-zero probability $p>0$ such that $p\rho$ is a refinement of $\omega$, i.e. $p\rho\in\mathsf{RefSet}(\omega)$. The adjective "internal" has a precise geometric connotation, since the state cannot belong to the border of $\st{A}$. An internal state describes a situation in which there is no definite knowledge about the system preparation, namely a priori we cannot in principle exclude any possible preparation. § QUANTUM THEORY AS AN OPT In this section we provide an overview of the six principles used for constructing quantum theory as an OPT. All features of quantum theory - ranging from the superposition principle, entanglement, no cloning, teleportation, Bell's inequalities violation, quantum cryptography - can be understood and proved using only the principles, without using Hilbert spaces. However, our aim is only to introduce the principle and analyse them from an operative point of view. All the six principles are operational, in that they stipulate whether or not certain tasks can be accomplished: they set the rules of the game for all the experiments and all the protocols that can be carried out in the theory. They also provide a great insight into the worldview at which quantum theory hints. We review the list of the principles: * Atomicity of composition * Perfect discriminability * Ideal compression * Causality * Local discriminability * Purification All six principles, with the exception of purification, express standard features that are shared by both classical and quantum theory. The principle of purification picks up uniquely quantum theory among the theories allowed by the first five, partly explaining the magic of quantum information. §.§ Atomicity of composition In the general framework we encountered the notions of coarse-grained and atomic operation. A coarse-grained operation is obtained by joining together outcomes of a test, corresponding to neglect some information. The inverse process of coarse-graining is that of refining. An atomic operation is one where no information has been neglected, namely an operation that cannot be refined. When the operation is atomic, the experimenter has maximal knowledge of what’s happening in the lab. A test consisting of atomic operations represents the highest level of control achievable according to our theory. The principle of atomicity of composition states that it possible to maintain such a level of control throughout a sequence of experiments, stating precisely what follows: [Atomicity of composition] The sequence of two atomic operations is an atomic operation. One of the immediate consequences granted by atomicity of composition is the following: Given two pure states $\alpha\in\st{A}$ and $\beta\in\st{B}$, the parallel composition of $\alpha$ and $\beta$ is a pure state of $\st{AB}$. §.§ Perfect discriminability Two deterministic states $\rho_0$ and $\rho_1$ are perfectly discriminable if there exists a measurement $\{m_y\}_{y\in\{0,1\}}$ such that $$(m_y|\rho_x)=\delta_{xy}\quad\forall x,y\in\{0,1\}.$$ The existence of perfectly discriminable states is important, because these states can be used to communicate classical information without errors. In a communication protocol, the sender can encode the value of a bit $x$ into the state $\rho_x$ and then transmit the system to the receiver, who can decode the value of the bit using the measurement $\{m_y\}_{y\in\{0,1\}}$. The perfect discriminability axiom ensures that our ability to discriminate states is as sharp as it could possibly be: except for trivial cases, every state can be perfectly discriminated from some other state. The "trivial cases" are those states that cannot be discriminated from anything else because they contain every other state in their convex decomposition. We can call them internal, or completely mixed. [Perfect discriminability] Every deterministic state that is not completely mixed is perfectly discriminable from some other state. As anticipated, the perfect discriminability axiom guarantees that every non-trivial system has at least two perfectly discriminable states: In a theory satisfying perfect discriminability, every physical system has at least two perfectly discriminable states, unless the system is trivial (i.e. it has only one deterministic state). Pick a pure state $\alpha\in\st{A}$. If $\alpha$ is not internal, then perfect discriminability guarantees that $\alpha$ is perfectly discriminable from some other state $\alpha^\prime$, hence A has two perfectly discriminable states. If $\alpha$ is internal every pure state belongs to its refinement set. Moreover, since it is also pure, i.e. extremal, one has that every other deterministic state $\rho_1\in\mathsf{St}_1(\rA)$ must be equal to $\alpha$, i.e. A has only one deterministic state. An easy consequence of this result is that the theory can describe noiseless classical communication. §.§ Ideal compression Ideal compression garantes that information can be transferred faithfully from one system to another. Namely, suppose that Alice has a preparation device, which prepares system A in some state $\alpha$. Alice does not know the state $\alpha$, but she knows that on average the device prepares the deterministic state $\rho\in\mathsf{St}_1(\rA)$. Now, suppose Alice wants to transfer the state of her system to Bob's laboratory, but unfortunately she cannot send system A directly. Instead, she has to encode the state $\alpha$ into the state of another system B, by applying a suitable deterministic operation $\tE$ (the encoding), which transforms the state $\alpha$ into the state We say that the encoding is lossless for the state $\rho$ iff there exists another deterministic operation $\tD$ (the decoding) such that $$\tD\tE\alpha=\alpha\quad\forall\alpha\in F_\rho$$ where $F_\rho$ is the refinement set of $\rho$, which is made of the set of all states $\alpha$ that are compatible with $\rho$ (on the convex set of states this would be the face to which $\rho$ belongs). This third axiom establishes the possibility of a particular type of lossless encoding, called ideal compression. The ultimate limit to the lossless compression of a given state $\rho$ is reached when every state of the encoding system B is a codeword for some state in $F_\rho$, namely when every state $\beta\in\st{B}$ is of the form $\tE\alpha$ for some $\alpha\in F_\rho$. When this is the case, we say that the compression is efficient, and we call the triple $(\rB,\tE,\tD)$ an ideal compression protocol. [Ideal Compression] Every state can be compressed in a lossless and efficient way. §.§ Causality The causality axiom identifies the input–output ordering of a circuit with the direction along which information flows, identifying such ordering with a proper-time arrow, corresponding to the request that future choices cannot influence the present. The probability of the outcome of a preparation test is independent of the choice of observation tests connected at its output. To better understand the statement it is useful to consider the joint test consisting of a preparation test $\tX=\{\rho_i\}_{i\in\rX}\subset\st{A}$ followed by the observation test $\tY=\{a_j\}_{j\in\rY}\subset\eff{A}$ performed on system A: \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { \prepareC{\tX}&\qw& \qw\poloFantasmaCn{\rA}&\qw& \measureD{\tY} \end{aligned} \end{equation*} The joint probability of preparation $\rho_i$ and observation $a_j$ is given by \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { \prepareC{\rho_i}&\qw& \qw\poloFantasmaCn{\rA}&\qw& \measureD{a_j} \end{aligned} \end{equation*} The marginal probability of the preparation alone does not depend on the outcome $j$. Yet, it generally depends on which observation test $\tY$ is performed, namely $$\sum_{a_j\in\rY}(a_j|\rho_i)\eqqcolon p(i|\tX,\tY)\,.$$ The marginal probability of preparation $\rho_i$ is then generally conditioned on the choice of the observation test $\tY$. What the causality axiom states is that $p(i|\tX,\tY)$ is actually independent of $\tY$, namely for any two different observation tests $\tY=\{a_j\}_{j\in\rY}$ and $\tZ=\{b_k\}_{k\in\rZ}$ one has In a causal OPT the choice of a test on a system can be conditioned on the outcomes of a preceding test, since causality guarantees that the probability distribution of the preceding test is independent of the choice of the following test. This leads us to introduce the notion of conditioned test. If $\{\tA_i\}_{i\in\rX}$ is a test from $\rA$ to $\rB$, and $\{\tB_j^{(i)}\}_{j\in\rY_i}$ is a test from $\rB$ to $\rC$ for every $i\in\rX$, then the conditioned test is a test from $\rA$ to $\rC$, with outcomes $(i,j)\in\rZ:=\cup_i\{\{i\}\times\rY_i\}$, and events $\{\tB_j^{(i)}\circ\tA_i\}_{(i,j)\in\rZ}$. Diagrammatically, the events $\tB_j^{(i)}\circ\tA_i$ are represented as follows: \begin{equation*} \begin{aligned} \Qcircuit @C=1em @R=.7em @! R { & \qw\poloFantasmaCn{\rA}& \gate{\tA_i}& \qw\poloFantasmaCn{\rB}& \gate{\tB_j^{(i)}}& \poloFantasmaCn{\rC}\qw& \qw \end{aligned} \end{equation*} Among conditioned test, a special role is played by the observe-and-prepare test,where the "connecting" system is the null system I. They are thus made of a preparation test conditioned by an observation test, as follows: \begin{equation*} \begin{aligned} \Qcircuit @C=.5em @R=.7em @! R { & \qw\poloFantasmaCn{\rA}& \measureD{l_i}& \poloFantasmaCn{\rC}\qw& \qw \end{aligned} \end{equation*} which can be also represented as $\{|\omega^{(i)})(l_i|\}_{i\in\rX}$. Another remarkable way to characterize causal theories is to require the unicity of the deterministic effect, the equivalence of this formulation with Axiom <ref> is given by the following lemma. An OPT is causal if and only if for every system $\rA$ there is a unique deterministic effect. We will prove the two directions separately, namely: (1) if the probability of preparation of states is independent of the observation test, then the deterministic effect is unique; (2) vice versa. (1) The probability of the preparation $\rho$ is given by the marginal of the joint probability with the observation, namely $p(\rho)=\sum_{i\in\rX}(a_i|\rho)$. Upon denoting the deterministic effects of two different tests as $a=\sum_{i\in\rX}a_i$ and $b=\sum_{j\in\rY}b_j$, the statement that the preparation probability is independent of the observation tests translates to $(a|\rho) = (b|\rho)$ for every preparation $\rho\in\st{A}$, which implies that $a=b$, since the set of states is separating for events. (2) Uniqueness of the deterministic effect implies that the preparation probability of each state is independent of the test, since the effect $a=\sum_{i\in\rX}a_i$ for any test $\{a_i\}_{i\in\rX}$ is deterministic, and $(a|\rho)$ for any deterministic effect $a\in\eff{a}$ is the probability of preparation $\rho$. We will denote the unique deterministic effect for system A as $e_\rA$, and the subindex will be dropped when no confusion can arise. In the following we will use the notation $\le$ to denote the partial ordering between effects, defined as follows: $$a,b\in\eff{A},\,a\le b\quad\Leftrightarrow\quad(a|\rho)\le(b|\rho),\,\forall\rho\in\st{A}\,.$$ It is immediate to show that the causality condition of Lemma <ref> spawns the following lemmas. Causality is equivalent to the following statements regarding tests: * Completeness of observation tests: For any system $\rA$ and for every observation test $\{a_i\}_{i\in\rX}$ one has * Completeness of tests: For any systems $\rA$, $\rB$ and for every test $\{\tC_i\}_{i\in\rX}$ from $\rA$ to $\rB$ one has * Domination of transformations: For any systems $\rA$, $\rB$ a transformation $\tC\in\transf{A}{B}$ satisfies the condition with the equality if and only if $\tC$ is a channel, i.e. a deterministic transformation corresponding to a single-outcome test. * Domination of effects: For any system $\rA$ all effects are dominated by a unique effect $e_\rA$ which is deterministic $$\forall a\in\eff{A},\quad0\le a\le e_\rA\,.$$ An immediate consequence of uniqueness of the deterministic effect is the identification of all transformations of the form for any observation test $\{a_i\}_{i\in\rX}$ of system A. In particular, we have the factorization of the deterministic effect of composite systems $$e_{\rA\rB}=e_\rA\otimes e_\rB.$$ The uniqueness of the deterministic effect naturally leads to the relevant notion of marginal state or also called local state. The marginal state of $|\sigma)_{\rA\rB}$ on system $\rA$ is the state represented by the diagram \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\sigma_{\rA\rB}}& \poloFantasmaCn{\rA}\qw& \qw \\ \pureghost{\sigma_{\rA\rB}}& \poloFantasmaCn{\rB}\qw& \measureD{e_\rB} \\ \end{aligned}\, \eqqcolon \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \prepareC{\rho_{\rA}}& \poloFantasmaCn{\rA}\qw& \qw \end{aligned} \end{equation*} Finally, the last implication of causality we would outline is the impossibility of signaling without interaction, i.e. by just performing local tests. In a causal OPT it is impossible to send signals by performing only local tests. Suppose the general situation in which two "distant" parties Alice and Bob share a bipartite state $|\Psi)_{\rA\rB}$ of systems A and B. Alice performs her local test $\{\tA_i\}_{i\in\rX}$ on system A and similarly Bob performs his local test $\{\tB_j\}_{j\in\rY}$ on system B. The joint probability of their outcomes is The marginal probabilities $p_i^\rA$ at Alice and $p_j^\rB$ at Bob are given by $$p_i^\rA\coloneqq\sum_{j}p_{ij},\quad p_j^\rB\coloneqq\sum_{i}p_{ij}.$$ Alice's marginal does not depend on the choice of test $\{\tB_j\}$ of Bob, since \begin{equation*} \begin{aligned} \end{aligned} \end{equation*} where we used Eq. (<ref>) and the normalization condition $\sum_{j}(e|_{\rB}\tB_j=(e|_{\rB}$. The same argument holds for Bob’s marginal. §.§ Local discriminability Now we introduce the principle of local discriminability, which stipulates the possibility of discriminating states of composite systems via local measurements on the component systems. [Local discriminability] It is possible to discriminate any pair of states of composite systems using only local measurements. Mathematically the axiom asserts that for every two joint states $\rho,\sigma\in\st{AB}$, with $\rho\ne\sigma$, there exist effects $a\in\eff{A}$ and $b\in\eff{B}$ such that the joint probabilities for the two states are different, namely, in circuits \begin{equation} \label{eq:loc-discr} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\rho}& \poloFantasmaCn{\rA}\qw& \qw \\ \pureghost{\rho}& \poloFantasmaCn{\rB}\qw& \qw \\ \end{aligned} \,\ne\, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\sigma}& \poloFantasmaCn{\rA}\qw& \qw \\ \pureghost{\sigma}& \poloFantasmaCn{\rB}\qw& \qw \\ \end{aligned} \, \Longrightarrow \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\rho}& \poloFantasmaCn{\rA}\qw& \measureD{a} \\ \pureghost{\rho}& \poloFantasmaCn{\rB}\qw& \measureD{b} \\ \end{aligned} \,\ne\, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\sigma}& \poloFantasmaCn{\rA}\qw& \measureD{a} \\ \pureghost{\sigma}& \poloFantasmaCn{\rB}\qw& \measureD{b} \\ \end{aligned}\,. \end{equation} We can now prove one of the main theorem following from the principle of local discriminability. A theory satisfies local discriminability if and only if, for every composite system $\rA\rB$, one has \begin{equation} \label{eq:product-rule-for-composite-systems} D_{\rA\rB}=D_\rA D_\rB\,. \end{equation} By Eq. (<ref>), a theory satisfies local discriminability if and only if local effects $a\otimes b\in\eff{AB}$, with $a\in\eff{A}$ and $b\in\eff{B}$, are separating for joint states $\st{AB}$. Equivalently, the set $T\coloneqq\{a\otimes b|a\in\eff{A},b\in\eff{B}\}$ is a spanning set for $\mathsf{Eff}_\mathbb{R}(\rA\rB)$. Since the dimension of $\mathsf{Span}_\mathbb{R}(T)$ is $D_\rA D_\rB$ and the spaces of states and effects have the same dimension, we have $D_{\rA\rB}=D_\rA D_\rB$. Conversely, if Eq. (<ref>) holds, then the product effects are a spanning set for the vector space $\mathsf{Eff}_\mathbb{R}(\rA\rB)$, hence they are separating, and local discriminability holds. Along with the axiom of local discriminability we introduce the notion of entangled and separable states, where entangled states are defined, by negation, as those states that are not separable. Given $n$ systems $\rA_1,\rA_2,\dots,\rA_n$, the separable states of the composite system $\rA_1\rA_2\dots\rA_n$ are those of the form \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{3}{\Sigma}& \poloFantasmaCn{\rA_1}\qw& \qw \\ \pureghost{\Sigma}& \poloFantasmaCn{\rA_2}\qw& \qw \\ \pureghost{\Sigma}& \vdots&& \\ \pureghost{\Sigma}& \poloFantasmaCn{\rA_n}\qw& \qw \\ \end{aligned} \,=\, \sum_{i\in\rX}p_i \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \prepareC{\alpha_{1_i}}& \poloFantasmaCn{\rA}\qw& \qw \\ \prepareC{\alpha_{2_i}}& \poloFantasmaCn{\rA}\qw& \qw \\ \vdots \\ \prepareC{\alpha_{n_i}}& \poloFantasmaCn{\rA}\qw& \qw \end{aligned} \end{equation*} where for $j=1,2,\dots,n$, $\alpha_{j_i}\in\st{A_j}$ $\forall i\in\rX$. §.§ Purification Purification is the really distinctive and fundamental trait of quantum theory, in the sense that purification allows to distinguish it between all the other possible theories (all the ones we can think of). The statement of the axiom is the following. For every system $\rA$ and for every state $\rho\in\st{A}$, there exists a system $\rB$ and a pure state $\Psi\in\mathsf{PurSt}(\rA\rB)$ such that \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \prepareC{\rho}& \poloFantasmaCn{\rA}\qw& \qw \end{aligned} \,=\, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi}& \poloFantasmaCn{\rA}\qw& \qw \\ \pureghost{\Psi}& \poloFantasmaCn{\rB}\qw& \measureD{e} \\ \end{aligned}\,. \end{equation*} If two pure states $\Psi$ and $\Psi^\prime$ satisfy \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi^\prime}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Psi^\prime}& \poloFantasmaCn{\rB}\qw& \measureD{e}\\ \end{aligned} \,=\, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi}& \poloFantasmaCn{\rA}\qw& \qw \\ \pureghost{\Psi}& \poloFantasmaCn{\rB}\qw& \measureD{e} \\ \end{aligned}\,, \end{equation*} then there exists a reversible transformation $\tU$, acting only on system $\rB$, such that \begin{equation} \label{eq:uniq purification} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi^\prime}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Psi^\prime}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned} \,=\, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi}& \qw& \poloFantasmaCn{\rA}\qw& \qw& \qw \\ \pureghost{\Psi}& \poloFantasmaCn{\rB}\qw& \gate{\tU}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned}\,\,. \end{equation} Here we say that $\Psi$ is a purification of $\rho$ and that $\rB$ is the purifying system. The property stated in Eq. <ref> is called uniqueness of purification and refers to the case in which the two purifications have the same purifying system. It can be easily generalized (the purifying systems are different): If two pure states $\Psi\in\mathsf{PurSt}(\rA\rB)$ and $\Psi^\prime\in\mathsf{PurSt}(\rA\rB^\prime)$ are purifications of the same mixed state, then \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi^\prime}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Psi^\prime}& \poloFantasmaCn{\rB^\prime}\qw& \qw\\ \end{aligned} \,=\, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi}& \qw& \poloFantasmaCn{\rA}\qw& \qw& \qw \\ \pureghost{\Psi}& \poloFantasmaCn{\rB}\qw& \gate{\tC}& \poloFantasmaCn{\rB^\prime}\qw& \qw\\ \end{aligned} \end{equation*} for some deterministic transformation $\tC$ transforming system $\rB$ into system $\rB^\prime$. Pick two pure states $\beta\in\mathsf{PurSt}(\rB)$ and $\beta^\prime\in\mathsf{PurSt}(\rB^\prime)$. Since $\Psi\otimes\beta^\prime$ and $\Psi^\prime\otimes\beta$ are purifications of the same state on $\rA$, the uniqueness of purification implies \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi^\prime}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Psi^\prime}& \poloFantasmaCn{\rB^\prime}\qw& \qw\\ \prepareC{\beta}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned} \,=\, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi}& \qw& \poloFantasmaCn{\rA}\qw& \qw& \qw\\ \pureghost{\Psi}& \poloFantasmaCn{\rB}\qw& \multigate{1}{\tU}& \poloFantasmaCn{\rB^\prime}\qw& \qw\\ \prepareC{\beta^\prime}& \poloFantasmaCn{\rB^\prime}\qw& \ghost{\tU}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned} \end{equation*} for some reversible transformation $\tU$ (we have also assumed local discriminability). Discarding system $\rB$ on both sides we then obtain $\Psi^\prime = (\tI_\rA\otimes\tC)\Psi$, where $\rC$ is the deterministic transformation defined by \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { & \poloFantasmaCn{\rB}\qw& \gate{\tC}& \poloFantasmaCn{\rB^\prime}\qw& \qw \end{aligned} \,\coloneqq \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { & \poloFantasmaCn{\rB}\qw& \multigate{1}{\tU}& \poloFantasmaCn{\rB^\prime}\qw& \qw\\ \prepareC{\beta}& \poloFantasmaCn{\rB^\prime}\qw& \ghost{\tU}& \poloFantasmaCn{\rB}\qw& \measureD{e}\\ \end{aligned}\,. \end{equation*} CHAPTER: BIT COMMITMENT Bit commitment is a cryptographic primitive involving two mistrustful parties, conventionally called Alice and Bob. Alice is supposed to submit an encoded bit of information to Bob in such a way that Bob has (in principle) no chance to identify the bit before Alice later decodes it for him, whereas Alice has (in principle) no way of changing the value of the bit once she has submitted it. In other words, Bob is interested in binding Alice to some commitment, whereas Alice would like to conceal her commitment from Bob. In the first two sections of this Chapter we will describe the protocol: we will start from an historical perspective, from the first article published by Blum in 1983 [14] to the most recent developments of the last decade that focus on the impossibility of bit commitment in quantum theory [15, 16]. Then, we will rigorously define the protocol in the language of OPTs. As every other cryptographic primitive, bit commitment does not need to be perfectly secure, i.e. probability of cheating equals to 0 for Alice and 1/2 for Bob (who can always randomly guess), to be efficient. In fact, even if a greater probability of error is admitted, iterating the protocol the error probability can be generally asymptotically reduced. A protocol that admits this possibility is called unconditionally secure, with a vast literature on the subject. However in our thesis we will only deal with perfectly secure bit commitment. Since we are considering the protocol within the operational framework, there are not yet the technical tools to analyse unconditionally secure bit commitment in OPTs and, anyway, the perfectly secure protocol should be the starting point for a rigorous analysis. The exceptional thing is that in our analysis we will be anyway able to build a cheating scheme also for some unconditionally secure bit commitment protocol, as we will see in Section <ref>. Finally, in the third part of the Chapter, we will study the proof of the impossibility of perfectly secure bit commitment in OP quantum theory done in Ref. [2]. We will review if all of the six axioms of quantum theory introduced in Section <ref> are really necessary conditions. Some of them will be neglected and other will be replaced with weaker hypothesis. With the new sufficient conditions that we will find, the proof of impossibility of perfectly secure bit commitment can be extended to other theories than the quantum one. Some applications hypothesis are the fermionic theory and the real quantum theory and in particular the PR-box theory. To the latter will be focused the next two Chapter, in fact, in literature numerous bit commitment protocols (and more generally quantum key distributions protocols) have been studied in the context of non-local correlated box, i.e. PR-boxes. So our analysis will provide a solid (operational probabilistic) point of view from which study all these protocols that have been prosed in the past years. § FROM COIN TOSSING TO NO-GO THEOREM The bit commitment protocol was conceived for the first time by Blum in 1983 as a building block for secure coin tossing. To cite the abstract of the original work [14]: Alice and Bob want to flip a coin by telephone. (They have just divorced, live in different cities, want to decide who gets the car.) Bob would not like to tell Alice HEADS and hear Alice (at the other end of the line) say “Here goes... I'm flipping the coin .... You lost!”. A standard example to illustrate bit commitment is for Alice to write the bit down on a piece of paper, which is then locked in a safe and sent to Bob, whereas Alice keeps the key. At a later time, she will unveil it by handing over the key to Bob. However, Bob has a well-equipped toolbox at home and may have been able to open the safe in the meantime. So while this scheme may offer reasonably good practical security, it is in principle insecure. Yet all bit commitment schemes that have wide currency today rely on such technological constraints: not on strongboxes and keys, but on unproven assumptions that certain computations are hard to perform. The first example of quantum bit commitment was first proposed by Bennet and Brassard in their famous paper of 1984 [17] as a primitive for implementing coin tossing. In their scheme, Alice commits to a bit value by preparing a sequence of photons in either of two mutually unbiased bases, in a way that the resulting quantum states are indistinguishable to Bob. The authors show that their protocol is secure against so-called passive cheating, in which Alice initially commits to the bit value $k$ and then tries to unveil $1-k$ later. However, they also prove that Alice can cheat with a more sophisticated strategy, in which she initially prepares pairs of maximally entangled states instead, keeps one particle of each pair in her laboratory and sends the second particle to Bob. For the first time, entanglement is recognized as a crucial factor in preventing perfectly secure bit commitment. Subsequent proposals for bit commitment schemes tried to evade this type of attack by forcing the players to carry out measurements and communicate classically as they go through the protocol. At a 1993 conference Brassard et al. presented a bit commitment protocol [18] that was claimed and generally accepted to be unconditionally secure. In 1996 Lo and Chau [19], and Mayers [20] realized that all previously proposed bit commitment protocols were vulnerable to a generalized version of the (EPR) attack that renders the BB84 proposal insecure, a result that they slightly extended to cover quantum bit commitment protocols in general. In OP terms, the determinant factor in the impossibility of bit commitment shifts from entanglement to the purification principle, this will be fully proved in Ref [2]. Their basic argument is the following. At the end of the commitment phase, Bob will hold one out of two quantum states $\psi_k$ as proof of Alice's commitment to the bit value $k\in\{0, 1\}$. Alice holds its purification $\Psi_k$, which she will later pass on to Bob to unveil. For the protocol to be concealing, the two states $k$ should be (almost) indistinguishable, $\psi_0\approx\psi_1$. But Uhlmann's theorem then implies the existence of a unitary transformation $\tU$ that (nearly) rotates the purification of $\psi_0$ into the purification of $\psi_1$. Since $\tU$ is localized on the purifying system only, which is entirely under Alice's control, Lo-Chau-Mayers argue that Alice can switch at will between the two states, and is not in any way bound to her commitment. As a consequence, any concealing bit commitment protocol is argued to be necessarily non-binding. Starting from 2000 the Lo-Chau-Mayers no-go theorem has been continually challenged, arguing that the impossibility proof does not exhaust all conceivable quantum bit commitment protocols. Several protocols have been proposed and claimed to circumvent the no-go theorem. These protocols seek to strengthen Bob's position with the help of "secret parameters" or "anonymous states", so that Alice lacks some information needed to cheat successfully: while Uhlmann's theorem would still imply the existence of a unitary cheating transformation as described above, this transformation might be unknown to Alice. However, the above attempts to build up a secure quantum bit commitment protocol have motivated the thorough analysis of Ref. [15], which provided a strengthened and explicit impossibility proof exhausting all conceivable protocols in which classical and quantum information is exchanged between two parties, including the possibility of protocol aborts and resets. This proof encompasses protocols even with unbounded number of communication rounds (it is only required that the expected number of rounds is finite), and with quantum systems on infinite-dimensional Hilbert spaces. However, the considerable length of this proof made it hard to follow, lacking a synthetic intuition of the impossibility proof. Finally in 2009 Chiribella et al. in Ref. [16] provided a new short impossibility proof of quantum bit commitment. In Ref. [2] a similar demonstrative structure is used to prove the impossibility of perfectly secure bit commitment in an operational framework, not only in quantum theory, but in a wide range of theories with purification. This proof will be the one that we will study in the third Section of this Chapter. § A FORMAL DEFINITION A rigorous definition is in order for a twofold reason. If it is true that it is necessary to define a framework where to operate, it is at the same time important to clearly remark the definition of bit commitment to which the statement "bit commitment is impossible" refers to. Despite we already referred to the word protocol, we did not linger to define what we mean with it, therefore we will start by the notion of protocol and we will state the bit commitment in the OPT language and its key properties only thereafter. §.§ The protocol A protocol regulates the exchange of messages between participants, defining what are the honest strategies that they can adopt, so that at every stage it is clear what type of message is expected from the participants, although, of course, their content is not fixed. The expected message types can be either classical or quantum or a combination thereof. In any bit commitment protocol, we can distinguish two main phases: the first is the commitment phase, in which Alice and Bob exchange classical and quantum messages in order to commit the bit. The second is the opening phase where Alice will send to Bob some classical or quantum information in order to to reveal the bit value. Commitment phase this phase can end either with a successful commitment, or with an abort, in which the two parties irrevocably give up the purpose of committing the bit (of course, in a well designed protocol, if both parties are honest the probability of abort should be vanishingly small). If no abort took place, the bit value is considered to be committed to Bob but, supposedly, concealed from him. Since bit commitment is a two-party protocol and trusted third parties are not allowed, the starting state necessarily has to be originated by one of the two parties. Moreover, since we can always include in the protocol null steps (in which no information, classical or quantum, is exchanged), without loss of generality, we can restrict our attention to protocols that are started by Alice. Opening phase in the case of abort during the commitment, this is just a null step, whereas, in the case of successful commitment, at the opening Alice will send to Bob some classical or quantum information in order to to reveal the bit value. Taking both Alice's message and his own (classical and quantum) records, Bob will then perform a suitable verification measurement. His measurement will result in either a successful readout of the committed bit, or in a failure, e.g. due to the detection of an attempted cheat. Again, in a well-designed protocol the probability of failure should be vanishingly small. §.§ OP bit commitment Alice wants to commit a classical bit $b\in\{0,1\}$ to Bob. * as we have seen before, the first phase is the commitment phase, in which Alice and Bob can perform any sequence of operations. Depending on whether Alice intended to commit $b=0$ or $b=1$, the state $\Psi_0,\,\Psi_1\in\st{AB}$ is selected, respectively. We will assume that Alice and Bob's systems at the end of the commitment phase are A and B respectively, so that the two possible pure states that can be transmitted to Bob are $(e|_\rA\Psi_0,\,(e|_\rA\Psi_1\in\st{B}$; * between the commitment and the opening phase Alice and Bob can perform only local operations on their systems A and B, respectively (together with every other ancillary system they control); * during the opening phase, Alice transmits her system A to Bob who can perform any measurement on the joint system AB to know the bit $b$ and to check if it is compatible with the commitment of Alice. In general, to do this Bob can perform a two outcome measurement (Positive Operator Valued Measure, POVM) $\{a_0,a_1\}$ on the joint system AB. From now on, when we will refer to bit commitment we mean a protocol that is included in the previous scheme. Already at the beginning of this Chapter we intuitively mention the two way of cheating that can occur. Being a two party protocol, we can have Alice's cheating, i.e. she changes the bit after the commitment (the protocol is not binding), and Bob's cheating, i.e. Bob discovers the bit committed before the opening (the protocol is not concealing). In this way we have now identified two key properties of any bit commitment protocol. However there is a third key property that is often omitted in the literature: the correctness of the protocol that guarantees the correct verification of the bit committed in the opening phase. To recapitulate with a properer language, we will say that a bit commitment protocol is * binding: if, for honest Bob, Alice should not be able to change the bit she committed. More precisely, assume that a possibly dishonest Alice committed $b$ but she wants to reveal $b^\prime\ne b$, then it must be \begin{equation} \label{eq:binding} \textsf{Pr}\left[\text{ Bob accepts }|\text{ Alice reveals }b^\prime\,\right]<1\,; \end{equation} * concealing: if, for honest Alice, Bob should not be able to know the bit that Alice committed until she reveals it; * correct: for honest Alice and Bob, if Alice commits $b$ and later reveals $b$ to Bob, then Bob accepts with probability grater then $1/2$: \begin{equation} \label{eq:correctness} \textsf{Pr}\left[\text{ Bob accepts }|\text{ Alice reveals }b\,\right]>\frac{1}{2}\,. \end{equation} Furthermore, we will say that a bit commitment protocol is perfect if * it is perfectly binding, namely for honest Bob, Alice cannot switch between $\Psi_0$ and $\Psi_1$ in such a way that Bob cannot detect the switch with certainty (i.e. Eq. (<ref>) becomes $\textsf{Pr}\left[\text{ Bob accepts }|\text{ Alice reveals }b^\prime\,\right]=0$). Namely it does not exists a reversible channel $\tU$ such that \begin{equation} \label{eq:perfectly-binding} \end{equation} * it is perfectly concealing, namely for honest Alice, Bob would not be able to perform some measurement of his system B and gain at least partial information about which bit Alice committed \begin{equation} \label{eq:perfectly-concealing} \end{equation} * it is correct with probability one, namely Bob accepts with probability one. So Eq. (<ref>) becomes \begin{equation*} \textsf{Pr}\left[\text{ Bob accepts }|\text{ Alice reveals }b\,\right]=1\,. \end{equation*} In the perfect implementation of the protocol the probability of cheating of Bob must be $1/2$, since in the worst case he can always make a random guess of the committed bit. The cheating probability of Alice is instead equal to zero. As we have anticipated in Section <ref>, perfectly secure bit commitment protocol is impossible in quantum and classic theory. However, if the perfect concealing and perfect biding conditions are relaxed as follows \begin{equation*} \begin{aligned} \end{aligned} \end{equation*} interesting level of security relevant for concrete applications emerges. Within this scenario the main effort is in quantifying the cheating probabilities and their trade-off in operational terms. The goal is to achieve a protocol that is asymptotically binding and concealing (both the Alice and Bob cheating probability can be made arbitrarily small) against an adversary that has no restrictions on the computational resources. In this case one has unconditionally secure bit commitment. § LIGHTENING NO BIT COMMITMENT In this Section we will analyse the necessary assumptions to prove the impossibility of perfectly secure bit commitment. In the literature, a no-go theorem in quantum theory has been proven in Ref. [15, 16] but we found that not all of the axioms of quantum theory are necessary. So, using weaker assumption, we will generalize the impossibility of bit commitment to a larger set of operational probabilistic theories. Clearly the causality Axiom <ref> is essential, otherwise the very protocol could not be defined properly (there would not be a given order in the succeeding of the phases). In our analysis we also assume Axiom <ref>, atomicity of composition, that is a sufficient condition to grant Corollary <ref>. Finally, instead of the purification Axiom <ref>, we take the following weaker assumption. Before stating it, we introduce the notion of dynamically faithful state. A state $\sigma\in\st{AC}$ is dynamically faithful for system $\rA$ if for any couple of transformations $\tA,\,\tA^\prime\in\transf{A}{B}$ one has \begin{equation*} \tA|\sigma)_{\rA\rC}=\tA^\prime|\sigma)_{\rA\rC}\Longrightarrow\tA=\tA^\prime\, . \end{equation*} We are now in position to state the "new" axiom. For every system $\rA$ there exist a system $\tilde{\rA}$ and a pure state $\Psi^{(\rA)}\in\mathsf{St}(\rA\tilde{\rA})$ that is dynamically faithful for system $\rA$. Furthermore, for every system $\rB$ and for every bipartite state $R\in\mathsf{St}(\rB\tilde{\rA})$ such that \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{R}& \poloFantasmaCn{\rB}\qw& \measureD{e}\\ \pureghost{R}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi^{(\rA)}}& \poloFantasmaCn{\rA}\qw& \measureD{e}\\ \pureghost{\Psi^{(\rA)}}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw\\ \end{aligned}\,, \end{equation*} there exists a purification of $R$. If $\Phi$, $\Phi^\prime\in\mathsf{St}(\rC\rB\tilde{\rA})$ are two purifications of $R$ then they are connected by a reversible transformation $\tU\in\mathsf{Transf}(\rC)$. In quantum theory, the existence of mixed faithful states is a direct consequence of local discriminability. In a theory with both local discriminability and purification there exist also dynamically faithful states that are pure. In Axiom <ref> we do not require local discriminability nor purification but the existence of dynamically faithful pure states. In addition we also require that a purification exists only for those states that have the same marginal of the dynamically faithful pure ones and that this purification is unique up to a reversible transformation on the purifying system. An immediate consequence of Axioms <ref> and <ref> is the following lemma on the uniqueness of purification. Let $\Psi\in\st{AB}$ and $\Psi^\prime\in\st{AC}$ be two purification of $\rho\in\st{A}$. Then there exists a channel $\tC\in\transf{B}{C}$ such that \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi^\prime}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Psi^\prime}& \poloFantasmaCn{\rC}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi}& \qw& \poloFantasmaCn{\rA}\qw& \qw&\qw\\ \pureghost{\Psi}& \poloFantasmaCn{\rB}\qw& \gate{\tC}& \poloFantasmaCn{C}\qw& \qw\\ \end{aligned}\,. \end{equation*} Moreover, channel $\tC$ has the form \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rB}\qw& \gate{\tC}& \poloFantasmaCn{\rC}\qw& \qw \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \prepareC{\varphi_0}& \poloFantasmaCn{\rC}\qw& \multigate{1}{\tU}& \poloFantasmaCn{\rB}\qw& \measureD{e}\\ \poloFantasmaCn{\rB}\qw& \ghost{\tU}& \poloFantasmaCn{C}\qw& \qw\\ \end{aligned}\,. \end{equation*} for some pure state $\varphi_0\in\st{C}$ and some reversible channel $\tU\in\transf{B}{C}$. Let $\eta$ and $\varphi_0$ be an arbitrary pure state of B and C, respectively. Then, due to Axiom <ref>, $|\Psi^\prime)_{\rA\rC}|\eta)_{\rB}$ and $|\Psi)_{\rA\rB}|\varphi_0)_{\rC}$ are still two pure states and so they are both purifications of $\rho$ with the same purifying system $\rB\rC$. Due to Axiom <ref>, we have \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi^\prime}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Psi^\prime}& \poloFantasmaCn{\rC}\qw& \qw\\ \prepareC{\eta}& \poloFantasmaCn{\rB}\qw& \qw \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi}& \qw& \poloFantasmaCn{\rA}\qw& \qw&\qw\\ \pureghost{\Psi}& \poloFantasmaCn{\rB}\qw& \multigate{1}{\tU}& \poloFantasmaCn{\rC}\qw& \qw\\ \prepareC{\varphi_0}& \poloFantasmaCn{\rC}\qw& \ghost{\tU}& \poloFantasmaCn{B}\qw& \qw\\ \end{aligned}\,. \end{equation*} Applying the deterministic effect $e_\rB$ on system $\rB$ we obtain the thesis, with $\tC=(e|_\rB\;\tU|\varphi_0)$. §.§ Reversible dilation of channels Before starting the proof of the impossibility of bit commitment it is in order to derive some useful results about reversible dilations of channels. Let $R\in\mathsf{St}(\rB\tilde{\rA})$ be a state such that \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{R}& \poloFantasmaCn{\rB}\qw& \measureD{e}\\ \pureghost{R}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi^{(\rA)}}& \poloFantasmaCn{\rA}\qw& \measureD{e}\\ \pureghost{\Psi^{(\rA)}}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw\\ \end{aligned}\,. \end{equation*} where $\Psi^{(\rA)}\in\mathsf{St}(\rA\tilde{\rA})$ is a pure dynamically faithful state for system $\rA$. Then there exist a system $\rC$, a pure state $\varphi_0\in\mathsf{St}(\rB\rC)$, and a reversible channel $\tU\in\mathsf{Transf}(\rA\rB\rC)$ such that \begin{equation} \label{eq:revchannel} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{R}& \poloFantasmaCn{\rB}\qw& \qw\\ \pureghost{R}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \prepareC{\varphi_0}& \poloFantasmaCn{\rB\rC}\qw& \multigate{1}{\tU}& \poloFantasmaCn{\rA\rC}\qw& \measureD{e}\\ \multiprepareC{1}{\Psi^{(\rA)}}& \poloFantasmaCn{\rA}\qw& \ghost{\tU}& \poloFantasmaCn{B}\qw& \qw\\ \pureghost{\Psi^{(\rA)}}& \qw& \poloFantasmaCn{\tilde{\rA}}\qw& \qw&\qw\\ \end{aligned}\,. \end{equation} Moreover the channel $\tV\in\mathsf{Transf}(\rA\rightarrow\rA\rB\rC)$ defined by $\tV:=\tU|\varphi_0)_{\rB\rC}$ is unique up to reversible channels on $\rA\rC$. Take a purification of $R$, say $\Psi_R\in\mathsf{St}(\rC\rB\tilde{\rA})$ for some purifying system $\rC$ (existence of such a purification is granted by Axiom <ref>). One has \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{2}{\Psi_R}& \poloFantasmaCn{\rC}\qw& \measureD{e}\\ \pureghost{\Psi_R}& \poloFantasmaCn{\rB}\qw& \measureD{e}\\ \pureghost{\Psi_R}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{R}& \poloFantasmaCn{\rB}\qw& \measureD{e}\\ \pureghost{R}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi^{(\rA)}}& \poloFantasmaCn{\rA}\qw& \measureD{e}\\ \pureghost{\Psi^{(\rA)}}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw\\ \end{aligned} \end{equation*} that is, the pure state $\Psi_R$ and $\Psi^{(\rA)}$ have the same marginal on system $\tilde{\rA}$. Applying the uniqueness of purification expressed by Lemma <ref> one then obtains \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{2}{\Psi_R}& \poloFantasmaCn{\rC}\qw& \qw\\ \pureghost{\Psi_R}& \poloFantasmaCn{\rB}\qw& \qw\\ \pureghost{\Psi_R}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \prepareC{\varphi_0}& \poloFantasmaCn{\rB\rC}\qw& \multigate{1}{\tU}& \poloFantasmaCn{\rA}\qw& \measureD{e}\\ \multiprepareC{1}{\Psi^{(\rA)}}& \poloFantasmaCn{\rA}\qw& \ghost{\tU}& \poloFantasmaCn{\rB\rC}\qw& \qw\\ \pureghost{\Psi^{(\rA)}}& \qw& \poloFantasmaCn{\tilde{\rA}}\qw& \qw&\qw \end{aligned}\, . \end{equation*} Applying the deterministic effect on system $\rC$ on both sides, one then proves Eq. (<ref>). Moreover, if $\tV^\prime\coloneqq\tU^\prime|\varphi^\prime_0)_{\rB\rC}$ is channel such that Eq. (<ref>) holds, then the pure states $\tV|\Psi^{(\rA)})_{\rA\tilde{\rA}}$ and $\tV^\prime|\Psi^{(\rA)})_{\rA\tilde{\rA}}$ have the same marginal on system $\rB\tilde{\rA}$. Uniqueness of purification then implies \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { & & \amultigate{1}{\tV^\prime}& \poloFantasmaCn{\rA\rC}\qw& \qw\\ \multiprepareC{1}{\Psi^{(\rA)}}& \poloFantasmaCn{\rA}\qw& \ghost{\tV^\prime}& \poloFantasmaCn{\rB}\qw& \qw\\ \pureghost{\Psi^{(\rA)}}& \qw& \poloFantasmaCn{\tilde{\rA}}\qw& \qw&\qw \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { & & \amultigate{1}{\tV}& \poloFantasmaCn{\rA\rC}\qw& \gate{\tW}& \poloFantasmaCn{\rA\rC}\qw& \qw\\ \multiprepareC{1}{\Psi^{(\rA)}}& \poloFantasmaCn{\rA}\qw& \ghost{\tV}& \poloFantasmaCn{\rB}\qw& \qw&\qw&\qw\\ \pureghost{\Psi^{(\rA)}}& \qw& \poloFantasmaCn{\tilde{\rA}}\qw& \qw&\qw&\qw&\qw \end{aligned} \end{equation*} for some reversible channel $\tW\in\mathsf{Transf}(\rA\rC)$. Since $\Psi^{(\rA)}$ is dynamically faithful for $\rA$, this implies $\tV^\prime=\tW\tV$. We now give the definition of dilatation and reversible dilatation. A dilatation of channel $\tC\in\transf{A}{B}$ is a channel $\tV\in\transf{A}{BE}$ such that \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA}\qw& \gate{\tC}& \poloFantasmaCn{\rB}\qw& \qw \end{aligned}\, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { & & \amultigate{1}{\tV}& \poloFantasmaCn{\rE}\qw& \measureD{e}\\ \poloFantasmaCn{\rA}\qw& \ghost{\tV}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned}\, . \end{equation*} We refer to system $\rE$ as to the environment. A dilatation $\tV\in\transf{A}{BE}$ is called reversible if there exists a system $\rE_0$ such that $\rA\rE_0\simeq\rB\rE$ and \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { & & \amultigate{1}{\tV}& \poloFantasmaCn{\rE}\qw& \qw\\ \poloFantasmaCn{\rA}\qw& \ghost{\tV}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \prepareC{\varphi_0}& \poloFantasmaCn{\rE_0}\qw& \multigate{1}{\tU}& \poloFantasmaCn{\rE}\qw& \qw\\ \poloFantasmaCn{\rA}\qw& \ghost{\tU}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned} \end{equation*} for some pure state $\varphi_0\in\mathsf{St}(\rE_0)$ and some reversible channel $\tU\in\mathsf{Transf}(\rA\rE_0\rightarrow\rB\rE)$. According to the above definitions, we have the following dilatation theorem: Every channel $\tC\in\transf{A}{B}$ has a reversible dilatation $\tV\in\transf{A}{BE}$. If $\tV$, $\tV^\prime\in\transf{A}{BE}$ are two reversible dilatations of the same channel, then they are connected by a reversible transformation on the environment, namely \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { & & \amultigate{1}{\tV^\prime}& \poloFantasmaCn{\rE}\qw& \measureD{e}\\ \poloFantasmaCn{\rA}\qw& \ghost{\tV^\prime}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { & & \amultigate{1}{\tV}& \poloFantasmaCn{\rE}\qw& \measureD{e}\\ \poloFantasmaCn{\rA}\qw& \ghost{\tV}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned} \end{equation*} \begin{equation*} \Longrightarrow \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { & & \amultigate{1}{\tV^\prime}& \poloFantasmaCn{\rE}\qw& \qw\\ \poloFantasmaCn{\rA}\qw& \ghost{\tV^\prime}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { & & \amultigate{1}{\tV}& \poloFantasmaCn{\rE}\qw& \gate{\tW}& \poloFantasmaCn{\rE}\qw& \qw\\ \poloFantasmaCn{\rA}\qw& \ghost{\tV}& \qw& \poloFantasmaCn{\rB}\qw& \qw&\qw\\ \end{aligned} \end{equation*} for some reversible channel $\tW\in\mathsf{Transf}(\rE)$. Let us store the channel $\tC$ in the faithful state $\Psi^{(\rA)}\in\mathsf{St}(\rA\tilde{\rA})$, thus getting the state $R_\tC$: \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{R_\tC}& \poloFantasmaCn{\rB}\qw& \qw\\ \pureghost{R_\tC}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw\\ \end{aligned} \, \coloneqq \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi}& \poloFantasmaCn{\rA}\qw& \gate{\tC}& \poloFantasmaCn{\rB}\qw& \qw\\ \pureghost{\Psi}& \qw& \poloFantasmaCn{\tilde{\rA}}\qw& \qw&\qw\\ \end{aligned}\, . \end{equation*} Since $\tC$ is a channel, it satisfy the normalization condition \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA}\qw& \gate{\tC}& \poloFantasmaCn{\rB}\qw& \measureD{e} \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA}\qw& \measureD{e} \end{aligned}\, , \end{equation*} which implies \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{R_\tC}& \poloFantasmaCn{\rB}\qw& \measureD{e}\\ \pureghost{R_\tC}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi^{(\rA)}}& \poloFantasmaCn{\rA}\qw& \gate{\tC}& \poloFantasmaCn{\rB}\qw& \measureD{e}\\ \pureghost{\Psi^{(\rA)}}& \qw& \poloFantasmaCn{\tilde{\rA}}\qw& \qw&\qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi^{(\rA)}}& \poloFantasmaCn{\rA}\qw& \measureD{e}\\ \pureghost{\Psi^{(\rA)}}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw\\ \end{aligned}\, . \end{equation*} Now, applying Lemma <ref> we obtain \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{R_\tC}& \poloFantasmaCn{\rB}\qw& \qw\\ \pureghost{R_\tC}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \prepareC{\varphi_0}& \poloFantasmaCn{\rB\rC}\qw& \multigate{1}{\tU}& \poloFantasmaCn{\rA\rC}\qw& \measureD{e}\\ \multiprepareC{1}{\Psi^{(\rA)}}& \poloFantasmaCn{\rA}\qw& \ghost{\tU}& \poloFantasmaCn{\rB}\qw& \qw\\ \pureghost{\Psi^{(\rA)}}& \qw& \poloFantasmaCn{\tilde{\rA}}\qw& \qw&\qw\\ \end{aligned}\, . \end{equation*} Since $\Psi^{(\rA)}$ is dynamically faithful for system $\rA$, this implies \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA}\qw& \gate{\tC}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \prepareC{\varphi_0}& \poloFantasmaCn{\rB\rC}\qw& \multigate{1}{\tU}& \poloFantasmaCn{\rA\rC}\qw& \measureD{e}\\ \poloFantasmaCn{\rA}\qw& \ghost{\tU}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned}\, . \end{equation*} Therefore, $\tV \coloneqq \tU|\varphi_0)_{\rB\rC}$ is a reversible dilatation of $\tC$, with $\rE_0\coloneqq\rB\rC$ and $\rE\coloneqq\rA\rC$. Finally, the uniqueness clause in Lemma <ref> implies uniqueness of the dilatation. Moreover, two reversible dilatations of the same channel with different environments are related as follows. Let $\tV\in\transf{A}{BE}$ and $\tV^\prime\in\mathsf{Transf}(\rA\rightarrow\rB\rE^\prime)$ be two reversible dilatations of the same channel $\tC$, with generally different environment $\tE$ and $\rE^\prime$. Then there is a channel $\tL$ from $\rE$ to $\rE\rE^\prime$ such that \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { & & \amultigate{1}{\tV^\prime}& \poloFantasmaCn{\rE^\prime}\qw& \qw\\ \poloFantasmaCn{\rA}\qw& \ghost{\tV^\prime}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { & & & & \amultigate{1}{\tL}& \poloFantasmaCn{\rE}\qw& \measureD{e}\\ & & \amultigate{1}{\tV}& \poloFantasmaCn{\rE}\qw& \ghost{\tL}& \poloFantasmaCn{\rE^\prime}\qw& \qw\\ \poloFantasmaCn{\rA}\qw& \ghost{\tV}& \qw& \poloFantasmaCn{\rB}\qw& \qw&\qw\\ \end{aligned}\, . \end{equation*} The channel $\tL$ has the form \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { & & \amultigate{1}{\tL}& \poloFantasmaCn{\rE}\qw& \qw\\ \poloFantasmaCn{\rE}\qw& \ghost{\tL}& \poloFantasmaCn{\rE^\prime}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \prepareC{\eta_0}& \poloFantasmaCn{\rE^\prime}\qw& \multigate{1}{\tU}& \poloFantasmaCn{\rE}\qw& \qw\\ \poloFantasmaCn{\rE}\qw& \ghost{\tU}& \poloFantasmaCn{\rE^\prime}\qw& \qw\\ \end{aligned} \end{equation*} for some pure state $\eta_0$ and some reversible transformation $\tU\in\mathsf{Transf}(\rE\rE^\prime)$. Apply $\tV$ and $\tV^\prime$ to the faithful state $\Psi^{(\rA)}$ and then use the uniqueness of purification stated in Lemma <ref>. §.§ Casually ordered channels and channels with memory The last step before the proof of the theorem is the notion of casually ordered channels and channels with memory. A bipartite channel $\tC$ from $\rA_1\rA_2$ to $\rB_1\rB_2$ is casually ordered if there is a channel $\tD$ from $\rA_1$ to $\rB_1$ such that $(e|_{\rB_2}\tC=\tD\otimes(e|_{\rA_2}$. Diagrammatically, \begin{equation} \label{eq:casuallyorderedbipartitechannels} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \multigate{1}{\tC}& \poloFantasmaCn{\rB_1}\qw& \qw\\ \poloFantasmaCn{\rA_2}\qw& \ghost{\tC}& \poloFantasmaCn{\rB_2}\qw& \measureD{e} \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \gate{\tD}& \poloFantasmaCn{\rB_1}\qw& \qw\\ \qw& \poloFantasmaCn{\rA_2}\qw& \qw& \measureD{e} \end{aligned}\,. \end{equation} Eq. (<ref>) means that the channel $\tC$ does not allow for signaling from the input system $\rA_2$ to the output system $\rB_1$. In a relativistic context, this can be interpreted as $\rB_1$ being outside the casual future of $\rA_2$. A bipartite channel $\tC$ from $\rA_1\rA_2$ to $\rB_1\rB_2$ can be realized as a sequence of two channels with memory if there exist two system $\rE_1,\,\rE_2$, called memory systems, and two channels $\tC_1\in\mathsf{Transf}(\rA_1\rightarrow\rB_1\rE_1)$ and $\tC_2\in\mathsf{Transf}(\rA_2\rE_1\rightarrow\rB_2\rE_2)$ such that $\tC=(e|_{\rE_2}(\tC_2\otimes\tI_{\rB_1})(\tI_{\rA_2}\otimes\tC_1)$. Diagrammatically, \begin{equation} \label{eq:channelwithmemory} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \multigate{1}{\tC}& \poloFantasmaCn{\rB_1}\qw& \qw\\ \poloFantasmaCn{\rA_2}\qw& \ghost{\tC}& \poloFantasmaCn{\rB_2}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \multigate{1}{\tC_1}& \poloFantasmaCn{\rB_1}\qw& \qw & & \poloFantasmaCn{\rA_2}\qw& \multigate{1}{\tC_2}& \poloFantasmaCn{\rB_2}\qw& \qw\\ & & \aghost{\tC_1}& \qw& \poloFantasmaCn{\rE_1}\qw& \qw&\qw& \ghost{\tC_2}& \poloFantasmaCn{\rE_2}\qw& \measureD{e}\\ \end{aligned}\,. \end{equation} For causally ordered bipartite channels the dilatation theorem implies the following result: A bipartite channel $\tC$ from $\rA_1\rA_2$ to $\rB_1\rB_2$ is casually ordered if and only if it can be realized as a sequence of two channels with memory. Moreover, the channels $\tC_1,\,\tC_2$ in Eq. (<ref>) can be always chosen such that $\tC_2\tC_1$ is a reversible dilatation of $\tC$. If Eq. (<ref>) holds, the channel $\tC$ is clearly casually ordered, with the channel $\tD$ given by $\tD\coloneqq(e|_{\rE_1}\tC_1$. Conversely, suppose that $\tC$ is casually ordered. Take a reversible dilatation of $\tC$, say $\tV\in\mathsf{Transf}(\rA_1\rA_2\rightarrow\rB_1\rB_2\rE)$, and a reversible dilatation of $\tD$, say $\tV_1\in\mathsf{Transf}(\rA_1\rightarrow\rB_1\rE_1)$. Now by the definition of casually ordered channels (Eq. (<ref>)) we have \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \multigate{2}{\tV}& \poloFantasmaCn{\rB_1}\qw& \qw\\ \poloFantasmaCn{\rA_2}\qw& \ghost{\tV}& \poloFantasmaCn{\rB_2}\qw& \measureD{e}\\ & & \aghost{\tV}& \poloFantasmaCn{\rE}\qw& \measureD{e} \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \multigate{1}{\tV_1}& \poloFantasmaCn{\rB_1}\qw& \qw\\ & & \aghost{\tV_1}& \poloFantasmaCn{\rE_1}\qw& \measureD{e}\\ \qw& \poloFantasmaCn{\rA_2}\qw& \qw& \measureD{e} \end{aligned}\, . \end{equation*} This means that $\tV$ and $\tV_1\otimes\tI_{\rA_2}$ are two reversible dilatations of the same channel. By the uniqueness of the reversible dilatation expressed by Lemma <ref> we then obtain \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \multigate{2}{\tV}& \poloFantasmaCn{\rB_1}\qw& \qw\\ \poloFantasmaCn{\rA_2}\qw& \ghost{\tV}& \poloFantasmaCn{\rB_2}\qw& \qw\\ & & \aghost{\tV}& \poloFantasmaCn{\rE}\qw& \qw \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \multigate{1}{\tV_1}& \qw& \poloFantasmaCn{\rB_1}\qw& \qw& \qw\\ & & \aghost{\tV_1}& \poloFantasmaCn{\rE_1}\qw& \multigate{2}{\tL}& \poloFantasmaCn{\rB_2}\qw& \qw\\ \qw& \poloFantasmaCn{\rA_2}\qw& \qw& \ghost{\tL}& \poloFantasmaCn{\rE}\qw& \qw\\ & & & & \aghost{\tL}& \poloFantasmaCn{\rE_1\rA_2}\qw& \measureD{e} \end{aligned}\, . \end{equation*} Once we have defined $\rE_2\coloneqq\rE\rE_1\rA_2$ it only remains to observe that the above diagram is nothing but the thesis, with $\tC_1=\tV_1$ and $\tC_2=\tL$. By construction, $\tC_2\tC_1$ is a reversible dilatation of $\tC$. The definition of casually ordered bipartite channel is easily extended to the multipartite case. Here we will only report the definition and the two main theorems of the theory. For their demonstrations we remand to the original article [2] since they are still valid also from our three Axiom as starting point. An N-partite channel $\tC^{(N)}$ from $\rA_1\dots\rA_N$ to $\rB_1\dots\rB_N$ is causally ordered if for every $k\le N$ there is a channel $\tC^{(k)}$ from $\rA_1\dots\rA_k$ to $\rB_1\dots\rB_k$ such that \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \multigate{5}{\tC^{(N)}}& \poloFantasmaCn{\rB_1}\qw& \qw\\ & \vdots & \aghost{\tC^{(N)}}& \vdots& \\ \poloFantasmaCn{\rA_k}\qw& \ghost{\tC^{(N)}}& \poloFantasmaCn{\rB_k}\qw& \qw\\ \poloFantasmaCn{\rA_{k+1}}\qw& \ghost{\tC^{(N)}}& \poloFantasmaCn{\rB_{k+1}}\qw& \measureD{e}\\ & \vdots & \aghost{\tC^{(N)}}& \vdots& \\ \poloFantasmaCn{\rA_{N}}\qw& \ghost{\tC^{(N)}}& \poloFantasmaCn{\rB_{N}}\qw& \measureD{e}\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \multigate{2}{\tC^{(k)}}& \poloFantasmaCn{\rB_1}\qw& \qw\\ & \vdots & \aghost{\tC^{(k)}}& \vdots& \\ \poloFantasmaCn{\rA_1}\qw& \ghost{\tC^{(k)}}& \poloFantasmaCn{\rB_1}\qw& \qw\\ \qw& \poloFantasmaCn{\rA_{k+1}}\qw& \qw& \measureD{e}\\ &\vdots& & \vdots&\\ \qw& \poloFantasmaCn{\rA_{N}}\qw& \qw& \measureD{e}\\ \end{aligned}\, . \end{equation*} The definition means that the output systems $\rB_1\dots\rB_k$ are outside the casual future of any input system $\rA_l$ with $l>k$. Causally ordered channels can be characterized as follows. An N-partite channels $\tC^{(N)}$ from $\rA_1\dots\rA_N$ to $\rB_1\dots\rB_N$ is causally ordered if and only if there exist a sequence of memory systems $\{\rE_k\}_{k=0}^N$ with $\rE_0=\rI$ and a sequence of channels $\{\tV_k\}_{k=1}^N$, with $\tV_k\in\mathsf{Transf}(\rA_k\rE_{k-1}\rightarrow\rB_k\rE_k)$ such that \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \multigate{2}{\tC^{(N)}}& \poloFantasmaCn{\rB_1}\qw& \qw\\ & \vdots & \aghost{\tC^{(N)}}& \vdots& \\ \poloFantasmaCn{\rA_N}\qw& \ghost{\tC^{(N)}}& \poloFantasmaCn{\rB_N}\qw& \qw \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.1em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \multigate{1}{\tV_1}& \poloFantasmaCn{\tB_1}\qw& \qw& & \poloFantasmaCn{\rA_2}\qw& \multigate{1}{\tV_2}& \poloFantasmaCn{\rB_2}\qw& \qw& \cdots& & \poloFantasmaCn{\rA_N}\qw& \multigate{1}{\tV_N}& \poloFantasmaCn{\rB_N}\qw& \qw\\ & & \aghost{\tV_1}& \qw& \poloFantasmaCn{\rE_1}\qw& \qw& \qw& \ghost{\tV_2}& \poloFantasmaCn{\rE_2}\qw& \qw& \cdots& & \poloFantasmaCn{\rE_{N-1}}\qw& \ghost{\tV_N}& \poloFantasmaCn{\rE_N}\qw& \measureD{e}\\ \end{aligned}\, . \end{equation*} Moreover, $\tV_N\dots\tV_1$ is a reversible dilation of $\tC^{(N)}$. We have also a uniqueness result: Let $\{\tV_k\}_{k=1}^N$, $\tV_k\in\mathsf{Transf}(\rA_k\rE_{k-1}\rightarrow\rB_k\rE_k)$ be a reversible realization of the casually ordered channel $\tC^{(N)}$ as a sequence of channels with memory, as in Theorem <ref>. Suppose that $\{\tV_k^\prime\}_{k=1}^N$, $\tV_k^\prime\in\mathsf{Transf}(\rA_k\rE_{k-1}^\prime\rightarrow\rB_k\rE_kì^\prime)$ is another reversible realization of $\tC^{(N)}$ as a sequence of channels with memory. Then there exist a channel $\tR$ from $\rE_N$ to $\rE_N^\prime$ such that \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \multigate{1}{\tV_1^\prime}& \poloFantasmaCn{\tB_1}\qw& \qw& & \poloFantasmaCn{\rA_2}\qw& \multigate{1}{\tV_2^\prime}& \poloFantasmaCn{\rB_2}\qw& \qw& \cdots& & \poloFantasmaCn{\rA_N}\qw& \multigate{1}{\tV_N^\prime}& \poloFantasmaCn{\rB_N}\qw& \qw\\ & & \aghost{\tV_1^\prime}& \qw& \poloFantasmaCn{\rE_1^\prime}\qw& \qw& \qw& \ghost{\tV_2^\prime}& \poloFantasmaCn{\rE_2^\prime}\qw& \qw& \cdots& & \poloFantasmaCn{\rE_{N-1}^\prime}\qw& \ghost{\tV_N}& \poloFantasmaCn{\rE_N^\prime}\qw& \qw\\ \end{aligned}\, =\\ \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \poloFantasmaCn{\rA_1}\qw& \multigate{1}{\tV_1}& \poloFantasmaCn{\tB_1}\qw& \qw& & \poloFantasmaCn{\rA_2}\qw& \multigate{1}{\tV_2}& \poloFantasmaCn{\rB_2}\qw& \qw& \cdots& & \poloFantasmaCn{\rA_N}\qw& \multigate{1}{\tV_N}& \poloFantasmaCn{\rB_N}\qw& \qw&\qw&\qw\\ & & \aghost{\tV_1}& \qw& \poloFantasmaCn{\rE_1}\qw& \qw& \qw& \ghost{\tV_2}& \poloFantasmaCn{\rE_2}\qw& \qw& \cdots& & \poloFantasmaCn{\rE_{N-1}}\qw& \ghost{\tV_N}& \poloFantasmaCn{\rE_N}\qw& \gate{\tR}& \poloFantasmaCn{\rE_N^\prime}\qw& \qw\\ \end{aligned}\, . \end{aligned} \end{equation*} §.§ No bit commitment The bit commitment protocol defined in Section <ref> is generally implemented with sequences of channels with memory, that can be used to describe sequences of moves of Alice or Bob. In this scenario, the memory systems are the private systems available to a party, while the other input-output systems are the systems exchanged in the communication with the other party. We recall that for every theory that satisfy the starting hypothesis, this proof has an absolutely general validity: for every kind of input states and every possible strategy adopted, including both atomic and non-atomic transformations. If a N-round protocol is perfectly concealing, then there is a perfect cheating. We first prove the impossibility for protocols that do not involve the exchange of classical information. Let $\tA_0,\,\tA_1\in\mathsf{Transf}(\rA_1...\rA_N\rightarrow\rB_1...\rB_N\rF_N)$ two causally ordered N-partite channels (here the last output system of the causally ordered channels is the bipartite system $\rB_N,\,F_N$), representing Alice's move to encode the bit value $b=0,\,1$, respectively. The system $\rF_N$ is the system sent from Alice to Bob at the opening phase in order to unveil the value of the bit. If the protocol is perfectly concealing, then the reduced channels before the opening phase must be indistinguishable, namely $(e|_{\rF_N}\tA_0=(e|_{\rF_N}\tA_1\coloneqq\tC$. Now, due to Theorem <ref>, there exist two reversible dilatations $\tV_0\in\mathsf{Transf}(\rA_1...\rA_N\rightarrow\rB_1...\rB_N\rF_N\rG_0)$ and $\tV_1\in\mathsf{Transf}(\rA_1...\rA_N\rightarrow\rB_1...\rB_N\rF_N\rG_1)$ for $\tA_0$ and $\tA_1$, respectively. Since $\tV_0$ and $\tV_1$ are also two dilatations of the channel $\tC$, due to Lemma <ref> there is a channel $\tR$ from $\rF_N\rG_0$ to $\rF_N\rG_1$ such that $\tV_1=\tR\tV_0$. Applying this channel to her private systems, Alice can switch from $\tV_0$ to $\tV_1$ just before the opening. Discarding the auxiliary system $\rG_1$, this yields channel $\tA_1$. The cheating is perfect, since Alice can play the strategy $\tV_0$ until the end of the commitment and decide the bit value before the opening without being detected by Bob. The above reasoning can be extended to N-round protocols involving the exchange of classical information. Indeed classical messages can be modeled by measure-and-prepare channels where the observation states are perfectly distinguishable. The fact that some systems can only be prepared in perfectly distinguishable states will be referred as to "communication interface" of the protocol. In this case, to construct Alice's cheating strategy we can first take the reversible dilatations $\tV_0$, $\tV_1$ and the channel $\tR$ such that $\tV_1=\tR\tV_0$. In order to comply with the communication interface protocol, one can compose $\tV_0$ and $\tV_1$ with classical channels on all system that must be "classical" before the opening, thus obtaining two channels $\tD_0$ and $\tD_1$ that are no longer reversible but still satisfy $\tD_1=\tR\tD_0$. Discarding the auxiliary system $\rG_1$ and, of required by the communication interface, applying a classical channels on $\rF_N$, Alice then obtains channel $\tA_1$. Again, this strategy allows Alice to decide the value of the bit just before the opening without being detected. CHAPTER: PR-BOXES In this Chapter we will analyse the probabilistic theory [21, 22, 23] corresponding to the popular PR-boxes model introduced in Ref. [7]. In particular, in the first Section we will retrace the crucial steps and the underlying reasons that led to the development of the PR-box model. Then, in the second Section, after the formalization of the PR-boxes model in the language of operational probabilistic theories, we will report our results: the general POVM that grants perfect discriminability between any two pure states in the bi-partite case, the existence and uniqueness of the purification exclusively for the maximally mixed state (again in the bi-partite scenario) and finally some consideration on the general case of N-partite boxes. § WHY PR-BOXES? One of the most striking feature of quantum theory is certentantly non-locality. In fact, since the very beginning of the theory the incompleteness of the Copenhagen interpretation of quantum mechanics in relation to the violation of local causality was one of the main discussed aspect. In 1935 Einstein Podolsky and Rosen published the famous article of the EPR paradox [24]. The thought experiment generated a great deal of interest in the following years. Their notion of a "complete description" was later formalized by the suggestion of hidden variables that determine the statistics of measurement results, but to which an observer does not have access. In 1964 John Bell proved that some predictions of quantum mechanics cannot be reproduced by any theory of local physical variables [3]. Although Bell worked within non-relativistic quantum theory, the definition of local variable is relativistic: a local variable can be influenced only by events in its backward light cone, not by events outside, and can influence events in its forward light cone only. Quantum mechanics, which does not allow us to transmit signals faster than light (super-luminal signalling), preserves relativistic causality. But quantum mechanics does not always allow us to consider distant systems as separate, as Einstein assumed. Now quantum non-locality has been experimentally verified under different physical assumptions. Any physical theory that aims at superseding or replacing quantum theory should account for such experiments and therefore must also be non-local in this sense; quantum non-locality is a property of the universe that is independent of our description of nature. So quantum non-locality is an essential feature of quantum theory but it often appear in a negative light. In 1994 Popescu and Rohrlich published a work [7] where they proposed to show quantum non-locality in a more positive light. They investigated the inversion of the logical approach to quantum mechanics, considering quantum non-locality as an axiom instead of as a theorem and wondering what non-locality together with relativistic causality would imply. They found that quantum theory is only one of a class of non-local theories consistent with causality, and not even the most non-local. In fact, in a certain sense, non-locality can be quantified. In 1969, John Clauser, Michael Horne, Abner Shimony, and Richard Holt reformulated Bell's inequality in a manner that best suites experimental testing, the homonymous CHSH inequality [5]. CHSH inequality, restricted to any classical theory, states that a particular algebraic combination of correlations lies between -2 and 2. This bound is obviously violated in quantum mechanics, where in fact the CHSH inequality allows a maximum value given by Cirel'son's theorem as $2\sqrt{2}$ [4]. However, Popescu and Rohrlich wrote down a set of correlations that return a value of 4 for the CHSH expression, the maximum value algebraically possible, and that yet are non(super-luminar)-signalling. A question that now rises spontaneously is why does quantum theory not allow these strongly non-local correlations. In the hope of making further progress with this question these correlations have been investigated in the context of a theory with well-defined dynamics. Abstractly this scenario may be described by introducing two observers that have access to a black box. Each observer selects an input from a range of possibilities and obtains an output. The box determines a joint probability for each output pair given each input pair. It is clear that a quantum state provides a particular example of such a box, with input corresponding to measurement choice and output to measurement outcome. More generally boxes can be divided into different types. Some will allow the observers to signal to one another via their choice of input, and correspond to two-way classical channels, as introduced by Shannon. Others will not allow signalling - it is well known, for example, that any box corresponding to an entangled quantum state will not. This is necessary for compatibility between quantum mechanics and special relativity. Among the non-signalling boxes, some will violate a Bell-type inequality, and we refer to any such a box as non-local. As we have described above, in terms of our boxes, there are some boxes that are non-signalling but are more non-local than any box allowed by quantum theory. § PR-BOXES AS OPT The model describes $N$ correlated boxes (in the original paper [7] it was $N=2$) in a casual context. Each box is represented by the same elementary system $\rA$, with the $N$ correlated boxes represented by the composite system $\rA^{\otimes N}$. We will consider the simplest situation where each box has both input and output as binary variables. On each elementary system $\rA_i$, $i=1,\ldots,N$ only two atomic binary observation tests are allowed, say $B^{(0)}\coloneqq\{b^{(0)},e_\rA-b^{(0)}\}$ or $B^{(1)}\coloneqq\{b^{(1)},e_\rA-b^{(1)}\}$, with $b^{(0)},b^{(1)}\in\eff{A}$ atomic effects, and $e_\rA$ the deterministic effect of system $\rA$. Notice that the since the model is causal, the deterministic effect $e_\rA$ is unique and it is $e_{\rA^{\otimes N}}=\otimes^N e_\rA$. The probabilistic model is typically presented in terms of the probability function $P:(b_1,b_2,\ldots, b_N|B_1,B_2,\ldots B_N)\mapsto [0,1]$, with $B_i\in\{B^{(0)},B^{(1)}\}$ and $b_i\in\{b^{(0)},b^{(1)}\}$, for every $i=1,\ldots,N$, which returns the probability of the outcomes $b_1,b_2,\ldots,b_N$ given the observation tests $B_1,B_2,\ldots B_N$. The constraint imposed on the function $P$ is no-signalling, i.e. \begin{equation} \label{eq:no-signalling} \begin{aligned} \sum_{b_k=b^{(0)},e-b^{(0)}} &P(b_1,\ldots, b_N|B_1,\ldots B_N)=\\ =&P(b_1,\ldots, b_{k-1},b_{k+1},\ldots,b_N|B_1,\ldots B_{k-1},B_{k+1},\ldots, B_N). \end{aligned} \end{equation} §.§ States, effects and transformations We will start our analysis with the boxes of the elementary system $\rA$ that are necessary local. They are described in terms of the following probability distribution \begin{equation} \label{eq:prbox1} \begin{cases} 1& a=\alpha x\oplus \beta\\ 0 & \text{otherwise} \end{cases}, \end{equation} with $\alpha,\,\beta=0,1$. This four probability distributions correspond to the four pure states of system $\rA$. In fact the elementary system $\rA$ has dimension $\dim(\rA)=3$, namely its states are described by vectors in $\mathbb{R}^3$ ($\mathsf{St}_\mathbb{R}(\rA)=\mathsf{Eff}_\mathbb{R}(\rA)=\mathbb{R}^3$). The four pure normalized states (<ref>) can be represented by the following vectors \begin{equation} \label{eq:local-states} \begin{aligned} \omega_0 = \begin{pmatrix} \end{pmatrix},\: \omega_1 = \begin{pmatrix} \end{pmatrix},\: \omega_2 = \begin{pmatrix} \end{pmatrix},\: \omega_3 = \begin{pmatrix} \end{pmatrix}, \end{aligned} \end{equation} where the correspondence between $\omega_n$ and the probability rule $p_{\alpha\beta}$ is given by $\alpha\beta=$(binary form of $n$). The convex set of states normalized is then represented by a square (see the square in the plane $z=1$ in Fig. <ref>). (15, 44) $\omega_1$ (67, 54) $\omega_3$ (29, 54) $\omega_0$ (74, 44) $\omega_2$ (48, 27) $b_1$ (73, 33) $b_2$ (22, 33) $b_0$ (50, 38) $b_3$ (45, 53) $\bar e$ (72.5, 60) $0$ (52.5, 64) $1$ (93, 56) $-1$ (-2.5, 52.5) $1$ (-2, 36.5) $0$ (-4.5, 21) $-1$ (0, 56) $-1$ (29, 61) $0$ Elementary system of the PR-boxes theory. This picture depicts the “squit” elementary system often considered in generalized probabilistic theories (in analogy to the “bit” and the “qubit” which are elementary systems of classical and quantum theory respectively). The system is specified by its sets of states and effects here represented as vectors in $\mathbb{R}^3$. The convex set of normalized states is represented by the square at the top, while the convex set of effects corresponds to the truncated cone. The set of effects of system $\rA$ is defined as the set of vectors $b$ such that $0 \le \Tr[b^T \omega] \le 1$ for every state $\omega\in\st{A}$, with $p_{y|x} = \Tr[{b^{(y)}}^T \omega_x]$ the rule providing the probability associated to an effect $b^{(y)}$ on a state $\omega_x$. This leads to the truncated cone of effects in Fig. <ref>, with extremal points given by \begin{align*} b^{(0)} = \begin{pmatrix} \frac{1}{2}\\[2pt]\frac{1}{2}\\[2pt]\frac{1}{2} \end{pmatrix},\; b^{(1)} = \begin{pmatrix} \end{pmatrix},\; b^{(2)} = \begin{pmatrix} \end{pmatrix},\; b^{(3)} = \begin{pmatrix} \tfrac{1}{2}\\[2pt]-\tfrac{1}{2}\\[2pt]\tfrac{1}{2} \end{pmatrix}. \end{align*} The deterministic effect (the effect $e_\rA$ such that $\Tr[e_\rA^T \omega]=1$ for every state $\omega\in\st{A}$) is the vector $e_\rA=(0,0,1)^T$ (when no ambiguity arises we will simply denote the deterministic effect with $e$). Our analysis now proceeds towards the composite system $\rA\otimes \rA$. In this case the probability function $p(a,b|x,y)$ form a table with $2^4$ entries, although these are not all independent due to the constraints of Eq. (<ref>). The dimension of the set of boxes is found by subtracting the number of independent constraints from $2^4$, and turns out to be 8 (we notice that $8=\dim\mathsf{St}_1(\rA\otimes\rA)$, while $\dim\mathsf{St}_\mathbb{R}(\rA\otimes \rA)=\dim\mathsf{St}_\mathbb{R}(\rA)\dim\mathsf{St}_\mathbb{R}(\rA)=9$, so due to Theorem <ref> it satisfies local discriminability). In this case we will have no more a square like in Fig. <ref> but a polytope with 24 vertices. The vertices will be boxes that satisfy all of the constraints and saturates a sufficient number of the positivity constraints to be uniquely determined. These 24 bilocal pure states may be divided into two classes. The local boxes, given by the following 16 probability distributions \begin{equation} \label{eq:14} \begin{cases} 1& a=\alpha x\oplus\beta\\ 1& b=\gamma y\oplus\delta\\ 0& \text{otherwise} \end{cases}, \end{equation} with $\alpha,\beta,\gamma,\delta=0,1$, and the non-local boxes, given by the 8 probability distributions \begin{equation} \label{eq:15} \begin{cases} 1/2& a\oplus b=xy\oplus \alpha x\oplus \beta y\oplus\gamma\\ 0 & \text{otherwise} \end{cases} \end{equation} with $\alpha,\beta,\gamma=0,1$. For convenience we can represent states and effects as $3\times 3$ real matrices rather than as vectors in $\mathbb{R}^9$. The 16 local state of the bipartite system are nothing else that the factorized pure states \begin{align} \label{eq:local-bipartite-states} \Omega_{4i+j} \coloneqq \omega_i \otimes \omega_j^T, \end{align} where $i, j \in \{ 0, 1, 2, 3 \}$, and the 8 non-local states (playing the role of entangled states) are represented by the following matrices \begin{equation} \label{eq:non-local-bipartite-states} \begin{aligned} \Omega_{16} & := \frac12 \begin{pmatrix} -1 & 1 & 0 \\ 1 & 1 & 0 \\ 0 & 0 & 2 \end{pmatrix}, & \Omega_{17} & := \frac12 \begin{pmatrix} -1 & -1 & 0 \\ -1 & 1 & 0 \\ 0 & 0 & 2 \end{pmatrix}, & \Omega_{18} := \frac12 \begin{pmatrix} 1 & -1 & 0 \\ -1 & -1 & 0 \\ 0 & 0 & 2 \end{pmatrix},\\ \Omega_{19} &:= \frac12 \begin{pmatrix} -1 & 1 & 0 \\ -1 & -1 & 0 \\ 0 & 0 & 2 \end{pmatrix}, & \Omega_{20} & := \frac12 \begin{pmatrix} -1 & -1 & 0 \\ 1 & -1 & 0 \\ 0 & 0 & 2 \end{pmatrix}, & \Omega_{21} := \frac12 \begin{pmatrix} 1 & -1 & 0 \\ 1 & 1 & 0 \\ 0 & 0 & 2 \end{pmatrix}, \\ \Omega_{22} & := \frac12 \begin{pmatrix} 1 & 1 & 0 \\ -1 & 1 & 0 \\ 0 & 0 & 2 \end{pmatrix}, & \Omega_{23} & := \frac12 \begin{pmatrix} 1 & 1 & 0 \\ 1 & -1 & 0 \\ 0 & 0 & 2 \end{pmatrix}.& \end{aligned} \end{equation} From a generic probability rule $p_{\alpha\beta\gamma}$ we can identify one of the above matrix $\Omega_n$ by the equation: $n=15+\left[(3+3\alpha+4\beta+6\gamma)\mod8\right]$. Again, the probability associated to a bipartite effect $B_y$ on the state $\Omega_x$ is given by $p_{y|x} = \Tr[B_y^T \Omega_x]$. Accordingly, the set of bipartite effects is easily derived via the consistency condition $\Tr[B_j^T \Omega_i] \ge 0$ for every $j \in [0, 23]$. It follows that the only admissible extremal effects are the 16 factorized matrices \begin{align} B_{4i+j} := b^{(i)} \otimes {b^{(j)}}^T. \end{align} This is a relevant feature of PR-boxes model, whose strong correlation incapsulated in the eight non-local states $\Omega_x$, $x\in[16,23]$ are incompatible with any in principle admissible non-factorized measurements. This feature has been firstly noticed in Ref. [25] and later in Ref. [26] where all possible theories compatible with the squit local system have been classified (among these theories is the dual version of PR-boxes that only have factorized states but eight non-local effects). Finally, the deterministic effect for the bipartite system is $e\otimes e^T$. Analogously one defines the convex set of states and the convex set of effects for the arbitrary $N$-partite system $\rA^{\otimes N}$. However, while for effect is nothing but a trivial generalization, for the states the discussion is not so straight and some interesting features arise. We will dedicate a following Section to discuss some of the most immediate aspect about the $N$-partite system, we will start our discussion from the tripartite boxes. We can now turn our attention to the transformations of the theory. We focus here on the reversible transformations which are of interest for the present paper results. The set $\mathcal{U}(\rA)$ of reversible transformations of the system $\rA$ coincides with the finite group of symmetries of the square (the dihedral group of order eight $D_8$ containing four rotations and four reflections). In the chosen representation we have \begin{equation} \label{eq:single-system-unitaries} \begin{aligned} \mathcal{U}(\rA)=\{U_k^s: k=0,\ldots,3, s=\pm\}\\ U_k^s = \begin{pmatrix} \cos \frac{\pi k}2 & -s \sin \frac{\pi k}2 & 0 \\ \sin \frac{\pi k}2 & s \cos \frac{\pi k}2 & 0 \\ 0 & 0 & 1 \end{pmatrix}. \end{aligned} \end{equation} The matrices $U_k^{+}$ and $U_k^{-}$ representing the four rotations and the four reflections of the square respectively. When we apply them to the four pure states of Eq. (<ref>) we have \begin{align} \label{eq:reversible-mapping-on-local-states} \omega_{j+k} & = U_k^+\, \omega_{j}, & \omega_{k} & = U_k^-\, \omega_{j+k}\,, \end{align} for $j\in\left[0,3\right]$ and where the sum is$\mod 3$, i.e. if $j=2$ and $k=2$ then $j+k=0$. We notice that these transformations are atomic. In Ref. [23] all atomic transformations of the squit system have been classified. The set of reversible transformations $\mathcal{U}(A\otimes A)$ of the composite system $\rA\otimes\rA$ (which has been derived in Ref. [27]) is \begin{equation} \label{eq:reversible} \mathcal{U}(\rA\otimes \rA)=\{ W^i(U_j^{s_1} \otimes U_k^{s_2})\}\quad i=0,1,\; 0\leq j,k\leq 3,\; s_1,s_2=\pm, \end{equation} with $W$ the swap map, namely the map that exchanges the two subsystems. This means that any reversible map corresponds to the tensor product of single system reversible transformations with, possibly, the application of the swap. As noticed in Ref. [27] reversible transformations cannot create entanglement. This result has been extended in Ref. [28] to the composition of an arbitrary number $N$ of systems $\rA^{\otimes N}$, showing that also in that case the set of reversible multipartite transformations is generated by local reversible operations and permutations of systems. We notice that in the PR-boxes theory any non-local bipartite pure state can be reversibly mapped to any other non-local bipartite pure state. Moreover, this mapping can be done via the local application of a single system reversible map. For example, starting from the state $\Omega_{16}$, one has \begin{align} \label{eq:reversible-mapping} \Omega_{16+k} & = (U_k^+\otimes I)\Omega_{16}, & \Omega_{23-k} & = (U_k^-\otimes I) \Omega_{16}. \end{align} We finally remark that the set of reversible transformations of system $\rA$, $\tU(\rA)$, in terms of description by the probability distributions in Eqs. (<ref>), (<ref>), (<ref>) correspond to a local relabelling defined by the operations \begin{equation} \label{eq:local-relabelling} \begin{aligned} x&\rightarrow x\oplus 1,\qquad a\rightarrow a\oplus\alpha x\oplus \gamma,\\ y&\rightarrow y\oplus 1,\qquad b\rightarrow b\oplus\beta y\oplus \gamma. \end{aligned} \end{equation} §.§ Discriminability between PR-boxes An important question to address when dealing with the PR-boxes theory is the following: given two deterministic state of the theory is it possible to discriminate between them? We will address to this question only for the extremal points of the polytope, i.e. the pure states, in order to analyse a way to perfect discriminate between them. For local boxes it is easy to show that there exist POVMs that are able to perfect discriminate between every two of the four possible state in Eq. (<ref>) [23]. For what concern bipartite boxes, while it is trivial to find perfectly discriminable POVMs for each pair of the 16 local boxes (since they are simply the tensor product of the 4 local boxes, so also the discriminable POVMs are just the tensor product of the perfectly discriminable POVMs of the local boxes) it is a bit more elaborate to explore the discriminability between non-local boxes. To help in our analysis we introduce the following table regarding non-local boxes labelled by $(\alpha,\beta,\gamma)$ as described in Eq. (<ref>): $x$ $y$ $a\oplus b$ 0 0 $\gamma$ 0 1 $\beta\oplus\gamma$ 1 0 $\alpha\oplus\gamma$ 1 1 $1\oplus\alpha\oplus\beta\oplus\gamma$ Correlations for a generic $(\alpha,\beta,\gamma)$ non-local box for all the possible given input combinations of $x$ and $y$. It is now easy to verify that for every two different non-local boxes, i.e. for every choice of two different combinations of $(\alpha,\beta,\gamma)$: $c_1=(\alpha_1,\beta_1,\gamma_1)$ and $c_2=(\alpha_2,\beta_2,\gamma_2)$, there is always at least one input combination $(x,y)$, whose output relation is equal to 0 for $c_1$ and to 1 for $c_2$. If we denote with $e$ the deterministic effect of the bipartite system $\rA\otimes\rA$ and $a\coloneqq b^{(3(1-x))} \otimes b^{(3(1-y))}+b^{(1+x)} \otimes b^{(1+y)}$, the POVM $\{a,e-a\}$ is able to perfectly discriminate between the two chosen non-local boxes. Finally, the last remark regards the discrimination between one local and one non-local bipartite box. We can help ourselves with the two TABLEs of Appendix <ref> that follows the notation of Eqs. (<ref>),(<ref>). For every pair of bipartite boxes (one local and one non-local) there is always a combination of $(x,y)$ such that in one case the outcome relation $a\oplus b$ is equal to 0 for one box and 1 for the other. So we can construct a perfectly discriminating POVM following the same strategy of above. §.§ Purification in the PR-box (bipartite restriction) model Until now nothing has been said about purification in the PR-box model. The following discussion represent an important result of the theory, valid in the general context of $N$-partite boxes (with $N$ finite), but we will operate in the significant restriction of admitting no more than bipartite correlated boxes. For the rest of the subsection when we will refer to PR-box model/theory we will mean the model under this limitation. Let begin with some general definitions and preliminary considerations. The ability to transform any pure state into any other by means of reversible transformations will be called transitivity, meaning that the action of the set of reversible transformations is transitive on the set of pure states. Based on this definition, PR-box model clearly enjoys this property. In fact the transformations of Eq. (<ref>) is transitive on the set of pure states of Eq. (<ref>). Among the numerous consequences that transitivity implies one will be of our interest, the uniqueness of the maximally mixed state. So it is in order to properly define what a maximally mixed state is. If a state is invariant under the action of every reversible transformation, then it is a maximally mixed state. We will not report here the demonstration of the uniqueness of the maximally mixed state from transitivity, for it we refer to Ref. [11]. From the transformations of Eq. (<ref>) and the vector representation of the pure state in Eq. (<ref>) it is not difficult to write down the maximally mixed state: \begin{equation} \label{eq:maximally-mixed-state} \mu = \begin{pmatrix} \end{pmatrix}\,. \end{equation} We are now in position to state the main result of this Section. Given a system $\otimes^N\rA$, the maximally mixed state $\mu^{\otimes N}\in\st{\otimes^NA}$ of Eq. (<ref>) is the unique internal state that is purificable and its purification is unique up to a reversible transformation on the purifying systems. The proof is divided in two part. In the first one we will prove the thesis for $N=1$, then in the second part it is extended to an arbitrary number of systems. Given the system $\rA_1$ we will consider a system $\rA_2$ as the purifying system. The pure states of $\rA_1\otimes\rA_2$ are the 24 pure states $\Omega_{i}$, for $i=0,\ldots,23$ of Eqs. (<ref>), (<ref>). We know that 16 of them (namely the local ones, that are expressed in Eq. (<ref>)) are separable states. So if we consider the marginal state obtained by applying the deterministic effect $e_{\rA_2}$ on the purifying system $\rA_2$ on these separable states we get one of the local 4 state represented in Eq. (<ref>), that are pure. However, if we repeat the same procedure on the 8 non-local pure states (represented in Eq. (<ref>)) we get the same state for all of them: the maximally mixed state $\mu\in\mathsf{St}(\rA_1)$. So the unique internal state that can be purificated is the maximally mixed state and since the 8 non-local bipartite pure states are all mapped to any other non-local bipartite pure state by the application of a single system reversible map, as shown in Eq. (<ref>), the purification is unique up to a reversible transformation on the purifying system. We now deal with the $N$-partite scenario. The more general state $\Psi\in\mathsf{St}(\otimes^N\rA)$ has the form: \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.8em @! R { \multiprepareC{2}{\Psi}& \poloFantasmaCn{\rA_1}\qw& \qw\\ \pureghost{\Psi}& \vdots&\\ \pureghost{\Psi_i}& \poloFantasmaCn{\rA_N}\qw& \qw\\ \end{aligned} \: = \: \begin{aligned} \Qcircuit @C=1.2em @R=.6em @! R { \multiprepareC{1}{\omega_1}& \poloFantasmaCn{\rA_1}\qw& \qw\\ \pureghost{\omega_1}& \poloFantasmaCn{\rA_2}\qw& \qw\\ \vdots&&\\ \multiprepareC{1}{\omega_{M/2}}& \poloFantasmaCn{\rA_{M-1}}\qw& \qw\\ \pureghost{\omega_{M/2}}& \poloFantasmaCn{\rA_M}\qw& \qw\\ \prepareC{\omega_{M/2+1}}& \poloFantasmaCn{\rA_{M+1}}\qw& \qw\\ \vdots&&\\ \prepareC{\omega_{N-M/2}}& \poloFantasmaCn{\rA_{N}}\qw& \qw\\ \end{aligned}\, . \end{equation*} In order to be purificable, every monopartite state has to be pure or purificable and every bipartite state to be pure (since we are admitting no more than bipartite correlations, a bipartite state can be purificable only if it is pure). But if one of the states $\omega_j$ for $j=1,\ldots,N-M/2$ is pure then $\Psi$ can not be internal. Since the only purificable internal state is $\mu$, we have that the only purificable internal $N$-partite state is $\mu\otimes\ldots\otimes\mu$ (thanks to local discriminability the parallel composition of internal states is still an internal state). Furthermore, for what we have seen before, the purification of $\mu\otimes\ldots\otimes\mu$ is unique up to a reversible transformation. This transformation is nothing else that the parallel composition of the maps $(U_k\otimes\tI)$ where $U_k$ are the maps of Eq. (<ref>). Remark: in the previous proof we noticed that the maximally mixed state of a system $\otimes^N\rA$ is not the unique mixed state that is purificable. In fact, the more general state $\Phi\in\mathsf{St}(\otimes^N\rA)$ that can be purificated is of the form: $\Phi=\phi_1\otimes\ldots\otimes\phi_N$ where \begin{equation*} \phi_i= \begin{cases} &\omega_{j}\quad\text{for }j\in\left[0,3\right]\\ &\Omega_{k}\quad\text{for }j\in\left[16,23\right] \end{cases}\quad\text{for }i=1,\ldots,N\,, \end{equation*} and its purification is still unique up to reversible transformations on the purifying systems. §.§ N-partite PR-boxes In the literature of PR-box theory a thorough and systematic study on $k$-partite correlated boxes, with $k\ge3$, has never been made. This represents an important absence within the model since it prevents the theory to be complete. In this Section we show some important consequences that emerge by just considering tripartite correlated boxes with some speculations about the complete $N$-partite model. In our analysis of PR-box model integrated with tripartite boxes we make use of the classification that has been made in Ref. [29]. In the article of Pironio et al., the no-signaling polytope is found to have 53856 extremal points, belonging to 46 inequivalent classes. The term inequivalent means that there not exist reversible local transformations that allow to move from a representative of one class to one of another class, while, inside the same class, all the extremal points are connected by local relabelling, see Eqs. (<ref>) (since we are dealing with three parties boxes, also permutation of the parties is a local relabelling, i.e. $x\rightarrow y$, $y\rightarrow z$, and $z\rightarrow x$ and so on for every possible permutation). Firstly, it is no more granted that the maximally mixed state is the unique internal state purificable in the theory. In fact it could happen that between the 53856 pure tripartite states, there will be one whose marginal state is an internal state different from the maximally mixed one. This leads to think that increasing the number of correlated systems that we are considering the number of internal states that are purificable will also increase. Even if there are not academic works in this matter, it is a very likely and reasonable possibility and we address to future studies to investigate in this direction. Secondly, even if the only internal purificable state would still be the maximally mixed one, the purification is no more unique. To see this it suffices to consider two states $\Psi_1,\Psi_2\in\mathsf{St}(\rA^{\otimes3})$ that have the following form: \begin{equation*} \Psi_1 \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.8em @! R { \multiprepareC{1}{\Omega}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Omega}& \poloFantasmaCn{\rA}\qw& \qw\\ \prepareC{\omega}& \poloFantasmaCn{\rA}\qw& \qw\\ \end{aligned} \quad \text{and}\quad \Psi_2 \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.8em @! R { \multiprepareC{2}{\Phi^{(44)}}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Phi^{(44)}}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Phi^{(44)}}& \poloFantasmaCn{\rA}\qw& \qw\\ \end{aligned}\, , \end{equation*} where $\omega\in\st{A}$ and $\Omega\in\mathsf{St}(\rA^{\otimes2})$ are two pure states and $\Phi^{(44)}$ is a pure state, representative of the $44^{th}$ class described in Ref. [29]. If $\Omega$ is one of the non-local bipartite states then they have the same marginal, namely \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.8em @! R { \multiprepareC{1}{\Omega}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Omega}& \poloFantasmaCn{\rA}\qw& \multimeasureD{1}{e}\\ \prepareC{\omega}& \poloFantasmaCn{\rA}\qw& \ghost{e}\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.8em @! R { \prepareC{\mu}& \poloFantasmaCn{\rA}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.8em @! R { \multiprepareC{2}{\Phi^{(44)}}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Phi^{(44)}}& \poloFantasmaCn{\rA}\qw& \multimeasureD{1}{e}\\ \pureghost{\Phi^{(44)}}& \poloFantasmaCn{\rA}\qw& \ghost{e}\\ \end{aligned}\, , \end{equation*} where $\mu$ is the maximally mixed state of Eq. (<ref>) and the second equality derives straightforward once the probability rule of the $44^{th}$ class is written explicitly, as we will see in Eq. (<ref>). So we found that $\Psi_1$ and $\Psi_2$ are two purification of the same state but since every reversible transformation $U\in\tU(\rA^{\otimes2})$ is the composition of local reversible maps, "correlation" can not be created and so \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.8em @! R { \multiprepareC{2}{\Psi_1}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Psi_1}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Psi_1}& \poloFantasmaCn{\rA}\qw& \qw\\ \end{aligned} \, \ne \, \begin{aligned} \Qcircuit @C=1.2em @R=.8em @! R { \multiprepareC{2}{\Psi_2}& \qw& \poloFantasmaCn{\rA}\qw& \qw& \qw\\ \pureghost{\Psi_2}& \poloFantasmaCn{\rA}\qw& \multigate{1}{U}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Psi_2}& \poloFantasmaCn{\rA}\qw& \ghost{U}& \poloFantasmaCn{\rA}\qw& \qw\\ \end{aligned}\quad\forall\,U\in\tU(\rA^{\otimes2})\,. \end{equation*} Furthermore, restricting our attention to the pure bipartite states, we have noticed that when we pick up two of these states that have the same marginal, than there is always a local reversible map from one to the other and vice-versa. This is the case for the 8 non-local bipartite boxes, that are all connected by local transformations in $\tU(\rA)$ of Eq. (<ref>). This mechanism will turn out to be exactly the one responsible for the impossibility of perfectly secure bit commitment, as we will see in detail in the next Chapter. Since in the tripartite scenario not all the tripartite non-local boxes are connected by local transformation (we remind that local relabelling is not enough to change from a class to another), it is reasonable to think that perfect bit commitment will be possible, or at least a completely new way of cheating has to be thought. With this purpose we will propose a scheme of bit commitment as the conclusive Section of the next Chapter. CHAPTER: NO BIT COMMITMENT IN PR-BOXES In the past years numerous protocols have been proposed to realize bit commitment using PR-boxes. However, as outlined by A.J. Short, N. Gisin, and S. Popescu in Ref. [9], “it is surprising that the possibility that non-local correlations which are stronger than those in quantum mechanics could be used for bit commitment, because it is the very existence of non-local correlations which in quantum mechanics prevents bit commitment”. In that article they particularly referred to the protocol prosed by S. Wolf and J. Wullschleger [8] and showed that it was erroneous by argument of causality. After that also Buhrman et al. [10] proposed a bit commitment protocol in PR-box theory that was claimed to be unconditionally secure and where the counter-proof of Short, Gisin and Popescu did not work anymore. From a OPT point of view, when dealing with PR-boxes, some issues arise since they still not have a complete and closed theory. In fact, $k$-partite boxes with $k\ge3$ have been studied only roughly and a coherent and comprehensive theory has not been proposed. As we have seen in Section <ref>, simplistic generalizations are not adequate since admitting more than bipartite correlations alters significantly the theory. Nevertheless, in the literature not only PR-box model is generally considered admitting no more than bipartite states, but also local transformations (that are admissible in the theory) are ignored. In this final Chapter we propose a proof of impossibility of perfectly secure bit commitment in PR-boxes (even if under two important limitations: pure input states and bipartite boxes, the proof includes almost all the protocols proposed in literature that make use of PR-boxes). Furthermore we will explicitly describe a cheating protocol, contextualized in OPTs, that confute both the scheme proposed by Wolf and Wullschleger and the one by Buhrman et al.. We will show that just admitting local atomic reversible transformations the protocols proposed in literature can be cheated. Our proof joins the work published by Barnum, Dahlsten, Leifer, and Toner [30] where, in the framework of probabilistic theories, they prove that in all theories that are locally non-classical but do not have entanglement, there exists a bit commitment protocol that is exponentially secure in the number of systems used. If the protocol of Buhrman et al. would have been turned out to be correct then it would have represented the first example of an unconditionally secure bit commitment protocol valid in a theory with entanglement. However the question if a theory with entanglement admits perfectly secure bit commitment is still open. Finally we will sketch at the end of the Chapter a bit commitment scheme that make use of tripartite non-local boxes that is not more cheatable by local transformations and it could satisfy perfectly secure bit commitment. But advancements in the theory are necessary in order to give a definitive answer. § NO-PERFECTLY SECURE BIT COMMITMENT In this Section we provide an explicit proof of the impossibility of perfectly secure bit commitment in PR-box theory. However, two important limitation will be adopted. Even if we will consider arbitrary $N$-partite systems (with $N$ finite), we will not admit more than bipartite correlations (it will be taken for granted for the rest of the Section). Furthermore $\Psi_0$ and $\Psi_1$ will always be selected between the pure states. This is due to the fact that discrimination has never been studied in PR-box theory and the only results we rely on are those of Section <ref> that refers to pure states. If it would turn out that the strategy exposed in Section <ref> is the unique one to grant perfect discriminability between two arbitrary bipartite states then the protocol could be easily extended also to $\Psi_0,\,\Psi_1$ as arbitrary mixed states. Actually we will prove our theorem in two different way. In the first proof, we will state the property of perfect bit commitment and we will show that inconsistencies arise. In the second, that we will call alternative proof, we will show that, with some shrewdness, the proof of Section <ref> can be used also in this context. §.§ First Proof In this Section we make use of the definition of the protocol given in Section <ref> and nomenclature of states and transformations given in Section <ref>. Before the main theorem a preliminary lemma is in order. If a bit commitment protocol is correct with probability one then the two input states $\Psi_0,\,\Psi_1\in\mathsf{St}(\rA^{\otimes N}\rB^{\otimes N})$ shared by Alice ($\rA^{\otimes N}$) and Bob ($\rB^{\otimes N}$) are of the form \begin{equation} \label{eq:input-states} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{4}{\Psi_i}& \poloFantasmaCn{\rA_1}\qw& \qw\\ \pureghost{\Psi_i}& \poloFantasmaCn{\rB_1}\qw& \qw\\ \pureghost{\Psi_i}& \vdots&\\ \pureghost{\Psi_i}& \poloFantasmaCn{\rA_N}\qw& \qw\\ \pureghost{\Psi_i}& \poloFantasmaCn{\rB_N}\qw& \qw \end{aligned} \: = \: \begin{aligned} \Qcircuit @C=1.2em @R=.6em @! R { \multiprepareC{1}{\Omega_{k_1(i)}}& \poloFantasmaCn{\rA_1}\qw& \qw\\ \pureghost{\Omega_{k_1(i)}}& \poloFantasmaCn{\rB_1}\qw& \qw\\ \vdots&&\\ \multiprepareC{1}{\Omega_{k_M(i)}}& \poloFantasmaCn{\rA_M}\qw& \qw\\ \pureghost{\Omega_{k_M(i)}}& \poloFantasmaCn{\rB_M}\qw& \qw\\ \prepareC{\omega_{k_{M+1}(i)}}& \poloFantasmaCn{\rA_{M+1}}\qw& \qw\\ \prepareC{\omega_{k_{M+2}(i)}}& \poloFantasmaCn{\rB_{M+1}}\qw& \qw\\ \vdots&&\\ \prepareC{\omega_{k_{2N-M-1}(i)}}& \poloFantasmaCn{\rA_{N}}\qw& \qw\\ \prepareC{\omega_{k_{2N-M}(i)}}& \poloFantasmaCn{\rB_{N}}\qw& \qw\\ \end{aligned} \quad \text{ for }i=0,\,1 \end{equation} where $k_j(i)\in\left[16,23\right]$ for $j=1,\ldots,M$ and $k_j(i)\in\left[0,3\right]$ for $j=M+1,\ldots,2N-M$. The thesis follows immediately from the two assumptions we made: no more than bipartite correlated boxes and pure input states. In fact, for every two pure states a perfectly discriminating procedure always exists, as analysed in Section <ref>. For discriminate between two parallel compositions of pure states the parallel composition of the discriminating POVMs for each pair of states is sufficient. In the previous lemma it would be possible that Alice and Bob have in control also non-factorized bipartite states, i.e. $\Omega_{k_{M+1}}\in\mathsf{St}(\rA_{M+1}\rA_{M+3})$ for $k_{M+1}\in\left[16,23\right]$ instead of $\omega_{k_{M+1}}\otimes\omega_{k_{M+3}}\in\mathsf{St}(\rA_{M+1}\rA_{M+3})$ for $k_{M+1},k_{M+3}\in\left[0,3\right]$, however this does not carry any modification in the proof and so, to not make the notation even more troublesome, we will refer to Eq. (<ref>) as the more general input states. Perfect bit commitment is impossible in PR-box theory. If a bit commitment is perfect it means that it should be correct with probability one, perfectly concealing and perfectly binding. If it is correct with probability one then, by the previous Lemma, the input states $\Psi_0,\,\Psi_1\in\mathsf{St}(\rA^{\otimes N}\rB^{\otimes N})$ must have the form of Eq. (<ref>). If it is also perfectly concealing, then we have to impose the condition of Eq. (<ref>), i.e. $(e|_{\rA^{\otimes N}}|\Psi_0)_{\rA^{\otimes N}\rB^{\otimes N}}=(e|_{\rA^{\otimes N}}|\Psi_1)_{\rA^{\otimes N}\rB^{\otimes N}}$: \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi_0}& \poloFantasmaCn{\rA^{\otimes N}}\qw& \measureD{e_{\rA^{\otimes N}}}\\ \pureghost{\Psi_0}& \poloFantasmaCn{\rB^{\otimes N}}\qw& \qw\\ \end{aligned} \: = \: \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Omega_{k_1(0)}}& \poloFantasmaCn{\rA_1}\qw& \measureD{e_\rA}\\ \pureghost{\Omega_{k_1(0)}}& \poloFantasmaCn{\rB_1}\qw& \qw\\ \vdots&&\\ \multiprepareC{1}{\Omega_{k_M(0)}}& \poloFantasmaCn{\rA_M}\qw& \measureD{e_\rA}\\ \pureghost{\Omega_{k_M(0)}}& \poloFantasmaCn{\rB_M}\qw& \qw\\ \prepareC{\omega_{k_{M+2}(0)}}& \poloFantasmaCn{\rB_{M+1}}\qw& \qw\\ \vdots&&\\ \prepareC{\omega_{k_{2N-M}(0)}}& \poloFantasmaCn{\rB_{N}}\qw& \qw\\ \end{aligned} \: = \: \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Omega_{k_1(1)}}& \poloFantasmaCn{\rA_1}\qw& \measureD{e_\rA}\\ \pureghost{\Omega_{k_1(1)}}& \poloFantasmaCn{\rB_1}\qw& \qw\\ \vdots&&\\ \multiprepareC{1}{\Omega_{k_M(1)}}& \poloFantasmaCn{\rA_M}\qw& \measureD{e_\rA}\\ \pureghost{\Omega_{k_M(1)}}& \poloFantasmaCn{\rB_M}\qw& \qw\\ \prepareC{\omega_{k_{M+2}(1)}}& \poloFantasmaCn{\rB_{M+1}}\qw& \qw\\ \vdots&&\\ \prepareC{\omega_{k_{2N-M}(1)}}& \poloFantasmaCn{\rB_{N}}\qw& \qw\\ \end{aligned} \: = \: \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi_1}& \poloFantasmaCn{\rA^{\otimes N}}\qw& \measureD{e_{\rA^{\otimes N}}}\\ \pureghost{\Psi_1}& \poloFantasmaCn{\rB^{\otimes N}}\qw& \qw\\ \end{aligned}\,. \end{equation*} So, for the systems from $\rB_{M+1}$ to $\rB_{N}$ we have that $\omega_{k_{M+2s}(0)}=\omega_{k_{M+2s}(1)}$ for $s=1,\ldots,N-M$. For the systems from $\rB_{M+1}$ to $\rB_{N}$ we can not make any further deductions since these 8 pure states have all the same marginal. In any case if the protocol is correct with probability one and perfectly concealing it can not be perfectly binding. In fact, for the non-local bipartite states there will be one of the local transformations of Eq. (<ref>) such that \begin{equation} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Omega_{k_j(0)}}& \poloFantasmaCn{\rA}\qw& \gate{U_{k_j}}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Omega_{k_j(i)}}& \qw& \poloFantasmaCn{\rB}\qw& \qw&\qw\\ \end{aligned}\, \: = \: \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Omega_{k_j(1)}}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Omega_{k_j(1)}}& \poloFantasmaCn{\rB}\qw& \qw\\ \end{aligned}\,, \end{equation} as expressed by Eq. (<ref>), where $U_{k_j}\in\tU(\rA)$. Furthermore for all the other states, namely for the local states $\omega_{k_{M+2s+1}(i)}$ with $s=0,\ldots,N-M-1$, $i=0,1$, it is immediate that there will be local transformations $U_{k_{M+2s+1}}\in\tU(\rA)$, $s=0,\ldots,N-M-1$ chosen from the 4 of Eq. (<ref>) that will permit to Alice to switch unnoticed by Bob from $\Psi_0$ to $\Psi_1$ and vice-versa. §.§ Alternative Proof In Section <ref> we pointed out the three sufficient conditions that a theory has to satisfy in order to ensure the impossibility of perfectly secure bit commitment: causality, atomicity of composition and the one required in Axiom <ref>. Clearly PR-box theory is manifestly causal, since Eq. (<ref>) imposes exactly the no-signalling constraint. Furthermore, also atomicity of composition is fulfilled by PR-box theory. The parallel composition of atomic operations is still atomic due to local discriminability, see Ref. [31]. To verify that also the sequential composition of atomic operation is still atomic it only need to consider all the atomic operations in the theory, see Ref. [23], and straightforwardly compute their composition. For what concern Axiom <ref> some considerations are in order. We required the existence of at least one dynamically faithful pure state, $\Psi^{(\rA)}$, and the existence of a purification for every state $R$ that has the same marginal of $\Psi^{(\rA)}$. Actually the latter assumption is excessive. If we look at Theorem <ref>, where this assumption needs to work, it would be enough that every state $|R)_{\rB\tilde{\rA}}$ that is obtained from $|\Psi^{(\rA)})_{\rA\tilde{\rA}}$ by a local channel $\tC$, i.e. $|R)_{\rB\tilde{\rA}}\coloneqq(\tC\otimes\tI_{\tilde{\rA}})|\Psi^{(\rA)})_{\rA\tilde{\rA}}$, is purificable. In fact, if $R$ is so defined, it certainly has the same marginal of $\Psi^{(\rA)}$: \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{R}& \poloFantasmaCn{\rB}\qw& \measureD{e}\\ \pureghost{R}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi^{(\rA)}}& \poloFantasmaCn{\rA}\qw& \gate{\tC}& \poloFantasmaCn{\rB}\qw& \measureD{e}\\ \pureghost{\Psi^{(\rA)}}& \qw& \poloFantasmaCn{\tilde{\rA}}\qw& \qw&\qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.7em @! R { \multiprepareC{1}{\Psi^{(\rA)}}& \poloFantasmaCn{\rA}\qw& \measureD{e}\\ \pureghost{\Psi^{(\rA)}}& \poloFantasmaCn{\tilde{\rA}}\qw& \qw\\ \end{aligned}\, . \end{equation*} Now if we limit to consider $\tC$ as a local reversible channel (since in the theory the reversible transformations are atomic), being $\Psi^{(\rA)}$ pure by hypothesis, due to atomicity of composition also $R$ is pure and hence trivially purificable. So, under this further constraint, we only need to check the existence of a dynamically faithful pure state within the theory. We can help ourselves by the fact that the purification of an internal state is dynamically faithful (and obviously pure). This result was published in Ref. [2] for theories that satisfy local discriminability and purification but it can easily extended to PR-box theory given the existence of at least one internal state that is purificable: $\mu$. Now the answer is immediate, all the non-local bipartite extremal point of the 8-dimensional polytope are dynamically faithful pure states. At this point it is easy to see that we can just choose one of the pure states in Eq. (<ref>), as the dynamically faithful pure state in Axiom <ref>. Finally, it is straightforward to verify that the purification is unique up to a local reversible transformation on the purifying system, see Eq. (<ref>). The further limitation that we imposed, on the atomicity of $\tC$, practically requires that for every commitment protocol the two encodings $\tA_0,\,\tA_1\in\mathsf{Transf}(\rA_1...\rA_N\rightarrow\rB_1...\rB_N\rF_N)$ have the marginal atomic: $(e|_{\rF_N}\tA_0=(e|_{\rF_N}\tA_1\coloneqq\tC$. But from Eq. (<ref>) we know that every element of the set of reversible transformation of the composite system $\rA\otimes\rA$ is nothing else than the tensor product of local reversible transformations (with eventually the swap map). So if we limit to consider only atomic local transformations the initial new constraint is satisfied. Finally, thanks to local discriminability, given two pure faithful states $\Psi^{(\rA)}\in\mathsf{St}(\rA\tilde{\rA})$ and $\Psi^{(\rB)}\in\mathsf{St}(\rB\tilde{\rB})$ for system $\rA$ and $\rB$, respectively, then also $\Psi^{(\rA)}\otimes\Psi^{(\rB)}\in\mathsf{St}(\rA\tilde{\rA}\rB\tilde{\rB})$ is a pure dynamically faithful state for the compound system $\rA\rB$ (a rigorous proof can be found in Ref. [2]) and so the previous discussion is easily generalizable to any $N$-partite system. In conclusion, the proof of impossibility of perfectly secure bit commitment of Section <ref> can be easily extended to include PR-box theory (limited to no more than bipartite correlations and only reversible transformations). § UNCONDITIONALLY SECURE BIT COMMITMENT As outlined before, in Ref. [10] Buhrman et al. proposed a bit commitment protocol that was claimed to be unconditionally secure. They would like to show that superstrong non-local correlations in the form of non-local boxes enable to solve cryptographic problems otherwise known to be impossible. In particular, their result would imply that the no-signaling principle and secure computation are compatible in principle. However, we now prove how it would be possible for Alice to perfectly cheat, i.e. with null probability of being detected by Bob, making use of the local reversible atomic transformations of Eq. (<ref>). In our analysis, we begin from dealing with only one bipartite PR-box and we suppose that Alice's input is the committed bit (even if this is not a well defined bit commitment protocol it is as well an instructive example in order to simplify the following analysis). Alice and Bob share the non-local bipartite PR-box $\Omega_{18}\in\mathsf{St}_\mathbb{R}(\rA\rB)$ (to which corresponds the probability rule $p_{000}$) but they have access only to system $\rA$ and $\rB$, respectively. We can summarise the protocol as follows. * Alice select her committed bit $x$, inputs it and obtain output bit $a$; * Bob inputs a random bit $y$ and obtains output bit $b$. * Alice sends $x$ and $a$ to Bob; * Bob checks to see if $a\oplus b=xy$. If this relation is true, Bob accepts $x$ as the revealed bit, otherwise he knows that Alice has cheated and rejects Alice’s revelation. It is easy to see that Alice has a probability to cheat successfully equal to $\frac{1}{2}$. With the help of local transformation Alice can make the probability of successful cheating equal to 1. In fact, it is sufficient to find $(\alpha,\beta,\gamma)$ of Eq. (<ref>) such that, for given $x$ and $y$ and for the output couple $(a,b)$ (i.e. such that $p_{\alpha\beta\gamma}(a,b|x,y)\ne0$) exists a generic function $f:\{0,1\}\longrightarrow\{0,1\}$ such that, given $a^\prime=f(a)$ and $x^\prime=x\oplus1$, $p_{000}(a^\prime,b|x^\prime,y)\ne0$. Mathematically, to find suitable $(\alpha,\beta,\gamma)$ it is sufficient to resolve \begin{equation}\label{eq:cheating} \begin{cases} & a\oplus b=xy\oplus \alpha x\oplus \beta y\oplus\gamma\\ & a^\prime\oplus b=x^\prime y \end{cases}. \end{equation} We find that $\beta=1$ and $f(a)=a\oplus\alpha x\oplus\gamma$, for every choice of $\alpha$ and $\gamma$. In conclusion, if Alice perform a local transformation (given by Eq. (<ref>)) such that $p_{000}\longrightarrow p_{\alpha1\gamma}$ and then inputs her bit $x$ and gets output $a$, she can reveal $x^\prime$ and $a^\prime=a\oplus\alpha x\oplus\gamma$ to Bob, who will accept with probability 1. We can now consider the unconditionally secure bit-commitment protocol proposed in Ref. [10] where $2n+1$ non-local bipartite PR-boxes in the state $\Omega_{18}\in\st{AB}$ are shared between Alice and Bob. The authors found that Alice probability of successfully cheating is at maximum equal to 1/2 but can be asymptotically reduced if the protocol is repeated $k$ times. However, using local transformation Alice can cheat without being detected with probability 1. We refer to the original article about the commitment protocol and we outline only the "cheating procedure". * Alice wants to commit to bit $c$ but to send to Bob bit $c^\prime=c\oplus1$. So she chooses $x\in\{0,1\}^{2n+1}$ by choosing the first $2n$ bits such that $|x^\prime_1...x^\prime_{2n}|_{11}$ is even where $x_i^\prime=x_i\oplus1$ for $i=1,2,...,2n$ (given a string of even length $x$, $|x|_{11}$ is the number of substring "11" in $x$ starting at an odd position) and then choosing $x_{2n+1}=c$ (analogously she can choose $x\in\{0,1\}^{2n+1}$ such that $|x^\prime_1...x^\prime_{2n}|_{11}$ is odd and $x_{2n+1}=c\oplus1$); * Alice, for each of the $2n+1$ shared boxes, apply a local transformation changing the probability law from $p_{000}\longrightarrow p_{\alpha1\gamma}$, then se puts the bits $x_1,x_2,...,x_{2n+1}$ into the boxes $1,2,...,2n+1$. Let $a_1,a_2,...,a_{2n+1}$ be Alice's output bits from the boxes; * Alice computes the parity of all the "cheated" output bits $A^\prime=\oplus_{i=0}^{2n+1}a^\prime_i$ and send $A^\prime$ to Bob, where $a^\prime_i=a_i\oplus\alpha x_i\oplus\gamma$ for $i=1,2,...,2n+1$; * Bob randomly chooses a string $y\in_\mathbb{R}\{0,1\}^{2n+1}$ and puts the bits $y_1,y_2,...,y_{2n+1}$ into his boxes. We call the output bits from his boxes $b_1,b_2,...,b_{2n+1}$. Then the REVEAL phase: * Alice sends $c^\prime$, her string $x^\prime$ (where $x_i^\prime=x_i\oplus1$ for $i=1,2,...,2n+1$) and all her $2n+1$ "cheated" outputs bits (i.e. $a^\prime_i$) to Bob; * Bob checks if Alice's data is consistent: $\forall i\in\{0,1\}^{2n+1}$, $x^\prime_i\cdot y_i=a^\prime_i\oplus b_i$ and $|x^\prime_1...x^\prime_{2n}|_{11}+x^\prime_{2n+1}+c^\prime$ is even. Since he finds no error, he accepts $c^\prime$ as the committed bit. In fact, according to Eq. (<ref>), every couple $(x^\prime_i,a_i^\prime)$ for $i=1,2,...,2n+1$ sent by Alice to Bob satisfies $x_i^\prime\cdot y=a_i^\prime\oplus b$ and by the proposed choices of $x\in\{0,1\}^{2n+1}$ also the parity constraints are § BIT COMMITMENT IN TRIPARTITE SCENARIO We have stressed that the proof of impossibility of perfectly secure bit commitment in PR-box theory was limited by considering no more than bipartite correlated boxes in $N$-partite systems. In fact, if we take in consideration just only tripartite boxes, the scenario changes considerably. In this Section we highlight that a scheme of bit commitment protocol like the one in Ref. [10] that would make use of tripartite non-local boxes would not subject to the cheating by local reversible transformations and it would represent a possible perfectly (or at least unconditionally) secure bit commitment protocol. In Section <ref> we used the classification done in Ref. [29] where the tripartite correlated boxes were divided in 46 non equivalent classes, i.e. 46 classes whose states are not connected by local reversible transformations. In order to see it directly, it is sufficient to write explicitly the probability rules for those classes. As an example we write out the representatives of the probability rule for three non-local tripartite classes (number 44, 45, and 46 in Ref. [29]): \begin{equation} \label{eq:tripartite-box-prob-rule} \begin{aligned} \text{Class 44: }& p(a,b,c|x,y,z)= \begin{cases} 1/4& a\oplus b\oplus c=xyz\\ 0 & \text{otherwise} \end{cases}\\ \text{Class 45: }& p(a,b,c|x,y,z)= \begin{cases} 1/4& a\oplus b\oplus c=xy\oplus xz\\ 0 & \text{otherwise} \end{cases}\\ \text{Class 46: }& p(a,b,c|x,y,z)= \begin{cases} 1/4& a\oplus b\oplus c=xy\oplus xz\oplus yz\\ 0 & \text{otherwise} \end{cases}\\ \end{aligned}, \end{equation} and we note that the local relabelling of Eq. (<ref>) plus the permutations of the parties do not permit to move from every representative of one class to any one of any other class. Now to build our bit commitment protocol we decide to encode $b=0,1$ by choosing as input state $\Psi_0,\Psi_1\in\mathsf{St}(\rA^{\otimes3})$ one representative of the $44^{th}$ class and one of the $45^{th}$, say the ones in Eq. (<ref>) for the sake of simplicity. Following the strategy in Section <ref> we can construct a POVM that is able to discriminate between them, for example $\{a,e-a\}$ where $a\coloneqq b^{(0)} \otimes b^{(3)}\otimes b^{(0)} + b^{(0)}\otimes b^{(1)}\otimes b^{(2)}+ b^{(2)} \otimes b^{(1)}\otimes b^{(0)}+b^{(2)} \otimes b^{(3)}\otimes b^{(2)}$ and $e=e_\rA\otimes e_\rA\otimes e_\rA$. So the protocol is correct with probability one. Furthermore $\Psi_0$ and $\Psi_1$ are not connected by any local reversible transformation (as pointed out before) and so the protocol is also perfectly binding. At the same time the two input states have the same marginal on Bob system, and so it is also perfectly concealing. Namely, it is not difficult to derive from Eq. (<ref>) that \begin{equation*} \begin{aligned} \Qcircuit @C=1.2em @R=.8em @! R { \multiprepareC{2}{\Psi_0}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Psi_0}& \poloFantasmaCn{\rA}\qw& \multimeasureD{1}{e}\\ \pureghost{\Psi_0}& \poloFantasmaCn{\rA}\qw& \ghost{e}\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.8em @! R { \prepareC{\mu}& \poloFantasmaCn{\rA}\qw& \qw\\ \end{aligned} \, = \, \begin{aligned} \Qcircuit @C=1.2em @R=.8em @! R { \multiprepareC{2}{\Psi_1}& \poloFantasmaCn{\rA}\qw& \qw\\ \pureghost{\Psi_1}& \poloFantasmaCn{\rA}\qw& \multimeasureD{1}{e}\\ \pureghost{\Psi_1}& \poloFantasmaCn{\rA}\qw& \ghost{e}\\ \end{aligned}\, . \end{equation*} In conclusion, by simply choosing as our input states two suitable non-local tripartite boxes we build a protocol that is correct with probability 1, perfectly concealing and seems to be also perfectly binding (we are considering only reversible transformations). Naturally, since the theory is not complete, we would not like to fall in the same mistake of claiming our bit commitment protocol perfectly (or, when applied in Ref. [10], unconditionally) secure in PR-box theory but, less then unexpected turns in the theory, PR-box theory could really represent the first example of a theory with entanglement and bit commitment. CHAPTER: CONCLUSIONS In this thesis we have formalized the bit commitment protocol in the operational language to investigate its feasibility in a more general context than quantum theory. In this way we are willing to make the first step in the understanding of the relation that exists between bit commitment and the operational axioms of quantum theory. In particular we focused on the study of BC in PR-box theory, a theory that is more non-local than the quantum one but where the purification property does not hold. In performing this analysis we were forced to investigate new aspects of PR-box theory, since it is far away to be closed and complete. Some of the most remarkable results that we achieved are the following. Firstly, we described a strategy that grants to always find a POVM able to perfectly discriminate between any two bipartite pure states. In addition, we proved that the maximally mixed state $\mu$, is purificable. Furthermore, if the theory is limited to no more than bipartite correlated boxes, then $\mu$ is the unique internal state that is purificable and its purification is also unique up to reversible transformations on the purifying system. Then, we were able to show how simplistic generalizations to the arbitrary $N$-partite case are not appropriate. For example just admitting tripartite boxes we showed how the purification of the maximally mixed state is not more unique, and how it is not even granted that $\mu$ is still the unique internal state that is purificable. After the presentation of the PR-box theory in Chapter <ref>, in Chapter <ref> we presented the results about the impossibility of bit commitment in PR-boxes. In the literature of BC performed on non-local boxes, these have been rarely considered as part of a coherent theory and in fact tripartite correlated boxes or local reversible transformations (that are admissible in the theory) have always been neglected. By simply taking in consideration the reversible transformations we proposed a proof of impossibility of perfectly secure bit commitment in a PR-box theory limited by the following constraint: * no more than bipartite correlated boxes admitted; * only reversible transformations considered; * only pure input states. Even under these three important limitation our scenario is still enough general to include all the protocol proposed in literature. Furthermore we were also able to adapt the solid proof in Ref. [2] (in same context as above) to obtain an identical result for the impossibility of bit commitment in PR-boxes. In addition, even if we dealt only with perfectly secure bit commitment, we explicitly described a scheme in which Alice is able to cheat perfectly in the protocol proposed in Ref. [10], that was claimed to be unconditionally secure. Finally we relaxed the limitations that we imposed on the theory and we proposed a protocol that seems to be perfectly secure. However, since the theory is not complete, we address future studies to investigate this and the other questions that remain unanswered in this thesis. First of all, if the discriminating strategy could include not only pure states but all of them, then the proof of impossibility of perfectly secure BC could be extended to non-pure input states, too. Anyhow, the great unknown variable is still the integration of generic $N$-partite non-local correlated boxes in the theory. The consequences could be very surprising, for example it would even be possible that the number of internal states that are purificable asymptotically increases increasing $N$, and so that the purification principle could hold in the limit $N\rightarrow\infty$. In conclusion we presented some results on the impossibility of perfectly secure bit commitment in PR-boxes and we precisely pointed out their limit of validity, that, even if under considerable assumptions, still have an important comparison with the literature on the subject. However, as we have outlined many times in this work, to achieve definitive results, other progresses in the fundamental aspects of the theory are absolutely necessary. CHAPTER: TABLES ($\alpha\beta\gamma\delta$) $a\oplus b$ 0000 0 0001 1 0010 $y$ 0011 $y\oplus1$ 0100 1 0101 0 0110 $y\oplus1$ 0111 $y$ 1000 $x$ 1001 $x\oplus1$ 1010 $x\oplus y$ 1011 $x\oplus y\oplus1$ 1100 $x\oplus1$ 1101 $x$ 1110 $x\oplus y\oplus1$ 1111 $x\oplus y$ Outcome relations for the 16 bilocal-boxes. ($\alpha\beta\gamma$) $a\oplus b$ 000 $xy$ 001 $xy\oplus 1$ 010 $xy\oplus y$ 011 $xy\oplus y\oplus1$ 100 $xy\oplus x$ 101 $xy\oplus x\oplus1$ 110 $xy\oplus x\oplus y$ 111 $xy\oplus x\oplus y\oplus1$ Outcome relations for the 8 bipartite nonlocal-boxes. [1] G.M. D'Ariano, P. Perinotti, and A. Tosini, Information and disturbance in operational probabilistic theories, pre-print arXiv:1907.07043 [quant-ph], [2] G. Chiribella, G.M. D'Ariano, and P. Perinotti, Probabilistic Theories with Purification, Physical Review A 81, 062348 [3] J. Bell, On the Einstein Podolsky Rosen paradox, Physics Physique 1 (3): 195–200 [4] B.S. Cirel'son, Quantum generalizations of Bell's inequality, Lett. in Math. Phys. 4, 83 [5] J.F. Clauser, M.A. Horne, A. Shimony, and R.A. Holt, Proposed experiment to test local hidden-variable theories, Phys. Rev. Lett., 23 (15): 880–4 [6] W. van Dam, Nonlocality & Communication Complexity, Ph.D. thesis, University of Oxford, Department of Physics, [7] S. Popescu and D. Rohrlich, Quantum Nonlocality as an Axiom, Foundations of Physics 24, 379 [8] S. Wolf and J. Wullschleger Oblivious transfer and quantum non-locality, [9] A.J. Short, N. Gisin, and S. Popescu, The Physics of No-Bit-Commitment: Generalized Quantum Non-Locality Versus Oblivious Transfer, Quantum Information Processing volume 5, pages131–138 [10] H. Buhrman, M. Christandl, F. Unger, S. Wehner, and A. Winter, Implications of Superstrong Nonlocality for Cryptography, Proceedings of the Royal Society A, 462(2071), pages 1919-1932, [11] G.M. D'Ariano, G. Chiribella, and P. Perinotti, Quantum Theory From First Principles, An informational Approach, Cambridge University Press, [12] G.M. D'Ariano, F. Manessi, P. Perinotti, and A. Tosini, Fermionic computation is non-local tomographic and violates monogamy of entanglement, EPL (Europhysics Letters), 107(2), 20009, [13] G.M. D'Ariano, F. Manessi, P. Perinotti, and A. Tosini, The Feynman problem and fermionic entanglement: Fermionic theory versus qubit theory, International Journal of Modern Physics A Vol. 29 1430025 [14] S. Blum, Coin Flipping by telephone - A Protocol for Solving Impossible Problems, SIGACT News 15, 23 [15] G.M. D’Ariano, D. Kretschmann, D. Schlingemann, and R.F. Werner, Reexamination of quantum bit commitment: The possible and the impossible, Phys. Rev. A 76, 032328 [16] G. Chiribella, G.M. D'Ariano, P. Perinotti, D. Schlingemann, and R.F. Werner, A short impossibility proof of Quantum Bit Commitment, Phys. Lett. A 377 [17] C.H. Bennet and G. Brassard, Quantum Cryptography: Public Key Distribution and Coin Tossing, International Conference on Computers, Systems & Signal Processing [18] G. Brassard, C. Crépeau, R. Jozsa, and D. Langlois in Proceedings of the 34th Annual IEEE Symposium on the Foundations of Computer Science, edited by Leonidas Guibas (IEEE Computer Society Press, Los Alamitos, 1993), p. 362. [19] H.K. Lo and H.F. Chau, Is Quantum Bit Commitment Really Possible?, Phys. Rev. Lett. 78, 3410 [20] D. Mayers, Unconditionally Secure Quantum Bit Commitment is Impossible, Phys. Rev. Lett. 78, 3414 [21] J. Barret, Information processing in generalized probabilistic theories, Phys. Rev A 75, 032304 [22] J. Barret, N. Linden, S. Massar, S. Pironio, S. Popescu, and D. Roberts, Nonlocal correlations as an information-theoretic resource, Phys. Rev. A 71, 022101 [23] G.M. D'Ariano and A. Tosini, Testing axioms for quantum theory on probabilistic toy-theories, Quantum Information Processing 9, 95 [24] A. Einstein, B. Podolsky, and N. Rosen, Can Quantum-Mechanical Description of Physical Reality Be Considered Complete?, Physical Review 47 (10): 777–780 [25] A.J. Short and J. Barret, Strong nonlocality: a trade-off between states and measurements, New Journal of Physics 12, 033034 [26] M. Dall’Arno, S. Brandsen, A. Tosini, F. Buscemi, and V. Vedral, No-Hypersignaling Principle Phys. Rev. Lett. 119, 020401 [27] D. Gross, M. Müller, R. Colbeck, and O.C.O. Dahlsten, All Reversible Dynamics in Maximally Nonlocal Theories are Trivial Phys. Rev. Lett. 104, 080402 [28] S.W. Al-Safi and A.J. Short, Reversible dynamics in strongly non-local Boxworld systems Journal of Physics A: Mathematical and Theoretical 47, 325303 [29] S. Pironio, J.-D. Bancal, and V. Scarani, Extremal correlations of the tripartite no-signaling polytope, J. Phys. A: Math. Theor. 44, 065303 [30] H. Barnum, O.C.O. Dahlsten, M. Leifer, and B. Toner, Nonclassicality without entanglement enables bit commitment Proceedings of IEEE Information Theory Workshop, pp. 386-390, [31] G.M. D'Ariano, F. Manessi, and P. Perinotti, Determinism without causality, Physica Scripta, T163:014013
# Sparsistent filtering of comovement networks from high-dimensional data111ASC acknoweldges R&P grant from Indian Institute of Management Ahmedabad. We are grateful to Vikram Sarabhai Library for providing the data utilized in this paper. All remaining errors are ours. Arnab Chakrabarti Misra Centre for Financial Markets and Economy, Indian Institute of Management Ahmedabad, Gujarat 380015, India. Email: <EMAIL_ADDRESS>Anindya S. Chakrabarti (Corresponding author) Economics Area and Misra Centre for Financial Markets and Economy, Indian Institute of Management Ahmedabad, Gujarat 380015, India. Email<EMAIL_ADDRESS> ###### Abstract Network filtering is an important form of dimension reduction to isolate the core constituents of large and interconnected complex systems. We introduce a new technique to filter large dimensional networks arising out of dynamical behavior of the constituent nodes, exploiting their spectral properties. As opposed to the well known network filters that rely on preserving key topological properties of the realized network, our method treats the spectrum as the fundamental object and preserves spectral properties. Applying asymptotic theory for high dimensional data for the filter, we show that it can be tuned to interpolate between zero filtering to maximal filtering that induces sparsity and consistency while having the least spectral distance from a linear shrinkage estimator. We apply our proposed filter to covariance networks constructed from financial data, to extract the key subnetwork embedded in the full sample network. ## 1 Introduction Network representation of large dimensional complex systems has become a standard methodology to delineate the nature of linkages across a large number of constituent entities comprising the systems [33]. Examples range across systems varying widely in terms of nature and architecture: economic and financial networks [9, 5], social networks [44], biological networks like food webs [45], technological networks like world wide web [20] and transportation networks [40] among many others. Broadly speaking, there are two major strands of literature that starts from the analysis of the realized network. One strand of the literature utilizes networks to explore dynamics on it [32], using the realized network as the true representation of the linkages. The other literature goes backward to extract true linkages from the realized linkages [4, 36], maintaining the idea that some of the realized linkages in fact might be spurious. We are interested in the second stream of literature where the fundamental objective is to isolate and filter the key subnetwork out of a large dimensional realized network. In a complex dynamical system, the correlation matrix of time-varying responses of the constituent entities captures pairwise-linkages between the entities. A co-movement network is constructed by considering each response variable as a node of the graph and an undirected edge between two nodes exists if the corresponding correlation is nonzero. This kind of network construction out of observational multi-variate data has been very successful as a modeling paradigm in finance [31] and biology [12, 3] among others. However, such inference about existence of linkages from purely observational data has a problem. As the pairwise sample correlation is hardly equal to 0 (even when the true correlation is 0), the realized co-movement network will always be a complete graph. The size of the correlation matrix grows as square of the number of nodes. Therefore for real-life data, a complete graph constructed from a such a large correlation matrix might have edges carrying information that would be spurious in nature. Many of the edges, particularly the edges with very low correlation, contain very little information and a likely scenario is that they lead to false discovery of linkages. Hence, before carrying out the statistical analysis of a co-movement network, it is important to extract only the meaningful interactions or correlations. Prominent network filtering techniques, like _minimum spanning tree_ [28] (MST) or _planar maximally filtered Graph_ (PMFG) [43], aspire to do so by reducing the graph to a subgraph containing the maximum amount of information regarding the system’s collective behavior by preserving geometric properties of the realized network (connectivity in case of MST and closed loops with three or four nodes in case of PMFG). The second type of filtering emphasizes the statistical significance of edges [29]. The third type of filtering focuses on the spectral structure [19]. In this paper, we propose a new filtering technique for large networks constructed from high-dimensional data, utilizing the spectral properties. Drawing from statistical theory of high-dimensional covariance matrix estimation, we develop a flexible method to find out the sparse adjacency matrix that represents the key subnetwork of the full network. Theoretically, the filtered network retains maximum similarity with the true spectral structure. The method is quite flexible as it allows the tuning the degree of filtering within a range of zero to maximal permissible pruning of edges. Two important properties of the filter are as follows: first, the filter generates sparsity in the covariance matrix which makes the filtering possible, and secondly, the filter statistically consistently prunes spurious linkages leading to reduction on false discovery of linkages. The combination of these two properties lead to the sparsistence of the resultant filter. Fundamentally, our approach depends on the literature on large dimensional covariance matrix estimation. For a large interacting system, a comovement network ia a high-dimensional graph. Often the number of nodes is in the order of number of observations leading to a well recognized problem that the eigenvalues of the sample covariance matrix do not converge to their population counterpart [30]. This result dictates the fact that the sample covariance matrix is not an consistent estimate of the true covariance matrix [34, 14]. Therefore, efficient estimation of high-dimensional covariance matrix is a relevant problem in context to large network analysis. A broad class of well-conditioned shrinkage/ridge-type estimators were proposed to circumvent the problem [25]. Element-wise regularization methods were also proposed to achieve sparsity [7, 8, 37, 6]. Some of the methods require a natural ordering among the variables [39, 8]. Some of the proposed estimators fail to guaranty positive definiteness. Some form of _tapering_ matrix [17] and maximum likelihood estimator under positive definite and sparsity constraints [11] were proposed to ensure positive definite covariance estimator. We borrow the idea of consistent estimation of sparse covariance matrices from this literature. However, there are two ways of approaching the problem of inference of linkages. In this paper we deal with the graph implied by the covariance matrix and utilize a non-parametric approach. In particular, we leverage the properties of regularized covariance estimators in [25] and [38]. The complementary approach is through the application of graphical Lasso algorithm for Gaussian graphical model [10, 16]. However, graphical Lasso algorithm attempts to estimate the precision matrix and not the covariance matrix. Despite it’s popularity, graphical Lasso algorithm is a strongly parametric approach and applicable within a rather restricted class of models. Several attempts have been made to develop filtering methods while preserving large scale structure [18, 19]. [18] show that local filtering techniques can preserve network properties more that global filtering methods and propose a new sparsification technique that preserves edges leading to nodes of local hubs. [21] integrate spectral clustering and edge bundling for effective visual understanding. Under some distributional assumptions, statistical methods have been proposed to extract the backbone of the network [13, 29]. These works develop statistical tests for significance of edges. Some methods are proposed to find the irreducible backbone of a network from a sequence of temporal contacts between vertices [22]. Finally, we note that the proposed filter is more efficient than the filters based on random matrix theory, as those filters lead to shrinkage of all elements in the correlation matrices due to spectral decomposition without converting any of them to zero. Thus the resultant network is of the same dimension as the original network [41]. There are application of hard thresholding on the resultant network to reduce the size of the network by removing edges with low weights. However, such a technique is fundamentally ad-hoc as there is no intrinsic property that can fix the threshold [41]. In the present context, we avoid both problems by essentially targeting consistent covariance matrix estimator via sparsity and the distance between the eigenspectra of the target and the filtered matrix uniquely pins down the degree of thresholding and consequently, the degree of filtering. The rest of the paper is organized as follows: Section 2 introduces the essential notations used throughout and discusses necessary statistical background. Readers familiar with high-dimensional covariance matrix estimation problem can skip this part and can directly go to the next section 3. In section 4, we present some of the possible alternatives of the choices we made in the algorithm. We have presented applications of the filter to real-life data in section 5. Section 6 summarizes the paper and concludes. ## 2 Notations and Technical Background on Covariance Matrix Estimation Throughout this paper we maintain the following notations: * • $\mathcal{D}$: the $n\times p$ data matrix consisting of $p$ variables and $n$ independent observations. * • $\Sigma$: true (unobserved) covariance matrix ($p\times p$) of $p$ variables. * • $S$: sample covariance matrix of size $p\times p$ calculated from data matrix $\mathcal{D}$. * • $S_{LW}$: Ledoit-Wolf estimator of size $p\times p$ of covariance matrix $\Sigma$. * • $S_{\eta}$: Thresholded (sample) covariance matrix of size $p\times p$ corresponding to the threshold $\eta$. * • $S_{\eta^{*}}$: Maximally filtered covariance matrix of size $p\times p$ for optimally chosen threshold parameter $\eta^{*}$ with zero cost for filtering. * • $S_{\tilde{\eta}}$: Tuned filtering of covariance matrix of size $p\times p$ for optimally chosen threshold parameter $\tilde{\eta}$ for positive cost for filtering. * • $\Gamma(S_{\eta})$: Network corresponding to $S_{\eta}$. Following the above notation, the goal of our proposed methodology is to find an optimal threshold parameter $\eta^{*}$ such that the corresponding filtered network $\Gamma(S_{\eta^{*}})$ will have the sparsistence property. Below we define all concepts and discuss each of the steps in detail. With multivariate data, the population covariance matrix is estimated by its sample counterpart. The sample covariance matrix has unbiasedness and other useful large sample properties [2]. However, these properties are established under the assumption that the number of observations is large while the number of variables being constant. The difference between the multivariate statistical theory and high-dimensional statistics is that the latter considers the case where the number of variables ($p$) also grows with the number of observations ($n$). Under such assumption, the sample covariance matrix does not behave desirably and becomes inconsistent. When sample is drawn from a high-dimensional Gaussian distribution with true covariance matrix $I$, the difference between the true and sample spectra increases with the dimension to size ratio- as illustrated in Fig. 1. This fact is theoretically proved by _Marčenko-Pastur_ theorem and consequent developments [2]. For this reason, several attempts have been made to construct more efficient estimator of high-dimensional covariance matrix. Here, we will describe a few of these approaches which are relevant to this paper and used in section 3 to develop the algorithm. (a) p/n=0.1 (b) p/n=0.5 (c) p/n=1 (d) p/n=1.5 Figure 1: Plot of sorted eigenvalues corresponding to the true (dotted line) and sample covariance matrix (solid line) with $n$ draws from a $p$ dimensional normal distribution with covariance matrix $I$ (identity matrix). Since all eigenvalues of an identity matrix would be equal to one, the true spectrum is shown as a horizontal line at 1. Each panel represents a specific dimension to sample size ratio varying from 0.1 to 1.5. Higher $p/n$ ratio leads to larger deviation of the sample spectrum from the true spectrum [34], which necessitates spectral shrinkage as described in Sec. 2.1 and 2.2. ### 2.1 Stein’s approach Fig. 1 shows that under high-dimensional setup, the eigenvalues of the sample covariance matrix deviates considerably from their population counterparts. However, the problem itself suggests a possible way out. We can see (Fig. 1) that as the dimension to sample size ratio goes up, the sample spectra move further away from the true spectra. So shrinking the eigenvalues towards a central value may lead to a better estimator. Such a strategy was suggested by Stein [42] and the proposed covariance estimator takes the following form: $\hat{\Sigma}=\hat{\Sigma}(S)=P\psi(\Lambda)P^{\prime},$ (1) where the spectral decomposition of $S$ is given by $S=P\Lambda P^{\prime}$, with $\Lambda=\mathrm{diag}(\lambda_{1},\lambda_{2},...,\lambda_{p})$ being the diagonal matrix of eigenvalues of $S$ and $P$ being the matrix of eigenvectors; $\psi(\Lambda)=\mathrm{diag}(\psi(\lambda_{1}),\psi(\lambda_{2}),...,\psi(\lambda_{p}))$ is also a diagonal matrix. If $\psi(\lambda_{i})=\lambda_{i}\forall i$ then $\hat{\Sigma}$ is the usual sample covariance matrix $S$. $\psi$ shrinks the eigenvalues $\lambda$ and thus reduce the deviation from its true counterpart. Clearly, this approach only regularizes the eigenvalues and keep the eigen vectors of the sample covariance matrix unaltered. Due to this reason this type of estimators are also called _rotation equivariant_ covariance estimator. ### 2.2 Ledoit-Wolf estimator A problem with Stein’s original prescription is that it does not ensure monotonicity and nonnegativeness of the eigenvalues [34]. This problem had been addressed by Ref. [25] which formulated a general approach towards shrinkage by defining a rotation equivariant regularization based on the following minimization problem: $\underset{\Psi}{\text{min}}\|P\Psi P^{\prime}-\Sigma\|$ (2) where $\|.\|$ can be any matrix norm. Most widely considered norm is the Frobenius norm.222The Frobenius norm of an arbitrary matrix $A$ of order $r\times m$ is $\|A\|^{F}=\sqrt{\text{tr}(AA^{\prime})/r}$. A particularly useful solution for the optimal $\psi$ was proposed by Ledoit and Wolf [25] which is based on the observation that the sample covariance matrix is an unbiased estimator of the population covariance matrix. This fact remains true for high-dimensional data as well. But in high-dimensional setup, the sample covariance matrix becomes considerably unstable i.e. the deviation from the true covariance matrix can potentially be large. On the other hand if we use a structured covariance estimator- such as an identity matrix then the estimator, while being severely biased under misspecification of the structure, will have very little variability. They showed that a suitably chosen linear combination of these two types of estimators would outperform each of them where the coefficients/weights of linear combination is chosen to optimize the bias-variance trade off. Formally, the Ledoit-Wolf estimator $S_{LW}$ [25] is defined as $S_{LW}=\alpha_{1}I+\alpha_{2}S$ (3) where $I$ is a $p\times p$ identity matrix and $\alpha_{1}$ and $\alpha_{2}$ are chosen to minimize the risk corresponding to the loss function $p^{-1}\mathrm{tr}(S_{LW}-\Sigma)^{2}$. #### 2.2.1 Consistency of Ledoit-Wolf estimator Since $S$ is positive definite, $S_{LW}$ can also be shown to be positive- definite and consistency of such estimator depends on the growth rate of $p$, the moments and the association structure of the data [25]. More precisely, the $p(p+1)/2$ elements of the true covariance matrix can be consistently estimated if three conditions hold described below. In large dimensional covariance matrices, both $p$ and $n$ grows. Therefore it is a common practice to write $p$ as $p_{n}$ (function of $n$) and to consider $n$ in the limit. Let us define $Y=\mathcal{D}V$, where $\mathcal{D}$ is a $n\times p$ matrix and $V$ is the matrix whose columns are the normalized eigenvectors of $\mathcal{D}$. Denote the $i$th entry of any row by $Y_{i}$. Also, let us denote the set of all quadruples made of four distinct elements of $\\{1,~{}2,~{}3,~{}..,~{}p\\}$ as $Q_{n}$. The three conditions are the following: 1. C1: There exists a constant $K_{1}$ independent of $n$ such that $p_{n}/n\leq K_{1}$. 2. C2: There exists a constant $K_{2}$ independent of $n$ such that $\sum_{i=1}^{p_{n}}E(Y_{i})^{8}<K_{2}$. 3. C3: $\underset{n\rightarrow\infty}{\mathrm{lim}}\frac{p_{n}^{2}}{n^{2}}\times\frac{\sum_{(i,j,k,l)\in Q_{n}}(Cov(Y_{i}Y_{j},Y_{k}Y_{l}))^{2}}{\\#Q_{n}}=0,$ where $\\#Q$ denotes the cardinality of the set $Q$. C1 says that $p$ can either remain constant or grow with $n$. That means this method cannot be used (more specifically consistency cannot be achieved) for data for which $p>>n$. ### 2.3 Sparsity and threshold estimator Threshold estimator of high-dimensional covariance matrix regularizes both eigenvalues and eigenvectors as opposed to the Ledoit-Wolf estimator which only regularizes the eigenvalues of the sample covariance matrix. Threshold estimator is particularly useful when the true covariance matrix from the data generating process is _sparse_ , i.e. many of the non-diagonal entries of the covariance matrix are 0 or close to 0. This assumption is reasonable for a wide range of practical scenarios. Threshold estimator forces all the off- diagonal entries below a suitably chosen threshold to 0. Even if the corresponding entries of the true covariance matrix are nonzero, the threshold estimates of those entries entail only a bias but no dispersion as estimated by a fixed constant which is zero. The objective function would be the Frobenius norm (see footnote 2) of the difference between the thresholded matrix and empirical covariance matrix obtained from repeated sampling [7]. If the threshold is large, this method produces a sparse covariance matrix. We will denote the threshold matrix by $S_{\eta}$, where $\eta>0$ is the chosen threshold and $S$ is the usual sample covariance matrix; $\displaystyle S_{\eta}$ $\displaystyle\equiv$ $\displaystyle[s_{\eta}({i,j})]$ (4) $\displaystyle\equiv$ $\displaystyle[s_{\eta}(i,j)I(|s(i,j)|<\eta)]\quad~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}$ where $I(.)$ is the indicator function. So the entries of $S$ which are less than $\eta$ in magnitude are replaced by 0. The optimal threshold parameter $\eta$ can be chosen by cross validation. The resulting threshold estimator would be consistent under the assumption $\log(p)/n\rightarrow 0$, and it is shown to be uniform for a class of matrices satisfying a condition that captures a notion of “approximate sparsity”. One problem of such estimator is that it does not always preserves positive definiteness [7]. Threshold estimators can be further extended to a broader class of matrices called _generalized thresholding operators_ , which combines two regularization methods: thresholding and shrinkage [38]. When the true covariance is sparse, generalized thresholding estimators can identify the the true zero entries with probability 1. This property commonly called as _sparsistency_. The sufficient conditions required to achieve this are the following [34]: 1. C1: The data generating process is Gaussian. 2. C2: The variances are bounded above by a constant, i.e. $\sigma_{i,j}\leq C$ for a sufficiently large $C$. 3. C3: $C\sqrt{\frac{\text{log}~{}p}{n}}=o(1)$. ## 3 Sparsistent Filtering of Networks Fundamentally, our objective is to combine the feature of sparsity from the large dimensional covariance matrix estimation along with preservation of the underlying network topology. We elaborate on these two related but separate features below. Intuitively, the problem of filtering a complex network is equivalent to the problem of deleting a number of “less-important” edges from the original graph such that it become less complicated and reveals an underlying structure. Therefore sparsity is an essential property of the adjacency matrix of a filtered network. Evidently, this can be achieved by a threshold estimator with appropriately chosen threshold parameter. However, in the context of network filtering, the choice of a threshold is often ad-hoc and suffers from lack of robustness [41]. Therefore, the technical problem is what can be an efficient method for choosing the threshold that retains sparsity but is also statistically robust? One candidate would be cross-validation [35]. However, a threshold implied by cross-validation is purely numerical in nature and fully dependent on the realized covariance matrix where the realized covariance matrix is itself a random sample. As a direct implication, such a threshold does not allow inference on the underlying structure of the true covarince matrix (and therefore, the resulting network structure). Within a certain restricted class of data generating process, a threshold estimator can indeed be consistent [7]. However, the corresponding restrictions are too severe for direct applications to real-life data (e.g. the assumption of multivariate Gaussianity is often rejected in systems exhibiting large fluctuations, like in the case of stock market data [41]). Additionally, a cross-validated threshold estimator would exhibit theoretical consistency only when the true covariance matrix is sparse to begin with. However, a network filter should be flexible enough to consider a scenario where the linkages are of small strength, but non-zero. A further question arises here about the best way to capture the network topology. One can consider observable geometric properties as well as the spectral structure of a network. There are filters which focus on the geometric properties (like connectivity in case of minimum spanning tree or closed loops in case of planar maximally filtered graphs [43]). However, in the present context, the spectral structure of networks is the best candidate: one, the spectral structure by definition captures network topology, and two, it is amenable to asymptotic theories and links naturally with covariance matrix estimation. In sum, our goal is to find a way to retain the feature of sparsity in one hand (to make the estimator efficient) while preserving the network topology on the other. We require the filter should be flexible enough to interpolate between the two. We achieve it by combining the Ledoit-Wolf estimator as the target for retaining statistical consistency and imposing a threshold estimator that emulates the corresponding spectral structure. We show that the resulting filter inherits both sparsity as well as consistency. We explain the main idea in Fig. 2. The red thick line represents spectral distance of the candidate thresholded matrix from a target matrix by increasing the threshold from zero to a large enough value (we will provide the analytical details below). This distance represents a cost, indicating that a higher distance is less efficient. The distance as a function of the threshold is non-monotonic in nature. By increasing the threshold, initially the distance between the thresholded matrix from the target spectrum reduces and beyond a level, further increase in the threshold leads to an increase in the distance. The global minimum here corresponds to what we call maximal filtering, which is obtained by simply minimizing the spectral distance. However, we have to also consider a case where the true network is not necessarily sparse and there can be edges which can be of small magnitude. Thus in a general context, deleting them would entail a cost. We note that for a small threshold the edges being deleted would have small weights. But as the threshold is increased, we will filter edges with larger weights. This idea can be captured through a convex cost function, which indicates that as the threshold increases, the cost associated with deletion of edges with higher weights also increases. The blue dashed line shows a stylized cost curve. Therefore, the proper objective function is to minimize the total cost (adding spectral distance and the cost of edge deletion) indicated by the black dotted line. The final tuned filter would extract a threshold which is less than the threshold for maximal filtering as shown on the $x$-axis. Figure 2: Illustration of the distance function and cost function against the values of $y(\eta)$\- the number of deleted edges. The x-axis represents values of $y$, the black and red curves stand for $d(F^{S_{\eta}},F^{S_{LW}})$ (Eq. 5) and $C(\eta)$ (Eq. 9). The algorithm described below explains the steps. ### 3.1 Sequential steps of the algorithm Given the data matrix $\mathcal{D}$ of size $n\times p$, our algorithm goes through the following steps and return a thresholded matrix $S_{\tilde{\eta}}$ of size $p\times p$. 1. 1. Sample Covariance Matrix Construction: From the data $\mathcal{D}$, we calculate the sample covariance matrix $S$ of size $p\times p$. 2. 2. Construction of Ledoit-Wolf estimator: From the sample covariance matrix $S$, we calculate Ledoit-Wolf estimator $S_{LW}$ (following Eqn. 3).333For numerical implementation in R, we used the R package RiskPortfolios for Ledoit-Wolf matrix calculation (https://github.com/ArdiaD/RiskPortfolios) with type ‘oneparm’. For numerical implementation in Matlab, we have used the code titled ‘cov1para.m’ obtained from the code repository of Ledoit-Wolf estimator (https://www.econ.uzh.ch/en/people/faculty/wolf/publications.html#9). 3. 3. Finding the Spectrum of Ledoit-Wolf estimator: Eigenvalue decomposition of the covariation matrix444We choose to decompose the covariance matrix to obtain the spectrum. However one can certainly perform the identical analysis on the correlation matrix as well. Sometimes covariance can be too small and can exhibit some computational problem while implementation. Therefore we suggest it to be used on the correlation matrix. $S_{LW}$ gives us the spectrum denoted by $\lambda(S_{LW})$, which is a $p$ dimensional vector comprising $p$ eigenvalues. The empirical distribution function of $\lambda(S_{LW})$ is denoted by $F^{S_{LW}}$. 4. 4. Quantifying the Spectral Distance: The goal is to find a sparse thresholded matrix that is proximate to $S_{LW}$ in terms of spectrum. We define the spectral distance between two matrices $A$ and $B$ as $\text{d}(F^{A},F^{B})$ as the Euclidean distance between two spectra $\lambda(A)$ and $\lambda(B)$: $\text{d}(F^{A},F^{B})\equiv d(\lambda(A),\lambda(B))=\big{(}\sum_{i=1}^{p}\big{|}\lambda_{i}^{A}-\lambda_{i}^{B}\big{|}^{2}\big{)}^{1/2}.$ (5) 5. 5. Maximally Filtered Network: In this step, we obtain the “strong” or the “maximal” filter $S_{\eta}^{*}$ corresponding to a threshold $\eta^{*}$ that has the least spectral distance from $S_{LW}$. Formally, $S_{\eta^{*}}$ can be obtained by minimizing the distance (Eqn. 5) of the resultant thresholded matrix from the Ledoit-Wolf matrix: $\eta^{*}=\underset{\eta}{\text{argmin}}\\{d(F^{S_{\eta}},F^{S_{LW}})\\}.$ (6) Suppose, the number of edges being deleted in $S$ to reduce the matrix to $S_{\eta^{*}}$ is $y^{*}$.555$y^{*}$ is a function of the threshold $\eta^{*}$. A higher threshold $\eta^{*}$ leads to deletion of a higher number of edges, implying that $y^{*}$ would also be higher. Formally, we write $y^{*}=\underset{i\neq j}{\sum}I(S_{\eta^{*}}(i,j)=0)$ (7) where $I(.)$ is the identity function and $S_{\eta^{*}}(i,j)$ is the $(i,j)$th entry of $S_{\eta^{*}}$. 6. 6. Tuned Filtering with Costly Edge deletion: Now we impose a cost for deleting edges. The cost-adjusted optimal edge filtering leads to the following optimization: $\tilde{\eta}=\underset{\eta}{\text{argmin}}\\{d(F^{S_{\eta}},F^{S_{LW}})+C(\eta)\\}.$ (8) where $C(\eta)$ denotes the cost of deleting $y$ edges by implementing threshold $\eta$. For practical implementation, it is easier to work with the cost function on the edges to be deleted ($C(y)$) rather than the threshold ($C(\eta)$). A natural requirement for $C$ as a function of the number of edges to be filtered, is that it should be non-negative, continuous and potentially increasing in the first derivative leading to convexity. A typical candidate for a flexible functional form of $C$ is as follows666The parameters $\theta_{1}$ and $\theta_{2}$ in the cost function given by Eqn. 9 has to be specified by the user depending on the problem and the context of application.: $C(y)=\theta_{1}y^{\theta_{2}}\hskip 8.5359pt\text{where}\hskip 8.5359pt\theta_{1}\geq 0,~{}\theta_{2}>1~{}\text{and}~{}y\equiv y(\eta).$ (9) Eqn. 8 can be equivalently written in terms of deleted edges: $\tilde{y}=\underset{y}{\text{argmin}}\\{d(F^{S_{\eta}},F^{S_{LW}})+C(y)\\}.$ (10) The resulting tuned threshold is $\tilde{\eta}$, number of edges deleted is $\tilde{y}$ and finally, the thresholded matrix is $S_{\tilde{\eta}}$. 7. 7. Tuned Filtered Network with Sparsistence: We create the filtered network $\Gamma_{\tilde{\eta}}$ adjacency matrix from $S_{\tilde{\eta}}$. An edge is present if the corresponding element in $S_{\tilde{\eta}}$ is nonzero.777The covariances would not represent a metric since they can be negative. If we want to visualize the network in the metric space, we can convert covariances into correlations by dividing each covariance entry by the product of sample standard deviations of the corresponding pair of nodes, and these correlations can be transformed into a metric by using the transformation $\gamma_{ij}=\sqrt{2(1-\rho_{ij})}$ where $\rho_{ij}$ is the correlation between $i$ and $j$-th nodes [41] obtained from $S_{\eta}$ for any given $\eta$. By construction, the maximally filtered network $\Gamma(S_{\eta^{*}})$ would be a subset of the tuned filtered network $\Gamma(S_{\tilde{\eta}})$. $\Gamma(S_{\tilde{\eta}})$ in turn would be a subset of the original network $\Gamma(S)$ corresponding to the sample covariance matrix $S$. In our empirical studies, we have seen that the threshold $\eta^{*}$ is typically higher than the threshold chosen by the conventional threshold-estimator which is based on cross-validation (see Sec. 2.3). This implies that the sparsistency property (see Sec. 2.3 for sufficient conditions) will be maintained by maximal filtering because edges deleted for a threshold will also be deleted for all higher thresholds. In other words all spurious edges will be removed with probability one. ### 3.2 An illustrative example (a) True network: $\Gamma(\Sigma)$ (b) Realized network: $\Gamma(S)$ (c) Maximally filtered network: $\Gamma(S_{\eta^{*}})$ (d) Network with tuned filtering: $\Gamma(S_{\tilde{\eta}})$ Figure 3: Illustration of the proposed algorithm. Panel (a): True sparse network $\Gamma(\Sigma)$ for simulation study with 15 edges and 10 nodes. Panel (b): The realized network $\Gamma(S)$ from simulated data corresponding to the sample covariance matrix $S$. Due to finite sampling ($n=50$), the network is fully connected with 45 edges (for $p$ = 10, the maximum number of edges $p(p-1)/2$ = 45). Panel (c): The maximally filtered network $\Gamma(S_{\eta^{*}})$ when $\eta^{*}=0.499$, with 6 edges. Panel (d): The tuned filtered network $\Gamma(S_{\tilde{\eta}})$ where the optimally chosen threshold $\tilde{\eta}$ is 0.170. $\Gamma(S_{\tilde{\eta}})$ has 20 edges. Here we present an example of our proposed filtering method to illustrate (1) how the filter produces a sparse network, (2) how different is this filtered network compared to the _true_ underlying network and (3) how the filtered network changes with the choice of the cost parameters. We choose the true data-generating process to be a $p(=10)$-dimensional Gaussian distribution with mean $\bm{0}$ and covariance matrix $\Sigma$. We illustrate the method for a particular choice of $\Sigma$ (given in Appendix A.1). $50$ sample observations are drawn from this $p$-dimensional distribution. Therefore $p/n$ ratio is 10/50=0.2. The sample covariance matrix ($S$) is calculated from the simulated data (see Appendix A.1). Applying the proposed algorithm, we get the filtered network. Fig. 3 shows the true comovement network for our chosen covariance matrix $\Sigma$. Although it has a moderately sparse structure (15 undirected edges), the sample correlation matrix obtained from the simulated data is not a sparse matrix. The network constructed from the sample covariance matrix $S$, is shown in Fig. 3 (45 undirected edges indicating a fully connected network). We plot two filtered networks corresponding to two choices of the threshold parameters. If we ignore the cost due to deletion of edges- i.e. if we only aim to reduce the distance between the spectral structure of the _threshold_ -network and the network induced by Ledoit-Wolf estimator- then we get the maximally filtered network shown in Fig. 3. We can see that this is not a connected network but it is able to preserve the stronger edges and the corresponding subnetworks of the true network. On the other hand, Fig. 3 represents the filtered network obtained by introducing a positive cost, which preserves the stronger edges along with the property of connectedness. Figure 4: Illustration of the distance function against number of deleted edges corresponding to a chosen threshold- denoted by $y(\eta)$. The distance is minimized at $y=39$ which means that the maximal filtering will remove 39 edges from $\Gamma(S)$ (the corresponding network is plotted in Fig. 3). ### 3.3 Filtering with known data generating process: Information loss and spuriousness If we know the true data generating process, then tuned filtering on a sample covariance matrix gives us two informational statistics related to information loss due to edge deletion and spuriousness of edges generated by finite sampling fluctuations. Fundamentally, these two statistics are related to finding false negatives (deleting edges that are actually informative) and false positives (retaining edges that appear in the sample covariance matrix due to sampling fluctuation, but are not there in the true covariance matrix). The first measure we define allows us to characterize true positives. We construct the measure by the proportion of _true_ edges which are retained in filtered network: $\text{P}_{t}(\eta)=\frac{\\#\\{E_{\Gamma(\Sigma)}\cap E_{\Gamma(S_{\eta})}\\}}{\\#\\{E_{\Gamma({\Sigma})}\\}}$ (11) where $\\#$ denotes the cardinality of a set, $E_{\Gamma(\Sigma)}$ and $E_{\Gamma(S_{\eta})}$ are the edge sets of the true and filtered networks. The numerator and denominator of the above equation only count the number of edges and do not take into account the relative importance of the edges. The next measure replace the _total number_ of edges by _total weight_ of the edges where weight is captured by absolute value of correlation: $\text{P}^{\prime}_{t}(\eta)=\frac{\underset{(i,j)\in\\{E_{\Gamma(\Sigma)}\cap E_{\Gamma(S_{\eta})}\\}}{\sum}w_{i,j}}{\underset{(i,j)\in\\{E_{\Gamma(\Sigma)}\\}}{\sum}w_{i,j}}.$ (12) Clearly, both $\text{P}_{t}(\eta)$ and $\text{P}_{t}^{\prime}(\eta)$ are bounded above by 1. As a consequence of being a complete graph, the sample covariance matrix has both the quantities equal to 1. The main challenge is to obtain a sparse graph with significantly high $\text{P}_{t}(\eta)$ and $\text{P}_{t}^{\prime}(\eta)$. However, high rate of edge retention might lead to retaining spurious edges. Therefore, it is important also to note how many edges the filtered network contains which are not part of the true network. The following proportion measures the same: $\text{P}_{f}(\eta)=\frac{\\#\\{E_{\Gamma(S_{\eta})}\cap E^{c}_{\Gamma(\Sigma)}\\}}{\\#\\{E_{\Gamma(S_{\eta})}\\}}$ (13) where $E^{c}_{\Gamma(\Sigma)}$ denotes the set of edges that do not exist in the true network (but might arise due to sampling). We report these three statistics in Table 1 for different threshold parameters $\eta$ for the data generating process discussed in Sec. 3.2. We see that when the threshold increases, $\text{P}_{t}(\eta)$ and $\text{P}^{\prime}_{t}(\eta)$ decrease on average. This is intuitive because a higher threshold leads to higher number of edges being deleted and therefore, the chances of deleting true edges also go up. On the contrary, a higher threshold simultaneously makes it more likely that spurious edges will be deleted. Therefore, the chances of having a false positive goes down. This is consistent with the column for $\text{P}_{f}(\eta)$ which shows that with higher threshold, the value of $\text{P}_{f}(\eta)$ decreases steadily. $\eta$ | $\text{P}_{t}(\eta)$ | $\text{P}^{\prime}_{t}(\eta)$ | $\text{P}_{f}(\eta)$ ---|---|---|--- 0.170 | 0.742 ($\pm$ 0.09) | 0.917 ($\pm$ 0.04) | 0.445 ($\pm$ 0.07) 0.230 | 0.698 ($\pm$ 0.09) | 0.899 ($\pm$ 0.05) | 0.351 ($\pm$ 0.09) 0.288 | 0.568 ($\pm$ 0.07) | 0.827 ($\pm$ 0.05) | 0.157 ($\pm$ 0.10) 0.499 | 0.454 ($\pm$ 0.14) | 0.752 ($\pm$ 0.08) | 0.072 ($\pm$ 0.14) Table 1: Measures $\text{P}_{t}(\eta)$, $\text{P}^{\prime}_{t}(\eta)$, $\text{P}_{f}(\eta)$ are shown for different threshold parameters $\eta$ for the data generating process discussed in Sec. 3.2. We have computed the measures based on 100 draw of sample covariance matrices of length $n=50$ and $p=10$. The average estimates over these 100 simulations appear in the table along with the corresponding standard deviation within the adjacent brackets. ## 4 Extensions and Robustness In the following, we discuss extensions and robustness of the proposed filtering algorithm. ### 4.1 Spectral similarity in terms of subset of eigenmodes In the algorithm presented in section 3, we have optimized on the threshold $\eta$ to minimize the distance between the spectrum of the Ledoit-Wolf estimator and the threshold estimator. However, one may not be interested in the full spectrum of covariance matrix. This is pertinent in the context of financial networks which is known to possess an eigenvalue distribution with wide heterogeneity. An array of statistical analysis (see e.g. [41]) shows that the highest eigenvalue captures the fluctuations due to the market mode, whereas sectoral fluctuations are associated with the deviating eigenvalues (except the largest one) from the bulk of the spectrum. The bulk of the spectrum on the other hand represents idiosyncratic fluctuations, which is modelled well by a Marčenko-Pastur distribution. Therefore in this context, only the deviating eigenvalues are informative. So one can argue for considering only these few eigenvalues and choose the threshold that minimizes the distance between two vectors of deviating eigenvalues. This represents evaluating the distance on a smaller set of eigenmodes as opposed to all eigenmodes, and the distance between eigenmodes of two matrices $A$ and $B$ to (Eqn. 5) to be modified as follows: $d(\lambda(A),\lambda(B))=\big{(}\sum_{i=p_{l}}^{p_{h}}\big{|}\lambda_{i}^{A}-\lambda_{i}^{B}\big{|}^{2}\big{)}^{1/2}$ (14) where $1\leq p_{l}\leq p_{h}\leq p$ and the choice of $p_{l}$ and $p_{h}$ can be chosen according to the specific system under analysis. Specifically, the upper bound of the Marčenko-Pastur distribution can provide such a natural cut-off for the choice of $p_{h}$ to capture the deviating eigenmodes and $p_{l}$ can be unity. ### 4.2 Non-linear shrinkage estimator All the rotation-equivariant estimators we have discussed so far are linear. They are linear combination of the sample covariance (or correlation) matrix and a suitable shrinkage target. This means that regardless of their ranks, all the sample eigenvalues are shrunk by same intensity. However our objective is to minimize Eqn. 2 and there is no guarantee that a linear shrinkage estimator would be our best choice. [24, 26] show that linear shrinkage is a first order approximation of a nonlinear problem whose utility depends very much on the situation- particularly on the limit of $p/n$. If this ratio is high then linear shrinkage will be a substantial improvement but not otherwise. Attempts have been made to find nonlinear solution to the problem which essentially results in individualized shrinkage intensity to every sample eigenvalue. First we describe the role of random matrix theory and why it is instrumental in finding the solution. Results from random matrix theory illustrate that for high dimensional set up the eigenvalues of sample covariance matrix do not converge to its population counterparts. However, random matrix theory attempts to establish a link between the two. First attempts in nonlinear shrinkage estimators harnessed the established relation between the limiting spectral distribution of the sample eigenvalues and that of the population eigenvalues. Once the spectral distribution of the population eigenvalues are obtained, it can be numerically inverted to calculate the population eigenvalues [24, 15]. Below, we describe one such solution. Let us introduce the following quantities: 1. 1. $p/n\rightarrow c~{}(>0)$. 2. 2. If $G$ be the cumulative distribution function of eigenvalues $\lambda$, then the Stieltjes transform of the $G$ is defined as below: $m_{G}(z)=\int_{-\infty}^{\infty}\frac{1}{\lambda-z}dG(\lambda)\quad\forall z\in\mathbb{C}^{+}.$ Stieltjes transform is an important tool in random matrix theory because of its one-one relationship with the distribution function (empirical spectral distribution in our context). Therefore, to determine the limiting spectral distribution one only needs to show the convergence of corresponding Stieltjes transform. 3. 3. The limiting empirical spectral distribution of the sample covariance matrix is denoted as $F$. 4. 4. The Stieltjes transform of the Marčenko-Pastur law [2] is denoted by $m_{F}(z)$. 5. 5. Define $\tilde{m}_{F}(\lambda)=\underset{z\in\mathbb{C}^{+}\rightarrow\lambda}{\text{lim}}m_{F}(z)\quad\forall\lambda\in\mathbb{R}-\\{0\\}$ 6. 6. Define $m_{LF}(z)=1+zm_{F}(z)\quad\forall z\in\mathbb{C}^{+}$. Under some general assumptions, [24] proposed nonlinear shrinkage intensities and the derived form of $\psi$ (see Eqn. 1) is the following: $\psi_{i}=\frac{\lambda_{i}}{|1-c-c\lambda_{i}\tilde{m}_{F}(\lambda_{i})|^{2}}.$ (15) Note that $\psi_{i}$ is dependent on $\lambda_{i}$ (unlike the linear shrinkage estimator). For more detailed discussion, see [15, 24, 26]. There are also some other methods of nonlinear shrinkage estimation. By exploiting the connection between nonlinear shrinkage and nonparametric estimation of Hilbert transform of the sample spectral density, an analytical formula for nonlinear shrinkage has been proposed recently [27]. [1] proposed a method called Nonparametric Eigenvalue-Regularized COvariance matrix estimator (NERCOME; [23]) that splits the sample into two parts. One part is used to estimate the eigenvectors of the covariance matrix and the other part of sample to estimate the eigenvalues associated with these eigenvectors. Averaging over a sufficiently large number of sample split results in reasonably good estimation. All these methods can be used as alternative to the linear shrinkage estimator considered in the algorithm proposed in Sec. 3. However, the difference in filtering via linear and nonlinear shrinkage estimators is often sample-dependent and exhibits large fluctuations. We attempted to filter financial networks via nonlinear shrinkage estimator (details given in Sec. 5). Through empirical analysis, we saw that when the filtering is moderate via linear estimator, then the nonlinear estimator does not produce radically different filtering. However, we have observed extreme cases where linear shrinkage estimator leads to complete filtering of all edges whereas nonlinear shrinkage estimator led to very minor filtering. Therefore, while nonlinear shrinkage estimator has more flexibility [26], the corresponding impact on the strength of filtering is sample-dependent. ### 4.3 Alternative choices of the cost function A convex cost function captures the idea that higher number of edges being deleted would entail a higher per unit cost.888A concave cost function would lead to maximal filtering, since more filtering leads to lower per unit cost. The underlying idea is that the first few edges being deleted would have the lowest weights. However, as we increase the threshold, the edges being filtered out would have larger and larger weights. This observation leads to the assumed convexity of the cost curve. The proposed functional form $C(y)=\theta_{1}y^{\theta_{2}}$ is useful for its simplicity and ease of manipulation. In principle, many other cost functions can be considered as long as they are convex in nature. A more general set of choices is presented below. Suppose the $(i,j)$th element of $S$ is denoted by $s_{i,j}$. We define the total edge weight as $W=\sum_{i\neq j}f(s_{i,j})$, for a suitable nonnegative function $f$. The total weight removed by filtering can be denote by $W_{\eta^{*}}$, which is defined as follows: $W_{\eta}=\underset{(i,j)\in\bar{E}}{\sum}f(s_{i,j}),~{}\text{where}~{}\bar{E}=E_{\Gamma(S)}-E_{\Gamma(S_{\eta})}.$ (16) In principle, any convex function of $\frac{W_{\eta}}{W}$ is a valid choice for $C(\eta)$. As an example, if we choose $f(s_{i,j})=I(s_{i,j}\neq 0)/2$ then $W_{\eta^{*}}$ becomes $y^{*}$ as defined in Eqn. 7.999The cost function defined in Eqn. 10 is: $C(\eta)\equiv C(y(\eta))\equiv\theta_{1}W_{\eta}^{\theta_{2}}.$ (17) This choice of $f(.)$, although simple and easily understandable, does not directly incorporate the weight of each deleted edge. Therefore an exogenous cost function is needed to be defined and imposed (see Eqn. 9). An endogenous choice is the following: $f(s_{i,j})=|s_{i,j}|/2~{}\text{and}~{}C(\eta)\propto\frac{W_{\eta}}{W}.$ (18) Given the above discussion, we see that this function is convex. ### 4.4 Alternative choices of the distance function In the description of the algorithm, we have utilized Euclidean distance between the spectra (Eqn. 5). This is useful in terms of implementation as well as simplicity. In principle, one can look for alternative notions of distances as well to measure similarity or dissimilarity between two spectra. We consider them below and discuss the relative merits and demerits. Suppose $F(.)$ and $G(.)$ are two spectral distribution functions. For finite sample let us denote the vectors of ordered eigenvalues corresponding to two $p\times p$ matrices as $\lambda_{a}=(\lambda_{a,1},\lambda_{a,2},..,\lambda_{a,p})$ and $\lambda_{b}=(\lambda_{b,1},\lambda_{b,2},..,\lambda_{b,p})$. In this case, $F$ and $G$ are the discrete uniform distribution on $\lambda_{a}$ and $\lambda_{b}$. Three well known measures for distance are as follows: 1. 1. _Minkowski distance (for general $\kappa$)_: $d(F,G)=d(\lambda_{a},\lambda_{b})=\big{(}\sum_{i=1}^{p}\big{|}\lambda_{a,i}-\lambda_{b,i}\big{|}^{\kappa}\big{)}^{\nicefrac{{1}}{{\kappa}}}$, where $\kappa\geq 1$. 2. 2. _$L^{1}$ distance:_ $d(F,G)=d(\lambda_{a},\lambda_{b})=\sum_{i=1}^{p}\big{|}\lambda_{a,i}-\lambda_{b,i}\big{|}$. 3. 3. _$L^{\infty}$ distance:_ $d(F,G)=d(\lambda_{a},\lambda_{b})=\text{max}_{i}\big{|}\lambda_{a,i}-\lambda_{b,i}\big{|}$. Note that, Minkowski distance for $\kappa=2$ is Euclidean distance which we considered in the algorithm. The other two metrics ($L^{1}$ and $L^{\infty}$) are special cases of the Minkowski distance. While we can potentially evaluate the filter with $L^{1}$ or $L^{\infty}$, given the lack of smoothness in the derivatives, we consider the $L^{2}$ (i.e. Minkowski distance with $\kappa$ = 2) to be the most appropriate metric for ease of computation and exposition. ## 5 Real-life Data Analysis To demonstrate application of our proposed filter, we apply it on a real-life financial data set. Although the filter would be more useful for _big_ data sets with high number of variables, we use a data set with a moderate number of variables for the sake of visualization of the resulting network. Applicability of our method depends on three factors: 1) The data possess a high-dimensional covariance matrix. 2) There is a network representation based on the covariance matrix. 3) There is a _core_ underlying set of connections or _skeleton_ of the network that is captured by large enough covariances. ### 5.1 Application to financial networks We perform our method on historical NASDAQ data for 50 stocks with prices recorded over 70 consecutive days, from 2nd January, 2015 to 14th April, 2015. As is customary in the analysis of return comovement networks [41], we first construct the log return series for each of the stocks. If a stock’s price at time $t$ is $P_{t}$, then the log return at time $t$ is defined as the following: $r_{t}=\text{log}P_{t}-\text{log}P_{t-1}.$ (19) From the generated return series, the sample correlation matrix $S$ is calculated and the corresponding network $\Gamma(S)$ is generated (Fig. 5). As it is a complete graph we can see all possible edges are present (we have excluded the self-loops) as all pairs of stocks would exhibit non-zero covariance. Fig. 5 exhibits the maximally filtered network, which shows a drastic reduction in the number of edges(from 1225 to 256). We also observe that this filtered graph is not connected. In particular, the maximally filtered network produces 8 isolated vertices while the remaining 42 stocks create a giant component.101010In some sets of stocks, we have observed that the spectral distance between the true matrix and the Ledoit-Wolf analog is so large that the maximal filtering leads to fully diagonal matrix, i.e. all nodes become separate. In such cases, the spectral similarity is not an useful criterion for filtering. (a) Financial network (b) Maximally filtered network (c) Spectral distance Figure 5: Illustration for Financial network. Panel (a) shows the network ($\Gamma(S)$) obtained from sample correlation matrix. It is a complete graph with dense connectivity. Panel (b) shows the maximally filtered network with substantially fewer edges. The network splits into a giant connected component and unconnected peripheral nodes. Panel (c) shows the spectral distance function against number of deleted edges. The lowest distance corresponds to the maximal filtered network shown in panel (b). Finally, in Fig. 5 we plot the spectral distance between the Ledoit-Wolf estimator and the filtered matrices in an increasing number of edges being deleted (associated with increasing threshold). The global minimum for the distance function is reached at 1082 (=1225-143) number of edges being deleted. Therefore, the maximally filtered network would consist of 143 edges as shown in Fig. 5. Clearly, such strong filtering is associated with high thresholding and consequent loss of a large number of edges. For a less strong filtering, one can impose a positive cost of edge deletion (maximally filtering requires zero cost of edge deletion) following Eqn. 9. based on the choice of parameters, one can interpolate between zero filtering to maximal filtering. To complement the above analysis, we carry out the filtering on a large covariance matrix arising out of the largest $p$ = 300 stocks in NASDAQ in 2015 calendar year by varying the number of observations $n$ from 50, 200, 300 and 450. The values are chosen such that the $p/n$ ratio varies from a number smaller than one to larger than one. The resulting maximal thresholds are shown in Fig. 6 in the Appendix A.2. As can be seen, for large ratio of $p$ to $n$ (indicating very small number of observations for each stock) the filtering threshold is very high and a high fraction of edges get filtered. In the other extreme, when $p$ to $n$ ratio is small (indicating a large number of observations for each stock), then the filtering threshold is very low and therefore, very few edges are filtered. For the sake of completeness, we should mention that the filtering threshold is influenced by sampling fluctuations and therefore, such a monotonic relationship may not be found in all applications. However, in our numerical experiments we found that on an average a larger number of observations for each entity (stocks in this case) leads to smaller filtering threshold. ## 6 Summary and Conclusion Many large scale systems are best described as networks [3, 4, 5, 12, 20, 33, 31, 41]. A standard approach of network construction is to create covariance- based measures of interlinkages [41]. However, construction of the comovement network from an observed data set is a challenging problem because the resulting network is a complete graph and therefore resists any naive attempt to uncover the underlying network topology due to existence of spurious linkages. Statistically such networks suffer from false positives, i.e. false discovery of linkages. Therefore, a robust methodology is needed to identify and prune such non-informative linkages and isolate the key subnetwork embedded in the complete network. In this paper, we develop a filtering technique that attempts to resolve this problem utilizing spectral structure of the network. The existing filtering techniques have mainly two features. First, they are primarily based on some graph-theoretic constraints and not on explicit statistical motivation (e.g. minimum spanning tree or more general, hierarchical structures). Second, many of the filtering techniques are not tunable and often they lead to a drastic reduction in the number of edges (e.g. minimum spanning tree), which also makes the resulting network very unstable and sample-dependent. In this paper, we propose an new filter based on the properties of high-dimensional covariance estimators, utilizing the concept of sparsistence along with retaining flexibility for tuning the degree of filtering. We note that one can consider algorithms based on hypothesis testing of individual edge weights and prune statistically insignificant edges. However, this kind algorithms still suffer from the problem of false positives (i.e. false discovery of edges) as they do not account for joint hypothesis testing. We approach the problem in a new way by considering the spectral structure of the covariance network and sparsistent analogues of that, based on Ledoit-Wolf estimators which features predominantly in high dimensional covariance matrix estimation. Depending on the statistical properties of the Ledoit-Wolf estimator, we prescribe an endogenously thresholded covariance matrix estimator such that its spectrum is closest to that of the Ledoit-Wolf matrix. We complement the theoretical structure with numerical simulations along with applications to real world financial data. Our work is situated in the intersection of the literature on network filtering, covariance matrix estimation and large dimensional data. The proposed algorithm can be applied to any large dimensional data. We have demonstrated the usefulness of the filtering algorithm by applying it to financial stock return data. Further applications to various domains spanning biological, physical and technological comovement networks would lead to a more complete understanding of the corresponding topological structures and key linkages that contribute to the dynamics of the system. ## References * [1] Karim M Abadir, Walter Distaso, and Filip Žikeš. Design-free estimation of variance matrices. Journal of Econometrics, 181(2):165–180, 2014. * [2] Zhidong Bai and Jack W Silverstein. Spectral analysis of large dimensional random matrices, volume 20. Springer, 2010. * [3] Albert-Laszlo Barabasi and Zoltan N Oltvai. Network biology: understanding the cell’s functional organization. Nature reviews genetics, 5(2):101–113, 2004. * [4] Wolfram Barfuss, Guido Previde Massara, Tiziana Di Matteo, and Tomaso Aste. Parsimonious modeling with information filtering networks. Physical Review E, 94(6):062306, 2016. * [5] Stefano Battiston, J Doyne Farmer, Andreas Flache, Diego Garlaschelli, Andrew G Haldane, Hans Heesterbeek, Cars Hommes, Carlo Jaeger, Robert May, and Marten Scheffer. Complexity theory and financial regulation. Science, 351(6275):818–819, 2016. * [6] Peter J Bickel, Elizaveta Levina, et al. Some theory for fisher’s linear discriminant function,naive bayes’, and some alternatives when there are many more variables than observations. Bernoulli, 10(6):989–1010, 2004. * [7] Peter J Bickel, Elizaveta Levina, et al. Covariance regularization by thresholding. The Annals of Statistics, 36(6):2577–2604, 2008. * [8] Peter J Bickel, Elizaveta Levina, et al. Regularized estimation of large covariance matrices. The Annals of Statistics, 36(1):199–227, 2008. * [9] Spiros Bougheas and Alan Kirman. Complex financial networks and systemic risk: A review. In Complexity and geographical economics, pages 115–139. Springer, 2015. * [10] Tony Cai, Weidong Liu, and Xi Luo. A constrained l-1 minimization approach to sparse precision matrix estimation. Journal of the American Statistical Association, 106(494):594–607, 2011. * [11] Sanjay Chaudhuri, Mathias Drton, and Thomas S Richardson. Estimation of a covariance matrix with zeros. Biometrika, 94(1):199–216, 2007. * [12] Dong-Yeon Cho, Yoo-Ah Kim, and Teresa M Przytycka. Network biology approach to complex diseases. PLoS Comput Biol, 8(12):e1002820, 2012. * [13] Michele Coscia and Frank MH Neffke. Network backboning with noisy data. In 2017 IEEE 33rd International Conference on Data Engineering (ICDE), pages 425–436. IEEE, 2017. * [14] Chandler Davis and William Morton Kahan. The rotation of eigenvectors by a perturbation. iii. SIAM Journal on Numerical Analysis, 7(1):1–46, 1970. * [15] Noureddine El Karoui et al. Spectrum estimation for large dimensional covariance matrices using random matrix theory. The Annals of Statistics, 36(6):2757–2790, 2008. * [16] Jerome Friedman, Trevor Hastie, and Robert Tibshirani. Sparse inverse covariance estimation with the graphical lasso. Biostatistics, 9(3):432–441, 2008. * [17] Reinhard Furrer and Thomas Bengtsson. Estimation of high-dimensional prior and posterior covariance matrices in kalman filter variants. Journal of Multivariate Analysis, 98(2):227–255, 2007. * [18] Michael Hamann, Gerd Lindner, Henning Meyerhenke, Christian L Staudt, and Dorothea Wagner. Structure-preserving sparsification methods for social networks. Social Network Analysis and Mining, 6(1):22, 2016. * [19] Gecia Bravo Hermsdorff and Lee Gunderson. A unifying framework for spectrum-preserving graph sparsification and coarsening. In Advances in Neural Information Processing Systems, pages 7736–7747, 2019. * [20] Bernardo A Huberman and Lada A Adamic. Growth dynamics of the world-wide web. Nature, 401(6749):131–131, 1999. * [21] Martin Imre, Jun Tao, Yongyu Wang, Zhiqiang Zhao, Zhuo Feng, and Chaoli Wang. Spectrum-preserving sparsification for visualization of big graphs. Computers & Graphics, 87:89–102, 2020. * [22] Teruyoshi Kobayashi, Taro Takaguchi, and Alain Barrat. The structured backbone of temporal social ties. Nature communications, 10(1):1–11, 2019. * [23] Clifford Lam et al. Nonparametric eigenvalue-regularized precision or covariance matrix estimator. The Annals of Statistics, 44(3):928–953, 2016. * [24] Olivier Ledoit and Sandrine Péché. Eigenvectors of some large sample covariance matrix ensembles. Probability Theory and Related Fields, 151(1-2):233–264, 2011. * [25] Olivier Ledoit and Michael Wolf. A well-conditioned estimator for large-dimensional covariance matrices. Journal of multivariate analysis, 88(2):365–411, 2004. * [26] Olivier Ledoit, Michael Wolf, et al. Nonlinear shrinkage estimation of large-dimensional covariance matrices. The Annals of Statistics, 40(2):1024–1060, 2012. * [27] Olivier Ledoit, Michael Wolf, et al. Analytical nonlinear shrinkage of large-dimensional covariance matrices. Annals of Statistics, 48(5):3043–3065, 2020. * [28] Rosario N Mantegna. Hierarchical structure in financial markets. The European Physical Journal B-Condensed Matter and Complex Systems, 11(1):193–197, 1999. * [29] Riccardo Marcaccioli and Giacomo Livan. A pólya urn approach to information filtering in complex networks. Nature communications, 10(1):1–10, 2019. * [30] Vladimir A Marčenko and Leonid Andreevich Pastur. Distribution of eigenvalues for some sets of random matrices. Mathematics of the USSR-Sbornik, 1(4):457, 1967. * [31] Gautier Marti, Frank Nielsen, Mikołaj Bińkowski, and Philippe Donnat. A review of two decades of correlations, hierarchies, networks and clustering in financial markets. arXiv preprint arXiv:1703.00485, 2017. * [32] Mark Ed Newman, Albert-László Ed Barabási, and Duncan J Watts. The structure and dynamics of networks. Princeton university press, 2006. * [33] Mark EJ Newman. The structure and function of complex networks. SIAM review, 45(2):167–256, 2003. * [34] Mohsen Pourahmadi. High-dimensional covariance estimation: with high-dimensional data, volume 882. John Wiley & Sons, 2013. * [35] Yumou Qiu and Janaka SS Liyanage. Threshold selection for covariance estimation. Biometrics, 75(3):895–905, 2019. * [36] Filippo Radicchi, José J Ramasco, and Santo Fortunato. Information filtering in complex weighted networks. Physical Review E, 83(4):046101, 2011. * [37] Adam J Rothman, Peter J Bickel, Elizaveta Levina, Ji Zhu, et al. Sparse permutation invariant covariance estimation. Electronic Journal of Statistics, 2:494–515, 2008. * [38] Adam J Rothman, Elizaveta Levina, and Ji Zhu. Generalized thresholding of large covariance matrices. Journal of the American Statistical Association, 104(485):177–186, 2009. * [39] Adam J Rothman, Elizaveta Levina, and Ji Zhu. A new approach to cholesky-based covariance regularization in high dimensions. Biometrika, 97(3):539–550, 2010. * [40] Parongama Sen, Subinay Dasgupta, Arnab Chatterjee, PA Sreeram, G Mukherjee, and SS Manna. Small-world properties of the indian railway network. Physical Review E, 67(3):036106, 2003. * [41] Sitabhra Sinha, Arnab Chatterjee, Anirban Chakraborti, and Bikas K Chakrabarti. Econophysics: an introduction. John Wiley & Sons, 2010. * [42] Charles Stein. Inadmissibility of the usual estimator for the mean of a multivariate normal distribution. Technical report, Stanford University Stanford United States, 1956. * [43] Michele Tumminello, Tomaso Aste, Tiziana Di Matteo, and Rosario N Mantegna. A tool for filtering information in complex systems. Proceedings of the National Academy of Sciences, 102(30):10421–10426, 2005. * [44] Fernando Vega-Redondo. Complex social networks. Number 44. Cambridge University Press, 2007. * [45] Richard J Williams and Neo D Martinez. Simple rules yield complex food webs. Nature, 404(6774):180–183, 2000. ## Appendix A Appendix ### A.1 True and sample covariance matrix for the illustrative example The true correlation matrix for the simulated illustrative example (Sec. 3.2) is the following: $\begin{bmatrix}1.00&0.80&0.00&0.00&0.00&0.30&0.00&0.00&0.00&0.00\\\ 0.80&1.00&0.00&0.00&0.00&0.09&0.00&0.00&0.00&0.00\\\ 0.00&0.00&1.00&0.30&0.09&0.09&0.00&0.09&0.00&0.00\\\ 0.00&0.00&0.03&1.00&0.00&0.00&0.30&0.00&0.00&0.00\\\ 0.00&0.00&0.09&0.00&1.00&0.80&0.00&0.00&0.00&0.00\\\ 0.30&0.09&0.09&0.00&0.80&1.00&0.00&0.00&0.00&0.00\\\ 0.00&0.00&0.00&0.30&0.00&0.00&1.00&0.30&0.80&0.80\\\ 0.00&0.00&0.09&0.00&0.00&0.00&0.30&1.00&0.09&0.30\end{bmatrix}$ After generating the sample, the sample covariance matrix is: $\begin{bmatrix}14.36&-2.12&2.36&-1.44&0.38&-4.01&0.64&2.65&-3.38&-2.13\\\ -2.12&7.02&-0.63&-0.01&-0.80&1.20&0.12&-3.29&1.87&-5.86\\\ 2.36&-0.63&11.12&5.55&-0.06&-1.31&5.04&4.63&-3.83&2.14\\\ -1.44&-0.01&5.55&9.73&-3.03&-1.69&7.45&1.15&-4.97&1.85\\\ 0.38&-0.80&-0.06&-3.03&6.76&3.57&-2.23&-3.82&-2.84&-0.96\\\ -4.01&1.20&-1.31&-1.69&3.57&5.99&-4.84&-2.91&-2.08&-1.12\\\ 0.64&0.12&5.04&7.45&-2.23&-4.84&13.54&-1.80&-3.02&0.33\\\ 2.65&-3.29&4.63&1.15&-3.82&-2.91&-1.80&9.56&1.38&4.99\end{bmatrix}$ We see that many entries of the true correlation matrix is 0 and therefore, the corresponding covariance matrix would be a sparse matrix. However, it is noteworthy that the sample covariance matrix does not contain any 0 due to sampling fluctuations. So the resulting network representation will be a fully connected network although the underlying network is sparsely connected. ### A.2 Application on large dimensional financial covariance matrix Figure 6: Application of the sparsistent filter on large dimensional financial network. We have considered the covariance matrices of $p=300$ largest stocks from NASDAQ stock exchange (in terms of market capitalization) with varying number of days. Each data set begins on 2nd January, 2015 and continues for $n$ days where $n$ varies from 50, 200, 300 and 450. Therefore, in Panel (a) $p/n$ = 2/3, Panel (b) $p/n$ = 1, Panel (c) $p/n$ = 3/2 and finally, Panel (d) $p/n$ = 6. A larger $p/n$ ratio leads to larger threshold for maximal filtering.
∎ aFEMclipboard 11institutetext: A. Chambolle 22institutetext: CEREMADE, CNRS & Université Paris Dauphine, PSL Research University, Paris 22email<EMAIL_ADDRESS>ORCID: 0000-0002-9465-4659 33institutetext: R. Tovey 44institutetext: MOKAPLAN, INRIA Paris, Paris 44email<EMAIL_ADDRESS>ORCID: 0000-0001-5411-2268 # “FISTA” in Banach spaces with adaptive discretisations Antonin Chambolle Robert Tovey ###### Abstract FISTA is a popular convex optimisation algorithm which is known to converge at an optimal rate whenever a minimiser is contained in a suitable Hilbert space. We propose a modified algorithm where each iteration is performed in a subset which is allowed to change at every iteration. Sufficient conditions are provided for guaranteed convergence, although at a reduced rate depending on the conditioning of the specific problem. These conditions have a natural interpretation when a minimiser exists in an underlying Banach space. Typical examples are L1-penalised reconstructions where we provide detailed theoretical and numerical analysis. ###### Keywords: Convex optimization Multiscale Multigrid Sparsity Lasso ††journal: Computational Optimization and Applications ## 1 Introduction The Fast Iterative Shrinkage-Thresholding Algorithm (FISTA) was proposed by Beck and Teboulle Beck2009 as an extension of Nesterov’s fast gradient method Nesterov2004 and is now a very popular algorithm for minimising the sum of two convex functions. We write this as the problem of computing $\inf_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])\qquad\text{such that}\qquad\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])\coloneqq\operatorname{f}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])+\operatorname{g}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]),$ (1) for a Hilbert space $\mathds{H}$ where $\operatorname{f}\colon\mathds{H}\to\mathds{R}$ is a convex differentiable function with $L$-Lipschitz gradient and $\operatorname{g}\colon\mathds{H}\to\overline{\mathds{R}}$ is a “simple” convex function, whose “proximity operator” is easy to compute. Throughout this work we assume that $\operatorname{E}$ is bounded below so that the infimum is finite. The iterates of the FISTA algorithm will be denoted $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}\in\mathds{H}$. If, moreover the infimum is achieved, it has been shown that $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\inf_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$ converges at the optimum rate of $n^{-2}$ Beck2009 , and later (after a small modification) the convergence of the iterates was also shown in a general Hilbert space setting Chambolle2015 . Many further works have gone on to demonstrate faster practical convergence rates for slightly modified variants of FISTA Tao2016 ; Liang2017 ; Alamo2019 . In this work we address the case where the minimiser possibly fails to exist or lies in a larger space where $\mathds{H}$ is dense. There is much overlap between the techniques used in this work and those used in the literature of inexact optimisation, however, our interpretation is relatively novel. In particular, we emphasise the infinite-dimensional setting where errors come from “discretisation”, rather than random or decaying errors in $\mathds{H}$, which enables two new perspectives: * • Analytically, we prove new rates of convergence for FISTA when the minimum energy is not achieved (at least not in $\mathds{H}$). The exact rate can be computed by quantifying coercivity and regularity properties of $\operatorname{E}$. If there isn’t a minimiser in $\mathds{H}$, then this rate is strictly slower than $n^{-2}$. * • Numerically, we allow the optimisation domain to change on every iteration. This enables us to understand how FISTA behaves with adaptive discretisations. Adaptive finite-element methods are known to improve the efficiency of, for example, approximating the solutions of PDEs. Our analytical results show how to combine such tools with FISTA without reducing the guaranteed rate of convergence, and our numerical results confirm much improved time and computer memory efficiency in the Lasso example (Section 6). All the examples in this work, discussed from Section 5 onward, consider $\\{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}\operatorname{\;s.t.\;}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])<\infty\\}$ to be contained in some ambient Banach space $\mathds{U}$. The idea is that FISTA provides a minimising sequence in $\mathds{H}\cap\mathds{U}$, but further properties like rate of convergence (of $\operatorname{E}$ or the iterates) must come from the topology of $\mathds{U}$. It will not be necessary for $\mathds{H}\hookrightarrow\mathds{U}$ to be a continuous embedding, nor in fact the full inclusion $\mathds{H}\subset\mathds{U}$. Some other works for FISTA-like algorithms include Jiang2012 ; Villa2013 . Of particular note, our stability estimate for FISTA in Theorem 4.1 is very similar to (Schmidt2011, , Prop 2) and (Aujol2015, , Prop 3.3). This is then used to analyse the convergence properties in our more general Banach space setting, but where all sources of inexactness come from subspace approximations. The ideas in Parpas2017 are similar although in application to the proximal gradient method with an additional smoothing on the functional $\operatorname{g}$. The permitted refinement steps are also more broad in our work. Very recent work in Yu2021 proposes a “Multilevel FISTA” algorithm which allows similar coarse-to-fine refinement strategies, although only a finite number. We also allow for non-uniform refinement with a posteriori strategies. ### 1.1 Outline This work is organised as follows. Section 2 defines notation and the generic form of our proposed refining FISTA algorithm, Algorithm 1. The main theoretical contribution of this work is the convergence analysis of Algorithm 1 which is split into two parts: first we outline the proof structure in Section 3, then we state the specific results in the case of FISTA in Section 4. The main results are Theorems 4.2/4.3 which extend the convergence of FISTA to cases with un-attained minima with uniform/adaptively chosen subspaces $\mathds{U}^{n}$ respectively. Section 5 presents some general results for the application of Algorithm 1 in Banach spaces and Section 6 gives a much more detailed discussion of adaptive refinement for Lasso minimisation. In particular, we describe how to choose efficient refining discretisations to approximate $\inf_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$, estimate the convergence of $\operatorname{E}$, and identify the support of the minimiser. The numerical results in Section 7 demonstrate these techniques in four different models demonstrating the apparent sharpness of our convergence rates and the computational efficiency of adaptive discretisations. ## 2 Definitions and notation We consider optimisation of (1) over a Hilbert space $(\mathds{H},\left\langle\cdot,\cdot\right\rangle,{\left\lVert\cdot\right\rVert})$. In the more analytical section (Sections 3 and 4) it will be more convenient to use the translated energy $\operatorname{E}_{0}\colon\mathds{H}\to\mathds{R},\qquad\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])\coloneqq\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])-\inf_{\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\in\mathds{H}}\operatorname{E}(\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]})$ (2) so that $\inf_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])=0$, although access to this function is not assumed for numerical examples. The proposed generalised FISTA algorithm is stated in Algorithm 1 for an arbitrary choice of closed convex subsets $\mathds{U}^{n}\subset\mathds{H}$ for $n\in\mathds{N}$. The only difference from standard FISTA is that on iteration $n$, all computations are performed in the subset $\mathds{U}^{n}$. If $\mathds{U}^{n}=\mathds{H}$, then we recover the original algorithm. More generally, the idea is that $\mathds{U}^{n}$ are “growing”, for example $\mathds{U}^{n}\subset\mathds{U}^{n+1}$, but this assumption is not necessary in most of the results. Without loss of generality we will assume $L=1$, i.e. $\nabla\operatorname{f}$ is 1-Lipschitz. To get the general statement of any of the results which follow, replace $\operatorname{E}$ with $\frac{\operatorname{E}}{L}$. In particular, ${\left\lVert\nabla\operatorname{f}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])-\nabla\operatorname{f}(\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1])\right\rVert}\leq{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]\right\rVert}$ (3) for all $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0],\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]\in\mathds{H}$ and $\operatorname{g}$ is called “simple” if it is proper, convex, weakly lower-semicontinuous, and $\operatorname*{argmin}_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\widetilde{\mathds{U}}}\tfrac{1}{2}{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]\right\rVert}^{2}+\operatorname{g}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$ (4) is exactly computable for all $\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]\in\mathds{H}$ and all $\widetilde{\mathds{U}}\in\\{\mathds{U}^{n}\\}_{n=0}^{\infty}$. Closed subsets of $\mathds{H}$ are locally weakly compact, therefore this argmin is always non-empty. One defining property of the FISTA algorithm is an appropriate choice of inertia, dictated by $t_{n}$. In particular, we will say that $(t_{n})_{n=0}^{\infty}$ is a _FISTA stepsize_ if $t_{0}=1,\qquad t_{n}\geq 1,\qquad\text{and}\qquad\rho_{n}\coloneqq t_{n}^{2}-t_{n+1}^{2}+t_{n+1}\geq 0\qquad\text{ for all }n=0,1,\ldots.$ (5) The precise constants associated to a given rate are given in the statements of the theorems but, for convenience, are otherwise omitted from the text. For sequences $(a_{n})_{n=0}^{\infty}$,$(b_{n})_{n=0}^{\infty}$ we will use the notation: $\displaystyle a_{n}\lesssim b_{n}\qquad$ $\displaystyle\iff\qquad\exists C,N>0\operatorname{\;s.t.\;}a_{n}\leq Cb_{n}\text{ for all }n>N,$ $\displaystyle a_{n}\simeq b_{n}\qquad$ $\displaystyle\iff\qquad a_{n}\lesssim b_{n}\lesssim a_{n}.$ For $n\in\mathds{N}$ we use the abbreviation $[n]=\\{1,2,\ldots,n\\}$. When the subdifferential of $\operatorname{E}$ is set-valued, we will use the short-hand ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\partial\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}\coloneqq\inf_{\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]\in\partial\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])}{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ (6) for any specified norm ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$. Algorithm 1 Refining subset FISTA 1:Choose $(\mathds{U}^{n})_{n\in\mathds{N}}$, $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{0}\in\mathds{U}^{0}$ and some FISTA stepsize choice $(t_{n})_{n\in\mathds{N}}$ 2:$\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{0}\leftarrow\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{0},n\leftarrow 0$ 3:repeat 4: $\overline{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}_{n}\leftarrow(1-\tfrac{1}{t_{n}})\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}+\tfrac{1}{t_{n}}\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}$ 5: $\displaystyle\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n+1}\leftarrow\operatorname*{argmin}_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}^{n+1}}\tfrac{1}{2}{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\overline{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}_{n}+\nabla\operatorname{f}(\overline{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}_{n})\right\rVert}^{2}+\operatorname{g}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$ $\triangleright$ Only modification, $\mathds{U}^{n+1}\subset\mathds{U}$ 6: $\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n+1}\leftarrow(1-t_{n})\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}+t_{n}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n+1}$ 7: $n\leftarrow n+1$ 8:until some stopping criterion is met ## 3 General proof recipe In this section we give an intuitive outline of the full proof for convergence of Algorithm 1 before giving formal theorems and proofs in the next section. First we recall the classical FISTA convergence guarantee given by (Chambolle2015, , Thm 3.1); if there exists $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\in\operatorname*{argmin}_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$, then $t_{N}^{2}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})+\sum^{N-1}_{n=1}\rho_{n}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\tfrac{1}{2}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{N}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right\rVert}^{2}\leq\tfrac{1}{2}{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{0}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right\rVert}^{2}$ (7) for any FISTA stepsize choice $t_{N}\simeq N$ such that $\rho_{n}\geq 0$. ##### Step 1: Quantifying the stability The first step is to generalise (7) to account for the adapting subsets $\mathds{U}^{n}$. In the notation of Algorithm 1, Theorem 4.1 shows that $t_{N}^{2}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})+\sum_{n=1}^{N-1}\rho_{n}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\tfrac{1}{2}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{N}-\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{N}\right\rVert}^{2}\leq\tfrac{1}{2}{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{0}-\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{0}\right\rVert}^{2}+\tfrac{{\left\lVert\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{N}\right\rVert}^{2}-{\left\lVert\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{0}\right\rVert}^{2}}{2}+\sum^{N}_{n=1}t_{n}\operatorname{E}_{0}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n})+\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1},\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n-1}-\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rangle$ (8) for any $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\in\mathds{U}^{n}$. The similarities to (7) are clear. If $\mathds{U}^{n}=\mathds{H}$, then we can choose $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}=\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$ and the two estimates agree. These extra terms in (8) quantify the robustness to changing of discretisation. ##### Step 2: Quantifying the scaling properties To show that the extra terms in (8) are small, we need to quantify the approximation properties of $\mathds{U}^{n}$. The idea is that there is a sequence $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\in\mathds{U}^{n}$, $n\in\mathds{N}$ such that ${\left\lVert\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rVert}$ grows slowly and $\operatorname{E}_{0}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n})$ decreases quickly. To quantify this balance, we introduce a secondary sequence $n_{0}<n_{1}<\ldots$ and constants $a_{\operatorname{U}},a_{\operatorname{E}}\geq 1$ such that for each $k\in\mathds{N}$ $n\leq n_{k}\implies{\left\lVert\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rVert}\lesssim a_{\operatorname{U}}^{k},\qquad n\geq n_{k}\implies\operatorname{E}_{0}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n})\lesssim a_{\operatorname{E}}^{-k}.$ (9) A canonical example would be $\mathds{U}^{n}=\\{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}\operatorname{\;s.t.\;}{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}\leq a_{\operatorname{U}}^{k}\\}$ for $n\in[n_{k},n_{k+1})$, then $a_{\operatorname{E}}$ reflects the smoothness of $\operatorname{E}_{0}$. The choice of exponential scaling is introduced to improve stability of Algorithm 1. It is natural if we consider the $\mathds{U}^{n}$ to be the subspace of functions discretised on a uniform mesh. If that mesh is sequentially refined, then the resolution of the mesh will be of order $h^{k}$ after $k$ refinements and for some $h<1$. The integer $n_{k}$ is then the time at which the mesh has refined $k$ times. The trade-off between $a_{\operatorname{E}}$ and $a_{\operatorname{U}}$ dictates the final convergence rate of the algorithm. If $a_{\operatorname{U}}>1$, then we cannot guarantee the original $n^{-2}$ rate of convergence. ##### Step 3: Generalising the convergence bound In this step we combine the FISTA stability estimate with the subset approximation guarantees to provide a sharper estimate of stability with respect to the parameters $a_{\operatorname{E}}$ and $a_{\operatorname{U}}$. For example, if for each $k\in\mathds{N}$ $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}=\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n_{k}}\text{ for each }n=n_{k},n_{k}+1,\ldots,n_{k+1}-1,$ then many terms on the right-hand side of (8) telescope to 0. The result of this is presented in Lemma 3. The key idea is that the stability error in (8) has $K\ll N$ terms, rather than $N$. ##### Step 4: Sufficiently fast growth In Step 3 we develop a convergence bound, now we wish to show that it is only worse than the classical (7) by a constant factor. In particular, it is equivalent to either run Algorithm 1 for $N$ iterations, or the classical FISTA algorithm for $N$ iterations on the fixed subset $\mathds{U}^{N}$. The estimate from (7) provides the estimate $N^{2}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})\lesssim{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{0}-\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{N}\right\rVert}^{2}=O(a_{\operatorname{U}}^{2K})$ for $N\leq n_{K}$. Lemma 4 shows that Algorithm 1 can achieve the same order of approximation, so long as $\mathds{U}^{n}$ grow sufficiently quickly (in particular $n_{k}^{2}\lesssim a_{\operatorname{E}}^{k}a_{\operatorname{U}}^{2k}$). ##### Step 5: Sufficiently slow growth The result of Step 3 is sufficient to prove convergence, but not yet a rate. If the subsets grow too quickly, then the influence of ${\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}\right\rVert}\to\infty$ will slow the rate of convergence. If $n_{k}$ is too large, then we overfit to the discrete problem, but if $n_{k}$ is too small, then FISTA converges slowly. Lemma 5 balances these two factors in an optimal way ($n_{k}^{2}\simeq a_{\operatorname{E}}^{k}a_{\operatorname{U}}^{2k}$) for Algorithm 1 resulting in a convergence rate of $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})\lesssim\frac{a_{\operatorname{U}}^{2K}}{N^{2}}\lesssim\frac{N^{2\kappa}}{N^{2}}$ for all $N\in\mathds{N}$ and $\kappa=\frac{2\log a_{\operatorname{U}}}{\log a_{\operatorname{E}}+2\log a_{\operatorname{U}}}\in[0,1)$. In particular, if the minimum is attained in $\mathds{H}$, then we recover the classical rate with $\kappa=0$. ##### Step 6: Adaptivity Up to this point we have implicitly focused on the case where $\mathds{U}^{n}$ (and $n_{k}$) are chosen a priori. The main challenge for adaptive choice of $\mathds{U}^{n}$ is to guarantee (9) from Step 3 using a posteriori estimates. Combined with the partial telescoping requirement in Step 3, a natural choice is $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}=\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n_{k}-1}$ for $n\in[n_{k},n_{k+1})$, i.e. the value of $n_{k}$ is chosen to be $n+1$ once the iterate $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}$ is observed. Theorem 4.3 shows that a sufficient condition is $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n_{k}-1}\in\mathds{U}^{n_{k}}\cap\mathds{U}^{n_{k}+1}\cap\ldots\cap\mathds{U}^{n_{k+1}-1},\qquad{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n_{k}-1}\right\rVert}\lesssim a_{\operatorname{U}}^{k},\quad\text{and}\quad\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n_{k}-1})\lesssim a_{\operatorname{E}}^{-k}.$ Convergence is most stable if the approximation spaces $\mathds{U}^{n}$ satisfy a monotone inclusion, breaking the monotonicity requires more care. The only non-trivial property to verify is the energy gap $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})=\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\inf_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$. Lemma 6 proposes some sufficient conditions to guarantee the same overall rate of convergence as in Step 3, $\min_{n\leq N}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\lesssim\frac{N^{2\kappa}}{N^{2}}$ for all $N\in\mathds{N}$, with the same $\kappa\in[0,1)$ from Step 3. The penalty for accelerating the change of discretisation is a potential loss of stability or monotonicity in $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})$, although this behaviour has not been seen in numerical experiments. ## 4 Proof of convergence In this section we follow the recipe motivated in Section 3 to prove convergence of two variants of Algorithm 1. Each of the main theorems and lemmas will be stated with a sketch proof in this section. The details of the proofs are either trivial or very technical and are therefore placed in Section A to preserve the flow of the argument. ### 4.1 Computing the convergence bound For Step 3 of Section 3 we look to replicate the classical bound of the form in (7) for Algorithm 1. The proofs in this step follow the classical arguments Beck2009 ; Chambolle2015 very closely. Throughout this section we consider a sequence $(\mathds{U}^{n})_{n\in\mathds{N}}$ which generate the iterates $(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})_{n\in\mathds{N}}$ in Algorithm 1 such that $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{n}\in\mathds{U}^{n+1}\subset\mathds{H}\quad\text{where }\mathds{U}^{n}\text{ is a closed, convex subset for all }n\in\mathds{N}.$ (10) #### 4.1.1 Single iterations We first wish to understand a single iteration of Algorithm 1. This is done through the following two lemmas. ###### Lemma 1 (equivalent to (Chambolle2015, , Lemma 3.1)) thm: descent lemma Suppose $\nabla\operatorname{f}$ is 1-Lipschitz, for any $\overline{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\in\mathds{H}$ define $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\coloneqq\operatorname*{argmin}_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}^{n}}\tfrac{1}{2}{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\overline{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}+\nabla\operatorname{f}(\overline{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]})\right\rVert}^{2}+\operatorname{g}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]).$ Then, for all $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]\in\mathds{U}^{n}$, we have $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])+\tfrac{1}{2}{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]\right\rVert}^{2}\leq\operatorname{E}_{0}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2])+\tfrac{1}{2}{\left\lVert\overline{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}-\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]\right\rVert}^{2}.$ The proof is exactly the same as in Chambolle2015 on the subset $\mathds{U}^{n}$. Applying Lemma 1 to the iterates from Algorithm 1 gives a more explicit inequality. ###### Lemma 2 ((Chambolle2015, , (17)), (Beck2009, , Lemma 4.1)) thm: one step FISTA Let $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\in\mathds{U}^{n}$ be chosen arbitrarily and $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}$/$\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}$ be generated by Algorithm 1 for all $n\in\mathds{N}$. For all $n>0$, it holds that $\Copy{thm:eq:onestepFISTA}{t_{n}^{2}(\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\operatorname{E}_{0}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}))-(t_{n}^{2}-t_{n})(\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n-1})-\operatorname{E}_{0}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}))\leq\tfrac{1}{2}\left[{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1}\right\rVert}^{2}-{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}\right\rVert}^{2}\right]+\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}-\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1},\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rangle.}$ (11) The proof is given in Theorem A.1 and is a result of the convexity of $\operatorname{E}_{0}$ and $\mathds{U}_{n}$ for a well chosen $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]$ in Lemma 1. #### 4.1.2 Generic convergence bound Lemma 2 gives us an understanding of a single iteration of Algorithm 1, summing over $n$ then gives our generic convergence bound for any variant of Algorithm 1. ###### Theorem 4.1 (analogous to (Chambolle2015, , Thm 3.2), (Beck2009, , Thm 4.1)) thm: mini FISTA convergence Fix a sequence of subsets $(\mathds{U}^{n})_{n\in\mathds{N}}$ satisfying (10), arbitrary $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{0}\in\mathds{U}^{0}$, and FISTA stepsize choice $(t_{n})_{n\in\mathds{N}}$. Let $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}$ and $\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}$ be generated by Algorithm 1, then, for any choice of $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\in\mathds{U}^{n}$ and $N\in\mathds{N}$ we have $\Copy{thm:eq:miniFISTAconvergence}{t_{N}^{2}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})+\sum_{n=1}^{N-1}\rho_{n}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\frac{{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{N}-\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{N}\right\rVert}^{2}}{2}\leq\frac{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{0}-\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{0}\right\rVert}^{2}-{\left\lVert\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{0}\right\rVert}^{2}+{\left\lVert\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{N}\right\rVert}^{2}}{2}+\sum^{N}_{n=1}t_{n}\operatorname{E}_{0}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n})+\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1},\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n-1}-\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rangle.}$ (12) The proof is given in Theorem A.2. This result is the key approximation for showing convergence of FISTA with changing subsets. In the classical setting, we have $\mathds{U}^{n}=\mathds{H}$, $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}=\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{0}\in\operatorname*{argmin}_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$ and the extra terms on the right-hand side collapse to 0. If there exists a minimiser $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\in\operatorname*{argmin}_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$, then the natural choice in (12) is $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}=\mathsf{\Pi}_{n}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$ for some projection $\mathsf{\Pi}_{n}\colon\mathds{H}\to\mathds{U}^{n}$, however, there are simple counter-examples which give $\operatorname{E}_{0}(\mathsf{\Pi}_{n}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})=\infty$ and so this inequality becomes useless. For example, if $\operatorname{f}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])={\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{L^{2}([0,1])}^{2}$, $\operatorname{g}$ is the indicator on the set $\mathds{D}=\\{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in L^{1}([0,1])\operatorname{\;s.t.\;}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0](x)\geq x\\}$, and $\mathsf{\Pi}_{n}$ is the $L^{2}$ projection onto a set of piecewise constant functions, then $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}=x\mapsto x$. On the other hand, suppose one of the pixels of the discretisation is $[x_{0}-h,x_{0}+h]$, then $\mathsf{\Pi}_{n}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\left(x_{0}+\tfrac{h}{2}\right)=\operatorname*{argmin}_{\IfEqCase{5}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k5]\in\mathds{R}}\int_{x_{0}-h}^{x_{0}+h}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}(x)-\IfEqCase{5}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k5])^{2}\mathop{}\\!\mathrm{d}x=\operatorname*{argmin}_{\IfEqCase{5}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k5]\in\mathds{R}}\int_{x_{0}-h}^{x_{0}+h}(x-\IfEqCase{5}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k5])^{2}\mathop{}\\!\mathrm{d}x=x_{0}<x_{0}+\tfrac{h}{2}.$ In particular $\mathsf{\Pi}_{n}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\notin\mathds{D}$ therefore $\operatorname{E}_{0}(\mathsf{\Pi}_{n}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})=\infty$. The choice $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}=\operatorname*{argmin}_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}^{n}}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$ is much more robust and allows us to apply Algorithm 1 more broadly. The penalty for this flexibility is a more complicated analysis; each time the subset changes, because $\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}\in\mathds{U}_{n}$, the system receives a “shock” proportional to ${\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}\right\rVert}{\left\lVert\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}-\mathsf{\Pi}_{n}\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n+1}\right\rVert}$. ### 4.2 Convergence bound with milestones In standard FISTA, the right-hand side of (12) is a constant. The following lemma minimises the growth of the “constant” as a function of $N$ by partially telescoping the sum on the right-hand side. Before progressing to the content of Step 3, we will first formalise the definition of the constants $a_{\operatorname{U}}$ and $a_{\operatorname{E}}$ introduced in Step 3. ###### Definition 1 Fix $a_{\operatorname{U}},a_{\operatorname{E}}\geq 1$ and a sequence $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\in\mathds{H}$. We say that $(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})_{k\in\mathds{N}}$ is an _$(a_{\operatorname{U}},a_{\operatorname{E}})$ -minimising sequence of $\operatorname{E}$ _ if ${\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right\rVert}\lesssim a_{\operatorname{U}}^{k}\qquad\text{and}\qquad\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})\lesssim a_{\operatorname{E}}^{-k}$ for all $k\in\mathds{N}$. In this section we will simply assume that such sequences exist and in Section 5 we will give some more general examples. ###### Lemma 3 thm: mini exponential FISTA convergence Let $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}$, $\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}$ be generated by Algorithm 1 with $(\mathds{U}^{n})_{n\in\mathds{N}}$ satisfying (10), $(n_{k}\in\mathds{N})_{k\in\mathds{N}}$ be a monotone increasing sequence, and choose $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\in\mathds{U}^{n_{k}}\cap\mathds{U}^{n_{k}+1}\cap\ldots\cap\mathds{U}^{n_{k+1}-1}$ for each $k\in\mathds{N}$. If such a sequence exists, then for all $K\in\mathds{N}$, $n_{K}\leq N<n_{K+1}$ we have $\Copy{thm:eq:miniexponentialFISTAconvergence}{t_{N}^{2}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})+\sum_{n=1}^{N-1}\rho_{n}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\frac{{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{N}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K}\right\rVert}^{2}}{2}\leq C+\frac{{\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K}\right\rVert}^{2}}{2}+\frac{(N+1)^{2}-n_{K}^{2}}{2}\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K})\\\ +\sum_{k=1}^{K}\frac{n_{k}^{2}-n_{k-1}^{2}}{2}\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1})+\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{k}-1},\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right\rangle}$ thm:end: mini exponential FISTA convergence where $C=\frac{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{0}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{0}\right\rVert}^{2}-{\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{0}\right\rVert}^{2}}{2}$. The proof is given in Lemma 11. The introduction of $n_{k}$ has greatly compressed the expression of Theorem 4.1. On the right-hand side, we now only consider $\operatorname{E}_{0}$ evaluated on the sequence $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}$ and there are $K$ elements to the sum rather than $N$. ### 4.3 Refinement without overfitting The aim of Step 3 is to show that $n$ iterations of Algorithm 1 is no slower (up to a constant factor) than $n$ iterations of classical FISTA on the space $\mathds{U}^{n}$. In other words, we would like to ensure that $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})=\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\min_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}^{n}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])\lesssim\frac{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{0}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right\rVert}^{2}}{n^{2}}$ (13) uniformly for $n\in[n_{k},n_{k+1})$. If this condition is not satisfied, then it indicates that computational effort has been wasted by a poor choice of subsets. This can be interpreted as an overfitting to the discretisation of $\operatorname{E}_{0}|_{\mathds{U}^{n}}$ rather than the desired function $\operatorname{E}_{0}|_{\mathds{H}}$. Combining the assumptions given by Definition 1 and the result of Lemma 3, the following lemma proves the convergence of Algorithm 1 provided that the refinement times $n_{k}$ are sufficiently small (i.e. $\mathds{U}^{n}$ grows sufficiently quickly). ###### Lemma 4 thm: sufficiently fast Suppose $\mathds{U}^{n},\ \IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n},\ \IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}$ and $n_{k}$ satisfy the conditions of Lemma 3 and $(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})_{k\in\mathds{N}}$ forms an $(a_{\operatorname{U}},a_{\operatorname{E}})$-minimising sequence of $\operatorname{E}$ with $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\in\mathds{U}^{n_{k}}\cap\mathds{U}^{n_{k}+1}\cap\ldots\cap\mathds{U}^{n_{k+1}-1}.$ If either: * • $a_{\operatorname{U}}>1$ and $n_{k}^{2}\lesssim a_{\operatorname{E}}^{k}a_{\operatorname{U}}^{2k}$, * • or $a_{\operatorname{U}}=1$, $\sum_{k=1}^{\infty}n_{k}^{2}a_{\operatorname{E}}^{-k}<\infty$, and $\sum_{k=1}^{\infty}{\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k+1}\right\rVert}<\infty,$ then $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})\lesssim\frac{a_{\operatorname{U}}^{2K}}{N^{2}}\qquad\text{for all}\qquad n_{K}\leq N<n_{K+1}.$ The proof is given in Lemma 12. We make two observations of the optimality of Lemma 4: * • The convergence guarantee for $N\in[n_{K},n_{K+1})$ iterations of classical FISTA in the space $\mathds{U}^{N}$ is $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})\lesssim\frac{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{0}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K}\right\rVert}^{2}}{N^{2}}+\min_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}^{N}}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])\lesssim\frac{a_{\operatorname{U}}^{2K}}{N^{2}}+a_{\operatorname{E}}^{-K}.$ This is equivalent to Lemma 4 after the assumptions on $n_{k}$. * • If $\mathds{H}$ is finite dimensional, then the condition $a_{\operatorname{U}}=1$ is almost trivially satisfied. Norms in finite dimensions are equivalent and any discretisation can be achieved with a finite number of refinements (i.e. the sums over $k$ are finite). ### 4.4 Convergence rate In Lemma 4 we show that $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})$ converges at a rate depending on $k$ and $n$, so long as $k$ grows sufficiently quickly. On the other hand, as $k$ grows, the rate becomes worse and so we need to also put a lower limit on the growth of $n_{k}$. The following lemma completes Step 3 by computing the global convergence rate of $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})$ when $k$ grows at the minimum rate which is consistent with Lemma 4. As a special case, note that if $a_{\operatorname{U}}=1$ then Lemma 4 already gives the optimal $O(N^{-2})$ convergence rate. This is in fact a special case of that shown in (Aujol2015, , Prop 3.3). If the minimum is achieved in $\mathds{H}$, then it is not possible to refine “too quickly” and the following lemma is not needed. ###### Lemma 5 thm: sufficiently slow Suppose $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}$ and $n_{k}$ are sequences satisfying $\forall N\in[n_{K},n_{K+1}),\ \operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})\lesssim\frac{a_{\operatorname{U}}^{2K}}{N^{2}}\qquad\text{where}\qquad n_{K}^{2}\gtrsim a_{\operatorname{E}}^{K}a_{\operatorname{U}}^{2K},$ then $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})\lesssim\frac{1}{N^{2(1-\kappa)}}\qquad\text{ where }\qquad\kappa=\frac{\log a_{\operatorname{U}}^{2}}{\log a_{\operatorname{E}}+\log a_{\operatorname{U}}^{2}}.$ The proof is given in Lemma 13. #### 4.4.1 FISTA convergence with a priori discretisation We can summarise Lemmas 3 to 5 into a single theorem stating the convergence guarantees when $\mathds{U}^{n}$ and $n_{k}$ are chosen a priori. ###### Theorem 4.2 thm: exponential FISTA convergence Let $(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})_{k\in\mathds{N}}$ be an $(a_{\operatorname{U}},a_{\operatorname{E}})$-minimising sequence of $\operatorname{E}$ and choose any $\mathds{U}^{n}$ satisfying (10) such that $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\in\mathds{U}^{n_{k}}\cap\mathds{U}^{n_{k}+1}\cap\ldots\cap\mathds{U}^{n_{k+1}-1}$ for all $k\in\mathds{N}$. Compute $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}$ and $\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}$ by Algorithm 1. Suppose that either: * • $a_{\operatorname{U}}>1$ and $n_{k}^{2}\simeq a_{\operatorname{E}}^{k}a_{\operatorname{U}}^{2k}$, or * • $a_{\operatorname{U}}=1$, $\sum_{k=1}^{\infty}n_{k}^{2}a_{\operatorname{E}}^{-k}<\infty$ and $\sum_{k=1}^{\infty}{\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k+1}\right\rVert}<\infty$, then $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})\lesssim\frac{1}{N^{2(1-\kappa)}}\qquad\text{ where }\qquad\kappa=\frac{\log a_{\operatorname{U}}^{2}}{\log a_{\operatorname{E}}+\log a_{\operatorname{U}}^{2}}\qquad\text{uniformly for }N\in\mathds{N}.$ Analytically, this theorem gives new rates of convergence for FISTA when the minimiser is not achieved in $\mathds{H}$. Indeed for the original algorithm ($\mathds{U}^{n}=\mathds{H}$), if $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{0}=0$ for simplicity and $(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})_{k\in\mathds{N}}$ is _any_ $(a_{\operatorname{U}},a_{\operatorname{E}})$-minimising sequence of $\operatorname{E}$ exists, the result of Lemma 3 is $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})\leq\inf_{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]\in\mathds{H}}\frac{{\left\lVert\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]\right\rVert}^{2}+N^{2}\operatorname{E}_{0}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2])}{2t_{N}^{2}}\leq\min_{k\in\mathds{N}}\frac{{\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right\rVert}^{2}+N^{2}\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})}{2t_{N}^{2}}\lesssim\min_{k\in\mathds{N}}\frac{a_{\operatorname{U}}^{2k}+N^{2}a_{\operatorname{E}}^{-k}}{N^{2}}\lesssim N^{-2(1-\kappa)}.$ (14) In this sense, we could say that $\operatorname{E}_{0}$ converges at the rate $N^{-2(1-\kappa)}$ if and only if such a sequence exists. Nothing is lost (or gained) analytically by choosing $\mathds{U}_{n}\subsetneq\mathds{H}$. Numerically, it is easy to implement the strategy of Theorem 4.2 and requires very little knowledge of how to estimate $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})$. So long as $a_{\operatorname{U}}$ and $a_{\operatorname{E}}$ can be computed analytically, one can choose $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}$ implicitly to be the discrete minimisers of some “uniform” discretisations (e.g. $\mathds{U}^{n}=\\{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}\leq k\\}$ or finite element spaces with uniform mesh) to achieve the stated convergence rate. #### 4.4.2 FISTA convergence with adaptivity There are two properties of the sequence $(\mathds{U}^{n})_{n\in\mathds{N}}$ which we may wish to decide adaptively: the refinement times $n_{k}$ and the discretising spaces $\\{\mathds{U}^{n}\operatorname{\;s.t.\;}n_{k}\leq n<n_{k+1}\\}$. We will refer to these as temporal and spatial adaptivity respectively. Lemma 4 gives a sufficient condition on $n_{k}$ for converging at the rate $O(N^{2(\kappa-1)})$, but it is not necessary. Indeed for $n\leq n_{k}$ we have $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\geq\min_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}^{n}}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])=O(a_{\operatorname{E}}^{-k})=O(n^{2(\kappa-1)}),$ which suggests that to converge faster than $n^{2(\kappa-1)}$ requires choosing smaller $n_{k}$. As an example, in Section 7.2 we will see Algorithm 1 can converge at a near-linear rate, although this is not possible without adaptive refinement times. On the other hand, choice of spatial adaptivity has no impact on rate but can impact computational efficiency. It will be permitted to use greedy discretisation techniques so long as it is sufficient to estimate $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})$ accurately. Theorem 4.2 already allows for spatial adaptivity, so we focus on temporal adaptivity. Lemma 4 suggests that a good refinement time strategy is to choose $n_{k}$ to be the minimal integer such that $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n_{k}-1})\lesssim a_{\operatorname{E}}^{-k}$. However, the value of $\operatorname{E}_{0}$ may be hard to estimate and so we retain a “backstop” condition which guarantees that convergence is no slower than the rate given by Theorem 4.2. In the non- classical case of $a_{\operatorname{U}}>1$, we provide the following theorem. ###### Theorem 4.3 thm: stronger exponential FISTA convergence Let $(\mathds{U}^{n}\subset\mathds{H})_{n\in\mathds{N}}$ be a sequence of subsets satisfying (10), compute $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}$ and $\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}$ by Algorithm 1. Suppose that there exists a monotone increasing sequence $n_{k}\in\mathds{N}$ such that $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\coloneqq\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n_{k}-1}\in\mathds{U}^{n_{k}}\cap\mathds{U}^{n_{k}+1}\cap\ldots\cap\mathds{U}^{n_{k+1}-1}$ for all $k\in\mathds{N}$. If $(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})_{k\in\mathds{N}}$ is an $(a_{\operatorname{U}},a_{\operatorname{E}})$-minimising sequence of $\operatorname{E}$ with $a_{\operatorname{U}}>1$ and $n_{k}^{2}\lesssim a_{\operatorname{E}}^{k}a_{\operatorname{U}}^{2k}$, then $\min_{n\leq N}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})=\min_{n\leq N}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\inf_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])\lesssim\frac{1}{N^{2(1-\kappa)}}\qquad\text{ where }\qquad\kappa=\frac{\log a_{\operatorname{U}}^{2}}{\log a_{\operatorname{E}}+\log a_{\operatorname{U}}^{2}}$ uniformly for $N\in\mathds{N}$. The proof is given in Theorem A.3. If we directly compare Theorems 4.2 and 4.3, both are a direct result of Lemma 4 assuming a specific choice of $n_{k}$ or $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}$ respectively. We note that the convergence rate is the same in both theorems but the price for better adaptivity (i.e. only an upper bound on $n_{k}$) is a slightly weaker stability guarantee (now convergence of $\min_{n\leq N}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})$). In Theorem 4.2, as in the original FISTA algorithm, the sequence $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})$ is not monotone but the magnitude of oscillation is guaranteed to decay in time. This behaviour is lost in Theorem 4.3. Although we do not prove it here, it can be shown that the stronger condition $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\in\mathds{U}^{n_{k}}\cap\mathds{U}^{n_{k}+1}\cap\ldots\cap\mathds{U}^{n_{k+1}-1}\cap\ldots\cap\mathds{U}^{N}$ (15) is sufficient to restore the stronger last-iterate guarantee on $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})$. Again, monotonicity of $\mathds{U}^{n}$ corresponds with improved stability of Algorithm 1. To enable a more practical implementation of Theorem 4.3, the following lemma describes several refinement strategies which provide sufficient condition for $\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})\lesssim a_{\operatorname{E}}^{-k}$. ###### Lemma 6 thm: practical refinement criteria Let $(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})_{k\in\mathds{N}}$ be a sequence in $\mathds{H}$ with ${\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right\rVert}\lesssim a_{\operatorname{U}}^{k}$. Suppose $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\in\widetilde{\mathds{U}}^{k}\coloneqq\mathds{U}^{n_{k}}$ and denote $\operatorname{E}_{0}(\widetilde{\mathds{U}}^{k})\coloneqq\inf_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\widetilde{\mathds{U}}^{k}}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$. Any of the following conditions are sufficient to show that $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}$ is an $(a_{\operatorname{U}},a_{\operatorname{E}})$-minimising sequence of $\operatorname{E}$: 1. 1. Small continuous gap refinement: $\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})\leq\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]a_{\operatorname{E}}^{-k}$ for all $k\in\mathds{N}$, some $\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]>0$. 2. 2. Small discrete gap refinement: $\operatorname{E}_{0}(\widetilde{\mathds{U}}^{k})\leq\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]a_{\operatorname{E}}^{-k}$ and $\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})-\operatorname{E}_{0}(\widetilde{\mathds{U}}^{k-1})\leq\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]a_{\operatorname{E}}^{-k}$ for all $k>0$, some $\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]>0$. Otherwise, suppose there exists a Banach space $(\mathds{U},{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|})$ which contains each $\widetilde{\mathds{U}}^{k}$, $\sup_{k\in\mathds{N}}{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}<\infty$, and the sublevel sets of $\operatorname{E}$ are ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$-bounded. With the subdifferential $\partial\operatorname{E}\colon\mathds{U}\rightrightarrows\mathds{U}^{*}$, it is also sufficient if either: 1. 3. Small continuous gradient refinement: $\sup_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}}\inf_{\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]\in\partial\operatorname{E}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})}\frac{|\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1],\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rangle|}{{\left|\kern-0.75346pt\left|\kern-0.75346pt\left|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right|\kern-0.75346pt\right|\kern-0.75346pt\right|}}\leq\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]a_{\operatorname{E}}^{-k}$ for all $k\in\mathds{N}$, some $\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]>0$. 2. 4. Small discrete gradient refinement: $\operatorname{E}_{0}(\widetilde{\mathds{U}}^{k})\leq\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]a_{\operatorname{E}}^{-k}$ and $\sup_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0],\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\in\widetilde{\mathds{U}}^{k}}\inf_{\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]\in\mathds{V}^{k}}\frac{|\left\langle v,\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\right\rangle|}{{\left|\kern-0.75346pt\left|\kern-0.75346pt\left|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\right|\kern-0.75346pt\right|\kern-0.75346pt\right|}}\leq\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]a_{\operatorname{E}}^{-k}$ for all $k\in\mathds{N}$, some $\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]>0$, where $\mathds{V}^{k}\coloneqq\partial(\operatorname{E}|_{\widetilde{\mathds{U}}^{k}})(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})$. The proof is given in Lemma 14. The refinement criteria described by Lemma 6 can be split into two groups. Cases (1) and (3) justify that any choice of $\mathds{U}^{n_{k}}$ satisfies the required conditions, so long as $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\in\mathds{U}^{n_{k}}$. In cases (2) and (4), $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}$ is sufficient to choose the refinement time $n_{k}$, but an apriori bound is required on $\operatorname{E}_{0}(\widetilde{\mathds{U}}^{k})$. In these cases one could, for example, choose $\widetilde{\mathds{U}}^{k}$ to be a uniform discretisation with a priori estimates. Another splitting of the criteria is into gap and gradient computations. Typically, gradient norms (in (4) and (5)) should be easier to estimate than function gaps because they only require local knowledge rather than global, i.e. $\partial\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})$ rather than an estimate of $\inf_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$. Implicitly, the global information comes from an extra condition on $\operatorname{E}$ to assert that sublevel sets are bounded. ## 5 General examples We consider the main use of Algorithm 1 to be where there exists a Banach space $(\mathds{U},{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|})$ such that $\mathds{U}\supset\\{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}\operatorname{\;s.t.\;}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])<\infty\\}$ and $\inf_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])=\min_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])=\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})$ for some $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\in\mathds{U}$. The cases where $\mathds{H}$ has finite dimension or is separable are more straightforward; if the total number of refinements is finite (i.e. $\mathds{U}^{n}=\mathds{U}^{N}$ for all $n\geq N$, some $N\in\mathds{N}$), then $a_{\operatorname{U}}=1$. This holds for most finite dimensional problems as well as the countable example discussed in detail in Section 6. In this section we give explicit computations of $a_{\operatorname{U}}$ and $a_{\operatorname{E}}$ in the setting where $\mathds{H}=L^{2}(\Omega)$ for some domain $\Omega\subset\mathds{R}^{d}$ and the subsets $\mathds{U}^{n}$ will be finite dimensional finite-element–like spaces, as defined below. ###### Definition 2 Suppose ${\left\lVert\cdot\right\rVert}_{q}\lesssim{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ (i.e. $\mathds{U}\subset L^{q}(\Omega)$) for some $q\in[1,\infty]$ and connected, bounded, measurable domain $\Omega\subset\mathds{R}^{d}$. We say that a collection $\mathds{M}$ is a _mesh_ if $\bigcup_{\omega\in\mathds{M}}\omega\supset\Omega\qquad\text{and}\qquad|\omega\cap\omega^{\prime}|=0\qquad\text{for all $\omega,\omega^{\prime}\in\mathds{M},\ \omega\neq\omega^{\prime}$.}$ Furthermore, we say a sequence of meshes $(\mathds{M}^{k})_{k\in\mathds{N}}$ is _consistent_ if there exists $\omega_{0}\subset\Omega$ such that $\forall\omega\in\mathds{M}^{k}\quad\exists(\IfEqCase{0}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k0]_{\omega},\vec{\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]}_{\omega})\in\mathds{R}^{d\times d}\times\mathds{R}^{d}\quad\text{such that}\quad\vec{x}\in\omega_{0}\iff\IfEqCase{0}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k0]_{\omega}\vec{x}+\vec{\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]}_{\omega}\in\omega.$ Fix $h\in(0,1)$, linear subspaces $\widetilde{\mathds{U}}^{k}\subset\mathds{H}$, and consistent meshes $\mathds{M}^{k}$. We say that the sequence $(\widetilde{\mathds{U}}^{k})_{k\in\mathds{N}}$ is an _$h$ -refining sequence of finite element spaces_ if there exists $c_{\IfEqCase{0}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k0]}>0$ such that: $\forall(\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]},\omega)\in\widetilde{\mathds{U}}^{k}\times\mathds{M}^{k},\quad\operatorname{det}(\IfEqCase{0}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k0]_{\omega})\geq c_{\IfEqCase{0}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k0]}h^{kd}\quad\text{and}\quad\exists\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\widetilde{\mathds{U}}^{0}\quad\text{such that}\quad\forall\vec{x}\in\omega_{0},\ \IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0](\vec{x})=\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}(\IfEqCase{0}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k0]_{\omega}\vec{x}+\vec{\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]}_{\omega}).$ We say that $(\widetilde{\mathds{U}}^{k})_{k\in\mathds{N}}$ is _of order $p$_ if for any $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\in\operatorname*{argmin}_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$ there exists a sequence $(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})_{k\in\mathds{N}}$ such that $\forall k\in\mathds{N},\qquad\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\in\widetilde{\mathds{U}}^{k}\quad\text{and}\quad{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}\lesssim_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}}h^{kp}.$ (16) We allow the implicit constant to have any dependence on $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$ so long as it is finite. For example, in the case of Sobolev spaces we would expect an inequality of the form ${\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right\rVert}_{W^{0,2}}\lesssim h^{kp}{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right\rVert}_{W^{p,2}}$ Strang1972 . ###### Remark 1 To clarify this definition with an example, suppose we wish to approximate $L^{q}(\Omega)$ with piecewise linear finite elements with a triangulated mesh. Then, $\omega_{0}\subset\Omega$ is a single triangle of diameter $O(h)$ and all meshes $\mathds{M}^{k}$ must be triangulations of $\Omega$ with cell volumes scaling no faster than $O(h^{kd})$. The function $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]$ from the $h$-refining property is an arbitrary linear element, so that each $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\widetilde{\mathds{U}}^{k}$ is linear on each $\omega\in\mathds{M}^{k}$, which leads to an order $p=2$ if $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\in W^{1,2}(\Omega)$. We note that any piecewise polynomial finite element (or spline) space can be used to form a $h$-refining sequence of subspaces. Wavelets with a compactly supported basis behave like a multi-resolution finite element space as there is always overlap in the supports of basis vectors. Similarly, a Fourier basis does satisfy the scaling properties, but each basis vector has global support. Both of these exceptions are important and could be accounted for with further analysis but we focus on the more standard finite element case. In order to align these discretisation properties with the assumptions of Theorems 4.2 and 4.3, we make the following observation. ###### Lemma 7 Fix $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\in\operatorname*{argmin}_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$ and $p^{\prime},q^{\prime}>0$. If a sequence $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\in\mathds{H}$ satisfies ${\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right\rVert}\lesssim h^{-kq^{\prime}}\quad\text{and}\quad\operatorname{E}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})-\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})\lesssim h^{kp^{\prime}},$ then $(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})_{k\in\mathds{N}}$ is an $(a_{\operatorname{U}},a_{\operatorname{E}})$-minimising sequence of $\operatorname{E}$ for $a_{\operatorname{U}}=h^{-q^{\prime}}$ and $a_{\operatorname{E}}=h^{-p^{\prime}}$. This is precisely rewriting the statement of Definition 1 into terms of resolution $h$. The following theorem links $p$ and $q$ from Definition 2 with $p^{\prime}$ and $q^{\prime}$ from Lemma 7. ###### Theorem 5.1 Suppose $\mathds{H}=L^{2}(\Omega)$ for some connected, bounded domain $\Omega\subset\mathds{R}^{d}$ and ${\left\lVert\cdot\right\rVert}_{q}\lesssim{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ for some $q\in[1,\infty]$. For $p\geq 0$ and $h\in(0,1)$, if $(\widetilde{\mathds{U}}^{k})_{k\in\mathds{N}}$ is an $h$-refining sequence of finite element spaces of order $p$, then $(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})_{k\in\mathds{N}}$ is an $(a_{\operatorname{U}},a_{\operatorname{E}})$-minimising sequence of $\operatorname{E}$ for $\displaystyle a_{\operatorname{U}}$ $\displaystyle\leq\begin{cases}1&\text{if }q\geq 2,\\\ \sqrt{h^{-d}}&q<2\text{ and }\sup_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\widetilde{\mathds{U}}^{0}}\frac{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{L^{\infty}(\omega_{0})}}{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{L^{2}(\omega_{0})}}<\infty\end{cases},$ $\displaystyle a_{\operatorname{E}}$ $\displaystyle\geq\begin{cases}h^{-2p}&\text{if }\nabla\operatorname{E}\text{ is ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$-Lipschitz at }\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*},\\\ h^{-p}\qquad&\text{if }\operatorname{E}\text{ is ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$-Lipschitz at }\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*},\\\ 1&\text{otherwise.}\end{cases}$ The proof of this theorem is in Appendix B. Note that $\sup_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\widetilde{\mathds{U}}^{0}}{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{L^{\infty}(\omega_{0})}{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{L^{2}(\omega_{0})}^{-1}$ is finite whenever $\widetilde{\mathds{U}}^{0}\subset L^{\infty}(\Omega)$ is finite dimensional, so this is not a very strong assumption. The main take- home for this theorem is that the computation of $a_{\operatorname{U}}$ and $a_{\operatorname{E}}$ is typically very simple and clear given a particular choice of ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ and $\operatorname{E}$. We also briefly remark that the Lipschitz constants in this lemma do not need to be valid globally, only on the sequence $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}$. The same result holds under a local-Lipschitz assumption, for example on the ball of radius $\sup_{k\in\mathds{N}}{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ which is finite whenever $p\geq 0$. ## 6 L1 penalised reconstruction The canonical example for FISTA is the LASSO problem with a quadratic data fidelity and L1 regularisation. In this section we develop the necessary analytical tools for the variant with general smooth fidelity term which will be used for numerical results in Section 7. We consider three forms which will be referred to as the continuous, countable, and discrete problem depending on whether the space $\mathds{U}$ is $\mathcal{M}([0,1]^{d})$, $\ell^{1}(\mathds{R})$, or $\mathds{R}^{M}$ respectively. We choose $\mathds{H}$ to be $L^{2}([0,1]^{d}),\ \ell^{2}(\mathds{R}),$ or $\mathds{R}^{M}$ correspondingly. Let $\mathsf{A}\colon\mathds{U}\cap\mathds{H}\to\mathds{R}^{m}$ be a linear operator represented by the kernels $\psi_{j}\in\mathds{H}$ such that $\forall\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}\cap\mathds{H},\ j=1,\ldots,m,\qquad(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])_{j}=\left\langle\psi_{j},\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rangle.$ (17) In the continuous case we will assume the additional smoothness $\psi_{j}\in C^{1}([0,1]^{d})$. In Section 6.5 we will formally define and estimate several operator semi-norms for $\mathsf{A}$ of this form, for example Lemma 8 confirms that $\mathsf{A}$ is continuous on $\mathds{H}$ (without loss of generality ${\left\lVert\mathsf{A}\right\rVert}\leq 1$). In each case, the energy we consider is written as $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])=\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\eta)+\mu{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ (18) for some $\mu>0$ where ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}={\left\lVert\cdot\right\rVert}_{1}$. We assume $\operatorname{f}\in C^{1}(\mathds{R}^{m})$ is convex, bounded from below, and $\nabla\operatorname{f}$ is 1-Lipschitz. Let $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\in\operatorname*{argmin}_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$, which is non-empty so long as $\psi_{j}\in C([0,1]^{d})$, see the proof of (Bredies2013, , Prop. 3.1) when $\operatorname{f}$ is quadratic. The aim of this section is to develop all of the necessary tools for implementing Algorithm 1 on the energy (18) using the convergence guarantees of either Theorem 4.2 or Theorem 4.3. This includes computing the rates $a_{\operatorname{U}}$ and $a_{\operatorname{E}}$, estimating the continuous gap $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})$, and developing an efficient refinement choice for $\mathds{U}^{n}$. Below we will just describe the form of $\widetilde{\mathds{U}}^{k}$ under the assumption that $\mathds{U}^{n}\subset\widetilde{\mathds{U}}^{k}$ is chosen adaptively for $n=n_{k-1}+1,\ldots,n_{k}$. The index $k$ refers to the scale or resolution and $n$ refers to the iteration number of the reconstruction algorithm. ### 6.1 Continuous case We start by estimating rates in the case $\mathds{U}=\mathcal{M}(\Omega)$ where $\Omega=[0,1]^{d}$. In this case we choose $\widetilde{\mathds{U}}^{k}$ to be the span of all piecewise constant functions on a mesh of squares with maximum side length $2^{-k}$ (i.e. $h=\tfrac{1}{2}$) and $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\coloneqq\sum_{\omega\in\mathds{M}^{k}}\frac{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}(\omega)}{|\omega|}\mathds{1}_{\omega}\quad\text{where}\quad\mathds{1}_{\omega}(\vec{x})=\begin{cases}1&\vec{x}\in\omega\\\ 0&\text{else}\end{cases}.$ By construction $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\in\widetilde{\mathds{U}}^{k}$, however note that for any $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in L^{1}(\Omega)$ and Dirac mass $\delta$ supported in $(0,1)^{d}$, ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\delta\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}=\sup_{\varphi\in C(\Omega),{\left\lVert\varphi\right\rVert}_{L^{\infty}}\leq 1}\left\langle\varphi,\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\delta\right\rangle={\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}+{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\delta\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}\geq 1=h^{0}.$ (19) Because of this, application of Theorem 5.1 with $p=0$ gives $a_{\operatorname{U}}=2^{\frac{d}{2}}$ but only $a_{\operatorname{E}}\geq 1$. To improve our estimate of $a_{\operatorname{E}}$ requires additional assumptions on $\mathsf{A}$. Note that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}=\sum_{\omega\in\mathds{M}^{k}}|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}(\omega)|\leq{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$, therefore we have $\displaystyle\operatorname{E}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})-\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})$ $\displaystyle=\operatorname{f}(\mathsf{A}\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\eta)-\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}-\eta)+\mu\left({\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}-{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}\right)$ (20) $\displaystyle\leq\nabla\operatorname{f}(\mathsf{A}\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\eta)\vbox{\hbox{\leavevmode\resizebox{6.0pt}{}{$\cdot$}}}\mathsf{A}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})$ (21) $\displaystyle\leq\left[{\left\lVert\nabla\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}-\eta)\right\rVert}_{\ell^{2}}+{\left\lVert\mathsf{A}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})\right\rVert}_{\ell^{2}}\right]{\left\lVert\mathsf{A}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})\right\rVert}_{\ell^{2}}.$ (22) as $\operatorname{f}$ is convex with 1-Lipschitz gradient. Clearly ${\left\lVert\nabla\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}-\eta)\right\rVert}_{\ell^{2}}$ is a constant. For the other term, for all $\vec{r}\in\mathds{R}^{m}$ denote $\varphi\coloneqq\mathsf{A}^{*}\vec{r}$, then note that $\vec{r}\vbox{\hbox{\leavevmode\resizebox{6.0pt}{}{$\cdot$}}}\mathsf{A}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})=\left\langle\varphi,\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right\rangle=\sum_{\omega\in\mathds{M}^{k}}\int_{\omega}\varphi(\vec{x})\mathop{}\\!\mathrm{d}[\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}]=\sum_{\omega\in\mathds{M}^{k}}|\omega|^{-1}\iint_{\omega^{2}}[\varphi(\vec{x})-\varphi(\vec{y})]\mathop{}\\!\mathrm{d}\vec{x}\mathop{}\\!\mathrm{d}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}(\vec{y}).$ (23) With the pointwise bound $|\varphi(\vec{x})-\varphi(\vec{y})|\leq\operatorname{diam}(\omega){\left\lVert\nabla\varphi\right\rVert}_{L^{\infty}}=\sqrt{d}2^{-k}{\left\lVert\nabla[\mathsf{A}^{*}\vec{r}]\right\rVert}_{L^{\infty}}$, we deduce the estimate ${\left\lVert\mathsf{A}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})\right\rVert}_{\ell^{2}}=\sup_{\vec{r}\in\mathds{R}^{m}}{\left\lVert\vec{r}\right\rVert}_{\ell^{2}}^{-1}\left\langle\mathsf{A}^{*}\vec{r},\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right\rangle\leq\sqrt{d}2^{-k}{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}\sup_{\vec{r}\in\mathds{R}^{m}}{\left\lVert\vec{r}\right\rVert}_{\ell^{2}}^{-1}{\left\lVert\nabla[\mathsf{A}^{*}\vec{r}]\right\rVert}_{L^{\infty}}.$ (24) In Lemma 9 we will show that this last term, which we denote the semi-norm $|\mathsf{A}^{*}|_{\ell^{2}\to C^{1}}$, is bounded by $\sqrt{m}\max_{j\in[m]}{\left\lVert\nabla\Psi_{j}\right\rVert}_{\infty}$. We conclude that $\operatorname{E}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})-\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})\lesssim 2^{-k}$. In particular, this computation confirms two things. Firstly that the scaling constant is $a_{\operatorname{E}}=2$, and secondly that the required smoothness to achieve a good rate with Algorithm 1 is that $\mathsf{A}^{*}\colon\mathds{R}^{m}\to C^{1}(\Omega)$ is a bounded operator. This accounts for using the weaker topology of $\mathcal{M}(\Omega)$ rather than $L^{1}(\Omega)$. Inserting the computed rates into Theorem 4.2 or Theorem 4.3 gives the guaranteed convergence rate $\kappa=\frac{\log a_{\operatorname{U}}^{2}}{\log a_{\operatorname{E}}+\log a_{\operatorname{U}}^{2}}=\frac{d}{1+d}\quad\implies\quad\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\inf_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])\lesssim n^{-2(1-\kappa)}=n^{-\frac{2}{1+d}}.$ (25) This rate can be used to infer the required resolution at each iteration, in particular on iteration $n$ with $n^{2}\simeq(a_{\operatorname{E}}a_{\operatorname{U}}^{2})^{k}$ we expect the resolution to be $2^{-k}=\left(a_{\operatorname{E}}a_{\operatorname{U}}^{2}\right)^{\frac{k}{1+d}}\simeq n^{-\frac{2}{1+d}}.$ (26) ### 6.2 Countable and discrete case We now extend the rate computations to the case when $\mathds{U}=\ell^{1}(\mathds{R})$, or a finite dimensional subspace. The key fact here is that, even when $\mathds{U}$ is infinite dimensional, it is known (e.g. (Unser2016, , Thm 6) and (Boyer2019, , Cor 3.8)) that there exists $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\in\operatorname*{argmin}_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$ with at most $m$ non-zeros. If this is the case, then $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\in\ell^{2}(\mathds{R})$, indeed ${\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right\rVert}_{\ell^{2}}\leq\sqrt{m}{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right\rVert}_{\ell^{1}}$. This makes the estimates of $a_{\operatorname{E}}/a_{\operatorname{U}}$ much simpler than in the continuous case as we can stay in the finite-dimensional Hilbert-space setting. For countable dimensions we consider discretisation subspaces of the form $\widetilde{\mathds{U}}^{k}=\\{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\ell^{1}(\mathds{R})\operatorname{\;s.t.\;}i\notin J_{k}\implies\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{i}=0\\}$ for some sets $J_{k}\subset\mathds{N}$, i.e. infinite vectors with finitely many non-zeros. The key change in analysis from the continuous case is ${\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right\rVert}<\infty$, so $a_{\operatorname{U}}=1$ and the expected rate of $n^{-2}$, independent of $a_{\operatorname{E}}$ or any additional properties of $\mathsf{A}$. The number of refinements will also be finite, therefore $n_{k}=\infty$ for some $k$, the remaining conditions of Theorems 4.2 and 4.3 hold trivially. ### 6.3 Refinement metrics Lemma 6 shows that adaptive refinement can be performed based on estimates of the function gap or the subdifferential. In this subsection we provide estimates for the forth case of Lemma 6 which can be easily computed. In this case we consider $\partial\operatorname{E}\colon\mathds{H}\rightrightarrows\mathds{H}$ so that subdifferentials are well behaved, for example for explicit computation assuming validity of the chain/sum rules for differentiation. #### 6.3.1 Bounds for discretised functionals We start by computing estimates for discretised energies. This covers the cases when either the continuous/countable energy is projected onto $\mathds{U}^{n}$, or $\mathds{U}$ is finite dimensional. For notation we will use the continuous case, to recover the other cases just replace continuous indexing with discrete (i.e. $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0](\vec{x})\leadsto\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{i}$). Let $\mathsf{\Pi}_{n}\colon\mathds{H}\to\mathds{U}^{n}$ denote the orthogonal projection. We consider the discretised function $\operatorname{E}|_{\mathds{U}^{n}}\colon\mathds{U}^{n}\to\mathds{R}$ and its subdifferential $\partial_{n}\operatorname{E}(\cdot)=\mathsf{\Pi}_{n}\partial\operatorname{E}(\cdot)$ on $\mathds{U}^{n}$. In our case, the behaviour of $\operatorname{E}|_{\mathds{U}^{n}}$ is equivalent to replacing $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]$ with $\mathsf{\Pi}_{n}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]$, and $\mathsf{A}^{*}$ with $\mathsf{\Pi}_{n}\mathsf{A}^{*}$. ##### Discrete gradient We can use $\mathsf{\Pi}_{n}$ to compute the discrete subdifferential at $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}\in\mathds{U}^{n}$: $\displaystyle\partial_{n}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})(\vec{x})$ $\displaystyle=[\mathsf{\Pi}_{n}\mathsf{A}^{*}\nabla\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}-\eta)](\vec{x})+\begin{cases}\\{+\mu\\}&\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}(\vec{x})>0\\\ [-\mu,\mu]&\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}(\vec{x})=0\\\ \\{-\mu\\}&\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}(\vec{x})<0\end{cases}$ (27) $\displaystyle\eqqcolon[\mathsf{\Pi}_{n}\mathsf{A}^{*}\nabla\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}-\eta)](\vec{x})+\mu\mathsf{\Pi}_{n}\operatorname{sign}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}(\vec{x}))$ (28) where we define $s+\mu[-1,1]=[s-\mu,s+\mu]$ for all $s\in\mathds{R}$, $\mu\geq 0$. As ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}={\left\lVert\cdot\right\rVert}_{1}$, the natural metric for $\partial_{n}\operatorname{E}$ is ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}={\left\lVert\cdot\right\rVert}_{\infty}$ which we can estimate $\displaystyle{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\partial_{n}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}$ $\displaystyle=\max_{\vec{x}\in\Omega}\min_{\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]}\left\\{|\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]|\operatorname{\;s.t.\;}\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]\in\mathsf{\Pi}_{n}\mathsf{A}^{*}\nabla\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}-\eta)(\vec{x})+\mu\mathsf{\Pi}_{n}\operatorname{sign}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}(\vec{x}))\right\\}$ (29) $\displaystyle=\max_{\vec{x}\in\Omega}\begin{cases}|[\mathsf{\Pi}_{n}\mathsf{A}^{*}\nabla\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}-\eta)(\vec{x})+\mu|&\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}(\vec{x})>0\\\ |[\mathsf{\Pi}_{n}\mathsf{A}^{*}\nabla\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}-\eta)(\vec{x})-\mu|&\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}(\vec{x})<0\\\ \max\left(|\mathsf{\Pi}_{n}\mathsf{A}^{*}\nabla\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}-\eta)(\vec{x})|-\mu,0\right)&\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}(\vec{x})=0\end{cases}$ (30) which can be used directly in Lemma 6. ##### Discrete gap We now move on to the discrete gap, $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\min_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}^{n}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$. This can be computed with a dual representation (e.g. Duval2017a ), $\displaystyle\min_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}^{n}}\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\eta)+\mu{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ $\displaystyle=\min_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\max_{\vec{\varphi}\in\mathds{R}^{m}}(\mathsf{A}\mathsf{\Pi}_{n}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\eta)\vbox{\hbox{\leavevmode\resizebox{6.0pt}{}{$\cdot$}}}\vec{\varphi}+\mu{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\mathsf{\Pi}_{n}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}-\operatorname{f}^{*}(\vec{\varphi})$ (31) $\displaystyle=\max_{\vec{\varphi}\in\mathds{R}^{m}}\min_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}(\mathsf{A}\mathsf{\Pi}_{n}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\eta)\vbox{\hbox{\leavevmode\resizebox{6.0pt}{}{$\cdot$}}}\vec{\varphi}+\mu{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\mathsf{\Pi}_{n}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}-\operatorname{f}^{*}(\vec{\varphi})$ (32) $\displaystyle=\max_{\vec{\varphi}\in\mathds{R}^{m}}\begin{cases}-\eta\vbox{\hbox{\leavevmode\resizebox{6.0pt}{}{$\cdot$}}}\vec{\varphi}-\operatorname{f}^{*}(\vec{\varphi})&\qquad{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\mathsf{\Pi}_{n}\mathsf{A}^{*}\vec{\varphi}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}\leq\mu\\\ -\infty&\qquad\text{else}\end{cases}$ (33) $\displaystyle=-\min_{\vec{\varphi}\in\mathds{R}^{m}}\underbrace{\operatorname{f}^{*}(\vec{\varphi})+\eta\vbox{\hbox{\leavevmode\resizebox{6.0pt}{}{$\cdot$}}}\vec{\varphi}}_{\eqqcolon\operatorname{E}^{\dagger}(\vec{\varphi})}+\chi({\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\mathsf{\Pi}_{n}\mathsf{A}^{*}\vec{\varphi}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}\leq\mu).$ (34) In particular, $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])-\min_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}^{n}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])=\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])+\min_{\vec{\varphi}\in\mathds{R}^{m}\operatorname{\;s.t.\;}{\left|\kern-0.75346pt\left|\kern-0.75346pt\left|\mathsf{\Pi}_{n}\mathsf{A}^{*}\vec{\varphi}\right|\kern-0.75346pt\right|\kern-0.75346pt\right|}_{*}\leq\mu}\operatorname{E}^{\dagger}(\vec{\varphi})\leq\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])+\operatorname{E}^{\dagger}(\vec{\varphi})$ (35) for any feasible $\vec{\varphi}\in\mathds{R}^{m}$. We further derive the criticality condition, if $(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*},\vec{\varphi}^{*})$ is a saddle point, then $\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}-\eta\in\partial\operatorname{f}^{*}(\vec{\varphi}^{*}),\qquad\text{or equivalently}\qquad\vec{\varphi}^{*}=\nabla\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}-\eta).$ (36) We remark briefly that $\operatorname{E}^{\dagger}$ should be thought of as the dual of $\operatorname{E}$ but without the constraint. We choose to omit it here to highlight that it is only the constraint which changes between the discrete and continuous cases; the value of $\operatorname{E}^{\dagger}$ will remain the same. Given $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}\in\mathds{U}^{n}$, the optimality condition motivates a simple rule for choosing $\vec{\varphi}$: $\vec{\varphi}_{n}\coloneqq\nabla\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}-\eta),\qquad\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])-\min_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{\prime}\in\mathds{U}^{n}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{\prime})\leq\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])+\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]\vec{\varphi}_{n})$ (37) for some $0\leq\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]\leq\frac{\mu}{{\left|\kern-0.75346pt\left|\kern-0.75346pt\left|\mathsf{\Pi}_{n}\mathsf{A}^{*}\vec{\varphi}_{n}\right|\kern-0.75346pt\right|\kern-0.75346pt\right|}_{*}}$. In the case $\operatorname{f}(\cdot)=\frac{1}{2}{\left\lVert\cdot\right\rVert}_{\ell^{2}}^{2}$, one can use the optimal choice $\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]=\max\left(0,\min\left(\frac{-\eta\vbox{\hbox{\leavevmode\resizebox{6.0pt}{}{$\cdot$}}}\vec{\varphi}_{n}}{{\left\lVert\vec{\varphi}_{n}\right\rVert}_{\ell^{2}}^{2}},\frac{\mu}{{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\mathsf{\Pi}_{n}\mathsf{A}^{*}\vec{\varphi}_{n}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}}\right)\right).$ (38) To apply Algorithm 1, we are assuming that both $\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}-\eta)$ and $\mathsf{\Pi}_{n}\nabla\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}-\eta)$ are easily computable, therefore $\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]$ and $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]\vec{\varphi}_{n})$ are also easy to compute. #### 6.3.2 Bounds for countable functionals Extending the results of Section 6.3.1 to $\mathds{U}=\ell^{1}(\mathds{R})$ is analytically very simple but computationally relies heavily on the specific choice of $\mathsf{A}$. The computations of subdifferentials and gaps carry straight over replacing $\mathsf{\Pi}_{n}$ with the identity and adding the sets $J_{n}\subset\mathds{N}$ which define $\mathds{U}^{n}=\\{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\ell^{1}\operatorname{\;s.t.\;}i\notin J_{n}\implies\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{i}=0\\}$. Recall that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\partial\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}\coloneqq\inf_{s\in\operatorname{sign}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})}{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\mathsf{A}^{*}\vec{\varphi}_{n}+\mu s\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}$ where the $\operatorname{sign}$ function has the pointwise set-valued definition as indicated in (27)-(28). Where $[u_{n}]_{i}=0$, the choice $s_{i}=\min(1,\max(-1,-\mu^{-1}[\mathsf{A}^{*}\vec{\varphi}_{n}]_{i}))$ achieves the minimal value $\displaystyle{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\partial\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}$ $\displaystyle=\max_{i\in\mathds{N}}\begin{cases}|[\mathsf{A}^{*}\vec{\varphi}_{n}]_{i}+\mu|&[\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}]_{i}>0\\\ |[\mathsf{A}^{*}\vec{\varphi}_{n}]_{i}-\mu|&[\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}]_{i}<0\\\ \max\left(|[\mathsf{A}^{*}\vec{\varphi}_{n}]_{i}|-\mu,0\right)&[\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}]_{i}=0\end{cases}$ (39) $\displaystyle\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\inf_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$ $\displaystyle\leq\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n}),\qquad\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\in\left[0,\frac{\mu}{{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\mathsf{A}^{*}\vec{\varphi}_{n}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}}\right]$ (40) where $\vec{\varphi}_{n}=\nabla\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}-\eta)\in\mathds{R}^{m}$ is always exactly computable. In the countable case, the sets $J_{n}$ give a clear partition into known/unknown values in these definitions. For $i\in J_{n}$ the computation is the same as in Section 6.3.1, then for $i\notin J_{n}$ we know $[\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}]_{i}=0$ which simplifies the remaining computations. This leads to: $\displaystyle{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\partial\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}$ $\displaystyle=\max\left(\max_{i\in J_{n}}|[\partial\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})]_{i}|,\ \sup_{i\notin J_{n}}|[\partial\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})]_{i}|\right)=\max\left({\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\partial_{n}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*},\ \sup_{i\notin J_{n}}|[\mathsf{A}^{*}\vec{\varphi}_{n}]_{i}|-\mu\right)$ (41) $\displaystyle{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\mathsf{A}^{*}\vec{\varphi}_{n}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}$ $\displaystyle=\max\left(\max_{i\in J_{n}}|[\mathsf{A}^{*}\vec{\varphi}_{n}]_{i}|,\ \sup_{i\notin J_{n}}|[\mathsf{A}^{*}\vec{\varphi}_{n}]_{i}|\right)\hskip 19.0pt=\max\left({\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\mathsf{\Pi}_{n}\mathsf{A}^{*}\vec{\varphi}_{n}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*},\ \sup_{i\notin J_{n}}|[\mathsf{A}^{*}\vec{\varphi}_{n}]_{i}|\right).$ (42) Both estimates only rely on an upper bound of $\max_{i\notin J_{n}}|[\mathsf{A}^{*}\vec{\varphi}_{n}]_{i}|$. One example computing this value is seen in Section 7.2. #### 6.3.3 Bounds for continuous functionals Finally we extend the results of Section 6.3.1 to continuous problems. Similar to the countable case (39)-(40), the exact formulae can be written down immediately: $\displaystyle{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\partial\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}$ $\displaystyle=\max_{\vec{x}\in\Omega}\begin{cases}|[\mathsf{A}^{*}\vec{\varphi}_{n}](\vec{x})+\mu|&\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}(\vec{x})>0\\\ |[\mathsf{A}^{*}\vec{\varphi}_{n}](\vec{x})-\mu|&\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}(\vec{x})<0\\\ \max\left(|[\mathsf{A}^{*}\vec{\varphi}_{n}](\vec{x})|-\mu,0\right)&\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}(\vec{x})=0\end{cases}$ (43) $\displaystyle\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\inf_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$ $\displaystyle\leq\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n}),\qquad\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\in\left[0,\frac{\mu}{{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\mathsf{A}^{*}\vec{\varphi}_{n}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}}\right]$ (44) with $\operatorname{E}^{\dagger}$ as defined in (34). Recall that there is a mesh $\mathds{M}^{n}$ corresponding to $\mathds{U}^{n}$ such that $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}$ is constant on each $\omega\in\mathds{M}^{n}$, so we can rewrite these bounds: $\displaystyle{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\partial\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}$ $\displaystyle=\max_{\omega\in\mathds{M}^{n}}\begin{cases}{\left\lVert\mathsf{A}^{*}\vec{\varphi}_{n}+\mu\right\rVert}_{L^{\infty}(\omega)}&\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}|_{\omega}>0\\\ {\left\lVert\mathsf{A}^{*}\vec{\varphi}_{n}-\mu\right\rVert}_{L^{\infty}(\omega)}&\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}|_{\omega}<0\\\ \max(0,{\left\lVert\mathsf{A}^{*}\vec{\varphi}_{n}\right\rVert}_{L^{\infty}(\omega)}-\mu)&\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}|_{\omega}=0\end{cases}$ (45) $\displaystyle{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\mathsf{A}^{*}\vec{\varphi}_{n}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}$ $\displaystyle=\max_{\omega\in\mathds{M}^{n}}{\left\lVert\mathsf{A}^{*}\vec{\varphi}_{n}\right\rVert}_{L^{\infty}(\omega)}.$ (46) Now, both values can be estimated relying on pixel-wise supremum norms of $\mathsf{A}^{*}\vec{\varphi}_{n}$ which we have assumed is sufficiently smooth. We will therefore use a pixel-wise Taylor expansion to provide a simple and accurate estimate. For instance, let $\vec{x}_{i}$ be the midpoint of the pixel $\omega$, then ${\left\lVert\mathsf{A}^{*}\vec{\varphi}_{n}\right\rVert}_{L^{\infty}(\omega)}\leq|[\mathsf{A}^{*}\vec{\varphi}_{n}](\vec{x}_{i})|+\frac{\operatorname{diam}(\omega)}{2}|[\nabla\mathsf{A}^{*}\vec{\varphi}_{n}](\vec{x}_{i})|+\frac{\operatorname{diam}(\omega)^{2}}{8}|\mathsf{A}^{*}\vec{\varphi}_{n}|_{C^{2}}.$ (47) In this work we chose a first order expansion because we are looking for extrema of $\mathsf{A}^{*}\vec{\varphi}_{n}$, i.e. we are most interested in the squares $\omega$ such that $|[\mathsf{A}^{*}\vec{\varphi}_{n}](\vec{x}_{i})|\approx\mu,\qquad|[\nabla\mathsf{A}^{*}\vec{\varphi}_{n}](\vec{x}_{i})|\approx 0,\qquad[\nabla^{2}\mathsf{A}^{*}\vec{\varphi}_{n}](\vec{x}_{i})\preceq 0.$ (48) A zeroth order expansion would be optimally inefficient (approximating $|[\nabla\mathsf{A}^{*}\vec{\varphi}_{n}](\vec{x}_{i})|$ with $|\mathsf{A}^{*}\vec{\varphi}_{n}|_{C^{1}}$) and a second order expansion would possibly be more elegant but harder to implement. We found that a first order expansion was simple and efficient. The bounds presented here for continuous problems emphasise the twinned properties required for adaptive mesh optimisation. The mesh should be refined greedily to the structures of $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$, but also must be sufficiently uniform to provide a good estimate for $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})$. This is a classical exploitation/exploration trade-off; exploiting visible structure whilst searching for other structures which are not yet visible. ### 6.4 Support detection The main motivation for using L1 penalties in applications is because it recovers sparse signals, in the case of compressed sensing the support of $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$ is also provably close to the “true” support Duval2017a ; Poon2018 . If $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}\approx\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$ in the appropriate sense, then we should also be able to quantify the statement $\operatorname{supp}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\approx\operatorname{supp}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})$. Such methods are referred to as _safe screening_ rules ElGhaoui2010 which gradually identify the support and allow the optimisation algorithm to constrain parts of the reconstruction to 0. In this subsection we propose a new simple screening rule which is capable of generalising to our continuous subspace approximation setting. It is likely that more advanced methods Bonnefoy2015 ; Ndiaye2017 can also be adapted, although that is beyond the scope of this work. The key difference is the allowance of inexact computations resulting from estimates such as (47). The support of $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$ has already been characterised very precisely Duval2017a ; Poon2018 . In particular, the support is at most $m$ distinct points and are a subset of $\\{\vec{x}\in\Omega\operatorname{\;s.t.\;}|\mathsf{A}^{*}\vec{\varphi}^{*}|(\vec{x})=\mu\\}$ (an equivalent statement holds for the countable case). Less formally, this can also be seen from the the subdifferential computations in Section 6.3, for all $\vec{x}\in\operatorname{supp}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})$ we have $0\in\partial\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})(\vec{x})=[\mathsf{A}^{*}\vec{\varphi}^{*}](\vec{x})+\mu\operatorname{sign}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}(\vec{x})).$ (49) Heuristically, we will use strong convexity of $\operatorname{E}^{\dagger}$ from (34) and smoothness of $\mathsf{A}^{*}$ to quantify the statement: $\text{if}\quad\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n})\approx 0\quad\text{then}\quad\left\\{\vec{x}\operatorname{\;s.t.\;}|[\mathsf{A}^{*}\vec{\varphi}_{n}](\vec{x})|\ll\mu\right\\}\subset\\{\vec{x}\operatorname{\;s.t.\;}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}(\vec{x})=0\\}.$ Recall that $\nabla\operatorname{f}$ is 1-Lipschitz if and only if $\operatorname{f}^{*}$ is 1-strongly convex (Hiriart2013, , Chapter 10, Thm. 4.2.2). Therefore, if $\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n}$ and $\vec{\varphi}^{*}$ are both dual-feasible, then $\tfrac{1}{2}{\left\lVert\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n}-\vec{\varphi}^{*}\right\rVert}_{\ell^{2}}^{2}\leq\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n})-\operatorname{E}^{\dagger}(\vec{\varphi}^{*})=\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n})+\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})\leq\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n})+\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}),$ (50) which gives an easily computable bound on ${\left\lVert\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n}-\vec{\varphi}^{*}\right\rVert}_{\ell^{2}}$. Now we estimate $\mathsf{A}^{*}\vec{\varphi}_{n}$ on the support of $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$: $\displaystyle\min_{\vec{x}\in\operatorname{supp}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})}|[\mathsf{\Pi}_{n}\mathsf{A}^{*}\vec{\varphi}_{n}](\vec{x})|$ $\displaystyle\geq\min_{\vec{x}\in\operatorname{supp}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})}|[\mathsf{A}^{*}\vec{\varphi}_{n}](\vec{x})|$ (51) $\displaystyle=\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}^{-1}\min_{\vec{x}\in\operatorname{supp}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})}|[\mathsf{A}^{*}\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n}](\vec{x})|$ (52) $\displaystyle\geq\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}^{-1}\min_{\vec{x}\in\operatorname{supp}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})}|[\mathsf{A}^{*}\vec{\varphi}^{*}](\vec{x})|-|[\mathsf{A}^{*}\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n}-\mathsf{A}^{*}\vec{\varphi}^{*}](\vec{x})|$ (53) $\displaystyle=\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}^{-1}\min_{\vec{x}\in\operatorname{supp}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})}\mu-|[\mathsf{A}^{*}\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n}-\mathsf{A}^{*}\vec{\varphi}^{*}](\vec{x})|$ (54) $\displaystyle\geq\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}^{-1}\left(\mu-|\mathsf{A}^{*}|_{\ell^{2}\to L^{\infty}}{\left\lVert\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n}-\vec{\varphi}^{*}\right\rVert}_{\ell^{2}}\right).$ (55) Therefore, $|[\mathsf{\Pi}_{n}\mathsf{A}^{*}\vec{\varphi}_{n}](\vec{x})|<\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}^{-1}\left(\mu-\sqrt{2(\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n}))}|\mathsf{A}^{*}|_{\ell^{2}\to L^{\infty}}\right)\qquad\implies\qquad\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}(\vec{x})=0.$ (56) This equation is valid when $\vec{x}$ is either a continuous or countable index, the only distinction is to switch to $\ell^{\infty}$ in the norm of $\mathsf{A}^{*}$. To make the equivalent statement on the discretised problem, simply replace $\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}$ with $\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]$ and $\mathsf{A}^{*}$ with $\mathsf{\Pi}_{n}\mathsf{A}^{*}$. There are two short observations on this formula: * • The convergence guarantee from Theorem 4.2 is for the primal gap $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})$, rather than the primal-dual gap $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n})$ used here. Although there is no guaranteed rate for the primal-dual gap, it is much more easily computable than the primal gap. * • In Section 6.1, $|\mathsf{A}^{*}|_{\ell^{2}\to C^{1}}<\infty$ was required to compute a rate of convergence for $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})$, but only $|\mathsf{A}^{*}|_{\ell^{2}\to L^{\infty}}<\infty$ is needed to estimate the support. ### 6.5 Operator norms For numerical implementation of (18), we are required to accurately estimate several operator norms of $\mathsf{A}$ of the form in (17). In particular, there are kernels $\psi_{j}\in\mathds{H}$ such that $(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])_{j}=\left\langle\psi_{j},\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rangle$ for each $j\in[m]$. Verifying that ${\left\lVert\mathsf{A}\right\rVert}\leq 1$ can be performed by computing $|\mathsf{A}\mathsf{A}^{*}|_{\ell^{2}\to\ell^{2}}$, and the adaptivity described in Sections 6.1, 6.3.3, and 6.4 requires the values of $|\mathsf{A}^{*}|_{\ell^{2}\to L^{\infty}}$, $|\mathsf{A}^{*}|_{\ell^{2}\to C^{1}}$, and $|\mathsf{A}^{*}|_{\ell^{2}\to C^{2}}$. The aim for this section is to provide estimates of these norms and seminorms for the numerical examples presented in Section 7. The following lemma allows for exact computation of the operator norm of $\mathsf{A}$. ###### Lemma 8 If $\mathsf{A}\colon\mathds{H}\to\mathds{R}^{m}$ has kernels $\psi_{j}\in\mathds{H}$ for $j\in[m]$, then $\mathsf{A}\mathsf{A}^{*}\in\mathds{R}^{m\times m}$ has entries $(\mathsf{A}\mathsf{A}^{*})_{i,j}=\left\langle\psi_{i},\psi_{j}\right\rangle$, so the spectral norm ${\left\lVert\mathsf{A}^{*}\mathsf{A}\right\rVert}={\left\lVert\mathsf{A}\mathsf{A}^{*}\right\rVert}$ can be computed efficiently. ###### Proof To compute the entries of $\mathsf{A}\mathsf{A}^{*}\colon\mathds{R}^{m}\to\mathds{R}^{m}$, observe that for any $\vec{r}\in\mathds{R}^{m}$ $(\mathsf{A}\mathsf{A}^{*}\vec{r})_{i}=\left\langle\psi_{i},\mathsf{A}^{*}\vec{r}\right\rangle=\left\langle\psi_{i},\sum_{j=1}^{m}r_{j}\psi_{j}\right\rangle=\sum_{j=1}^{m}\left\langle\psi_{i},\psi_{j}\right\rangle r_{j}$ (57) as required. ∎ If ${\left\lVert\mathsf{A}^{*}\mathsf{A}\right\rVert}$ is not analytically tractable, then Lemma 8 enables it to be computed using standard finite dimensional methods. The operator $\mathsf{A}\mathsf{A}^{*}$ is always finite dimensional, and can be computed without discretisation error. In the continuous case, when $\mathds{H}=L^{2}(\Omega)$ we also need to estimate the smoothness properties of $\mathsf{A}^{*}$. A generic result for this is given in the following lemma. ###### Lemma 9 If $\mathsf{A}\colon L^{2}([0,1]^{d})\to\mathds{R}^{m}$ has kernels $\psi_{j}\in L^{2}(\Omega)\cap C^{k}(\Omega)$ for $j\in[m]$, then for all $\frac{1}{q}+\frac{1}{q^{*}}=1$, $q\in[1,\infty]$, we have $\displaystyle|\mathsf{A}^{*}\vec{r}|_{C^{k}}$ $\displaystyle\coloneqq\sup_{\vec{x}\in\Omega}|\nabla^{k}[\mathsf{A}^{*}\vec{r}]|(\vec{x})\leq\sup_{\vec{x}\in\Omega}{\left\lVert(\nabla^{k}\psi_{j}(\vec{x}))_{j=1}^{m}\right\rVert}_{\ell^{q^{*}}}{\left\lVert\vec{r}\right\rVert}_{\ell^{q}},$ (58) $\displaystyle|\mathsf{A}^{*}|_{\ell^{2}\to C^{k}}$ $\displaystyle\coloneqq\sup_{{\left\lVert\vec{r}\right\rVert}_{\ell^{2}}\leq 1}|\mathsf{A}^{*}\vec{r}|_{C^{k}}\leq\sup_{\vec{x}\in\Omega}{\left\lVert(\nabla^{k}\psi_{j}(\vec{x}))_{j=1}^{m}\right\rVert}_{\ell^{q^{*}}}\times\begin{cases}1&q\geq 2\\\ \sqrt{m^{2-q}}&q<2\end{cases}.$ (59) ###### Proof For the first inequality, we apply the Hölder inequality on $\mathds{R}^{m}$: $|\nabla^{k}[\mathsf{A}^{*}\vec{r}]|(\vec{x})=\left|\sum_{j=1}^{m}\nabla^{k}\psi_{j}(\vec{x})r_{j}\right|\leq\left(\sum_{j=1}^{m}|\nabla^{k}\psi_{j}(\vec{x})|^{q^{*}}\right)^{\frac{1}{q^{*}}}{\left\lVert\vec{r}\right\rVert}_{\ell^{q}}={\left\lVert(\nabla^{k}\psi_{j}(\vec{x}))_{j}\right\rVert}_{\ell^{q^{*}}}{\left\lVert\vec{r}\right\rVert}_{\ell^{q}}\;.$ For the second inequality, if $q\geq 2$ and $\sum_{j=1}^{m}r_{j}^{2}\leq 1$, then $|r_{j}|\leq 1$ for all $j$ and ${\left\lVert\vec{r}\right\rVert}_{\ell^{q}}^{q}\leq{\left\lVert\vec{r}\right\rVert}_{\ell^{2}}^{2}\leq 1$. If $q<2$ and ${\left\lVert\vec{r}\right\rVert}_{\ell^{2}}\leq 1$, then we again use Hölder’s inequality: $\sum_{j=1}^{m}r_{j}^{q}\leq\Big{(}\sum_{j=1}^{m}1^{Q^{*}}\Big{)}^{\frac{1}{Q^{*}}}\Big{(}\sum_{j=1}^{m}r_{j}^{qQ}\Big{)}^{\frac{1}{Q}}\leq m^{\frac{2-q}{2}}$ for $Q=\frac{2}{q}$. ∎ The examples in Section 7 require explicit computations of the expressions in Lemmas 8 and 9. These computations are provided in the appendix, Theorem C.1. ## 7 Numerical examples We present four numerical examples. The first two are in 1D to demonstrate the performance of different variants of Algorithm 1, both with and without adaptivity. In particular, we explore sparse Gaussian deconvolution and sparse signal recovery from Fourier data. We compare with the _continuous basis pursuit_ (CBP) discretisation Ekanadham2011 ; Duval2017b which is also designed to achieve super-resolution accuracy within a convex framework. More details of this method will be provided in Section 7.1. The next example is 2D reconstruction from Radon or X-ray data with wavelet- sparsity and a robust data fidelity. As the forward operator is not sufficiently smooth, we must optimise in $\ell^{1}(\mathds{R})$, which naturally leads to the choice of a wavelet basis. Finally, we process a dataset which represents a realistic application in biological microscopy, referred to as STORM microscopy. In essence, the task is to perform 2D Gaussian de-blurring/super-resolution and denoising to find the location of sparse spikes of signal. In this section, the main aim is to minimise $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})=\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})$, and so this will be our main metric for the success of an algorithm, referred to as the “continuous gap”. Lemma 6 only provides guarantees on the values of $\min_{n\leq N}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})$ so it is this monotone estimate which is plotted. As $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})$ is not known exactly, we always use the estimate $\min_{n\leq N}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\approx\min_{n\leq N}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\min_{n^{\prime}\leq n}\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n^{\prime}})$. Another quantity of interest is minimisation of the discrete energy $\min_{n\leq N}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\min_{n^{\prime}\leq n}\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]\vec{\varphi}_{n^{\prime}})$ which will be referred to as the “discrete gap”. Note that for the adaptive schemes the discrete gap may not be monotonic as the discrete dual problem changes with $N$. The code to reproduce these examples can be found online111https://github.com/robtovey/2020SpatiallyAdaptiveFISTA. ### 7.1 1D continuous LASSO In this example we choose $\mathds{U}=\mathcal{M}([0,1])$, $\mathds{H}=L^{2}([0,1])$, $\operatorname{f}(\cdot)=\frac{1}{2}{\left\lVert\cdot\right\rVert}_{\ell^{2}}^{2}$ and $\mathsf{A}\colon\mathds{U}\to\mathds{R}^{30}$ with either random Fourier kernels: $(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])_{j}=\int_{0}^{1}\cos(\IfEqCase{3}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k3]_{j}x)\mathop{}\\!\mathrm{d}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0](x),\qquad\IfEqCase{3}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k3]_{j}\sim\operatorname{Uniform}[-100,100],\ j=1,2,\ldots,30,\ \mu=0.02,$ (60) or Gaussian kernels on a regular grid: $(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])_{j}=(2\pi\sigma^{2})^{-\frac{1}{2}}\int_{0}^{1}\exp\left(-\frac{(x-(j-1)\Delta)^{2}}{2\sigma^{2}}\right)\mathop{}\\!\mathrm{d}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0](x),\quad\sigma=0.12,\ \Delta=\tfrac{1}{29},\ j=1,2,\ldots,30,\ \mu=0.06.$ (61) Several variants of FISTA are compared for these examples but the key alternative shown here is the CBP discretisation. For this choice of $\operatorname{f}$, we call (18) the continuous LASSO problem, for which there are many numerical methods (c.f. Bredies2013 ; Castro2016 ; Boyd2017 ; Catala2019 ) however, most require the solution of a non-convex problem. We have focused on CBP because it approximates $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$ through a convex discrete optimisation problem which is asymptotically exact in the limit $h\to 0$. It can also be optimised with FISTA which allows for direct comparison with the uniform and adaptive mesh approaches. The idea is that for a fixed mesh, the kernels of $\mathsf{A}$ are expanded to first order on each pixel and a particular first order basis is also chosen Ekanadham2011 ; Duval2017b . If $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$ has only one Dirac spike in each pixel, then the zeroth order information should correspond to the mass of the spike, and additional first order information should determine the location. As shown in Section 6, in 1D we have $a_{\operatorname{U}}=a_{\operatorname{E}}=2$. The estimates given in (25) and (26) in dimension $d=1$ predict that the adaptive energy will decay at a rate of $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})\lesssim\frac{1}{n}$ so long as the pixel size also decreases at a rate of $h\sim\frac{1}{n}$. To achieve these rates, we implement a refinement criterion from Lemma 6 with guarantee of $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n_{k}-1})-\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})\lesssim 2^{-k}$ using the estimates made in Section 6.3. We choose subspaces $\mathds{U}^{n}$ to approximately enforce $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n})\leq 2(\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]\vec{\varphi}_{n})),$ (62) i.e. the continuous gap is bounded by twice the discrete gap. In particular, note that for $\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\approx\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]$, $\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n})=\tfrac{1}{2}{\left\lVert\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\vec{\varphi}_{n}\right\rVert}^{2}+\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}\eta\vbox{\hbox{\leavevmode\resizebox{6.0pt}{}{$\cdot$}}}\vec{\varphi}_{n}=\frac{\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}}{\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]}\left(\frac{\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}}{\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]}\tfrac{1}{2}{\left\lVert\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]\vec{\varphi}_{n}\right\rVert}^{2}+\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]\eta\vbox{\hbox{\leavevmode\resizebox{6.0pt}{}{$\cdot$}}}\vec{\varphi}_{n}\right)\approx\frac{\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}}{\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]}\operatorname{E}^{\dagger}(\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]\vec{\varphi}_{n}).$ (63) Converting this into a spatial refinement criteria, recall $\frac{\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}}{\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]}\approx\frac{{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\mathsf{A}^{*}\vec{\varphi}_{n}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}}{{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\mathsf{\Pi}_{n}\mathsf{A}^{*}\vec{\varphi}_{n}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}}=\frac{\max_{\omega\in\mathds{M}^{n}}{\left\lVert\mathsf{A}^{*}\vec{\varphi}_{n}\right\rVert}_{L^{\infty}(\omega)}}{\max_{\omega\in\mathds{M}^{n}}|\mathsf{\Pi}_{n}\mathsf{A}^{*}\vec{\varphi}_{n}(\omega)|}\approx\max_{\omega\in\mathds{M}^{n}}\frac{{\left\lVert\mathsf{A}^{*}\vec{\varphi}_{n}\right\rVert}_{L^{\infty}(\omega)}}{|\mathsf{\Pi}_{n}\mathsf{A}^{*}\vec{\varphi}_{n}(\omega)|}$ (64) is the maximum ratio of second vs. zeroth order Taylor approximations of $\mathsf{A}^{*}\vec{\varphi}_{n}$ on pixel $\omega$. This was found to be an efficient method of selecting pixels for refinement using quantities which had already been computed. Note briefly that this greedy strategy directly targets uncertainty, refinements also happen outside of the support of $u_{n}$ to guarantee that this is representative of $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$. Such refinement is necessary to avoid discrete minimisers of $\operatorname{E}$ which are not global minimisers. Figure 1: Rates of continuous/discrete gap convergence for different LASSO algorithms with 128, 256, or 512 pixels. The “adaptive” method uses the proposed algorithm. Both “fixed” and “CBP” use standard FISTA with a uniform discretisation. Figure 2: Convergence plots for solving 1D problems with different algorithms. “Adaptive” methods use Algorithm 1 with fewer than 1024 pixels and the remaining methods use a uniform discretisation of 1024 pixels. Figure 3: Example reconstruction from the algorithms considered in Fig. 3. Pixel boundaries are indicated on the $x$-axis and the filtering method of Section 6.4 allows us to exclude the red shaded regions from $\operatorname{supp}(u^{*})$. Values on the $y$-axis are normalised to units of mass, i.e. a Dirac mass would have height 1. ##### Comparison of discretisation methods In Fig. 3 we compare the three core approaches: fixed uniform discretisation, adaptive discretisation, and CBP. In particular, we wish to observe their convergence properties as the number of pixels is allowed to grow. In each case we use a FISTA stepsize of $t_{n}=\frac{n+19}{20}$. The adaptive discretisation is started with one pixel and limited to 128, 256, or 512 pixels while the fixed and CBP discretisations have uniform discretisations with the maximum number of pixels. The main observations are: * • The adaptive scheme is much more efficient, in both examples the adaptive scheme with 128 pixels is at least as good as both fixed discretisations with 512 pixels. In fact, only a maximum of 214 pixels were needed by the adaptive method in either example. * • With Fourier kernels the uniform piecewise constant discretisation is more efficient than CBP but in the Gaussian case this is reversed. This suggests that the performance of CBP depends on the smoothness of $\mathsf{A}$. * • The discrete gaps for non-adaptive optimisation behave as is common for FISTA, initial convergence is polynomial until a locally linear regime activates Tao2016 . CBP is always slower to converge than the piecewise constant discretisation. * • The adaptive refinement criterion succeeds in keeping the continuous/discrete gaps close for all $n$, i.e. (62). It is not completely fair to judge CBP with the continuous gap because, although it generates a continuous representation, this continuous representation is not necessarily consistent with the discrete gap being optimised, unlike when discretised with finite element methods. On the other hand, this is still the intended interpretation of the algorithm and we have no more appropriate metric for success in this case. ##### Comparison of FISTA variants Fig. 3 compares many methods with either fixed or adaptive discretisations. Each adaptive scheme is allowed up to 1024 pixels and each uniform discretisation uses exactly 1024. An example of each reconstruction method is shown in Fig. 3. The adaptive method better identifies the support of $u^{*}$ and clearly localises pixels on that support. The reconstruction with uniform grid fails to provably identify the support of $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$, despite having found a qualitatively accurate discrete minimiser. The “Greedy FISTA” implementation was proposed by in Liang2018 and we include the adaptive variant despite a lack of convergence proof. The remaining FISTA algorithms use a FISTA time step of $t_{n}=\frac{n+a-1}{a}$ for the given value of $a$, as proposed in Chambolle2015 . In this example CBP used the greedy FISTA implementation which gave faster observed convergence. Fig. 3 compares the discrete gaps because it is the accurate metric for fixed discretisations, and for the adaptive discretisation it should also be an accurate predictor of the continuous gap. The main observations are: * • Each algorithm displays very similar convergence properties. The main difference is that the reconstructions with fixed discretisations accelerate after $10^{4}$-$10^{5}$ iterations. * • During the initial “slow” phase, adaptive and fixed discretisations appear to achieve very similar (discrete) convergence rates. The coarse-to-fine adaptivity is not slower than fixed discretisations in this regime. * • Lemma 6 accurately predicts the $\frac{1}{n}$ rate of the adaptive methods, mirrored in the fixed discretisations. This suggests that high-resolution discretisations are also initially limited by this $\frac{1}{n}$ rate before entering the asymptotic regime, consistent with (14). * • The fastest FISTA stepsize choice is consistently the greedy variant, although $a=20$ is very comparable. * • While each adaptive algorithm is allowed to use up to 1024 pixels, in Fig. 3 the most used was 235. ##### Comparison of fixed and adaptive discretisation Motivated by the findings in Fig. 3, we now look more closely at the performance of the $a=20$ and the greedy FISTA schemes. We have convergence results for the former, but the latter typically performs the best for non- adaptive optimisation and is never worse than $a=20$ in the adaptive setting. The question is whether it is faster/more efficient to use the proposed adaptive scheme, or to use a classical scheme at sufficiently high uniform resolution. The fixed discretisations use 1024 pixels (i.e. constant pixel size of $2^{-10}$ in Fig. 6) and the adaptive discretisation starts with two pixels with an upper limit of 1024. As expected, the fixed discretisation starts with a smaller continuous gap before plateauing to a sub-optimal gap around $\operatorname{E}=\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})+0.1$. Fig. 6 shows convergence of pixel size and continuous gap with respect to number of iterations. Fig. 6 shows the more practical attributes of continuous gap and number of pixels against execution time. We see that the adaptive discretisation is consistently capable of computing lower energies with fewer pixels and in less time than the uniform discretisation. The convergence behaviour is very consistent with respect to number of iterations. Suppose that the numerical aim is to find a function $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n}$ with $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})\leq 0.1$, all methods would converge after $O(10^{3})$ iterations, demonstrating some equivalence between the two FISTA algorithms. For $n\in[10^{3},10^{4}]$, in both problems, the adaptive schemes coincide with the fixed schemes in both energy and minimum pixel size. On the other hand, we also see that the adaptive scheme achieves this energy in almost an order of magnitude less time and fewer pixels. Figure 4: Continuous convergence of adaptive (coarse-to-fine pixel size) compared with uniform discretisation (constant pixel size) with respect to number of iterations. Figure 5: Continuous convergence of adaptive compared with uniform discretisation with respect to wall-clock time and total number of pixels (memory requirement). $\begin{aligned} J_{n}&=\\{(0,0),(0,1),(0,2),(1,2),(1,1)\\}\\\ \operatorname{leaf}(J_{n})&=\\{(0,2),(1,2),(1,1)\\}\\\ \mathds{M}^{n}&=\left\\{[0,\tfrac{1}{4}),[\tfrac{1}{4},\tfrac{1}{2}),[\tfrac{1}{2},1)\right\\}\end{aligned}$ $(0,0)$ $[0,1]$ $(0,1)$ $[0,\frac{1}{2}]$ $(0,2)$ $[0,\frac{1}{4}]$ $(1,2)$ $[\frac{1}{4},\frac{1}{2}]$ $(1,1)$ $[\frac{1}{2},1]$ Figure 6: Example tree representation of 1D wavelets. Left: nodes, leaves, and mesh of discretisation. Right: arrangement into a tree with index $(j,k)$ and corresponding support of wavelet $w_{j,k}$ underneath. ### 7.2 2D robust sparse wavelet reconstruction In this example we consider $\mathsf{A}$ to be a 2D Radon transform. In particular, the rows of $\mathsf{A}$ correspond to integrals over the sets $\mathds{X}^{I}_{i}$ where $\mathds{X}_{i}^{I}=\left\\{\vec{x}\in[-\tfrac{1}{2},\tfrac{1}{2}]^{2}\operatorname{\;s.t.\;}\vec{x}\vbox{\hbox{\leavevmode\resizebox{6.0pt}{}{$\cdot$}}}\begin{pmatrix}\cos\theta_{I}\\\ \sin\theta_{I}\end{pmatrix}\in\left[-\tfrac{1}{2}+\tfrac{i-1}{100},-\tfrac{1}{2}+\tfrac{i}{100}\right)\right\\},\quad\theta_{I}=\frac{180^{\circ}}{51}I$ (65) for $i=\in[100]$, $I\in[50]$. This is not exactly in the form analysed by Theorem C.1, only the sets $\\{\mathds{X}^{I}_{i}\operatorname{\;s.t.\;}i\in[100]\\}$ for each $I$ are disjoint, therefore we apply Theorem C.1 block-wise to estimate ${\left\lVert\mathsf{A}\right\rVert}_{L^{2}\to\ell^{2}}\leq\sqrt{\sum_{I\in[50]}\max_{i\in[100]}|\mathds{X}^{I}_{i}|}=\sqrt{\sum_{I\in[50]}\max_{i\in[100]}\int_{\mathds{X}^{I}_{i}}1\mathop{}\\!\mathrm{d}\vec{x}}=\sqrt{\sum_{I\in[50]}\max_{i\in[100]}\ (\mathsf{A}\mathds{1})_{i,I}}\;.$ (66) $\mathsf{A}$ is not smooth, therefore we can’t bound $|\mathsf{A}^{*}|_{C^{k}}$ for $k>0$, and so we must look to minimise over $\ell^{1}$ rather than $L^{1}$. The natural choice is to promote sparsity in a wavelet basis which can be rearranged into the form of (18): $\min_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}}\operatorname{f}(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]-\eta)+\mu{\left\lVert\mathsf{W}^{-1}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{\ell^{1}}=\min_{\widehat{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\in\ell^{1}(\mathds{R})}\operatorname{f}(\mathsf{A}\mathsf{W}\widehat{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}-\eta)+\mu{\left\lVert\widehat{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rVert}_{\ell^{1}}.$ (67) The minimisers are related by $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}=\mathsf{W}\widehat{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}^{*}$ and, for wavelet bases, $\mathsf{W}$ is orthonormal so ${\left\lVert\mathsf{A}\mathsf{W}\right\rVert}_{\ell^{2}\to\ell^{2}}={\left\lVert\mathsf{A}\right\rVert}_{L^{2}\to\ell^{2}}$. In this example we consider the smoothed robust fidelity Rosset2007 $\operatorname{f}(\vec{\varphi})=\sum_{i=1}^{m}\begin{cases}10^{-4}|\varphi_{i}|&|\varphi_{i}|\geq 10^{-4}\\\ \tfrac{1}{2}|\varphi_{i}|^{2}+\tfrac{1}{2}10^{-8}&\text{else}\end{cases}\approx 10^{-4}{\left\lVert\vec{\varphi}\right\rVert}_{\ell^{1}}.$ (68) From Section 6.3 we know that to track convergence and perform adaptive refinement, it is sufficient to accurately bound $|[\mathsf{W}^{\top}\mathsf{A}^{*}\vec{\varphi}_{n}]_{j}|$ for all $j\notin J_{n}$. If $\mathsf{W}$ is a wavelet transformation then its columns, $w_{j}\in L^{2}$, are simply the wavelets themselves and we can use the bound $|\left\langle w_{j},\mathsf{A}^{*}\vec{\varphi}_{n}\right\rangle|=\left|\left\langle w_{j},\mathds{1}_{\operatorname{supp}(w_{j})}\mathsf{A}^{*}\vec{\varphi}_{n}\right\rangle\right|\leq{\left\lVert\mathds{1}_{\operatorname{supp}(w_{j})}\mathsf{A}^{*}\vec{\varphi}_{n}\right\rVert}_{L^{2}}\leq{\left\lVert\mathds{1}_{\mathds{X}}\mathsf{A}^{*}\vec{\varphi}_{n}\right\rVert}_{L^{2}}$ (69) for all $\mathds{X}\supset\operatorname{supp}(w_{j})$. In the case of the Radon transform, we can compute the left-hand side explicitly for the finitely many $j\in J_{n}$, but we wish to use the right-hand side in a structured way to avoid computing the infinitely many $j\notin J_{n}$. To do this, we will take a geometrical perspective on the construction of wavelets to view them in a tree format. ##### Tree structure of wavelets Finite elements are constructed with a mesh which provided a useful tool for adaptive refinement in Section 6.3.3. For wavelets, we will associate a tree with every discretisation and the leaves of the tree correspond to a mesh. This perspective comes from the multi-resolution interpretation of wavelets. An example is seen in Fig. 6 for 1D Haar wavelets, $w_{j,k}(x)=\sqrt{2}^{k}\psi(2^{k}x-j)$ where $\psi=\mathds{1}_{[0,1)}-\mathds{1}_{[-1,0)}$. In higher dimensions, the only two things which change are the number of children ($2^{d}$ for non-leaves) and at each node you store the coefficients of $2^{d}-1$ wavelets. The support on each node is still a disjoint partition of unity consisting of regular cubes of side length $2^{-k}$ at level $k$. The only change in our own implementation is to translate the support to $[-\tfrac{1}{2},\tfrac{1}{2}]^{2}$. We briefly remark that the tree structuring of wavelets is not novel and appears more frequently in the Bayesian inverse problems literature Castillo2019 ; Kekkonen2021 . ##### Continuous gradient estimate In Section 7.1 we used the continuous gap as a measure for convergence, for wavelets we will use the continuous subdifferential. With the tree structure we can easily adapt the results of Section 6.3 to estimate subdifferentials (or function gaps). In particular, $\displaystyle{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\partial\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}$ $\displaystyle=\max\left({\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\partial_{n}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*},\max_{j\notin J_{n}}|\left\langle w_{j},\mathsf{A}^{*}\vec{\varphi}_{n}\right\rangle|-\mu\right)$ (70) $\displaystyle\leq\max\left({\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\partial_{n}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*},\max_{j\in\operatorname{leaf}(J_{n})}{\left\lVert\mathds{1}_{\operatorname{supp}(w_{j})}\mathsf{A}^{*}\vec{\varphi}_{n}\right\rVert}_{L^{2}}-\mu\right).$ (71) ##### Numerical results We consider two phantoms where the ground-truth is either a binary disc or the Shepp-Logan phantom. Both examples are corrupted with $2\text{\,}\mathrm{\char 37\relax}$ Laplace distributed noise. This is visualised in Fig. 9. All optimisations shown are spatially adaptive using Haar wavelets and initialised with $\mathds{U}^{0}=\\{x\mapsto c\operatorname{\;s.t.\;}c\in\mathds{R}\\}$. The gradient metric shown throughout is the $\ell^{\infty}$ norm. Motivated by (71), the spatial adaptivity is chosen to refine nodes $j\in\operatorname{leaf}(J_{n})$ to ensure that ${\left\lVert\mathds{1}_{\operatorname{supp}(w_{j})}\mathsf{A}^{*}\vec{\varphi}_{n}\right\rVert}_{L^{2}}-\mu\leq 10{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\partial_{n}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{*}$ for all $j$ and $n$ (i.e. so that the continuous gradient is less than 10 times the discrete gradient). We do not expect wavelet regularisation to have state-of-the-art performance in the examples of Fig. 9. What they demonstrate is the preference Haar wavelets have to align large discontinuities with a coarse grid, even when the discretisation is allowed to be as fine as necessary. There is an average of $2\cdot 10^{6}$ wavelet coefficients in each discretised reconstruction, although the higher frequencies have much smaller intensities. In limited data scenarios, wavelet regularisation automatically selects a local “resolution” which reflects the quality of data. Particularly in the Shepp-Logan reconstruction, we see that the outer ring is detected with a finer precision than the dark interior ellipses. The first numerical results shown in Fig. 9 compare the same adaptive FISTA variants as shown in Fig. 3. In these examples we see that the greedy FISTA and the $a=20$ algorithms achieve almost linear convergence while $a=2$ is significantly slower. Interestingly, in both examples the $a=20$ variant uses half as many wavelets as the Greedy variant, and therefore converges slightly faster in time. Figure 7: Phantoms, data and reconstructions for wavelet-sparse tomography optimisation. Both examples are corrupted with $2\text{\,}\mathrm{\char 37\relax}$ Laplace distributed noise. Figure 8: Convergence of different implementations of Algorithm 1 with an unlimited number of pixels for sparse wavelet optimisation. Figure 9: Example images from STORM dataset. ### 7.3 2D continuous LASSO Our final application is a super-resolution/de-blurring inverse problem from biological microscopy. In mathematical terms, the observed data is a large number of sparse images which are corrupted by blurring and a large amount of noise, examples are seen in Fig. 9. The task is to compute the centres of the spikes of signal in each image and then re-combine into a single super- resolved image, as in Fig. 11. This technique is referred to as _Single Molecule Localisation Microscopy_ (SMLM), of which we consider the specific example of _Stochastic Optical Reconstruction Microscopy_ (STORM). Readers are directed to the references Sage2015 ; Sage2019 ; Schermelleh2019 for further details. The LASSO formulation ($\operatorname{f}(\cdot)=\frac{1}{2}{\left\lVert\cdot\right\rVert}_{\ell^{2}}^{2}$) has previously been shown to be effective in the context of STORM Huang2017 ; Denoyelle2019 . Here we use a simulated dataset provided as part of the 2016 SMLM challenge222http://bigwww.epfl.ch/smlm/challenge2016/datasets/MT4.N2.HD/Data/data.html for benchmarking software in this application. The corresponding LASSO formulation is $(\mathsf{A}\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])_{i}=(2\pi\sigma^{2})^{-1}\int_{[0,6.4]^{2}}\exp\left(-\frac{1}{2\sigma^{2}}\left|\vec{x}-\Delta\begin{pmatrix}i_{1}+\tfrac{1}{2}&i_{2}+\tfrac{1}{2}\end{pmatrix}^{\top}\right|^{2}\right)\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0](\vec{x})\mathop{}\\!\mathrm{d}\vec{x},\qquad\sigma=0.2,\ \Delta=0.1$ (72) for $i_{1},i_{2}=1,2,\ldots,64$, $\mathds{U}=\mathcal{M}([0,6.4]^{2})$ and $\mathds{H}=L^{2}([0,6.4]^{2})$ with lengths in $\text{\,}\mathrm{\SIUnitSymbolMicro m}$. 3020 frames are provided, examples of which are shown in Fig. 9. To process this dataset, image intensities were normalised to $[0,1]$ then a constant was subtracted to approximate 0-mean noise. The greedy FISTA algorithm was used for optimisation with $\mu=0.15$, $10^{3}$ iterations, and a maximum of $10^{5}$ pixels per image. Finally, all the reconstructions were summed and the result shown in Fig. 11. The adaptive scheme used fewer than $10^{4}$ pixels per frame, a fixed discretisation with equivalent resolution of $1.3\text{\,}\mathrm{nm}$ would have required more than $3\cdot 10^{6}$ per frame. LASSO is compared with ThunderSTORM Ovesny2014 , a popular ImageJ plugin Schindelin2012 which finds the location of signal using Fourier filtering. The performance of ThunderSTORM was rated very highly in the initial SMLM challenge Sage2015 . Both methods compared here demonstrate the key structures of the reconstruction, however, both are sensitive to tuning parameters. In this examples, LASSO has possibly recovered too little signal and ThunderSTORM contains spurious signal. Fig. 11 shows various convergence metrics for the adaptive reconstructions. The magenta line in the first panel shows that the continuous gap converges slightly faster than the $n^{-2/3}$ predicted by (26) in dimension $d=2$. In this example we also implement the suggestion of Section 6.4 to remove pixels outside of the support of $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$. From (56), any pixel $\omega\in\mathds{M}^{n}$ satisfying $\IfEqCase{2}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k2]_{0}{\left\lVert\mathsf{\Pi}_{n}\mathsf{A}^{*}\vec{\varphi}_{n}\right\rVert}_{L^{\infty}(\omega)}\leq(1-\operatorname{threshold}_{n})\mu$ (73) guarantees that $\omega\cap\operatorname{supp}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})=\emptyset$. This threshold is plotted in red in the first panel of Fig. 11. Once the value becomes less than 1, we can start reducing the number of pixels instead of continual refinement. We see that the resolution decreases steadily (second panel), but the total number of pixels (final panel) stops increasing after around 30 iterations. Figure 10: Convergence of adaptive FISTA for STORM dataset. Lines indicate the median value over 3020 STORM frames. Shaded regions indicate the $25\text{\,}\mathrm{\char 37\relax}75\text{\,}\mathrm{\char 37\relax}$ interquartile range. Pixel width is scaled $[0,1]$ rather than $[0,$6.4\text{\,}\mathrm{\SIUnitSymbolMicro m}$]$. Figure 11: Processed results of the STORM dataset. Top left: LASSO optimisation with Algorithm 1. Top right: Comparison with ThunderSTORM plugin. Bottom: Average data, no super-resolution or de-blurring. ## 8 Conclusions and outlook In this work we have proposed a new adaptive variant of FISTA and provided convergence analysis. This algorithm allows FISTA to be applied outside of the classical Hilbert space setting, still with a guaranteed rate of convergence. We have presented several numerical examples where convergence with the refining discretisation is at least as fast as a uniform discretisation, although more efficient with regards to both memory and computation time. In 1D we see good agreement with the theoretical rate. This rate also seems to be a good predictor for all variants of FISTA tested, although this is yet to be proven. Even the classical methods with a fixed discretisation are initially limited to the slower adaptive rate for small $n$. The results in 2D are similar, all tested FISTA methods converge at least at the guaranteed rate. The wavelet example was most impressive, achieving nearly linear convergence in energy. This is similar to the behaviour for classical FISTA although it is also yet to be formally proven. An interesting observation over all of the adaptive LASSO examples is that the standard oscillatory behaviour of FISTA has not occurred. With the monotone gaps plotted, oscillatory convergence should correspond to a piecewise constant descending gap. Either this behaviour only emerges for larger $n$, or the adaptivity provides a dampening effect for this oscillation. Moving forward, it would be interesting to see how far the analysis extends to other optimisation algorithms. Other variants of FISTA, such as the “greedy” implementation used here or the traditional Forward-Backward algorithm, should also be receptive to the analysis performed here. Furthermore, it would also interesting to attempt to replicate this refinement argument to extend the primal-dual algorithm Chambolle2011 or the Douglas-Rachford algorithm Douglas1956 . ###### Acknowledgements. R.T. acknowledges funding from EPSRC grant EP/L016516/1 for the Cambridge Centre for Analysis, and the ANR CIPRESSI project grant ANR-19-CE48-0017-01 of the French Agence Nationale de la Recherche. Most of this work was done while A.C. was still in CMAP, CNRS and Ecole Polytechnique, Institut Polytechnique de Paris, Palaiseau, France. Both authors would like to thank the anonymous reviewers who put in so much effort to improving this work. * Data availability statement The synthetic STORM dataset was provided as part of the 2016 SMLM challenge, http://bigwww.epfl.ch/smlm/challenge2016/datasets/MT4.N2.HD/Data/data.html. The remaining examples used in this work can be generated with the supplementary code, https://github.com/robtovey/2020SpatiallyAdaptiveFISTA. * Conflict of interest The authors have no conflicts of interest to declare which are relevant to the content of this article. ## References * (1) Alamo, T., Limon, D., Krupa, P.: Restart fista with global linear convergence. In: 2019 18th European Control Conference (ECC), pp. 1969–1974. IEEE (2019) * (2) Aujol, J.F., Dossal, C.: Stability of over-relaxations for the forward-backward algorithm, application to fista. SIAM Journal on Optimization 25(4), 2408–2433 (2015) * (3) Beck, A., Teboulle, M.: A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM Journal on Imaging Sciences 2(1), 183–202 (2009) * (4) Bonnefoy, A., Emiya, V., Ralaivola, L., Gribonval, R.: Dynamic screening: Accelerating first-order algorithms for the lasso and group-lasso. IEEE Transactions on Signal Processing 63(19), 5121–5132 (2015) * (5) Boyd, N., Schiebinger, G., Recht, B.: The alternating descent conditional gradient method for sparse inverse problems. SIAM Journal on Optimization 27(2), 616–639 (2017) * (6) Boyer, C., Chambolle, A., Castro, Y.D., Duval, V., De Gournay, F., Weiss, P.: On representer theorems and convex regularization. SIAM Journal on Optimization 29(2), 1260–1281 (2019) * (7) Bredies, K., Pikkarainen, H.K.: Inverse problems in spaces of measures. ESAIM: Control, Optimisation and Calculus of Variations 19(1), 190–218 (2013) * (8) Castillo, I., Rockova, V.: Multiscale analysis of bayesian cart. University of Chicago, Becker Friedman Institute for Economics Working Paper (2019-127) (2019) * (9) Catala, P., Duval, V., Peyré, G.: A low-rank approach to off-the-grid sparse superresolution. SIAM Journal on Imaging Sciences 12(3), 1464–1500 (2019) * (10) Chambolle, A., Dossal, C.: On the convergence of the iterates of the “fast iterative shrinkage/thresholding algorithm”. Journal of Optimization Theory and Applications 166(3), 968–982 (2015) * (11) Chambolle, A., Pock, T.: A First-Order Primal-Dual Algorithm for Convex Problems with Applications to Imaging. Journal of Mathematical Imaging and Vision 40(1), 120–145 (2011). DOI 10.1007/s10851-010-0251-1 * (12) De Castro, Y., Gamboa, F., Henrion, D., Lasserre, J.B.: Exact solutions to super resolution on semi-algebraic domains in higher dimensions. IEEE Transactions on Information Theory 63(1), 621–630 (2016) * (13) Denoyelle, Q., Duval, V., Peyré, G., Soubies, E.: The sliding frank–wolfe algorithm and its application to super-resolution microscopy. Inverse Problems 36(1), 014001 (2019) * (14) Douglas, J., Rachford, H.H.: On the numerical solution of heat conduction problems in two and three space variables. Transactions of the American Mathematical Society 82(2), 421–439 (1956) * (15) Duval, V., Peyré, G.: Sparse spikes super-resolution on thin grids i: the lasso. Inverse Problems 33(5), 055008 (2017) * (16) Duval, V., Peyré, G.: Sparse spikes super-resolution on thin grids ii: the continuous basis pursuit. Inverse Problems 33(9), 095008 (2017) * (17) Ekanadham, C., Tranchina, D., Simoncelli, E.P.: Recovery of sparse translation-invariant signals with continuous basis pursuit. IEEE transactions on signal processing 59(10), 4735–4744 (2011) * (18) El Ghaoui, L., Viallon, V., Rabbani, T.: Safe feature elimination in sparse supervised learning. Tech. Rep. UCB/EECS-2010–126, EECS Department, University of California, Berkeley (2010) * (19) Hiriart-Urruty, J.B., Lemaréchal, C.: Convex Analysis and Minimization Algorithms II: Advanced Theory and Bundle Methods, vol. 305. Springer-Verlag, Berlin, Heidelberg (1993) * (20) Huang, J., Sun, M., Ma, J., Chi, Y.: Super-resolution image reconstruction for high-density three-dimensional single-molecule microscopy. IEEE Transactions on Computational Imaging 3(4), 763–773 (2017) * (21) Jiang, K., Sun, D., Toh, K.C.: An inexact accelerated proximal gradient method for large scale linearly constrained convex sdp. SIAM Journal on Optimization 22(3), 1042–1064 (2012) * (22) Kekkonen, H., Lassas, M., Saksman, E., Siltanen, S.: Random tree besov priors–towards fractal imaging. arXiv preprint arXiv:2103.00574 (2021) * (23) Liang, J., Fadili, J., Peyré, G.: Activity identification and local linear convergence of forward–backward-type methods. SIAM Journal on Optimization 27(1), 408–437 (2017) * (24) Liang, J., Schönlieb, C.B.: Improving fista: Faster, smarter and greedier. arXiv preprint arXiv:1811.01430 (2018) * (25) Ndiaye, E., Fercoq, O., Gramfort, A., Salmon, J.: Gap safe screening rules for sparsity enforcing penalties. The Journal of Machine Learning Research 18(1), 4671–4703 (2017) * (26) Nesterov, Y.: Introductory Lectures on Convex Optimization: A Basic Course. Kluwer Academic Publishers Boston, Dordrecht, London (2004) * (27) Ovesnỳ, M., Křížek, P., Borkovec, J., Švindrych, Z., Hagen, G.M.: Thunderstorm: a comprehensive imagej plug-in for palm and storm data analysis and super-resolution imaging. Bioinformatics 30(16), 2389–2390 (2014) * (28) Parpas, P.: A multilevel proximal gradient algorithm for a class of composite optimization problems. SIAM Journal on Scientific Computing 39(5), S681–S701 (2017) * (29) Poon, C., Keriven, N., Peyré, G.: The geometry of off-the-grid compressed sensing. arXiv preprint arXiv:1802.08464 (2018) * (30) Rosset, S., Zhu, J.: Piecewise linear regularized solution paths. The Annals of Statistics pp. 1012–1030 (2007) * (31) Sage, D., Kirshner, H., Pengo, T., Stuurman, N., Min, J., Manley, S., Unser, M.: Quantitative evaluation of software packages for single-molecule localization microscopy. Nature Methods 12(8), 717–724 (2015) * (32) Sage, D., Pham, T.A., Babcock, H., Lukes, T., Pengo, T., Chao, J., Velmurugan, R., Herbert, A., Agrawal, A., Colabrese, S., et al.: Super-resolution fight club: assessment of 2d and 3d single-molecule localization microscopy software. Nature Methods 16(5), 387–395 (2019) * (33) Schermelleh, L., Ferrand, A., Huser, T., Eggeling, C., Sauer, M., Biehlmaier, O., Drummen, G.P.: Super-resolution microscopy demystified. Nature cell biology 21(1), 72–84 (2019) * (34) Schindelin, J., Arganda-Carreras, I., Frise, E., Kaynig, V., Longair, M., Pietzsch, T., Preibisch, S., Rueden, C., Saalfeld, S., Schmid, B., et al.: Fiji: an open-source platform for biological-image analysis. Nature Methods 9(7), 676–682 (2012) * (35) Schmidt, M., Roux, N.L., Bach, F.R.: Convergence rates of inexact proximal-gradient methods for convex optimization. In: Advances in Neural Information Processing Systems, pp. 1458–1466 (2011) * (36) Strang, G.: Approximation in the finite element method. Numerische Mathematik 19(1), 81–98 (1972) * (37) Tao, S., Boley, D., Zhang, S.: Local linear convergence of ista and fista on the lasso problem. SIAM Journal on Optimization 26(1), 313–336 (2016) * (38) Unser, M., Fageot, J., Gupta, H.: Representer theorems for sparsity-promoting $\ell^{1}$ regularization. IEEE Transactions on Information Theory 62(9), 5167–5180 (2016) * (39) Villa, S., Salzo, S., Baldassarre, L., Verri, A.: Accelerated and inexact forward-backward algorithms. SIAM Journal on Optimization 23(3), 1607–1633 (2013) * (40) Yu, J., Lai, R., Li, W., Osher, S.: A fast proximal gradient method and convergence analysis for dynamic mean field planning. arXiv preprint arXiv:2102.13260 (2021) ## Appendix A Proofs for FISTA convergence This section contains all of the statements and proofs of the results contained in Section 4. Recall that the subsets $\mathds{U}^{n}\subset\mathds{H}$ satisfy (10). ### A.1 Proofs for Step 3 ###### Theorem A.1 (Lemma 2) thm: one step FISTA $t_{n}^{2}(\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\operatorname{E}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}))-(t_{n}^{2}-t_{n})(\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n-1})-\operatorname{E}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}))\leq\tfrac{1}{2}\left[{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1}\right\rVert}^{2}-{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}\right\rVert}^{2}\right]+\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}-\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1},\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rangle.$ (74) ###### Proof Modifying (Chambolle2015, , Thm 3.2), for $n\geq 1$ we apply Lemma 1 with $\overline{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}=\overline{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}_{n-1}$ and $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]=(1-\frac{1}{t_{n}})\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n-1}+\frac{1}{t_{n}}\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}$. By (10), $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n-1}\in\mathds{U}^{n}$ is convex so $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]\in\mathds{U}^{n}$. This gives $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})+\tfrac{1}{2}{\left\lVert\tfrac{1}{t_{n}}\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}-\tfrac{1}{t_{n}}\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rVert}^{2}\leq\operatorname{E}\left((1-\tfrac{1}{t_{n}})\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n-1}+\tfrac{1}{t_{n}}\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right)+\tfrac{1}{2}{\left\lVert\tfrac{1}{t_{n}}\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1}-\tfrac{1}{t_{n}}\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rVert}^{2}.$ (75) By the convexity of $\operatorname{E}$, this reduces to $\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-\operatorname{E}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n})-(1-\tfrac{1}{t_{n}})[\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n-1})-\operatorname{E}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n})]\leq\tfrac{1}{2t_{n}^{2}}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1}-\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rVert}^{2}-\tfrac{1}{2t_{n}^{2}}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}-\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rVert}^{2}=\tfrac{1}{2t_{n}^{2}}\left[{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1}\right\rVert}^{2}-{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}\right\rVert}^{2}\right]+\tfrac{1}{t_{n}^{2}}\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}-\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1},\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rangle.$ (76) Multiplying through by $t_{n}^{2}$ gives the desired inequality. ∎ ###### Theorem A.2 (Theorem 4.1) thm: mini FISTA convergence $\Paste{thm:eq:miniFISTAconvergence}$ (77) ###### Proof Theorem A.2 is just a summation of (74) over all $n=1,\ldots,N$. To see this: first add and subtract $\inf_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{H}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$ to each term on the left-hand side to convert $\operatorname{E}$ to $\operatorname{E}_{0}$, then move $\operatorname{E}_{0}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n})$ to the right- hand side. Now (74) becomes $t_{n}^{2}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})-(t_{n}^{2}-t_{n})\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n-1})\leq t_{n}\operatorname{E}_{0}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n})+\tfrac{1}{2}\left[{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1}\right\rVert}^{2}-{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}\right\rVert}^{2}\right]+\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}-\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1},\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rangle.$ (78) Summing this inequality from $n=1$ to $n=N$ gives $t_{N}^{2}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})+\sum_{n=1}^{N-1}(\underbrace{t_{n}^{2}-t_{n+1}^{2}+t_{n+1}}_{=\rho_{n}})\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{n})\leq\frac{{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{0}\right\rVert}^{2}-{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{N}\right\rVert}^{2}}{2}+\sum_{n=1}^{N}t_{n}\operatorname{E}_{0}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n})+\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}-\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1},\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rangle.$ (79) The final step is to flip the roles of $\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}$/$\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}$ in the final inner product term. Re-writing the right-hand side gives $\sum_{n=1}^{N}\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n}-\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1},\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rangle=\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{N},\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{N}\right\rangle-\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{0},\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{0}\right\rangle+\sum_{n=1}^{N}\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1},\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n-1}-\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rangle.$ (80) Noting that $\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{0}=\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{0}$, the previous two equations combine to prove the statement of Theorem A.2. ∎ The following lemma is used to produce a sharper estimate on sequences $t_{n}$. ###### Lemma 10 If $\rho_{n}=t_{n}^{2}-t_{n+1}^{2}+t_{n+1}\geq 0$, $t_{n}\geq 1$ for all $n\in\mathds{N}$ then $t_{n}\leq n-1+t_{1}$. ###### Proof This is trivially true for $n=1$. Suppose true for $n-1$, the condition on $\rho_{n-1}$ gives $t_{n}^{2}-t_{n}\leq t_{n-1}^{2}\leq(n-2+t_{1})^{2}=(n-1+t_{1})^{2}-2(n-1+t_{1})+1.$ (81) Assuming the contradiction, if $t_{n}>n-1+t_{1}$ then the above equation simplifies to $n-1+t_{1}<1$. However, $t_{1}\geq 1$ implying that $n<1$ which completes the contradiction. ∎ ###### Lemma 11 (Lemma 3) thm: mini exponential FISTA convergence $\Paste{thm:eq:miniexponentialFISTAconvergence}$ (82) thm:end: mini exponential FISTA convergence ###### Proof This is just a telescoping of the right-hand side of (77) with the introduction of $n_{k}$ and simplification $\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}=\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}$, $\tfrac{1}{2}{\left\lVert\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{N}\right\rVert}^{2}+\sum^{N}_{n=1}t_{n}\operatorname{E}_{0}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n})+\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n-1},\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n-1}-\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]_{n}\right\rangle=\tfrac{1}{2}{\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K}\right\rVert}^{2}+\sum_{n=n_{K}}^{N}t_{n}\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K})+\sum_{k=1}^{K}\sum_{n=n_{k-1}}^{n_{k}-1}t_{n}\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1})+\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{k}-1},\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right\rangle.$ (83) By Lemma 10, $t_{n}\leq n$ so we can further simplify $\sum_{n=\IfEqCase{3}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k3]}^{\IfEqCase{4}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k4]-1}t_{n}\leq\sum_{n=\IfEqCase{3}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k3]}^{\IfEqCase{4}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k4]-1}n=(\IfEqCase{4}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k4]-\IfEqCase{3}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k3])\frac{\IfEqCase{4}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k4]-1+\IfEqCase{3}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k3]}{2}\leq\frac{\IfEqCase{4}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k4]^{2}-\IfEqCase{3}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k3]^{2}}{2}$ to get the required bound. ∎ ### A.2 Proof for Step 4 ###### Lemma 12 (Lemma 4) thm: sufficiently fast ###### Proof Starting from Lemma 11 we have $\displaystyle t_{N}^{2}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})+\tfrac{1}{2}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{N}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K}\right\rVert}^{2}$ $\displaystyle\leq C+\frac{{\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K}\right\rVert}^{2}}{2}+\frac{(N+1)^{2}-n_{K}^{2}}{2}\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K})$ $\displaystyle\hskip 70.0pt+\sum_{k=1}^{K}\frac{n_{k}^{2}-n_{k-1}^{2}}{2}\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1})+\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{k}-1},\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k+1}\right\rangle$ (84) $\displaystyle\leq C+\frac{{\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K}\right\rVert}^{2}}{2}+\frac{n_{K+1}^{2}}{2}\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K})$ $\displaystyle\hskip 40.0pt+\sum_{k=1}^{K}\frac{n_{k}^{2}}{2}\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1})+\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{k}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1}+\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1},{\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}}\right\rangle.$ (85) The inductive step now depends on the value of $a_{\operatorname{U}}$. Case $a_{\operatorname{U}}>1$: We simplify the inequality $\displaystyle t_{N}^{2}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})+\tfrac{1}{2}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{N}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K}\right\rVert}^{2}$ $\displaystyle\lesssim a_{\operatorname{U}}^{2K}+n_{K+1}^{2}a_{\operatorname{E}}^{-K}+\sum_{k=1}^{K}n_{k}^{2}a_{\operatorname{E}}^{-k}+a_{\operatorname{U}}^{k}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{k}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1}\right\rVert}+a_{\operatorname{U}}^{2k}$ (86) $\displaystyle\leq C_{1}\left[a_{\operatorname{U}}^{2K+2}+\sum_{k=1}^{K}a_{\operatorname{U}}^{2k}+a_{\operatorname{U}}^{k}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{k}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1}\right\rVert}\right]$ (87) for some $C_{1}>C$. Choose $C_{2}\geq{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{1}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{1-1}\right\rVert}a_{\operatorname{U}}^{-1}$ such that $\frac{1}{2}C_{2}^{2}\geq\frac{C_{1}}{a_{\operatorname{U}}^{2}-1}(C_{2}+a_{\operatorname{U}}^{2}).$ (88) Assume ${\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{k}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1}\right\rVert}\leq C_{2}a_{\operatorname{U}}^{k}$ for $1\leq k\leq K$ (trivially true for $K=1$), then for $N=n_{K+1}-1$ we have $\displaystyle\tfrac{1}{2}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{K+1}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K}\right\rVert}^{2}$ $\displaystyle\leq C_{1}\left[a_{\operatorname{U}}^{2K+2}+\sum_{k=1}^{K}a_{\operatorname{U}}^{2k}+a_{\operatorname{U}}^{k}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{k}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1}\right\rVert}\right]$ (89) $\displaystyle\leq C_{1}\left[a_{\operatorname{U}}^{2K+2}+(1+C_{2})\frac{a_{\operatorname{U}}^{2K+2}}{a_{\operatorname{U}}^{2}-1}\right]$ (90) $\displaystyle\leq\frac{C_{1}a_{\operatorname{U}}^{2K+2}}{a_{\operatorname{U}}^{2}-1}\left(a_{\operatorname{U}}^{2}+C_{2}\right)\leq\tfrac{1}{2}(C_{2}a_{\operatorname{U}}^{K+1})^{2}.$ (91) Case $a_{\operatorname{U}}=1$: Denote $\IfEqCase{4}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k4]_{k}={\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k+1}\right\rVert}$ and note that ${\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1}\right\rVert}\leq{\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{0}\right\rVert}+\sum_{0}^{\infty}\IfEqCase{4}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k4]_{k}\lesssim 1$. We therefore bound $\displaystyle t_{N}^{2}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})+\tfrac{1}{2}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{N}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K}\right\rVert}^{2}$ $\displaystyle\lesssim 1+n_{K+1}^{2}a_{\operatorname{E}}^{-K}+\sum_{k=1}^{K}n_{k}^{2}a_{\operatorname{E}}^{-k}+({\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{k}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1}\right\rVert}+1)b_{k}$ (92) $\displaystyle\leq C_{1}\left[1+\sum_{k=1}^{K}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{k}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1}\right\rVert}\IfEqCase{4}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k4]_{k-1}\right]$ (93) for some $C_{1}>0$. Choose $C_{2}\geq\frac{{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{1}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{1-1}\right\rVert}}{\sum_{0}^{\infty}\IfEqCase{4}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k4]_{k}}$ such that $\frac{1}{2}C_{2}^{2}\geq C_{1}\left(1+C_{2}\sum_{0}^{\infty}\IfEqCase{4}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k4]_{k}\right).$ (94) Assume ${\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{k}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1}\right\rVert}\leq C_{2}$ for $1\leq k\leq K$ (trivially true for $K=1$), then for $N=n_{K+1}-1$ we have $\tfrac{1}{2}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{K+1}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K}\right\rVert}^{2}\leq C_{1}\left[1+\sum_{k=1}^{K}{\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{k}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1}\right\rVert}\IfEqCase{4}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k4]_{k-1}\right]\leq C_{1}\left(1+C_{2}\sum_{0}^{\infty}\IfEqCase{4}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k4]_{k}\right)\leq\frac{C_{2}^{2}}{2}$ (95) In both cases, the induction on ${\left\lVert\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]_{n_{K+1}-1}-\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{K}\right\rVert}$ holds for all $K$, and we have $t_{N}^{2}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})\leq\frac{1}{2}C_{2}^{2}a_{\operatorname{U}}^{2K}$ for all $N<n_{K}-1$. ∎ ### A.3 Proof for Step 5 ###### Lemma 13 (Lemma 5) thm: sufficiently slow ###### Proof The proof is direct computation, note that $(a_{\operatorname{E}}a_{\operatorname{U}}^{2})^{\kappa}=\exp\left(\kappa\log(a_{\operatorname{E}}a_{\operatorname{U}}^{2})\right)=\exp(\log a_{\operatorname{U}}^{2})=a_{\operatorname{U}}^{2},$ (96) therefore $a_{\operatorname{U}}^{2K}=\left((a_{\operatorname{E}}a_{\operatorname{U}}^{2})^{K}\right)^{\kappa}\lesssim n_{K}^{2\kappa}\leq N^{2\kappa},$ (97) so $\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})\lesssim N^{-2(1-\kappa)}$ as required. ∎ ### A.4 Proofs for Step 6 ###### Theorem A.3 (Theorem 4.3) thm: stronger exponential FISTA convergence ###### Proof Let $C>0$ satisfy $n_{k}^{2}\leq Ca_{\operatorname{E}}^{k}a_{\operatorname{U}}^{2k}$ for each $k\in\mathds{N}$. Fix $N>C$ and choose $k$ such that $Ca_{\operatorname{E}}^{k-1}a_{\operatorname{U}}^{2k-2}\leq N<Ca_{\operatorname{E}}^{k}a_{\operatorname{U}}^{2k}$. By construction, and using the equality from (96), we have $\min_{n\leq N}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]_{N})\leq\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k-1})\lesssim a_{\operatorname{E}}^{-k}=(a_{\operatorname{E}}a_{\operatorname{U}}^{2})^{-k(1-\kappa)}<C^{\kappa-1}N^{-2(1-\kappa)}$ (98) as required. ∎ ###### Lemma 14 (Lemma 6) thm: practical refinement criteria ###### Proof The conditions for $a_{\operatorname{U}}$ in Definition 1 are already met, it remains to be shown that $\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})\leq Ca_{\operatorname{E}}^{-k}$ for some fixed $C>0$. For cases (3) and (4), fix $R>0$ such that both $\\{\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\\}_{k\in\mathds{N}}$ and the sublevel set $\\{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}\operatorname{\;s.t.\;}\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])\leq 1+\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]\\}$ are contained in the ball of radius $R$. Any minimising sequences of $\operatorname{E}$ in $\mathds{U}$ or $\widetilde{\mathds{U}}^{k}$ are contained in this ball. We can therefore compute $C$ in each case: * (1) $\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})\leq\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]a_{\operatorname{E}}^{-k}$, so $C=\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]$ suffices. * (2) $\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})\leq\operatorname{E}_{0}(\widetilde{\mathds{U}}^{k})+\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]a_{\operatorname{E}}^{-k}\leq(a_{\operatorname{E}}+1)\beta a_{\operatorname{E}}^{-k}$, so $C=(a_{\operatorname{E}}+1)\beta$ suffices. * (3) $\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})-\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])\leq\inf_{\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]\in\partial\operatorname{E}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})}\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1],\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rangle\leq 2R\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]a_{\operatorname{E}}^{-k}$ for any $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}$ with ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}\leq R$. Maximising over $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]$ gives $C=2R\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]$ * (4) $\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})-\operatorname{E}_{0}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])\leq\inf_{\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1]\in\partial\operatorname{E}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})}\left\langle\IfEqCase{1}{{0}{u}{1}{v}{2}{w}}[u1],\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rangle\leq 2R\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]a_{\operatorname{E}}^{-k}$ for any $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\widetilde{\mathds{U}}^{k}$ with ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}\leq R$, so $\operatorname{E}_{0}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})\leq\operatorname{E}_{0}(\widetilde{\mathds{U}}^{k})+2R\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]a_{\operatorname{E}}^{-k}$ and $C=(1+2R)\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]$. This completes the requirements of Definition 1. ∎ ## Appendix B Proof of Theorem 5.1 First we recall the setting of Definition 2, fix: $p\geq 0$, $q\in[1,\infty]$, $h\in(0,1)$, $N\in\mathds{N}$, connected and bounded domain $\Omega\subset\mathds{R}^{d}$, and $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\in\operatorname*{argmin}_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\mathds{U}}\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0])$. We assume that $\mathds{H}=L^{2}(\Omega)$, ${\left\lVert\cdot\right\rVert}_{q}\lesssim{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$, and there exist spaces $(\widetilde{\mathds{U}}^{k})_{k\in\mathds{N}}$ with $\widetilde{\mathds{U}}^{k}\subset\mathds{U}$ containing a sequence $(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\in\widetilde{\mathds{U}}^{k})_{k\in\mathds{N}}$ such that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}\lesssim h^{kp}$, c.f. (16). Furthermore, there exists constant $c_{\alpha}>0$ and meshes $\mathds{M}^{k}$ such that: $\displaystyle\exists\omega_{0}\subset\Omega\quad\text{such that}\quad\forall\omega\in\mathds{M}^{k}\quad\exists(\IfEqCase{0}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k0]_{\omega},\vec{\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]}_{\omega})\in\mathds{R}^{d\times d}\times\mathds{R}^{d}\quad\text{such that}\quad\vec{x}\in\omega_{0}\iff\IfEqCase{0}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k0]_{\omega}\vec{x}+\vec{\IfEqCase{1}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k1]}_{\omega}\in\omega,\qquad\text{ and}$ (99) $\displaystyle\forall(\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]},\omega)\in\widetilde{\mathds{U}}^{k}\times\mathds{M}^{k},\quad\exists\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\widetilde{\mathds{U}}^{0}\quad\text{such that}\quad\operatorname{det}(\alpha_{\omega})\geq c_{\alpha}h^{kd}\quad\text{and}\quad\forall\vec{x}\in\omega_{0},\ \IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0](\vec{x})=\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}(\alpha_{\omega}\vec{x}+\vec{\beta}_{\omega}).$ (100) In this section, these assumptions will be summarised simply by saying that $\mathds{H}$ and $(\widetilde{\mathds{U}}^{k})_{k\in\mathds{N}}$ satisfy Definition 2. We prove Theorem 5.1 as a consequence of Lemma 7, namely we compute exponents $p^{\prime},q^{\prime}$ with $a_{\operatorname{U}}=h^{-q^{\prime}}$ and $a_{\operatorname{E}}=h^{-p^{\prime}}$. These values are computed as the result of the following three lemmas. The first, Lemma 15, is a quantification of the equivalence between $L^{q}$ and $L^{2}$ norms on general sub-spaces. Lemma 16 applies this result to finite-element spaces to compute the value of $q^{\prime}$. Finally, Lemma 17 then performs the computations for $p^{\prime}$ depending on the smoothness properties of $\operatorname{E}$. ###### Lemma 15 (Equivalence of norms for fixed $k$) Suppose $\mathds{H}=L^{2}(\Omega)$ for some connected, bounded domain $\Omega\subset\mathds{R}^{d}$ and ${\left\lVert\cdot\right\rVert}_{q}\leq C{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ for some $q\in[1,\infty]$, $C>0$. For any linear subspace $\widetilde{\mathds{U}}\subset\mathds{U}$ and $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\in\widetilde{\mathds{U}}$, ${\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\right\rVert}\leq\sup_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0],\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\in\widetilde{\mathds{U}}}\frac{\left\langle\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0],\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rangle}{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}\,{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}}{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}\quad\text{where}\quad\sup_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0],\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\in\widetilde{\mathds{U}}}\frac{\left\langle\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0],\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rangle}{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}\,{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}}\leq C^{-1}\begin{cases}|\Omega|^{\frac{1}{2}-\frac{1}{q}}&\text{ if }q\geq 2\text{, otherwise}\\\ |\Omega|^{1-\frac{1}{q}}\sup_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\widetilde{\mathds{U}}}{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{\infty}/{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}\qquad&\text{ if }q\in[1,2).\end{cases}$ (101) ###### Proof The first statement of the result is by definition, for each $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\in\widetilde{\mathds{U}}\subset L^{\infty}(\Omega)\subset\mathds{H}$ we have ${\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\right\rVert}=\frac{\left\langle\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]},\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\right\rangle}{{\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\right\rVert}}\leq\sup_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\widetilde{\mathds{U}}}\frac{\left\langle\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0],\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\right\rangle}{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}\,{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}}{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}\leq\sup_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0],\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\in\widetilde{\mathds{U}}}\frac{\left\langle\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0],\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rangle}{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}\,{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}}{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}.$ Recall ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}\geq C^{-1}{\left\lVert\cdot\right\rVert}_{q}$. To go further we use Hölder’s inequality. If $\frac{1}{q}+\frac{1}{q^{*}}=1$, then for any $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0],\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\in\widetilde{\mathds{U}}$ $\frac{\left\langle\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0],\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rangle}{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}\,{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}}\leq C^{-1}\frac{\left\langle\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0],\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rangle}{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}\,{\left\lVert\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rVert}_{q}}\leq C^{-1}\frac{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{q^{*}}}{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}}.$ (102) If $q\geq 2$ we use Hölder’s inequality a second time: $\int_{\Omega}|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0](\vec{x})|^{q^{*}}\mathop{}\\!\mathrm{d}\vec{x}\leq\left(\int_{\Omega}1\mathop{}\\!\mathrm{d}\vec{x}\right)^{1-q^{*}/2}\left(\int_{\Omega}|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0](\vec{x})|^{2}\mathop{}\\!\mathrm{d}\vec{x}\right)^{q^{*}/2}=\left(|\Omega|^{\frac{1}{2}-\frac{1}{q}}{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}\right)^{q^{*}}.$ (103) This confirms the inequality when $q\geq 2$. If $q<2$, we can simply upper bound ${\left\lVert\cdot\right\rVert}_{q^{*}}\leq|\Omega|^{\frac{1}{q^{*}}}{\left\lVert\cdot\right\rVert}_{\infty}$ as required. ∎ ###### Lemma 16 Suppose $\mathds{H}$ and $(\widetilde{\mathds{U}}^{k})_{k\in\mathds{N}}$ satisfy Definition 2, then 1. 1. If $q\geq 2$, then ${\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right\rVert}\lesssim 1$ (i.e. $q^{\prime}=0$). 2. 2. If $q<2$ and $\sup_{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\widetilde{\mathds{U}}^{0}}\frac{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{L^{\infty}(\omega_{0})}}{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{L^{2}(\omega_{0})}}<\infty$, then ${\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right\rVert}\lesssim h^{-\frac{kd}{2}}$ (i.e. $q^{\prime}=-\frac{d}{2}$). ###### Proof Most of the conditions of Lemma 15 are already satisfied. Furthermore observe that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}\lesssim{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}+h^{kp}\lesssim 1$. For the $q\geq 2$ case, this is already sufficient to conclude ${\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right\rVert}\lesssim 1$ from Lemma 15, as required. For the case $q<2$, from Lemma 15 recall that we are required to bound $\sup_{\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\in\widetilde{\mathds{U}}^{k}}\frac{{\left\lVert\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rVert}_{\infty}}{{\left\lVert\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rVert}}=\sup_{\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\in\widetilde{\mathds{U}}^{k}}\sup_{\omega\in\mathds{M}^{k}}\frac{{\left\lVert\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rVert}_{L^{\infty}(\omega)}}{{\left\lVert\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rVert}_{L^{2}(\Omega)}}\leq\sup_{\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\in\widetilde{\mathds{U}}^{k}}\sup_{\omega\in\mathds{M}^{k}}\frac{{\left\lVert\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rVert}_{L^{\infty}(\omega)}}{{\left\lVert\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rVert}_{L^{2}(\omega)}}.$ (104) However, due to the decomposition property (100), for each $\omega\in\mathds{M}^{k}$ and $\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\in\widetilde{\mathds{U}}^{k}$ there exists $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\in\widetilde{\mathds{U}}^{0}$ such that ${\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{L^{\infty}(\omega_{0})}={\left\lVert\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rVert}_{L^{\infty}(\omega)},\qquad{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{L^{2}(\omega_{0})}^{2}=\int_{\omega_{0}}|\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0](\vec{x})|^{2}\mathop{}\\!\mathrm{d}\vec{x}=\int_{\omega_{0}}|\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}(\alpha\vec{x}+\vec{\beta})|^{2}\mathop{}\\!\mathrm{d}\vec{x}=\operatorname{det}(\alpha)^{-1}{\left\lVert\widetilde{\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]}\right\rVert}_{L^{2}(\omega)}^{2}.$ (105) Combining these two equations with the assumed bound on $\frac{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{L^{\infty}(\omega_{0})}}{{\left\lVert\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]\right\rVert}_{L^{2}(\omega_{0})}}$ confirms ${\left\lVert\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}\right\rVert}\lesssim\sqrt{\operatorname{det}(\alpha)^{-1}}\leq c_{\alpha}^{-\frac{1}{2}}h^{-\frac{kd}{2}}$ as required. ∎ ###### Lemma 17 Suppose $\mathds{H}$ and $(\widetilde{\mathds{U}}^{k})_{k\in\mathds{N}}$ satisfy Definition 2 and $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$ is the minimiser of $E$ such that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}\lesssim h^{kp}$. 1. 1. If $\operatorname{E}$ is ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$-Lipschitz at $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$, then $\operatorname{E}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})-\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})\lesssim h^{kp}$ (i.e. $p^{\prime}=p$). 2. 2. If $\nabla\operatorname{E}$ is ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$-Lipschitz at $\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}$, then $\operatorname{E}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})-\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})\lesssim h^{2kp}$ (i.e. $p^{\prime}=2p$). ###### Proof Both statements are direct by definition, observe $\displaystyle\operatorname{E}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})-\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})\leq\operatorname{Lip}(\operatorname{E}){\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|},$ (106) $\displaystyle\operatorname{E}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})-\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*})\leq\left\langle\nabla\operatorname{E}(\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]),\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right\rangle=\left\langle\nabla\operatorname{E}(\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k})-\nabla\operatorname{E}(\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}),\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right\rangle\leq\operatorname{Lip}(\nabla\operatorname{E}){\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}-\IfEqCase{0}{{0}{u}{1}{v}{2}{w}}[u0]^{*}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}^{2}.$ (107) The proof is concluded by using the approximation bounds of $\widetilde{\IfEqCase{2}{{0}{u}{1}{v}{2}{w}}[u2]}_{k}$ in Definition 2. ∎ ## Appendix C Operator norms for numerical examples ###### Theorem C.1 Suppose $\mathsf{A}\colon\mathds{H}\to\mathds{R}^{m}$ has kernels $\psi_{j}\in L^{\infty}([0,1]^{d})$ for $j\in[m]$. 1. Case 1: If $\psi_{j}(\vec{x})=\begin{cases}1&\vec{x}\in\mathds{X}_{j}\\\ 0&\text{ else}\end{cases}$ for some collection $\mathds{X}_{j}\subset\Omega$ such that $\mathds{X}_{i}\cap\mathds{X}_{j}=\emptyset$ for all $i\neq j$, then ${\left\lVert\mathsf{A}\right\rVert}_{L^{2}\to\ell^{2}}=\max_{j\in[m]}\sqrt{|\mathds{X}_{j}|}.$ 2. Case 2: If $\psi_{j}(\vec{x})=\cos(\vec{\IfEqCase{3}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k3]}_{j}\vbox{\hbox{\leavevmode\resizebox{6.0pt}{}{$\cdot$}}}\vec{x})$ for some frequencies $\vec{\IfEqCase{3}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k3]}_{j}\in\mathds{R}^{d}$ with $|\vec{\IfEqCase{3}{{0}{\alpha}{1}{\beta}{2}{\gamma}{3}{a}{4}{b}{5}{c}}[k3]}_{j}|\leq A$, then ${\left\lVert\mathsf{A}\right\rVert}_{L^{2}\to\ell^{2}}\leq\sqrt{m},\qquad|\mathsf{A}^{*}\vec{r}|_{C^{k}}\leq m^{1-\frac{1}{q}}A^{k}{\left\lVert\vec{r}\right\rVert}_{q},\quad\text{and}\quad|\mathsf{A}^{*}|_{\ell^{2}\to C^{k}}\leq\sqrt{m}A^{k}$ for all $\vec{r}\in\mathds{R}^{m}$ and $q\in[1,\infty]$. 3. Case 3: Suppose $\psi_{j}(\vec{x})=(2\pi\sigma^{2})^{-\frac{d}{2}}\exp\left(-\frac{|\vec{x}-\vec{x}_{j}|^{2}}{2\sigma^{2}}\right)$ for some regular mesh $\vec{x}_{j}\in[0,1]^{d}$ and separation $\Delta$. i.e. $\\{\vec{x}_{j}\operatorname{\;s.t.\;}j\in[m]\\}=\\{\vec{x}_{0}+(j_{1}\Delta,\ldots,j_{d}\Delta)\operatorname{\;s.t.\;}j_{i}\in[\widehat{m}]\\}$ for some $\vec{x}_{0}\in\mathds{R}^{d}$, $\widehat{m}\coloneqq\sqrt[d]{m}$. For all $\frac{1}{q}+\frac{1}{q^{*}}=1$, $q\in(1,\infty]$, we have $\displaystyle{\left\lVert\mathsf{A}\right\rVert}_{L^{2}\to\ell^{2}}$ $\displaystyle\leq\bigg{(}(4\pi\sigma^{2})^{-\frac{1}{2}}\sum_{j=-2\widehat{m},\ldots,2\widehat{m}}\exp(-\tfrac{\Delta^{2}}{4\sigma^{2}}j^{2})\bigg{)}^{d},$ (108) $\displaystyle|\mathsf{A}^{*}\vec{r}|_{C^{0}}$ $\displaystyle\leq(2\pi\sigma^{2})^{-\frac{d}{2}}\bigg{(}\sum_{\vec{j}\in J}\exp\left(-\tfrac{q^{*}\Delta^{2}}{2\sigma^{2}}\max(0,|\vec{j}|-\delta)^{2}\right)\bigg{)}^{\frac{1}{q^{*}}}{\left\lVert\vec{r}\right\rVert}_{q},$ (109) $\displaystyle|\mathsf{A}^{*}\vec{r}|_{C^{1}}$ $\displaystyle\leq\frac{(2\pi\sigma^{2})^{-\frac{d}{2}}}{\sigma}\frac{\Delta}{\sigma}\bigg{(}\sum_{\vec{j}\in J}(|\vec{j}|+\delta)^{q^{*}}\exp\left(-\tfrac{q^{*}\Delta^{2}}{2\sigma^{2}}\max(0,|\vec{j}|-\delta)^{2}\right)\bigg{)}^{\frac{1}{q^{*}}}{\left\lVert\vec{r}\right\rVert}_{q},$ (110) $\displaystyle|\mathsf{A}^{*}\vec{r}|_{C^{2}}$ $\displaystyle\leq\frac{(2\pi\sigma^{2})^{-\frac{d}{2}}}{\sigma^{2}}\bigg{(}\sum_{\vec{j}\in J}\left(1+\tfrac{\Delta^{2}}{\sigma^{2}}(|\vec{j}|+\delta)^{2}\right)^{q^{*}}\exp\left(-\tfrac{q^{*}\Delta^{2}}{2\sigma^{2}}\max(0,|\vec{j}|-\delta)^{2}\right)\bigg{)}^{\frac{1}{q^{*}}}{\left\lVert\vec{r}\right\rVert}_{q},$ (111) where $\delta=\frac{\sqrt{d}}{2}$ and $J=\\{\vec{j}\in\mathds{Z}^{d}\operatorname{\;s.t.\;}{\left\lVert\vec{j}\right\rVert}_{\ell^{\infty}}\leq 2\widehat{m}\\}$. The case for $q=1$ can be inferred from the standard limit of ${\left\lVert\cdot\right\rVert}_{{q^{*}}}\to{\left\lVert\cdot\right\rVert}_{\infty}$ for $q^{*}\to\infty$. ###### Proof (Case 1.) From Lemma 8 we have $(\mathsf{A}\mathsf{A}^{*})_{i,j}=\left\langle\mathds{1}_{\mathds{X}_{i}},\mathds{1}_{\mathds{X}_{j}}\right\rangle=|\mathds{X}_{i}\cap\mathds{X}_{j}|=\begin{cases}|\mathds{X}_{i}|&i=j\\\ 0&i\neq j\end{cases}.$ (112) Therefore, $\mathsf{A}\mathsf{A}^{*}$ is a diagonal matrix and ${\left\lVert\mathsf{A}\mathsf{A}^{*}\right\rVert}_{\ell^{2}\to\ell^{2}}=\max_{j\in[m]}|\mathds{X}_{j}|$ completes the result. ∎ ###### Proof (Case 2.) $\psi_{j}$ are not necessarily orthogonal however $|\left\langle\psi_{i},\psi_{j}\right\rangle|\leq 1$ therefore we can estimate ${\left\lVert\mathsf{A}\mathsf{A}^{*}\right\rVert}_{\ell^{2}\to\ell^{2}}\leq{\left\lVert\mathsf{A}\mathsf{A}^{*}\right\rVert}_{\ell^{\infty}\to\ell^{\infty}}\leq m.$ (113) Now looking to apply Lemma 9, note ${\left\lVert\nabla^{k}\psi_{j}\right\rVert}_{\infty}\leq A^{k}$, therefore $|\mathsf{A}^{*}\vec{r}|_{C^{k}}\leq A^{k}m^{\frac{1}{q^{*}}}{\left\lVert\vec{r}\right\rVert}_{q}=A^{k}m^{1-\frac{1}{q}}{\left\lVert\vec{r}\right\rVert}_{q}\quad\text{and}\quad|\mathsf{A}^{*}|_{\ell^{2}\to C^{k}}\leq A^{k}\min_{q\in[1,\infty]}m^{1-\frac{1}{q}}\sqrt{m}^{\max(0,2-q)}=\sqrt{m}A^{k}.$ (114) ∎ ###### Proof (Case 3.) In the Gaussian case, we build our approximations around the idea that sums of Gaussians should converge very quickly. The first example can be used to approximate the operator norm. Computing the inner products gives $\left\langle\psi_{i},\psi_{j}\right\rangle=(2\pi\sigma^{2})^{-d}\int_{[0,1]^{d}}\exp\left(-\tfrac{|\vec{x}-\vec{x}_{i}|^{2}}{2\sigma^{2}}-\tfrac{|\vec{x}-\vec{x}_{j}|^{2}}{2\sigma^{2}}\right)\mathop{}\\!\mathrm{d}\vec{x}\leq(2\pi\sigma^{2})^{-d}(\pi\sigma^{2})^{\frac{d}{2}}\exp\left(-\tfrac{|\vec{x}_{i}-\vec{x}_{j}|^{2}}{4\sigma^{2}}\right).$ (115) Estimating the operator norm, $\displaystyle{\left\lVert\mathsf{A}\mathsf{A}^{*}\right\rVert}_{\ell^{2}\to\ell^{2}}$ $\displaystyle\leq{\left\lVert\mathsf{A}\mathsf{A}^{*}\right\rVert}_{\ell^{\infty}\to\ell^{\infty}}=\max_{i\in[m]}\sum_{j=1}^{m}|\left\langle\psi_{i},\psi_{j}\right\rangle|$ (116) $\displaystyle=\max_{i\in[m]}(4\pi\sigma^{2})^{-\frac{d}{2}}\sum_{j_{1},\ldots,j_{d}\in[\widehat{m}]}\exp\left(-\frac{(j_{1}\Delta- i_{1}\Delta)^{2}+\ldots+(j_{d}\Delta-i_{d}\Delta)^{2}}{4\sigma^{2}}\right)$ (117) $\displaystyle\leq(4\pi\sigma^{2})^{-\frac{d}{2}}\sum_{\vec{j}\in\mathds{Z}^{d}\cap[-\widehat{m},\widehat{m}]^{d}}\exp\left(-\frac{(j_{1}\Delta)^{2}+\ldots+(j_{d}\Delta)^{2}}{4\sigma^{2}}\right)=\left[(4\pi\sigma^{2})^{-\frac{1}{2}}\sum_{j=-\widehat{m}}^{\widehat{m}}\exp\left(-\frac{\Delta^{2}j^{2}}{4\sigma^{2}}\right)\right]^{d}.$ (118) This is a nice approximation because it factorises simply over dimensions. Applying the results from Lemma 9, note $\begin{array}[]{rll}\displaystyle|\psi_{j}(\vec{x})|&\displaystyle=\left|\psi_{j}(\vec{x})\right|&\displaystyle=(2\pi\sigma^{2})^{-\frac{d}{2}}\exp\left(-\frac{|\vec{x}-\vec{x}_{j}|^{2}}{2\sigma^{2}}\right),\\\ \displaystyle|\nabla\psi_{j}(\vec{x})|&\displaystyle=\left|\frac{\vec{x}-\vec{x}_{j}}{\sigma^{2}}\psi_{j}(\vec{x})\right|&\displaystyle=\frac{(2\pi\sigma^{2})^{-\frac{d}{2}}}{\sigma}\frac{|\vec{x}-\vec{x}_{j}|}{\sigma}\exp\left(-\frac{|\vec{x}-\vec{x}_{j}|^{2}}{2\sigma^{2}}\right),\\\ |\nabla^{2}\psi_{j}(\vec{x})|&\displaystyle=\left|\frac{1}{\sigma^{2}}+\frac{(\vec{x}-\vec{x}_{j})(\vec{x}-\vec{x}_{j})^{\top}}{\sigma^{4}}\right|\psi_{j}(\vec{x})&\displaystyle=\frac{(2\pi\sigma^{2})^{-\frac{d}{2}}}{\sigma^{2}}\left(1+\frac{|\vec{x}-\vec{x}_{j}|^{2}}{\sigma^{2}}\right)\exp\left(-\frac{|\vec{x}-\vec{x}_{j}|^{2}}{2\sigma^{2}}\right).\end{array}$ We now wish to sum over $j=1,\ldots,m$ and produce an upper bound on these, independent of $t$. To do so we will use the following lemma. ###### Lemma 18 Suppose $q>0$. If the polynomial $p(|\vec{x}|)=\sum p_{k}|\vec{x}|^{k}$ has non-negative coefficients and $\vec{x}\in[-m,m]^{d}$, then $\sum_{{\left\lVert\vec{j}\right\rVert}_{\ell^{\infty}}\leq m}p(|\vec{j}-\vec{x}|)\exp\left(-\tfrac{q|\vec{j}-\vec{x}|^{2}}{2}\right)\leq\sum_{{\left\lVert\vec{j}\right\rVert}_{\ell^{\infty}}\leq 2m}p(|\vec{j}|+\delta)\exp\left(-\frac{q\max(0,|\vec{j}|-\delta)^{2}}{2}\right)$ where $\delta\coloneqq\frac{\sqrt{d}}{2}$ and $\vec{j}\in\mathds{Z}^{d}$. ###### Proof There exists $\widehat{\vec{x}}\in[-\tfrac{1}{2},\tfrac{1}{2}]^{d}$ such that $\vec{x}+\widehat{\vec{x}}\in\mathds{Z}^{d}$, therefore $\displaystyle\sum_{{\left\lVert\vec{j}\right\rVert}_{\ell^{\infty}}\leq m}p(|\vec{j}-\vec{x}|)\exp\left(-\tfrac{q|\vec{j}-\vec{x}|^{2}}{2}\right)$ $\displaystyle=\sum_{{\left\lVert\vec{j}\right\rVert}_{\ell^{\infty}}\leq m}p(|\vec{j}-(\vec{x}+\widehat{\vec{x}})+\widehat{\vec{x}}|)\exp\left(-\tfrac{q|\vec{j}-(\vec{x}+\widehat{\vec{x}})+\widehat{\vec{x}}|^{2}}{2}\right)$ $\displaystyle\leq\sum_{{\left\lVert\vec{j}\right\rVert}_{\ell^{\infty}}\leq 2m}p(|\vec{j}+\widehat{\vec{x}}|)\exp\left(-\tfrac{q|\vec{j}+\widehat{\vec{x}}|^{2}}{2}\right)$ $\displaystyle\leq\sum_{\begin{subarray}{c}\vec{j}\in\mathds{Z}^{d}\\\ {\left\lVert\vec{j}\right\rVert}_{\ell^{\infty}}\leq 2m\end{subarray}}p(|\vec{j}|+\delta)\exp\left(-\tfrac{q\max(0,|\vec{j}|-\delta)^{2}}{2}\right)$ as $|\widehat{\vec{x}}|\leq\delta$ and $p$ has non-negative coefficients. ∎ Now, continuing the proof of Theorem C.1, for $\widehat{m}=\sqrt[d]{m}$, $\delta=\frac{\sqrt{d}}{2}$ and $J=\\{\vec{j}\in\mathds{Z}^{d}\operatorname{\;s.t.\;}{\left\lVert\vec{j}\right\rVert}_{\ell^{\infty}}\leq 2\widehat{m}\\}$, Lemma 18 bounds $\displaystyle\sum_{j=1}^{m}|\psi_{j}(\vec{x})|^{q^{*}}$ $\displaystyle\leq(2\pi\sigma^{2})^{-\frac{dq^{*}}{2}}\left[\sum_{\vec{j}\in J}\exp\left(-\frac{q^{*}\Delta^{2}}{2\sigma^{2}}\max(0,|\vec{j}|-\delta)^{2}\right)\right]$ $\displaystyle\sum_{j=1}^{m}|\nabla\psi_{j}(\vec{x})|^{q^{*}}$ $\displaystyle\leq\frac{(2\pi\sigma^{2})^{-\frac{dq^{*}}{2}}}{\sigma^{q^{*}}}\frac{\Delta^{q^{*}}}{\sigma^{q^{*}}}\left[\sum_{\vec{j}\in J}(|\vec{j}|+\delta)^{q^{*}}\exp\left(-\frac{q^{*}\Delta^{2}}{2\sigma^{2}}\max(0,|\vec{j}|-\delta)^{2}\right)\right]$ $\displaystyle\sum_{j=1}^{m}|\nabla^{2}\psi_{j}(\vec{x})|^{q^{*}}$ $\displaystyle\leq\frac{(2\pi\sigma^{2})^{-\frac{dq^{*}}{2}}}{\sigma^{2q^{*}}}\left[\sum_{\vec{j}\in J}\left(1+\frac{\Delta^{2}}{\sigma^{2}}(|\vec{j}|+\delta)^{2}\right)^{q^{*}}\exp\left(-\frac{q^{*}\Delta^{2}}{2\sigma^{2}}\max(0,|\vec{j}|-\delta)^{2}\right)\right]$ for all $\vec{x}\in\Omega$. In a worst case, this is $O(2^{d}m)$ time complexity however the summands all decay faster than exponentially and so should converge very quickly. ∎
# X-ray Scatter Estimation Using Deep Splines Philipp Roser, Annette Birkhold, Alexander Preuhs, Christopher Syben, Lina Felsner, Elisabeth Hoppe, Norbert Strobel, Markus Korwarschik, Rebecca Fahrig, Andreas Maier P. Roser, A. Preuhs, C. Syben, L. Felsner, E. Hoppe, and A. Maier are with the Pattern Recognition Lab, Department of Computer Science, Friedrich-Alexander Universität Erlangen-Nürnberg, Erlangen, Germany. P. Roser is funded by the Erlangen Graduate School in Advanced Optical Technologies (SAOT), Friedrich-Alexander Universität Erlangen-Nürnberg, Erlangen, Germany. A. Maier is principal investigator at the SAOT. A. Birkhold, M. Kowarschik, and R. Fahrig are employees of Siemens Healthcare GmbH, 91301 Forchheim, Germany. N. Strobel is with the Institute of Medical Engineering Schweinfurt, University of Applied Sciences Würzburg-Schweinfurt, 97421 Schweinfurt, Germany. ###### Abstract Algorithmic X-ray scatter compensation is a desirable technique in flat-panel X-ray imaging and cone-beam computed tomography. State-of-the-art U-net based image translation approaches yielded promising results. As there are no physics constraints applied to the output of the U-Net, it cannot be ruled out that it yields spurious results. Unfortunately, those may be misleading in the context of medical imaging. To overcome this problem, we propose to embed B-splines as a known operator into neural networks. This inherently limits their predictions to well-behaved and smooth functions. In a study using synthetic head and thorax data as well as real thorax phantom data, we found that our approach performed on par with U-net when comparing both algorithms based on quantitative performance metrics. However, our approach not only reduces runtime and parameter complexity, but we also found it much more robust to unseen noise levels. While the U-net responded with visible artifacts, our approach preserved the X-ray signal’s frequency characteristics. ###### Index Terms: Approximation, B-spline, Neural network, X-ray scatter. ## I Introduction Xray transmission imaging enables advanced diagnostic imaging or facilitates instrument guidance during minimally-invasive interventions. Unfortunately, primary photons scattered by the patient’s body impair image quality at the detector. Especially cone-beam X-ray systems using flat-panel detectors suffer from this phenomenon due to the large X-ray field and the resulting high scatter-to-primary ratio. Besides contrast degradation in X-ray projections, artifacts in cone-beam computed tomography (CBCT) deteriorate the image quality in the reconstructed volume. Since the advent of CBCT, various methods to compensate for scatter have been developed [1, 2]. These can be categorized into physical and algorithmic scatter compensation methods. ### I-A Physical Scatter Compensation Physical scatter compensation refers to the direct manipulation of the X-ray field to either suppress scattered photons from reaching the detector, or modulate them to enable differentiation into scattered and primary photons. A simple yet effective approach is to increase the patient to detector distance, the so-called air gap [3, 4, 5, 6]. Since the number of scattered photons depends on the irradiated volume, the scatter-to-primary ratio can be reduced by shaping the X-ray field to a small volume of interest. In slit scanning, the stitching of these smaller volumes of interest yields a normal-sized reconstruction [7, 8]. Although slit scanning reduces patient dose and scattered radiation, the prolonged acquisition time complicates motion compensation and X-ray tube heat management [9]. Since relying on large air gaps or strict collimation limits the flexibility of the imaging system, clinical systems are equipped with anti-scatter grids that physically block scattered photons [5, 10, 11, 12, 13, 14]. Although state-of-the-art for clinical CBCT, anti-scatter grids have disadvantages. For example, they can increase radiation exposure [15] and lead to Moiré and grid line artifacts [16]. Overall, the efficiency and effectiveness of anti-scatter grids depends on multiple factors, e.g., air gap, photon energy, source-to-detector distance, and detector resolution [2, 17, 18]. In contrast to detector-side grids, primary modulation follows a different operation principle [19, 20, 21]. In the shadow of this modulator grid, almost no primary radiation is measured. Instead, information of the scatter properties of the irradiated volume is encoded. Using demodulation algorithms, the X-ray projection and the associated scatter distribution can be distinguished. Due to its complex structure, primary modulation, especially with C-arm systems, has not made its way into the clinical practice. In conclusion, hardware-based solutions induce additional manufacturing costs and limit flexibility. Also, since most approaches need algorithmic support or look-up tables, software-only solutions are highly desirable [1, 2]. ### I-B Algorithmic Scatter Compensation Algorithmic scatter compensation approaches typically try to estimate the scatter signal from the acquired X-ray projections first. The estimated scatter image is then subtracted to obtain the primary signal. For CBCT, using Monte Carlo (MC) methods or Boltzmann transport solvers are common approaches [22, 23, 24]. Both require an initial reconstruction. Their computational complexity can partially be coped with using dedicated hardware and variance reduction techniques [25] or by using a coarse simulation to derive the scattering model online [26]. Trading off effectiveness for efficiency, model- based approaches are typically preferred for clinical applications. Such approaches rely on either simplified physical or analytical models [27, 28, 29] or convolution kernels [30, 31, 32] in combination with an iterative correction scheme. Although being fast, their ability to generalize to different imaging settings is limited. Recently, learning-based methods found their way to modeling physical processes [33, 34, 35]. Especially the application of the U-net [36] to the task of scatter estimation, referred to as deep scatter estimation, yields superior results compared to kernel-based or MC methods, respectively [37]. The U-net based method has the potential to become the new gold standard in the domain of scatter correction. However, there are several drawbacks to employing a deep U-net. First, deep convolutional neural networks are challenging to comprehend in their operating principle. Given the high parameter complexity of such a U-net, large amounts of training samples are mandatory to arrive at a robust solution. Second, to maintain a fast performance, the U-net requires a dedicated top tier graphics processing unit (GPU), which might not be universally available. Third, the U-net is a universal function approximator without including any known physical characteristics of scattered radiation. To come up with an alternative, it appears attractive to build on the rich prior knowledge about X-ray scatter. Also, it has been shown that the incorporation of prior knowledge into neural networks reduces error bounds [38] and simplifies the analysis of the deployed network [39]. Although subject to photon Poisson noise, X-ray scatter is mainly a low-frequency signal in the diagnostic energy regime [40, 41]. In a previous proof-of- concept study, we exploited this property by approximating X-ray scatter using a smooth bivariate B-spline and directly inferred spline coefficients using a lean convolutional encoder architecture [42]. This (a) allowed us to reduce the number of parameters and computational complexity tremendously, (b) ensured that no high-frequency details of diagnostic value can get manipulated, while (c) still meeting the high accuracy of the U-net. ### I-C Contributions In this work, we extend the idea of approximating X-ray scatter as bivariate B-spline by integrating its evaluation into the computational graph of a neural network. We analyze our approach in depth and compare it to several U-net realizations in a nested cross-validation study using synthetic head and thorax X-ray image data. Besides benchmarking the parameter and computational complexity, we investigate all networks in terms of their frequency response, the power spectral density of the co-domain, and the response to unseen noise levels. Finally, we demonstrate how our method performs when applied to real data in an anthropomorphic thorax phantom study. ## II Materials and Methods ### II-A X-ray Projection Formation using B-splines In general, in both techniques, X-ray fluoroscopy and CBCT it is assumed that primary photons either reach the detector along a straight line or are absorbed completely. Therefore, the observed X-ray projection is typically described in terms of its primary component $\boldsymbol{I}_{\text{p}}\in\mathbb{R}^{w\times h}$ with image width $w$ and height $h$ in pixels. The intensity at pixel $(u,v)$ is given by the Beer- Lambert law ${I}_{\text{p}}\left(u,v\right)=\int_{E}{I}_{0}\left(u,v,E\right)e^{-\int_{\lambda}\mu\left(\boldsymbol{s}+\lambda\cdot\boldsymbol{r}\left(u,v\right),E\right)d\lambda}dE\enspace.$ (1) The polychromatic flat-field projection or X-ray spectrum $\boldsymbol{I}_{0}$ as well as the linear attenuation coefficient $\mu\left(\boldsymbol{l},E\right)$ depend on the photon energy $E$. Note that, due to the cone-beam geometry, $\boldsymbol{I}_{0}$ decreases towards the borders and thus depends on the pixel position $(u,v)$. The linear photon attenuation further depends on the media along the straight line $\boldsymbol{l}:\mathbb{R}\mapsto\mathbb{R}^{3}$ from the X-ray source $\boldsymbol{s}\in\mathbb{R}^{3}$ in direction $\boldsymbol{r}\in\mathbb{R}^{3}$ to the pixel $(u,v)$. Unfortunately, in reality, besides being photoelectrically absorbed, X-ray photons also undergo Compton scattering, Rayleigh scattering, or multiple occurrences of both effects. Therefore, a more realistic formation model $\boldsymbol{I}$ adds scattered photons $\boldsymbol{I}_{\text{s}}$ leading to ${I}\left(u,v\right)={I}_{\text{p}}\left(u,v\right)+{I}_{\text{s}}\left(u,v\right)\enspace.$ (2) Since, for CBCT, we are mostly interested in the low-frequency components of the scatter, we neglect photon shot noise and detector noise below. In a recent study [42], we experimentally showed that the main scatter components, which are low-frequency for diagnostic X-rays, can be well approximated by sparse bivariate B-splines with error rates below $1\text{\,}\mathrm{\char 37\relax}$. The domain and co-domain of a B-spline of degree $k$ and order $n=k+1$ are fully characterized by the knot vectors $\boldsymbol{t}_{u}=(t_{u,1},t_{u,2},\dots t_{u,w_{c}-n})$ and $\boldsymbol{t}_{v}=(t_{v,1},t_{v,2},\dots t_{v,h_{c}-n})$, as well as the coefficient matrix $\boldsymbol{C}\in\mathbb{R}^{w_{\text{c}}\times h_{\text{c}}}$ with width $w_{\text{c}}$ and height $h_{\text{c}}$, respectively. Based on this, we can approximate the spatial scatter distribution as a tensor product B-spline $\tilde{\boldsymbol{I}}_{\text{s},n}$ $\tilde{I}_{\text{s},n}\left(u,v\right)=\sum_{i=1}^{w_{\text{c}}}\sum_{j=1}^{h_{\text{c}}}C\left(i,j\right)B_{i,n,\boldsymbol{t}_{u}}\left(u\right)B_{j,n,\boldsymbol{t}_{v}}\left(v\right)\,,$ (3) with the basis splines $B$, which are zero for knots that do not affect the spline at the pixel $(u,v)$. The basis splines are recursively defined by $\displaystyle B_{i,n,\boldsymbol{t}}(x)$ $\displaystyle={}$ $\displaystyle\frac{x-t_{i}}{t_{i+n}-t_{i}}$ $\displaystyle\cdot B_{i,n-1,\boldsymbol{t}}(x)$ (4) $\displaystyle+{}$ $\displaystyle\frac{t_{i+n+1}-x}{t_{i+n+1}-t_{i+1}}$ $\displaystyle\cdot B_{i+1,n-1,\boldsymbol{t}}(x)\enspace,$ with $B_{i,0,\boldsymbol{t}}(x)=\begin{cases}1,&\text{if }t_{i}\leq x<t_{i+1}\\\ 0,&\text{otherwise}\end{cases}\enspace.$ (5) Since the basis functions $B$ are equal to zero for coefficients that do not contribute to a pixel $(u,v)$, only a $n\times n$ sub-grid of $\boldsymbol{C}$ needs to be considered for each pixel. Thus, using matrix notation, the tensor product can be reformulated to $\tilde{I}_{\text{s},n}\left(u,v\right)=\boldsymbol{u}^{\text{T}}\cdot\left[\boldsymbol{M}_{n,\boldsymbol{t}_{u}}\left(u\right)\right]\cdot\boldsymbol{C}_{\text{uv}}\cdot\left[\boldsymbol{M}_{n,\boldsymbol{t}_{v}}\left(v\right)\right]^{\text{T}}\cdot\boldsymbol{v}\enspace,$ (6) with the coefficient patch $\boldsymbol{C}_{\text{uv}}\in\mathbb{R}^{n\times n}$ impacting the pixel $(u,v)$, the general matrix representation of a univariate B-spline $\boldsymbol{M}_{n,\boldsymbol{t}}\left(\cdot\right)\in\mathbb{R}^{n\times n}$ [43], and the vectors $\boldsymbol{u}=\begin{pmatrix}u^{0}&u^{1}&\cdots&u^{k}\end{pmatrix}^{\text{T}}$ and $\boldsymbol{v}=\begin{pmatrix}v^{0}&v^{1}&\cdots&v^{k}\end{pmatrix}^{\text{T}}$. The general matrix representation of B-splines allows for arbitrarily spaced knot grids, i.e., in our case, it allows for endpoint interpolation, which exposes a more convenient behavior towards the borders of the evaluation grid. Besides the borders of the grid, we currently limit our approach to uniform knot grids and cubic B-splines ($k=3$). Thus, for most pixels $(u,v)$, the B-spline is evaluated using $\boldsymbol{M}_{4}=\frac{1}{6}\begin{pmatrix}1&4&1&0\\\ -3&0&3&0\\\ 3&-6&3&0\\\ 1&3&-3&1\end{pmatrix}\enspace.$ (7) With using a fixed knot grid and evaluation grid, we can pre-calculate both $\boldsymbol{u}^{\text{T}}\cdot\left[\boldsymbol{M}_{n,\boldsymbol{t}_{u}}\left(u\right)\right]$ and $\boldsymbol{v}^{\text{T}}\cdot\left[\boldsymbol{M}_{n,\boldsymbol{t}_{v}}\left(v\right)\right]$ for each pixel $(u,v)$. Padding the resulting vectors with zeros to account for all coefficients in $\boldsymbol{C}$ and stacking them row-wisely, we obtain the evaluation matrices $\boldsymbol{U}_{n}\in\mathbb{R}^{w\times w_{\text{c}}}$ and $\boldsymbol{V}_{n}\in\mathbb{R}^{h\times h_{\text{c}}}$, respectively. Thus, we can calculate the cubic ($n=4$) scatter distribution given the B-spline coefficients $\boldsymbol{C}$ via $\tilde{\boldsymbol{I}}_{\text{s},4}=\boldsymbol{U}_{4}\cdot\boldsymbol{C}\cdot\boldsymbol{V}_{4}^{\text{T}}\enspace.$ (8) Since neural networks can extract scatter distributions from a measured X-ray projection either directly [37] or indirectly via B-spline coefficients [42] we aim to combine deep learning with the matrix evaluation scheme. Using the Kronecker product ‘$\otimes$’, the derivative $\frac{\partial\tilde{\boldsymbol{I}}_{\text{s},4}}{\partial\boldsymbol{C}}$ is $\frac{\partial\tilde{\boldsymbol{I}}_{\text{s},4}}{\partial\boldsymbol{C}}=\boldsymbol{V}_{4}\otimes\boldsymbol{U}_{4}\enspace.$ (9) Thus, we can straightforwardly embed the B-spline evaluation into a computational graph $f_{\boldsymbol{\theta}}:\mathbb{R}^{w\times h}\mapsto\mathbb{R}^{w\times h}$ with parameters to train $\boldsymbol{\theta}$ without breaking differentiability and thus back- propagation. ### II-B Network Architecture Figure 1: The general architecture of our proposed approach. First, a (convolutional) encoder extracts the latent variables $\boldsymbol{Z}$ from the input X-ray projection $\boldsymbol{I}$. Second, a bottleneck network maps the latent space to bivariate spline coefficients $\boldsymbol{C}$. The different choices for both networks are discussed later in more detail. Third, we obtain the scatter estimate $\tilde{\boldsymbol{I}}_{\text{s},n}$ by evaluating the spline coefficients $\boldsymbol{C}$ using the pre-calculated evaluation grid defined by the pre-computed matrices $\boldsymbol{U}$ and $\boldsymbol{V}$. The top (green) path refers to the forward pass, and the bottom (orange) path to the backward pass, respectively. In the following, we describe the devised neural network. The overall architecture comprises a generic convolutional encoder followed by a bottleneck network to infer spline coefficients from measured X-ray projections, as proposed in our previous study [42]. This architecture is depicted in Fig. 1. We employ a lean convolutional encoder $f_{\boldsymbol{\theta}}:\mathbb{R}^{w\times h}\mapsto\mathbb{R}^{w_{\text{c}}\times h_{\text{c}}}$ that consists of $d$ convolutional blocks. A block comprises two convolutional layers with $c$ feature channels and $3\times 3$ kernels. Each convolutional layer is followed by a rectified linear unit (ReLU) [44] activation, and between two blocks, $2\times 2$ average pooling is applied. As X-ray scatter is low-frequency and to potentially limit the computational complexity of our model, we allow for additional pooling layers before the first convolutional block. The encoder $f_{\boldsymbol{\theta}}$ is completed by a $1\times 1$ convolution to build the weighted sum of all channels $c$. Overall, based on convolutions only, $f_{\boldsymbol{\theta}}$ only encodes a local scatter representation $\boldsymbol{Z}$, since its receptive field does not necessarily cover the whole input image. To establish a global context, we employ different types of bottleneck networks $g_{\boldsymbol{\phi}}$ to map $\boldsymbol{Z}$ to spline coefficients $\boldsymbol{C}$: (1) a constrained weighting matrix $\boldsymbol{W}$ with $W_{i,j}>0$, (2) an unconstrained fully-connected layer with ReLU activation (which merely relates to an unconstrained weighting matrix with enforced non-negativity constraint), (3) two fully-connected layers with ReLU activation, or (4) two additional convolutional blocks followed by a fully-connected layer. ### II-C Synthetic Dataset Acquiring raw scatter-free X-ray projections and their scatter-contaminated counterparts is tedious and time-consuming, especially at the scale to carry out deep learning. As a solution, we leveraged the X-ray transport code MC-GPU [25] to generate artificial pairs of scatter-contaminated and scatter-free X-ray projections. We used openly available CT scans from The Cancer Imaging Archive (TCIA) [45] as inputs to the simulation. In order to keep the simulation time manageable, we selected 20 head scans from the HNSCC-3DCT-RT dataset [46] and 15 thorax scans from the CT Lymph Nodes dataset [47], respectively. To prepare the phantoms for MC simulation, we employed a basic pre-processing pipeline [48] based on tissue and density estimation [49], and connected component labeling [50]. For each CT scan, we simulated a stack of 260 X-ray projections ($w=1152$, $h=768$) over an angular range of $200\text{\,}\mathrm{\SIUnitSymbolDegree}$. The source-to-isocenter and source-to-detector distances are $785\text{\,}\mathrm{mm}$ and $1300\text{\,}\mathrm{mm}$, respectively. Per X-ray projection, we simulated $5\text{\times}{10}^{10}$ photons sampled from an $85\text{\,}\mathrm{kV}$ peak voltage tungsten spectrum. All projections were flat-field normalized. Note that we used the data as is, without registering similar anatomies in a common reference frame. To be more comparable to the related work [37], speed- up training times, and suppress simulation noise, we down-sampled the projections to $384\times 256$ pixels. Before the down-sampling, we applied Gaussian filtering to the primary and scatter projections independently ($\sigma_{\text{p}}=2$, $\sigma_{\text{s}}=30$). Corresponding cross-sectional slices were reconstructed on an isotropic ${256}^{3}$ grid with $1\text{\,}{\mathrm{mm}}^{3}$ voxels. ### II-D Real Dataset To evaluate the proposed method on real data, we scanned the thorax of an anthropomorphic phantom (PBU-60, Kyoto Kagaku Co., Ltd., Kyoto, Japan) using a C-arm CBCT system (ARTIS icono floor, Siemens Healthineers AG, Forchheim, Germany). In total, we acquired 12 datasets, three short scans for each full view grid (referred to as full), and maximum collimation (referred to as slit), both with and without an anti-scatter grid. Each short scan consists of 397 projections ($648\text{\times}472$ pixels, $616\text{\,}\mathrm{\SIUnitSymbolMicro m}$ isotropic spacing) over an angular range of $197.5\text{\,}\mathrm{\SIUnitSymbolDegree}$ using $85\text{\,}\mathrm{kV}$ peak tube voltage. The source-to-isocenter and source-to-detector distances are $750\text{\,}\mathrm{mm}$ and $1200\text{\,}\mathrm{mm}$, respectively. We reconstructed $512\text{\times}512$ slices on an $484\text{\,}\mathrm{\SIUnitSymbolMicro m}$ isotropic voxel grid using an in-house reconstruction pipeline. We regard the slit scan in conjunction with the anti-scatter grid as ground truth. ## III Experiments In the following, we describe the experiments carried out to evaluate our proposed method. Since the U-net based approach outperformed other computational scatter estimation methods by far [37], we considered different configurations of the U-net as a baseline method. To account for our relatively small training corpora, we evaluated our method and the baseline using a nested cross-validation approach. For the head dataset (20 subjects), we used a $4^{*}3$-fold cross-validation approach and, for the thorax dataset (15 subjects), we used a $5^{*}4$-fold cross-validation approach. The real dataset was only used for testing. To keep the training procedure manageable, we divided the evaluation into meta-parameter search based on the scatter mean absolute percentage error (MAPE), and further in-depth analysis. ### III-A Meta-Parameter Search Since clinic CBCT systems usually have fine-tuned acquisition protocols for each anatomic region, we evaluated each network architecture for head and thorax data separately. To fix the meta-parameters, we only used the synthetic head dataset and the scatter MAPE as metric. We validated both our approach and the U-net for different combinations of depths $d$ and feature channels $c$ using Glorot initialization [51], which was also used by the baseline [37]. Furthermore, we distinguished between a deep U-net (DU-net) and a shallow U-net (SU-net), in which the number of feature maps is not doubled at each level. For all experiments in this section, we performed a $4^{*}3$-fold cross-validation and trained all networks using the adaptive moments optimizer (Adam) [52] with an initial learning rate of ${10}^{-4}$ for 100 epochs. In total, we trained 12 networks for each configuration and used the averaged MAPE to assess their quality. To reduce the total computation time, we stopped each training procedure when no significant performance increase was observed for 20 epochs. In a first step, we scanned different parametrizations of the network architectures to find the most promising ones for in-depth comparison. Overall, we tested both U-nets with $c=16$ and $d\in\\{4,5,6,7\\}$, and our approach with $c=16$, $d\in\\{4,5,6\\}$, and additional pre-pooling $p\in\\{0,1,2\\}$. In addition, to decouple the architecture of our convolutional encoder $f_{\boldsymbol{\theta}}$ and the spline coefficient dimensionality, we investigated the four bottleneck architectures $g_{\boldsymbol{\phi}}$ as described in Sec. II-B. ### III-B Qualitative and Quantitative Results Based on the findings of the previous experiments, we adapted the learning rate to ${10}^{-5}$ but kept the overall training routine. We separately performed 4∗3-fold and 5∗4-fold cross-validations for the head and thorax datasets, respectively. In addition to the scatter MAPE, we included the structural similary index (SSIM) of scatter-compensated reconstructions with respect to the ground truth in our evaluation. ### III-C In-Depth Analysis #### III-C1 Spectral Analysis For clinical applications, data integrity is of utmost importance to ensure that automated systems do not alter diagnostically relevant content. While the predictions of neural networks may appear reasonable at first glance, unrealistic perturbations can be unveiled by investigating the spectral properties of (a) the predicted images [53] or (b) the neural network itself. Therefore, we first investigate all networks’ performance concerning the predicted scatter distributions’ power spectral density. Second, from scatter estimation theory, we know that the scatter distribution can be recovered from the measured signal by the convolution with a so-called scatter kernel [30]. This allows us to interpret a neural network for a specific pair of scatter distribution and X-ray signal as a filtering operation and assess its frequency response. #### III-C2 Noise Analysis Assessing a network’s robustness is inherently difficult given small training corpora. Therefore, adversarial attacks are often used to expose weaknesses deliberately. Since we trained all networks on noise-free data, testing them on data with different unseen noise levels appeared appropriate. To this end, we applied Poisson noise associated with different photon counts ranging from ${10}^{3}{10}^{5}$ to our head dataset and investigated the accuracy of the networks’ predictions. #### III-C3 Runtime Analysis For interventional applications, the fast execution speed of computer programs is essential. Therefore, we benchmarked the average execution time for all networks for different batch sizes using a 12-core CPU (Intel(R) Xeon(R) Silver 4116 CPU 2.10GHz). #### III-C4 Real Data Analysis To confirm the generalizability to real data, we tested the networks, which performed best on our synthetic thorax dataset, on the real thorax dataset. For reference, we considered various different scatter suppression techniques, namely using an anti-scatter grid or slit scanning. As an almost scatter free baseline, we considered the configuration to use both, an anti-scatter grid and slit scanning. Note that we neither performed an intensity or a geometry calibration in between the scans to account for the missing anti-scatter grid. ## IV Results ### IV-A Meta-Parameter Search #### IV-A1 Network Parametrization $\displaystyle 10^{5}$$\displaystyle 10^{6}$$\displaystyle 10^{7}$$\displaystyle 10^{8}$No. parameters$\displaystyle 6$$\displaystyle 7$$\displaystyle 8$$\displaystyle 9$MAPE [$\displaystyle\%$]DU-netSU-netOurs Figure 2: Distribution of different network configurations for our approach and the U-net in terms of absolute percentage errors averaged over all folds and patients. Note that, due to visualization purposes, the standard deviation encoded by the circular margins does not correspond to absolute values but rather indicates the relative spread between the networks. The standard deviation is in the range of $1\text{\,}\mathrm{\char 37\relax}5\text{\,}\mathrm{\char 37\relax}$. The x-axis refers to the number of parameters in the convolutional layers. Figure 2 establishes the relationship between the number of convolutional parameters to train for each network to the averaged error rates of all folds and patients. All networks’ error rates range between $6\text{\,}\mathrm{\char 37\relax}9\text{\,}\mathrm{\char 37\relax}$, and we observed that overall the more compact networks outperform the DU-nets. Based on these findings, we selected two spline networks ($c=16$, $(d,p)\in\\{(4,1),(5,0)\\}$) and four U-nets (deep and shallow, $c=16$, $d\in\\{6,7\\}$) for further investigations. #### IV-A2 Bottleneck 1234Fold05101520MAPE [$\displaystyle\%$]constrainedunconstrainedfc-netconv-net Figure 3: Boxplots of the mean absolute percentage error (MAPE) of our proposed method for different bottleneck architectures averaged over the validation folds for each training fold. The horizontal lines indicate the mean value over all test and validation folds. $\displaystyle 0$$\displaystyle 200$$\displaystyle 0$$\displaystyle 50$constrained$\displaystyle 0$$\displaystyle 200$$\displaystyle 0$$\displaystyle 50$unconstrained Figure 4: Normalized constrained (left) and unconstrained (right) weighting matrices used to map the latent variables $\boldsymbol{Z}$ to spline coefficients $\boldsymbol{C}$. Figure 3 shows the results for the different bottleneck architectures. Our proposed constrained weighting matrix, homogeneously initialized, achieves the best results on average, even surpassing the convolutional bottleneck followed by a fully-connected layer. The unconstrained fully-connected layer architectures yield considerably worse results. Figure 4, which shows the weighting matrix for the constrained and unconstrained case, substantiates this finding. While the constrained matrix converges to a block circulant matrix, which merely relates to a convolution, the unconstrained one hardly resembles a sensible operation and is overall noisy. ### IV-B Qualitative and Quantitative Results Head data | Thorax data ---|--- 1234Fold05101520MAPE [$\displaystyle\%$] | 12345Fold05101520MAPE [$\displaystyle\%$] 1234Fold0.950.960.970.991.00SSIM | 12345Fold0.950.960.970.991.00SSIM Figure 5: Quantitative boxplots for our synthetic datasets. The top row shows the mean absolute percentage errors (MAPE) with respect to the scatter ground truth for each test fold. The bottom row shows structural similarity indices (SSIM) between reconstructed volumes from the simulated ideal primary signal and the scatter-corrected ones using neural networks. The horizontal lines indicate the overall average across all folds. For the sake of clarity, we only depict the best performing networks in each category. For the head dataset, we depict the DU-net with $d=6$, SU-net with $d=7$, and our method with $d=4$ and $p=1$. For the thorax dataset, we depict the DU-net with $d=7$, SU-net with $d=6$, and our method with $d=5$ and $p=0$. Figure 6: Selected subjects of both datasets. The results of the learning- based approaches are given in terms of error maps using the absolute percentage error (APE) for scatter and the absolute error (AE) for the reconstructed slices. For convenience, the associated mean APE (MAPE) and mean AE (MAE) are also provided below each output. Figure IV-B shows fold-wise boxplots for the two synthetic datasets. We observe similar error rates of approximately $5\text{\,}\mathrm{\char 37\relax}$ across all folds and network configurations regarding the predicted scatter distributions for the head dataset. In general, all networks achieve high SSIM values above 0.99 when comparing the reconstructed volumes from the simulated ideal primary signal to their counterparts obtained using the scatter-corrected projections. Note that for the head data, our approach performed equally well for all folds, whereas the results obtained with the U-nets varied more widely. Processing the thorax datasets, on the other hand, was more challenging. Again, all networks performed comparably well with error rates of approximately $7.5\text{\,}\mathrm{\char 37\relax}$. In comparison to the head dataset, larger error margins and outliers were found. This can also be seen with the SSIM, which is, on average, just below 0.98. Again, we find that our proposed method is on par with U-net-like structures from a quantitative point of view. Looking at selected subjects of both datasets in Fig. 6 reveals the potential advantages of the proposed approach. Modeling the predicted scatter as a B-spline intrinsically limits the network output to smooth surfaces. In contrast, both U-nets preserve some details of the input, especially the shallow architecture. However, the reconstructed slices show no systematic trend, and all methods can adequately recover the desired signal. ### IV-C In-Depth Analysis #### IV-C1 Spectral Analysis $\displaystyle 10^{-2}$$\displaystyle 10^{-1}$$\displaystyle 10^{0}$Normalized frequency$\displaystyle 10^{-2}$$\displaystyle 10^{-1}$$\displaystyle 10^{0}$Normalized powerGTDU-netSU-netOurs Figure 7: Power spectral density plots for the X-ray scatter distributions of the ground truth (black solid), ours (light orange dashed), and the U-net (light purple dotted) averaged over all test folds. Note that the graphs associated with both U-nets are almost identical. We calculated the power spectral density by azimuthally averaging the magnitudes of the 2D Fourier transform of X-ray scatter distributions. Also, we averaged all power spectral densities of all projection, patients, and test folds. The resulting densities plotted in Fig. 7 support our previous findings. While the U-nets yield numerically sound predictions, they systematically boost the high frequencies in the scatter distributions. Our approach, however, preserves the real power spectral density over the whole frequency spectrum. AmplitudeGTDU-netSU- netOurs$\displaystyle-3$$\displaystyle-2$$\displaystyle-1$$\displaystyle 0$Phase$\displaystyle 0$$\displaystyle 1$$\displaystyle 2$ Figure 8: Frequency responses of the systems for one test patient averaged over all validation folds. From top to bottom: Normalized $\log$-amplitude, phase in radians. From left to right: ideal system (GT), our spline-net (Ours), deep U-net (DU-net), shallow U-net (SU-net). As mentioned above, conventional convolutional neural networks can be interpreted for a single input image in terms of a filtering operation. Thus, we divided the Fourier transform of the output by the Fourier transform of the input to obtain the respective frequency responses. Figure 8 shows the amplitude and phase of an ideal system’s frequency responses, for our method and both U-nets, averaged over all projections for one patient. Our spline- net’s frequency response was closer to the ideal frequency response in both, amplitude and phase. However, we observed a noticeable intensity shift in our method. The U-nets, in contrast, both exhibited larger deviations in the patterns of amplitude and phase, indicating that their represented operation is less predictable. #### IV-C2 Noise Analysis $\displaystyle 10^{3}$$\displaystyle 10^{4}$$\displaystyle 10^{5}$Photon count$\displaystyle 5.0$$\displaystyle 7.5$$\displaystyle 10.0$$\displaystyle 12.5$$\displaystyle 15.0$$\displaystyle 17.5$MAPE [$\displaystyle\%$]DU-netSU- netOurs Figure 9: Error rates of predicted scatter distributions for different noise levels averaged over all validation and test fold networks. All networks have been trained with noise-free data. Note that a lower photon count relates to a higher noise level. Figure 9 shows plots of the networks’ error rates when confronted with noise. We could confirm that the U-net is very sensitive to unseen noise levels in both configurations, whereas our approach performs more robustly. #### IV-C3 Runtime Analysis 12481632Batch size$\displaystyle 0$$\displaystyle 25$$\displaystyle 50$$\displaystyle 75$$\displaystyle 100$$\displaystyle 125$$\displaystyle 150$Runtime [ms]DU-net 6DU-net 7SU-net 6SU-net 7Ours 4\1Ours 5\0 Figure 10: Runtimes per projection in $\mathrm{ms}$ for different architectures (specified by the depth and optional pre-pooling, $d$ \ $p$) and batch sizes. Figure 10 compiles the inference speed of all considered networks. As expected from the parameter complexity, our approach is the fastest with $4\text{\,}\mathrm{ms}30\text{\,}\mathrm{ms}$ and therefore $1.78.5$ times faster than the SU-net with $34\text{\,}\mathrm{ms}50\text{\,}\mathrm{ms}$. The DU-net is the slowest with $89\text{\,}\mathrm{ms}144\text{\,}\mathrm{ms}$. #### IV-C4 Real Data Analysis ReconstructionGrid+SlitGrid+FullSlitFullDU-netSU- netOursDifference$\displaystyle 39.31\pm 0.21$$\displaystyle 58.46\pm 0.09$$\displaystyle 123.84\pm 0.32$$\displaystyle 62.86\pm 0.25$$\displaystyle 64.52\pm 0.23$$\displaystyle 63.97\pm 0.19$ Figure 11: Central slices of reconstructed volumes for different scatter compensation strategies and errors taken with respect to the reconstruction result resulting from the grid + slit data acquisition. Grid refers to employing a conventional anti-scatter grid. Slit refers to the most narrow collimator setting available on the X-ray imaging system. Full refers to using no collimation at all. Both, the reconstructed slices as well as the difference images are shown using a gray level window $[-1000,1000]$ HU. For convenience the absolute average HU errors and the associated standard deviations with respect to three consecutive measurements are provided. As shown in Fig. 11, all methods were able to compensate for scatter artifacts well. However, employing an anti-scatter grid still yielded the lowest overall error in Hounsfield units (HU) as compared to using slit collimation in addition. The learning-based approaches performed about as well as the slit scanning without an anti-scatter grid, which is in accordance to previous findings [37]. Overall, the networks achieved similar error rates, and no systematic trend was observable. ## V Discussion X-ray scatter is a major source of artifacts in interventional CBCT. Deep- learning-based approaches have shown the potential to outperform conventional physical or algorithmic approaches to scatter compensation. Without incorporating prior information, scatter distributions can be inferred from the measured X-ray signal [37]. However, data integrity and robustness are critical aspects of clinical imaging, which can be violated by deep neural networks [54]. To ensure sound scatter estimates, we proposed to embed bivariate B-splines in neural networks to constrain their co-domain to smooth results [42]. By reformulating the spline evaluation in terms of matrix multiplications, we were able to integrate B-splines with neural networks without further ado such that end-to-end training was feasible. In an extensive cross-validation using synthetic data, we showed that our proposed lean convolutional encoder using B-spline evaluation performs on par with several U-net based architectures. We substantiated this finding in a first phantom study. There, our approach performed basically as well as the U-net and the slit scanning technique (without anti-scatter grid), which was used as a baseline before [37]. Note, however, that the proposed method offers several advantages not present in the U-net architectures. First, we considerably lowered the parameter and runtime complexity, rendering our method suitable for a variety of hardware. More importantly, we verified that our spline-based approach ensures data integrity concerning the power spectral density of scatter estimates and the overall network’s frequency response. This property ensures that no high- frequency details, which relate to anatomic structures or pathologies, are altered. In comparison, the U-net considerably changes the concerning spectral contents, which was already shown for neural networks containing up- convolutions [53]. As our spline network corresponds to a low-pass filtering operation, it is robust towards noise even when trained on noise-free data. While we implemented the U-net baselines to the best of our knowledge, our error rates are higher than previously reported [37]. Potential reasons include but are not limited to (a) different simulation codes, (b) our smaller training corpora, and (c) the heterogeneity of our data. Our simulation setup currently assumes an ideal detector, and we do not consider the domain shift between synthetic and real data. For future work, we indicate several research directions for either method or data and experiments. First, since our preferred bottleneck weighting matrix converges to a block circulant matrix, a fixed representation is desirable to further reduce the number of trainable parameters and increase the plausibility of our approach. For instance, training both the encoder and the bottleneck separately or introducing additional constraints are promising approaches to do so. Second, replacing the bottleneck fully-connected layer with a small U-net reduces the number of parameters while still covering the entire latent space with its receptive field. Third, since our network is end- to-end trainable, it appears reasonable to include the reconstruction into the computational graph to calculate the loss function in the CBCT domain [55, 56]. Last but not least, we believe our approach applies to low-frequency signal estimation and correction in general, e.g., bias field correction in magnetic resonance imaging, ultrasound imaging, or microscopy techniques. ## VI Conclusion Embedding B-splines in neural networks ensures data integrity for low- frequency signals. This reduces the number of network parameters needed to arrive at physically sensible results, and thanks to the reduced parameter set, network inference can be made faster. ## Disclaimer The concepts and information presented in this article are based on research and are not commercially available. ## References * [1] E.-P. Rührnschopf and K. Klingenbeck, “A general framework and review of scatter correction methods in x-ray cone-beam computerized tomography. Part 1: Scatter compensation approaches,” _Med Phys_ , vol. 38, no. 7, pp. 4296–4311, 2011. * [2] ——, “A general framework and review of scatter correction methods in cone-beam CT. Part 2: Scatter estimation approaches,” _Med Phys_ , vol. 38, no. 9, pp. 5186–5199, 2011. * [3] F. Groedel and R. Wachter, “Bedeutung der Röhren-Fern- und Platten-Abstandsaufnahmen,” _Verh d Dt Röntgengesellschaft_ , vol. 17, pp. 134–135, 1926. * [4] F. Janus, “Die Bedeutung der gestreuten ungerichteten Röntgenstrahlen fur die Bildwirkung und ihre Beseitigung,” _Verh d Dt Röntgengesellschaft_ , vol. 17, pp. 136–139, 1926. * [5] U. Neitzel, “Grids or air gaps for scatter reduction in digital radiography: A model calculation,” _Med Phys_ , vol. 19, no. 2, pp. 475–481, 1992. * [6] J. Persliden and G. A. Carlsson, “Scatter rejection by air gaps in diagnostic radiology. calculations using a Monte Carlo collision density method and consideration of molecular interference in coherent scattering,” _Phys Med Biol_ , vol. 42, no. 1, pp. 155–175, 1997. * [7] G. T. Barnes, “Contrast and scatter in x-ray imaging.” _RadioGraphics_ , vol. 11, no. 2, pp. 307–323, 1991. * [8] R. Bhagtani and T. G. Schmidt, “Simulated scatter performance of an inverse-geometry dedicated breast CT system,” _Med Phys_ , vol. 36, no. 3, pp. 788–796, 2009. * [9] T. G. Schmidt, R. Fahrig, N. J. Pelc, and E. G. Solomon, “An inverse-geometry volumetric CT system with a large-area scanned source: A feasibility study,” _Med Phys_ , vol. 31, no. 9, pp. 2623–2627, 2004. * [10] G. Bucky, “Über die Ausschaltung der im Objekt entstehenden Sekundärstrahlen bei Röntgenstrahlen,” _Verh d Dt Röntgengesellschaft_ , vol. 9, pp. 30–32, 1913. * [11] W. A. Kalender, “Calculation of x-ray grid characteristics by Monte Carlo methods,” _Phys Med Biol_ , vol. 27, no. 3, pp. 353–361, 1982. * [12] H.-P. Chan and K. Doi, “Investigation of the performance of antiscatter grids: Monte Carlo simulation studies,” _Phys Med Biol_ , vol. 27, no. 6, pp. 785–803, 1982. * [13] H.-P. Chan, Y. Higashida, and K. Doi, “Performance of antiscatter grids in diagnostic radiology: Experimental measurements and Monte Carlo simulation studies,” _Med Phys_ , vol. 12, no. 4, pp. 449–454, 1985. * [14] H.-P. Chan, K. L. Lam, and Y. Wu, “Studies of performance of antiscatter grids in digital radiography: Effect on signal-to-noise ratio,” _Med Phys_ , vol. 17, no. 4, pp. 655–664, 1990. * [15] H. Aichinger, J. Dierker, S. Joite-Barfuss, and M. Saebel, _Radiation exposure and image quality in X-ray diagnostic radiology_. Berlin, New York: Springer, 2004. * [16] D. M. Gauntt and G. T. Barnes, “Grid line artifact formation: A comprehensive theory,” _Med Phys_ , vol. 33, no. 6, pp. 1668–1677, 2006. * [17] V. Singh, A. Jain, D. R. Bednarek, and S. Rudin, “Limitations of anti-scatter grids when used with high resolution image detectors,” in _Proc SPIE Int Soc Opt Eng_ , B. R. Whiting and C. Hoeschen, Eds., vol. 9033, International Society for Optics and Photonics. SPIE, 2014, pp. 1618–1626. * [18] R. Rana, A. Jain, A. Shankar, D. R. Bednarek, and S. Rudin, “Scatter estimation and removal of anti-scatter grid-line artifacts from anthropomorphic head phantom images taken with a high resolution image detector,” in _Proc SPIE Int Soc Opt Eng_ , D. Kontos and T. G. Flohr, Eds., vol. 9783, International Society for Optics and Photonics. SPIE, 2016, pp. 1619–1628. * [19] A. Bani-Hashemi, E. Blanz, J. Maltz, D. Hristov, and M. Svatos, “Tu-d-i-611-08: Cone beam x-ray scatter removal via image frequency modulation and filtering,” _Med Phys_ , vol. 32, no. 6, pp. 2093–2093, 2005\. * [20] L. Zhu, N. R. Bennett, and R. Fahrig, “Scatter correction method for x-ray CT using primary modulation: Theory and preliminary results,” _IEEE Trans Med Imaging_ , vol. 25, no. 12, pp. 1573–1587, 2006. * [21] B. Bier, M. Berger, A. Maier, M. Kachelrieß, L. Ritschl, K. Müller, J.-H. Choi, and R. Fahrig, “Scatter correction using a primary modulator on a clinical angiography c-arm CT system,” _Med Phys_ , vol. 44, no. 9, pp. 125–137, 2017. * [22] W. Zbijewski and F. J. Beekman, “Fast scatter estimation for cone-beam x-ray CT by combined Monte Carlo tracking and richardson-lucy fitting,” in _IEEE Nucl Sci Symp Conf Rec_ , vol. 5, 2004, pp. 2774–2777. * [23] G. Poludniowski, P. M. Evans, V. N. Hansen, and S. Webb, “An efficient Monte Carlo-based algorithm for scatter correction in keV cone-beam CT,” _Phys Med Biol_ , vol. 54, no. 12, pp. 3847–3864, 2009. * [24] A. Wang, A. Maslowski, P. Messmer, M. Lehmann, A. Strzelecki, E. Yu, P. Paysan, M. Brehm, P. Munro, J. Star-Lack, and D. Seghers, “Acuros cts: A fast, linear boltzmann transport equation solver for computed tomography scatter – part ii: System modeling, scatter correction, and optimization,” _Med Phys_ , vol. 45, no. 5, pp. 1914–1925, 2018. * [25] A. Badal and A. Badano, “Accelerating Monte Carlo simulations of photon transport in a voxelized geometry using a massively parallel graphics processing unit,” _Med Phys_ , vol. 36, no. 11, pp. 4878–4880, 2009. * [26] M. Baer and M. Kachelrieß, “Hybrid scatter correction for CT imaging,” _Phys Med Biol_ , vol. 57, no. 21, pp. 6849–6867, 2012. * [27] W. Swindell and P. M. Evans, “Scattered radiation in portal images: A Monte Carlo simulation and a simple physical model,” _Med Phys_ , vol. 23, no. 1, pp. 63–73, 1996. * [28] W. Yao and K. W. Leszczynski, “An analytical approach to estimating the first order scatter in heterogeneous medium. ii. a practical application,” _Med Phys_ , vol. 36, no. 7, pp. 3157–3167, 2009. * [29] M. Meyer, W. A. Kalender, and Y. Kyriakou, “A fast and pragmatic approach for scatter correction in flat-detector CT using elliptic modeling and iterative optimization,” _Phys Med Biol_ , vol. 55, no. 1, pp. 99–120, 2009\. * [30] B. Ohnesorge, T. Flohr, and K. Klingenbeck-Regn, “Efficient object scatter correction algorithm for third and fourth generation CT scanners,” _Eur Radiol_ , vol. 9, no. 3, pp. 563–569, Mar 1999. * [31] H. Li, R. Mohan, and X. R. Zhu, “Scatter kernel estimation with an edge-spread function method for cone-beam computed tomography imaging,” _Phys Med Biol_ , vol. 53, no. 23, pp. 6729–6748, 2008. * [32] M. Sun and J. M. Star-Lack, “Improved scatter correction using adaptive scatter kernel superposition,” _Phys Med Biol_ , vol. 55, no. 22, pp. 6695–6720, 2010. * [33] T. Q. Chen, Y. Rubanova, J. Bettencourt, and D. K. Duvenaud, “Neural ordinary differential equations,” in _Adv Neural Inf Process Syst_ , S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, Eds. Curran Associates, Inc., 2018, pp. 6571–6583. * [34] P. Roser, X. Zhong, A. Birkhold, N. Strobel, M. Kowarschik, R. Fahrig, and A. Maier, “Physics-driven learning of x-ray skin dose distribution in interventional procedures,” _Med Phys_ , vol. 46, no. 10, pp. 4654–4665, 2019. * [35] P. Roser, X. Zhong, A. Birkhold, A. Preuhs, C. Syben, E. Hoppe, N. Strobel, M. Kowarschik, R. Fahrig, and A. Maier, “Simultaneous estimation of x-ray back-scatter and forward-scatter using multi-task learning,” in _International Conference on Medical Image Computing and Computer-Assisted Intervention_. Springer, 2020, pp. 199–208. * [36] O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional networks for biomedical image segmentation,” in _Med Image Comput Comput Assist Interv_ , ser. Lecture notes in computer science, vol. 9351. Springer, 2015, pp. 234–241. * [37] J. Maier, E. Eulig, T. Vöth, M. Knaup, J. Kuntz, S. Sawall, and M. Kachelrieß, “Real-time scatter estimation for medical CT using the deep scatter estimation: Method and robustness analysis with respect to different anatomies, dose levels, tube voltages, and data truncation,” _Med Phys_ , vol. 46, no. 1, pp. 238–249, 2019. * [38] A. Maier, C. Syben, B. Stimpel, T. Würfl, M. Hoffmann, F. Schebesch, W. Fu, L. Mill, L. Kling, and S. Christiansen, “Learning with known operators reduces maximum error bounds,” _Nat Mach Intell_ , vol. 1, pp. 373–380, 2019\. * [39] B. Stimpel, C. Syben, F. Schirrmacher, P. Hoelter, A. Dörfler, and A. Maier, “Multi-modal deep guided filtering for comprehensible medical image processing,” _IEEE Trans Med Imaging_ , vol. 39, no. 5, pp. 1703–1711, 2019. * [40] G. H. Glover, “Compton scatter effects in CT reconstructions,” _Med Phys_ , vol. 9, no. 6, pp. 860–867, 1982. * [41] J. M. Boone and J. A. Seibert, “An analytical model of the scattered radiation distribution in diagnostic radiology,” _Med Phys_ , vol. 15, no. 5, pp. 721–725, 1988. * [42] P. Roser, A. Birkhold, A. Preuhs, C. Syben, N. Strobel, M. Korwarschik, R. Fahrig, and A. Maier, “Deep scatter splines: Learning-based medical x-ray scatter estimation using b-splines,” in _The 6th International Conference on Image Formation in X-Ray Computed Tomography (CT-Meeting)_ , 2020\. * [43] K. Qin, “General matrix representations for b-splines,” _The Visual Computer_ , vol. 16, no. 3-4, pp. 177–186, 2000. * [44] X. Glorot, A. Bordes, and Y. Bengio, “Deep sparse rectifier neural networks,” in _Proceedings of the fourteenth international conference on artificial intelligence and statistics_ , 2011, pp. 315–323. * [45] K. Clark, B. Vendt, K. Smith, J. Freymann, J. Kirby, P. Koppel, S. Moore, S. Phillips, D. Maffitt, M. Pringle, L. Tarbox, and F. Prior, “The Cancer Imaging Archive (TCIA): Maintaining and operating a public information repository,” _J Digit Imaging_ , vol. 26, pp. 1045–1057, 07 2013. * [46] T. Bejarano, M. De Ornelas Couto, and I. Mihaylov, “Head-and-neck squamous cell carcinoma patients with CT taken during pre-treatment, mid-treatment, and post-treatment dataset. the cancer imaging archive,” 2018. [Online]. Available: http://doi.org/10.7937/K9/TCIA.2018.13upr2xf * [47] H. Roth, L. Le, S. Ari, K. Cherry, J. Hoffman, S. Wang, and R. Summers, “A new 2.5 d representation for lymph node detection in ct. the cancer imaging archive,” 2018. [Online]. Available: http://doi.org/10.7937/K9/TCIA.2015.AQIIDCNM * [48] P. Roser, A. Birkhold, A. Preuhs, B. Stimpel, C. Syben, N. Strobel, M. Kowarschik, R. Fahrig, and A. Maier, “Fully-automatic CT data preparation for interventional x-ray skin dose simulation,” in _Bildverarbeitung für die Medizin 2020_. Springer, 2020, pp. 125–130. * [49] W. Schneider, T. Bortfeld, and W. Schlegel, “Correlation between CT numbers and tissue parameters needed for Monte Carlo simulations of clinical dose distributions,” _Phys Med Biol_ , vol. 45, no. 2, pp. 459–478, 2000. * [50] L. He, X. Ren, Q. Gao, X. Zhao, B. Yao, and Y. Chao, “The connected-component labeling problem: A review of state-of-the-art algorithms,” _Pattern Recognit_ , vol. 70, pp. 25–43, 2017. * [51] X. Glorot and Y. Bengio, “Understanding the difficulty of training deep feedforward neural networks,” in _Proceedings of the thirteenth international conference on artificial intelligence and statistics_ , 2010, pp. 249–256. * [52] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” in _3rd Int. Conf. on Learning Representations_ , Y. Bengio and Y. LeCun, Eds., 2015. * [53] R. Durall, M. Keuper, and J. Keuper, “Watch your up-convolution: Cnn based generative deep neural networks are failing to reproduce spectral distributions,” in _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 2020, pp. 7890–7899. * [54] Y. Huang, T. Würfl, K. Breininger, L. Liu, G. Lauritsch, and A. Maier, “Some investigations on robustness of deep learning in limited angle tomography,” in _Med Image Comput Comput Assist Interv_. Springer, 2018, pp. 145–153. * [55] C. Syben, M. Michen, B. Stimpel, S. Seitz, S. Ploner, and A. K. Maier, “Pyro-NN: Python reconstruction operators in neural networks,” _Med Phys_ , vol. 46, no. 11, pp. 5110–5115, 2019. * [56] C. Syben, B. Stimpel, P. Roser, A. Dörfler, and A. Maier, “Known operator learning enables constrained projection geometry conversion: Parallel to cone-beam for hybrid MR/x-ray imaging,” _IEEE Trans Med Imaging_ , 2020\.
# On the Negativity of Wigner Function as a measure of entanglement under quantum polarization converter devices Mustapha<EMAIL_ADDRESS>and Morad El <EMAIL_ADDRESS> Université Mohammed V, Faculté des Sciences Equipe Sciences de la Matière et du Rayonnement Av. Ibn Battouta, B.P. 1014, Agdal, Rabat, Morocco ###### Abstract We study the behaviour of the Negativity of Wigner Function (NWF) as a measure of entanglement in non-Gaussian states under quantum polarisation converter devices. We analyze comparatively this quantity with other measures of entanglement in a system prepared in a superposition of two-mode coherent states. We show that the (WF) can be identified as a quantifier of non- Gaussian entanglement. ## Keywords Wigner function, Q-function, Negativity, Non-Gaussian state, Non-gaussianity, Coherent states, Quantum polarization. ## Introduction Quantum entanglement, is recognised as one of quantum resources for several applications in quantum information theory including quantum computing, quantum communications and quantum cryptography. However, quantifying quantum correlations particularly, entanglement is one of the most relevant challenges in quantum information theory. On the other hand polarization play a central role in a large number of optical phenomena. It appear in several paradigmatic application including remote sensing and light scattering [., 3 abdo] In recent years, a considerable attention has been paid to the polarization of quantum light due to its application in quantum information protocols since the light can be extensively used in the quantum information-coding. There is a wide consensus that coherent states are the most classical quantum states. It has been shown that these states minimize the Heisenberg uncertainty relation for position and momentums operators. In addition the dynamic of their expectation values has the same form as their classical counterpart. These properties, make coherent states least quantum states. For this reason, they are called quasi-classical states. These states describe coherent optical light and can be generated in the laboratory. In this way, the quantum polarization formalism is extended from that of classical light polarization through replacing the Stokes parameters by the associated stocks operators. In addition, the degree of polarization of classical light does not depend on its intensity. However, the degree of polarization of coherent states decreases with decreasing values of optical power. Furthermore, the quantum Stokes parameters $\hat{S}_{1},\hat{S}_{2}$ and $\hat{S}_{3}$ do not commute with each other. Then, it is not possible to know the values of any two of them simultaneously without uncertainties. This fact has been used in [1] in order to create a continuous variable quantum key distribution system. Basically, polarization of coherent states has extensively been used in quantum information processing, precisely in quantum key distribution protocols with continuous variables. [1, 2, 3, 4]. Recently, one of the most important applications of the Wigner function in quantum information theory is the classification of classical and non- classical states based on its non-Gaussianity and its negativity [5, 6]. In fact, a Gaussian state is associated with a Gaussian Wigner function in phase space of one mode or multimode pure state in continuous variables systems [7]. On the other hand, a non-Gaussian state with negative Wigner function reveals that the system possesses non-classical correlations [8]. Recent works showed that the negativity of the Wigner function can detect the entanglement and it is not sensitive to all kinds of quantum correlations and can be a best quantifier of genuine entanglement in tripartite systems [9, 10]. In this work, we study the strength of the NWF as a quantifier of entanglement under a quantum polarization converter devices. In this direction we describe in section (1) the Stokes parameters in order to analyse the quantum polarization of the superposition of two bi-mode coherent states; As well in section (3), we use the entanglement of formation and the NWF to analyse the entanglement behavior before and after the polarization converter devices; Finally, in section (4) we discuss our results and provide conclusions. ## 1 Review of quantum polarization In classical optics the polarization of light beams is determined by computing the four Stokes parameters [11]. Similarly, In quantum optics the degree of polarization of states of light is quantified by the calculation of the mean values of the associated Stocks operators[12, 13]. For a monochromatic plane wave propagating in the $z$-direction, whose electric field lines in the $xy$ plane. In terms of the annihilation and creation operators of horizontally and vertically polarized modes noted by $\hat{a}_{H}$ and $\hat{a}_{V}$, respectively, the Stokes operators can be expressed as[12, 13] $\begin{matrix}\hat{S}_{0}&=&\hat{a}_{H}^{+}\hat{a}_{H}+\hat{a}_{V}^{+}\hat{a}_{V},\\\ \hat{S}_{1}&=&\hat{a}_{H}^{+}\hat{a}_{H}-\hat{a}_{V}^{+}\hat{a}_{V},\\\ \hat{S}_{2}&=&\hat{a}_{H}^{+}\hat{a}_{V}+\hat{a}_{V}^{+}\hat{a}_{H},\\\ \hat{S}_{3}&=&i\left(\hat{a}_{V}^{+}\hat{a}_{H}+\hat{a}_{H}^{+}\hat{a}_{V}\right)\end{matrix}$ (1) and the Stokes parameters are given by calculation of the corresponding average values $<\hat{S}_{k}>$. Using the bosonic commutation relations $\left[\hat{a}_{i},\hat{a}_{j}^{+}\right]=\hat{\textbf{1}}\delta_{ij}\quad;\quad\\{i,j\\}\in\\{H,V\\}.$ (2) It has been shown that, the operators $\hat{S}_{1}$, $\hat{S}_{2}$ and $\hat{S}_{3}$ all commute with $\hat{S}_{0}$ and satisfy the corresponding commutation relations, $\left[\hat{S}_{k},\hat{S}_{l}\right]=2i\hat{S}_{m}\quad;\quad\\{k,l,m\\}\in\\{1,2,3\\}.$ (3) The Stokes operators $\hat{S}_{1}$, $\hat{S}_{2}$ and $\hat{S}_{3}$ thus from $SU\left(2\right)$ algebra and generate all transformations from this group; * • $\hat{S}_{2}$ is the infinitesimal generator of geometric rotations around the direction of propagation, * • $\hat{S}_{3}$ is the differential phase shifts between the two modes. For a quasi-classical two-mode coherent state $\lvert\alpha,\beta\rangle$ defined as $\lvert\alpha,\beta\rangle=\mathrm{e}^{-\dfrac{\lvert\alpha\rvert+\lvert\beta\rvert}{2}}\sum_{n,m}\dfrac{\left(\alpha\right)^{n}}{\sqrt{n!}}\dfrac{\left(\beta\right)^{m}}{\sqrt{m!}}\lvert n,m\rangle,$ (4) the mean values $<\hat{S}_{i}>$ of the three Stokes operators and the variances $V_{i}$ are expressed as a function of the field amplitudes: $\begin{matrix}<\hat{S}_{1}>=&\lvert\alpha\rvert^{2}-\lvert\beta\rvert^{2},&<\hat{S}_{1}^{2}>=&\left(\left(\lvert\alpha\rvert\right)^{2}-\left(\lvert\beta\rvert\right)^{2}\right)^{2}+\left(\lvert\alpha\rvert\right)^{2}+\left(\lvert\beta\rvert\right)^{2},\\\ <\hat{S}_{2}>=&\alpha^{*}\beta-\alpha\beta^{*},\quad&<\hat{S}_{2}^{2}>=&\left(\alpha^{*}\beta\right)^{2}+\left(\alpha\beta^{*}\right)^{2}+\lvert\alpha\rvert^{2}+\lvert\beta\rvert^{2}+2\lvert\alpha\rvert^{2}\lvert\beta\rvert^{2},\\\ <\hat{S}_{3}>=&i\left(\alpha\beta^{*}-\alpha^{*}\beta\right),\quad&<\hat{S}_{3}^{2}>=&-\left(\alpha^{*}\beta\right)^{2}-\left(\alpha\beta^{*}\right)^{2}+\lvert\alpha\rvert^{2}+\lvert\beta\rvert^{2}+2\lvert\alpha\rvert^{2}\lvert\beta\rvert^{2},\\\ \end{matrix}$ (5) and $V_{1}=V_{2}=V_{3}=\lvert\alpha\rvert^{2}+\lvert\beta\rvert^{2}.$ The average values of the quantum Stokes parameters of coherent state are equal to of classical light Stokes parameters values and the variance of the three parameters increase while the optical power increases. ## 2 Quantum degree of polarization of superposition of two mode coherent states Classically, the light is considered unpolarized if it’s Stokes parameters vanish. In quantum mechanics, this is condition necessary but not sufficient. A quantum light beam can be considered unpolarized if its observable do not change after a geometric rotation and/or a phase shift between the components. These condition are mathematically described by [14]: $\left[\hat{\rho},\hat{S}_{1}\right]=\left[\hat{\rho},\hat{S}_{3}\right]=0,$ (6) where $\hat{\rho}$ is the density matrix of the quantum state. By analogy with the classical definition, many measures of the quantum polarization degree have been proposed [13, 12]. Here we consider a measure based on Q-function [13]: $P=\dfrac{D}{1+D}$ (7) with $D=4\pi\int_{0}^{2\pi}d\Omega\int_{0}^{\pi}\left[Q\left(\theta,\phi\right)-\dfrac{1}{4\pi}\right]^{2}\sin\left(\theta\right)d\theta d\phi$ where $d\Omega=\sin\left(\theta\right)d\theta d\phi$ is the differential of solid angle and $Q\left(\theta,\phi\right)$ is the Q-function of the light. For the two-mode coherent state $\lvert\alpha e^{i\phi_{\alpha}},\beta e^{i\phi_{\beta}}\rangle$ the Q-function reads [15]: $Q\left(\theta,\phi\right)=\dfrac{e^{-\left(\lvert\alpha\rvert^{2}+\lvert\beta\rvert^{2}\right)}}{4\pi}\left(1+z\right)e^{2}$ (8) where $\displaystyle z$ $\displaystyle=\left[\lvert\alpha\rvert\cos\left(\dfrac{\theta}{2}\right)\cos\left(\phi_{\alpha}+\phi\right)+\beta\sin\left(\dfrac{\theta}{2}\right)\cos\left(\phi_{\beta}\right)\right]^{2}$ $\displaystyle+\left[\lvert\alpha\rvert\cos\left(\dfrac{\theta}{2}\right)\sin\left(\phi_{\alpha}+\phi\right)+\beta\sin\left(\dfrac{\theta}{2}\right)\sin\left(\phi_{\beta}\right)\right]^{2}.$ Using (7) and (8), the quantum polarization degree of the quantum state $\lvert\alpha,0\rangle$ is $P=1-\dfrac{4\lvert\alpha\rvert^{2}}{1+2\lvert\alpha\rvert}.$ (9) When $\rvert\alpha\lvert^{2}\gg 1$, equation (9) can be approximated by $P\simeq 1-\dfrac{2}{\rvert\alpha\lvert^{2}}$ (10) Showing that the quantum degree of polarization of two-mode coherent state increase with increasing of the light power. In Figure (2) we show the quantum degree of polarization of the states $\left\lvert 0,\pm\alpha\right\rangle,$ $\left\lvert\pm\alpha,0\right\rangle,$ $\left\lvert\pm\alpha,\mp\alpha\right\rangle$ and $\left\lvert\pm\alpha,\pm\alpha\right\rangle$ that are equivalent to vertical polarization $\left\lvert V\right\rangle$, horizontal polarization $\left\lvert H\right\rangle$, anti-diagonal polarization $\left\lvert-\dfrac{\pi}{4}\right\rangle$ and diagonal polarization $\left\lvert\dfrac{\pi}{4}\right\rangle$, respectively. In Figure (2) the diagonal states have a larger quantum degree of polarization, compared with horizontal and vertical states, because they have a larger mean photon number (in total). In order to discuss the quantum degree of polarization, we consider the quantum Stokes parameters of quantum state composed by the superposition of bimodal coherent states defined by $\left\lvert\psi_{\pm}\right\rangle=N\left(\left\lvert\alpha,\beta\right\rangle\pm\left\lvert\gamma,\lambda\right\rangle\right)$ (11) where $\rvert N\lvert^{2}=\left\\{2+\left(\zeta+\zeta^{*}\right)\emph{Exp}\left[-\left(\lvert\alpha\rvert^{2}+\lvert\beta\rvert^{2}+\lvert\gamma\rvert^{2}+\lvert\lambda\rvert^{2}\right)/2\right]\right\\}^{-1}$ and $\zeta=\emph{Exp}\left(\alpha^{*}\gamma+\beta^{*}\lambda\right)$. The average of the quantum Stokes parameters and of their squared values and the Q-function of the state (11) are given in Appendix (A). For the raison of simplification, we choose to consider the following particular cases of the state (11), $\displaystyle\left\lvert\psi_{1}\right\rangle=$ $\displaystyle N_{1}\left(\left\lvert\alpha,\beta\right\rangle+\left\lvert\beta,\alpha\right\rangle\right)$ (12a) $\displaystyle\left\lvert\psi_{2}\right\rangle=$ $\displaystyle N_{2}\left(\left\lvert-\alpha,-\alpha\right\rangle+\left\lvert\alpha,\alpha\right\rangle\right)$ (12b) $\displaystyle\left\lvert\psi_{3}\right\rangle=$ $\displaystyle N_{3}\left(\left\lvert\alpha,0\right\rangle+\left\lvert 0,\alpha\right\rangle\right)$ (12c) with $N_{1}=\left\\{2\left[1+\textnormal{Exp}\left(2\alpha\beta-\lvert\alpha\rvert^{2}-\lvert\beta\rvert^{2}\right)\right]\right\\}^{-1/2}$, $N_{2}=\left\\{2\left[1+\textnormal{Exp}\left(-4\lvert\alpha\rvert^{2}\right)\right]\right\\}^{-1/2}$ and $N_{3}=\left\\{2\left[1+\textnormal{Exp}\left(-\lvert\alpha\rvert^{2}\right)\right]\right\\}^{-1/2}$. The averages and the covariances of the quantum Stokes parameters of the states $\left\lvert\psi_{1}\right\rangle$, $\left\lvert\psi_{2}\right\rangle$ and $\left\lvert\psi_{3}\right\rangle$ are calculated from their general expressions of superposed bi-mode coherent state and given in Appendix (A). we found that $<\hat{S}_{1}>$ and $<\hat{S}_{3}>$ vanish for the states (12a), (12b) and (12c). That is consequence of the fact that the average of optical powers in horizontal and vertical polarizations are equal for these cases. In the separable bi-mode state given in (4), the Stockes parameters and the variances expressed propationally to optical power contrary to the superposed entangled state defined in (11). The entangled state gives an interesting behavior of the variances of the Stokes parameters. The variances can increase and decrease when the total mean photon number increases, as are shown in Figure (1). Figure 1: Variances $V_{1}$, $V_{2}$, and $V_{3}$ of $\hat{S}_{1}$, $\hat{S}_{2}$ and $\hat{S}_{3}$ (respectivly) versus $\lvert\alpha\rvert^{2}$ for $\left\lvert\psi_{1}\right\rangle$ having $\lvert\beta\rvert^{2}=4$. Figure 2: Quantum degree of polarization for the states $\left\lvert 0,\pm\alpha\right\rangle$ (Solidligne) and $\left\lvert\pm\alpha,\mp\alpha\right\rangle$ (Dashedligne). ## 3 Wigner function and entanglement under polarization converter devices Wigner function is an important tool in physics, especially in quantum physics to detect the non-classical behavior of light by studying its negativity. Recently, in quantum information theory the NWF is used as a measure of entanglement[9]. In this section we study the NWF and the entanglement behavior of the superposition of two-mode coherent states before and after pass by a polarization converter devices, in order to testing the strength of the NWF as a measure of entanglement. For a single quantum system described by a density matrix $\hat{\rho}$, the associated Wigner function is defined by $\mathcal{W}\left(q,p\right)=\dfrac{1}{2\pi}\int\exp{(\dfrac{-ipy}{\hbar})}\left\langle q+\dfrac{y}{2}\right\rvert\hat{\rho}\left\lvert q-\dfrac{y}{2}\right\rangle dy,$ (13) where $\lvert q\pm\dfrac{y}{2}\rangle$ are the eigenkets of the position operator. If the state in question is a pure state $\hat{\rho}=\lvert\psi\rangle\langle\psi\rvert$ then $\mathcal{W}\left(q,p\right)=\dfrac{1}{2\pi\hbar}\int\psi^{*}\left(q-\dfrac{y}{2}\right)\psi\left(q+\dfrac{y}{2}\right)\exp{(\dfrac{-ipy}{\hbar})}dy.$ (14) Hence, the doubled volume of the integrated negative part of the Wigner function may be written as [5] $\delta\left(\rho\right)=\iint\left|\mathcal{W}\left(q,p\right)\right|\,dqdp-1.$ (15) (a) Wigner function of $\psi_{1}$ for $\left|\alpha\right|=\left|\beta\right|=1$. (b) Wigner function of $\psi_{+}$ for $\left|\alpha\right|=\left|\beta\right|=2$. Figure 3: Wigner function of superposed coherent state ((12a)). It is clear from the plot in Figure (3) that the Wigner function of the bi- mode superposed state (12a) is not positive on the all phase space. The volume of the negative part of the Wigner function is plotted in Figure (4) versus $\lvert\alpha\lvert^{2}$ for $\lvert\beta\lvert=2.$ The quantum entanglement of the same state ((11)) can be measured by the concurrence [16, 17] $C=\dfrac{\sqrt{\left(1-\lvert\langle\alpha|\gamma\rangle\rvert^{2}\right)\left(1-\lvert\langle\lambda|\beta\rangle\rvert^{2}\right)}}{1+Re\left(\langle\alpha|\gamma\rangle\langle\beta|\lambda\rangle\right)}.$ (16) The concurrence $C_{\psi_{1}}\left(\alpha\right)$ of the particular state $\left\lvert\psi_{1}\right\rangle$ defined in (12a) is showen in Figure (5) . Figure 4: The NWF of the state $\left\lvert\psi_{1}\right\rangle$ versus $\lvert\alpha\rvert$ for $\lvert\beta\rvert=2$. Figure 5: The concurrence of the state $\left\lvert\psi_{1}\right\rangle$ versus $\lvert\alpha\rvert$ and $\lvert\beta\rvert$. Now let as assume that, the state (12a) pass by $\textbf{C}\left(\phi_{2}\right)\textbf{R}\left(\theta\right)\textbf{C}\left(\phi_{1}\right)$ device (Compensator-Rotationer-Compensator device), where $\textbf{C}\left(\phi\right)=\textnormal{Exp}\left(i\dfrac{\phi}{2}\hat{S}_{1}\right)$ is the application of phase shift $\phi$ between the horizontal and vertical modes and $\textbf{R}\left(\theta\right)=\textnormal{Exp}\left(i\dfrac{\theta}{2}\hat{S}_{3}\right)$ is a geometric rotation by angle $\theta$ in the polarization. The quantum state at the output is given by $\displaystyle\left\lvert\psi_{out}\right\rangle=N_{1}$ $\displaystyle\biggl{(}\left\lvert\beta\sin\left(\theta\right)e^{i\left(\phi_{2}-\phi_{1}\right)/2}+\alpha\cos\left(\theta\right)e^{i\left(\phi_{2}+\phi_{1}\right)/2}\right\rangle\left\lvert\beta\cos\left(\theta\right)e^{-i\left(\phi_{2}+\phi_{1}\right)/2}-\alpha\sin\left(\theta\right)e^{-i\left(\phi_{2}-\phi_{1}\right)/2}\right\rangle$ $\displaystyle+\left\lvert\alpha\sin\left(\theta\right)e^{i\left(\phi_{2}-\phi_{1}\right)/2}+\beta\cos\left(\theta\right)e^{i\left(\phi_{2}+\phi_{1}\right)/2}\right\rangle\left\lvert\alpha\cos\left(\theta\right)e^{-i\left(\phi_{2}+\phi_{1}\right)/2}-\beta\sin\left(\theta\right)e^{-i\left(\phi_{2}-\phi_{1}\right)/2}\right\rangle\biggr{)}.$ (17) The NWF and the concurrence of the output state (3) are plotted in the Figure (6) versus $\lvert\alpha\rvert^{2}$ for $\lvert\beta\lvert^{2}=2$. (a) (b) Figure 6: The concurrence (a) and the NWF (b) versus rotator’s angle $\theta$ for $\phi_{1}\in\\{0,\dfrac{\pi}{8},\dfrac{\pi}{6},\dfrac{\pi}{4}\\}$ and $\phi_{2}=0$. ## 4 Discussion and conclusion In this section, we will discuss comparatively the behavior of the NWF as a measure of entanglement and the concurrence under a polarization converter device, $\textbf{C}\left(\phi_{2}\right)\textbf{R}\left(\theta\right)\textbf{C}\left(\phi_{1}\right)$. For this purpose we chose to plot the NWF and the concurrence for the input and output states $\left\lvert\psi_{1}\right\rangle$ and $\left\lvert\psi_{out}\right\rangle$ ((12a) and (3)). Figure (6) shows the entanglement and the NWF in the input state (12a) that are dependent to the parameters $\theta$ and $\phi$ according to $\lvert\alpha\rvert$ and $\lvert\beta\rvert$. We see that the entanglement increases with increasing values of the parameter $\alpha$ to reach its maximum when $\lvert\alpha\rvert^{2}\geq 1.5$ that is a result of the fact that at the limit of large values of the parameter $\lvert\alpha-\beta\rvert$ the coherent states $\left\lvert\alpha\right\rangle$ and $\left\lvert\beta\right\rangle$ become orthogonal. Thus the behavior of the bi-mode superposed coherent state is, as expected, exactly that of the Bell state. After the state passed through the CRC device, we show in Figure (6) its entanglement as a function of $\theta$ for different values of $\phi_{1}$ fixing $\lvert\alpha-\beta\rvert^{2}=4$. It is interesting to say that the rotation of $\dfrac{\pi}{4}$ applied on the input state $\left\lvert\psi_{1}\right\rangle$ destroys completely the entanglement. This implies that, the $\textbf{R}\left(\theta\right)$ device can be a perfect entangler$/$disentangler gate. In figure (6), the NWF is plotted as a function of $\theta$ for $\lvert\alpha-\beta\rvert^{2}=4$ for different values of $\phi_{1}$. For a specific values of $\phi_{1}$, we see that, the NWF decrease with increasing values of $\theta$ to reach its minimum and vanish for $\phi_{1}=\dfrac{\pi}{4}$. Then, it increases again with increasing values of $\theta$. This allows to show that the NWF and the concurrence behave identically and they have the same inflection points which does confirm that the NWF is a true measure of entanglement in non-Gaussian states. As conclusion, in this paper we have studied the behavior of the entanglement and the polarization degree in superposition of two-mode coherent states. We have confirm that the NWF can be used as a good quantifier of entanglement in non-Gaussian systems. As matter of fact, it turn out that the volume of the negative part of Wigner function is in fact a best quantifier of bipartite entanglement in non- Gaussian systems. This work allows as to describe the Wigner function and the polarization of superposition of two-mode coherent states and the important use of the WF to study the entanglement in non-Gaussian systems. Consequently, the NWF can be considered as a measure of entanglement in non-Gaussian systems. We believe that this result will be efficienct in quantum information theory, mostly in quantum computing [18], because the Wigner function can be measured experimentally, [19, 20], including the measurements of its negative values [5]. The interest point on such experiments has triggered a search for operational definitions of the Wigner functions, based on experimental setup [21, 22]. It does represent a major step forward in the detection and the quantification of non-Gaussian entanglement in bipartite systems. ## References * [1] A. Vidiella-Barranco and L. Borelli, “Continuous variable quantum key distribution using polarized coherent states,” International Journal of Modern Physics B, vol. 20, no. 11n13, pp. 1287–1296, 2006. * [2] G. A. Barbosa, “Fast and secure key distribution using mesoscopic coherent states of light,” Physical Review A, vol. 68, no. 5, p. 052307, 2003. * [3] Z.-Q. Yin, Z.-F. Han, F.-W. Sun, and G.-C. Guo, “Decoy state quantum key distribution with modified coherent state,” Physical Review A, vol. 76, no. 1, p. 014304, 2007. * [4] W.-H. Kye, C.-M. Kim, M. Kim, and Y.-J. Park, “Quantum key distribution with blind polarization bases,” Physical review letters, vol. 95, no. 4, p. 040501, 2005. * [5] A. Kenfack and K. Życzkowski, “Negativity of the wigner function as an indicator of non-classicality,” Journal of Optics B: Quantum and Semiclassical Optics, vol. 6, no. 10, p. 396, 2004. * [6] M. G. Genoni, M. G. Paris, and K. Banaszek, “Measure of the non-gaussian character of a quantum state,” Physical Review A, vol. 76, no. 4, p. 042327, 2007. * [7] J. Wenger and R. Tualle-Brouri, “Non-gaussian statistics from individual pulses of squeezed light,” Physical review letters, vol. 92, no. 15, p. 153601, 2004. * [8] S. L. Braunstein and H. J. Kimble, “Teleportation of continuous quantum variables,” Physical Review Letters, vol. 80, no. 4, p. 869, 1998. * [9] M. Ziane and M. El Baz, “Direct measure of genuine tripartite entanglement independent from bipartite constructions,” Quantum Information Processing, vol. 17, no. 8, p. 196, 2018. * [10] M. Ziane, F.-Z. Siyouri, M. El Baz, and Y. Hassouni, “The negativity of partial transpose versus the negativity of wigner function,” International Journal of Geometric Methods in Modern Physics, vol. 16, no. 06, p. 1930003, 2019. * [11] G. Stokes, “Gg stokes, trans. cambridge philos. soc. 9, 399 (1852).,” Trans. Cambridge Philos. Soc., vol. 9, p. 399, 1852. * [12] A. Klimov, L. Sánchez-Soto, E. Yustas, J. Söderholm, and G. Björk, “Distance-based degrees of polarization for a quantum field,” Physical Review A, vol. 72, no. 3, p. 033813, 2005. * [13] A. Luis, “Degree of polarization in quantum optics,” Physical review A, vol. 66, no. 1, p. 013806, 2002. * [14] G. Agarwal, J. Lehner, and H. Paul, “Invariances for states of light and their quasi-distributions,” Optics communications, vol. 129, no. 5-6, pp. 369–372, 1996. * [15] R. Viana Ramos*, “Mixture of two-mode unpolarized and pure quantum light states: quantum polarization and application in quantum communication,” Journal of Modern Optics, vol. 52, no. 15, pp. 2093–2103, 2005. * [16] P. Rungta, V. Bužek, C. M. Caves, M. Hillery, and G. J. Milburn, “Universal state inversion and concurrence in arbitrary dimensions,” Physical Review A, vol. 64, no. 4, p. 042315, 2001. * [17] L.-M. Kuang and L. Zhou, “Generation of atom-photon entangled states in atomic bose-einstein condensate via electromagnetically induced transparency,” Physical Review A, vol. 68, no. 4, p. 043606, 2003. * [18] T. Forcer, A. Hey, D. Ross, and P. Smith, “Superposition, entanglement and quantum computation,” Quantum Information and Computation, vol. 2, no. 2, pp. 97–116, 2002. * [19] D. Smithey, M. Beck, M. Raymer, and A. Faridani, “Measurement of the wigner distribution and the density matrix of a light mode using optical homodyne tomography: Application to squeezed states and the vacuum,” Physical review letters, vol. 70, no. 9, p. 1244, 1993. * [20] K. Banaszek, C. Radzewicz, K. Wódkiewicz, and J. Krasiński, “Direct measurement of the wigner function by photon counting,” Physical Review A, vol. 60, no. 1, p. 674, 1999. * [21] P. Lougovski, E. Solano, Z. Zhang, H. Walther, H. Mack, and W. Schleich, “Fresnel representation of the wigner function: An operational approach,” Physical review letters, vol. 91, no. 1, p. 010401, 2003. * [22] U. Leonhardt, Measuring the quantum state of light, vol. 22. Cambridge university press, 1997. ## Appendix A Average values and variances of quantum Stokes parameters and Q-function: For each Quantum Stokes parameter $\hat{S}_{i}$ the variance is defined by $V_{i}=\left\langle\hat{S}_{i}^{2}\right\rangle-\left\langle\hat{S}_{i}\right\rangle^{2}$, where the averages of the quantum Stokes parameters and of their squared values in the state $\left\lvert\psi_{\pm}\right\rangle$ defined in (11) are: $\left\langle\hat{S}_{1}\right\rangle=\lvert N\rvert^{2}\Big{(}\left(\lvert\alpha\rvert^{2}-\lvert\beta\rvert^{2}\right)+\left(\lvert\gamma\rvert^{2}-\lvert\lambda\rvert^{2}\right)+\big{[}\left(\alpha^{*}\gamma-\beta*\lambda\right)+\left(\alpha\gamma^{*}-\beta\lambda^{*}\right)\big{]}\delta\Big{)}$ (A.1) $\left\langle\hat{S}_{2}\right\rangle=\lvert N\rvert^{2}\Big{(}\left(\alpha^{*}\beta+\alpha\beta^{*}\right)+\left(\gamma^{*}\lambda+\gamma\lambda^{*}\right)+\big{[}\left(\alpha^{*}\lambda+\gamma\beta^{*}\right)+\left(\gamma^{*}\beta+\alpha\gamma^{*}\right)\big{]}\delta\Big{)}$ (A.2) $\left\langle\hat{S}_{3}\right\rangle=\lvert N\rvert^{2}\Big{(}\left(\alpha\beta^{*}-\alpha^{*}\beta\right)+\left(\gamma\lambda^{*}-\gamma^{*}\lambda\right)+\big{[}\left(\gamma\beta^{*}-\alpha^{*}\lambda\right)+\left(\alpha\lambda^{*}-\beta\gamma^{*}\right)\big{]}\delta\Big{)}$ (A.3) $\left\langle\hat{S}_{1}^{2}\right\rangle=\lvert N\rvert^{2}\left\\{\begin{array}[]{cc}\lvert\alpha\rvert^{2}+\lvert\beta\rvert^{2}+\lvert\gamma\rvert^{2}+\lvert\lambda\rvert^{2}+\left(\lvert\alpha\rvert^{2}-\lvert\beta\rvert^{2}\right)^{2}+\left(\lvert\gamma\rvert^{2}-\lvert\lambda\rvert^{2}\right)^{2}+\\\ \big{[}\alpha^{*}\gamma+\beta^{*}\lambda+\alpha\gamma^{*}+\beta\lambda^{*}+\left(\alpha^{*}\gamma-\beta^{*}\lambda\right)^{2}+\left(\alpha\gamma^{*}-\beta\lambda^{*}\right)^{2}\big{]}\delta\end{array}\right\\}$ (A.4) $\left\langle\hat{S}_{2}^{2}\right\rangle=\lvert N\rvert^{2}\left\\{\begin{array}[]{cc}\lvert\alpha\rvert^{2}+\lvert\beta\rvert^{2}+\lvert\gamma\rvert^{2}+\lvert\lambda\rvert^{2}+2\left(\lvert\alpha\rvert^{2}\lvert\beta\rvert^{2}+\lvert\gamma\rvert^{2}\lvert\lambda\rvert^{2}\right)+\left(\alpha^{*}\beta\right)^{2}+\left(\alpha\beta^{*}\right)^{2}\left(\gamma^{*}\lambda\right)^{2}+\\\ \left(\gamma\lambda^{*}\right)^{2}+\big{[}\alpha^{*}\gamma+\beta^{*}\lambda+\alpha\gamma^{*}+\beta\lambda^{*}+\left(\alpha^{*}\lambda+\beta^{*}\gamma\right)^{2}+\left(\gamma^{*}\beta+\alpha\lambda^{*}\right)^{2}\big{]}\delta\par\end{array}\right\\}$ (A.5) $\left\langle\hat{S}_{3}^{2}\right\rangle=\lvert N\rvert^{2}\left\\{\begin{array}[]{cc}\lvert\alpha\rvert^{2}+\lvert\beta\rvert^{2}+\lvert\gamma\rvert^{2}+\lvert\lambda\rvert^{2}+2\left(\lvert\alpha\rvert^{2}\lvert\beta\rvert^{2}-\lvert\gamma\rvert^{2}\lvert\lambda\rvert^{2}\right)-\left(\alpha^{*}\beta\right)^{2}-\left(\alpha\beta^{*}\right)^{2}\left(\gamma^{*}\lambda\right)^{2}-\\\ \left(\gamma\lambda^{*}\right)^{2}+\big{[}\alpha^{*}\gamma+\beta^{*}\lambda+\alpha\gamma^{*}+\beta\lambda^{*}-\left(\alpha^{*}\lambda-\beta^{*}\gamma\right)^{2}-\left(\gamma^{*}\beta-\alpha\lambda^{*}\right)^{2}\big{]}\delta\end{array}\right\\}$ (A.6) where $\delta=\textnormal{exp}\big{[}\alpha^{*}\gamma+\beta^{*}\lambda-\left(\lvert\alpha\rvert^{2}+\lvert\beta\rvert^{2}+\lvert\gamma\rvert^{2}+\lvert\lambda\rvert^{2}\right)/2\big{]}$ and the Q-function of the same state ((11)) is $Q\left(\theta,\phi\right)=\dfrac{\lvert N\rvert^{2}}{4\pi}\left\\{\begin{array}[]{cc}e^{-\left(\lvert\alpha\rvert^{2}+\lvert\beta\rvert^{2}\right)}\Big{(}1+z_{1}\Big{)}e^{z_{1}}+e^{-\left(\lvert\gamma\rvert^{2}+\lvert\lambda\rvert^{2}\right)}\Big{(}1+z_{2}\Big{)}e^{z_{2}}+\\\ e^{-\left(\lvert\alpha\rvert^{2}+\lvert\beta\rvert^{2}+\lvert\gamma\rvert^{2}+\lvert\lambda\rvert^{2}\right)/2}\left[\Big{(}1+z_{12}\Big{)}e^{z_{12}}+\Big{(}1+z_{12}^{*}\Big{)}e^{z_{12}^{*}}\right]\end{array}\right\\}$ (A.7)
# The extended Bregman divergence and parametric estimation Sancharee Basak Ayanendranath Basu (Interdisciplinary Statistical Research Unit Indian Statistical Institute, Kolkata, INDIA ) ###### Abstract Minimization of suitable statistical distances (between the data and model densities) has proved to be a very useful technique in the field of robust inference. Apart from the class of $\phi$-divergences of [1] and [2], the Bregman divergence ([3]) has been extensively used for this purpose. However, since the data density must have a linear presence in the cross product term of the Bregman divergence involving both the data and model densities, several useful divergences cannot be captured by the usual Bregman form. In this respect, we provide an extension of the ordinary Bregman divergence by considering an exponent of the density function as the argument rather than the density function itself. We demonstrate that many useful divergence families, which are not ordinarily Bregman divergences, can be accommodated within this extended description. Using this formulation, one can develop many new families of divergences which may be useful in robust inference. In particular, through an application of this extension, we propose the new class of the GSB divergence family. We explore the applicability of the minimum GSB divergence estimator in discrete parametric models. Simulation studies as well as conforming real data examples are given to demonstrate the performance of the estimator and to substantiate the theory developed. Keywords— Bregman divergence; $S$-divergence; B-exponential divergence; GSB divergence; discrete model; robustness ## 1 Introduction In the domain of statistical inference, there is, generally, an inherent trade-off between the model efficiency and the robustness of the procedure. Often we compromise the efficiency of the procedure to a certain allowable extent for achieving better robustness properties. In the present age of big data, the robustness angle of statistical inference has to be dealt with greater care than ever before. Such methods do exist in the literature which allow full asymptotic efficiency simultaneously with strong robustness properties; see, e.g., [4], [5] and [6]. In practice, however, one would be hard-pressed to find a procedure which matches the likelihood-based methods in terms of efficiency in small to moderate samples without inheriting any of the robustness limitations of the latter. Many of these trade-off issues are discussed in the canonical texts on robustness such as [7], [8] and [9]; for a minimum divergence view of this issue, see [6]. There are several types of divergences which are used in minimum distance inference. Most of them are not mathematical metrics. They may not satisfy the triangle inequality or may not even be symmetric in their arguments. The only properties we demand of these measures are that they are non-negative and are equal to zero if and only if the two arguments are identically equal. Sometimes we will refer to these divergence measures as ‘statistical distances’ or, loosely, as ‘distances’ without any claim to metric properties. Most of the density-based divergences in the literature belong to either the class of chi-square type distances (formally called $\phi$-divergences, $f$-divergences or disparities) or Bregman divergences. See [1], [2] and [5] for a description of the divergences of the chi-square type and [3] for Bregman divergences. Although [3] introduced the Bregman divergence in order to be used in convex programming, it has, because of its flexible characteristics, been used in many branches of natural science as well as in areas like information theory and computational geometry. The class of chi- square type distances between two densities $g$ and $f$ includes, for example, the likelihood disparity (LD), the Kullback-Leibler divergence (KLD) and the (twice, squared) Hellinger distance (HD), given by ${\rm LD}(g,f)=\int g\log\left(\frac{g}{f}\right),~{}{\rm KLD}(g,f)=\int f\log\left(\frac{f}{g}\right),~{}{\rm HD}(g,f)=\frac{1}{2}\int(f^{1/2}-g^{1/2})^{2},$ (1) respectively. Representative members of the class of Bregman divergences include the LD and the squared $L_{2}$ distance, where $L_{2}(g,f)=\int(g-f)^{2}.$ (2) The LD is the only common member between the class of chi-square type distances and Bregman divergences. In the parametric estimation scheme that we consider, the estimator corresponds to the parameter of the model density which is closest to the observed data density in terms of the given divergence, the observed data density being a non-parametric representative of the true unknown density based on the given sample. In case of chi-square type distances, the construction of the data density inevitably requires the use of an appropriate non-parametric smoothing technique, like kernel density estimation, in continuous models (the LD is the only exception). This makes the derivation of the asymptotic properties far more involved, and complicates the computational aspect of this estimation. On the other hand, all the minimum Bregman divergence estimators are M-estimators and, hence, they avoid this density estimation component. In this paper, our primary aim is to extend the scope of the Bregman divergence by utilizing the powers of densities as arguments, rather than the densities themselves; this leads to the generalized class of the extended Bregman divergences which can then be used to generate new divergences which could provide more refined tools for minimum divergence inference compared to the current state of the art. This is the key idea of this work. Note that the use of the Bregman divergence in statistics is relatively recent; the class of density power divergences by [10], defined in Section 2, is a prominent example of Bregman divergences having significant applications in statistical inference. Many minimum divergence procedures have natural robustness properties against data contamination and outliers. The extended Bregman divergence allows us to express several existing divergence families as special cases of it, which is not possible through the ordinary Bregman divergence. Consequently, the extended Bregman idea can be used to generate large super-families of divergences containing, together with the existing divergences, many new and useful divergence families as special cases. In Section 2, we propose the extended Bregman divergence family. We demonstrate how it allows us to capture well known divergences that are not within the ordinary Bregman class and give a potential route for constructing new divergences. In Section 3, by considering a specific form of the convex function along with a particular choice of exponent of densities, we construct a large super-family of divergences within the extended Bregman family. Several known divergence families are obtained as special cases of this super- family. Section 4 introduces the corresponding minimum distance estimator, while Section 5 studies its asymptotic properties. Section 6 explores the robustness properties of the estimator based on its influence function. A large scale simulation study is taken up in Section 7, and a tuning parameter selection strategy is discussed in Section 8. The final section has some concluding remarks. Before concluding this section, we summarize what we believe to be the main achievements in this paper. 1. 1. We provide a simple extension of the Bregman divergence by considering powers of densities (instead of the densitites themselves) as arguments. Many divergence families (which are ordinarily not members of the class of Bregman divergences) can now be looked upon as members of this extended family, and the properties of the corresponding minimum distance estimators may be obtained from the general properties common to all the members of the extended family. 2. 2. Ordinarily, all minimum Bregman divergence estimators are also M-estimators. But, through this extension, several minimum distance estimators which are not M-estimators also become a part of this extended family. 3. 3. Consideration of exponentiated arguments with a specific choice of the convex function introduces a generalized super-family which we refer to as the GSB divergence family. The power divergence family of [11], the density power divergence (DPD) family of [10], the Bregman exponential divergence (BED) of [12] and the $S$-divergence family of [13] can all be brought under the umbrella of this super-family. 4. 4. This GSB divergence family consists of three tuning parameters $\alpha$, $\beta$ and $\lambda$. By simultaneously varying all these three parameters, we can generate new divergences (and hence new minimum divergence estimators) which are outside the union of the BED and S-divergence family, but can potentially provide improved performance over both of these classes. ## 2 The extended Bregman divergence and special cases Being motivated by the problem of convex programming, [3] introduced the Bregman divergence, a measure of dissimilarity between any two vectors in the Euclidean space. In $\mathbb{R}^{p}$, it has the form $\displaystyle D_{\psi}\left(\boldsymbol{x},\boldsymbol{y}\right)=\Bigg{\\{}\psi\left(\boldsymbol{x}\right)-\psi\left(\boldsymbol{y}\right)-\langle\nabla\psi\left(\boldsymbol{y}\right),\boldsymbol{x}-\boldsymbol{y}\rangle\Bigg{\\}},$ (3) for any strictly convex function $\psi:\mathcal{S}\rightarrow\mathbb{R}$ and for any two $p$-dimensional vectors $\boldsymbol{x},\boldsymbol{y}\in\mathcal{S}$, where $\mathcal{S}$ is a convex subset of $\mathbb{R}^{p}$. Here, $\nabla\psi\left(\boldsymbol{y}\right)$ denotes the gradient of $\psi$ with respect to its argument at $\boldsymbol{y}=\left(y_{1},y_{2},\ldots,y_{p}\right)^{T}$. It is evident that only the convexity criterion of the function $\psi\left(\cdot\right)$ is necessary for the non-negativity property of the divergence $D_{\psi}\left(\boldsymbol{x},\boldsymbol{y}\right)$ to hold. One could, therefore, consider other quantities as the arguments rather than the points themselves in this measure. Hence, as long as $\psi$ remains convex, any set of arguments whose equivalence translates to the equivalence of $\boldsymbol{x}$ and $\boldsymbol{y}$ can be used in the distance expression. This observation may be used to extend the Bregman divergence to have the form $\displaystyle D_{\psi}\left(\boldsymbol{x},\boldsymbol{y}\right)=\Bigg{\\{}\psi\left(\boldsymbol{x}^{k}\right)-\psi\left(\boldsymbol{y}^{k}\right)-\langle\nabla\psi\left(\boldsymbol{y}^{k}\right),\boldsymbol{x}^{k}-\boldsymbol{y}^{k}\rangle\Bigg{\\}}.$ (4) Here, $\nabla\psi\left(\boldsymbol{y}^{k}\right)$ denotes the gradient of $\psi$ with respect to its argument, evaluated at $\boldsymbol{y}^{k}=\left(y_{1}^{k},y_{2}^{k},\ldots,y_{d}^{k}\right)^{T}$ and $\psi$ is a strictly convex function, mapping $\mathcal{S}$ to $\mathbb{R}$, $\mathcal{S}$ being a convex subset of $\mathbb{R^{+}}^{p}$. Since our main purpose is to utilize this extension in the field of statistics where the arguments, being probability density functions, are inherently non-negative, restricting the domain of $\psi$ to $\mathbb{R^{+}}^{p}$ does not cause any difficulty. It is also not difficult to see that many of the properties of the Bregman divergence in Equation (3), as described by [14], are retained by the extended version in Equation (4). However, we will not make use of these properties in this paper, so we do not discuss them here any further. The Bregman divergence has significant applications in the domain of statistical inference for both discrete and continuous models. Given two densities $g$ and $f$, the Bregman divergence between these densities (associated with the convex function $\psi$) is given by $\displaystyle D_{\psi}\left(g,f\right)=\displaystyle\int\Bigg{\\{}\psi\left(g\left(x\right)\right)-\psi\left(f\left(x\right)\right)-\left(g\left(x\right)-f\left(x\right)\right)\nabla\psi\left(f\left(x\right)\right)\Bigg{\\}}dx.$ (5) By the strict convexity of the function $\psi$, the integrand in the above Equation (5) is non-negative and, therefore, so is the integral. It is also clear that the divergence equals zero if and only if the arguments $g$ and $f$ are identically equal. Well-known examples include the LD and the (squared) $L_{2}$ distance which correspond to $\psi\left(x\right)=x\log x$ and $\psi\left(x\right)=x^{2}$ respectively. In a real scenario, one uses $g$ as the true data generating density and $f=f_{\theta}$ as the parametric model density. In [10], the important class of density power divergences (DPDs) has been proposed, which is a subfamily of the class of Bregman divergences. This family is generated by the function $\psi\left(x\right)=\frac{x^{\alpha+1}-x}{\alpha}$, indexed by a non-negative tuning parameter $\alpha$. As a function of $\alpha$, the density power divergence may be expressed as $\displaystyle DPD_{\alpha}\left(g,f\right)=\displaystyle\int\left\\{f^{\alpha+1}\left(x\right)-\left(1+\frac{1}{\alpha}\right)g\left(x\right)f^{\alpha}\left(x\right)+\frac{1}{\alpha}g^{\alpha+1}\left(x\right)\right\\}dx.$ (6) It is a simple matter to check that for $\alpha=1$, the above reduces to the (squared) $L_{2}$ distance between $g$ and $f$, whereas when $\alpha\rightarrow 0$, one recovers the likelihood disparity defined in Equation (1). The DPD class should not be confused with the power divergence (PD) class of divergences (see [11]) which has the form $\displaystyle PD_{\lambda}\left(g,f\right)=\displaystyle\frac{1}{\lambda\left(\lambda+1\right)}\int\left\\{g\left(x\right)\left(\frac{g\left(x\right)}{f\left(x\right)}\right)^{\lambda}-1\right\\}dx,\lambda\in\mathbb{R}.$ (7) The PD class is a subfamily of chi-square type distances. The latter class of divergences has the form $\rho\left(g,f\right)=\displaystyle\int C\left(\delta\left(x\right)\right)f\left(x\right)dx,$ (8) where $C$ is a strictly convex function and $\delta\left(x\right)=\frac{g\left(x\right)}{f\left(x\right)}-1$. The power divergence corresponds to the specific convex function $C\left(\delta\right)=\frac{\left(\delta+1\right)^{\lambda+1}-\left(\delta+1\right)}{\lambda\left(\lambda+1\right)}-\frac{\delta}{\lambda+1}.$ (9) Important special cases of the PD class include the LD (obtained in the limit as $\lambda\rightarrow 0$) and the twice, squared HD (obtained for $\lambda=-\frac{1}{2}$). The LD is the only common member between the PD and the DPD classes. The Bregman exponential divergence (BED) class ([12]), on the other hand, has the form $\displaystyle BED_{\beta}\left(g,f\right)=\displaystyle\frac{2}{\beta}\displaystyle\int\left\\{e^{\beta f\left(x\right)}\left(f\left(x\right)-\frac{1}{\beta}\right)-e^{\beta f\left(x\right)}g\left(x\right)+\frac{e^{\beta g\left(x\right)}}{\beta}\right\\}dx.$ (10) The defining convex function is $\psi\left(x\right)=\frac{2\left(e^{\beta x}-\beta x-1\right)}{\beta^{2}}$ which is indexed by the real parameter $\beta$. This family generates the (squared) $L_{2}$ distance in the limit $\beta\rightarrow 0$. A list of some Bregman divergences useful in the context of statistical inference is presented in Table 1. Table 1: Different divergences as special cases of the Bregman divergence Choice of convex function | Divergences ---|--- $B\left(x\right)=\displaystyle x^{2}$ | (squared) $L_{2}$ Distance $B\left(x\right)=\displaystyle x\log\left(x\right)$ | Likelihood Disparity $B\left(x\right)=\displaystyle\frac{x^{1+\alpha}-x}{\alpha}$ | Density Power Divergence (DPD) $B\left(x\right)=\displaystyle-\frac{\log\left(x\right)}{2\pi}$ | Itakura-Saito Distance $B\left(x\right)=\frac{2\left(e^{\beta x}-\beta x-1\right)}{\beta^{2}}$ | Bregman Exponential Divergence Consider the standard set up of parametric estimation where $G$ is the true data generating distribution which is modeled by the parametric family ${\cal F}=\\{F_{\theta}:\theta\in\Theta\subset{\mathbb{R}}^{p}\\}$. Let $g$ and $f_{\theta}$ be the corresponding density functions. Further we assume that both $G$ and $F_{\theta}$ belong to $\mathcal{G}$, the class of all cumulative distribution functions having densities with respect to some appropriate dominating measure. Our aim is to estimate the unknown parameter $\theta$ by choosing the model density closest to the true density in the Bregman sense. The definition of ordinary Bregman divergences as given in Equation (5), useful as it is, does not include many well-known and popular divergences which are extensively used in the literature for different purposes including parameter estimation. The PD family is a prominent example. An inspection of the Bregman form in Equation (5) indicates that the term which involves both densities $g$ and $f$ is of the form $\int g\left(x\right)\nabla\psi\left(f\left(x\right)\right)dx.$ (11) Here, the density $g$ is present only as a linear term having exponent one. Given a random sample $X_{1},X_{2},\ldots,X_{n}$ from the true distribution $G$, the term in Equation (11) can be empirically estimated by $\frac{1}{n}\sum\nabla\psi(f_{\theta}(X_{i}))$ (with $f=f_{\theta}$ under the parametric model) so that one can construct an empirical version of the divergence without any non-parametric density estimation. On the other hand, this restricts the class of divergences that are expressible in the Bregman form. Using an extension in the spirit of Equation (4) may allow the construction of richer classes of divergences. With this aim, we define the extended Bregman divergence between two densities $g$ and $f$ as $\displaystyle D^{(k)}_{\psi}\left(g,f\right)=\displaystyle\int\left\\{\psi\left(g^{k}\left(x\right)\right)-\psi\left(f^{k}\left(x\right)\right)-\left(g^{k}\left(x\right)-f^{k}\left(x\right)\right)\nabla\psi\left(f^{k}\left(x\right)\right)\right\\}dx.$ Apart from the requirement of strict convexity of the function $\psi$, this formulation also depends on a positive index $k$ with which the density is exponentiated. For the rest of the paper, the notation $D^{(k)}_{\psi}(\cdot,\cdot)$ will refer to this general form in Equation (LABEL:ss1), of which the divergence in Equation (5) is a special case for $k=1$. Evidently, $D^{(k)}_{\psi}\left(g,f\right)\geq 0$ for any choices of densities $f$ and $g$ with respect to the same measure. Moreover, the fact that $D^{(k)}_{\psi}\left(g,f\right)=0$ if and only if $g=f$, holds true in this case due to non-negativity property of a density as well as the consideration of strict convexity of the function $\psi\left(\cdot\right)$. In the following, we present some special cases of extended Bregman divergence. 1. 1. $S$-Hellinger family ([13]) If we take $\psi\left(x\right)=\frac{2e^{\beta x}}{\beta^{2}}$ with $k=\frac{1+\alpha}{2}$, $\alpha\in\left(0,1\right)$ in Equation (LABEL:ss1), it will generate an extension of the BED family having the form $BED^{(k)}_{\beta}\left(g,f\right)=\frac{2}{\beta}\int\left\\{e^{\beta f^{k}\left(x\right)}\left(f^{k}\left(x\right)-\frac{1}{\beta}\right)-e^{\beta f^{k}\left(x\right)}g^{k}\left(x\right)+\frac{e^{\beta g^{k}\left(x\right)}}{\beta}\right\\}dx.$ (13) It can be easily shown that, as $\beta\rightarrow 0$ and $k=\frac{1+\alpha}{2}$, $\alpha\in\left(0,1\right)$, the application of L’Hospital’s rule leads to the S-Hellinger Distance (SHD) family with the form $SHD_{\alpha}\left(g,f\right)=\frac{2}{1+\alpha}\int\left(g^{\frac{1+\alpha}{2}}\left(x\right)-f^{\frac{1+\alpha}{2}}\left(x\right)\right)^{2}dx.$ (14) This was introduced by [13] as a special case of the $S$-divergence family. This family cannot be expressed through the normal expression of the Bregman divergence, but through this extension, we can express this member of the $S$-divergence family as a (limiting) member of the extended BED class. 2. 2. PD family ([11]) If we take $\psi\left(x\right)=\frac{x^{1+\frac{B}{A}}}{B}$, $A=1+\lambda$, $B=-\lambda$ and $\lambda\in\mathbb{R}$, with $k=A$ in Equation (LABEL:ss1), we get the PD family introduced in Equation (7). 3. 3. $S$-divergence family ([13]) If we take $\psi\left(x\right)=\frac{x^{1+\frac{B}{A}}}{B}$, $A=1+\lambda\left(1-\alpha\right)$, $B=\alpha-\lambda\left(1-\alpha\right)$, $A+B=1+\alpha$, $\alpha\geq 0$, $\lambda\in\mathbb{R}$ and $k=A$ in Equation (LABEL:ss1), we get the $S$-divergence having the following form $SD_{(\alpha,\lambda)}\left(g,f\right)=\int\left\\{\frac{1}{B}\left(g^{A+B}\left(x\right)-f^{A+B}\left(x\right)\right)-\left(g^{A}\left(x\right)-f^{A}\left(x\right)\right)\frac{A+B}{AB}f^{B}\left(x\right)\right\\}dx.$ (15) This is one of the most useful divergence families in the domain of robust inference due to its capacity to generate much more robust estimator(s) than the DPD and PD families can generate. Through this extension of Bregman divergence, it is now possible to express this divergence as a special case of the extended family. Note that, for $k\neq 1$, $f^{k}$ and $g^{k}$ will generally no longer represent probability densities, and by extending the divergence idea to general positive measures (beyond probability measures), [15] has suggested certain constructions where the power divergence has been exhibited in the Bregman divergence form for general measures. Also, see the discussion in [16]. We differ from the interpretation in these papers in the sense that we still view the family of divergences presented here in Equation (LABEL:ss1) as divergences between valid probability densities. Given any two probability densities, the expression in Equation (LABEL:ss1) is non-negative, and equals zero if and only if the densities $g$ and $f$ are identical, irrespective of the value of $k$. For the ordinary Bregman divergence, the term in Equation (11), with $f=f_{\theta}$, may be approximated by $\frac{1}{n}\sum_{i=1}^{n}\nabla\psi\left(f_{\theta}\left(X_{i}\right)\right),$ (16) through the replacement of $dG\left(x\right)$ by $dG_{n}\left(x\right)$, where $G_{n}$ is the empirical distribution function obtained from the random sample $X_{1},X_{2},\ldots,X_{n}$. It is evident that the minimizer of the empirical version of the Bregman divergence is an M-estimator. But there are several useful divergences (or divergence families) where the empirical representation of the term involving $f$ and $g$ is not possible using the above trick, and such divergences generate estimators beyond the M-estimator class. See [17] for more discussion on this issue. In the above we have given several examples where the extended Bregman class contains such divergences which are not covered by the ordinary Bregman form. Thus the structure of the extended class allows us to extend the scope much beyond that of the ordinary Bregman divergence. ## 3 Introducing a new divergence family Our aim here is to exploit the extended Bregman idea and generate rich new super families of divergences by choosing a suitable convex generating function and a suitable exponent. In particular, we use the convex function $\psi(x)=e^{\beta x}+\frac{x^{1+\frac{B}{A}}}{B}$, $A=1+\lambda\left(1-\alpha\right)$, $B=\alpha-\lambda\left(1-\alpha\right)$, $A+B=1+\alpha$, $\alpha\geq-1$, $\beta,\lambda\in\mathbb{R}$, which, together with the exponent $k=A$, generates the divergence $D^{*}\left(g,f\right)=\int\left\\{e^{\beta f^{A}}\left(\beta f^{A}-\beta g^{A}-1\right)+e^{\beta g^{A}}+\frac{1}{B}\left(g^{A+B}-f^{A+B}\right)-\left(g^{A}-f^{A}\right)\frac{A+B}{AB}f^{B}\right\\}dx,$ (17) which we refer to as the GSB divergence (being the abbreviated form of generalized S-Bregman divergence). The divergence measure $D^{*}$ is also a function of $\alpha,\lambda$ and $\beta$, which we suppress for brevity. If we put $A+B=0$ in the above expression with $A\neq 0$ and $B\neq 0$, we will get the extended BED family with parameter $\beta$ and exponent $k=A$. Moreover, if $A=1$, i.e., $\lambda=0$ then it will lead to the ordinary BED family with parameter $\beta$. On the contrary, if we put $\beta=0$, it will lead to the $S$-divergence family with parameters $\alpha$ and $\lambda$ (in terms of $A$ and $B$). More specifically, when $\alpha=0$ and $\beta=0$, it leads to the power divergence (PD) family. On the other hand, $\beta=0$ and $\lambda=0$ lead us to the density power divergence (DPD) family. Thus, it acts as a connector between the BED and the $S$-divergence family. ### 3.1 Special cases We will get several well-known divergences or divergence families from the general form of GSB for particular choices of the three tuning parameters $\alpha$, $\lambda$ and $\beta$. Some such choices are given in Table 2. Table 2: Different divergences as special cases of GSB divergence $\alpha$ | $\lambda$ | $\beta$ | Divergences ---|---|---|--- $\alpha=-1$ | $\lambda=0$ | $\beta\in\mathbb{R}$ | Bregman Exponential Divergence1 $\alpha=0$ | $\lambda\in\mathbb{R}$ | $\beta=0$ | Power Divergence $\alpha\in\mathbb{R}$ | $\lambda=0$ | $\beta=0$ | Density Power Divergence $\alpha\in\mathbb{R}$ | $\lambda\in\mathbb{R}$ | $\beta=0$ | $S$-Divergence $\alpha=0$ | $\lambda=-1$ | $\beta=0$ | Kullback-Liebler Divergence $\alpha=0$ | $\lambda=0$ | $\beta=0$ | Likelihood Disparity $\alpha=0$ | $\lambda=-.5$ | $\beta=0$ | Hellinger Distance $\alpha\in\mathbb{R}$ | $\lambda=-.5$ | $\beta=0$ | S-Hellinger Distance $\alpha=0$ | $\lambda=1$ | $\beta=0$ | Pearson’s Chi-square Divergence $\alpha=0$ | $\lambda=-2$ | $\beta=0$ | Neyman’s Chi-square Divergence $\alpha=1$ | $\lambda\in\mathbb{R}$ | $\beta=0$ | (squared) $L_{2}$ Distance * 1 This is a constant time B-exponential divergence. It basically generates all the members of BED family corresponding to the same $\beta$ except (squared) $L_{2}$ distance, which occurs when $\beta\rightarrow 0$. However, as seen above, the (squared) $L_{2}$ distance remains a member of the GSB class for other choices of the tuning parameters. ## 4 The minimum GSB divergence estimator Under the parametric set-up described in Section 2, we would like to identify the best fitting parameter $\theta^{g}$ by choosing the element of the model family of distributions which provides the closest match to the true density $g$ in terms of the given divergence. The minimum GSB divergence functional $T_{\alpha,\lambda,\beta}:\mathcal{G}\rightarrow\Theta$ is defined by the relation $D^{*}\left(g,f_{T_{\alpha,\lambda,\beta}}\right)=\min\\{D^{*}\left(g,f_{\theta}\right):\theta\in\Theta\\},$ provided the minimum exists. If the parametric model family is identifiable, it follows from the definition of the divergence that $D^{*}\left(g,f_{\theta}\right)=0$, if and only if $g=f_{\theta}$. Thus, $T_{\alpha,\lambda,\beta}\left(F_{\theta}\right)=\theta$, uniquely. Hence, the functional $T_{\alpha,\lambda,\beta}$ is Fisher consistent. Given the density $g$, a straightforward differentiation of the GSB divergence of Equation (17) leads to the estimating equation $\displaystyle\displaystyle\int\left\\{A^{2}\beta^{2}e^{\beta f_{\theta}^{A}\left(x\right)}f_{\theta}^{A}\left(x\right)+\left(A+B\right)f_{\theta}^{B}\left(x\right)\right\\}\left(f_{\theta}^{A}\left(x\right)-g^{A}\left(x\right)\right)u_{\theta}\left(x\right)dx=0.$ (18) In practice, the true density $g$ is unknown, so one has to use a suitable non-parametric density estimator $\hat{g}$ for $g$, depending on the situation. Under a discrete parametric set-up, the natural choice for $\hat{g}$ is the vector of relative frequencies as obtained from the sample data. Thus, assuming a discrete parametric model, and assuming, without loss of generality, that the support of the random variable is $\\{0,1,2,\ldots,\\}$, the estimating equation becomes $\displaystyle\displaystyle\sum_{x=0}^{\infty}A^{2}\beta^{2}e^{\beta f_{\theta}^{A}\left(x\right)}f_{\theta}^{2A}\left(x\right)u_{\theta}\left(x\right)+\sum_{x=0}^{\infty}\left(A+B\right)f_{\theta}^{A+B}\left(x\right)u_{\theta}\left(x\right)$ (19) $\displaystyle=$ $\displaystyle\displaystyle\sum_{x=0}^{\infty}A^{2}\beta^{2}e^{\beta f_{\theta}^{A}\left(x\right)}f_{\theta}^{A}\left(x\right)\hat{g}^{A}\left(x\right)u_{\theta}\left(x\right)+\displaystyle\sum_{x=0}^{\infty}\left(A+B\right)f_{\theta}^{B}\left(x\right)\hat{g}^{A}\left(x\right)u_{\theta}\left(x\right).$ For continuous models, on the other hand, some suitable non-parametric smoothing technique such as kernel density estimation is inevitable unless the exponent $A$ equals 1. In the latter case, $g$ appears as a linear term in the estimating equation (18). In that case, we can use $G_{n}$, the empirical distribution function, as an estimator of $G$. Hence, for $A=1$, Equation (19) can be reduced as $\displaystyle\displaystyle\sum_{x=0}^{\infty}\beta^{2}e^{\beta f_{\theta}\left(x\right)}f_{\theta}^{2}\left(x\right)u_{\theta}\left(x\right)+\sum_{x=0}^{\infty}\left(1+B\right)f_{\theta}^{1+B}\left(x\right)u_{\theta}\left(x\right)$ (20) $\displaystyle=$ $\displaystyle\frac{1}{n}\displaystyle\sum_{i=1}^{n}\beta^{2}e^{\beta f_{\theta}\left(X_{i}\right)}f_{\theta}\left(X_{i}\right)u_{\theta}\left(X_{i}\right)+\frac{1}{n}\sum_{i=1}^{n}\left(1+B\right)f_{\theta}^{B}\left(X_{i}\right)u_{\theta}\left(X_{i}\right).$ Since the left hand side of the above equation is non-random and the right hand side is a sum of independent and identically distributed terms, it is of the form $\sum_{i=1}^{n}\psi\left(X_{i},\theta\right)=0$ and the corresponding estimator belongs to the M-estimator class. In accordance with the information on the first three rows of Table 2, we will refer to the parameters $\alpha$, $\lambda$ and $\beta$ as the DPD parameter, the PD parameter and the BED parameter, respectively. ## 5 Asymptotic properties of the GSB divergence In this section, we concentrate on the asymptotic properties of our proposed minimum divergence estimator. As mentioned, we will focus on the discrete set- up throughout the rest of the paper. Let $X_{1},X_{2},\ldots,X_{n}$ be independent and identically distributed observations from an unknown distribution $G$ with support $\chi=\\{0,1,2,3,\ldots,\\}$. On the other hand, we consider a parametric family of distributions $\mathcal{F}=\\{F_{\theta}:\theta\in\Theta\subseteq\mathbb{R}^{p}\\}$, also supported on $\chi$, to model the true data generating distribution $G$. In this set-up, we assume both $G$ and $\mathcal{F}$ to have densities $g$ and $f_{\theta}$ with respect to the counting measure. Let $\theta^{g}=T_{\alpha,\beta,\lambda}(G)$ be the best fitting parameter. Since $G$ is unknown, we are going to use the vector of relative frequencies, obtained from the data, as an estimate of $g$ throughout the rest of this paper. Let $r_{n}(x)$ be the relative frequency of the value $x$ in the sample. The minimum GSB divergence estimator is obtained as a root of the estimating equation $\displaystyle\displaystyle\sum_{x=0}^{\infty}\left\\{A^{2}\beta^{2}e^{\beta f_{\theta}^{A}\left(x\right)}f_{\theta}^{A}\left(x\right)+\left(A+B\right)f_{\theta}^{B}\left(x\right)\right\\}\left(f_{\theta}^{A}\left(x\right)-\hat{g}^{A}\left(x\right)\right)u_{\theta}\left(x\right)=0$ (21) $\displaystyle\Rightarrow$ $\displaystyle\displaystyle\sum_{x=0}^{\infty}\left\\{A^{2}\beta^{2}e^{\beta f_{\theta}^{A}\left(x\right)}f_{\theta}^{2A}\left(x\right)+\left(A+B\right)f_{\theta}^{A+B}\left(x\right)\right\\}\frac{\left(1-\frac{\hat{g}^{A}\left(x\right)}{f_{\theta}^{A}\left(x\right)}\right)}{A}u_{\theta}\left(x\right)=0$ $\displaystyle\Rightarrow$ $\displaystyle\displaystyle\sum_{x=0}^{\infty}K\left(\delta\left(x\right)\right)\left(A^{2}\beta^{2}e^{\beta f_{\theta}^{A}\left(x\right)}f_{\theta}^{2A}\left(x\right)+\left(A+B\right)f_{\theta}^{A+B}\left(x\right)\right)u_{\theta}\left(x\right)=0,$ where, $\delta\left(x\right)=\delta_{n}\left(x\right)=\frac{\hat{g}\left(x\right)}{f_{\theta}\left(x\right)}-1=\frac{r_{n}\left(x\right)}{f_{\theta}\left(x\right)}-1$, $K\left(\delta\right)=\frac{\left(\delta+1\right)^{A}-1}{A}$ and $u_{\theta}\left(x\right)$ is the likelihood score function at $x$. We denote the minimum GSB divergence estimator obtained as a solution of the above equation as $\hat{\theta}$. Let $\displaystyle J_{g}$ $\displaystyle=$ $\displaystyle J_{\alpha,\beta,A}\left(g\right)$ $\displaystyle=$ $\displaystyle\displaystyle E_{g}\left(u_{\theta^{g}}\left(X\right)u^{T}_{\theta^{g}}\left(X\right)K^{\prime}\left(\delta_{g}^{g}\left(X\right)\right)\left(\left(A+B\right)f_{\theta^{g}}^{\alpha}\left(X\right)+A^{2}\beta^{2}e^{\beta f_{\theta^{g}}^{A}\left(X\right)}f_{\theta^{g}}^{2A-1}\left(X\right)\right)\right)$ $\displaystyle+$ $\displaystyle\sum_{x=0}^{\infty}K\left(\delta_{g}^{g}\left(x\right)\right)\left(\left(A+B\right)f_{\theta^{g}}^{1+\alpha}\left(x\right)+A^{2}\beta^{2}e^{\beta f_{\theta^{g}}^{A}\left(x\right)}f_{\theta^{g}}^{2A}\left(x\right)\right)i_{\theta^{g}}\left(x\right)$ $\displaystyle-$ $\displaystyle\sum_{x=0}^{\infty}K\left(\delta_{g}^{g}\left(x\right)\right)\left(\left(A+B\right)^{2}f_{\theta^{g}}^{1+\alpha}\left(x\right)+A^{3}\beta^{2}e^{\beta f_{\theta^{g}}^{A}\left(x\right)}f_{\theta^{g}}^{2A}\left(x\right)\left(2+\beta f_{\theta^{g}}^{A}\left(x\right)\right)\right)u_{\theta^{g}}\left(x\right)u^{T}_{\theta^{g}}\left(x\right)$ $\displaystyle V_{g}$ $\displaystyle=$ $\displaystyle\displaystyle Var_{g}\left(u_{\theta^{g}}\left(X\right)K^{\prime}\left(\delta_{g}^{g}\left(X\right)\right)\left(\left(A+B\right)f_{\theta^{g}}^{\alpha}\left(X\right)+A^{2}\beta^{2}e^{\beta f_{\theta^{g}}^{A}\left(X\right)}f_{\theta^{g}}^{2A-1}\left(X\right)\right)\right),$ (22) where, $X$ is a random variable having density $g$, $Var_{g}$ represents variance under the density $g$, $\delta_{g}\left(x\right)=\frac{g\left(x\right)}{f_{\theta}\left(x\right)}-1$, $K^{\prime}\left(\cdot\right)$ is the derivative of $K\left(\cdot\right)$ with respect to its argument, $\delta_{g}^{g}\left(x\right)=\frac{g\left(x\right)}{f_{\theta}^{g}\left(x\right)}-1$ and $i_{\theta}\left(x\right)=-u^{\prime}_{\theta}\left(x\right)$, the negative of the derivative of the score function with respect to the parameter. ###### Theorem 1. Under the above-mentioned set-up and certain regularity assumptions given in the Online Supplement, there exists a consistent sequence of roots $\hat{\theta}_{n}$ of the estimating equation (18). Moreover, the asymptotic distribution of $\sqrt{n}\left(\hat{\theta}_{n}-\theta^{g}\right)$ is p-dimensional normal with mean $0$ and $J_{g}^{-1}V_{g}J_{g}^{-1}$. ###### Corollary 5.1. When $g=f_{\theta}$ for some $\theta\in\Theta$, then $\sqrt{n}\left(\theta_{n}-\theta\right)\sim N\left(0,J^{-1}VJ^{-1}\right)$ asymptotically, where, $\displaystyle J$ $\displaystyle=$ $\displaystyle E_{f_{\theta}}\left\\{u_{\theta}\left(X\right)u^{T}_{\theta}\left(X\right)\left(\left(A+B\right)f^{\alpha}\left(x\right)+A^{2}\beta^{2}e^{\beta f_{\theta}^{A}\left(X\right)}f_{\theta}^{2A-1}\left(X\right)\right)\right\\}$ $\displaystyle=$ $\displaystyle\sum_{x=0}^{\infty}\left\\{u_{\theta}\left(x\right)u^{T}_{\theta}\left(x\right)\left(\left(A+B\right)f^{\alpha}\left(x\right)+A^{2}\beta^{2}e^{\beta f_{\theta}^{A}\left(x\right)}f_{\theta}^{2A-1}\left(x\right)\right)\right\\}f_{\theta}\left(x\right),$ $\displaystyle V$ $\displaystyle=$ $\displaystyle V_{f_{\theta}}\left\\{u_{\theta}\left(X\right)\left(\left(A+B\right)f_{\theta}^{\alpha}\left(X\right)+A^{2}\beta^{2}e^{\beta f_{\theta}^{A}\left(X\right)}f_{\theta}^{2A-1}\left(X\right)\right)\right\\}$ (24) $\displaystyle=$ $\displaystyle\left(A+B\right)^{2}\sum_{x=0}^{\infty}f^{1+2\alpha}\left(x\right)u_{\theta}\left(x\right)u^{T}_{\theta}\left(x\right)$ $\displaystyle+$ $\displaystyle A^{4}\beta^{4}\sum_{x=0}^{\infty}e^{2\beta f_{\theta}^{A}\left(x\right)}f_{\theta}^{4A-1}\left(x\right)u_{\theta}\left(x\right)u^{T}_{\theta}\left(x\right)$ $\displaystyle+$ $\displaystyle 2\left(A+B\right)A^{2}\beta^{2}\sum_{x=0}^{\infty}e^{\beta f_{\theta}^{A}\left(x\right)}f_{\theta}^{2A+\alpha}\left(x\right)u_{\theta}\left(x\right)u^{T}_{\theta}\left(x\right)-\zeta\zeta^{\prime},$ where, $\zeta=\displaystyle\sum_{x=0}^{\infty}u_{\theta}\left(x\right)\left(\left(A+B\right)f^{A+B}\left(x\right)+A^{2}\beta^{2}e^{\beta f_{\theta}^{A}\left(x\right)}f_{\theta}^{2A}\left(x\right)\right).$ ## 6 Influence analysis of the minimum GSB estimator Here we study the stability of our proposed class of estimators on the basis of the influence function (IF), which measures the effect of adding an infinitesimal mass to the distribution and is one of the most important heuristic tools of robustness. A simple differentiation of a contaminated version of the estimating equation (18) leads to the expression $IF\left(y,G,T_{\alpha,\lambda,\beta}\right)=J_{G}^{-1}N_{G}\left(y\right),~{}~{}\rm{where},$ (25) $\displaystyle N_{G}\left(y\right)$ $\displaystyle=$ $\displaystyle\displaystyle\left(A^{2}\beta^{2}e^{\beta f^{A}_{\theta^{g}}\left(y\right)}f^{A}_{\theta^{g}}\left(y\right)+(A+B)f^{B}_{\theta^{g}}\left(y\right)\right)g^{A-1}\left(y\right)u_{\theta^{g}}\left(y\right)$ $\displaystyle-$ $\displaystyle\sum_{x=0}^{\infty}\left(A^{2}\beta^{2}e^{\beta f^{A}_{\theta^{g}}\left(x\right)}f^{A}_{\theta^{g}}\left(x\right)+(A+B)f^{B}_{\theta^{g}}\left(x\right)\right)g^{A}\left(x\right)u_{\theta^{g}}\left(x\right),$ $\displaystyle J_{G}$ $\displaystyle=$ $\displaystyle\displaystyle A^{2}\beta^{2}\sum_{x=0}^{\infty}e^{\beta f^{A}_{\theta^{g}}\left(x\right)}f^{A}_{\theta^{g}}\left(x\right)\left(2f^{A}_{\theta^{g}}\left(x\right)-g^{A}\left(x\right)\right)u_{\theta^{g}}\left(x\right)u^{T}_{\theta^{g}}\left(x\right)$ $\displaystyle+$ $\displaystyle A^{3}\beta^{3}\sum_{x=0}^{\infty}e^{\beta f^{A}_{\theta^{g}}\left(x\right)}f^{2A}_{\theta^{g}}\left(x\right)\left(f^{A}_{\theta^{g}}\left(x\right)-g^{A}\left(x\right)\right)u_{\theta^{g}}\left(x\right)u^{T}_{\theta^{g}}\left(x\right)$ $\displaystyle+$ $\displaystyle(A+B)\sum_{x=0}^{\infty}f^{B}_{\theta^{g}}\left(x\right)\left(\left(A+B\right)f^{A}_{\theta^{g}}\left(x\right)-Bg^{A}\left(x\right)\right)u_{\theta^{g}}\left(x\right)u^{T}_{\theta^{g}}\left(x\right)$ $\displaystyle+$ $\displaystyle A^{2}\beta^{2}\sum_{x=0}^{\infty}e^{\beta f^{A}_{\theta^{g}}\left(x\right)}f^{A}_{\theta^{g}}\left(x\right)\left(g^{A}\left(x\right)-f^{A}_{\theta^{g}}\left(x\right)\right)i_{\theta^{g}}\left(x\right)$ $\displaystyle-$ $\displaystyle(A+B)\sum_{x=0}^{\infty}f^{B}_{\theta^{g}}\left(f^{A}_{\theta^{g}}\left(x\right)-g^{A}\left(x\right)\right)i_{\theta^{g}}\left(x\right).$ If the distribution $G$ belongs to the model family $\mathcal{F}$ with $g=f_{\theta}$, then the influence function reduces to, $\displaystyle IF\left(y,F_{\theta},T_{\alpha,\lambda,\beta}\right)$ $\displaystyle=$ $\displaystyle J_{F_{\theta}}^{-1}N_{F_{\theta}}\left(y\right),\rm{where},$ (26) $\displaystyle J_{F_{\theta}}$ $\displaystyle=$ $\displaystyle\displaystyle\sum_{x=0}^{\infty}\left(A^{2}\beta^{2}e^{\beta f^{A}_{\theta}\left(x\right)}f^{2A}_{\theta}\left(x\right)+(A+B)f^{A+B}_{\theta}\left(x\right)\right)u_{\theta}\left(x\right)u^{T}_{\theta}\left(x\right),$ $\displaystyle N_{F_{\theta}}\left(y\right)$ $\displaystyle=$ $\displaystyle\displaystyle A^{2}\beta^{2}e^{\beta f^{A}_{\theta}\left(y\right)}f^{2A-1}_{\theta}\left(y\right)u_{\theta}\left(y\right)+(A+B)f^{A+B-1}_{\theta}\left(y\right)u_{\theta}\left(y\right)$ $\displaystyle-$ $\displaystyle\sum_{x=0}^{\infty}A^{2}\beta^{2}e^{\beta f^{A}_{\theta}\left(x\right)}f^{2A}_{\theta}\left(x\right)u_{\theta}\left(x\right)-\sum_{x=0}^{\infty}(A+B)f^{A+B}_{\theta}\left(x\right)u_{\theta}\left(x\right).$ Evidently, the influence function is dependent on all the three tuning parameters. Whenever the matrix $J_{F_{\theta}}$ is non singular, the boundedness of the influence function depends on the ability of the coefficients to control the score function $u_{\theta}(y)$ in the first two terms of the numerator. In most parametric models including all exponential family models, $f_{\theta}^{\tau}(y)u_{\theta}(y)$ remains bounded for any $\tau>0$; in the case $\tau=0$, however the expression equals $u_{\theta}(y)$ and there is no control over it to keep it bounded. For the second term of the numerator in Equation (6), this is achieved when $A+B>1$, i.e., when $\alpha>0$. The first term of the numerator contains an additional exponential term. However, given that $f_{\theta}(y)\leq 1$ for any value $y$ in the support of a discrete random variable, the first term of the numerator is easily seen to be bounded for any fixed non-zero real $\beta$ when $2A-1>0$, i.e., $A>1/2$. We now list the different possible cases for boundedness of the influence function as follows: 1. 1. $\beta=0$; here the first and third terms of the numerator vanish, and the only other condition necessary is $A+B>1$, i.e., $\alpha>0$. This is essentially the $S$-divergence case, and shows that all minimum $S$-divergence functionals with $\alpha>0$ have bounded influence (irrespective of the value of $\lambda$). In this case the allowable region for the triplet $(\alpha,\lambda,\beta)$ for bounded influence is ${\mathbb{S}}_{1}=\left(\alpha>0,\lambda\in{\mathbb{R}},\beta=0\right)$. 2. 2. $\beta\neq 0$, $A=0$. In this case also the first and third terms of the numerator drop out and the additional required condition is $\alpha>0$. However, since $A=1+\lambda(1-\alpha)=0$, this implies $\lambda=-\frac{1}{1-\alpha}$. In this case the influence function is independent of $\beta$. Now the relevant region for the triplet is ${\mathbb{S}}_{2}=\left(\alpha>0,\lambda=-\frac{1}{1-\alpha},\beta\neq 0\right)$. 3. 3. Now suppose $A+B=0$, without the components being individually zero. In this case the second and fourth terms get eliminated and we have $\alpha=-1$. In this case the condition $2A-1>0$ translates to $\lambda>-\frac{1}{4}$. Here the corresponding region for the triplet is ${\mathbb{S}}_{3}=\left(\alpha=-1,\lambda\geq-\frac{1}{4},\beta\neq 0\right)$. 4. 4. Now we allow all the terms $\beta$, $A$ and $A+B$ to be non-zero. In this case all the four terms of the numerator are non-vanishing. Then, beyond the condition on $\beta$, the required conditions are $\alpha>0$ and $\lambda\left(1-\alpha\right)>-\frac{1}{2}$. The region here is ${\mathbb{S}}_{4}=\left(\alpha>0,\lambda(1-\alpha)>-\frac{1}{2},\beta\neq 0\right)$. Combining all the cases, we see that the IF will be bounded if the triplet $\left(\alpha,\lambda,\beta\right)\in\mathbb{S}={\mathbb{S}}_{1}\cup{\mathbb{S}}_{2}\cup{\mathbb{S}}_{3}\cup{\mathbb{S}}_{4}$. It is easily seen that the four constituent subregions are disjoint. For illustration, we present some plots for bounded and unbounded influence functions for the minimum GSB functional under the Poisson($\theta$) model in Figure 1, where the true data distribution is Poisson(3). In the four rows of the right panel we give examples of triplets belonging to the four disjoint components of ${\mathbb{S}}$. In the first two rows of the right panel, the $\alpha$ value alone determines the shape of the curve. On the $i$-th row of the left panel, on the other hand, the triplets are slightly different from the triplets of $i$-th row on the right, but far enough to be pushed out of ${\mathbb{S}}_{i}$. Accordingly, all the plots on the left correspond to unbounded influence functions. Generally, it may also be observed that for increasing $\beta$ the curves get flatter in each plot, where IF varies over different $\beta$. We will provide further illustration of the bounded influence region of the triplet through three-dimensional graphs at the end of the simulation section. ## 7 Simulation results In the simulation section our aim is to demonstrate that by choosing non-zero values of the parameter $\beta$, we may be able to generate procedures which, in a suitable sense, improve upon the estimators that are provided by the existing standard, the class of $S$-divergences. We consider the Poisson ($\theta$) model, and choose samples of size 50 from the $(1-\epsilon){\rm Poisson}(3)+\epsilon{\rm Poisson}(10)$ mixture, where the second component is the contaminant and $\epsilon\in[0,1)$ is the contaminating proportion. The values 0, 0.05, 0.1 and 0.2 are considered for $\epsilon$, and at each contamination level, the samples are replicated 1000 times. The Poisson parameter is estimated in each of the 1000 replications, for each contamination level, and at each of several $(\alpha,\lambda,\beta)$ triplets considered in our study. Subsequently we construct the empirical mean square error (MSE) against the target value of 3, for each tuning parameter triplet and each contamination level over the 1000 replications. In case of the minimum $S$-divergence estimator, [13] have empirically identified a subset of $\left(\alpha,\lambda\right)$ collections which represent good choices. According to them, the zone of “best” estimators correspond to an elliptical subset of the tuning parameter space, with $\alpha\in[0.1,0.6]$ and $\lambda\in[-1,-0.3]$. We hope to show that for most of the $(\alpha,\lambda)$ combinations (including the best ones) there is a corresponding better or competitive $\left(\alpha,\lambda,\beta\right)$ combination with a non-zero $\beta$, thus providing an option which appears to perform better, at least to the extent of the findings in these simulations. We begin with an exploration of the $S$-divergence, since this is the basis for comparison. The MSEs are presented in Table 3 over a cross-classified grid with $\alpha$ values in {0.1, 0.25, 0.4, 0.5, 0.6, 0.8, 1} and $\lambda$ values in {-1, -0.7, -0.5, -0.3, 0, 0.2, 0.5, 0.8, 1}, a total of 63 cells. In each cell the empirical MSEs for $\epsilon=0,0.05,0.1$ and $0.2$ are presented in a column of four elements, in that order, followed by the corresponding combination of tuning parameters $(\alpha,\lambda,\beta=0)$. We have carried the $\beta=0$ parameter in each triplet of parameters, to indicate that the $S$-divergence is indeed a special case of the GSB divergence. It may be noted that between all the cells, there is no unique $(\alpha,\lambda)$ combination which produces an overall best result (in terms of smallest MSE) over all the four columns (levels of contamination). We now expand the exploration by considering, in addition, a grid of possible non-zero $\beta$ values at each $(\alpha,\lambda)$ combination to see if the results can be improved. To be conservative about our definition of improvement, we declare the existence of a “better” triplet in the GSB sense if all the four mean square errors corresponding to a ${(\alpha,\lambda)}$ combination within the $S$-divergence family in Table 3 are improved (reduced) by a suitable member of the GSB divergence class which is strictly outside the $S$-divergence family (corresponding to a non-zero $\beta$). | ---|--- | | | Figure 1: Examples of unbounded influence functions (left panel) and bounded influence functions (right panel) corresponding to $\left(\alpha,\lambda,\beta\right)\in$ each disjoint subsets contained in $\mathbb{S}$. Our exploration indicates that in a large majority of the 63 cells there is a member of the GSB divergence with a non-zero $\beta$ parameter which improves (over all the four cells) the performance of the corresponding $S$-divergence estimator with the same $(\alpha,\lambda)$ combination. Interestingly it turns out that in practically all the cases where an improvement is observed it happens for a negative value of $\beta$ (it is observed to be zero in rare cases, but is never positive). A more detailed inspection indicates that in many of these cases, the improvement occurs at the value $\beta=-4$. In order to summarize the findings of this rather large exploration (presented in Table 4) in a meaningful manner, we first note the following different cases, 1. 1. (First Case) These are the cells where all the four mean square errors for the $S$-divergence case are reduced by the minimum GSB divergence estimator with the same values of $(\alpha,\lambda)$ and $\beta=-4$. These cells are highlighted with the blue colour in Table 4. (There are 18 such cells). 2. 2. (Second Case) These are the cells where all the four MSEs for the $S$-divergence case are reduced by a minimum GSB divergence estimator with $\beta=-4$ but with a different $(\alpha,\lambda)$ combination than that for the corresponding cell. These cells are highlighted in red in Table 4. (There are 39 such cells). 3. 3. (Third Case) These are the cells where all the four MSEs are reduced by a minimum GSB divergence estimator outside the $S$-divergence family, but with $\beta\neq-4$, and not necessarily the same $(\alpha,\lambda)$. These cells are highlighted in orange in Table 4. (There is one such cell). 4. 4. (Fourth Case) These are the cells where some triplet within the minimum GSB divergence class can improve upon the three MSEs under contamination ($\epsilon=0.05,0.1,0.2$) but not all the four MSEs simultaneously. While these are not “better” triplets in the sense described earlier in the section, the pure data MSEs (not reported here) for these triplets are close to those of the S-Divergence MSEs for these cells; in this sense these triplets are at least competitive. These cells are highlighted in green in Table 4. (There are three such cases). 5. 5. (Fifth Case) These are the cells where no $(\alpha,\lambda,\beta)$ provides an improvement over the S-divergence results in the sense of any of the previous four cases (although there are competitive alternatives). These cells remain in black in Table 4. There are 2 such cells. On the whole, therefore, it turns out that we observe improvements in 57 out of the 63 cells in all four rows of the column of MSEs in that cell by choosing $\beta=-4$ together with the $S$-divergence parameters. Even in the handful of cases (cells) where we do not have an improvement in all the rows of the column, there generally are competitive (although not strictly better) options within the minimum GSB divergence class with a negative value of $\beta$. In Table 4, in each cell, we also present the particular $(\alpha,\lambda,\beta)$ combination which generates the mean square errors (improved over Table 3 in most cases, as we have seen) reported in that cell. In Figure 2, we provide a three-dimensional plot (as described in that section) in the three-dimensional $(\alpha,\lambda,\beta)$ plane, where the region ${\cal S}$ has been expressed as a union of several colour-coded subregions representing the individual components. The triplets corresponding to the improved MSE solutions reported in the cells of Table 4 all belong to the blue subregion of this figure, indicating that all improved solutions are provided by bounded influence estimators. Table 3: MSEs of the minimum divergence estimators within the $S$-divergence family for pure and contaminated data 0.1968 | 0.0836 | 0.0704 | 0.0708 | 0.0733 | 0.0802 | 0.0876 ---|---|---|---|---|---|--- 0.1974 | 0.0981 | 0.0855 | 0.0852 | 0.0869 | 0.0926 | 0.0994 0.1753 | 0.1063 | 0.1012 | 0.1028 | 0.1054 | 0.1116 | 0.1118 0.3099 | 0.2245 | 0.2119 | 0.2113 | 0.2130 | 0.2200 | 0.2298 (0.1, $-1$, 0) | (0.25, -1, 0) | (0.4, -1, 0) | (0.5, -1, 0) | (0.6, -1, 0) | (0.8, -1, 0) | (1, -1, 0) 0.0751 | 0.0666 | 0.0673 | 0.0698 | 0.0729 | 0.0800 | 0.0876 0.0893 | 0.0830 | 0.0831 | 0.0847 | 0.0869 | 0.0927 | 0.0994 0.1081 | 0.1044 | 0.1045 | 0.1056 | 0.1073 | 0.1121 | 0.1118 0.2830 | 0.2505 | 0.2328 | 0.2264 | 0.2231 | 0.2233 | 0.2298 (0.1, -0.7, 0) | (0.25, -0.7, 0) | (0.4, -0.7, 0) | (0.5, -0.7, 0) | (0.6, -0.7, 0) | (0.8, -0.7, 0) | (1, -0.7, 0) 0.0638 | 0.0635 | 0.0665 | 0.0694 | 0.0727 | 0.0799 | 0.0876 0.0836 | 0.0821 | 0.0832 | 0.0849 | 0.0871 | 0.0927 | 0.0994 0.1203 | 0.1120 | 0.1087 | 0.1083 | 0.1089 | 0.1125 | 0.1118 0.3715 | 0.2958 | 0.2559 | 0.2408 | 0.2319 | 0.2258 | 0.2298 (0.1, -0.5, 0) | (0.25, -0.5, 0) | (0.4, -0.5, 0) | (0.5, -0.5, 0) | (0.6, -0.5, 0) | (0.8, -0.5, 0) | (1, -0.5, 0) 0.0600 | 0.0622 | 0.0660 | 0.0691 | 0.0725 | 0.0798 | 0.0876 0.0895 | 0.0846 | 0.0843 | 0.0856 | 0.0875 | 0.0928 | 0.0994 0.1554 | 0.1264 | 0.1149 | 0.1118 | 0.1109 | 0.1129 | 0.1118 0.5669 | 0.3709 | 0.2904 | 0.2605 | 0.2424 | 0.2286 | 0.2298 (0.1, -0.3, 0) | (0.25, -0.3, 0) | (0.4, -0.3, 0) | (0.5, -0.3, 0) | (0.6, -0.3, 0) | (0.8, -0.3, 0) | (1, -0.3, 0) 0.0592 | 0.0617 | 0.0657 | 0.0688 | 0.0721 | 0.0796 | 0.0876 0.1415 | 0.0971 | 0.0880 | 0.0873 | 0.0883 | 0.0929 | 0.0994 0.3491 | 0.1774 | 0.1308 | 0.1196 | 0.1147 | 0.1136 | 0.1118 1.3860 | 0.6555 | 0.3832 | 0.3061 | 0.2655 | 0.2333 | 0.2298 (0.1, 0, 0) | (0.25, 0, 0) | (0.4, 0, 0) | (0.5, 0, 0) | (0.6, 0, 0) | (0.8, 0, 0) | (1, 0, 0) 0.0608 | 0.0621 | 0.0657 | 0.0687 | 0.0721 | 0.0794 | 0.0876 0.3302 | 0.1231 | 0.0930 | 0.0892 | 0.0890 | 0.0930 | 0.0994 0.8565 | 0.2745 | 0.1508 | 0.1276 | 0.1181 | 0.1141 | 0.1118 2.5938 | 1.0853 | 0.5550 | 0.3553 | 0.2867 | 0.2370 | 0.2298 (0.1, 0.2, 0) | (0.25, 0.2, 0) | (0.4, 0.2, 0) | (0.5, 0.2, 0) | (0.6, 0.2, 0) | (0.8, 0.2, 0) | (1, 0.2, 0) 0.0671 | 0.0638 | 0.0658 | 0.0685 | 0.0718 | 0.0792 | 0.0876 1.1434 | 0.3251 | 0.1115 | 0.0943 | 0.0907 | 0.0931 | 0.0994 2.3829 | 0.8165 | 0.2234 | 0.1489 | 0.1255 | 0.1151 | 0.1118 4.8817 | 2.4261 | 0.8641 | 0.4847 | 0.3338 | 0.2434 | 0.2298 (0.1, 0.5, 0) | (0.25, 0.5, 0) | (0.4, 0.5, 0) | (0.5, 0.5, 0) | (0.6, 0.5, 0) | (0.8, 0.5, 0) | (1, 0.5, 0) 0.0778 | 0.0676 | 0.0665 | 0.0685 | 0.0716 | 0.0790 | 0.0876 1.9928 | 0.9339 | 0.1951 | 0.1068 | 0.0936 | 0.0933 | 0.0994 3.7890 | 1.9909 | 0.4869 | 0.1994 | 0.1378 | 0.1162 | 0.1118 6.6731 | 4.2592 | 1.6784 | 0.7520 | 0.4130 | 0.2511 | 0.2298 (0.1, 0.8, 0) | (0.25, 0.8, 0) | (0.4, 0.8, 0) | (0.5, 0.8, 0) | (0.6, 0.8, 0) | (0.8, 0.8, 0) | (1, 0.8, 0) 0.0863 | 0.0717 | 0.0673 | $0.0686$ | 0.0714 | 0.0789 | 0.0876 2.4449 | 1.3987 | 0.3803 | $0.1283$ | 0.0967 | 0.0934 | 0.0994 4.5000 | 2.7992 | 0.9117 | $0.2793$ | 0.1514 | 0.1171 | 0.1118 7.5320 | 5.3554 | 2.5215 | $1.0745$ | 0.4969 | 0.2572 | 0.2298 (0.1, 1, 0) | (0.25, 1, 0) | (0.4, 1, 0) | (0.5, 1, 0) | (0.6, 1, 0) | (0.8, 1, 0) | (1, 1, 0) Table 4: MSEs of the minimum GSB divergence estimators under pure and contaminated data 0.0623 | 0.0696 | 0.0704 | 0.0708 | 0.0687 | 0.0720 | 0.0763 ---|---|---|---|---|---|--- 0.0816 | 0.0843 | 0.0855 | 0.0852 | 0.0833 | 0.0859 | 0.0892 0.1115 | 0.1056 | 0.1012 | 0.1028 | 0.1042 | 0.1060 | 0.1077 0.2831 | 0.2207 | 0.2119 | 0.2113 | 0.2110 | 0.2162 | 0.2115 (0.4, $-0.4$, -4) | (0.8, -0.5, -4) | (0.4, -1, 0) | (0.5, -1, 0) | (0.8, 0, -7.5) | (0.8, -0.3, -4) | (1, -1, -4) 0.0642 | 0.0681 | 0.0681 | 0.0696 | 0.0696 | 0.0720 | 0.0763 0.0816 | 0.0826 | 0.0826 | 0.0843 | 0.0843 | 0.0859 | 0.0892 0.1076 | 0.1043 | 0.1043 | 0.1055 | 0.1055 | 0.1060 | 0.1077 0.2514 | 0.2135 | 0.2135 | 0.2207 | 0.2207 | 0.2162 | 0.2115 (0.6, -0.5, -4) | (0.8, 0, -8) | (0.8, 0, -8) | (0.8, -0.5, -4) | (0.8, -0.5, -4) | (0.8, -0.3, -4) | (1, -0.7, -4) 0.0623 | 0.0623 | 0.0659 | 0.0678 | 0.0678 | 0.0696 | 0.0763 0.0816 | 0.0816 | 0.0825 | 0.0834 | 0.0834 | 0.0843 | 0.0892 0.1115 | 0.1115 | 0.1071 | 0.1061 | 0.1061 | 0.1055 | 0.1077 0.2831 | 0.2831 | 0.2417 | 0.2295 | 0.2295 | 0.2207 | 0.2115 (0.4, -0.4, -4) | (0.4, -0.4, -4) | (0.5, -0.3, -4) | (0.6, -0.3, -4) | (0.6, -0.3, -4) | (0.8, -0.5, -4) | (1, -0.5, -4) 0.0600 | 0.0619 | 0.0642 | 0.0659 | 0.0678 | 0.0720 | 0.0763 0.0845 | 0.0822 | 0.0819 | 0.0825 | 0.0834 | 0.0859 | 0.0892 0.1294 | 0.1154 | 0.1091 | 0.1071 | 0.1061 | 0.1060 | 0.1077 0.4049 | 0.3080 | 0.2602 | 0.2417 | 0.2295 | 0.2162 | 0.2115 (0.1, -0.3, -4) | (0.25, -0.3, -4) | (0.4, -0.3, -4) | (0.5, -0.3, -4) | (0.6, -0.3, -4) | (0.8, -0.3, -4) | (1, -0.3, -4) 0.0600 | 0.0600 | 0.0644 | 0.0644 | 0.0644 | 0.0755 | 0.0763 0.0845 | 0.0845 | 0.0817 | 0.0817 | 0.0817 | 0.0887 | 0.0892 0.1294 | 0.1294 | 0.1073 | 0.1073 | 0.1073 | 0.1077 | 0.1077 0.4049 | 0.4049 | 0.2492 | 0.2492 | 0.2492 | 0.2139 | 0.2115 (0.1, -0.3, -4) | (0.1, -0.3, -4) | (0.8, -1, -4) | (0.8, -1, -4) | (0.8, -1, -4) | (0.8, 0, -4) | (1, 0, -4) 0.0600 | 0.0600 | 0.0644 | 0.0644 | 0.0644 | 0.0779 | 0.0763 0.0845 | 0.0845 | 0.0817 | 0.0817 | 0.0817 | 0.0907 | 0.0892 0.1294 | 0.1294 | 0.1073 | 0.1073 | 0.1073 | 0.1092 | 0.1077 0.4049 | 0.4049 | 0.2492 | 0.2492 | 0.2492 | 0.2146 | 0.2115 (0.1, -0.3, -4) | (0.1, -0.3, -4) | (0.8, -1, -4) | (0.8, -1, -4) | (0.8, -1, -4) | (0.8, 0.2, -4) | (1, 0.2, -4) 0.0600 | 0.0600 | 0.0644 | 0.0644 | 0.0644 | 0.0696 | 0.0763 0.0845 | 0.0845 | 0.0817 | 0.0817 | 0.0817 | 0.0843 | 0.0892 0.1294 | 0.1294 | 0.1073 | 0.1073 | 0.1073 | 0.1055 | 0.1077 0.4049 | 0.4049 | 0.2492 | 0.2492 | 0.2492 | 0.2207 | 0.2115 (0.1, -0.3, -4) | (0.1, -0.3, -4) | (0.8, -1, -4) | (0.8, -1, -4) | (0.8, -1, -4) | (0.8, -0.5, -4) | (1, 0.5, -4) 0.0600 | 0.0600 | 0.0644 | 0.0644 | 0.0644 | 0.0696 | 0.0763 0.0845 | 0.0845 | 0.0817 | 0.0817 | 0.0817 | 0.0843 | 0.0892 0.1294 | 0.1294 | 0.1073 | 0.1073 | 0.1073 | 0.1055 | 0.1077 0.4049 | 0.4049 | 0.2492 | 0.2492 | 0.2492 | 0.2207 | 0.2115 (0.1, -0.3, -4) | (0.1, -0.3, -4) | (0.8, -1, -4) | (0.8, -1, -4) | (0.8, -1, -4) | (0.8, -0.5, -4) | (1, 0.8, -4) 0.0600 | 0.0600 | 0.0644 | 0.0644 | 0.0644 | 0.0696 | 0.0763 0.0845 | 0.0845 | 0.0817 | 0.0817 | 0.0817 | 0.0843 | 0.0892 0.1294 | 0.1294 | 0.1073 | 0.1073 | 0.1073 | 0.1055 | 0.1077 0.4049 | 0.4049 | 0.2492 | 0.2492 | 0.2492 | 0.2207 | 0.2115 (0.1, -0.3, -4) | (0.1, -0.3, -4) | (0.8, -1, -4) | (0.8, -1, -4) | (0.8, -1, -4) | (0.8, -0.5, -4) | (1, 1, -4) --- --- Figure 2: The figure shows the region needed for bounded IF. Here, the grey, the orange, the green and the blue planes represent the boundaries of the sets ${\mathbb{S}}_{1}$, ${\mathbb{S}}_{2}$, ${\mathbb{S}}_{3}$ and ${\mathbb{S}}_{4}$, respectively. ## 8 Selection of tuning parameters Our simulations in the previous section seem to suggest that the minimum divergence estimators within the GSB class with $\beta=-4$ often provides good options for data analysis. To take full advantage of this observation, this subclass of the GSB family should be explored further. However, we want to fully exploit the flexibility of the three parameter system, and noting that in some cases the optimal is outside the $\beta=-4$ subclass, including some which generate the most competitive solutions in the full system, we want to use an overall data-based tuning parameter selection rule in which all the three parameters are allowed to vary over reasonable supports. The aim is to select the “best” tuning parameter combination depending on the amount of anomaly in the data. Thus datasets which show very close compatibility to the model should be analyzed by a triplet providing an efficient solution, while a more anomalous one should have a more robust member of the GSB class to deal with it. The current literature contains some suggestions for choosing data-driven tuning parameters in specific situations which we can make use of. The works of [18], [19] and [20] are relevant in this connection. Different algorithms for generating the “optimal” tuning parameter in case of the DPD have been described in these papers, which we will denote as the HK (for Hong and Kim), the OWJ (for one-step Warwick-Jones) and the IWJ (for the iterated Warwick- Jones) algorithms, respectively. The essential idea here is to construct an empirical approximation to the mean square error as a function of the tuning parameters (and a pilot estimator) and minimize it over the tuning parameter. The IWJ algorithm of [20] refines the OWJ algorithm of [19] by choosing the solution at a particular stage as the pilot for the next stage and going through an iterative process, thus eliminating the dependence on the pilot estimator as long as one has a good robust estimator to start with. The Hong and Kim algorithm does not consider the bias part of the mean square error and occasionally throws up highly non-robust solutions. See [20] for a full description and a comparative discussion of all the three algorithms. We will implement the same algorithms here, but for the GSB parameters rather than just the DPD parameter. In the following we have taken up two real data examples and considered the problem of selecting the “optimal” tuning parameters in each case. The OWJ algorithm considered here uses the minimum $L_{2}$ distance estimator as the pilot. Although the IWJ algorithm is pilot independent, for computational purposes it needs to commence from some suitable robust pilot for which also we utilize the minimum $L_{2}$ distance estimator. While the IWJ algorithm is our preferred method, we demonstrate the use of all the three algorithms in the following data sets. ###### Example 8.1. (Peritonitis Data): This example involves the incidence of peritonitis in $390$ kidney patients. The data are available in Chapter 2 of [6], and are also presented in the Online Supplement. The observations at 10 and 12 may be regarded as mild outliers. A geometric model with success probability $\theta$ has been fitted to these frequency data. Here, the IWJ solution coincides with the HK solution where the estimate of success probability is 0.5110 corresponding to $\left(\alpha,\lambda,\beta\right)=\left(0.41,-0.84,-3.5\right)$. The OWJ solution gives a slightly different success probability of 0.5105 corresponding to $\left(\alpha,\lambda,\beta\right)=\left(0.17,-0.60,-3\right)$. In case of clean data these IWJ, OWJ and HK estimates will be 0.5044, 0.5061 and 0.5029 corresponding to $\left(\alpha,\lambda,\beta\right)=\left(0.47,-1,-2\right)$, $\left(0.29,-1,-1\right)$ and $\left(0.55,-1,-3\right)$, respectively, being slightly different from each other. On the contrary, the MLEs for the full dataset and the (two) outlier deleted dataset are $0.4962$ and $0.5092$, respectively. Now we consider a more recent dataset for the implementation of our new proposal. ###### Example 8.2. (Stolen Bases Data): In “Major League Baseball (MLB) Player Batting Stats” for the 2019 MLB Regular Season, obtained from the ESPN.com website, one variable of interest is the number of Stolen Bases (SB) awarded to the top 40 Home Run (HR) scorers of the American League (AL). This dataset, containing three extreme and six moderate outliers, could be well-modelled by the Poisson distribution if not for the outliers. We are interested to estimate $\theta$, the average number of Stolen Bases (SB) awarded to the MLB batters of the AL throughout the whole regular season. The “optimal” estimates, derived from the implementation of the three algorithms under the Poisson model, are presented in Table 5. The fitted curves corresponding to some of these optimal estimates are given in Figure 3. It is clear that except for the full data MLE, all the other estimators primarily model the main model conforming part of the data and sacrifice the outliers. Table 5: Optimal estimates in different cases for the Stolen Bases Data data | method | optimal $\hat{\theta}$ | optimal $\left(\alpha,\lambda,\beta\right)$ ---|---|---|--- Full data (with outliers) | IWJ | $2.6270$ | $\left(0.65,-0.98,-8\right)$ | OWJ | $2.5086$ | $\left(0.73,-1,-8\right)$ | HK | $2.6409$ | $\left(0.65,-1,-8\right)$ | MLE | $4.875$ | $\left(0,0,0\right)$ excluding 9 outliers | IWJ | $2.3949$ | $\left(0.01,1.00,0\right)$ | OWJ | $2.3229$ | $\left(0.25,1.00,0\right)$ | HK | $2.6918$ | $\left(0.45,-1,-8\right)$ | MLE | $2.3871$ | $\left(0,0,0\right)$ --- Figure 3: Some significant fits for the Stolen Bases Data under the Poisson model. Here “clean data” refer to the modified data after removing all 9 outliers. ## 9 Concluding remarks In this paper, we have provided an extension of the ordinary Bregman divergence which has direct applications to developing new classes of divergence measures, and, in turn, in providing more options for minimum distance inference with better mixes of model efficiency and robustness. In the second part of the paper, we have made use of the suggested approach in generating a particular super-family of divergences which seems to work very well in practice and provides new minimum divergence techniques which appear to improve the performance of the $S$-divergence based procedures in many cases. Since the results presented here are based on a single study, more research will certainly be needed to decide to what extent the observed advantages of the procedures considered here can be generalized, but clearly there appears to be enough evidence to suggest such explorations are warranted. Even apart from the search for other divergences, several possible follow ups of this research immediately present themselves. This paper is restricted to discrete parametric models. An obvious follow up step is to suitably handle the case of continuous models, where the construction of the density and the divergence are more difficult. Another obvious extension will be to extend the procedures to more complicated data structures beyond the simple independently and identically distributed data scenarios. Yet another extension would be to apply this and other similarly developed super-divergences in the area of robust testing of hypothesis. We hope to take up all of these extension in our future work. The subfamily of the class of GSB divergences with $\beta=-4$ also needs some attention, and we hope to take it up in the future. For the time being we have presented the results for our real data examples for the $\beta=-4$ subfamily of GSB in the Online Supplement. ## Disclosure statement No potential conflict of interest was reported by the authors. ## References * [1] Csiszár I. Eine informations theoretische ungleichung und ihre anwendung auf den beweis der ergodizitat von Markoffschen ketten. Publ. Math. Inst. Hungar. Acad. Sci. 1963;3:85–107. * [2] Ali SM, Silvey SD. A general class of coefficients of divergence of one distribution from another. Journal of the Royal Statistical Society B (Methodological). 1966;28:131–142. * [3] Bregman LM. The relaxation method of finding the common point of convex sets and its application to the solution of problems in convex programming. USSR Computational Mathematics and Mathematical Physics. 1967;7:200–217. * [4] Beran R. Minimum Hellinger distance estimates for parametric models. The Annals of Statistics. 1977;5:445–463. * [5] Lindsay BG. Efficiency versus robustness: the case for minimum Hellinger distance and related methods. The Annals of Statistics. 1994;22:1081–1114. * [6] Basu A, Shioya H, Park C. Statistical inference: the minimum distance approach. Boca Raton (FL): CRC Press; 2011. * [7] Hampel FR, Ronchetti EM, Rousseeuw PJ, et al. Robust statistics: the approach based on influence functions. New York (NY): John Wiley and Sons; 1986. * [8] Huber PJ, Ronchetti EM. Robust statistics. Hoboken (NJ): John Wiley and Sons; 2009. * [9] Maronna RA, Martin RD, Yohai VJ. Robust statistics. Chichester: John Wiley and Sons; 2006. * [10] Basu A, Harris IR, Hjort NL, et al. Robust and efficient estimation by minimising a density power divergence. Biometrika. 1998;85:549–559. * [11] Cressie N, Read TRC. Multinomial goodness‐of‐fit tests. Journal of the Royal Statistical Society B (Methodological). 1984;46:440–464. * [12] Mukherjee T, Mandal A, Basu A. The B-exponential divergence and its generalizations with applications to parametric estimation. Statistical Methods and Applications. 2019;28:241–257. * [13] Ghosh A, Harris IR, Maji A, et al. A generalized divergence for statistical inference. Bernoulli. 2017;23:2746–2783. * [14] Banerjee A, Merugu S, Dhillon IS, et al. Clustering with Bregman divergences. Journal of Machine Learning Research. 2005;6:1705–1749. * [15] Amari S. Alpha-divergence is unique, belonging to both $f$-divergence and Bregman divergence classes. IEEE Transactions on Information Theory. 2009;55:4925–4931. * [16] Gutmann M, Hirayama J. Bregman divergence as general framework to estimate unnormalized statistical models. arXiv preprint arXiv:1202.3727. 2012. * [17] Jana S, Basu A. A characterization of all single-integral, non-kernel divergence estimators. IEEE Transactions on Information Theory. 2019;65:7976–7984. * [18] Hong C, Kim Y. Automatic selection of the tuning parameter in the minimum density power divergence estimation. J. Korean Stat. Soc. 2001;30:453–465. * [19] Warwick J, Jones MC. Choosing a robustness tuning parameter. J. Stat. Comput. Simul. 2005;75:581–588. * [20] Basak S, Basu A, Jones MC. On the ‘optimal’ density power divergence tuning parameter. Journal of Applied Statistics. 2020. doi:10.1080/02664763.2020.1736524.
# Comment on “Strong Quantum Darwinism and Strong Independence are Equivalent to Spectrum Broadcast Structure” Alexandre Feller1 Benjamin Roussel1 Irénée Frérot2,3 Pascal Degiovanni4 (1) European Space Agency - Advanced Concepts Team, ESTEC, Keplerlaan 1, 2201 AZ Noordwijk, The Netherlands. (2) ICFO-Institut de Ciencies Fotoniques, The Barcelona Institute of Science and Technology, Av. Carl Friedrich Gauss 3, 08860 Barcelona, Spain (3) Max-Planck-Institut für Quantenoptik, D-85748 Garching, Germany (4) Univ Lyon, Ens de Lyon, Université Claude Bernard Lyon 1, CNRS, Laboratoire de Physique, F-69342 Lyon, France ###### Abstract In a recent Letter [Phys. Rev. Lett. 122, 010403 (2019)], an equivalence is proposed between the so-called Spectrum Broadcast Structure for a system- multienvironment quantum state, and the conjunction of two information-theory notions: (a) Strong Quantum Darwinism; and (b) Strong Independence. Here, we show that the mathematical formulation of condition (b) by the authors (namely, the pairwise independence of the fragments of the environment, conditioned on the system), is necessary but not sufficient to ensure the equivalence. We propose a simple counter-example, together with a strengthened formulation of condition (b), ensuring the equivalence proposed by the authors. In their paper Castro-2019 , the authors introduce the notions of: (a) “strong quantum Darwinism”; and (b) “strong independence”. As the main result of the paper, it is proposed that, taken together, conditions (a) and (b) are equivalent to the so-called “spectrum broadcast structure” (SBS) for the system–environment quantum state $\rho_{{\cal S}{\cal E}}$ (throughout this Comment, we follow the notations and definitions used in the paper). By definition, if $\rho_{{\cal S}{\cal E}}$ has a SBS, the state of ${\cal E}={\cal E}_{1}\cdots{\cal E}_{F}$ relative to ${\cal S}$ is fully factorized, and has no correlations whatsoever; and the authors aimed at capturing this fully-factorized structure by an information-theory criterion of “strong independence”. In this Comment, we show that condition (b), while ensuring pairwise independence of the fragments ${\cal E}_{i}$ when conditioning on the system (a necessary condition for having the SBS), does not rule out the existence of higher-order correlations, and as such is insufficient to imply a SBS. We first construct a (completely classical) counter-example of a state with no SBS yet satisfying conditions (a) and (b). We then propose a stronger condition (b’), ensuring the equivalence of conditions (a) and (b’) with SBS. At a conceptual level, the main conclusion of the paper, namely the equivalence between, on the one hand, the SBS, and on the other, the conjunction of “strong quantum Darwinism” (as defined by the authors) and “strong independence” (as now defined by (b’)), remains therefore unaltered. Counter-example. We consider a qubit system ${\cal S}:\\{|0\rangle,|1\rangle\\}$, and the environment ${\cal E}={\cal E}_{1}{\cal E}_{2}{\cal E}_{3}$, with three qutrits fragments ${\cal E}_{i}:\\{|0_{i}\rangle,|1_{i}\rangle,|2_{i}\rangle\\}$ ($i\in\\{1,2,3\\}$). We define the projectors $\Pi_{a}^{({\cal S})}=|a\rangle\langle a|$ and $\Pi_{a}^{(i)}=|a_{i}\rangle\langle a_{i}|$. Finally, we define the projectors $\Pi_{abc}=\Pi_{a}^{(1)}\otimes\Pi_{b}^{(2)}\otimes\Pi_{c}^{(3)}$. Our counter-example is the state ($0\leq p\leq 1$): $\rho_{{\cal S}{\cal E}}=(1-p)~{}\Pi_{0}^{({\cal S})}\otimes\rho_{{\cal E}}^{(0)}+p~{}\Pi_{1}^{({\cal S})}\otimes\rho_{{\cal E}}^{(1)}~{},$ (1) with the relative states: $\displaystyle\rho_{{\cal E}}^{(0)}$ $\displaystyle=$ $\displaystyle\Pi_{000}$ (2) $\displaystyle\rho_{{\cal E}}^{(1)}$ $\displaystyle=$ $\displaystyle\frac{1}{4}[\Pi_{111}+\Pi_{122}+\Pi_{212}+\Pi_{221}]$ (3) $\displaystyle\rho_{{\cal E}_{i}{\cal E}_{j}}^{(1)}$ $\displaystyle=$ $\displaystyle\frac{1}{2}[\Pi_{1}^{(i)}+\Pi_{2}^{(i)}]\otimes\frac{1}{2}[\Pi_{1}^{(j)}+\Pi_{2}^{(j)}]$ (4) $\displaystyle\rho_{{\cal E}_{i}}^{(1)}$ $\displaystyle=$ $\displaystyle\frac{1}{2}[\Pi_{1}^{(i)}+\Pi_{2}^{(i)}]$ (5) As the relative states $\rho_{{\cal E}_{i}}^{(0)}=\Pi_{0}^{(i)}$ and $\rho_{{\cal E}_{i}}^{(1)}=[\Pi_{1}^{(i)}+\Pi_{2}^{(i)}]/2$ have disjoint supports, it is clear that each fragment has full access to the system’s state. Therefore: $I({\cal S}:{\cal E}_{i})=I_{\rm acc}({\cal S}:{\cal E}_{i})=\chi({\cal S}^{\Pi}:{\cal E}_{i})=H({\cal S})$ [condition (a)]. The state also fulfills condition (b) (pairwise-independence conditioned on ${\cal S}$). Indeed, the relative states $\rho_{{\cal E}_{i}{\cal E}_{j}}^{(0)}=\Pi_{0}^{(i)}\otimes\Pi_{0}^{(j)}$ and $\rho_{{\cal E}_{i}{\cal E}_{j}}^{(1)}$ [Eq. (4)] are product states, which implies that $I({\cal E}_{i}:{\cal E}_{j}|{\cal S})=0$ (as can be checked by direct computation on $\rho_{{\cal S}{\cal E}_{i}{\cal E}_{j}}$). Yet, the state $\rho_{{\cal S}{\cal E}_{1}{\cal E}_{2}{\cal E}_{3}}$ does not admit a SBS. Indeed, the relative state $\rho_{{\cal E}}^{(1)}$ [Eq. (3)] does not factorize (while pairs of fragments are uncorrelated, the parity of the total number of fragments in state $|1_{i}\rangle$ is odd, yielding genuine 3-partite correlations). This concludes our proof that the conditions (a) and (b) are insufficient to imply a SBS. Proposed correct formulation of the theorem. The above counter-example suggests to introduce a stronger condition (b’) ensuring the complete factorization of the fragments when conditioning on the system, which is a defining property of the SBS. Noticing that a multipartite state is factorized, $\rho_{{\cal E}_{1},\dots\cal E_{F}}=\otimes_{i=1}^{F}\rho_{{\cal E}_{i}}$, iff the multipartite mutual information vanishesYang-2009 : $I({\cal E}_{1},\dots\cal E_{F}):=\sum_{i=1}^{F}H(\rho_{{\cal E}_{i}})-H(\rho_{{\cal E}_{1},\dots\cal E_{F}})=0$, we are led to propose the following: Theorem. The state $\rho_{{\cal S}{\cal E}}$ has the SBS iff: * (a1) $I({\cal S}:{\cal E})=\chi({\cal S}^{\Pi},{\cal E})$ (classical–quantum state), * (a2) $I_{\rm acc}({\cal S}:{\cal E}_{i})=H({\cal S})$ for all $i$ (the information about ${\cal S}$ can be fully recovered by measuring any fragment), * (b’) $I({\cal E}_{1},\cdots{\cal E}_{F}|{\cal S})=0$ (totally-factorized relative states). ###### Proof. If the state has the SBS, it is clear that conditions (a1), (a2) and (b’) are fulfilled. Conversely, condition (a1) (vanishing discord) implies that the system–environment state is of the form $\rho_{{\cal S}{\cal E}}=\sum_{s}p_{s}\Pi_{s}^{({\cal S})}\otimes\rho_{\cal E}^{(s)}$ with $\\{\Pi_{s}^{({\cal S})}\\}$ forming mutually orthogonal projectors for the system. Condition (b’) then implies that the relative states factorize: $\rho_{\cal E}^{(s)}=\otimes_{i=1}^{F}\rho_{{\cal E}_{i}}^{(s)}$. Finally, condition (a2) implies that for each $i$, the states $\rho_{{\cal E}_{i}}^{(s)}$ are pairwise orthogonal. ∎ ###### Acknowledgements. We thank Thao P. Le and Alexandra Olaya-Castro for constructive feedback on our manuscript. This work has been supported by the European Space Agency (Ariadna study 1912-01). IF acknowledges support from the Government of Spain (FIS2020-TRANQI and Severo Ochoa CEX2019-000910-S), Fundació Cellex and Fundació Mir-Puig through an ICFO-MPQ Postdoctoral Fellowship, Generalitat de Catalunya (CERCA, AGAUR SGR 1381 and QuantumCAT). ## References * (1) Thao P. Le and Alexandra Olaya-Castro, _Strong quantum darwinism and strong independence are equivalent to spectrum broadcast structure_ , Phys. Rev. Lett. 122 (2019), 010403. * (2) Dong Yang, Karol Horodecki, Michal Horodecki, Pawel Horodecki, Jonathan Oppenheim, and Wei Song, _Squashed entanglement for multipartite states and entanglement measures based on the mixed convex roof_ , IEEE Transactions on Information Theory 55 (2009), no. 7, 3375–3387.
# Third-harmonic light polarization control in magnetically-resonant silicon metasurfaces Andrea Tognazzi 1,2 Kirill I. Okhlopkov 3 Attilio Zilli 4 Davide Rocco 1,2 Luca Fagiani 4,5 Erfan Mafakheri 5 Monica Bollani 5 Marco Finazzi 4 Michele Celebrano 4 Maxim R. Shcherbakov 3 Andrey A. Fedyanin 3 and Costantino De Angelis1,2 1Department of Information Engineering, University of Brescia, Via Branze 38, 25123 Brescia, Italy 2CNR-INO (National Institute of Optics), Via Branze 45, 25123 Brescia, Italy 3Faculty of Physics, Lomonosov Moscow State University, Moscow 119991, Russia 4Department of Physics, Politecnico di Milano, Piazza Leonardo Da Vinci 32, 20133 Milano, Italy 5CNR-IFN, LNESS laboratory, Via Anzani 42, 22100 Como, Italy <EMAIL_ADDRESS> ###### Abstract Nonlinear metasurfaces have become prominent tools for controlling and engineering light at the nanoscale. Usually, the polarization of the total generated third harmonic is studied. However, diffraction orders may present different polarizations. Here, we design an high quality factor silicon metasurface for third harmonic generation and perform back focal plane imaging of the diffraction orders, which present a rich variety of polarization states. Our results demonstrate the possibility of tailoring the polarization of the generated nonlinear diffraction orders paving the way to a higher degree of wavefront control. ††journal: oe ## 1 Introduction Manipulation of light is of paramount importance in many fields such as opto- electronics, image processing, sensing and cryptography[1, 2]. The 2D nature of metasurfaces, which are composed by an array of resonators, makes them suitable candidates for compact photonic devices[3, 4]. The optical properties of such structures can be tailored by tuning the geometrical parameters of each resonator in the periodic array or by changing the material[5, 6, 7]. Thanks to the improved accuracy of nanofabrication techniques, it is nowadays possible to obtain high quality nano-objects consisting of metals or dielectric materials to manipulate light in the visible and near-infrared regimes[8]. The applications of metasurfaces include beam steering [9, 10], light focusing[11, 12], holography[13], and sensing[14]. Implementing nonlinear optics at the nanoscale is very challenging because one cannot exploit phase matching, which can be achieved only over mesoscopic scales. In this frame, the added value of metasurfaces consists in the possibility of exploiting collective modes stemming from the interactions between neighboring nanoresonators to enhance the local electric field, improve the conversion efficiency, and tailor the emitted radiation[15, 16, 17, 18]. The low losses make dielectrics more suitable than metals for second- and third-harmonic generation (THG), all-optical switching and modulation of visible and near-infrared light[19, 20, 21, 22, 23, 24, 25, 26]. In the past few years, high-refractive index dielectric materials were employed to build nanoresonators to improve nonlinear frequency conversion [27, 28] and manipulate light emission [29, 30]. One of the most attractive materials for nanophotonics applications is silicon due to its well-established fabrication technology, high-refractive index and technological relevance[31, 32]. Previously, nonlinear beam deflection has been achieved by inducing a phase shift using different building blocks[33]. However, the polarization of the diffraction orders is usually an overlooked property when studying nonlinear gratings [34]. In this article, we report the design and fabrication of high quality factor ($Q$-factor) metasurfaces and we propose a simple electromagnetic model to explain the polarization of the third harmonic (TH) diffraction orders. Orthogonal polarizations are measured for different diffraction orders depending on the dominant multipolar component at resonance. Our results pave the way to the realization of a higher degree of polarization-controlled nonlinear diffractive metasurfaces. Figure 1: (a) Sketch of the metasurface and of the electric field distribution at normal incidence showing the magnetic quadrupole behaviour of the fundamental frequency. The near field distribution can be used to predict the diffraction orders polarization. The unitary cell is constituted by a silicon cuboid laying on a SiO2 substrate. (b) SEM planar view image of a dielectric metasurface image. Figure 2: Simulated reflectance for $p$ (a) and $s$ (b) incident polarization as a function of the wavelength and the incidence angle $\vartheta$. The dashed white lines delimit the FWHM bandwidth ($1554\pm 8$ $\mathrm{n}\mathrm{m}$) of the pump used in the experiments. For incident p-polarization, the high quality factor is preserved when $\vartheta$ increases and the resonance blue shifts. For incident s-polarization, the high quality factor resonance fades away when $\vartheta$ increases and a broader resonance appears at shorter wavelengths. ## 2 Design and fabrication We employ a commercial finite element solver (Comsol Multiphysics) to optimize the design of high-$Q$ metasurfaces made of silicon cuboids arranged in a periodic rectangular lattice (see Fig. 1a). We created a waveguide-like system with a channel coupling the light and the structure to obtain a metasurface with different quality factors depending on the excitation geometry. In the experiments, we achieve the resonant condition by changing the angle of incidence, this allows us to excite two different modes under incident p\- and s-polarization. The metasurfaces are realized on a silicon-on-insulator (SOI) substrate with a device layer of $H=125$ $\mathrm{n}\mathrm{m}$ on 2 $\mathrm{\SIUnitSymbolMicro}\mathrm{m}$ of buried oxide (see Fig. 1b). Arrays of rectangles (width $W=428$ $\mathrm{n}\mathrm{m}$, length $L=942$ $\mathrm{n}\mathrm{m}$ and periodicity $P_{x}$ and $P_{y}$ 1065–1060 $\mathrm{n}\mathrm{m}$, respectively) aligned along the [110] direction are patterned by means of e-beam lithography (EBL) and reactive ion etching (RIE). The resist is spin-coated on the SOI substrate and then exposed to the electron beam of a converted scanning electron microscope (SEM) along the designed pattern (acceleration voltage of 30 $\mathrm{k}\mathrm{V}$). A double layer of PMMA diluted in chlorobenzene, respectively, at 3.5% and 1.5% is employed. The dose used for the structures is 350 $\mathrm{\SIUnitSymbolMicro}\mathrm{C}\mathrm{/}\mathrm{c}\mathrm{m}\textsuperscript{2}$. After exposure, PMMA is developed in a solution of methyl isobutyl ketone (MIBK) and isopropanol (IPA) in a 1:3 ratio; MIBK is diluted in order to obtain well-defined profiles. The sample is immersed in this solution and agitated manually for 90 $\mathrm{s}$; a pure IPA solution is used for 1 minute to stop the development of the resist. Then, the pattern is transferred to the thin Si film by RIE in a CF4 plasma, using 80 $\mathrm{W}$ of radio frequency power and a total gas pressure of 5.4 $\mathrm{m}\mathrm{T}\mathrm{o}\mathrm{r}\mathrm{r}$. Finally, the resist is removed using acetone and the sample surface is exposed to O2 plasma in order to remove any residual resist. In the simulations, we model the silicon refractive index as reported in [35] and we assume a wavelength-independent refractive index (n=1.45) for the SiO2 substrate. The spatial period $P_{x}=P_{y}=P$ of the through-notches has been chosen to satisfy the matching condition resulting from momentum conservation between a normal incident plane wave with in-plane modes, i.e. $2\pi/P=\beta(\omega)$ where $\beta(\omega)$ is the propagation wavevector of the mode[15]. Possible deviations of $\beta(\omega)$ in the fabricated device from the simulated value can be matched by tuning the wavelength $\lambda_{0}$ or the angle $\theta$ of the incident plane wave. Fig. 2 shows sketches of the incident polarization and the reflectance (R) at the fundamental frequency (FF) as a function of $\lambda_{0}$ and $\theta$ for p and s polarized excitation with $\vec{E}_{0}(\omega)\perp\hat{x}$. When $\vec{E}_{0}(\omega)$ is p polarized the metasurface shows a sharp resonance ($Q=399$) which blue-shifts when $\vartheta$ is varied, albeit maintaining a narrow spectral width. When the impinging wave is s polarized, it excites a magnetic dipole mode with a lower $Q$ factor ($Q=29$). We then simulate the TH field $\vec{E}(3\omega)$ by evaluating in a second step of the computation the nonlinear current generated within the structures by the fundamental field $\vec{E}_{0}(\omega)$ through the third-order susceptibility as reported in [36]. The diffraction orders are calculated by performing the Fourier transform of the near field simulated at the TH frequency. ## 3 Experiments We employ a pulsed laser (160 fs) centered at 1554 $\mathrm{n}\mathrm{m}$ (FWHM=17 $\mathrm{n}\mathrm{m}$) and focus the beam in the back focal plane (BFP) of a 60x objective (Nikon, CFI Plan Fluor 60XC, NA=0.85) to obtain a loosely focused beam on the sample. We shift the pump beam in the BFP plane of the objective to change the angle of incidence. We collect the emitted TH through the same objective used for the excitation, and chromatically filter it. A Bertrand lens in the detection path focuses the TH beam in the BFP of the objective to image the TH diffraction orders, while a polarizer is used to analyze the polarization of the TH. A cooled CCD camera is used to acquire BFP images such as the ones in Fig. 3, where $\vartheta$ is the angle of incidence and $\phi$ is the polarizer angle with respect to the x-axis. ## 4 Results In Figs. 4(a,b) the intensity of TH light as function of the analyzer angle of different diffraction orders is reported for experiments (dashed) and simulations (continuous) for s and p-polarized fundamental wavelength light illuminating the sample at the incidence angle leading to maximum THG. Normalized experimental data in Fig. 4a refer to p-polarized light impinging on the sample at 41∘, corresponding to the maximum THG signal, while simulations corresponds to 32°. All the THG diffraction orders are polarized along the $y$ axis as predicted by the simulations. THG is maximum for incident s-polarization at 14∘ in the experiment and 22∘ in the simulations. The discrepancies in the resonant angles and the systematic tilt of the (-1,0) diffraction order polarization may be due to the uncertainty in the experimental pump beam angle of incidence and fabrication defects. In order to have a better insight on the polarization of diffraction orders we performed a cartesian multipolar decomposition of the TH field (see Fig. 4c,d). For incident p-polarization, the main multipolar component is always a magnetic quadrupole, $Q_{xz}$, whose amplitude is maximum at the resonant angle, leading to no variations of the diffraction order polarization when the angle of incidence is changed . For s-polarization, the main multipolar component changes at resonance and becomes a magnetic dipole along the z-axis, with a spatially non uniform far field polarization (see Fig.4d inset). This corresponds to a variation of the polarization of all the diffraction orders $(m,n)$ with $n\neq 1$. The polarization of the diffraction orders can be described by a simple formula, which takes into account the electromagnetic field distribution of each scatterer and the periodic structure. In the far- field region the total electric field radiated ($E_{t}$) by the metasurface is proportional to the far field radiated by the single array element ($E_{s}$) through the array factor ($AF$): $\vec{E}_{t}^{3\omega}=\vec{E}_{s}^{3\omega}AF(P,3\omega),$ (1) where the $AF(P,3\omega)$ is a function that depends only on the periodicity of the array and the TH frequency. Here, each cuboid can be envisioned as an antenna whose emission pattern is determined by the superposition of all the multipolar components and the diffraction orders are determined according to the $AF$ as described in [37]. This formalism enables one to tailor the polarization state of the nonlinear diffraction orders by engineering the main multipolar components at the TH frequency describing the single antenna behaviour. It is worth noting that this approach can be applied also for closely packed unit-cells once the multipolar decomposition is completely resolved for the meta-units that form the metasurface under test. Figure 3: (a,c) Experimental BFP images under normal incidence excitation. The red circle represents the numerical aperture of the objective (NA=0.85). (b,d) BFP images with tilted illumination at 41° for p and 14° for s polarized light. When the angle of incidence is increased, the (0,0) order is not emitted perpendicularly to the metasurface due to the in-plane components. For p-polarization, the diffraction orders move along $k_{y}$ since the incident beam wavevector lies in the yz-plane, while, for s-polarization, they move along $k_{x}$ since the incident wavevector is in the xz-plane. As a consequence, at certain angles, some diffraction orders disappear and others fall in the NA view window. Figure 4: Experimental (dashed) and simulated (continuous) polarization-resolved TH power excited at the angle with maximum THG for $p$ (a) and $s$ (b) polarization, respectively. The vertical double arrows represent the incident pump polarization. (c,d) Cartesian multipolar decomposition for incident $p$ and $s$ polarization, respectively. The insets in (c,d) represent the far field polarization of the magnetic quadrupole and of the magnetic dipole, respectively. ## 5 Conclusions We showed, both experimentally and numerically, a complex behaviour of the polarization of the TH diffraction orders as a function of the incidence angle of the fundamental pump beam. We applied a cartesian multipolar decomposition and a simple formula to describe the polarization of the diffraction orders and provide a method to tailor the far field properties of the metasurface. Our results demonstrate that the polarization of the diffraction orders is solely influenced by the near field distribution within each single element, which can be modelled by considering the multipolar decomposition, while the far field is determined by the period of the metasurface relative to the exciting wavelength. Our description can be readily applied to second-harmonic generation and to any type of periodic metasurface. ## Funding We acknowledge the financial support by the European Commission through the FET-OPEN projects "Narciso" (828890) and "METAFAST" (899673), by the Italian Ministry of University and Research (MIUR) through the PRIN project “NOMEN” (2017MP7F8F) and by Russian Ministry of Science and Higher Education (No 14.W03.31.0008). ## Disclosures The authors declare no conflicts of interest. ## References * [1] P. Ma, L. Gao, P. Ginzburg, and R. E. Noskov, “Ultrafast cryptography with indefinitely switchable optical nanoantennas,” Light: Science & Applications 7, 77 (2018). * [2] N. Liu, M. L. Tang, M. Hentschel, H. Giessen, and A. P. Alivisatos, “Nanoantenna-enhanced gas sensing in a single tailored nanofocus,” Nature Materials 10, 631–636 (2011). * [3] I. Kim, G. Yoon, J. Jang, P. Genevet, K. T. Nam, and J. Rho, “Outfitting next generation displays with optical metasurfaces,” ACS Photonics 5, 3876–3895 (2018). * [4] J. Scheuer, “Optical metasurfaces are coming of age: Short- and long-term opportunities for commercial applications,” ACS Photonics 7, 1323–1354 (2020). * [5] A. E. Minovich, A. E. Miroshnichenko, A. Y. Bykov, T. V. Murzina, D. N. Neshev, and Y. S. Kivshar, “Functional and nonlinear optical metasurfaces,” Laser & Photonics Reviews 9, 195–213 (2015). * [6] Q. He, S. Sun, and L. Zhou, “Tunable/reconfigurable metasurfaces: Physics and applications,” Research 2019, 1849272 (2019). * [7] S. M. Kamali, E. Arbabi, A. Arbabi, and A. Faraon, “A review of dielectric optical metasurfaces for wavefront control,” Nanophotonics 7, 1041 – 1068 (2018). * [8] M. Naffouti, R. Backofen, M. Salvalaglio, T. Bottein, M. Lodari, A. Voigt, T. David, A. Benkouider, I. Fraj, L. Favre, A. Ronda, I. Berbezier, D. Grosso, M. Abbarchi, and M. Bollani, “Complex dewetting scenarios of ultrathin silicon films for large-scale nanoarchitectures,” Science Advances 3, eaao1472 (2017). * [9] Z. Wei, Y. Cao, X. Su, Z. Gong, Y. Long, and H. Li, “Highly efficient beam steering with a transparent metasurface,” Optics Express 21, 10739–10745 (2013). * [10] F. A. and M. H., “Tunable two dimensional optical beam steering with reconfigurable indium tin oxide plasmonic reflectarray metasurface,” Journal of Optics 18, 125003 (2016). * [11] C. Schlickriede, S. S. Kruk, L. Wang, B. Sain, Y. Kivshar, and T. Zentgraf, “Nonlinear imaging with all-dielectric metasurfaces,” Nano Letters 20, 4370–4376 (2020). * [12] A. Benali, J.-B. Claude, N. Granchi, S. Checcucci, M. Bouabdellaoui, M. Zazoui, M. Bollani, M. Salvalaglio, J. Wenger, L. Favre, D. Grosso, A. Ronda, I. Berbezier, M. Gurioli, and M. Abbarchi, “Flexible photonic devices based on dielectric antennas,” Journal of Physics: Photonics 2, 015002 (2020). * [13] Y. Hu, X. Luo, Y. Chen, Q. Liu, X. Li, Y. Wang, N. Liu, and H. Duan, “3D-integrated metasurfaces for full-colour holography,” Light: Science & Applications 8, 86 (2019). * [14] F. A. A. Nugroho, D. Albinsson, T. J. Antosiewicz, and C. Langhammer, “Plasmonic metasurface for spatially resolved optical sensing in three dimensions,” ACS Nano 14, 2345–2353 (2020). * [15] M. R. Shcherbakov, K. Werner, Z. Fan, N. Talisa, E. Chowdhury, and G. Shvets, “Photon acceleration and tunable broadband harmonics generation in nonlinear time-dependent metasurfaces,” Nature Communications 10, 1345 (2019). * [16] G. Li, S. Zhang, and T. Zentgraf, “Nonlinear photonic metasurfaces,” Nature Reviews Materials 2, 17010 (2017). * [17] N. Xu, R. Singh, and W. Zhang, “Collective coherence in nearest neighbor coupled metamaterials: A metasurface ruler equation,” Journal of Applied Physics 118, 163102 (2015). * [18] G. Sun, L. Yuan, Y. Zhang, X. Zhang, and Y. Zhu, “Q-factor enhancement of Fano resonance in all-dielectric metasurfaces by modulating meta-atom interactions,” Scientific Reports 7, 8128 (2017). * [19] L. Carletti, D. Rocco, A. Locatelli, C. De Angelis, M. Gili, V. F.and Ravaro, I. Favero, G. Leo, M. Finazzi, L. Ghirardini, M. Celebrano, G. Marino, and A. V. Zayats, “Controlling second-harmonic generation at the nanoscale with monolithic AlGaAs-on-AlOx antennas,” Nanotechnology 28, 114005 (2017). * [20] D. Rocco, M. Vincenti, and C. De Angelis, “Boosting second harmonic radiation from AlGaAs nanoantennas with epsilon-near-zero materials,” Applied Science 8, 2212 (2018). * [21] I. Staude, A. E. Miroshnichenko, M. Decker, N. T. Fofang, S. Liu, E. Gonzales, J. Dominguez, T. S. Luk, D. N. Neshev, I. Brener, and Y. Kivshar, “Tailoring directional scattering through magnetic and electric resonances in subwavelength silicon nanodisks,” ACS Nano 7, 7824–7832 (2013). * [22] V. F. Gili, L. Ghirardini, D. Rocco, G. Marino, I. Favero, I. Roland, G. Pellegrini, L. Duò, M. Finazzi, L. Carletti, A. Locatelli, A. Lemaître, D. Neshev, C. De Angelis, G. Leo, and M. Celebrano, “Metal–dielectric hybrid nanoantennas for efficient frequency conversion at the anapole mode,” Beilstein J. Nanotechnology 9, 2306–2314 (2018). * [23] R. Camacho-Morales, M. Rahmani, S. Kruk, L. Wang, L. Xu, D. A. Smirnova, A. S. Solntsev, A. Miroshnichenko, H. H. Tan, F. Karouta, S. Naureen, K. Vora, L. Carletti, C. De Angelis, C. Jagadish, Y. S. Kivshar, and D. N. Neshev, “Nonlinear generation of vector beams from AlGaAs nanoantennas,” Nano Letters 16, 7191–7197 (2016). * [24] G. Marino, C. Gigli, D. Rocco, A. Lemaître, I. Favero, C. De Angelis, and G. Leo, “Zero-order second harmonic generation from AlGaAs-on-insulator metasurfaces,” ACS Photonics 6, 1226–1231 (2019). * [25] M. R. Shcherbakov, P. P. Vabishchevich, A. S. Shorokhov, K. E. Chong, D.-Y. Choi, I. Staude, A. E. Miroshnichenko, D. N. Neshev, A. A. Fedyanin, and Y. S. Kivshar, “Ultrafast all-optical switching with magnetic resonances in nonlinear dielectric nanostructures,” Nano Letters 15, 6985–6990 (2015). * [26] M. R. Shcherbakov, S. Liu, V. V. Zubyuk, A. Vaskin, P. P. Vabishchevich, G. Keeler, T. Pertsch, T. V. Dolgova, I. Staude, I. Brener, and A. A. Fedyanin, “Ultrafast all-optical tuning of direct-gap semiconductor metasurfaces,” Nature Communications 8, 17 (2017). * [27] S. Liu, P. P. Vabishchevich, A. Vaskin, J. L. Reno, G. A. Keeler, M. B. Sinclair, I. Staude, and I. Brener, “An all-dielectric metasurface as a broadband optical frequency mixer,” Nature Communications 9, 2507 (2018). * [28] M. R. Shcherbakov, D. N. Neshev, B. Hopkins, A. S. Shorokhov, I. Staude, E. V. Melik-Gaykazyan, M. Decker, A. A. Ezhov, A. E. Miroshnichenko, I. Brener, A. A. Fedyanin, and Y. S. Kivshar, “Enhanced third-harmonic generation in silicon nanoparticles driven by magnetic response,” Nano Letters 14, 6488–6492 (2014). * [29] L. Ghirardini, G. Marino, V. F. Gili, I. Favero, D. Rocco, L. Carletti, A. Locatelli, C. De Angelis, M. Finazzi, M. Celebrano, D. N. Neshev, and G. Leo, “Shaping the nonlinear emission pattern of a dielectric nanoantenna by integrated holographic gratings,” Nano Letters 18, 6750–6755 (2018). * [30] L. Carletti, G. Marino, L. Ghirardini, V. F. Gili, D. Rocco, I. Favero, A. Locatelli, A. V. Zayats, M. Celebrano, M. Finazzi, G. Leo, C. De Angelis, and D. N. Neshev, “Nonlinear goniometry by second-harmonic generation in AlGaAs nanoantennas,” ACS Photonics 5, 4386–4392 (2018). * [31] I. Staude and J. Schilling, “Metamaterial-inspired silicon nanophotonics,” Nature Photonics 11, 274–284 (2017). * [32] K.-T. Lee, M. Taghinejad, J. Yan, A. S. Kim, L. Raju, D. K. Brown, and W. Cai, “Electrically biased silicon metasurfaces with magnetic Mie resonance for tunable harmonic generation of light,” ACS Photonics 6, 2663–2670 (2019). * [33] L. Wang, S. Kruk, K. Koshelev, I. Kravchenko, B. Luther-Davies, and Y. Kivshar, “Nonlinear wavefront control with all-dielectric metasurfaces,” Nano Letters 18, 3978–3984 (2018). * [34] F. J. F. Löchner, A. N. Fedotova, S. Liu, G. A. Keeler, G. M. Peake, S. Saravi, M. R. Shcherbakov, S. Burger, A. A. Fedyanin, I. Brener, T. Pertsch, F. Setzpfandt, and I. Staude, “Polarization-dependent second harmonic diffraction from resonant GaAs metasurfaces,” ACS Photonics 5, 1786–1793 (2018). * [35] M. A. Green, “Self-consistent optical parameters of intrinsic silicon at 300 K including temperature coefficients,” Solar Energy Materials and Solar Cells 92, 1305–1310 (2008). * [36] D. A. Smirnova, A. B. Khanikaev, L. A. Smirnov, and Y. S. Kivshar, “Multipolar third-harmonic generation driven by optically induced magnetic resonances,” ACS Photonics 3, 1468–1476 (2016). * [37] C. A. Balanis, _Antenna Theory: Analysis and Design_ (John Wiley & Sons, Inc., 2016), 4th ed.
11institutetext: Anadijiban Das 22institutetext: Department of Mathematics, Simon Fraser University, Burnaby, British Columbia, V5A 1S6, Canada 22email<EMAIL_ADDRESS>33institutetext: Rupak Chatterjee 44institutetext: Center for Quantum Science and Engineering Department of Physics, Stevens Institute of Technology, Hoboken, NJ 07030, USA 44email<EMAIL_ADDRESS> # Discrete phase space and continuous time relativistic quantum mechanics II: Peano circles, hyper-tori phase cells, and fibre bundles Anadijiban Das Rupak Chatterjee (Received: date / Accepted: date) ###### Abstract The discrete phase space and continuous time representation of relativistic quantum mechanics is further investigated here as a continuation of paper I DasRCI . The main mathematical construct used here will be that of an area- filling Peano curve. We show that the limit of a sequence of a class of Peano curves is a Peano circle denoted as $\bar{S}^{1}_{n}$, a circle of radius $\sqrt{2n+1}$ where $n\in\\{0,1,\cdots\\}$. We interpret this two-dimensional Peano circle in our framework as a phase cell inside our two-dimensional discrete phase plane. We postulate that a first quantized Planck oscillator, being very light, and small beyond current experimental detection, occupies this phase cell $\bar{S}^{1}_{n}$. The time evolution of this Peano circle sweeps out a two-dimensional vertical cylinder analogous to the world-sheet of string theory. Extending this to three dimensional space, we introduce a $(2+2+2)$-dimensional phase space hyper-tori $\bar{S}^{1}_{n^{1}}\times\bar{S}^{1}_{n^{2}}\times\bar{S}^{1}_{n^{3}}$ as the appropriate phase cell in the physical dimensional discrete phase space. A geometric interpretation of this structure in state space is given in terms of product fibre bundles. We also study free scalar Bosons in the background $[(2+2+2)+1]$-dimensional discrete phase space and continuous time state space using the relativistic partial difference-differential Klein-Gordon equation. The second quantized field quantas of this system can cohabit with the tiny Planck oscillators inside the $\bar{S}^{1}_{n^{1}}\times\bar{S}^{1}_{n^{2}}\times\bar{S}^{1}_{n^{3}}$ phase cells for eternity. Finally, a generalized free second quantized Klein-Gordon equation in a higher $[(2+2+2)N+1]$-dimensional discrete state space is explored. The resulting discrete phase space dimension is compared to the significant spatial dimensions of some of the popular models of string theory. ###### Keywords: Discrete phase space Peano curves Partial difference-differential equations Fibre bundles ###### pacs: 11.10Ef 11.15Ha 02.70Bf 03.65Fd ## 1 Introduction We begin in section 2 by introducing the concept of a Peano curve Clark ; Gelbaum . In 1890, G. Peano startled the mathematical world with the introduction of an area-filling curve. After that, many area-filling curves have been discovered including those by David Hilbert. To our knowledge, very few papers in the mathematical physics arena have used this concept DasHaldar . We begin by discussing the original Peano curve that fills a square $\bar{D}$ of unit area inside $\mathbb{R}^{2}$. In section 3, we derive other Peano curves filling a sequence $\\{\bar{D}_{M}\\}^{\infty}_{M=1}$ of unit areas each in the shape of a rectangle. Next, we introduce a double sequence $\\{\bar{D}_{Mn}\\}^{\infty}_{M=1}$ for $n\in\\{0,1,2,...\\}$. Each of closed regions $\bar{D}_{Mn}$ for a fixed $n$ is endowed with a unit area and a rectangular shape. Moreover, each of these regions $\bar{D}_{Mn}$ is covered by a Peano curve. Next, we discuss another double sequence $\\{\bar{A}_{Mn}\\}$ of closed regions where each region $\bar{A}_{Mn}$ is annular in shape and possesses a unit area. This annular region $\bar{A}_{Mn}$ is also covered by a Peano curve. Finally, in the limiting process of $M\rightarrow\infty$ for a fixed $n$, the annular region $\bar{A}_{Mn}$ collapses into a circle $\bar{S}^{1}_{n}$ of unit area. Since this cirlce contains a ’squashed’ Peano curve, one calls $\bar{S}^{1}_{n}$ a Peano circle. Section 4 details more analysis on this Peano circle. The simplified Klein-Gordon equation in the background of a $[(1+1)+1]$-dimensional discrete phase space and continuous time is discussed in section 5. Some physically relevant results are also derived. In section 6, we introduce the abstract concept of fibre bundles Choquet and discuss their applications to certain physical problems. We apply this concept in section 7 to the physically important $[(2+2+2)+1]$-dimensional phase space. This discrete phase space and continuous time arena has been investigated with respect to the second quantization of many free relativistic field theories DasI ; DasII . The corresponding formulations of interacting relativistic second quantized fields yielding an $S^{\\#}$-matrix series were given in DasIII . Second order expansion terms of this $S^{\\#}$-matrix involving Møller scattering of quantum electrodynamics Jauch ; Peskin produced a non-singular Coulomb potential in DasBen ; DasRC . In section 7, we define a new set-theoretic mapping Goldberg by the direct or Cartesian product Lightstone as: $\begin{array}[]{c}\bar{S}^{1}_{n^{1}}\times\bar{S}^{1}_{n^{2}}\times\bar{S}^{1}_{n^{3}}=\displaystyle{\lim_{M\to\infty}}\\\ \\\ \left([g^{M}_{n^{1}}\times h^{M}_{n^{1}}\times h^{M}_{1}]\times[g^{M}_{n^{2}}\times h^{M}_{n^{2}}\times h^{M}_{2}]\times[g^{M}_{n^{3}}\times h^{M}_{n^{3}}\times h^{M}_{3}]\right)\left(\bar{D}_{1}\times\bar{D}_{2}\times\bar{D}_{3}\right)\\\ \\\ \subset\mathbb{R}^{2}\times\mathbb{R}^{2}\times\mathbb{R}^{2}.\end{array}$ (1) The left hand side of this equation ($\bar{S}^{1}_{n^{1}}\times\bar{S}^{1}_{n^{2}}\times\bar{S}^{1}_{n^{3}}$) is a hyper-torus Massey of physical dimension $\left[\dfrac{ML^{2}}{T}\right]^{3}$. It will be shown that these hyper-tori can represent a phase cell in the usual $(2+2+2)$-dimensional discrete phase space. In section 8, the second quantization of the scalar field $\Phi(n^{1},n^{2},n^{3};t)=:\Phi(\mathbf{n};t)$ is discussed in the arena of $[(2+2+2)+1]$-dimensional discrete phase space and continuous time using the relativistic Klein-Gordon operator equation. A partial difference-differential version of the Klein-Gordon equation is solved for a special class of solutions involving Fourier-Hermite transforms DasII . The total energy $\mathcal{H}$, momentum component $\mathcal{P}_{j}$ and electric charge $\mathcal{Q}$ are worked out for the second quantized Klein-Gordon scalar field operator $\Phi(\mathbf{n};t)$. Finally, in section 9, we study discrete $(2+2+2)N$-dimensional phase space many-particle systems and the corresponding Peano hyper-tori $\bar{S}^{1}_{n^{1}}\times\bar{S}^{1}_{n^{2}}\times\cdots\bar{S}^{1}_{n^{N}}$. Each of these Peano hyper-tori has the physical dimension of $\left[\dfrac{ML^{2}}{T}\right]^{N}$ and represent appropriate phase cells inside the discrete $(2+2+2)N$-dimensional phase space. The generalized Klein- Gordon equation within this discrete $[(2+2+2)N+1]$-dimensional phase space and continuous time arena is also discussed. As our $2N$-dimensional Peano hyper-tori is a phase-cell inside a discrete $6N$-dimensional phase space, we compare our structure with that of the spatial dimension of string theory Green ; PolchinskiI ; PolchinskiII . ## 2 The area filling curve of Peano Consider a continuous, piece-wise linear, oriented, parametrized curve Clark $f^{1}(u)$ inside a $(1+1)$-dimensional plane $\mathbb{R}^{2}$ as depicted in figure 1. Figure 1: The graph of the parametrized curve $f^{1}(u)$. Note that the curve $f^{1}(u)$ is defined over nine closed intervals $\left[\dfrac{k-1}{9},\dfrac{k}{9}\right]$ for $k\in\\{1,2,\cdots,9\\}\subset\mathbb{R}$. The image of the curve of figure 1 is a continuous, piece-wise zigzag, oriented, parametrized curve over a closed square $\bar{D}\subset\mathbb{R}$ of unit area inside the $x-y$ plane $\mathbb{R}^{2}$. The image of the function $f^{1}(u)$ in figure 1 has for the closed domain$[0,1/9]$ the closed range marked as the first linear segment with an arrow and denoted with a ”1”. Similarly, the closed domain $[1/9,2/9]$ is mapped to the linear segment denoted ”2”. The ten corners of the zig-zag curve are situated at $f^{1}(k/9),k=0,1,...,9$. Figure 2: The graph of the parametrized curve $f^{2}(u)$. The image of the next parametrized curve $f^{2}(u)$, depicted in figure 2, is continuous, oriented, and piece-wise linear in the $x-y$ plane $\mathbb{R}^{2}$. It has $9^{2}=81$ linear segments inside the unit square $\bar{D}$ within $\mathbb{R}^{2}$. This figure is constructed by replicating figure 1 nine times within the unit square area resulting in $81$ oriented linear pieces. Let us define a distance function $d(f^{1}(u),f^{2}(u))$ abstractly as $d(f^{1}(u),f^{2}(u)):=\displaystyle{\textit{sup}_{u\in[0,1]}}||f^{1}(u)-f^{2}(u)||\leq\dfrac{\sqrt{2}}{3}.$ (2) Here, the number $\dfrac{\sqrt{2}}{3}$ is the diameter of a square with each side being $\dfrac{1}{3}$. The sequence of pre-Peano parametrized curves $\\{f^{j}(u)\\}^{\infty}_{j=1}$ is obtained by replicating the above process over and over again. Therefore, one may obtain the result that $d(f^{j}(u),f^{j+1}(u))\leq\dfrac{\sqrt{2}}{3^{j}}.$ (3) It follows by the uniform Cauchy criterion Clark that for each $\epsilon>0$, there exists an integer $N>0$ such that $d(f^{j}(u),f^{m}(u))<\epsilon$ for all $j$ and $m\geq N$. Therefore, there exists a uniformly continuous function $f(u)$ such that $\displaystyle{\lim_{j\rightarrow\infty}}f^{j}(u)=f(u)$. Moreover, on can prove that this parametrized curve $f(u)$ passes through every point of the closed unit square $\bar{D}$. This is the first example of a Peano curve filling in the area of $\bar{D}$ Clark . ## 3 Other Peano curves filling a $(1+1)$-dimensional physical phase space We can interpret physically the areas $\bar{D}$ in both figures 1 and 2 as phase cells inside a $(1+1)$-dimensional phase plane. Furthermore, the uncertainty of quantum physics Dirac $|\Delta x||\Delta y|\geq 1$ $(i.e.\,|\Delta q||\Delta p|\geq 1,\hbar=1)$ automatically holds inside a phase cell $\bar{D}$ for a quantized particle occupying $\bar{D}$ (here, $\Delta$ stands for an increment and not a finite difference operator). Any possible movement of the quantized particle along the Peano curve $f(u)$ covering $\bar{D}$ is physically unobservable. Now, we shall define an infinite sequence of functions $\\{h^{M}\\}^{\infty}_{M=1}$ from the region $\bar{D}\subset\mathbb{R}^{2}$ into regions $\\{\bar{D}_{M}\\}^{\infty}_{M=1}$ as depicted in figure 3. Figure 3: The graph of the function $h^{M}$ for fixed $M$. The function $h^{M}$ is defined for a specific $M$ as the following linear transformation: $\begin{array}[]{c}\rho=\left(\dfrac{1}{2M\pi}\right)x+\dfrac{1}{2},\\\ \\\ \alpha=(2M\pi)y-M\pi,\\\ \\\ M\geq 1.\end{array}$ (4) The Jacobian of these transformations is $\dfrac{\partial(\rho,\alpha)}{\partial(x,y)}=1,$ (5) making these transformations canonical transformations of Hamiltonian mechanics Lanczos ; Goldstein . Therefore, the area of each of the regions $\\{\bar{D}_{M}\\}$ is furnished by the double integral of the following differential 2-form Spivak ; DasTA $\begin{array}[]{c}Area(\bar{D}^{M}):=\displaystyle{{\int\int}_{\bar{D}^{M}}}d\rho\wedge d\alpha=\int_{\frac{1}{2}}^{\frac{1}{2M\pi}+\frac{1}{2}}\int_{-M\pi}^{M\pi}d\rho d\alpha=1,\\\ \\\ M\in\\{1,2,3,...\\}.\end{array}$ (6) Remarks: (i) In the limiting process $M\rightarrow\infty$, the sequence of regions $\\{\bar{D}^{M}\\}^{\infty}_{M=1}$ collapses into the open vertical line given by $\\{(\rho,\alpha)\in\mathbb{R}^{2}:\rho=1/2,\alpha\in\mathbb{R}\\}$. (ii) This infinite vertical line is an open string-like phase cell of unit area as seen by using (6) above. (iii) The infinite vertical line in remark (i) is consistent with the uncertainty principle $|\Delta\rho||\Delta\alpha|\geq 1$. (iv) The string-like phase cell may or may not contain any quanta of a second quantized relativistic wave field (see section 5). (v) Moreover, unlike in conventional string theory Green ; PolchinskiI ; PolchinskiII , in our approach of string-like phase cells, an open string-like phase cell cannot be of finite length. Consider another sequence of transformations $\\{h^{M}_{n}\\}^{\infty}_{M=1}$ where $n$ is chosen to be a fixed non-negative integer. A typical mapping $h^{M}_{n}$ maps the closed domain $\bar{D}_{M}$ into the closed region $\bar{D}_{Mn}$ by the linear transformation $\begin{array}[]{c}r=\rho+n,\\\ \\\ \theta=\alpha,\\\ \\\ \dfrac{\partial(r,\theta)}{\partial(\rho,\alpha)}=1.\end{array}$ (7) This transformation is exhibited in figure 4. Figure 4: The mapping $h^{M}_{n}$ from $\bar{D}_{M}$ to $\bar{D}_{Mn}$. The area of each domain $\\{\bar{D}_{Mn}\\}$ is given by $\begin{array}[]{c}Area(\bar{D}_{Mn}):=\displaystyle{{\int\int}_{\bar{D}_{Mn}}}dr\wedge d\theta=\int_{n+\frac{1}{2}}^{n+\frac{1}{2M\pi}+\frac{1}{2}}\int_{-M\pi}^{M\pi}drd\theta=1,\\\ \\\ M\in\\{1,2,3,...\\},\,\,\,\ n\in\\{0,1,2,....\\}.\end{array}$ (8) Now, consider a third sequence of canonical transformations $\\{g^{M}_{n}\\}^{\infty}_{M=1}$ for a fixed $n$ as follows, $\begin{array}[]{c}q=\sqrt{2\rho}\cos\theta,\\\ \\\ p=\sqrt{2\rho}\sin\theta,\\\ \\\ \dfrac{\partial(q,p)}{\partial(\rho,\theta)}=1,\\\ \\\ q^{2}+p^{2}=2r>0,\\\ \\\ \left(\dfrac{p}{q}\right)=\tan\theta,\,\,\,q\neq 0.\end{array}$ (9) The closed domain of each of the mappings of $\\{g^{M}_{n}\\}^{\infty}_{M=1}$ is furnished by $\bar{D}_{Mn}:=\left\\{(r,\theta)\in\mathbb{R}^{2}:\left(n+\frac{1}{2}\right)\leq r\leq\left(n+\frac{1}{2M\pi}+\frac{1}{2}\right),-M\pi\leq\theta\leq M\pi\right\\}.$ (10) The corresponding co-domain $\bar{A}_{Mn}$ in the $q-p$ phase plane is provided by the annular region: $\begin{array}[]{c}\bar{A}_{Mn}:=\left\\{(q,p)\in\mathbb{R}^{2}:2n+1\leq q^{2}+p^{2}\leq\left(2n+1+\frac{1}{M\pi}\right)\right\\},\\\ \\\ Area(\bar{A}_{Mn}):=\displaystyle{\int\int_{\bar{A}_{Mn}}}dp\wedge dq=1.\end{array}$ (11) The mappings $\\{g^{M}_{n}\\}^{\infty}_{M=1}$ are depicted in the figure 5. Figure 5: The mapping $g^{M}_{n}$ with topological identifications of the two horizontal sides of $\bar{D}_{Mn}$. Note that the annular region $\bar{A}_{Mn}$ is not a simply connected region of the circular disk $q^{2}+p^{2}=2n+1+\frac{1}{M\pi}$. Furthermore, the winding number of each boundary of the annular region $\bar{A}_{Mn}$ is exactly $M$. ## 4 The Peano circle $\bar{S}^{1}_{n}$ in the $(1+1)$-dimensional phase plane We shall now investigate the closed domain $\bar{A}_{Mn}$ of the mapping $g^{M}_{n}$ for a fixed $n$ as exhibited in figure 5. Using (9) and (11), the inner and outer boundaries of the closed co-domain $\bar{A}_{Mn}$ are found to be $\begin{array}[]{c}\partial_{-}[\bar{A}_{Mn}]:=\\{(q,p)\in\mathbb{R}:q^{2}+p^{2}=2n+1\\},\\\ \\\ \partial_{+}[\bar{A}_{Mn}]:=\left\\{(q,p)\in\mathbb{R}:q^{2}+p^{2}=2n+1+\frac{1}{M\pi}\right\\}.\end{array}$ (12) Consider the limiting map $\displaystyle{\lim_{M\rightarrow\infty}}\\{g^{M}_{n}\\}_{M=1}^{\infty}$. Clearly, one has for this limiting map $\displaystyle{\lim_{M\rightarrow\infty}}\left(\left\\{(q,p)\in\mathbb{R}:q^{2}+p^{2}=2n+1+\frac{1}{M\pi}\right\\}\right)=\partial_{-}[\bar{A}_{Mn}]$ (13) which is depicted in figure 6. Figure 6: The limiting map $\displaystyle{\lim_{M\rightarrow\infty}}\\{g^{M}_{n}\\}$ and the closed range as the Peano circle $\bar{S}^{1}_{n}$. Remarks: (i) Note that by (11), $Area(\bar{A}_{Mn})=1$ for all $M\in\\{1,2,3,...\\}$. (ii) Since the closed range of the mapping $\displaystyle{\lim_{M\rightarrow\infty}}\\{g^{M}_{n}\\}=\bar{S}^{1}_{n}$, the Peano circle $\bar{S}^{1}_{n}$ itself is endowed with the unit $(1+1)$-dimensional area. (iii) The Peano circle can act as a string-like Green ; PolchinskiI ; PolchinskiII phase-cell. (iv) The quantum mechanical uncertainty principle $|\Delta q||\Delta p|\geq 1$ holds for each quanta inside the phase cell $\bar{S}^{1}_{n}$. (v) The original Peano-curve $\displaystyle{\lim_{j\rightarrow\infty}}f^{j}(u)=f(u)$ is squashed inside the Peano circle $\bar{S}^{1}_{n}$. (vi) The first quantized oscillator introduced in part I of this work DasRCI follows a zig-zag motion or zitter-bewengung Greiner along the squashed Peano curve inside $\bar{S}^{1}_{n}$. (vii) Unlike the Peano circle $\bar{S}^{1}_{n}$ here, the one-dimensional circle of figure 2 in part I DasRCI cannot act as a two-dimensional phase cell within the $q-p$ phase plane. Consider now the infinite $[(1+1)+1]$-dimensional hyper-circular cylinder $\bar{S}^{1}_{n}\times\mathbb{R}$ inside the $[(1+1)+1]$-dimensional discrete phase plane plus continuous time $t\in\mathbb{R}$ called the state space Lanczos . Within this state space is a world-sheet like object Green ; PolchinskiI ; PolchinskiII depicted in figure 7. Figure 7: The $[(1+1)+1]$-dimensional infinite hyper-cylinder $\bar{S}^{1}_{n}\times\mathbb{R}$ inside the $[(1+1)+1]$-dimensional discrete phase plane plus continuous time. The $[(1+1)+1]$-dimensional circular cylinder $S^{1}_{n}\times\mathbb{R}$ inside the state space in figure 4 in part I DasRCI cannot represent an evolving phase cell unlike the hyper-circular cylinder $\bar{S}^{1}_{n}\times\mathbb{R}$ depicted in figure 7. Let us summarize what we have achieved so far. Figures 1 and 2 have described the area filling curve of Peano, which completely fills the area of a unit square. We identify this unit area as a possible phase cell within a $(p,q)$-phase plane. The area filling Peano curve is identified as the trajectory of one quanta of certain quantized simple harmonic oscillators introduced in DasRCI . The trajectory of the Peano curve is not observable due to the uncertainty principle of quantum mechanics Greiner ; Bethe . Equations (4), (5), (7), and (9) indicate sequences of area preserving mathematical mappings physically representing canonical transformations of Hamiltonian mechanics Lanczos ; Goldstein . Therefore, in figure 5, the original unit square has yielded a sequence of circular, annular domains each of unit area. In figure 6, this sequence of circular annular domains collapses into a Peano circle $\bar{S}^{1}_{n}$ of unit area. The quanta of the Planck oscillator DasRCI still goes around the Peano circle with constant energy $\sqrt{2n+1}$ in a zig-zag pattern but is undetectable by external observations. Note that a Peano circle $\bar{S}^{1}_{n}$ of unit area in the $(1+1)$-phase plane is analogous to a closed string of prototypical string theory Green ; PolchinskiI ; PolchinskiII . Figure 7 depicts the $[(1+1)+1]$-dimensional hyper-circular-cylinder inside the three-dimensional state space that is analogous to the two-dimensional world-sheet evolution of closed strings. ## 5 The Klein-Gordon equation in a $[(1+1)+1]$-dimensional discrete phase- plane and continuous time Now, we investigate the second quantization of the Klein-Gordon equation in a $[(1+1)+1]$-dimensional discrete phase-plane and continuous time scenario. The relativistic Klein-Gordon equation in a $[(2+2+2)+1]$-dimensional discrete phase-plane and continuous time was discussed in section 7 of part I of this paper DasRCI . The second quantized scalar field linear operator, denoted by $\Phi(n,t)$, acts on a Hilbert space bundle Choquet ; DasTA . The corresponding Klein-Gordon equation is $\begin{array}[]{c}[\mathbf{P}\cdot\mathbf{P}-(\mathbf{P_{t}})^{2}+m^{2}\mathbf{I}]\overrightarrow{\mathbf{\Psi}}=\overrightarrow{\mathbf{0}},\\\ or\\\ (\Delta^{\\#})^{2}\Phi(n,t)-(\partial_{t})^{2}\Phi(n,t)-m^{2}\Phi(n,t)=0.\end{array}$ (14) Consider the complex valued functions $\xi_{n}(k)$ involving Hermite polynomials $H_{n}(k)$ Olver , $\begin{array}[]{c}\xi_{n}(k):=\dfrac{(i)^{n}e^{-k^{2}/2}H_{n}(k)}{\pi^{1/4}2^{n/2}\sqrt{n!}},\\\ \\\ \xi_{n}(-k)=\overline{\xi_{n}(k)},\\\ \xi_{2n+1}(0)=0,\\\ \displaystyle{\int_{\mathbb{R}}}\overline{\xi_{m}(k)}\xi_{n}(k)dk=\delta_{mn},\\\ \displaystyle{\sum_{n=0}^{\infty}}\overline{\xi_{n}(k)}\xi_{n}(\hat{k})=\delta(k-\hat{k}),\\\ \Delta^{\\#}\xi_{n}(k)=ik\xi_{n}(k).\end{array}$ (15) Here, $\delta(k-\hat{k})$ indicates the Dirac delta distribution function Zemanian . A special class of exact solutions to the partial difference- differential equations of (14) are given by $\begin{array}[]{c}\Phi^{-}(n,t):=\displaystyle{\int_{\mathbb{R}}}\dfrac{1}{\sqrt{2\omega(k)}}\left[A(k)\xi_{n}(k)e^{-i\omega t}\right]dk,\\\ \\\ \Phi^{+}(n,t):=\displaystyle{\int_{\mathbb{R}}}\dfrac{1}{\sqrt{2\omega(k)}}\left[B^{\dagger}(k)\overline{\xi_{n}(k)}e^{i\omega t}\right]dk,\\\ \\\ \Phi(n,t):=\Phi^{-}(n,t)+\Phi^{+}(n,t),\\\ \\\ \omega=\omega(k):=+\sqrt{k^{2}+m^{2}}>0.\end{array}$ (16) The Fourier-Hermite integrals in (16) are supposed to be uniformly convergent Buck . Moreover, the linear operators $A(k),A^{\dagger}(k),B(k)$ and $B^{\dagger}(k)$ (creation and annihilation operators in momentum space) acting linearly on a Hilbert space bundle must satisfy the following commutation relations: $\begin{array}[]{c}[A(k),A^{\dagger}(\hat{k})]=[B(k),B^{\dagger}(\hat{k})]=\delta(k-\hat{k})I(k),\\\ \\\ \,[A(k),A(\hat{k})]=[A^{\dagger}(k),A^{\dagger}(\hat{k})]=[B(k),B(\hat{k})]=[B^{\dagger}(k),B^{\dagger}(\hat{k})]=0,\\\ \\\ N^{+}(k):=A^{\dagger}(k)A(k),\\\ \\\ N^{-}(k):=B^{\dagger}(k)B(k),\end{array}$ (17) where the eigenvalues of the number operators $N^{\pm}(k)$ are the set $\\{0,1,2,...\\}$. Physically speaking, the scalar field operator $\Phi(n,t)$ represents a collection of massive, spin-less, electrically charged, second quantized scalar particle excitations. One can show the following relations for total energy, total momentum and total electric charge respectively DasI ; DasII , $\begin{array}[]{c}\mathcal{H}:=\displaystyle{\int_{\mathbb{R}}}[N^{+}(k)+N^{-}(k)+\delta(0)I(k)]\omega(k)dk,\\\ \\\ \mathcal{P}:=\displaystyle{\int_{\mathbb{R}}}[N^{+}(k)+N^{-}(k)]kdk,\\\ \\\ \mathcal{Q}:=e\displaystyle{\int_{\mathbb{R}}}[N^{+}(k)-N^{-}(k)]dk.\end{array}$ (18) The divergent null point energy term $\delta(0)I(k)$ may be ignored for physical interpretations but cannot be rectified directly. ## 6 Fibre Bundles Let $M$ and $M^{\\#}$ be two topological manifolds Choquet ; DasTA . The Cartesian product $M\times M^{\\#}$ and the projection mapping $\Pi$ from $M\times M^{\\#}$ into $M$ constitute a product or trivial bundle $(M\times M^{\\#},M,\Pi)$ depicted in figure 8. The vertical linear segment inside $M\times M^{\\#}$ is called a fibre Choquet ; DasTA . Figure 8: The trivial or product bundle $(M\times M^{\\#},M,\Pi)$. Figure 9 provides an explicit example of a product bundle using our Peano cirle $\bar{S}^{1}_{n}$ to create a vertical circular cylinder $\bar{S}^{1}_{n}\times I_{t}$, Figure 9: The product bundle $(\bar{S}^{1}_{n}\times I_{t},\bar{S}^{1}_{n},\Pi_{t})$. The closed interval $I_{t}:=[0,T]\subset\mathbb{R}$ and $(\chi,[-\pi,\pi])$ is a coordinate chart DasTA in figure 9. The second quantized scalar field operators defined in (16) are restricted for fixed numbers $n$ and $T$ of the product bundle in figure 9. Another product bundle associated with the linear operators (16) is created by using our Peano circle $\bar{S}^{1}_{n}$ and the closed linear line segment $I_{k}:=[K_{1},K_{2}]\subset\mathbb{R}$ as in figure 10. Figure 10: The product bundle $(\bar{S}^{1}_{n}\times I_{k},\bar{S}^{1}_{n},\Pi_{k})$. Remarks: (i) Figure 9 represents geometrically the occupation of a tiny first quantized Planck oscillator of figures 6 and 7 in the interval $[0,T]\subset\mathbb{R}$ (the full actual physics takes place in the time interval $-\infty<t<\infty$). (ii) Moreover, in figure 9, the second quantized scalar field $\Phi(n,t)$ quanta can cohabit with a tiny first quantized Planck oscillator during the time interval $I_{t}:=[0,T]\subset\mathbb{R}$. (iii) Furthermore, in figure 10, the second quantized scalar field quanta with a range of momentum $K_{1}<k<K_{2}$ can cohabit with a tiny first quantized Planck oscillator. (iv) The complete mathematical treatment of this phenomena involves the product bundles $(\bar{S}^{1}_{n}\times I_{t},\bar{S}^{1}_{n},\Pi_{t})\times(\bar{S}^{1}_{n}\times I_{k},\bar{S}^{1}_{n},\Pi_{k})$ (which is difficult to depict). ## 7 Discrete phase space and hyper-tori like phase cells Here, we shall extend the various mappings of figures 3,4, and 5 to the $(2+2+2)$-dimensional discrete phase space arena. Choosing the fixed indices $a\in\\{1,2,3\\}$ and $M\in\\{1,2,3,\cdots\\}$, the relevant mappings are illustrated in figure 11. It depicts the composite map $g^{M}_{n^{a}}\circ h^{M}_{n^{a}}\circ h^{M}_{a}$ for fixed indices $a$ and $M$. Figure 11: The discrete phase space mappings $h^{M}_{a},h^{M}_{n^{a}}$, and $g^{M}_{n^{a}}$. The closed domains and the closed range of the composite map $g^{M}_{n^{a}}\circ h^{M}_{n^{a}}\circ h^{M}_{a}$ are provided by $\begin{array}[]{c}closed\,\,\ domain\,\left[g^{M}_{n^{a}}\circ h^{M}_{n^{a}}\circ h^{M}_{a}\right]=\bar{D}_{a}\subset\mathbb{R}^{2},\\\ \\\ closed\,\,\ range\,\left[g^{M}_{n^{a}}\circ h^{M}_{n^{a}}\circ h^{M}_{a}\right]=\bar{A}_{Mn^{a}}\subset\mathbb{R}^{2},\\\ or\\\ \bar{A}_{Mn^{a}}=\left[g^{M}_{n^{a}}\circ h^{M}_{n^{a}}\circ h^{M}_{a}\right](\bar{D}_{a}).\end{array}$ (19) where the last relation illustrates a set theoretic mapping Goldberg . Consider the sequence of composite mappings $\left\\{g^{M}_{n^{a}}\circ h^{M}_{n^{a}}\circ h^{M}_{a}\right\\}_{M=1}^{\infty}$ for a fixed $a$ and $n^{a}$. The limiting set-theoretic mapping from figure 6 and the last relation of (19) is given by $\bar{S}^{1}_{n}=\displaystyle{\lim_{M\rightarrow\infty}}\left\\{\left[g^{M}_{n^{a}}\circ h^{M}_{n^{a}}\circ h^{M}_{a}\right](\bar{D}_{a})\right\\}$ (20) From this, one can derive the set-theoretic Cartesian product mapping Goldberg ; Lightstone : $\begin{array}[]{c}\bar{S}^{1}_{n^{1}}\times\bar{S}^{1}_{n^{2}}\times\bar{S}^{1}_{n^{3}}=\displaystyle{\lim_{M\rightarrow\infty}}\left(\left\\{\bar{A}_{Mn^{1}}\right\\}\times\left\\{\bar{A}_{Mn^{2}}\right\\}\times\left\\{\bar{A}_{Mn^{3}}\right\\}\right),\\\ or\\\ \bar{S}^{1}_{n^{1}}\times\bar{S}^{1}_{n^{2}}\times\bar{S}^{1}_{n^{3}}=\displaystyle{\lim_{M\rightarrow\infty}}\left(\left\\{\left[g^{M}_{n^{1}}\circ h^{M}_{n^{1}}\circ h^{M}_{1}\right](\bar{D}_{1})\right\\}\times\left\\{\left[g^{M}_{n^{2}}\circ h^{M}_{n^{2}}\circ h^{M}_{2}\right](\bar{D}_{2})\right\\}\right.\\\ \left.\times\left\\{\left[g^{M}_{n^{3}}\circ h^{M}_{n^{3}}\circ h^{M}_{3}\right](\bar{D}_{3})\right\\}\right).\end{array}$ (21) Furthermore, one has $\begin{array}[]{c}\left[g^{M}_{n^{1}}\times h^{M}_{n^{1}}\times h^{M}_{1}\right]\times\left[g^{M}_{n^{2}}\times h^{M}_{n^{2}}\times h^{M}_{2}\right]\times\left[g^{M}_{n^{3}}\times h^{M}_{n^{3}}\times h^{M}_{3}\right]\left(\bar{D}_{1}\times\bar{D}_{2}\times\bar{D}_{3}\right):=\\\ \\\ \left\\{\left[g^{M}_{n^{1}}\circ h^{M}_{n^{1}}\circ h^{M}_{1}\right](\bar{D}_{1})\right\\}\times\left\\{\left[g^{M}_{n^{2}}\circ h^{M}_{n^{2}}\circ h^{M}_{2}\right](\bar{D}_{2})\right\\}\times\left\\{\left[g^{M}_{n^{3}}\circ h^{M}_{n^{3}}\circ h^{M}_{3}\right](\bar{D}_{3})\right\\}\end{array}$ (22) From the above two relations, a new set-theoretic mapping is furnished by $\begin{array}[]{c}\bar{S}^{1}_{n^{1}}\times\bar{S}^{1}_{n^{2}}\times\bar{S}^{1}_{n^{3}}=\displaystyle{\lim_{M\rightarrow\infty}}\left\\{\left[g^{M}_{n^{1}}\times h^{M}_{n^{1}}\times h^{M}_{1}\right]\times\left[g^{M}_{n^{2}}\times h^{M}_{n^{2}}\times h^{M}_{2}\right]\times\right.\\\ \\\ \left.\left[g^{M}_{n^{3}}\times h^{M}_{n^{3}}\times h^{M}_{3}\right]\left(\bar{D}_{1}\times\bar{D}_{2}\times\bar{D}_{3}\right)\right\\}\subset\mathbb{R}^{2}\times\mathbb{R}^{2}\times\mathbb{R}^{2}.\end{array}$ (23) Remarks: (i) The $(2+2+2)$-dimensional hyper-sphere or hyper-torus Massey $\bar{S}^{1}_{n^{1}}\times\bar{S}^{1}_{n^{2}}\times\bar{S}^{1}_{n^{3}}$ for three fixed integers $(n^{1},n^{2},n^{3})$ denotes a region where a tiny first quantized Planck oscillator inhabits this region for all time. (ii) The whole of the $(2+2+2)$-dimensional discrete phase space contains a denumerably infinite number of concentric hyper-tori each containing a single tiny first quantized Planck oscillator. (iii) Each $\bar{S}^{1}_{n^{1}}\times\bar{S}^{1}_{n^{2}}\times\bar{S}^{1}_{n^{3}}$ constitutes a phase cell of physical dimension $\left[\dfrac{ML^{2}}{T}\right]^{3}$. Therefore, in the fundamental physical units, each of these phase cells is endowed with a hyper-volume of $1\left[\dfrac{ML^{2}}{T}\right]^{3}$. (iv) Each of these phase-cells bears a resemblance to a $D$-dimensional hyper- torus in standard string theory Green ; PolchinskiI ; PolchinskiII . (v) In the arena of $[(2+2+2)+1]$-dimensional discrete phase space and continuous time, the hyper-torus $\bar{S}^{1}_{n^{1}}\times\bar{S}^{1}_{n^{2}}\times\bar{S}^{1}_{n^{3}}$ sweeps out a world-sheet like vertical, circular hyper-cylinder. Such a vertical, circular cylinder always contains just one first quantized tiny Planck oscillator for all time. (vi) Finally, one or more second quantized scalar field $\Phi(\mathbf{n};t)$ quanta may cohabit the tiny first quantized Planck oscillator temporarily or forever. ## 8 Second quantized, relativistic Klein-Gordon field equations in $[(2+2+2)+1]$-dimensional discrete phase space and continuous time From part I of this work DasRCI , the second quantized, relativistic Klein- Gordon field equations are given by $\begin{array}[]{c}a,b\in\\{1,2,3\\},\\\ \\\ n^{a}\in\\{0,1,2,...\\},\\\ \\\ \mathbf{n}:=(n^{1},n^{2},n^{3}),\\\ \\\ \left[(\delta^{ab}\Delta^{\\#}_{a}\Delta^{\\#}_{b})-(\partial_{t})^{2}-m^{2}\right]\Phi(\mathbf{n};t)=\mathbf{0}.\end{array}$ (24) The three dimensional extensions of the functions $\xi_{n}(k)$ in (15) imply that Olver $\begin{array}[]{c}\xi_{n^{a}}(k_{a}):=\dfrac{(i)^{n^{a}}e^{-k_{a}^{2}/2}H_{n^{a}}(k_{a})}{\pi^{1/4}2^{n^{a}/2}\sqrt{n^{a}!}},\\\ \\\ \mathbf{k}:=(k_{1},k_{2},k_{3}),\\\ \\\ \Delta^{\\#}_{a}\xi_{n^{a}}(k_{a})=ik_{a}\xi_{n^{a}}(k_{a}),\\\ \\\ \displaystyle{\sum_{n^{1}=0}^{\infty}\sum_{n^{2}=0}^{\infty}\sum_{n^{3}=0}^{\infty}}\left[\overline{\xi_{n^{1}}}(k_{1})\xi_{n^{1}}(\hat{k}_{1})\right]\left[\overline{\xi_{n^{2}}}(k_{2})\xi_{n^{2}}(\hat{k}_{2})\right]\left[\overline{\xi_{n^{3}}}(k_{3})\xi_{n^{3}}(\hat{k}_{3})\right]\\\ =\delta({k_{1}}-{\hat{k}_{1}})\delta({k_{2}}-{\hat{k}_{2}})\delta({k_{3}}-{\hat{k}_{3}})=:\delta^{3}(\mathbf{k}-\mathbf{\hat{k}}).\end{array}$ (25) A special class of exact solutions for these relativistic partial difference- differential operator equations (24) are given by $\begin{array}[]{c}\Phi^{-}(\mathbf{n};t):=\displaystyle{\int_{\mathbb{R}^{3}}}\dfrac{1}{\sqrt{2\omega(\mathbf{k})}}\left[A(\mathbf{k})\xi_{n^{1}}(k_{1})\xi_{n^{2}}(k_{2})\xi_{n^{3}}(k_{3})e^{-i\omega t}\right]dk_{1}dk_{2}dk_{3},\\\ \\\ \Phi^{+}(\mathbf{n};t):=\displaystyle{\int_{\mathbb{R}^{3}}}\dfrac{1}{\sqrt{2\omega(\mathbf{k})}}\left[B^{\dagger}(\mathbf{k})\overline{\xi_{n^{1}}}(k_{1})\overline{\xi_{n^{2}}}(k_{2})\overline{\xi_{n^{3}}}(k_{3})e^{i\omega t}\right]dk_{1}dk_{2}dk_{3},\\\ \\\ \Phi(\mathbf{n};t):=\Phi^{-}(\mathbf{n};t)+\Phi^{+}(\mathbf{n};t),\\\ \\\ \omega=\omega(\mathbf{k}):=+\sqrt{k_{1}^{2}+k_{3}^{2}+k_{3}^{2}+m^{2}}>0.\end{array}$ (26) The Fourier-Hermite integrals in (26) are supposed to be uniformly convergent Buck . Moreover, the linear operators $A(\mathbf{k}),A^{\dagger}(\mathbf{k}),B(\mathbf{k})$ and $B^{\dagger}(\mathbf{k})$ (creation and annihilation operators in momentum space) acting linearly on a Hilbert space bundle Choquet ; DasTA must satisfy the following commutation relations: $\begin{array}[]{c}[A(\mathbf{k}),A^{\dagger}(\hat{\mathbf{k}})]=[B(\mathbf{k}),B^{\dagger}(\hat{\mathbf{k}})]=\delta(\mathbf{k}-\hat{\mathbf{k}})\mathbf{I}(\mathbf{k}),\\\ \\\ \,[A(\mathbf{k}),A(\hat{\mathbf{k}})]=[A^{\dagger}(\mathbf{k}),A^{\dagger}(\hat{\mathbf{k}})]=[B(\mathbf{k}),B(\hat{\mathbf{k}})]=[B^{\dagger}(\mathbf{k}),B^{\dagger}(\hat{\mathbf{k}})]=\mathbf{0},\\\ \\\ N^{+}(\hat{\mathbf{k}}):=A^{\dagger}(\mathbf{k})A(\mathbf{k}),\\\ \\\ N^{-}(\hat{\mathbf{k}}):=B^{\dagger}(\mathbf{k})B(\mathbf{k}),\end{array}$ (27) where the eigenvalues of the number operators $N^{\pm}(\hat{\mathbf{k}})$ are the set $\\{0,1,2,...\\}$. The second quantized scalar field operator $\Phi(\mathbf{n};t)$ represents a collection of massive, spin-less, electrically charged, physical particles or quantas. One can show the following relations for total energy, total momentum and total electric charge respectively DasI ; DasII , $\begin{array}[]{c}\mathcal{H}:=\displaystyle{\int_{\mathbb{R}^{3}}}[N^{+}(\mathbf{k})+N^{-}(\mathbf{k})+\delta^{3}(\mathbf{0})\mathbf{I}(\mathbf{k})]\omega(\mathbf{k})dk_{1}dk_{2}dk_{3},\\\ \\\ \mathcal{P}_{j}:=\displaystyle{\int_{\mathbb{R}^{3}}}[N^{+}(\mathbf{k})+N^{-}(\mathbf{k})]k_{j}dk_{1}dk_{2}dk_{3},\\\ \\\ \mathcal{Q}:=e\displaystyle{\int_{\mathbb{R}^{3}}}[N^{+}(\mathbf{k})-N^{-}(\mathbf{k})]dk_{1}dk_{2}dk_{3}.\end{array}$ (28) The divergent null point energy term $\delta^{3}(\mathbf{0})\mathbf{I}(\mathbf{k})$ may be ignored for physical interpretations but cannot be directly remedied. ## 9 Many-particle systems, $(2+2+2)N$-dimensional phase space and $(2+2+2)N$-dimensional hyper-tori as phase cells Recall Hamilton’s canonical equations of motion for non-relativistic classical mechanics Lanczos ; Goldstein $\begin{array}[]{c}A\in\\{1,2,3,...,N\\},\\\ \\\ \dot{q}^{A}:=\dfrac{d\mathcal{Q}^{A}(t)}{dt},\\\ \\\ \dot{p}_{A}:=\dfrac{d\mathcal{P}_{A}(t)}{dt},\\\ \\\ \dot{q}^{A}:=\dfrac{\partial H}{\partial p_{A}}(q^{1},\cdots,q^{3N};p_{1},\cdots,p_{3N};t),\\\ \\\ \dot{p}_{A}:=-\dfrac{\partial H}{\partial q^{A}}(q^{1},\cdots,q^{3N};p_{1},\cdots,p_{3N};t).\end{array}$ (29) The corresponding Schrödinger wave equation for the first quantized physical system is furnished by $\begin{array}[]{c}i\dfrac{\partial}{\partial_{t}}\psi\left(q^{1},\cdots,q^{3N};t\right)=\\\ H\left(q^{1},\cdots,q^{3N};-i\frac{\partial}{\partial q^{1}},\cdots,-i\frac{\partial}{\partial q^{3N}};t\right)\psi(q^{1},\cdots,q^{3N};t),\end{array}$ (30) which has become ubiquitous with standard quantum mechanical systems being verified experimentally for a multitude of specific systems. From part I of this paper DasRCI , the $N=1$ relativistic Klein-Gordon equation in four-dimensional space is given by ($\eta_{\mu\nu}:=[1,1,1,-1]$) $\begin{array}[]{c}\eta^{\mu\nu}\dfrac{\partial^{2}}{\partial q^{\mu}\partial q^{\mu}}\psi\left(q^{1},q^{2},q^{3},q^{4}\right)-m^{2}\psi\left(q^{1},q^{2},q^{3},q^{4}\right)=0,\\\ or\\\ \delta^{ab}\dfrac{\partial^{2}}{\partial q^{a}\partial q^{b}}\psi\left(\mathbf{q};t\right)-(\partial_{t})^{2}\psi\left(\mathbf{q};t\right)-m^{2}\psi\left(\mathbf{q};t\right)=0.\\\ \end{array}$ (31) The relativistic wave equation for the case of two spin-$\frac{1}{2}$ particles $q_{(1)}^{\mu}$ and $q_{(2)}^{\nu}$ is furnished by the Bethe- Salpeter equation Bethe $\begin{array}[]{c}\left\\{\gamma_{(1)}^{\mu}\left[\left(\dfrac{m_{1}}{m_{1}+m_{2}}P_{\mu}+P_{\mu}\right)-im_{1}\right]\right\\}\left\\{\gamma_{(2)}^{\nu}\left[\left(\dfrac{m_{1}}{m_{1}+m_{2}}P_{\nu}+P_{\nu}\right)-im_{2}\right]\right\\}\\\ \\\ \cdot\psi\left(q_{(1)}^{1}-q_{(2)}^{1},q_{(1)}^{2}-q_{(2)}^{2},q_{(1)}^{3}-q_{(2)}^{3},q_{(1)}^{4}-q_{(2)}^{4}\right)=\\\ \\\ i\bar{G}\left(q_{(1)}^{1}-q_{(2)}^{1},q_{(1)}^{2}-q_{(2)}^{2},q_{(1)}^{3}-q_{(2)}^{3},q_{(1)}^{4}-q_{(2)}^{4}\right)\psi\left(q_{(1)}^{1}-q_{(2)}^{1},\cdots\right)\end{array}$ (32) where $\bar{G}$ is the appropriate Green’s function. In $N+1$-dimensional state space, the generalization to the relativistic Klein-Gordon wave equation is $\begin{array}[]{c}\left[\delta^{AB}\dfrac{\partial^{2}}{\partial q^{A}\partial q^{B}}-(\partial_{t})^{2}-m^{2}\right]\psi\left(q^{1},\cdots,q^{N};t\right)=0,\\\ A,B\in\\{1,\cdots,N\\}.\\\ \end{array}$ (33) with a group invariance of $\mathcal{I}[O(N,1)]_{+}^{+}$. Similarly, we can express within a $[(2+2+2)N+1]$-dimensional discrete phase space and continuous time arena, the first quantized partial differential- difference Klein-Gordon equation as $\left[\delta^{AB}\Delta^{\\#}_{A}\Delta^{\\#}_{B}-(\partial_{t})^{2}-m^{2}\right]\phi(n^{1},\cdots,n^{N};t)=0.$ (34) The second quantized version of this generalized Klein-Gordon equation is given by $\left[\delta^{AB}\Delta^{\\#}_{A}\Delta^{\\#}_{B}-(\partial_{t})^{2}-m^{2}\right]\Phi(n^{1},\cdots,n^{N};t)=\mathbf{0}$ (35) The group invariance of the above equation is provided by $\mathcal{I}[O(N,1)]_{+}^{+}$ (see DasRCI ). A special class of exact solutions for these relativistic partial difference- differential operator equations (35) are given by $\begin{array}[]{c}\Phi^{-}(n^{1},\cdots,n^{N};t):=\displaystyle{\int_{\mathbb{R}^{N}}}\dfrac{1}{\sqrt{2\omega(\mathbf{k})}}\left[A(\mathbf{k})\xi_{n^{1}}(k_{1})\cdots\xi_{n^{N}}(k_{N})e^{-i\omega t}\right]dk_{1}\cdots dk_{N},\\\ \\\ \Phi^{+}(n^{1},\cdots,n^{N};t):=\displaystyle{\int_{\mathbb{R}^{N}}}\dfrac{1}{\sqrt{2\omega(\mathbf{k})}}\left[B^{\dagger}(\mathbf{k})\overline{\xi_{n^{1}}}(k_{1})\cdots\overline{\xi_{n^{N}}}(k_{N})e^{i\omega t}\right]dk_{1}\cdots dk_{N},\\\ \\\ \Phi(n^{1},\cdots,n^{N};t):=\Phi^{-}(n^{1},\cdots,n^{N};t)+\Phi^{+}(n^{1},\cdots,n^{N};t),\\\ \\\ \omega=\omega(k_{1},\cdots,k_{N}):=+\sqrt{\delta^{AB}k_{A}k_{B}+m^{2}}>0.\end{array}$ (36) The second quantized linear operators $A(k_{1},\cdots,k_{N}),B^{\dagger}(k_{1},\cdots,k_{N})$, etc. (creation and annihilation operators in momentum space) acting linearly on a Hilbert space bundle must satisfy very similar commutation relations as those of (27). The linear operator $\Phi(n^{1},\cdots,n^{N};t)$ is defined over the geometric configuration $\bar{S}_{n^{1}}^{1}\times\bar{S}_{n^{2}}^{1}\times\cdots\times\bar{S}_{n^{N}}^{1}\times\mathbb{R}$ where $\bar{S}_{n^{A}}^{1}$ is a Peano circle of physical dimension $\left[\dfrac{ML^{2}}{T}\right]$. Table 1: Discrete Phase Space and Popular String Theory Dimension Comparison Finally, we would like to make a final comparison with our method to two popular approaches to quantizing gravity: string theory and loop quantum gravity. Table 1 consists of a higher dimensional comparison between our discrete phase space theory put forward here and that of various popular string theories Green ; PolchinskiI ; PolchinskiII . There appears to be a striking similarity between the spatial or gauge group dimension of string theory and the dimension of our discrete phase space. One clear advantage of our method is that we have gone further in understanding the scattering matrix of our theory, the $S^{\\#}$-matrix, as delineated in DasI ; DasII ; DasIII . In DasRC , we have explicitly calculated new Feynman rules for computations of the $S^{\\#}$-matrix elements and have derived an exact singularity free Coulomb-type potential within our discrete phase space approach. See Green ; PolchinskiI ; PolchinskiII for further information on the actual calculation abilities of string theory. Loop quantum gravity Rovelli , and its spin network structure of space-time composed of extremely fine but finite loops has a similarity to our Peano circle $\bar{S}^{1}_{n}$ of unit area in the $(1+1)$ dimensional phase plane. The area filling Peano curve is identified as the trajectory of one quanta as shown here and in DasRCI . It would be interesting to further investigate this analogy to possibly find a quantum theory of gravity within our framework or find Peano type motion within the framework of loop quantum gravity. ## 10 Concluding Remarks In this paper, we have shown how an area filling Peano curve represents a possible particle trajectory in the unit phase cell of a discrete phase space and continuous time relativistic quantum mechanical system. This is one of the first uses of this fascinating mathematical structure as a physical construct. Both first quantized Planck oscillators, first explored in part I of this work DasRCI , and second quantized Klein-Gordon excitations were explored with respect to the Peano curve formalism. Furthermore, the state space evolution of our Peano circle was shown to be analogous to the world-sheet evolution of closed strings. The geometric framework of this evolution was interpreted in terms of a product fibre bundle structure. Finally, extensions of our model to higher dimensions and striking similarities to popular dimensions used in traditional string theory were explored. ## Acknowledgements A.D. thanks Dr. Jack Gegenberg for some informal discussions. ## References * (1) A. Das and R. Chatterjee, Discrete phase space and continuous time relativistic quantum mechanics I: Planck oscillators and closed string-like circular orbits, arXiv:2012.14256, (2020). * (2) C. Clark, The Theoretical Side of Calculus, Wadsworth Publ. Co., Belmont, (1972). * (3) B.R. Gelbaum and J.M.H. Olmstead Counterexamples in Analysis, Holden-Day, Inc., San Francisco, (1964). * (4) A. Das and S. Haldar, Physical Science International Journal 17(2), 1 (2018). * (5) Y. Choquet-Bruhat, C. Dewitt-Morette, and M. Dillard-Bleck Analysis, Manifolds, and Physics North-Holland, Amsterdam, (1978). * (6) A. Das, Can. J. Phys. 88, 73 (2010). * (7) A. Das, Can. J. Phys. 88, 93 (2010). * (8) A. Das, Can. J. Phys. 88, 111 (2010). * (9) J.M. Jauch and F. Rohrlich, The Theory of Photons and Electrons, Addison-Wesley, Cambridge, MA, (1955). * (10) M.E. Peskin and D.V. Schroeder, An Introduction to Quantum Field Theory, Addison-Wesley, Cambridge, MA, (1995). * (11) A. Das and A. DeBenedictis, Scientific Voyage, 1, 45 (2015). * (12) A. Das, R. Chatterjee, and T. Yu, Mod. Phys. Lett. A 35, 24 (2020). * (13) R.R. Goldberg Methods of Real Analysis John-Wiley & Sons, New York, (1976). * (14) A.H. Lightstone Symbolic Logic and the Real Number System Harper & Row, New York, (1965). * (15) W.S. Massey Algebraic Topology: An Introduction Harcourt, Brace & World, , New York, (1967). * (16) M.B. Green, J.H. Schwarz, and E. Witten, Superstring Theory Vol 1 & 2, Cambridge University Pres, Cambridge, (1987). * (17) J. Polchinski, String Theory Vol 1, Cambridge University Pres, Cambridge, (1998). * (18) J. Polchinski, String Theory Vol 2, Cambridge University Pres, Cambridge, (1998). * (19) P.A.M. Dirac, The Principles of Quantum Mechanics, 4th Edition Oxford University Press, London, (1967). * (20) C. Lanczos, The Variational Principles of Mechanics, University of Toronto Press, Toronto, (1970). * (21) H. Goldstein, Classical Mechanics, 2nd, Addison-Wesley, Reading, Mass., (1980). * (22) M. Spivak, Calculus on Manifolds, Benjamin-Cummings, Melno Park, CA, (1965). * (23) A. Das, Tensors:The Mathematics of Relativity Theory and Continuum Mechanics, Springer-Verlag, Berlin, (2007). * (24) W. Greiner, Relativistic Quantum Mechanics, 3rd Edition Springer-Verlag, Berlin, (2000). * (25) F.W.J. Olver, Introduction to Asymptotics and Special Functions Academic Press, New York, (1974). * (26) A.H.. Zemanian, Distribution Theory and Transform Analysis Dover Publications, New York, (1965). * (27) R.C. Buck and E.F. Buck, Advanced Calculus McGraw-Hill, New York, (1965). * (28) H.A. Bethe and E.E. Salpeter, Quantum Mechanics of One and Two Electron Atoms Springer-Verlag, Berlin, (1957). * (29) C. Rovelli, Quantum Gravity Cambridge University Press, Cambridge, (2007).
]www.polyphys.mat.ethz.ch # Coordinate conditions and field equations for pure composite gravity Hans Christian Öttinger<EMAIL_ADDRESS>[ ETH Zürich, Department of Materials, CH-8093 Zürich, Switzerland ###### Abstract Whenever an alternative theory of gravity is formulated in a background Minkowski space, the conditions characterizing admissible coordinate systems, in which the alternative theory of gravity may be applied, play an important role. We here propose Lorentz covariant coordinate conditions for the composite theory of pure gravity developed from the Yang-Mills theory based on the Lorentz group, thereby completing this previously proposed higher derivative theory of gravity. The physically relevant static isotropic solutions are determined by various methods, the high-precision predictions of general relativity are reproduced, and an exact black-hole solution with mildly singular behavior is found. ## I Introduction Only two years after the discovery of Yang-Mills theories Yang and Mills (1954), it has been recognized that that there is a striking formal relationship between the Riemann curvature tensor of general relativity and the field tensor of the Yang-Mills theory based on the Lorentz group Utiyama (1956). However, developing this particular Yang-Mills theory into a consistent and convincing theory of gravity is not at all straightforward. The ideas of Utiyama (1956) have been found to be “unnatural” by Yang (see footnote 5 of Yang (1974)), whose work has later been criticized massively in Chap. 19 of Blagojević and Hehl (2013). Nevertheless, the pioneering work Utiyama (1956) may be considered as the origin of what is now known as gauge gravitation theory Capozziello and De Laurentis (2011); Ivanenko and Sardanashvily (1983). An obvious problem with the Yang-Mills theory based on the Lorentz group is that it has the large number of $48$ degrees of freedom, half of which are physically relevant. One is faced with six four-vector fields satisfying second-order evolution equations. For the pure field theories, the physical degrees of freedom are essentially given by the two transverse components of the four-vector fields, like in electrodynamics with its single vector potential. In view of this enormous number of degrees of freedom we need an almost equally large number of constraints to keep only a few degrees of freedom in a theory of gravity. In other words, we need a structured principle for selecting just a few ones among all the solutions of the Yang-Mills theory based on the Lorentz group. A powerful selection principle can be implemented by means of the tool of composite theories Öttinger (2018a, 2019). The basic idea is to write the gauge vector fields of the Yang-Mills theory in terms of fewer, more fundamental variables and their derivatives. The admission of derivatives in this so-called composition rule implies that the composite theory involves higher than second derivatives. The power of the tool of composite theories results from the fact that, in their Hamiltonian formulations Öttinger (2018a, 2019), the structure of the constraints providing the selection principle is highly transparent. As the composite theory of gravity Öttinger (2020a), just like the underlying Yang-Mills theory, is formulated in a background Minkowski space, the question arises how to characterize the “good” coordinate systems in which the theory may be applied. This characterization should be Lorentz invariant, but not invariant under more general coordinate transformations, that is, it shares the formal properties of coordinate conditions in general relativity. However, the unique solutions obtained from Einstein’s field equations only after specifying coordinate conditions are all physically equivalent, whereas the coordinate conditions in composite gravity characterize physically preferred systems. From a historical perspective, it is remarkable that Einstein in 1914 still believed that the metric should be completely determined by the field equations and, therefore, a generally covariant theory of gravity was not desirable (see Giovanelli (2020) for a detailed discussion). The important task of characterizing the preferred systems in composite gravity is addressed in the present paper. Once it is solved, we can provide a canonical Hamiltonian formulation of composite theory of gravity beyond the weak-field approximation Öttinger (2020b) and we obtain the static isotropic black-hole solution in a proper coordinate system. The structure of the paper is as follows. As a preparatory step, we present the various variables and relations between them (Sec. II) and discuss their gauge transformation behavior (Sec. III). A cornerstone of the development is the close relationship between the covariant derivatives associated (i) with a connection with torsion and (ii) with the Yang-Mills theory based on the Lorentz group. The core of the composite theory of gravity consists of the field equations presented for several sets of variables (Sec. IV) and the coordinate conditions characterizing the admissible coordinate systems (Sec. V). As an application, we determine the static isotropic solutions and provide the results for the high-precision tests of gravity as well as an exact black- hole solution (Sec. VI). We finally offer a detailed summary of our results and draw a number of conclusions (Sec. VII). A number of detailed results and arguments are provided in six appendices. ## II Various variables and relations between them For the understanding of composite theories, it is important to introduce different kinds of variables and to clarify the relations between them. On the one hand, we have the metric tensors, tetrad variables, connections and curvature tensors familiar from general relativity and other theories of gravity. On the other hand, we have the gauge vectors and field tensors of the Yang-Mills theory based on the Lorentz group. An important step is the decomposition of metric tensors in terms of tetrad or _vierbein_ variables, $g_{\mu\nu}=\eta_{\kappa\lambda}\,{b^{\kappa}}_{\mu}{b^{\lambda}}_{\nu}={b^{\kappa}}_{\mu}\,b_{\kappa\nu},$ (1) where $\eta_{\kappa\lambda}=\eta^{\kappa\lambda}$ is the Minkowski metric with signature $(-,+,+,+)$. Throughout this paper, the Minkowski metric is used for raising or lowering space-time indices. For the inverses of the metric and the tetrad variables we introduce the components $\bar{g}^{\mu\nu}$ and $\mbox{$\bar{b}^{\mu}$}_{\kappa}$. Note that they are not obtained by raising or lowering indices of $g_{\mu\nu}$ and ${b^{\kappa}}_{\mu}$, respectively. Equation (1) may be regarded as the characterization of metric tensors by symmetry and definiteness properties. A general metric tensor may also be regarded as the result of transforming the Minkowski metric. The decomposition of a metric $g_{\mu\nu}$ into tetrad variables ${b^{\kappa}}_{\mu}$ is not unique. If we multiply ${b^{\kappa}}_{\mu}$ from the left with any Lorentz transformation, the invariance of the Minkowski metric under Lorentz transformations implies that we obtain another valid decomposition. This observation reveals the origin of the underlying gauge symmetry of the composite theory of gravity. The key role of the metric tensor in the present theory is the characterization of the momentum-velocity relation, so that it can be interpreted as an indication of tensorial properties of mass. While this is also the case in general relativity, Einstein’s theory of gravity goes much further in the geometric interpretation of the metric by assuming that it characterizes the underlying space-time. In contrast, the present theory is developed in an underlying Minkowski space, which is the standard situation for Yang-Mills theories. As a next step, we introduce the vector fields $A_{(\kappa\lambda)\rho}$ in terms of the tetrad variables (the pair $(\kappa,\lambda)$ of space-time indices should be considered as a label associated with the Lorentz group, $\rho$ as a four-vector index), $\displaystyle{b^{\kappa}}_{\mu}{b^{\lambda}}_{\nu}\,A_{(\kappa\lambda)\rho}$ $\displaystyle=$ $\displaystyle\frac{1}{2}\left(\frac{\partial g_{\nu\rho}}{\partial x^{\mu}}-\frac{\partial g_{\mu\rho}}{\partial x^{\nu}}\right)$ (2) $\displaystyle+$ $\displaystyle\frac{1}{2\tilde{g}}\left({b^{\kappa}}_{\mu}\,\frac{\partial b_{\kappa\nu}}{\partial x^{\rho}}-\frac{\partial{b^{\kappa}}_{\mu}}{\partial x^{\rho}}\,b_{\kappa\nu}\right).\qquad$ From the Yang-Mills perspective, $\tilde{g}$ is the coupling constant. From a metric viewpoint, $\tilde{g}\neq 1$ implies torsion (see Eq. (6) below). The antisymmetry of the right-hand side of Eq. (2) in $\mu$ and $\nu$ leads, after resolving for $A_{(\kappa\lambda)\rho}$, to antisymmetry in $\kappa$ and $\lambda$. We have thus introduced six vector fields associated with six pairs $(\kappa,\lambda)$, or with a label $a$ taking the values from $1$ to $6$ according to Table 1. The pairs $(0,1)$, $(0,2)$, $(0,3)$ correspond to Lorentz boosts in the respective directions (involving also time) and the pairs $(2,3)$, $(3,1)$, $(1,2)$ correspond to rotations in the respective planes, as can be recognized by analyzing the gauge transformation behavior of the fields $A_{(\kappa\lambda)\rho}$ resulting from the freedom of acting with Lorentz transformations on ${b^{\kappa}}_{\mu}$ (see Sec. III for details). $a$ | $1$ | $2$ | $3$ | $4$ | $5$ | $6$ ---|---|---|---|---|---|--- $(\kappa,\lambda)$ | $(0,1)$ | $(0,2)$ | $(0,3)$ | $(2,3)$ | $(3,1)$ | $(1,2)$ Table 1: Correspondence between the label $a$ for the base vectors of the six- dimensional Lie algebra ${\rm so}(1,3)$ and ordered pairs $(\kappa,\lambda)$ of space-time indices. Following standard procedures for Yang-Mills theories (see, e.g., Sect. 15.2 of Peskin and Schroeder (1995), Chap. 15 of Weinberg (2005), or Öttinger (2018b)), we can introduce a field tensor in terms of the vector fields, $F_{a\mu\nu}=\frac{\partial A_{a\nu}}{\partial x^{\mu}}-\frac{\partial A_{a\mu}}{\partial x^{\nu}}+\tilde{g}f^{bc}_{a}A_{b\mu}A_{c\nu},$ (3) where $f^{bc}_{a}$ are the structure constants of the Lorentz group. A Lie algebra label, say $a$, can be raised or lowered by raising or lowering the indices in the pairs associated with $a$ according to Table 1. The structure constants can then be specified as follows: $f^{abc}$ is $1$ ($-1$) if $(a,b,c)$ is an even (odd) permutation of $(4,5,6)$, $(1,3,5)$, $(1,6,2)$ or $(2,4,3)$ and $0$ otherwise (see also Eq. (51)). The definition (2) suggests the following general passage from quantities labeled by a Lie algebra index to a quantity with space-time indices, $\tilde{X}_{\mu\nu}={b^{\kappa}}_{\mu}{b^{\lambda}}_{\nu}\,X_{(\kappa\lambda)}.$ (4) One then gets a deep relation between covariant derivatives associated with metrics and connections on the one hand and covariant derivatives associated with a Yang-Mills theory based on the Lorentz group on the other hand (for a proof of this fundamental relation based on the structure of the Lorentz group, see Appendix A), $\displaystyle\frac{\partial\tilde{X}_{\mu\nu}}{\partial x^{\rho}}-\Gamma^{\sigma}_{\rho\mu}\tilde{X}_{\sigma\nu}-\Gamma^{\sigma}_{\rho\nu}\tilde{X}_{\mu\sigma}$ $\displaystyle=$ (5) $\displaystyle{b^{\kappa}}_{\mu}{b^{\lambda}}_{\nu}\left[\frac{\partial X_{(\kappa\lambda)}}{\partial x^{\rho}}+\tilde{g}\,f_{(\kappa\lambda)}^{bc}A_{b\rho}X_{c}\right],$ where the connection $\Gamma^{\rho}_{\mu\nu}$ is given by $\Gamma^{\rho}_{\mu\nu}=\frac{1}{2}\,\bar{g}^{\rho\sigma}\left[\frac{\partial g_{\sigma\nu}}{\partial x^{\mu}}+\tilde{g}\left(\frac{\partial g_{\mu\sigma}}{\partial x^{\nu}}-\frac{\partial g_{\mu\nu}}{\partial x^{\sigma}}\right)\right]=\bar{g}^{\rho\sigma}\,\bar{\Gamma}_{\sigma\mu\nu}.$ (6) Unlike the Christoffel symbols obtained for $\tilde{g}=1$, $\Gamma^{\rho}_{\mu\nu}$ is not symmetric in $\mu$ and $\nu$ for $\tilde{g}\neq 1$. This lack of symmetry indicates the presence of torsion. Note, however, that the connection is metric-compatible for all $\tilde{g}$ Jiménez _et al._ (2019), that is, $\frac{\partial g_{\mu\nu}}{\partial x^{\rho}}-\Gamma^{\sigma}_{\rho\mu}g_{\sigma\nu}-\Gamma^{\sigma}_{\rho\nu}g_{\mu\sigma}=0,$ (7) which can be recast in the convenient form $\frac{\partial g_{\mu\nu}}{\partial x^{\rho}}=\bar{\Gamma}_{\mu\rho\nu}+\bar{\Gamma}_{\nu\rho\mu}.$ (8) From the connection $\Gamma^{\rho}_{\mu\nu}$, we can further construct the Riemann curvature tensor (see, e.g. Jiménez _et al._ (2019) or Weinberg (1972)) ${R^{\mu}}_{\nu\mu^{\prime}\nu^{\prime}}=\frac{\partial\Gamma^{\mu}_{\mu^{\prime}\nu}}{\partial x^{\nu^{\prime}}}-\frac{\partial\Gamma^{\mu}_{\nu^{\prime}\nu}}{\partial x^{\mu^{\prime}}}+\Gamma^{\sigma}_{\mu^{\prime}\nu}\Gamma^{\mu}_{\nu^{\prime}\sigma}-\Gamma^{\sigma}_{\nu^{\prime}\nu}\Gamma^{\mu}_{\mu^{\prime}\sigma}.$ (9) In Appendix B, it is shown that the field tensor (3) can be written in the alternative form $\displaystyle\tilde{F}_{\mu\nu\mu^{\prime}\nu^{\prime}}$ $\displaystyle=$ (10) $\displaystyle\frac{1}{2}\left(\frac{\partial^{2}g_{\nu\nu^{\prime}}}{\partial x^{\mu}\partial x^{\mu^{\prime}}}-\frac{\partial^{2}g_{\nu\mu^{\prime}}}{\partial x^{\mu}\partial x^{\nu^{\prime}}}-\frac{\partial^{2}g_{\mu\nu^{\prime}}}{\partial x^{\nu}\partial x^{\mu^{\prime}}}+\frac{\partial^{2}g_{\mu\mu^{\prime}}}{\partial x^{\nu}\partial x^{\nu^{\prime}}}\right)$ $\displaystyle+\,\frac{1}{\tilde{g}}\,\bar{g}^{\rho\sigma}(\bar{\Gamma}_{\rho\mu^{\prime}\mu}\bar{\Gamma}_{\sigma\nu^{\prime}\nu}-\bar{\Gamma}_{\rho\nu^{\prime}\mu}\bar{\Gamma}_{\sigma\mu^{\prime}\nu}).$ This explicit expression for $\tilde{F}_{\mu\nu\mu^{\prime}\nu^{\prime}}$ reveals its symmetry properties: antisymmetry under $\mu\leftrightarrow\nu$ and $\mu^{\prime}\leftrightarrow\nu^{\prime}$ and, more surprisingly, symmetry under $(\mu\nu)\leftrightarrow(\mu^{\prime}\nu^{\prime})$. A comparison between the expressions (9) and (10) yields a remarkable relationship between the Riemann curvature tensor and the field tensor of the Yang-Mills theory based on the Lorentz group, $\tilde{g}\,\bar{g}^{\mu\rho}\tilde{F}_{\rho\nu\mu^{\prime}\nu^{\prime}}={R^{\mu}}_{\nu\mu^{\prime}\nu^{\prime}},$ (11) which holds for all values of the coupling constant $\tilde{g}$. ## III Gauge transformation behavior As a consequence of the decomposition (1), there exists the gauge freedom of acting with a Lorentz transformation from the left on the tetrad variables ${b^{\kappa}}_{\mu}$. In its infinitesimal version, this possibility corresponds to the transformation $\delta b_{\kappa\mu}=\tilde{g}\,\Lambda_{(\kappa\lambda)}{b^{\lambda}}_{\mu},$ (12) where $\Lambda_{(\kappa\lambda)}$ is antisymmetric in $\kappa$ and $\lambda$ and can hence be understood as $\Lambda_{a}$ according to Table 1. For $\kappa=0$, time is mixed with a spatial dependence in one of the coordinate directions so that we deal with the respective Lorentz boosts. If both $\kappa=k$ and $\lambda=l$ are both spatial indices, the antisymmetric matrix $\Lambda_{(\kappa\lambda)}$ describes rotations in the corresponding $(k,l)$ plane. For the inverse of ${b^{\kappa}}_{\mu}$, Eq. (12) implies $\delta\mbox{$\bar{b}^{\mu}$}_{\kappa}=-\tilde{g}\,\Lambda_{(\kappa\lambda)}\bar{b}^{\mu\lambda}.$ (13) By using Eq. (12) in the composition rule (2), we obtain $\delta A_{(\kappa\lambda)\rho}-\tilde{g}\,\eta^{\kappa^{\prime}\lambda^{\prime}}\Big{[}A_{(\kappa^{\prime}\lambda)\rho}\Lambda_{(\kappa\lambda^{\prime})}-\Lambda_{(\kappa^{\prime}\lambda)}A_{(\kappa\lambda^{\prime})\rho}\Big{]}=\frac{\partial\Lambda_{(\kappa\lambda)}}{\partial x^{\rho}},$ (14) which, by means of Eq. (49), can be written as $\delta A_{a\rho}=\frac{\partial\Lambda_{a}}{\partial x^{\rho}}+\tilde{g}f^{bc}_{a}\,A_{b\rho}\,\Lambda_{c}.$ (15) This result demonstrates that the six vector fields $A_{a\rho}$ indeed possess the proper gauge transformation behavior for the vector fields of the Yang- Mills theory based on the Lorentz group. By means of the Jacobi identity for the structure constants, $f^{sb}_{a}f^{cd}_{s}+f^{sc}_{a}f^{db}_{s}+f^{sd}_{a}f^{bc}_{s}=0,$ (16) we further obtain the gauge transformation behavior of the field tensor, $\delta F_{a\mu\nu}=\tilde{g}f^{bc}_{a}\,F_{b\mu\nu}\,\Lambda_{c}.$ (17) Finally, we look at the gauge transformation behavior obtained for the Yang- Mills variables transformed according to Eq. (4). From Eqs. (12) and (14) we obtain $\delta\tilde{A}_{\mu\nu\rho}={b^{\kappa}}_{\mu}{b^{\lambda}}_{\nu}\,\frac{\partial\Lambda_{(\kappa\lambda)}}{\partial x^{\rho}}.$ (18) As the metric is gauge invariant (gauge degrees of freedom result only from its decomposition), the representations (6) and (10) imply the gauge invariance properties $\delta\Gamma^{\rho}_{\mu\nu}=\delta\bar{\Gamma}_{\sigma\mu\nu}=0,$ (19) and $\delta\tilde{F}_{\mu\nu\mu^{\prime}\nu^{\prime}}=0.$ (20) ## IV Field equations With the help of Eq. (5), the standard field equations for our Yang-Mills theory based on the Lorentz group (see, e.g., Sect. 15.2 of Peskin and Schroeder (1995), Chap. 15 of Weinberg (2005), or Öttinger (2018b)) can be written in the manifestly gauge invariant form $\eta^{\mu^{\prime}\mu^{\prime\prime}}\left(\frac{\partial\tilde{F}_{\mu\nu\mu^{\prime\prime}\nu^{\prime}}}{\partial x^{\mu^{\prime}}}-\Gamma^{\sigma}_{\mu^{\prime}\mu}\tilde{F}_{\sigma\nu\mu^{\prime\prime}\nu^{\prime}}-\Gamma^{\sigma}_{\mu^{\prime}\nu}\tilde{F}_{\mu\sigma\mu^{\prime\prime}\nu^{\prime}}\right)=0.$ (21) By means of Eq. (11), these field equations can be rewritten in terms of the Riemann curvature tensor, $\eta^{\rho\nu^{\prime}}\left(\frac{\partial{R^{\mu}}_{\nu\mu^{\prime}\nu^{\prime}}}{\partial x^{\rho}}+\Gamma^{\mu}_{\rho\sigma}{R^{\sigma}}_{\nu\mu^{\prime}\nu^{\prime}}-\Gamma^{\sigma}_{\rho\nu}{R^{\mu}}_{\sigma\mu^{\prime}\nu^{\prime}}\right)=0.$ (22) In view of Eq. (9), this latter equation is entirely in terms of the variables $\Gamma^{\rho}_{\mu\nu}$. The explicit form of the resulting equation is given in Appendix C. This observation offers the option of the following two-step procedure: one first determines the most general solution of the second-order differential equations (LABEL:compactGameq) for $\Gamma^{\rho}_{\mu\nu}$ and then, in a post-processing step, one obtains the metric by solving the first- order differential equations (6). The post-processing step selects those solutions $\Gamma^{\rho}_{\mu\nu}$ that can actually be expressed in terms of the metric. Finally, we write the field equations directly as third-order differential equations for the metric. As the solutions of these third-order equations can be understood in terms of selected solutions of the Yang-Mills theory found by post-processing, there is no reason to be concerned about the potential instabilities resulting from higher-order differential equations, known as Ostrogradsky instabilities Ostrogradsky (1850); Woodard (2015). Avoiding such instabilities is an important topic, in particular, in alternative theories of gravity j. Chen _et al._ (2013); Raidal and Veermäe (2017); Stelle (1977, 1978); Krasnikov (1987); Grosse-Knetter (1994); Becker _et al._ (2017); Salvio (2019). We write all the third and second derivatives of the metric explicitly, whereas the first derivatives are conveniently combined into connection variables. The result is the following set of equations for the composite theory of gravity obtained by expressing the gauge vector fields of the Yang-Mills theory based on the Lorentz group in terms of the tetrad variables obtained by decomposing a metric, $\displaystyle\Xi_{\mu\nu\mu^{\prime}}$ $\displaystyle=$ $\displaystyle\frac{1}{2}\frac{\partial}{\partial x^{\mu}}\square g_{\mu^{\prime}\nu}-\frac{1}{2}\frac{\partial^{2}}{\partial x^{\mu}\partial x^{\mu^{\prime}}}\frac{\partial g_{\nu\rho}}{\partial x_{\rho}}$ (23) $\displaystyle-\,\frac{1}{2}\Gamma^{\sigma}_{\mu^{\prime}\mu}\bigg{(}\frac{1}{\tilde{g}}\square g_{\sigma\nu}+\frac{\partial}{\partial x^{\nu}}\frac{\partial g_{\sigma\rho}}{\partial x_{\rho}}-\frac{\partial}{\partial x^{\sigma}}\frac{\partial g_{\nu\rho}}{\partial x_{\rho}}\bigg{)}$ $\displaystyle+\,\frac{\eta^{\rho\rho^{\prime}}}{2}\Gamma^{\sigma}_{\rho\nu}\bigg{(}\frac{\partial^{2}g_{\sigma\rho^{\prime}}}{\partial x^{\mu}\partial x^{\mu^{\prime}}}-\frac{\partial^{2}g_{\mu\rho^{\prime}}}{\partial x^{\sigma}\partial x^{\mu^{\prime}}}$ $\displaystyle\hskip 5.0pt+\,2\frac{\partial^{2}g_{\mu\mu^{\prime}}}{\partial x^{\sigma}\partial x^{\rho^{\prime}}}-2\frac{\partial^{2}g_{\sigma\mu^{\prime}}}{\partial x^{\mu}\partial x^{\rho^{\prime}}}-\frac{1}{\tilde{g}}\frac{\partial^{2}g_{\sigma\mu}}{\partial x^{\mu^{\prime}}\partial x^{\rho^{\prime}}}\bigg{)}$ $\displaystyle+\,\frac{\eta^{\rho\rho^{\prime}}}{\tilde{g}}\bigg{[}\Gamma^{\alpha}_{\mu^{\prime}\mu}\bigg{(}2\bar{\Gamma}_{\alpha\rho^{\prime}\beta}+\bar{\Gamma}_{\beta\rho^{\prime}\alpha}\bigg{)}-\Gamma^{\alpha}_{\rho^{\prime}\mu}\bar{\Gamma}_{\alpha\mu^{\prime}\beta}\bigg{]}\Gamma^{\beta}_{\rho\nu}$ $\displaystyle\hskip 80.00012pt-\;\boxed{\mu\leftrightarrow\nu}=0.$ In view of the antisymmetry of $\Xi_{\mu\nu\mu^{\prime}}$ implied by the last line of the above equation, we can assume $\mu<\nu$ so that Eq. (23) provides a total of $24$ equations for the ten components of the symmetric matrix $g_{\mu\nu}$. If we wish to determine the time evolution of the metric from the third-order differential equations (23), we need $30$ initial conditions for the matrix elements $g_{\mu\nu}$ and their first and second time derivatives as well as expressions for the third time derivatives. Closer inspection of the third-order terms in Eq. (23) reveals that the six equations $\Xi_{0mn}=0$ for $m\leq n$ provide the derivatives $\partial^{3}g_{mn}/\partial t^{3}$, but that the remaining equations do not contain any information about $\partial^{3}g_{0\mu}/\partial t^{3}$. Therefore, the remaining $18$ equations constitute constraints for the initial conditions, and we are faced with two tasks: (i) find equations for the time evolution of $g_{0\mu}$, and (ii) show that the constraints are satisfied at all times if they hold initially (or count the additional constraints that need to be satisfied otherwise). It is not at all trivial to find the number of further constraints arising from the dynamic invariance of the constraints contained in Eq. (23). A controlled handling of constraints is more straightforward in a Hamiltonian setting. As the canonical Hamiltonian formulation has been elaborated only in the weak-field approximation Öttinger (2020b), we sketch the generalizations required for the full, nonlinear theory of composite pure gravity in Appendix E. As a conclusion, we expect (at least) four physical degrees of freedom remaining in the field equations (23) for $g_{\mu\nu}$. Note that the Hamiltonian approach also provides the natural starting point for a generalization to dissipative systems. In particular, this approach allows us to formulate quantum master equations Breuer and Petruccione (2002); Weiss (2008); Öttinger (2011); Taj and Öttinger (2015) and to make composite gravity accessible to the robust framework of dissipative quantum field theory Öttinger (2017); Oldofredi and Öttinger (2021). The issue of missing evolution equations is addressed in the subsequent section. As in the weak-field approximation, coordinate conditions characterizing those coordinate systems in which the composite theory of gravity can be applied provide the missing evolution equations. ## V Coordinate conditions As we have assumed an underlying Minkowski space for developing composite gravity, we need to characterize those coordinate systems in which the theory actually holds. These characteristic coordinate conditions should clearly be Lorentz covariant. Furthermore, the coordinate conditions should provide evolution equations for $g_{0\mu}$ because the field equations (23) determine the third-order time derivatives of $g_{mn}$, but not of $g_{0\mu}$. Therefore, the formulation of appropriate coordinate conditions is an important task. The status of coordinate conditions in composite theory is very different from their status in general relativity, where they have no influence on the physical predictions. The coordinate conditions should be a set of four Lorentz covariant equations. An appealing form is given by $\frac{\partial g_{\mu\rho}}{\partial x_{\rho}}=\frac{\partial\phi}{\partial x^{\mu}},$ (24) where the potential $\phi$ is often assumed to be proportional to the trace of the metric. To eliminate the need of specifying a potential, we can write the second-order integrability conditions $\frac{\partial}{\partial x^{\nu}}\frac{\partial g_{\mu\rho}}{\partial x_{\rho}}=\frac{\partial}{\partial x^{\mu}}\frac{\partial g_{\nu\rho}}{\partial x_{\rho}}.$ (25) After taking the derivatives with respect to $x_{\nu}$ and summing over $\nu$, we arrive at the four Lorentz covariant coordinate conditions $\square\frac{\partial g_{\mu\rho}}{\partial x_{\rho}}=K\frac{\partial}{\partial x^{\mu}}\frac{\partial^{2}g_{\rho\sigma}}{\partial x_{\rho}\partial x_{\sigma}},$ (26) actually with $K=1$. Note that it is very appealing to use third-order equations as coordinate conditions because we actually need only expressions for the third time derivatives of $g_{0\mu}$ (stronger, first-order conditions are needed for the Hamiltonian formulation; see Appendix E). For $K=1$, we would obtain such equations for $g_{0m}$, but not for $g_{00}$. This is the reason why we have introduced the factor $K$ in Eq. (26). For any $K\neq 1$, we obtain the desired four evolution equations for $g_{0\mu}$. Formally, we could stick to the first-order conditions (24), but then the potential $\phi$ would be described by the second-order differential equations $\square\phi=K\frac{\partial^{2}g_{\rho\sigma}}{\partial x_{\rho}\partial x_{\sigma}},$ (27) where suitable space-time boundary conditions would be required. Note, however, that for $K\neq 1$, Eqs. (24) and (27) imply $\frac{\partial^{2}g_{\rho\sigma}}{\partial x_{\rho}\partial x_{\sigma}}=0,$ (28) whereas Eq. (26) implies the weaker requirement $\square\frac{\partial^{2}g_{\rho\sigma}}{\partial x_{\rho}\partial x_{\sigma}}=0.$ (29) The coordinate conditions (26) are an essential new ingredient into the composite theory of gravity. Of course, these coordinate conditions take a particularly simple form for $K=0$, which is a possible choice. Alternatively, we could choose $K=\tilde{g}/(1+\tilde{g})$ because we can then express the coordinate conditions as $\frac{\partial^{2}\bar{\Gamma}_{\mu\rho\sigma}}{\partial x_{\rho}\partial x_{\sigma}}=0.$ (30) In the following, we leave the particular choice of $K\neq 1$ open. From a structural point of view, the coordinate conditions (26) have the important advantage that they can be implemented in exactly the same way as the gauge conditions in Yang-Mills theories: one can add a term to the Lagrangian that does not lead to any modification of the field equations, provided that the desired (coordinate or gauge) conditions are imposed as constraints. For the coordinate conditions (26), the additional contribution to the Lagrangian is given in Appendix D. ## VI Static isotropic solution The study of static isotropic solutions of composite gravity is of great importance because these solutions provide the predictions for the high- precision tests of general relativity (deflection of light by the sun, anomalous precession of the perihelion of Mercury, gravitational redshift of spectral lines from white dwarf stars, travel time delay for radar signals reflecting off other planets) and the properties of black holes. Therefore, we here discuss these solutions in great detail. We assume that the static isotropic solutions are of the general form, $g_{\mu\nu}=\left(\begin{matrix}-\beta&0\\\ 0&\alpha\,\delta_{mn}+\xi\,\frac{x_{m}x_{n}}{r^{2}}\end{matrix}\right),$ (31) with inverse $\bar{g}^{\mu\nu}=\left(\begin{matrix}-\frac{1}{\beta}&0\\\ 0&\frac{\delta_{mn}}{\alpha}-\frac{\xi}{\alpha(\alpha+\xi)}\,\frac{x_{m}x_{n}}{r^{2}}\end{matrix}\right),$ (32) where $\alpha$, $\beta$ and $\xi$ are functions of the single variable $r=(x_{1}^{2}+x_{2}^{2}+x_{3}^{2})^{1/2}$. The static isotropic metric (31) is given in terms of the three real-valued functions $\alpha$, $\beta$ and $\xi$. In the original work on the composite theory of gravity (see Sec. V of Öttinger (2020a)), we had parametrized these three functions in terms of only two functions $A$ and $B$: $\alpha=1$, $\beta=B$, and $\xi=A-1$. This particular parametrization corresponds to standard quasi-Minkowskian coordinates. A problem with these quasi-Minkowskian coordinates is that it is unclear how they can be generalized to full coordinate conditions for general metrics. The more general form (31) of the metric is consistent with the coordinate conditions (26). In particular, we do not need to introduce a further function for characterizing the components $g_{0m}$. In general relativity, the form (31) of the metric (with $g_{0m}=0$) can be achieved by shifting time by a function depending on $r$ (see Sec. 8.1 of Weinberg (1972)). Nonzero $g_{0m}$ arise by Lorentz transformation of the metric (31) so that the form (31) belongs to a particularly simple solution of coordinate conditions and field equations. The field equations (23) provide two third-order ordinary differential equations involving all three functions $\alpha$, $\beta$ and $\xi$. For $K\neq 1$, the coordinate conditions (26) lead to another third-order differential equation relating $\alpha$ and $\xi$, which is actually independent of $K$; only for $K=1$, no further condition arises. In the remainder of this section, we solve the three differential equations for our three unknown functions for $K\neq 1$ by various methods. ### VI.1 Robertson expansion The high-precision tests for theories of gravity depend on the behavior of the static isotropic solutions at large distances. We therefore construct the so- called Robertson expansion in terms of $1/r$. One obtains the following results, $\alpha=1+\alpha_{1}\frac{r_{0}}{r}+\alpha_{3}\frac{r_{0}^{3}}{r^{3}}+\ldots,$ (33) $\xi=\xi_{1}\frac{r_{0}}{r}+\ldots,$ (34) and $\beta=1-2\frac{r_{0}}{r}+\big{[}2+(\tilde{g}-1)(\alpha_{1}+\xi_{1})\big{]}\,\frac{r_{0}^{2}}{2r^{2}}+\ldots,$ (35) where all higher terms indicated by $\ldots$ in these Robertson expansions are uniquely determined by the dimensionless parameters $\alpha_{1}$, $\alpha_{3}$, $\xi_{1}$ and the coupling constant $\tilde{g}$. However, $\alpha_{1}$ and $\xi_{1}$ are not independent but rather related by a cubic algebraic equation with a single real solution establishing a one-to-one relation between $\alpha_{1}$ and $\xi_{1}$ (see Appendix F). The parameter $r_{0}$ with dimension of length is determined by the mass at the center creating the static isotropic field, as can be shown by reproducing the limit of Newtonian gravity (see, e.g., Sec. 3.4 of Weinberg (1972)). An obvious strategy for finding the dimensionless parameters is to make sure that the high-precision predictions of general relativity are reproduced. This is achieved by choosing $\alpha_{1}+\xi_{1}=2,\qquad\alpha_{1}=\tilde{g}.$ (36) Imposing a further relation between $\alpha_{1}$ and $\xi_{1}$ is subtle as we have already established the cubic relationship between these parameters given explicitly in Eq. (69). This implies that the first part of Eq. (36) can be satisfied only for particular values of the coupling constant $\tilde{g}$. By using Eq. (36) for eliminating $\alpha_{1}$ and $\xi_{1}$ from Eq. (69), we obtain the following equation for $\tilde{g}$, $(4+4\tilde{g}-\tilde{g}^{2}-5\tilde{g}^{3})(2-\tilde{g})=0.$ (37) Two of the roots of this polynomial equation of degree four are real. In addition to the obvious root $2$, implying $\alpha_{1}=2$ and $\xi_{1}=0$, one finds the further real-valued root $\frac{1}{15}\big{[}(1259+30\sqrt{1509})^{1/3}+(1259-30\sqrt{1509})^{1/3}-1\big{]},$ which is approximately equal to $1.13164$; although closer to unity, this irrational number seems to be less appealing than the integer $2$. Only for these two values of $\tilde{g}$ composite gravity with the coordinate conditions (26) for $K\neq 1$ can reproduce the high-precision predictions of general relativity. Note that the parameter $\alpha_{3}$ in the expansions (33)–(35) remains undetermined as it is the only term among the listed ones that does not affect the high-precision tests of gravity. ### VI.2 Short-distance singularity We next focus on singular behavior at small distances, which we expect to describe black holes. A glance at the field equations (23) reveals that any fixed multiple of a solution is another solution of the field equations. For the “equidimensional” third-order differential equations determining the functions $\alpha$, $\beta$ and $\xi$ of $r$, we assume the following form, $\alpha=\frac{c_{\alpha}}{r^{x}},\qquad\beta=\frac{c_{\beta}}{r^{x}},\qquad\xi=\frac{c_{\xi}}{r^{x}},$ (38) with constants $c_{\alpha}$, $c_{\beta}$, $c_{\xi}$ and an exponent $x$. We further assume that $c_{\alpha}$, $c_{\beta}$, and $x$ are different from zero. For $\tilde{g}=2$, we then find that the field equations and coordinate conditions are equivalent to $c_{\xi}=0$ and $x=1$. For general $\tilde{g}$, one can verify that the values $x=\frac{2}{\tilde{g}},\qquad c_{\xi}=0,$ (39) lead to a static isotropic solution of both field equations and coordinate conditions. Of course, this solution is physically unacceptable as a global solution because it does not converge to the Minkowski metric at large distances. It does, however, characterize the asymptotic singular behavior of physical solutions at short distances. The exponent $x$ given in Eq. (39) speaks strongly in favor of choosing $\tilde{g}=2$ (rather than an irrational value). We then obtain a solution decaying according to a $1/r$ power law, the spatial part of which is a multiple of the three-dimensional unit matrix. ### VI.3 Numerical solution After discussing the static isotropic solutions at large and small distances from the center, we would now like to consider their behavior over the entire range of $r$. In particular, we are interested in the influence of the so far undetermined parameter $\alpha_{3}$ in Eq. (33) on the behavior of the solutions. To explore the full solutions, we solve the field equations and coordinate conditions by numerical integration, starting from a large initial distance $r_{\rm i}$ and then proceeding to smaller values of $r$. Assuming $\tilde{g}=2$, the initial conditions at $r_{\rm i}$ are given by the truncated third-order expansions $\alpha=1+2\frac{r_{0}}{r}+\alpha_{3}\frac{r_{0}^{3}}{r^{3}},$ (40) $\xi=-3\alpha_{3}\frac{r_{0}^{3}}{r^{3}},$ (41) $\beta=1-2\frac{r_{0}}{r}+2\frac{r_{0}^{2}}{r^{2}}-2\frac{r_{0}^{3}}{r^{3}}.$ (42) These expressions do not only provide the values of the coefficient functions at $r_{\rm i}$, but also their first and second derivatives required for solving the third-order differential equations for the functions $\alpha$, $\beta$ and $\xi$ of $r$. The actual numerical solution is performed with an implicit Runge-Kutta scheme of Mathematica. Figure 1: The functions $\beta$ (dashed line) and $\xi$ (continuous lines) characterizing the temporal and off-diagonal components of the isotropic metric (31) obtained from the composite theory for gravity for $\tilde{g}=2$ and $\alpha_{3}=\pm 0.25$. Positive and negative values of $\xi$ correspond to $\alpha_{3}=-0.25$ and $\alpha_{3}=0.25$, respectively. If $r_{\rm i}$ is sufficiently large, that is, in the range of validity of the asymptotic solutions (40)-(42), the numerical solutions are expected to be independent of the choice of $r_{\rm i}$. This expectation is scrutinized in Figure 1. This figure displays the functions $\beta$ and $\xi$ for the values $\alpha_{3}=\pm 0.25$ in the conditions (40), (41). The numerical solutions have been calculated for $r_{\rm i}=50$ and $r_{\rm i}=500$, so that each curve for $\xi$ actually consists of two overlapping curves and the anticipated independence of the results of $r_{\rm i}$ is confirmed. The result for $\beta$ actually consists of four curves, which implies that $\xi$ has remarkably little influence on the function $\beta$ until it touches the $r$ axis. Figure 1 suggests that $\xi$ diverges around the value $r$ at which $\beta$ touches the $r$ axis (and numerical difficulties arise). According to Eqs. (38), (39), $\xi$ must go to zero for small $r$. The real function $\xi$ might actually end in a cusp singularity and develop a complex branch at smaller $r$ that reaches zero at $r=0$ (see Sec. V C of Öttinger (2020a)). Alternatively, $\xi$ might jump from $+\infty$ to $-\infty$, or vice versa, to return as a real function to zero at $r=0$, where it started at large $r$ (this kind of behavior is found for the Schwarzschild solution of general relativity; see Sec. VI.5). To avoid singularities at finite $r$ we from now on assume $\alpha_{3}=0$, for which $\xi(r)$ is found to be identically zero. Note that singularities would be much more alarming in the composite theory of gravity than in general relativity because they cannot be considered as artifacts (“coordinate singularities”) removable by general coordinate transformations. ### VI.4 An exact solution As we have by now fixed the values of the coupling constant ($\tilde{g}=2$) and all the free parameters in the Robertson expansions (33)-(35) ($\alpha_{1}=2$, $\xi_{1}=0$, $\alpha_{3}=0$), there should be a unique static isotropic solution, which is the counterpart of the Schwarzschild solution in general relativity. The Robertson expansions suggest that all higher coefficients $\alpha_{n}$, $\xi_{n}$ for $n\geq 2$ vanish, so that $\alpha$ consists of only two terms and $\xi$ vanishes identically, as already noted in the numerical solutions. Then, a closed-form expression for $\beta$ can be found from the field equation $4r_{0}^{2}\,\beta=r^{4}\left(1+2\frac{r_{0}}{r}\right){\beta^{\prime}}^{2},$ (43) so that we arrive at the complete solution $\alpha=1+2\frac{r_{0}}{r},\quad\xi=0,\quad\beta=\left(2-\sqrt{1+2\frac{r_{0}}{r}}\right)^{2}.$ (44) These functions $\alpha$ and $\beta$ are shown in Figure 2. The present results are qualitatively similar to what was found in previous work on the composite theory of gravity for different coordinate conditions (see Fig. 1 of Öttinger (2020a)). Figure 2: The exact solutions (44) for the functions $\alpha$ and $\beta$ characterizing the diagonal components of the isotropic metric (31) in the composite theory for gravity. Note that $\beta$ is non-negative, vanishes at $r=(2/3)r_{0}$, and that $\sqrt{\alpha}\pm\sqrt{\beta}=2$, where the $+$ sign holds for $r\geq(2/3)r_{0}$ and the $-$ sign for $r\leq(2/3)r_{0}$. The only singularities occur at the origin, and they are of the Newtonian $1/r$ type. The most remarkable feature is that $\beta$ reaches a local minimum at $r=(2/3)r_{0}$, where $\beta$ becomes zero. The observation that the proper time stands still at this distance from the origin is the essence of black- hole behavior in the composite theory of gravity. An interesting consequence of $\beta=0$ is revealed by considering the curvature scalar $R=\bar{g}^{\nu\nu^{\prime}}{R^{\mu}}_{\nu\mu\nu^{\prime}}=\tilde{g}\,\bar{g}^{\mu\mu^{\prime}}\bar{g}^{\nu\nu^{\prime}}\tilde{F}_{\mu\nu\mu^{\prime}\nu^{\prime}}=\tilde{g}\,\mbox{$\bar{b}^{\mu}$}_{\kappa}\mbox{$\bar{b}^{\nu}$}_{\lambda}{F^{(\kappa\lambda)}}_{\mu\nu}.$ (45) For the isotropic solution given in Eq. (44), we find $R=\frac{16r_{0}^{2}}{r^{4}\left(1+2\frac{r_{0}}{r}\right)^{3}\left(\sqrt{1+2\frac{r_{0}}{r}}-2\right)},$ (46) which implies infinite curvature at $r=(2/3)r_{0}$ where $\beta$ vanishes, and a change of sign at that point. This is an important insight because, in the weak-field approximation, the curvature scalar and tensor have been explored for the coupling of gravitational field and matter Öttinger (2020b). If we want to keep geodesic motion of a mass point in a gravitational field, however, the coupling should be done in terms of a scalar or tensor quantity that is given in terms of second-derivatives of the metric and vanishes, at least for the static isotropic metric. In this context, the scalar identity (28) holding for the static isotropic solution might be useful. A tensorial coupling could be based on the following identity for the static isotropic solution, $\frac{\partial^{2}g_{\mu\nu}}{\partial x^{\rho}\partial x_{\rho}}+\frac{1}{2}\bar{g}^{\rho\rho^{\prime}}\,\frac{\partial g_{\mu\nu}}{\partial x^{\rho}}\,\frac{\partial g_{\rho^{\prime}\sigma}}{\partial x_{\sigma}}=\frac{2r_{0}^{2}}{(r+2r_{0})r^{3}}\,\eta_{\mu\nu},$ (47) which implies that the trace-free part of the tensor on the left-hand side vanishes. ### VI.5 Comparison to Schwarzschild solution The Schwarzschild solution of general relativity in harmonic coordinates is given by (see, e.g., Eq. (8.2.15) of Weinberg (1972)) $\alpha=\left(1+\frac{r_{0}}{r}\right)^{2},\quad\xi=\frac{r+r_{0}}{r-r_{0}}\,\frac{r_{0}^{2}}{r^{2}},\quad\beta=\frac{r-r_{0}}{r+r_{0}}.$ (48) We here compare this solution to the static isotropic solution (44) of composite gravity. The functions $\alpha$ in Eqs. (44) and (48) differ by the term $r_{0}^{2}/r^{2}$. This term does not matter for the high-precision tests. Whereas the singularity at the origin is $1/r$ for composite gravity, it is $1/r^{2}$ for general relativity. This observation goes nicely with the exponent $x$ in Eqs. (38), (39), for $\tilde{g}=2$ and $\tilde{g}=1$, respectively, where general relativity corresponds to the torsion-free case $\tilde{g}=1$. Whereas $\xi$ vanishes in composite gravity, it has a singularity at $r=r_{0}$ for the Schwarzschild solution, with a jump from $+\infty$ for $r=r_{0}^{+}$ to $-\infty$ for $r=r_{0}^{-}$. While this may be considered as a coordinate singularity in general relativity, this would not be possible for a theory in Minkowski space. For the high-precision tests, the absence of a $1/r$ contribution to $\xi$ is crucial. Also $\beta$ is remarkably different for the two solutions. Whereas $\beta$ is non-negative in composite gravity, it changes sign at $r_{0}$ for the Schwarzschild solution. Although the two solutions look so different, their truncated third-order expansions (42) coincide. The coincidence of these expansions to order $1/r^{2}$ is crucial for satisfying the high-precision tests. ## VII Summary and conclusions Yang-Mills theories are formulated on a background Minkowski space, and so is the composite theory of gravity that selects a small subset of solutions from the Yang-Mills theory based on the Lorentz group. Such theories are covariant under Lorentz transformations but, unlike general relativity, not under general coordinate transformations. Therefore, it is important to characterize the coordinate systems, in which the composite theory of gravity should be valid, by coordinate conditions. We here propose the Lorentz covariant third- order equations (26) for the metric as appealing coordinate conditions that nicely supplement the third-oder differential equations for the composite theory of gravity. Their alternative formulation in Eqs. (24) and (27) shows that we essentially introduce a potential for the divergence of the metric, where the potential itself satisfies a second-order differential equation. In the original work on composite gravity Öttinger (2020a), no general coordinate conditions were given. The static isotropic solution was determined for quasi-Minkowskian coordinates, which are defined only for solutions of this particular type and do not satisfy the new coordinate conditions. Also the coordinate conditions previously used in the complete Hamiltonian formulation of the linearized theory, or weak-field approximation, of composite gravity Öttinger (2020b) differ from the present proposal. Therefore, previous results are qualitatively similar but quantitatively different from our previous results. The coordinate conditions (26) complete the nonlinear theory of pure composite gravity proposed in Öttinger (2020a). The field equations for pure composite gravity can be expressed in a number of different ways. One option is to solve the field equations of the Yang-Mills theory based on the Lorentz group and, in a post-processing step, select those solutions that can be properly expressed in terms of the derivatives of the tetrad variables obtained by decomposing the metric. Alternatively, one can introduce a gauge-invariant connection with torsion and formulate second-order differential equations entirely in terms of those. One is then interested in the solutions for the connection that can be properly expressed in terms of first derivatives of the metric. A final possibility is to write third-order evolution equations directly for the metric. In the various formulations of the field equations, it is difficult to count the number of degrees of freedom of composite gravity. This difficulty is a consequence of the primary constraints arising from the composition rule of composite theories and serving as a selection principle for the relevant solutions of the underlying Yang-Mills theory. A canonical Hamiltonian formulation on the combined spaces of tetrad and Yang-Mills variables provides the most structured form of both field equations and coordinate conditions. This formulation suggests that composite gravity has four degrees of freedom (whereas the Yang-Mills theory based on the Lorentz group has $24$ degrees of freedom). The Hamiltonian formulation suggests that we deal with two types of constraints: (i) constraints resulting from the composition rule and (ii) gauge constraints. As the former can be handled by Dirac brackets Dirac (1950, 1958a, 1958b) and the latter by the BRST methodology (the acronym derives from the names of the authors of the original papers Becchi _et al._ (1976); Tyutin (1975); see also Nemeschansky _et al._ (1988); Öttinger (2018b)), the path to quantization of composite gravity is clear. This is a major advantage of an approach starting from the class of Yang-Mills theories, which so successfully describe electro-weak and strong interactions and for which quantization is perfectly understood, and imposing Dirac-type constraints. In addition, this background reveals why composite theories, although they are higher derivative theories, are not prone to Ostrogradsky instabilities. The fact that just a few degrees of freedom of the Yang-Mills theory based on the Lorentz group survive in the composite theory of gravity is also reflected in its static isotropic solutions. Its Robertson expansion has two free dimensionless parameters in addition to the Yang-Mills coupling constant. For reproducing the high-precison predictios of general relativity, one of the free parameters and the coupling constant ($\tilde{g}=2$) need to be fixed. The remaining dimensionless parameter can be chosen to avoid singularities at finite distances from the origin. A closed-form solution for the static isotropic metric, which plays the same role in composite gravity as the Schwarzschild solution in general relativity, has been found. The solution displays a $1/r$ singularity at the origin but remains finite at all finite values of $r$. The only remarkable feature is $g_{00}=0$ at a particular distance from the origin, which is of the order of the Schwarzschild radius; for all other values of $r$, we have $g_{00}<0$. This paper develops only the pure theory of gravity. The coupling to matter still needs to be elaborated. For the linearized composite theory of gravity, we had proposed scalar and tensorial coupling mechanisms Öttinger (2020b). As the curvature tensor for the static isotropic metric no longer vanishes for the nonlinear theory, which would lead to a deviation from geodesic motion for a coupling based on the curvature tensor, an alternative scalar [see, e.g., Eq. (28)] or tensor [see, e.g., Eq. (47)] must be identified for the coupling of the gravitational field to the energy-momentum tensor of matter. ###### Acknowledgements. I am grateful for the opportunity to do this work during my sabbatical at the _Collegium Helveticum_ in Zürich. ## Appendix A Relation between covariant derivatives The reformulation of equations for the Yang-Mills theory based on the Lorentz group in the metric language is based on the identity $f_{(\kappa\lambda)}^{bc}B_{b}C_{c}=\eta^{\kappa^{\prime}\lambda^{\prime}}\Big{[}B_{(\kappa^{\prime}\lambda)}C_{(\kappa\lambda^{\prime})}-C_{(\kappa^{\prime}\lambda)}B_{(\kappa\lambda^{\prime})}\Big{]},$ (49) which, in view of the definition (4), can be rewritten in the alternative form ${b^{\kappa}}_{\mu}{b^{\lambda}}_{\nu}f_{(\kappa\lambda)}^{bc}B_{b}C_{c}=\bar{g}^{\rho\sigma}\Big{(}\tilde{B}_{\rho\mu}\tilde{C}_{\sigma\nu}+\tilde{B}_{\rho\nu}\tilde{C}_{\mu\sigma}\Big{)}.$ (50) These remarkably simple identities follow from the form of the structure constants of the Lorentz group. After writing the structure constants in the following explicit form (see Table 1 for the index conventions), $\displaystyle f^{abc}$ $\displaystyle=$ $\displaystyle\eta^{\kappa_{a}\lambda_{c}}\eta^{\kappa_{b}\lambda_{a}}\eta^{\kappa_{c}\lambda_{b}}-\eta^{\kappa_{a}\lambda_{b}}\eta^{\kappa_{b}\lambda_{c}}\eta^{\kappa_{c}\lambda_{a}}$ (51) $\displaystyle+$ $\displaystyle\eta^{\kappa_{a}\kappa_{b}}\big{(}\eta^{\kappa_{c}\lambda_{a}}\eta^{\lambda_{b}\lambda_{c}}-\eta^{\kappa_{c}\lambda_{b}}\eta^{\lambda_{a}\lambda_{c}}\big{)}$ $\displaystyle+$ $\displaystyle\eta^{\kappa_{a}\kappa_{c}}\big{(}\eta^{\kappa_{b}\lambda_{c}}\eta^{\lambda_{a}\lambda_{b}}-\eta^{\kappa_{b}\lambda_{a}}\eta^{\lambda_{b}\lambda_{c}}\big{)}$ $\displaystyle+$ $\displaystyle\eta^{\kappa_{b}\kappa_{c}}\big{(}\eta^{\kappa_{a}\lambda_{b}}\eta^{\lambda_{a}\lambda_{c}}-\eta^{\kappa_{a}\lambda_{c}}\eta^{\lambda_{a}\lambda_{b}}\big{)},$ the result (49) is obtained by straightforward calculation. We can now use Eq. (50) to evaluate the right-hand side of Eq. (5), $\displaystyle{b^{\kappa}}_{\mu}{b^{\lambda}}_{\nu}\left[\frac{\partial X_{(\kappa\lambda)}}{\partial x^{\rho^{\prime}}}+\tilde{g}\,f_{(\kappa\lambda)}^{bc}A_{b\rho^{\prime}}X_{c}\right]$ $\displaystyle=$ $\displaystyle\frac{\partial\tilde{X}_{\mu\nu}}{\partial x^{\rho^{\prime}}}$ (52) $\displaystyle+\,\bar{g}^{\rho\sigma}\bigg{[}\left(\tilde{g}\tilde{A}_{\rho\mu\rho^{\prime}}-{b^{\kappa}}_{\rho}\,\frac{\partial b_{\kappa\mu}}{\partial x^{\rho^{\prime}}}\right)\tilde{X}_{\sigma\nu}$ $\displaystyle+\,\left(\tilde{g}\tilde{A}_{\rho\nu\rho^{\prime}}-{b^{\kappa}}_{\rho}\,\frac{\partial b_{\kappa\nu}}{\partial x^{\rho^{\prime}}}\right)\tilde{X}_{\mu\sigma}\bigg{]}.\qquad$ By using the composition rule (2) we recover the fundamental relationship (5) with the definition (6) of the connection following from $\bar{\Gamma}_{\mu\rho\nu}={b^{\kappa}}_{\mu}\,\frac{\partial b_{\kappa\nu}}{\partial x^{\rho}}-\tilde{g}\tilde{A}_{\mu\nu\rho}.$ (53) ## Appendix B Alternative expression for field tensor From the definitions (3) and (4) and the fundamental relations (5) and (50), we obtain $\displaystyle\tilde{F}_{\mu\nu\mu^{\prime}\nu^{\prime}}$ $\displaystyle=$ $\displaystyle\frac{\partial\tilde{A}_{\mu\nu\nu^{\prime}}}{\partial x^{\mu^{\prime}}}-\Gamma^{\sigma}_{\mu^{\prime}\mu}\tilde{A}_{\sigma\nu\nu^{\prime}}+\Gamma^{\sigma}_{\mu^{\prime}\nu}\tilde{A}_{\sigma\mu\nu^{\prime}}$ (54) $\displaystyle-$ $\displaystyle\frac{\partial\tilde{A}_{\mu\nu\mu^{\prime}}}{\partial x^{\nu^{\prime}}}+\Gamma^{\sigma}_{\nu^{\prime}\mu}\tilde{A}_{\sigma\nu\mu^{\prime}}-\Gamma^{\sigma}_{\nu^{\prime}\nu}\tilde{A}_{\sigma\mu\mu^{\prime}}$ $\displaystyle-$ $\displaystyle\tilde{g}\,\bar{g}^{\rho\sigma}\Big{(}\tilde{A}_{\rho\mu\mu^{\prime}}\tilde{A}_{\sigma\nu\nu^{\prime}}-\tilde{A}_{\rho\nu\mu^{\prime}}\tilde{A}_{\sigma\mu\nu^{\prime}}\Big{)}.\qquad$ By means of Eq. (53), we obtain $\displaystyle\tilde{g}\left(\frac{\partial\tilde{A}_{\mu\nu\nu^{\prime}}}{\partial x^{\mu^{\prime}}}-\frac{\partial\tilde{A}_{\mu\nu\mu^{\prime}}}{\partial x^{\nu^{\prime}}}\right)$ $\displaystyle=$ $\displaystyle\frac{\partial\bar{\Gamma}_{\mu\mu^{\prime}\nu}}{\partial x^{\nu^{\prime}}}-\frac{\partial\bar{\Gamma}_{\mu\nu^{\prime}\nu}}{\partial x^{\mu^{\prime}}}$ (55) $\displaystyle+\,\frac{\partial{b^{\kappa}}_{\mu}}{\partial x^{\mu^{\prime}}}\frac{\partial b_{\kappa\nu}}{\partial x^{\nu^{\prime}}}-\frac{\partial{b^{\kappa}}_{\mu}}{\partial x^{\nu^{\prime}}}\frac{\partial b_{\kappa\nu}}{\partial x^{\mu^{\prime}}},$ and, again Eq. (53), gives $\frac{\partial{b^{\kappa}}_{\mu}}{\partial x^{\mu^{\prime}}}\frac{\partial b_{\kappa\nu}}{\partial x^{\nu^{\prime}}}=\bar{g}^{\rho\sigma}(\bar{\Gamma}_{\rho\mu^{\prime}\mu}+\tilde{g}\tilde{A}_{\rho\mu\mu^{\prime}})(\bar{\Gamma}_{\sigma\nu^{\prime}\nu}+\tilde{g}\tilde{A}_{\sigma\nu\nu^{\prime}}).$ (56) By combining Eqs. (54)–(56), we finally arrive at $\tilde{F}_{\mu\nu\mu^{\prime}\nu^{\prime}}=\frac{1}{\tilde{g}}\bigg{(}\frac{\partial\bar{\Gamma}_{\mu\mu^{\prime}\nu}}{\partial x^{\nu^{\prime}}}-\frac{\partial\bar{\Gamma}_{\mu\nu^{\prime}\nu}}{\partial x^{\mu^{\prime}}}+\bar{\Gamma}_{\sigma\mu^{\prime}\mu}\Gamma^{\sigma}_{\nu^{\prime}\nu}-\bar{\Gamma}_{\sigma\nu^{\prime}\mu}\Gamma^{\sigma}_{\mu^{\prime}\nu}\bigg{)}.$ (57) This expression for the field tensor coincides with the one given in Eq. (10) when the definition (6) of the connection is used. ## Appendix C Field equation for connection By inserting the expression (9) for the Riemann curvature tensor in terms of the connection, the field equation (22) for the composite theory of gravity can be written as a second-order differential equation for the connection, $\displaystyle\frac{\partial^{2}\Gamma^{\mu}_{\mu^{\prime}\nu}}{\partial x_{\rho}\partial x^{\rho}}$ $\displaystyle-$ $\displaystyle\frac{\partial^{2}\Gamma^{\mu}_{\rho\nu}}{\partial x_{\rho}\partial x^{\mu^{\prime}}}+\eta^{\rho\rho^{\prime}}\bigg{[}\Gamma^{\sigma}_{\mu^{\prime}\nu}\frac{\partial\Gamma^{\mu}_{\rho\sigma}}{\partial x^{\rho^{\prime}}}-\Gamma^{\mu}_{\mu^{\prime}\sigma}\frac{\partial\Gamma^{\sigma}_{\rho\nu}}{\partial x^{\rho^{\prime}}}$ $\displaystyle+\,\Gamma^{\mu}_{\rho\sigma}\bigg{(}2\frac{\partial\Gamma^{\sigma}_{\mu^{\prime}\nu}}{\partial x^{\rho^{\prime}}}-\frac{\partial\Gamma^{\sigma}_{\rho^{\prime}\nu}}{\partial x^{\mu^{\prime}}}\bigg{)}-\Gamma^{\sigma}_{\rho\nu}\bigg{(}2\frac{\partial\Gamma^{\mu}_{\mu^{\prime}\sigma}}{\partial x^{\rho^{\prime}}}-\frac{\partial\Gamma^{\mu}_{\rho^{\prime}\sigma}}{\partial x^{\mu^{\prime}}}\bigg{)}\quad\;\;$ $\displaystyle+\,\Gamma^{\mu}_{\rho^{\prime}\sigma}\Gamma^{\sigma}_{\rho\sigma^{\prime}}\Gamma^{\sigma^{\prime}}_{\mu^{\prime}\nu}+\Gamma^{\mu}_{\mu^{\prime}\sigma}\Gamma^{\sigma}_{\rho\sigma^{\prime}}\Gamma^{\sigma^{\prime}}_{\rho^{\prime}\nu}-2\Gamma^{\mu}_{\rho\sigma}\Gamma^{\sigma}_{\mu^{\prime}\sigma^{\prime}}\Gamma^{\sigma^{\prime}}_{\rho^{\prime}\nu}\bigg{]}=0.$ Note that $\eta^{\rho\rho^{\prime}}$ occurs rather than $\bar{g}^{\rho\rho^{\prime}}$, so that there is no need to know the metric for solving this equation. ## Appendix D Modified Lagrangian The Lagrangian for a pure Yang-Mills theory, including a covariant but gauge breaking term for removing degeneracies associated with gauge invariance (the particular form corresponds to the convenient Feynman gauge), is given by $L=-\int\left(\frac{1}{4}F^{a}_{\mu\nu}F_{a}^{\mu\nu}+\frac{1}{2}\frac{\partial A^{a}_{\mu}}{\partial x_{\mu}}\frac{\partial A_{a}^{\nu}}{\partial x^{\nu}}\right)d^{3}x.$ (59) We propose to add the further term $L_{\rm cc}=\frac{1}{2}\int\left(\frac{\partial^{2}g_{\mu\nu}}{\partial x_{\mu}\partial x^{\sigma}}\frac{\partial^{2}{g_{\rho}}^{\nu}}{\partial x_{\rho}\partial x_{\sigma}}-K\frac{\partial^{2}g_{\mu\nu}}{\partial x_{\mu}\partial x_{\nu}}\frac{\partial^{2}g_{\rho\sigma}}{\partial x_{\rho}\partial x_{\sigma}}\right)d^{3}x,$ (60) implying the functional derivative $\frac{\delta L_{\rm cc}}{\delta g_{\mu\nu}}=\frac{\partial}{\partial x_{\mu}}\left(\square\frac{\partial{g_{\rho}}^{\nu}}{\partial x_{\rho}}-K\frac{\partial}{\partial x_{\nu}}\frac{\partial^{2}g_{\rho\sigma}}{\partial x_{\rho}\partial x_{\sigma}}\right),$ (61) which vanishes upon imposing the coordinate conditions (26) as constraints. If the gauge conditions and the coordinate conditions are imposed as constraints, the above modifications of the Lagrangian for the pure Yang-Mills theory have no effect on the field equations. ## Appendix E Hamiltonian formulation For the weak-field approximation of composite gravity, a canonical Hamiltonian formulation with a detailed analysis of all constraints has been given in Öttinger (2020b). We here sketch how that approach can be generalized to a full, nonlinear theory of pure gravity selected from the Yang-Mills theory based on the Lorentz group. The underlying space of the Hamiltonian formulation consists of the tetrad variables ${b^{\kappa}}_{\mu}$ and the gauge vector fields $A_{a\nu}$ associated with the Lorentz group as configurational variables, together with their conjugate momenta ${p_{\kappa}}^{\mu}$ and $E^{a\nu}$ (where $E_{aj}=F_{aj0}$ and $E_{a0}=\partial A_{a\mu}/\partial x_{\mu}$) Öttinger (2019, 2020b). This space consists of $80$ fields, but massive constraints arise from the composition rule and gauge invariance so that, in the end, the composite theory of pure gravity turns out to possess only four degrees of freedom. The generalization of the Hamiltonian (25)–(27) of Öttinger (2020b) is obtained by introducing the Hamiltonian for the full, nonlinear version of Yang-Mills theory, $\displaystyle H_{\rm pure}$ $\displaystyle=$ $\displaystyle\int\bigg{[}\frac{1}{2}E^{a\mu}E_{a\mu}+\frac{1}{4}F_{aij}F^{aij}-E^{a0}\frac{\partial A_{aj}}{\partial x_{j}}$ (62) $\displaystyle-\,E^{aj}\left(\frac{\partial A_{a0}}{\partial x^{j}}+\tilde{g}f_{a}^{bc}A_{bj}A_{c0}\right)+\mbox{$\dot{b}^{\kappa}$}_{\mu}\,{p_{\kappa}}^{\mu}\bigg{]}d^{3}x,\qquad$ where the functional form of the $16$ time derivatives $\mbox{$\dot{b}^{\kappa}$}_{\mu}$ in terms of the configurational variables ${b^{\kappa}}_{\mu}$ and $A_{a\nu}$ is obtained from $12$ components of the composition rule (2) and the four coordinate conditions (24) (the potential $\phi$ is assumed to be a functional of $g_{\mu\nu}$). For pure gravity without external sources, we can impose the $16$ constraints ${p_{\kappa}}^{\mu}=0$ so that the composite theory consists of selected solutions of the Yang-Mills theory based on the Lorentz group Öttinger (2019, 2020b). The terms involving $E^{a0}$ in the Hamiltonian (62) are associated with the gauge breaking term in the Lagrangian (59). Of course, this Hamiltonian implies the canonical evolution equations for the entire set of $80$ fields. The generalization of the weak-field approximation becomes particularly simple if we introduce the following variables eliminating the nonlinear effects of the coupling constant, $\breve{A}_{\mu\nu\rho}=\tilde{A}_{\mu\nu\rho}-\frac{1}{2\tilde{g}}\,\Omega_{\mu\nu/\rho},$ (63) with $\Omega_{\mu\nu/\rho}={b^{\kappa}}_{\mu}\,\frac{\partial b_{\kappa\nu}}{\partial x^{\rho}}-\frac{\partial{b^{\kappa}}_{\mu}}{\partial x^{\rho}}\,b_{\kappa\nu},$ (64) and $\breve{E}_{\mu\nu 0}=\tilde{E}_{\mu\nu 0}-\frac{\eta^{\rho\rho^{\prime}}}{2\tilde{g}}\Big{(}\Gamma^{\sigma}_{\rho\nu}\bar{\Gamma}_{\mu\rho^{\prime}\sigma}-\Gamma^{\sigma}_{\rho\mu}\bar{\Gamma}_{\nu\rho^{\prime}\sigma}\Big{)},$ (65) $\breve{E}_{\mu\nu j}=\tilde{E}_{\mu\nu j}-\frac{1}{\tilde{g}}\bigg{(}\Gamma^{\sigma}_{j\mu}\bar{\Gamma}_{\sigma 0\nu}-\Gamma^{\sigma}_{j\nu}\bar{\Gamma}_{\sigma 0\mu}\bigg{)},$ (66) as further modifications of the variables $\tilde{A}_{\mu\nu\rho}$ and $\tilde{E}_{\mu\nu\rho}$ defined in Eq. (4). For example, the composition rule (2) takes the linear form $\breve{A}_{\mu\nu\rho}=\frac{1}{2}\left(\frac{\partial g_{\nu\rho}}{\partial x^{\mu}}-\frac{\partial g_{\mu\rho}}{\partial x^{\nu}}\right),$ (67) which corresponds to Eq. (7) of Öttinger (2020b) in the symmetric gauge and includes $12$ primary constraints. Also the evolution equations for $\breve{A}_{\mu\nu\rho}$ and hence also the $12$ secondary constraints keep the same form as in the linearized theory (cf. Eqs. (39), (40) and (46), (47) of Öttinger (2020b)). The $12$ tertiary constraints can be obtained by acting with the operator $\square$ on the primary constraints. The invariance of the tertiary constraints follows from ${p_{\kappa}}^{\mu}=0$. In order to verify the above statements, one needs the identity $\frac{\partial\Omega_{\mu\nu/\rho}}{\partial x_{\rho}}-\eta^{\rho\rho^{\prime}}\left(\Gamma^{\sigma}_{\rho\mu}\Omega_{\sigma\nu/\rho^{\prime}}+\Gamma^{\sigma}_{\rho\nu}\Omega_{\mu\sigma/\rho^{\prime}}\right)=0,$ (68) which is the counterpart of Eq. (16) of Öttinger (2020b) and can be inferred from the gauge invariance of the left-hand side of Eq. (68). Finally, the $24$ evolution equations for $\breve{E}_{\mu\nu\rho}$ correspond to the field equations given in various forms in Sec. IV. As the structure of the Hamiltonian and the constraints for the full, nonlinear theory is so similar (mostly even formally identical) to the case of the linear weak-field approximation, we expect the same count of $24+3\cdot 12+16=76$ constraints for $2\cdot(16+24)=80$ variables. Half of the $24$ constraints associated with gauge invariance result from the gauge conditions $E_{a0}=\partial A_{a\mu}/\partial x_{\mu}=0$, which establish a relationship between the (unphysical) temporal and longitudinal modes of the four-vector potentials. The above arguments suggest that pure composite gravity possesses (at least) four physical degrees of freedom, just as in the thoroughly elaborated special case of the weak-field approximation Öttinger (2020b). ## Appendix F A cubic equation The coefficients $\alpha_{1}$ and $\xi_{1}$ in the Robertson expansions (33), (34) are related by the following cubic equation, $\displaystyle 10\xi_{1}^{3}$ $\displaystyle+$ $\displaystyle 10\tilde{g}\xi_{1}(4\alpha_{1}^{2}+5\alpha_{1}\xi_{1}+2\xi_{1}^{2})$ (69) $\displaystyle-$ $\displaystyle 5\tilde{g}^{2}\big{[}4\xi_{1}-(\alpha_{1}+\xi_{1})(8\alpha_{1}^{2}+9\alpha_{1}\xi_{1}-\xi_{1}^{2})\big{]}$ $\displaystyle-$ $\displaystyle 5\tilde{g}^{3}\big{[}4+3(\alpha_{1}+\xi_{1})^{2}\big{]}(3\alpha_{1}+2\xi_{1})$ $\displaystyle+$ $\displaystyle\tilde{g}^{4}(\alpha_{1}+\xi_{1})(36+11(\alpha_{1}+\xi_{1})^{2})=0.$ Its only real solution for $\alpha_{1}$ in terms of $\xi_{1}$ is given by $\displaystyle\alpha_{1}$ $\displaystyle=$ $\displaystyle\Big{[}\Big{(}w_{3}+\sqrt{w_{3}^{2}-w_{2}^{3}}\Big{)}^{1/3}+w_{2}\Big{(}w_{3}+\sqrt{w_{3}^{2}-w_{2}^{3}}\Big{)}^{-1/3}$ $\displaystyle-\,\xi_{1}(40+85\tilde{g}-120\tilde{g}^{2}+33\tilde{g}^{3})\Big{]}/\big{[}3\tilde{g}(40-45\tilde{g}+11\tilde{g}^{2})\big{]},$ with $\displaystyle w_{2}$ $\displaystyle=$ $\displaystyle 36\tilde{g}^{3}\big{(}200-345\tilde{g}+190\tilde{g}^{2}-33\tilde{g}^{3}\big{)}$ (71) $\displaystyle+$ $\displaystyle 5\big{(}320+160\tilde{g}-85\tilde{g}^{2}-282\tilde{g}^{3}+111\tilde{g}^{4}\big{)}\xi_{1}^{2},\qquad$ and $\displaystyle w_{3}$ $\displaystyle=$ $\displaystyle-5\xi_{1}\Big{[}108\tilde{g}^{4}\big{(}520-985\tilde{g}+633\tilde{g}^{2}-155\tilde{g}^{3}+11\tilde{g}^{4})$ (72) $\displaystyle+$ $\displaystyle\big{(}12800+52800\tilde{g}-82200\tilde{g}^{2}-10735\tilde{g}^{3}$ $\displaystyle+63045\tilde{g}^{4}-33273\tilde{g}^{5}+5427\tilde{g}^{6}\big{)}\xi_{1}^{2}\Big{]}.$ ## References * Yang and Mills (1954) C. N. Yang and R. L. Mills, “Conservation of isotopic spin and isotopic gauge invariance,” Phys. Rev. 96, 191–195 (1954). * Utiyama (1956) R. Utiyama, “Invariant theoretical interpretation of interaction,” Phys. Rev. 101, 1597–1607 (1956). * Yang (1974) C. N. Yang, “Integral formalism for gauge fields,” Phys. Rev. Lett. 33, 445–447 (1974). * Blagojević and Hehl (2013) M. Blagojević and F. W. Hehl, eds., _Gauge Theories of Gravitation: A Reader with Commentaries_ (Imperial College Press, London, 2013). * Capozziello and De Laurentis (2011) S. Capozziello and M. De Laurentis, “Extended theories of gravity,” Phys. Rep. 509, 167–321 (2011). * Ivanenko and Sardanashvily (1983) D. Ivanenko and G. Sardanashvily, “The gauge treatment of gravity,” Phys. Rep. 94, 1–45 (1983). * Öttinger (2018a) H. C. Öttinger, “Hamiltonian formulation of a class of constrained fourth-order differential equations in the Ostrogradsky framework,” J. Phys. Commun. 2, 125006 (2018a). * Öttinger (2019) H. C. Öttinger, “Natural Hamiltonian formulation of composite higher derivative theories,” J. Phys. Commun. 3, 085001 (2019). * Öttinger (2020a) H. C. Öttinger, “Composite higher derivative theory of gravity,” Phys. Rev. Research 2, 013190 (2020a). * Giovanelli (2020) M. Giovanelli, “Nothing but coincidences. The point-coincidence argument and Einstein’s struggle with the meaning of coordinates in physics,” Euro. Jnl. Phil. Sci. 10, under review (2020). * Öttinger (2020b) H. C. Öttinger, “Mathematical structure and physical content of composite gravity in weak-field approximation,” Phys. Rev. D 102, 064024 (2020b). * Peskin and Schroeder (1995) M. E. Peskin and D. V. Schroeder, _An Introduction to Quantum Field Theory_ (Perseus Books, Reading, MA, 1995). * Weinberg (2005) S. Weinberg, _Modern Applications_, The Quantum Theory of Fields, Vol. 2 (Cambridge University Press, Cambridge, 2005). * Öttinger (2018b) H. C. Öttinger, “BRST quantization of Yang-Mills theory: A purely Hamiltonian approach on Fock space,” Phys. Rev. D 97, 074006 (2018b). * Jiménez _et al._ (2019) J. B. Jiménez, L. Heisenberg, and T. S. Koivisto, “The geometrical trinity of gravity,” Universe 5, 173 (2019). * Weinberg (1972) S. Weinberg, _Gravitation and Cosmology, Principles and Applications of the General Theory of Relativity_ (Wiley, New York, 1972). * Ostrogradsky (1850) M. Ostrogradsky, “Mémoires sur les équations différentielles, relatives au problème des isopérimètres,” Mem. Acad. St. Petersbourg 6, 385–517 (1850). * Woodard (2015) R. P. Woodard, “Ostrogradsky’s theorem on Hamiltonian instability,” Scholarpedia 10, 32243 (2015). * j. Chen _et al._ (2013) T. j. Chen, M. Fasiello, E. A. Lim, and A. J. Tolley, “Higher derivative theories with constraints: Exorcising Ostrogradski’s ghost,” J. Cosmol. Astropart. Phys. 02, 042 (2013). * Raidal and Veermäe (2017) M. Raidal and H. Veermäe, “On the quantisation of complex higher derivative theories and avoiding the Ostrogradsky ghost,” Nucl. Phys. B 916, 607–626 (2017). * Stelle (1977) K. S. Stelle, “Renormalization of higher-derivative quantum gravity,” Phys. Rev. D 16, 953–969 (1977). * Stelle (1978) K. S. Stelle, “Classical gravity with higher derivatives,” Gen. Relat. Gravit. 9, 353–371 (1978). * Krasnikov (1987) N. V. Krasnikov, “Nonlocal gauge theories,” Theor. Math. Phys. 73, 1184–1190 (1987). * Grosse-Knetter (1994) C. Grosse-Knetter, “Effective Lagrangians with higher derivatives and equations of motion,” Phys. Rev. D 49, 6709–6719 (1994). * Becker _et al._ (2017) D. Becker, C. Ripken, and F. Saueressig, “On avoiding Ostrogradski instabilities within asymptotic safety,” J. High Energy Phys. 12, 121 (2017). * Salvio (2019) A. Salvio, “Metastability in quadratic gravity,” Phys. Rev. D 99, 103507 (2019). * Breuer and Petruccione (2002) H.-P. Breuer and F. Petruccione, _The Theory of Open Quantum Systems_ (Oxford University Press, Oxford, 2002). * Weiss (2008) U. Weiss, _Quantum Dissipative Systems_ , 3rd ed., Series in Modern Condensed Matter Physics, Volume 13 (World Scientific, Singapore, 2008). * Öttinger (2011) H. C. Öttinger, “The geometry and thermodynamics of dissipative quantum systems,” Europhys. Lett. 94, 10006 (2011). * Taj and Öttinger (2015) D. Taj and H. C. Öttinger, “Natural approach to quantum dissipation,” Phys. Rev. A 92, 062128 (2015). * Öttinger (2017) H. C. Öttinger, _A Philosophical Approach to Quantum Field Theory_ (Cambridge University Press, Cambridge, 2017). * Oldofredi and Öttinger (2021) A. Oldofredi and H. C. Öttinger, “The dissipative approach to quantum field theory: Conceptual foundations and ontological implications,” Euro. Jnl. Phil. Sci. 11, 18 (2021). * Dirac (1950) P. A. M. Dirac, “Generalized Hamiltonian dynamics,” Canad. J. Math. 2, 129–148 (1950). * Dirac (1958a) P. A. M. Dirac, “Generalized Hamiltonian dynamics,” Proc. Roy. Soc. A 246, 326–332 (1958a). * Dirac (1958b) P. A. M. Dirac, “The theory of gravitation in Hamiltonian form,” Proc. Roy. Soc. A 246, 333–343 (1958b). * Becchi _et al._ (1976) C. Becchi, A. Rouet, and R. Stora, “Renormalization of gauge theories,” Ann. Phys. (N.Y.) 98, 287–321 (1976). * Tyutin (1975) I. V. Tyutin, “Gauge invariance in field theory and statistical physics in operator formalism,” (1975), preprint of P. N. Lebedev Physical Institute, No. 39, 1975, arXiv:0812.0580. * Nemeschansky _et al._ (1988) D. Nemeschansky, C. Preitschopf, and M. Weinstein, “A BRST primer,” Ann. Phys. (N.Y.) 183, 226–268 (1988).
See Tesi_RossiP-twoside-frn.pdf ## Abstract Equivariant localization theory is a powerful tool that has been extensively used in the past thirty years to elegantly obtain exact integration formulas, in both mathematics and physics. These integration formulas are proved within the mathematical formalism of equivariant cohomology, a variant of standard cohomology theory that incorporates the presence of a symmetry group acting on the space at hand. A suitable infinite-dimensional generalization of this formalism is applicable to a certain class of Quantum Field Theories (QFT) endowed with _supersymmetry_. In this thesis we review the formalism of equivariant localization and some of its applications in Quantum Mechanics (QM) and QFT. We start from the mathematical description of equivariant cohomology and related localization theorems of finite-dimensional integrals in the case of an Abelian group action, and then we discuss their formal application to infinite-dimensional path integrals in QFT. We summarize some examples from the literature of computations of partition functions and expectation values of supersymmetric operators in various dimensions. For 1-dimensional QFT, that is QM, we review the application of the localization principle to the derivation of the Atiyah- Singer index theorem applied to the Dirac operator on a twisted spinor bundle. In 3 and 4 dimensions, we examine the computation of expectation values of certain Wilson loops in supersymmetric gauge theories and their relation to 0-dimensional theories described by “matrix models”. Finally, we review the formalism of non-Abelian localization applied to 2-dimensional Yang-Mills theory and its application in the mapping between the standard “physical” theory and a related “cohomological” formulation. ## Acknowledgments First and foremost, I would like to express my sincere gratitude to professor Diego Trancanelli, who supervised me during the draft of this thesis. I thank him for his constant availability, his valuable advice and his sincere interest for my learning process, as well as his great patience in meticulously reviewing my work step by step. It is also a pleasure for me to thank professor Olindo Corradini, who initiated me to the wonderful world of QFT with two brilliant courses, and was always available to discuss and answer my questions. I would like to thank all the friends and colleagues that grew up with me in this journey. Even though our paths and interests separated, they have always been a precious source of inspiration to me. A heartfelt thanks goes to Giulia, whose presence alone makes everything easier. Finally I would like to thank my family, for the love and support, no matter what, during all these years. ###### Contents 1. Abstract 2. 1 Introduction 3. 2 Equivariant cohomology 1. 2.1 A brief review of standard cohomology theory 2. 2.2 Group actions and equivariant cohomology 3. 2.3 The Weil model and equivariant de Rham’s theorem 4. 2.4 The Cartan model 5. 2.5 The BRST model 4. 3 Localization theorems in finite-dimensional geometry 1. 3.1 Equivariant localization principle 2. 3.2 The ABBV localization formula for Abelian actions 3. 3.3 Equivariant cohomology on symplectic manifolds 1. 3.3.1 Pills of symplectic geometry 2. 3.3.2 Equivariant cohomology for Hamiltonian systems 5. 4 Supergeometry and supersymmetry 1. 4.1 Gradings and superspaces 1. 4.1.1 Definitions 2. 4.1.2 Integration 2. 4.2 Supergeometric proof of ABBV formula for a circle action 3. 4.3 Introduction to Poincaré-supersymmetry 1. 4.3.1 Super-Poincaré algebra and superspace 2. 4.3.2 Chiral superspace and superfields 3. 4.3.3 Supersymmetric actions and component field expansion 4. 4.3.4 R-symmetry 5. 4.3.5 Supersymmetry multiplets 6. 4.3.6 Euclidean 3d N=2 supersymmetric gauge theories 7. 4.3.7 Euclidean 4d N=4,2,2* gauge theories 4. 4.4 From flat to curved space 1. 4.4.1 Coupling to background SUGRA 2. 4.4.2 Supercurrent multiplets and metric multiplets 3. 4.4.3 N=2 gauge theories on the round 3-sphere 4. 4.4.4 N=4,2,2* gauge theories on the round 4-sphere 5. 4.4.5 Trial and error method 5. 4.5 BRST cohomology and equivariant cohomology 6. 5 Localization for circle actions in supersymmetric QFT 1. 5.1 Localization principle in Hamiltonian QM 2. 5.2 Localization and index theorems 3. 5.3 Equivariant structure of supersymmetric QFT and supersymmetric localization principle 4. 5.4 Localization of N=4,2,2* gauge theory on the 4-sphere 1. 5.4.1 The action and the supersymmetric Wilson loop 2. 5.4.2 Quick localization argument 3. 5.4.3 The equivariant model 4. 5.4.4 Localization formulas 5. 5.4.5 The Matrix Model for N=4 SYM 5. 5.5 Localization of N=2 Chern-Simons theory on the 3-sphere 1. 5.5.1 Matter-coupled N=2 Euclidean SCS theory on the 3-sphere 2. 5.5.2 The supersymmetric Wilson loop 3. 5.5.3 Localization: gauge sector 4. 5.5.4 Localization: matter sector 5. 5.5.5 The ABJM matrix model 7. 6 Non-Abelian localization and 2d YM theory 1. 6.1 Prelude: moment maps and YM theory 2. 6.2 A localization formula for non-Abelian actions 3. 6.3 “Cohomological” and “physical” YM theory 4. 6.4 Localization of 2-dimensional YM theory 8. 7 Conclusion 9. A Some differential geometry 1. A.1 Principal bundles, basic forms and connections 2. A.2 Spinors in curved spacetime 10. B Mathematical background on equivariant cohomology 1. B.1 Equivariant vector bundles and equivariant characteristic classes 2. B.2 Universal bundles and equivariant cohomology 3. B.3 Fixed point sets and Borel localization 4. B.4 Equivariant integration and Stokes’ theorem ## Chapter 1 Introduction Quantum Field Theory (QFT) is the framework in which modern theoretical physics describes fundamental interactions between elementary particles and it is also central in the study of condensed matter physics and statistical mechanics. QFT has made the most precise predictions ever in the history of science and it has been tested against a huge amount of experimental data. Nowadays, the most useful formulation of QFT is made in terms of _path integrals_ , which are integrals over the space of all possible field configurations. A “field” mathematically speaking can be thought roughly as a function over spacetime, so these integrals are computed over functional spaces, that are infinite-dimensional. This makes the exact computation of such objects a complicated task, except for some very special cases, and in fact their precise mathematical formulation is still an open problem. Despite these formal difficulties, many interesting results can be extracted from these objects, that can describe _partition functions_ and _expectation values_ of physical observables in QFT. The favorite approach to deal with such computations is perturbation theory, applied in the case in which the QFT is weakly coupled. In this regime, one can compute approximately the expectation values in a perturbative expansion, order by order in the coupling constant. This method, applied to the computation of the partition function, is the infinite-dimensional analogous of a “saddle-point” or “stationary- phase” approximation. Intuitively it represents a semi-classical approach to the quantum dynamics. There are however many cases in which perturbation theory is not applicable, mainly when the QFT is strongly coupled, i.e. the coupling constant is of order 1. This is not a very rare situation. For example we know that one of the fundamental interactions of the Standard Model of particle physics, the strong nuclear force, is well described by “quantum chromodynamics” (QCD), a QFT that is strongly coupled at low energies (so in the “phenomenological” regime). Understanding the behavior of QFT in the strong coupling regime is then a major problem from the physical point of view, and one is lead to develop techniques that permit to study the path integral in a non- perturbative approach. In this thesis we describe features of one of these techniques, that has been exploited in the last few decades for a class of special QFTs, those who exhibit some kind of _supersymmetry_. This technique is called _supersymmetric localization_ , or _equivariant localization_ , or simply localization. Its name derives from the fact that, when one is able to use this method, the path integral of the QFT at hand simplifies (so “localizes”) to an integral over a smaller domain, sometimes even a finite-dimensional integral over constant field configurations. This result can be viewed as an exact stationary-phase approximation, so that the full quantum spectrum of the localized theory is completely determined by its semi-classical limit. Without entering in the technical details of this localization phenomenon, we just point out that this method fundamentally relies on the presence of a large amount of symmetry of the theory, that can be described by the presence of a _group action_ on the space of fields. When the space of fields is graded, i.e. there is a distinction between “bosonic” and “fermionic” degrees of freedom, the symmetry group action can exchange these two types of fields and in this case it is called _super_ symmetry. This situation arises mainly in BRST-fixed and topological field theories, where the grading is regarded as the _ghost number_ , and in Poincaré-supersymmetric QFT, where the grading distinguishes between bosons and fermions in the standard sense of particle physics. In both cases, the action of a supersymmetry transformation “squares” to a canonical (bosonic) one, i.e. a gauge transformation or a Poincaré transformation. It is reasonable that such a huge amount of symmetry can simplify the dynamics of the theory, but it can be non-trivial a priori how to translate this in a simplification of the path integral. At this point, one wishes to understand if there is a theoretical framework that allows to systematically understand why and when such a drastic simplification of the path integral can occur, and at which level this is related to (super-)symmetries in QFT. To be mathematically more rigorous, we can think in terms of integration over finite-dimensional spaces, and then try to extrapolate and generalize the important results to the infinite- dimensional case. Integrals of differential forms over manifolds are built up technically from the smooth (so local) structure of the space, but it is a well-known fact that their result can describe and is regulated by topological (so global) properties. It is a consequence of de Rham’s theorem and Stokes’ theorem that they really depend on the _cohomology class_ of the integrand and not on the particular differential form that represents the class. It is thus reasonable that a theory of integration that embeds the presence of a symmetry group action should arise from a topological construction. Indeed, the mathematical framework in which the localization formulas were firstly derived is a suitable modification of the standard cohomology theory. This is called _equivariant cohomology_. As de Rham’s theorem relates the usual cohomology to differential forms on a smooth manifold, equivariant cohomology can be associated to a modification of them, called _equivariant differential forms_. Equivariant cohomology theory was initiated in the mathematical literature by Cartan, Borel and others during the 50’s [1, 2, 3, 4], but the first instance of a localization formula was presented by Duistermaat and Heckman in 1982 [5]. In this paper, they proved the exactness of a stationary-phase approximation in the context of symplectic geometry and Hamiltonian Abelian group actions. Subsequently, Atiyah and Bott realized that the Duistermaat- Heckman localization formula can be viewed as a special case of a more general theorem, that they proved in the topological language of equivariant cohomology [6]. Almost at the same time, Berline and Vergne derived an analogous localization formula valid for Killing vectors on general compact Riemannian manifolds [7]. Roughly, the Atiyah-Bott-Berline-Vergne (ABBV) formula says that the integral over a manifold that is acted upon by an Abelian group localizes as a sum of contributions arising only from the fixed points of the group action. The first infinite-dimensional generalization of this localization formula was given soon after by Atiyah and Witten in 1985 [8], applied to supersymmetric Quantum Mechanics (QM). This turns out to be an example of topological theory, since their localization formula relates the partition function to the index of a Dirac operator. Many generalizations of this approach followed, first mainly in the context of topological field theories. Here localization allows to get closed formulas relating partition functions of physical QFTs to topological invariants of the spaces where they live. In all these cases, the BRST cohomology is interpreted as the equivariant structure of the theory, and is responsible for the localization of the path integral. In recent years, the formal application of the ABBV localization formula for Abelian symmetry actions was employed in the context of Poincaré- supersymmetric theories, whose explicit supersymmetry is an extension of the spacetime symmetry that is common to all QFTs. In this case, the equivariant structure is generated by the cohomology of a _supersymmetry charge_ , that acts as a differential on the space of Poincaré symmetric field configurations. This formal structure of supersymmetric field theories was firstly realized by Niemi, Palo and Morozov in the 90’s [9, 10]. Starting from the work of Pestun in 2007 [11], localization has been applied to the computation of partition functions and expectation values of supersymmetric operators on curved compact manifolds [12]. In many of these cases, the partition function localizes to a finite-dimensional integral over matrices, a so-called _matrix model_. To carry out this procedure, a number of technical difficulties have to be overcome, the most urgent being to understand how to define Poincaré-supersymmetric theory on curved spaces. Nowadays there is a well-defined and well-understood procedure that allows to do that, essentially deforming an original theory defined in flat space through the coupling to a non-trivial “rigid” supersymmetric background. This in general reduces the degree of supersymmetry of the original theory, but if some of it is preserved on the new background then one is able in principle to perform localization. Beside the general and abstract motivation of understanding strongly coupled QFT, the importance of these computations can be viewed in a string theory perspective, and in particular as possible tests of the so called _AdS/CFT correspondence_ [13]. Some generalizations of the ABBV formula for non-Abelian group actions have been proposed both in the mathematical and physical literature. The first generalization of the Duistermaat-Heckman theorem to non-Abelian group actions was presented by Guillemin and Prato [14], restricting the localization principle to the action of the maximal Abelian subgroup. An infinite- dimensional generalization of the localization principle was proposed by Witten in 1992 [15], applied to the study of 2-dimensional Yang-Mills theory, and a more rigorous proof appeared in the mathematical literature in 1995 by Jeffrey and Kirwan [16]. Of course these localization formulas find many interesting applications not only in physics but also in pure mathematics, but this side of the story is far from the purpose of this work. ### An exact saddle-point approximation To give a feeling of what we mean by “exact saddle-point approximation”, we present a very simple but instructive example here, in the finite-dimensional setting. The saddle-point (or stationary-phase) method is applied to oscillatory integrals of the form $I(t)=\int_{-\infty}^{+\infty}dx\ e^{itf(x)}g(x),$ (1.1) when one is interested in the asymptotic behavior of $I(t)$ at positive large values of the real parameter $t$. In this limit, the integral is dominated by the critical points of $f(x)$, where its first derivative vanishes and it can be expanded in Taylor series as $f(x)=f(x_{0})+\frac{1}{2}f^{\prime\prime}(x_{0})(x-x_{0})^{2}+\cdots.$ (1.2) If $F\subset\mathbb{R}$ is the set of critical points, that for simplicity we assume to be discrete, the leading contribution to (1.1) is then given by a Gaussian integral, $\displaystyle I(t)$ $\displaystyle\approx\sum_{x_{0}\in F}g(x_{0})e^{itf(x_{0})}\int_{-\infty}^{+\infty}dx\ e^{\frac{it}{2}f^{\prime\prime}(x_{0})(x-x_{0})^{2}}$ (1.3) $\displaystyle=\sum_{x_{0}\in F}g(x_{0})e^{itf(x_{0})}e^{\frac{i\pi}{4}\mathrm{sign}(f^{\prime\prime}(x_{0}))}\sqrt{\frac{2\pi}{t|f^{\prime\prime}(x_{0})|}}.$ If the integral is performed over $\mathbb{R}^{n}$, this last formula generalizes easily to $I(t)\approx\left(\frac{2\pi}{t}\right)^{n/2}\sum_{x_{0}\in F}g(x_{0})e^{itf(x_{0})}\frac{e^{\frac{i\pi}{4}\sigma(x_{0})}}{|\det(\mathrm{Hess}_{f}(x_{0}))|^{1/2}},$ (1.4) where $\mathrm{Hess}_{f}(x)$ is the matrix of second derivatives of $f$ at $x\in\mathbb{R}^{n}$, and $\sigma(x)$ denotes the sum of the signs of its eigenvalues. Of course, there is no reason for the RHS of (1.4) to be the exact answer for $I(t)$, but the claimed property of localization is that in some cases this turns out to be true! To see this, let us consider the integration over the 2-sphere $\mathbb{S}^{2}$, defined by its embedding in $\mathbb{R}^{3}$ as the set of points whose distance from the origin is 1. For this example we chose $f(x,y,z)=z$, the “height function”, and $g(x,y,z)=1$. The resulting oscillatory integral is then $I(t)=\int_{\mathbb{S}^{2}}dA\ e^{itz},$ (1.5) where $dA$ is the volume form on the sphere, normalized such that $\int_{\mathbb{S}^{2}}dA=4\pi$. The critical points of the height function are the North and the South poles, where $z\approx\pm\left(1-\frac{1}{2}(x^{2}+y^{2})\right).$ (1.6) The volume form at the poles is just $dA=dxdy$, so if we apply the saddle- point approximation to (1.5) we get $\displaystyle\int_{\mathbb{S}^{2}}dA\ e^{itz}$ $\displaystyle\approx e^{it}\int dxdy\ e^{-\frac{it}{2}(x^{2}+y^{2})}+e^{-it}\int dxdy\ e^{\frac{it}{2}(x^{2}+y^{2})}$ (1.7) $\displaystyle=\frac{2\pi}{it}e^{it}-\frac{2\pi}{it}e^{-it}$ $\displaystyle=4\pi\frac{\sin(t)}{t}.$ Now, since this integral is rather easy, in this case we can actually compare the result of the saddle-point approximation with its exact value. Using spherical coordinates, $I(t)=\int_{-1}^{+1}d\cos(\theta)\int_{0}^{2\pi}d\varphi\ e^{it\cos(\theta)}=4\pi\frac{\sin(t)}{t}.$ (1.8) As promised, the result coincides with the stationary-phase result (1.7). This is the simplest example of equivariant localization! The scope of the first chapters of the thesis is to describe in general the structure underlying this result, how it can be related to symmetry properties of the specific function and space under consideration, and the localization theorems that generalize this specific computation to possibly more complicated examples. In the remaining part, we will deal instead with examples of the infinite-dimensional analog of this exact stationary-phase approximation. ### Structure of the thesis The aim of this project is to summarize some results in the context of equivariant localization applied to physics. We will draw a line from the first mathematical results concerning the theory of equivariant cohomology and associated powerful localization theorems of finite-dimensional integrals, to the generalization to path integration in quantum mechanical models and finally to applications in quantum field theories. One of the purposes of this work is to provide a suitable reference for other students in theoretical and mathematical physics with a background at the level of a master degree, who are interested in approaching the subject. In this spirit, we will try to expose the material in a pedagogical order, being as self-contained as possible, and otherwise giving explicit references to background material. In Chapter 2, we will review the basics of equivariant cohomology theory, starting from its construction in algebraic topology. After having recalled some notions of basic homology and cohomology, we will introduce group actions, and define equivariant cohomology with the so-called _Borel construction_. Then, we will describe the most common algebraic models that generalize de Rham’s theorem in the equivariant setup, the _Weil_ , _Cartan_ and _BRST models_. These give a description of equivariant cohomology in terms of a suitable modification of the complex of differential forms. In Chapter 3, we will describe the common rationale behind the localization property of equivariant integrals in finite-dimensional geometry, the so- called _equivariant localization principle_. Then we will state and explain the Abelian localization formula derived by Atiyah-Bott and Berline-Vergne. In the final part of the chapter, we will connect the discussion with the context of symplectic geometry that, as we will see, can be rephrased in terms of equivariant cohomology. We will review the basic notions of symplectic manifolds, symmetries and Hamiltonian systems, and then state the Duistermaat- Heckman localization formula as a special case of the ABBV theorem. Chapter 4 can be viewed as a long technical aside. Here we will review some notions about supergeometry that are needed to understand a proof of the ABBV theorem, and then specialize the discussion to Poincaré-supersymmetric theories. We will discuss their construction from the perspective of superspace, give some practical examples, and then generalize their description over general curved backgrounds. This is achieved by coupling the given theory to the supersymmetric version of Einstein gravity, _supergravity_ , and then requiring the gravitational sector of the resulting theory to decouple from the rest in a “rigid limit”, analogous to $G_{N}\to 0$. This method will bring up to the notion of _Killing spinors_ , a special type of spinorial fields whose existence ensures the preservation of some supersymmetry on the curved background. Finally, we will comment about a possible “super-interpretation” of the models of equivariant cohomology described in Chapter 2, that connects them to the usual BRST formalism for quantization of constrained Hamiltonian systems. In Chapter 5 we will discuss examples of Abelian supersymmetric localization of path integrals in the infinite-dimensional setting of QFT. The first case we will report is 1-dimensional, i.e. QM. In an Hamiltonian formulation on phase space, we will describe how it is possible to give a supersymmetric (so equivariant) interpretation to the path integral in a model-independent way, the supersymmetry arising as a “hidden” BRST symmetry which is linked to the Hamiltonian dynamics. This results in a localization of the path integral over the space of classical field configurations or over constant field configurations, that applied to supersymmetric QM gives an alternative proof the Atiyah-Singer index theorem for the Dirac operator over a twisted spinor bundle. Next, we will review two modern applications of the localization principle to the computation of expectation values of _Wilson loop operators_ in supersymmetric gauge theories. The first application computes the expectation value in $\mathcal{N}=4,2,2^{*}$ Super Yang-Mills theory on the 4-sphere $\mathbb{S}^{4}$, the second one in $\mathcal{N}=2$ Super Chern- Simons theory on the 3-sphere $\mathbb{S}^{3}$. In both cases, the partition function and the Wilson loop expectation value can be reduced to finite- dimensional integrals over the Lie algebra of the gauge group of the theory. This makes the supersymmetric theories in exam equivalent to a suitable “matrix model”, whose path integral can be computed exactly with some special regularization. We will give an example of computations of such a matrix model, since this class of objects arises in many important areas of modern theoretical physics. In Chapter 6 we will introduce Witten’s non-Abelian localization formula, and briefly describe its possible application in the study of 2-dimensional Yang- Mills theory. This more general formalism is able to show the mapping between the standard “physical” version of Yang-Mills theory and its “cohomological” (i.e. topological, in some sense) formulation, and is at the base of the localization of the 2-dimensional Yang-Mills partition function. Some technical asides are relegated to the appendices. Appendix A is devoted to some background in differential geometry, concerning principal bundles and the definition of spinors in curved spacetime. In Appendix B we report more details about equivariant cohomology, equivariant vector bundles and characteristic classes. This can be seen as a completion of the discussion of Chapter 2, from a more mathematical point of view. ## Chapter 2 Equivariant cohomology In this chapter we review the theory of _equivariant cohomology_ , as a modification of the standard cohomology theory applied to spaces that are equipped with the action of a _symmetry group_ $G$ on them, the so-called $G$-manifolds. First, we will review the basic notions about cohomology and homology, from their algebraic definition to the application in topology and differential geometry. The main result that we need to care about, and extend to the equivariant case, is _de Rham’s theorem_ [17], that gives an _algebraic model_ for the cohomology of a smooth manifold in terms of the complex of its differential forms. This and Stokes’ theorem relate the theory of cohomology classes to the integration on smooth manifolds. Next, we will extend this to the equivariant setting, giving a topological definition of equivariant cohomology, and then discussing, in the smooth case, an equivariant version of de Rham’s theorem. This, analogously to the standard case, will give an equivalence between the topological definition of equivariant cohomology and the cohomology of some suitable differential complex built from the smooth structure of the space at hand. There are different, but equivalent, possibilities of such algebraic models for the equivariant cohomology of a $G$-manifold: we will see the _Weil model_ , the _Cartan model_ and the _BRST model_ , and discuss how they are related one to each other, since at the end they have to describe the same equivariant cohomology. The purpose of all this, from the physics point of view, is that with equivariant cohomology we can describe a theory of cohomology and integration over manifolds that are acted upon by a symmetry group, the standard setup of classical mechanics and QFT. In the next chapter we will review one of the climaxes of this theory applied to the problem of integration over $G$-manifolds: the famous _localization formulas_ of Berline-Vergne [7] and Atiyah-Bott [6], that permit to highly simplify a large class of integrals thanks to the equivariant structure of the underlying manifold. The aim and the core of this thesis will be then the description of some generalizations and application of those theorems to the context of QM and QFT, where the integrals of interest are the infinite-dimensional _path integrals_ describing partition functions or expectation values of operators. For this introductory chapter, we follow mainly [18, 19, 20, 21]. Another classical reference is [22]. Some background tools from differential geometry that are needed can be found in Appendix A. ### 2.1 A brief review of standard cohomology theory In this section we will review some of the basic facts about standard homology and cohomology theory, and in particular its application to topological spaces with the definition of _singular_ homology and cohomology groups. Since this is after all standard material, we refer to any book of topology/geometry/algebra (for example [23, 24, 25, 26]) for the various proofs, while we will give some intuitive examples to help making concrete the various abstract definitions. The main result that we aim to recall is _de Rham’s theorem_ , that relates the cohomology theory to differential forms over smooth manifolds, and that will be extended in the next sections to the modified equivariant setup. We start with the abstract definition of homology and cohomology as algebraic constructions. From this point of view, (co)homology groups are defined in relation to _(co)chain complexes_ (or _differential complexes_). ###### Definition 2.1.1. Given a ring $R$, a _chain complex_ is an ordered sequence $A=(A_{p},d_{p})_{p\in\mathbb{N}}$ of $R$-modules $A_{p}$ and homomorphisms $d_{p}:A_{p}\to A_{p-1}$ such that $d_{p-1}\circ d_{p}=0$. A _cochain complex_ has the same structure but with homomorphisms $d_{p}:A_{p}\to A_{p+1}$, and $d_{p+1}\circ d_{p}=0$. Chain complex $\displaystyle:\qquad\cdots A_{p-1}\xleftarrow{d_{p}}A_{p}\xleftarrow{d_{p+1}}A_{p+1}\cdots$ Cochain complex $\displaystyle:\qquad\cdots A_{p-1}\xrightarrow{d_{p-1}}A_{p}\xrightarrow{d_{p}}A_{p+1}\cdots$ An element $\alpha$ of a (co)chain complex is called _(co)cycle_ or _closed_ if $\alpha\in\mathrm{Ker}(d_{p})$ for some $p$. It is instead called _(co)boundary_ or _exact_ if $\alpha\in\mathrm{Im}(d_{p})$ for some $p$. By definition $\mathrm{Im}(d_{p})\subseteq\mathrm{Ker}(d_{p\pm 1}),$ where the $-$ is for chain and the $+$ for cochain complexes, so the quotient sets $\faktor{\mathrm{Ker}}{Im}$ of $p$-(co)cycles modulo $p$-(co)boundries are well defined. ###### Definition 2.1.2. Given a (co)chain complex $A$, the _$p^{th}$ (co)homology group_ of $A$ is $\left(H^{p}(A):=\faktor{\mathrm{Ker}(d_{p})}{\mathrm{Im}(d_{p+1})}\right)\qquad H_{p}(A):=\faktor{\mathrm{Ker}(d_{p})}{\mathrm{Im}(d_{p-1})}.$ (Co)homology groups are called like that because they inherit a natural Abelian group structure (or equivalently $\mathbb{Z}$-module structure) from the sum in the original chain complex.111Notice that $[\alpha]+[\beta]:=[\alpha+\beta]$ is well defined. As usual, once we have a definition of a class of mathematical objects, a prime interest lies in the study of structure preserving maps between them. A morphism between (co)chain complexes $A$ and $B$ is then a sequence of homomorphisms $\left(f_{p}:A_{p}\to B_{p}\right)_{p\in\mathbb{N}}$ such that, schematically, $f\circ d^{(A)}=d^{(B)}\circ f$. It is easy to see that every such a morphism induces an homomorphism of (co)homology groups, since for example $f^{*}:H^{p}(A)\to H^{p}(B)$ such that $f^{*}([\alpha]):=[f(\alpha)]$ is well defined. Notice that, in many applications one considers (co)chain complexes defined by _graded_ modules or algebras with a suitable _differential_. An example of this type is the complex of differential forms $(\Omega(M),d)$ over a smooth manifold, that we will recover later on. For a general (co)chain complex $(A_{p},d_{p})_{p\in\mathbb{N}}$, we can always see $A:=\bigoplus_{p}A_{p}$ as a _graded_ $R$-module, whose elements as $\alpha\in A_{p}\subset A$ are said to have pure _degree_ $\textrm{deg}(\alpha):=p$. A generic element will be a sum of elements of pure degree. ###### Definition 2.1.3. A _differential graded algebra_ (_dg-algebra_ for short) over $R$ is then an $R$-algebra with the decomposition (grading) $A=\bigoplus_{p}A_{p}$, the product satisfying $A_{p}A_{q}\subseteq A_{p+q}$, and a _differential_ $d:A\to A$ such that 1. (i) it has degree $\textrm{deg}(d)=\pm 1$, meaning that for every $\alpha$ of degree $p$, $\textrm{deg}(d\alpha)=p\pm 1$; 2. (ii) it is nilpotent, $d^{2}=0$; 3. (iii) it satisfies the graded Leibniz rule, $d(\alpha\beta)=(d\alpha)\beta+(-1)^{\textrm{deg}(\alpha)}\alpha(d\beta)$. Every such an algebra clearly defines an underlying complex (cochain if $\textrm{deg}(d)=1$, chain if $\textrm{deg}(d)=-1$) and thus has associated (co)homology groups. Even if the algebra structure (the product and the Leibniz rule) are not needed to define the complex, we included it in the definition because this is the kind of structure that arises in physics or in differential geometry. Morphisms of dg-algebras are naturally defined as structure preserving maps between them, analogously to the above discussion. Beside the abstract algebraic definitions of above, one of the most important applications of homology and cohomology groups is in the study and classification of topological spaces. In order to define these groups in a topological setup, the complexes one takes into consideration are the _simplicial complexes_ , that intuitively represents a formal way of constructing “polyhedra” over $\mathbb{R}^{n}$, and that can be used in turn to study properties of topological spaces. Given $\mathbb{R}^{\infty}$ with the standard basis $\\{e_{i}\\}_{i=0,1,\cdots}$ ($e_{0}=0$), a _standard q-simplex_ is $\Delta_{q}:=\left\\{\left.x=\sum_{i=0}^{q}\lambda_{i}e_{i}\right|\sum_{i=0}^{q}\lambda_{i}=1,\lambda_{i}\in[0,1]\ \forall i=0,\cdots,q\right\\}.$ (2.1) Although this definition takes into account any possible dimensionality, we can embed these simplices in in finite-dimensional Euclidean spaces, giving them a more practical interpretation. Given $q+1$ points $v_{0},\cdots,v_{q}\in\mathbb{R}^{n}$, the associated _affine singular q-simplex_ in $\mathbb{R}^{n}$ is the map $\displaystyle\left[v_{0}\cdots v_{q}\right]:\Delta_{q}$ $\displaystyle\to\mathbb{R}^{n}$ (2.2) $\displaystyle\sum_{i=0}^{q}\lambda_{i}e_{i}$ $\displaystyle\mapsto\sum_{i=0}^{q}\lambda_{i}v_{i}.$ This is the convex hull in $\mathbb{R}^{n}$ generated by the _vertices_ $(v_{i})$. Geometrically, the 0-simplex is just the point $0\in\mathbb{R}^{n}$, 1-simplices are line segments, 2-simplices are triangles and so on. Notice that $\Delta_{q-1}\subset\Delta_{q}$, and its image through $[v_{0}\cdots v_{q}]$ is a “face” of the resulting polygon. More precisely, $[v_{0}\cdots\hat{v_{i}}\cdots v_{q}]:\Delta_{q-1}\to\Delta_{q}$ (the hat means we take away that point from the list) is regarded as the _$i^{th}$ face map_, denoted concisely as $F^{(i)}_{q}$. (a) (b) Figure 2.1: (a) First standard simplices. (b) An oriented affine 2-simplex, its face maps and a singular 2-simplex $\sigma_{2}$ on a 2-dimensional topological space. The same idea can be used to embed the simplices in a generic topological space $M$, changing the codomain of the simplex map. A _singular q-simplex_ in $M$ is then a continuous map222The standard topology on $\mathbb{R}^{q}$ is induced on $\Delta_{q}$. $\sigma_{q}:\Delta_{q}\to M$ (2.3) where now $\\{\sigma_{q}(e_{0}),\cdots,\sigma_{q}(e_{q})\\}$ are the _vertices_ of $\sigma_{q}$. Two simplices are said to have the same/opposite _orientation_ if the vertex sets are respectively even/odd permutations of each other. The word “singular” is there because only continuity is required, thus from a “smooth” point of view these simplices can present singularities. With this setup, we can construct chain complexes on topological spaces in terms of singular simplices. In fact, defining the sum of two singular simplices $\sigma_{q},\rho_{p}$ as $\displaystyle(\sigma_{q}+\rho_{p}):\Delta_{q}\sqcup\Delta_{p}$ $\displaystyle\to M$ (2.4) $\displaystyle\lambda$ $\displaystyle\mapsto\begin{cases}\sigma_{q}(\lambda)&\text{if}\ \lambda\in\Delta_{q}\\\ \rho_{q}(\lambda)&\text{if}\ \lambda\in\Delta_{p},\end{cases}$ whose image is the (disjoint) union in $M$ of the images of the two starting simplices. Since the “+” is clearly commutative, $\mathcal{C}_{q}(M):=C^{0}(\Delta_{q},M)$ is an Abelian group, called the _(singular) q-chain group_ of $M$. We can define a _boundary operator_ as a group homomorphism $\displaystyle\partial:\mathcal{C}_{q}(M)$ $\displaystyle\to\mathcal{C}_{q-1}(M)$ (2.5) $\displaystyle\sigma$ $\displaystyle\mapsto\partial\sigma:=\sum_{i=0}^{q}(-1)^{i}\left(\sigma\circ F^{(i)}_{q}\right),$ that restricts to the (oriented) sum of faces of a given simplex, and happens to satisfy the nilpotency condition $\partial\circ\partial=0$. This means that $\mathcal{C}(M):=\bigoplus_{q=0}^{\infty}\mathcal{C}_{q}(M)$ with the operator $\partial$ defines a dg-module over $\mathbb{Z}$, and an associated chain complex, that we use to define the homolgy groups of $M$. ###### Definition 2.1.4. The _singular $q^{th}$ homology group_ of $M$ is $H_{q}(M;\mathbb{Z}):=\faktor{\mathrm{Ker}(\partial_{q})}{\mathrm{Im}(\partial_{q+1})}.$ It is often useful to work with homology groups _with coefficients_ in some $\mathbb{Z}$-module $A$ (like the real numbers), that is considering $H_{q}(M;A)$ as defined from the simplicial complex $\mathcal{C}(M)\otimes A$. ###### Example 2.1.1 (Homology of spheres). In practice, the strategy to get $H_{*}(M)$ is the so-called _triangulation_ of $M$, i.e. constructing a suitable simplicial complex $K$ in $\mathbb{R}^{\dim(M)}$ as a set of standard simplices, whose union gives a polyhedron that is homeomorphic to $M$. Then, one can count and classify all the cycles and the boundaries in $K$, and then get $H_{*}(K)\cong H_{*}(M)$. Some examples of this rigorous approach can be found in [23]. We can still give some examples, less rigorously, by looking directly at simple topological spaces, just to help building some intuition. Remember that a $q$-cycle $\sigma$ on $M$ is a boundary-less singular $q$-simplex _up to continuous deformations_ , and it is also a boundary if it can be seen as the border of a $(q+1)$-simplex. 1. $(\mathbb{S}^{1})$ On the circle there is no place for simplices of dimension higher than 1, so we look at the 1-simplices. There are two inequivalent ways of deforming the standard 1-simplex onto the circle: it can join at the end points covering all $\mathbb{S}^{1}$ or not. In the first case, that we call $\sigma_{1}$, we have $\partial\sigma_{1}=0$ since $\mathbb{S}^{1}$ has no boundaries, in the second the boundaries are the end points of the singular 1-simplex. The first boundary-less case cannot be seen as a boundary of something else, by dimensionality, so the Abelian group $H_{1}(\mathbb{S}^{1};\mathbb{Z})=\mathrm{Ker}(\partial_{1})/\mathrm{Im}(\partial_{2})$ is generated by a single element, $[\sigma_{1}]$. In other words, $H_{1}(\mathbb{S}^{1};\mathbb{Z})\cong\mathrm{span}_{\mathbb{Z}}\\{[\sigma_{1}]\\}\cong\mathbb{Z}$. The case $q=0$ is trivial, since we have only one way of drawing a point on the circle, and every point is boundary-less. We have just said that the boundary of a 1-simplex is either zero or two points, so a single point is never a boundary. Thus the Abelian group $H_{0}(\mathbb{S}^{1};\mathbb{Z})\cong\mathbb{Z}$ since it is generated by only one element. Generalizing a little, we can already see from this example that the homology group in 0-degree will always follow this trend for _connected_ topological spaces. If the space has $n$ connected components, there will be $n$ inequivalent ways of drawing a point on it, so $n$ generators for the homology group, giving $H_{0}(M_{(n)};\mathbb{Z})\cong\mathbb{Z}\oplus\cdots\oplus\mathbb{Z}$ ($n$ factors). 2. $(\mathbb{S}^{2})$ The 2-sphere does not necessitate of much more work, at least with this level of rigor. Again, by dimensionality the homology groups in degrees higher than $\dim(\mathbb{S}^{2})=2$ are empty. For $q=2$, the only way we can construct a boundary-less figure on the sphere from $\Delta_{2}$ is joining the vertices and the edges together and cover the whole sphere. All other singular 2-simplices have boundaries, and the 2-sphere cannot be seen as a boundary of something else by dimensionality, so analogously to the previous case $H_{2}(\mathbb{S}^{2};\mathbb{Z})\cong\mathbb{Z}$. For the 1-simplices, we notice that the only two ways of drawing a segment on the sphere (up to continuous deformations) is to close it or not at the end points. In the first case, the 1-simplex has no boundary, but can be seen as the boundary of its internal area, so it is in fact exact. In the second case, the 1-simplex has boundaries so it is outside $\mathrm{Ker}(\partial)$. This means that every 1-cycle is also a boundary, and thus $H_{1}(\mathbb{S}^{2};\mathbb{Z})\cong 0$. In 0-degree we can argue in the same way as for the circle that $H_{0}(\mathbb{S}^{2};\mathbb{Z})\cong\mathbb{Z}$. 3. $(\mathbb{S}^{n})$ It turns out that all spheres follow this trend, giving homology groups $H_{q}(\mathbb{S}^{n};\mathbb{Z})\cong\begin{cases}\mathbb{Z}&q=0,n\\\ 0&\text{otherwise}.\end{cases}$ If interested in the case with real coefficient, the homology of spheres are again very simple, since $\mathbb{Z}\otimes\mathbb{R}\cong\mathbb{R}$. Now we can turn to the construction of singular cohomology groups on topological spaces. This is done considering the dual spaces $\text{Hom}(\mathcal{C}_{q}(M),A)$ with values in a $\mathbb{Z}$-module $A$. The simplest choice is of course $A=\mathbb{Z}$. Notice that $\text{Hom}(\mathcal{C}_{q}(M),A)$ itself is a $\mathbb{Z}$-module. The _coboundary operator_ in this case is defined as the $\mathbb{Z}$-module homomorphism $\displaystyle\delta:\text{Hom}(\mathcal{C}_{q}(M),A)$ $\displaystyle\to\text{Hom}(\mathcal{C}_{q+1}(M),A)$ (2.6) $\displaystyle f$ $\displaystyle\mapsto\delta f\quad s.t.\quad\delta f(\sigma_{q+1}):=f(\partial\sigma_{q+1})$ and from the nilpotency $\partial^{2}=0$ we get easily $\delta^{2}=0$. ###### Definition 2.1.5. The _singular $q^{th}$ cohomology group_ of $M$, _with coefficients in $A$_, is $H^{q}(M;A):=\faktor{\mathrm{Ker}(\delta_{q})}{\mathrm{Im}(\delta_{q-1})}.$ Note that for a commutative ring $A$ (as for example $\mathbb{R}$), the cohomology groups are naturally $A$-modules. Although the definition is less practical than the one for homology groups, there is an important theorem that allows to relate the two, so that homology computations can be used to infer the structure of singular cohomology groups. This is the so-called _universal coefficient theorem_ [24]. Since for applications to smooth manifolds we will be primarily interested in cohomology groups with real coefficients (as it will become clearer later) we take $A=\mathbb{R}$. For this special case, the theorem says $H^{q}(M;\mathbb{R})\cong H^{q}(M;\mathbb{Z})\otimes\mathbb{R}\cong H_{q}(M;\mathbb{R})^{*},$ (2.7) so that the cohomolgy groups are exactly the dual spaces of the homology groups. Notice that, from the example above, $H^{q}(\mathbb{S}^{n};\mathbb{R})\cong\mathbb{R}$ in degree $q=0,n$. The above construction relates the topological properties of the space $M$ to the algebraic concept of (co)homology groups. In general we said that morphisms of complexes induce morphisms of associated (co)homologies, and this extends to the present topological case: if we consider a continuous map (morphism of topological spaces) $F:M\to N$ between the topological spaces $M,N$, we can lift it to $F_{\\#}:\mathcal{C}_{q}(M)\to\mathcal{C}_{q}(N)$ such that $F_{\\#}(\sigma):=F\circ\sigma$, that is a morphism of dg-modules. This in turn induces the morphism of homology groups as descrbed above. For the cohomology groups we have, analogously, the lifted map in the opposite direction $F^{\\#}:\text{Hom}(\mathcal{C}_{q}(N),A)\to\text{Hom}(\mathcal{C}_{q}(M),A)$ such that $F^{\\#}(f):=f\circ F_{\\#}$, giving the morphism of cochain complexes. This induces $F^{*}:H^{q}(N)\to H^{q}(M)$ such that $F^{*}([f]):=[F^{\\#}(f)]$.333In the context of smooth manifolds and de Rham cohomology, this is analogous to the _pull-back_ of differential forms. Also, we notice that if we have two continuous maps $F,G$ between the topological spaces, then444In category theory language, we can summarize these properties saying that singular homology $H_{*}(\cdot)$ is a _covariant functor_ between the categories Top of topological spaces and Ab of Abelian groups, and singular cohomology $H^{*}(\cdot)$ is a _contravariant functor_ between Top and Ab. Anyway, we will not need such a terminology for what follows. See for example [27], Appendix A, for a quick introduction to the subject. $\displaystyle(F\circ G)_{\\#}=F_{\\#}\circ G_{\\#}\quad\Rightarrow\quad(F\circ G)_{*}=F_{*}\circ G_{*}$ (2.8) $\displaystyle(F\circ G)^{\\#}=G^{\\#}\circ F^{\\#}\quad\Rightarrow\quad(F\circ G)_{*}=G_{*}\circ F_{*}.$ An important fact that permits to use homology and cohomology groups to classify and characterize topological spaces, is that these objects are _topological invariants_ , meaning that isomorphic spaces have the same (co)homology groups. Moreover, a stricter result holds: two homotopy- equivalent topological spaces have the same cohomology and homology groups. We recall that two continuous maps $F,G:M\to N$ between topological spaces are _homotopic_ if it exists a continuous map $H:[0,1]\times M\to N$ that deforms continuously $F$ in $G$, i.e. $H(0,x)=F(x)$ and $H(1,x)=G(x)$ for every $x\in M$. Homotopy of maps is an equivalent relation, and we denote it by $F\sim G$. Two topological spaces $M,N$ are said to be _homotopy-equivalent_ , or of the same homotopy type, if there exist two maps $F:M\to N$ and $G:N\to M$ such that $(G\circ F)\sim id_{M}$. Homotopy-equivalence is also an equivalence relation, that we denote also as $M\sim N$. The result stated above is then, for cohomologies $M\sim N\Rightarrow H^{*}(M)\cong H^{*}(N).$ (2.9) ###### Example 2.1.2 (More singular homologies). 1. $(pt)$ We can consider the very trivial case of $M$ being just a point. In this case, the informal discussion of Example 2.1.1 can be carried out just for the 0-dimensional simplices: $H_{*}(pt;\mathbb{Z})\cong\mathbb{Z}$ in degree 0. It follows by definition and by the universal coefficient theorem that for any commutative ring $A$ (as $\mathbb{R}$), $H^{0}(pt;A)\cong H_{0}(pt;A)\cong A$ and $H^{q}(pt;A)\cong H_{q}(pt;A)\cong 0$ for $q>0$. By the homotopy- invariance property discussed above, any contractible space will have the same trivial cohomology and homology as the point! 2. $(\mathcal{C})$ Let us look at another simple case, the cylinder $\mathcal{C}=\mathbb{S}^{1}\times[0,1]$. To compute its homology groups we could follow the intuitive discussion of Example 2.1.1, or we can just notice that since the interval $[0,1]$ is contractible, $\mathbb{S}^{1}\times[0,1]\sim\mathbb{S}^{1}.$ This means that $H^{q}(\mathbb{S}^{1}\times[0,1];\mathbb{R})\cong H^{q}(\mathbb{S}^{1};\mathbb{R})\cong H_{q}(\mathbb{S}^{1};\mathbb{R})$. 3. $(T)$ A less trivial example is the 2-torus $T=\mathbb{S}^{1}\times\mathbb{S}^{1}$. In this case no one of the factors is contractible, so we cannot use the homotopy invariance to get the result from a simpler space. We can anyway get the answer using the same method of Example 2.1.1. Starting from the top- degree homology group, we notice that the only boundary-less surface on the torus is the torus itself. Thus analogously to all the other cases, $H_{2}(T;\mathbb{Z})\cong\mathbb{Z}$ or $H_{2}(T;\mathbb{R})\cong\mathbb{R}$. Since the torus is connected, in degree 0 we get trivially $H_{0}(T;\mathbb{R})\cong\mathbb{R}$. In degree 1 we see the difference with the other cases. On the torus there are two inequivalent ways of drawing a closed line that is not a boundary of any 2-dimensional surface, following essentially the two factors of $\mathbb{S}^{1}$ (see figure 2.2). This means that the $1^{st}$ homology group is generated by two elements, and thus $H_{1}(T;\mathbb{Z})\cong\mathbb{Z}\oplus\mathbb{Z}$. The same reasoning can be applied to higher genus surfaces $\Sigma_{g}$, giving $H_{1}(\Sigma_{g};\mathbb{Z})\cong(\mathbb{Z})^{\oplus 2g}$. (a) (b) Figure 2.2: (a) The 2 inequivalent non-exact 1-cycles $\sigma_{1}$ and $\sigma_{1}^{\prime}$ on the 2-torus. (b) The genus-$g$ surface $\Sigma_{g}$ has $2g$ inequivalent non-exact 1-cycles. Figures adapted from [23]. We leave now the purely topological setup, since in physics we are mostly interested in studying local properties, i.e. from the differential geometry point of view. We assume to work in the smooth setting and consider $M$ to be a $d$-dimensional $C^{\infty}$-manifold. With $TM$ we denote its _tangent bundle_ , and with $T^{*}M$ its _cotangent bundle_.555Sections of any bundle $E\to M$ over $M$ will be denoted in the following with $\Gamma(M,E)$, or $\Gamma(E)$ when the base space is clear from the context. For example, vector fields are elements of $\Gamma(TM)$. At every point $p\in M$, $T_{p}M$ and $T^{*}_{p}M$ are the dual vector spaces of tangent vectors and 1-forms at $p$, respectively. We consider the _exterior algebra_ $\bigwedge(T^{*}_{p}M)$, with the wedge product $\wedge$ making it in a graded-commutative algebra, and the exterior derivative $d:\bigwedge(T^{*}_{p}M)\to\bigwedge(T^{*}_{p}M)$ acting as a graded derivation of $\textrm{deg}(d)=+1$. Extending these operations point-wise for every point $p\in M$, we have the bundle of _differential forms_ over $M$, $\Omega(M):=\bigoplus_{k=0}^{d}\Omega^{k}(M)\qquad\text{with}\ \Omega^{k}(M):=\bigsqcup_{p\in M}\bigwedge^{k}(T^{*}_{p}M).$ (2.10) $(\Omega(M),\wedge,d)$ is thus a dg-algebra over the commutative ring $C^{\infty}(M)$, and naturally defines a cochain complex called the _de Rham complex_. The associated cohomology groups are the _de Rham cohomology groups_ , constituting the graded-commutative ring666Analogously to the singular cohomology, in category theory language the de Rham cohomology $H_{dR}(\cdot)$ is a contravariant functor between the categories Man of smooth manifolds and Ab of Abelian groups. $H_{dR}(M)=\bigoplus_{k=0}^{d}H_{dR}^{k}(M)\qquad\text{with}\quad H_{dR}^{k}(M):=H^{k}(\Omega(M),d)=\faktor{\mathrm{Ker}(d_{k})}{\mathrm{Im}(d_{k-1})}.$ (2.11) The ring structure of $H_{dR}(M)$ is naturally inherited from the wedge product of differential forms, that lifts at the level of cohomology classes. In fact, for two closed forms $\omega,\eta\in\Omega(M)$, $[\omega]\wedge[\eta]:=[\omega\wedge\eta]$ (2.12) is well-defined.777This result can be seen also in the topological setup for singular cohomology groups, as it should be by de Rham’s theorem. The operation that corresponds to the wedge product between singular cohomology classes is called _cup-product_ [24]. It is important to remember that cohomology in general has a ring structure. The final important result that we state, and that will be crucial to extend to the equivariant setting in the following section, is the so called _de Rham’s theorem_ : ###### Theorem 2.1.1 (de Rham). The de Rham cohomology of the smooth manifold $M$ is isomorphic to its singular cohomology with real coefficients: $H_{dR}(M)\cong H^{*}(M;\mathbb{R}).$ The power of this theorem is that it allows to study topological properties of the manifold (recall that $H^{*}(M;\mathbb{R})$ are homotopy-invariants) using differential geometric (so local) objects, the differential forms. We say that the de Rham complex $(\Omega(M),d)$ constitute an _algebraic model_ for the singular cohomology of $M$. Notice that, by dimensionality reasons, we get trivially also in this case that the cohomology groups $H_{dR}^{q}(M)$ for $q>\dim(M)$ are automatically zero. Another important property that is intuitively very clear from the de Rham complex is the _Poincaré duality_. For a closed connected manifold $M$ this states that, as vector spaces $H^{k}_{dR}(M)\cong H^{\dim(M)-k}_{dR}(M).$ (2.13) A crucial tool for the proof of de Rham’s theorem is the so-called _Stokes’ theorem_ , that relates the integral of an exact $d$-form over a $d$-dimensional manifold to the integral of its primitive over the $(d-1)$-dimensional boundary, $\int_{M}d\omega=\int_{\partial M}\omega.$ (2.14) Notice that integration over $M$ when $\partial M=\emptyset$ can be regarded as a function on the $d^{th}$ de Rham cohomology $\int:H^{d}_{dR}(M)\to\mathbb{R}$. ###### Example 2.1.3 (Cohomology rings). With the help of de Rham’s theorem, we can compute some of the previous example directly at the level of cohomology using differential forms and integration.888Another powerful tool to practically compute cohomology groups and rings, at the topological level, goes by the name of _spectral sequences_. See for example [18]. Let us consider the case of the tours $T=(\mathbb{S}^{1})^{2}$. We can parametrize it with coordinates $(x,y)$ taking values in $[0,1)^{2}\subset\mathbb{R}^{2}$. If we call $\alpha:=dx$ and $\beta:=dy$ in $\Omega^{1}(T)$, a natural choice of volume form is $\omega:=\alpha\wedge\beta$, that gives $\mathrm{vol}(T)=1$. The volume form is of course closed by dimensionality, but it cannot be exact since otherwise by Stokes’ theorem the volume of the torus would be 0, so it defines a non- trivial cohomology class $[\alpha\wedge\beta]$. Any other 2-form is of the type $\omega^{\prime}=f\omega$ for some $f\in C^{\infty}(T)$, but closed forms must satisfy $df=0$, so $f\in\mathbb{R}$ constant. We conclude that any other independent closed 2-form has to be “cohomologous” to $[\alpha\wedge\beta]$, so that in top-degree $H^{2}_{dR}(T)\cong\mathrm{span}\\{[\alpha\wedge\beta]\\}\cong\mathbb{R}$. In degree 1, any closed form must be a combination of $\alpha$ and $\beta$ with real coefficients (since again $d(f\alpha)=0\ \Leftrightarrow\ df=0$), so they are the only independent closed 1-forms (they correspond to the volume forms for the two $\mathbb{S}^{1}$ factors). To see whether or not they are exact, we can use Stokes’ theorem: if they are, then their integral over _any_ closed curve on $T$ must be zero. But we can take the two curves $\sigma_{1}(t)=(t,0)$ and $\sigma_{1}^{\prime}(t)=(0,t)$ of Figure 2.2 and see that $\int_{\sigma_{1}}\alpha=1=\int_{\sigma_{1}^{\prime}}\beta,$ so they define two independent cohomology classes $[\alpha]$ and $[\beta]$. This means that $H^{1}_{dR}(T)\cong\mathrm{span}\\{[\alpha],[\beta]\\}\cong\mathbb{R}\oplus\mathbb{R}$. Since the torus is connected, the only closed 0-form is a constant number, that we can chose to be $1$. Thus, $H^{0}_{dR}\cong\mathbb{R}$. We can further easily get the ring structure of $H_{dR}(T)$ by looking at the multiplication rules between the generators. If we call $a:=[\alpha]$ and $b:=[\beta]$, the wedge product of differential forms gives the following rules $a^{2}\sim 0,\qquad b^{2}\sim 0,\qquad ab+ba\sim 0.$ Thus we can rewrite the cohomology ring as a polynomial ring over the indeterminates $(a,b)$, taken in degree 1, that satisfy the above rules: $H^{*}(T;\mathbb{R})\cong\faktor{\mathbb{R}[a,b]}{(a^{2},b^{2},ab+ba)},$ where $(a^{2},b^{2},ab+ba)$ denotes the quotient by the ideal generated by the corresponding expressions. In the same fashion we can rewrite the cohomology rings of the other examples that we gave above for the $n$-sphere. Introducing an indeterminate $u$ of degree $n$, and the multiplication rule $u^{2}\sim 0$, its cohomology ring can be expressed as $H^{*}(\mathbb{S}^{n};\mathbb{R})\cong\faktor{\mathbb{R}[u]}{u^{2}}.$ We quote another example, that will enter in the case of equivariant cohomology with respect to a circle action by $U(1)\cong\mathbb{S}^{1}$. For the complex projective plane $\mathbb{C}P^{n}$, it turns out that $H^{*}(\mathbb{C}P^{n};\mathbb{R})\cong\faktor{\mathbb{R}[u]}{u^{(n+1)}}$ where $\mathrm{deg}(u):=2$. In the limiting case $n\to\infty$, one has thus $H^{*}(\mathbb{C}P^{\infty};\mathbb{R})\cong\mathbb{R}[u]$, the polynomials in $u$. ### 2.2 Group actions and equivariant cohomology As already mentioned, equivariant cohomology is an extension of the standard cohomology theory, partly reviewed in the last section, to the cases in which the space $M$ is acted upon by some group $G$. This is the common setup in physics, from the finite-dimensional cases of classical Lagrangian or Hamiltonian mechanics to the infinite dimensional case of Quantum Field Theory, where $M$ can be the configuration space, the phase space, or the space of fields, and $G$ is a Lie group representing a _symmetry_ of the physical system. In gauge theory for example, we want to identify those physical configurations that are equivalent modulo a gauge transformations, so the moduli space of gauge orbits $M/G$. In Poincaré-supersymmetric theories, the group $G$ is actually the Poincaré group of spacetime symmetries. In all these cases we are interested in the cohomology of $M$ modulo these symmetry transformations, since many primary objects of study (partition functions, expectation values…) are usually given in terms of integrals over $M$. Before moving to the technical definition of $G$-equivariant cohomology of $M$, we recall some terminology about group actions. ###### Definition 2.2.1. 1. (i) Given a group $G$ and a topological space $M$,999We are going to work practically always with smooth manifolds and (compact) Lie groups, but for the moment we do not need this level of structure on $M$ and $G$. a _$G$ -action_ on $M$ is given by a group homomorphism (_left_ action) or anti-homomorphism (_right_ action) $\rho:G\to\mbox{Homeo}(M)\quad(\text{or Diff}(M)\text{ for smooth manifolds}).$ If $m\in M,g\in G$, the left action of $g$ on $m$ can be denoted $\rho(g)m\equiv g\cdot m$, and the right action $m\cdot g$, if this causes no confusion. $M$ is said to be a (left or right) _$G$ -space_. 2. (ii) If $M,N$ are two $G$-spaces, on the product $M\times N$ it is canonically defined the _diagonal $G$-action_ $\rho^{M\times N}(g)(m,n):=(\rho^{M}(g)m,\rho^{N}(g)n)\quad\text{for }m\in M,n\in N,g\in G.$ 3. (iii) Given a point $m\in M$, the _orbit_ of $m$ is the subset of $M$ of all points that are reached from $m$ by the action of $G$. The _orbit space_ with respect to the $G$-action is $M/G$.101010It is easy to check that $m\sim m^{\prime}\Leftrightarrow m^{\prime}=g\cdot m$ for some $g\in G$ is an equivalence relation. 4. (iv) The _stabilizer_ (or _isotropy group_ , or _little group_) of $m$ is the subgroup of $G$ of all elements that act trivially on $m$, i.e. $g\cdot m=m$. The $G$-action is called _free_ if the stabilizer of every point in $M$ is given by the identity of $G$. The $G$-action is called _locally free_ if the stabilizer of every point is _discrete_. The _fixed point set_ $F\subseteq M$, is the set of all points that are stabilized by the entire $G$. 5. (v) Morphisms of $G$-spaces are called _$G$ -equivariant_ functions. $f:M\to N$ is $G$-equivariant if $f(g\cdot m)=g\cdot f(m)$, for every $m\in M,g\in G$. In the following we will not care much about distinguishing between left and right actions, and assume all $G$-actions are from the left, unless otherwise stated. For the first part of the discussion it is not needed, but we are going to assume $M$ and $G$ to be at least topological manifolds, and then specialize to the case of smooth manifolds, since these are the most common structures arising in physics. Since, as we said above, we are interested in identifying those elements in $M$ that are equivalent up to a “symmetry” transformation by $G$, the first candidate for the $G$-equivariant cohomology of $M$ could be simply the cohomology of the orbit space $M/G$, $H_{G}^{*}(M):=H^{*}\left(\faktor{M}{G}\right).$ (2.15) This definition has the problem that, if the $G$-action is not free and has fixed points on $M$, the orbit space is singular: in the neighborhood of those fixed points there is no well-defined notion of dimensionality. This kind of singular quotient spaces are called _orbifolds_. ###### Example 2.2.1 (Some group actions and orbit spaces). 1. (i) Let us consider the Euclidean space $\mathbb{R}^{n}$. $O(n)$ rotations (and reflections) act naturally on it, with the only fixed point being the origin. Any point but the origin identifies a direction in the Euclidean space, and thus is stabilized by the subgroup of $n-1$ rotations, $O(n-1)$. For example we see that, without considering parity transformations, the standard $SO(2)$-action is free on $\mathbb{R}^{2}\setminus\\{0\\}$. If we bring translations into the game, considering the Euclidean space as an affine space acted upon by $ISO(n)=\mathbb{R}^{n}\rtimes O(n)$, then the stabilizer of any point is the entire $O(n)$, since any point can be considered an origin after translation. So the action of $ISO(n)$ is neither free nor locally free, but has no fixed points on the entire $\mathbb{R}^{n}$. Considering only rotations, the orbit space $\mathbb{R}^{n}/O(n)$ is the space of points identified up to their angular coordinates, that is an half-line starting from the origin, $\mathbb{R}^{n}/O(n)\cong[0,\infty)$. This is not a manifold, since the interval is closed on the left, giving a “singularity” on the original fixed point of the action. Notice that, by embedding $\mathbb{S}^{n-1}$ in $\mathbb{R}^{n}$, an $O(n)$-action descends on it, and the orbit of any point of $\mathbb{S}^{n-1}$ is the sphere itself. The orbit of any point can be seen as the quotient of $O(n)$ by the stabilizer of that point, so that one has $\faktor{O(n)}{O(n-1)}\cong\mathbb{S}^{n-1}.$ This quotient describes the common situation of spontaneous symmetry braking inside $O(n)$-models, in Statistical Mechanics. 2. (ii) One can always consider circle actions on the spheres $\mathbb{S}^{n}$. Starting with $n=1$, and considering the circle as embedded in the $\mathbb{C}$-plane, $U(1)$ acts on itself by multiplication: $e^{i\varphi}\mapsto e^{ia}e^{i\varphi}$ for some $a\in[0,2\pi)$. This action has clearly no fixed points, and it is also free. Thus the quotient is well defined, giving simply $\mathbb{S}^{1}/U(1)\cong pt$. The $U(1)$-action on the 2-sphere is already more interesting. Rotations around a given axis fix two points on $\mathbb{S}^{2}$, that we identify with the North and the South poles. If we exclude the poles the resulting space is homeomorphic to a cylinder, the $U(1)$-action becomes free and indeed we have that $\mathbb{S}^{1}\times(0,1)\to(0,1)$ is a trivial principal $U(1)$-bundle. But considering the poles, the quotient space is singular since $\mathbb{S}^{2}/U(1)\cong[0,1]$. This is an elementary example of an orbifold. Figure 2.3: The circle acting on the 2-sphere, the orbits being the parallels. The orbit space $\mathbb{S}^{2}/U(1)$ is a meridian, homeomorphic to the interval $[0,1]$. Let us consider also the case $n=3$. The 3-sphere $\mathbb{S}^{3}\cong SO(2)$ can be parametrized by a pair of complex numbers $(z_{1},z_{2})$ such that $|z_{1}|^{2}+|z_{2}|^{2}=1$. The circle then acts naturally by diagonal multiplication: $(z_{1},z_{2})\mapsto(e^{ia}z_{1},e^{ia}z_{2})$ for some $e^{ia}\in U(1)$. This action is clearly free, since the two coordinates $z_{i}$ cannot be simultaneously zero on the sphere, thus the quotient is well defined, giving the 3-sphere the structure of a principal $U(1)$-bundle known as the _Hopf bundle_. The equivalence classes $[z_{1},z_{2}]\in\mathbb{S}^{3}/U(1)$ describe, by definition, points on the complex projective line $\mathbb{C}P^{1}\cong\mathbb{S}^{2}$, that is isomorphic to the Riemann 2-sphere. The Hopf bundle can be thus seen as $\mathbb{S}^{3}\to\mathbb{S}^{2}$, with typical fiber $\mathbb{S}^{1}$. Clearly this is not a trivial bundle, since $\mathbb{S}^{3}\neq\mathbb{S}^{2}\times\mathbb{S}^{1}$. The above case generalizes to any odd-dimensional sphere $\mathbb{S}^{2n+1}$, since they all can be embedded in complex spaces $\mathbb{C}^{n+1}$. The circle acts always by diagonal multiplication, and the resulting action is free. The bundles $\mathbb{S}^{2n+1}\to\mathbb{S}^{2n+1}/U(1)\cong\mathbb{C}P^{n}$ are all principal $U(1)$-bundles over the complex projective spaces $\mathbb{C}P^{n}$. 3. (iii) The last case we mention is the possible $U(1)$-action on a torus $T=(\mathbb{S}^{1})^{2}$, by rotations along one of the two factors. This is the only possible free action on a closed surface, giving the well defined quotient $T/U(1)\cong\mathbb{S}^{1}$. More examples can be found in [28]. The example above showed that also in very simple cases singularities can appear in quotient spaces, so that one cannot define cohomology in a smooth way using the powerful de Rham theorem. It is thus more convenient to set up a definition of equivariant cohomology that automatically avoids this problem. This more clever definition is given by the _Borel construction_ for the $G$-space $M$. ###### Definition 2.2.2. Considering a $G$-space $M$, its associated _Borel construction_ , or _homotopy quotient_ , is $M_{G}:=\faktor{(M\times EG)}{G}\equiv M\times_{G}EG$ where $EG$ is some contractible space on which $G$ acts freely, called the _universal bundle of $G$_ (see Appendix B.2 for the precise definition). The _$G$ -equivariant cohomology_ of $M$ is then defined as111111From now on we always consider cohomologies with coefficients in $\mathbb{R}$, unless otherwise stated. $H_{G}^{*}(M):=H^{*}(M_{G}).$ We assume the action on the product $M\times EG$ to be the diagonal action. Notice that, since $G$ acts freely on $EG$, the action on the product is automatically free. Indeed, if in the worst case $p\in M$ is a fixed point, for every point $e\in EG$, $g\cdot(p,e)=(p,g\cdot e)\neq(p,e)$. This means that the homotopy quotient defines a smooth manifold, and we can hope for a generalization of de Rham’s theorem, allowing to study this topological definition from its smooth structure in terms of something analogous to the differential forms on $M$. We will discuss this result in the next section. Since the space $EG$ does not need to either exist or be unique a priori, one could think that the above definition contains some degree of arbitrariness, so a natural question is: is equivariant cohomology well defined? The answer is of course yes, and the crucial fact allowing this stands in the contractibility of the space $EG$. The arguments that lead to this conclusion are summarized in Appendix B.2, together with some examples of universal bundles. The important property that one has to keep in mind is that, intuitively, to get it acted freely by $G$ and being contractible, one has to define it “so big” that for _any_ other principal $G$-bundle $P$, there is a copy of $P$ sitting inside $EG$. This is why it is called “universal”. Here we just notice that, if we assume a contractible free $G$-space $EG$ to exist, we have an homotopy equivalence $M\times EG\sim M$, that descends also to the homotopy quotient $\faktor{(M\times EG)}{G}\sim\faktor{M}{G},$ (2.16) since one can show that $M\times_{G}EG\to M/G$ is a fiber bundle with typical fiber $EG$ [18]. From homotopy invariance of cohomology, we see that at least in the case in which $G$ acts freely on $M$ and $M/G$ is well defined, the equivariant cohomology reduces to the naive definition above, $H_{G}^{*}(M)=H^{*}\left(M\times_{G}EG\right)\cong H^{*}\left(\faktor{M}{G}\right).$ (2.17) Notice that the contractible space $EG$ alone has a very simple cohomology. Indeed, for what we pointed out in Example 2.1.2, it must be $H^{*}(EG)\cong H^{*}(pt)\cong\mathbb{R}$ in degree zero. When we take the quotient, the base space $BG:=EG/G$ can have a less trivial cohomology. This space is called _classifying space_ of the Lie group $G$. When, for example, the $G$-action on $M$ is trivial (all points are fixed points), the homotopy quotient is just $M\times_{G}EG\cong M\times(EG/G)=M\times BG$, and in this case we have121212This is an application of the so-called _Künneth theorem_ [24]. $H^{*}_{G}(M)=H^{*}(M\times BG)\cong H^{*}(M)\otimes H^{*}(BG),$ (2.18) so that the homotopy quotient by a trivial action does not bring any further information to the cohomology of $M$ but for tensoring it with the cohomology of the classifying space. In Section 2.4 we will see that the latter can be described in general by a very simple algebraic model, while here we carry on the example of the case $G=U(1)$. ###### Example 2.2.2 (A few $U(1)$-equivariant cohomologies). To search for a suitable principal $U(1)$-bundle whose total space is contractible, we can first notice from Example 2.2.1 that we already described a class of principal $U(1)$-bundles, $\mathbb{S}^{2n+1}\to\mathbb{C}P^{n}$, whose total spaces are the odd-dimensional spheres. The bad news is that any of these total spaces are contractible, but this problem can be solved considering the limiting case $n\to\infty$, since it turns out that $\mathbb{S}^{\infty}=\bigcup_{n}S^{2n+1}\equiv\bigcup_{n}S^{n}$ is contractible [18]!131313Notice that any sphere $\mathbb{S}^{n}$ can be embedded as the equator of $\mathbb{S}^{n+1}$. Thus there is a sequence of inclusions $\mathbb{S}^{1}\subset\mathbb{S}^{3}\subset\cdots$ as well as $\mathbb{C}P^{1}\subset\mathbb{C}P^{3}\cdots$, and the circle action is compatible with the inclusion. Thus, in the limit, a free circle action induces on $\mathbb{S}^{\infty}$. Thus the universal bundle for $U(1)$ can be chosen to be $EU(1)=\mathbb{S}^{\infty}$, and the classifying space $BU(1)=\mathbb{C}P^{\infty}$. Being infinite-dimensional, they are strictly speaking not manifolds, but $\mathbb{S}^{\infty}\to\mathbb{C}P^{\infty}$ is still a topological bundle, and this is enough for the definition of an homotopy quotient. 1. (i) In Example 2.1.2 we quoted the resulting cohomology ring of the complex projective planes, $H^{*}(\mathbb{C}P^{n})\cong\mathbb{R}[\phi]/\phi^{n+1}$ and $H^{*}(\mathbb{C}P^{\infty})=H^{*}(BU(1))\cong\mathbb{R}[\phi]$, with $\phi$ in degree 2. Thus the $U(1)$-equivariant cohomology of any space $M$ on which $U(1)$ acts trivially is, from (2.18), $H^{*}_{U(1)}(M)=H^{*}(M)\otimes\mathbb{R}[\phi].$ In particular if $M$ is contractible, $H^{*}_{U(1)}(M)=H^{*}_{U(1)}(pt)=\mathbb{R}[\phi]$. 2. (ii) The opposite case is the one of a free action, for example the circle acting on itself. As we pointed out above, $\mathbb{S}^{1}/U(1)\cong pt$, so simply $H^{*}_{U(1)}(\mathbb{S}^{1})\cong\mathbb{R}$. 3. (iii) The last example that we mention is the case of $U(1)$ acting on $\mathbb{S}^{2}$. The equivariant cohomology $H^{*}_{U(1)}(\mathbb{S}^{2})$ is non-trivial a priori, as we remarked above, and it can be calculated easily for example using spectral sequences. We will not enter in the detail of the calculation but only describe the result. Consider first the standard cohomology of the 2-sphere that we already saw in various examples, being $H^{*}(\mathbb{S}^{2})\cong\mathbb{R}\oplus\mathbb{R}y$, where we explicitly wrote a generator $y$ for the term in degree 2, that can be identified in the de Rham model by a volume form $y=[\omega],\omega\in\Omega^{2}(\mathbb{S}^{2})$. It turns out that its equivariant version can be obtained simply by tensoring with the polynomial ring $H^{*}(BU(1))=\mathbb{R}[\phi]$, $H^{*}_{U(1)}(\mathbb{S}^{2})\cong\mathbb{R}[\phi]\oplus\mathbb{R}[\phi]y,$ although the generator $y$ has now a different interpretation, that we will give in terms of an equivariant version of the de Rham model in the next sections.141414In terms of differential forms, $\omega$ will have to be _equivariantly extended_ in the Cartan model, as discussed at the end of Section 2.4. This equivariant cohomology actually can be given a ring structure, defining the multiplication $y\cdot y=a\phi y+b\phi^{2}$ for some constants $a,b$. It turns out [18] that the correct constants are $a=1,b=0$, making $\faktor{\mathbb{R}[y,\phi]}{(y^{2}-\phi^{2})}\to H_{U(1)}^{*}(\mathbb{S}^{2})$ into a ring isomorphism, where the denominator stands for the ideal generated by the expression $(y^{2}-\phi^{2})$ in $\mathbb{R}[y,\phi]$. Notice that the cohomology groups $H^{n}_{U(1)}(\mathbb{S}^{2})$ are now non-empty in every even-degree (while in odd-degree they are all trivial), even when $n$ is bigger than the dimension of the sphere! This intuitively matches the fact that the quotient $\mathbb{S}^{2}/U(1)$ is singular, and thus simple dimensionality arguments do not make sense anymore at the fixed points. ### 2.3 The Weil model and equivariant de Rham’s theorem From now on, we specialize the equivariant cohomological theory to $G$ being a Lie group with Lie algebra $\mathfrak{g}$,151515Some of what follows is only rigorous if $G$ is compact, but the formal discussion can be applied generically. and $M$ being a smooth $G$-manifold. We saw that de Rham’s theorem provides an _algebraic model_ for the singular cohomology (with real coefficients) of the smooth manifold $M$, through the complex of differential forms. We now describe a way to obtain an algebraic model for the homotopy quotient $(M\times EG)/G$, the so-called _Weil model_ for the $G$-equivariant cohomology of $M$. From the discussion of the last sections, it is already imaginable that this will contain in some way the de Rham complex of $M$, but modifying it through a somewhat “trivial” extension, in the sense of the triviality of the cohomology of $EG$. This is thus the most natural model that is connected to the topological definition of the last section, but we will see that it is also overly complicated. In fact, in the next section we will describe a simpler but equivalent way to obtain the same equivariant cohomology, the _Cartan model_ , that is more intuitive from the differential geometry point of view, and that we will use to generalize the theory of integration to the equivariant setting. This is what we often use in physics for practical calculations. Before defining the Weil model, we notice that, in presence of a $G$-action, the de Rham complex $(\Omega(M),d)$ of differential forms on $M$ has more structure than being a dg algebra. In fact, if $\rho:G\to\text{Diff}(M)$ is the $G$-action, this induces an infinitesimal action of the Lie algebra $\mathfrak{g}$ on any tensor space via the Lie algebra homomorphism161616Usually, in physics conventions, in the action of the exponential map one collects a factor of $i$ at the exponent, in order to consider the Lie algebra element Hermitian for the most commonly considered group actions. Left and right actions should be taken with different signs at the exponent. $\displaystyle\mathfrak{g}$ $\displaystyle\to\Gamma(TM)$ (2.19) $\displaystyle X$ $\displaystyle\mapsto\underline{X}:=\left.\frac{d}{dt}\right|_{t=0}\rho\left(e^{-tX}\right)^{*}$ that defines for any $X\in\mathfrak{g}$ the corresponding _fundamental vector field_ $\underline{X}\in\Gamma(TM)$. Then $\mathfrak{g}$ acts infinitesimally on $\Omega(M)$ via the _Lie derivative_ and the _interior multiplication_ with respect to the fundamental vector fields, $\begin{aligned} \mathcal{L}_{X}(\alpha)&:=\mathcal{L}_{\underline{X}}(\alpha)\\\ \iota_{X}\alpha&:=\iota_{\underline{X}}\alpha\end{aligned}\qquad\text{for}\ X\in\mathfrak{g},\alpha\in\Omega(M),$ (2.20) with the additional property (_Cartan’s magic formula_) $\mathcal{L}_{X}=d\circ\iota_{X}+\iota_{X}\circ d.$ (2.21) This makes $\Omega(M)$ into a so-called _$\mathfrak{g}$ -gd algebra_. In general, a $\mathfrak{g}$-gd algebra is defined as a differential graded algebra (cf. definition 2.1.3) with two actions of $\mathfrak{g}$, denoted by analogy as $\iota$ and $\mathcal{L}$, such that for any $X\in\mathfrak{g}$ 1. (i) $\iota_{X}$ acts as an _antiderivation_ of degree $-1$, satisfying $(\iota_{X})^{2}=0$; 2. (ii) $\mathcal{L}_{X}$ acts as a derivation (of degree 0); 3. (iii) the Cartan’s magic formula holds: $\mathcal{L}_{X}=d\circ\iota_{X}+\iota_{X}\circ d$. Morphisms of $\mathfrak{g}$-dg algebras are naturally defined as maps between $\mathfrak{g}$-dg algebras that commute with all the above stated operations. It is not difficult to show that $G$-equivariant maps of $G$-manifolds induce pull-backs of differential forms that preserve the $\mathfrak{g}$-dg algebra structure. Now we can define the Weil model via an extension of $(\Omega(M),d)$ that preserves this new structure. We want this extension to be an algebraic analog of $EG$, so its cohomology must be trivial, but carrying information about $\mathfrak{g}$. To do this, we associate it to the characteristic differential structure of a generic principal $G$-bundle $P$ (remember that any principal $G$-bundle sits inside $EG$): its connection 1-form $A\in\Omega^{1}(P)\otimes\mathfrak{g}$ and the associated curvature $F=dA+\frac{1}{2}[A\stackrel{{\scriptstyle\wedge}}{{,}}A]\in\Omega^{2}(P)\otimes\mathfrak{g}$, satisfying the Bianchi identity $dF=[F,A]$. We first notice that the connection 1-form and the curvature 2-form can be seen as linear maps $\begin{aligned} A:\mathfrak{g}^{*}&\to\Omega^{1}(P)\\\ \eta&\mapsto A(\eta):=(\eta\circ A),\end{aligned}\qquad\begin{aligned} F:\mathfrak{g}^{*}&\to\Omega^{2}(P)\\\ \eta&\mapsto F(\eta):=(\eta\circ F).\end{aligned}$ (2.22) These maps can be extended multi-linearly to the whole $\Omega(P)$, if we start from algebras constructed by $\mathfrak{g}^{*}$ that respect the commutativity of 1- and 2-forms, respectively. This means that $A$ has to “eat” an element of the (anticommutative) exterior algebra $\bigwedge(\mathfrak{g}^{*})$, while $F$ has to “eat” an element of the (commutative) symmetric algebra $S(\mathfrak{g}^{*})$: $\begin{aligned} A:\bigwedge(\mathfrak{g}^{*})&\to\Omega(P)\\\ \eta_{1}\wedge\cdots\wedge\eta_{k}&\mapsto A(\eta_{1})\wedge\cdots\wedge A(\eta_{k}),\end{aligned}\qquad\begin{aligned} F:S(\mathfrak{g}^{*})&\to\Omega(P)\\\ \eta_{1}\cdots\eta_{k}&\mapsto F(\eta_{1})\wedge\cdots\wedge F(\eta_{k}).\end{aligned}$ (2.23) We can combine the two maps in the homomorphism of graded-algebras $\displaystyle f:S(\mathfrak{g}^{*})\otimes\bigwedge(\mathfrak{g}^{*})$ $\displaystyle\to\Omega(P)$ (2.24) $\displaystyle\eta\otimes\xi$ $\displaystyle\mapsto F(\eta)\wedge A(\xi).$ This captures the fact that a connection of $P$ could be _defined_ as a map $S(\mathfrak{g}^{*})\otimes\bigwedge(\mathfrak{g}^{*})\to\Omega(P)$, and motivates the following definition. ###### Definition 2.3.1. The _Weil algebra_ of $\mathfrak{g}$ is the graded algebra $W(\mathfrak{g}):=S(\mathfrak{g}^{*})\otimes\bigwedge(\mathfrak{g}^{*})$ and the map $f:W(\mathfrak{g})\to\Omega(P)$ is called the _Weil map_. We define the graded structure of $W(\mathfrak{g})$ by assigning to the generators $\\{\phi^{a}\\}$ of $S(\mathfrak{g}^{*})$ degree $\textrm{deg}(\phi^{a})=2$, and to the generators $\\{\theta^{a}\\}$ of $\bigwedge(\mathfrak{g}^{*})$ degree $\textrm{deg}(\theta^{a})=1$. The generators $\\{\phi^{a},\theta^{a}\\}$ are two copies of a basis set for $\mathfrak{g}^{*}$, but taken in different degrees. With respect to this _graded_ basis, the Weil algebra can also be written as $W(\mathfrak{g})=\bigwedge\left(\mathbb{R}[\phi^{1},\cdots,\phi^{\dim\mathfrak{g}}]\oplus\mathbb{R}[\theta^{1},\cdots,\theta^{\dim{\mathfrak{g}}}]\right),$ (2.25) and a generic element will be expanded as $\alpha=\alpha_{0}+\alpha_{a}\theta^{a}+\frac{1}{2}\alpha_{ab}\theta^{a}\theta^{b}+\cdots+\alpha^{(top)}\theta^{1}\theta^{2}\cdots\theta^{\dim\mathfrak{g}}\quad\text{with}\quad\alpha_{I}\in\mathbb{R}[\phi^{1},\cdots,\phi^{\dim\mathfrak{g}}],$ (2.26) since higher order terms vanish by the anticommutativity of the $\theta$’s. Here we suppressed tensor and wedge products to simplify the notation, as we will often do in the following. On this basis, the Weil map projects simply the connection and the curvature on the given Lie algebra components, $f(\theta^{a})=\theta^{a}\circ A=A^{a},\qquad f(\phi^{a})=\phi^{a}\circ\Omega=\Omega^{a}.$ (2.27) The Weil algebra is the central object to define an algebraic model for $EG$. We need to define a $\mathfrak{g}$-dg algebra structure on it to properly take its cohomology, but this is naturally done requiring the Weil map to be a morphism of $\mathfrak{g}$-dg algebras. This means introducing a differential $d_{W}:W(\mathfrak{g})\to W(\mathfrak{g})$ and two $\mathfrak{g}$-actions $\mathcal{L},\iota$ such that the following diagram commutes for all the three operations separately, ${W(\mathfrak{g})}$${\Omega(P)}$${W(\mathfrak{g})}$${\Omega(P).}$$\scriptstyle{d_{W}\ \iota\ \mathcal{L}}$$\scriptstyle{f}$$\scriptstyle{d\ \iota\ \mathcal{L}}$$\scriptstyle{f}$ (2.28) One can check that, defining the _Weil differential_ $d_{W}$ on the generators as171717In the second column we introduced $\theta:=\theta^{i}\otimes T_{i}$ and $\phi:=\phi^{i}\otimes T_{i}$ in $W(\mathfrak{g})\otimes\mathfrak{g}$, where $\\{T_{i}\\}$ is a basis of $\mathfrak{g}$ dual to the generators. Notice that this notation make the formulas independent on a choice of basis. Also, these are the objects that are really correspondent to the connection $A$ and the curvature $F$ on $P$, respectively. $\begin{aligned} d_{W}\theta^{a}&=\phi^{a}-\frac{1}{2}f^{a}_{bc}\theta^{b}\theta^{c}\\\ d_{W}\phi^{a}&=f^{a}_{bc}\phi^{b}\theta^{c}\end{aligned}\quad\text{or}\quad\begin{aligned} d_{W}\theta&=\phi-\frac{1}{2}[\theta\stackrel{{\scriptstyle\wedge}}{{,}}\theta]\\\ d_{W}\phi&=[\phi\stackrel{{\scriptstyle\wedge}}{{,}}\theta]\end{aligned}$ (2.29) where $f_{ab}^{c}$ are the structure constants of $\mathfrak{g}$, it commutes with $f$ giving correctly the definition of curvature and the Bianchi identity. Moreover, extending the differential on $W(\mathfrak{g})$ as an antiderivation of degree $+1$, it gives $d_{W}^{2}=0$ (since $d_{W}^{2}$ is a derivation, it is enough to check it on the generators). To be compatible with the properties of the connection and the curvature $\iota_{X}A=A(\underline{X})=X,\qquad\iota_{X}F=0\qquad\forall X\in\mathfrak{g},$ (2.30) the interior multiplication must be defined as $\begin{aligned} \iota_{X}\theta^{a}&:=\theta^{a}(X)=X^{a}\\\ \iota_{X}\phi^{a}&:=0\end{aligned}\quad\text{or}\quad\begin{aligned} \iota_{X}\theta&:=\theta(X)=X\\\ \iota_{X}\phi&:=0\end{aligned}$ (2.31) and extended as an antiderivation of degree $-1$. Then the Lie derivative is simply defined via Cartan’s magic formula. We finally have defined the Weil algebra as a $\mathfrak{g}$-dg algebra. ###### Theorem 2.3.1. The cohomology of the Weil algebra is $H^{0}(W(\mathfrak{g}),d_{W})\cong\mathbb{R},\qquad H^{k}(W(\mathfrak{g}),d_{W})\cong 0\ \text{for}\ k>0.$ ###### Proof. The full proof can be found in [18]. Schematically it follows the proof of the Poincaré lemma: one has to find an _cochain homotopy_ , i.e. a map $K:W(\mathfrak{g})\to W(\mathfrak{g})$ of degree -1, such that $[K,d_{W}]_{+}=id$. Then any cocycle ($d_{W}\alpha=0$) is also a coboundary, since $\alpha=[K,d_{W}]_{+}\alpha=d_{W}(K\alpha)$. This can be found for any degree $k>0$. In degree zero $W^{0}(\mathfrak{g})\cong\mathbb{R}$ by definition, so every element is a cocycle, and no one is a coboundary for degree reasons. ∎ ###### Example 2.3.1 (Weil model for torus and circle actions). Consider the case of a compact Abelian group, i.e. a torus $T=U(1)^{l}$ for some $l$. Remember that a possible purpose of the Weil algebra is to describe the connection and the curvature of any principal $T$-bundle, so we are somewhat analyzing the structure of $l$ electromagnetic fields, from the point of view of the Lie algebra $\mathfrak{t}$. Since the structure constants are all zero, the Weil differential (2.29) and the $\mathfrak{t}$-actions (2.31) on the generators $(\theta^{a},\phi^{a})$ simplify as $\begin{array}[]{ll}d_{W}\theta^{a}=\phi^{a},&d_{W}\phi^{a}=0,\\\ \iota_{b}\theta^{a}=\delta^{a}_{b},&\iota_{b}\phi^{a}=0,\\\ \mathcal{L}_{b}\theta^{a}=0,&\mathcal{L}_{b}\phi^{a}=0,\end{array}$ where we denoted $\iota_{a}\equiv\iota_{T_{a}}$ and $\mathcal{L}_{a}\equiv\mathcal{L}_{T_{a}}$, with $\\{T_{a}\\}$ the basis of $\mathfrak{t}$ dual to the generators of $W(\mathfrak{t})$. We jump ahead a little and notice that the first line really resembles the structure of a “supersymmetry” transformation, with $\phi^{a}$ being the “bosonic partner” of $\theta^{a}$. The remaining non-Abelian piece of the generic case can be viewed as the action of a Chevalley-Eilemberg differential, so that $d_{W}=d_{susy}+d_{CE}$.181818Remember that the C-E differential is the one that appears in BRST quantization of gauge theories. We will return to this point in Section 4.5, after having introduced some technology about supergeometry and supersymmetry. Let us simplify again and prove theorem 2.3.1 for $l=1$. In the case of $U(1)$, the Lie algebra has only one generator $T\cong i$, and the symmetric algebra is the algebra of polynomials in the indeterminate $\phi\in\mathfrak{g}^{*}$, $S(\mathfrak{g}^{*})=\mathbb{R}[\phi]$, while the exterior algebra reduces to $\bigwedge(\theta)=\mathbb{R}\oplus\mathbb{R}\theta$ by anticommutativity. The Weil algebra is thus $W(\mathfrak{u}(1))=\mathbb{R}[\phi]\otimes\left(\mathbb{R}\oplus\mathbb{R}\theta\right)=\mathbb{R}[\phi]\oplus\mathbb{R}[\phi]\theta.$ The cohomology of $W(\mathfrak{u}(1))^{0}=\mathbb{R}$ in degree zero is as always trivial, since all constant numbers are closed, and none of them is exact, giving $H^{0}(W(\mathfrak{u}(1)),d_{W})\cong\mathbb{R}$. In degree 1 we have $W(\mathfrak{u}(1))^{1}=\mathbb{R}\theta$, thus no one element (but zero) is closed. This extends to any odd-degree, since $W(\mathfrak{u}(1))^{2n+1}=\mathbb{R}\phi^{n}\theta$, and $d_{W}(\phi^{n}\theta)=\phi^{n+1}\neq 0$. This means that $H^{2n+1}(W(\mathfrak{u}(1)),d_{W})\cong 0$. In degree 2, we have $W(\mathfrak{u}(1))^{2}=\mathbb{R}\phi$, so that any element is closed but also exact, since $\phi=d_{W}\theta$. This extends to any even-degree, since $W(\mathfrak{u}(1))^{2n}=\mathbb{R}\phi^{n}$, and $\phi^{n}=d_{W}\theta\phi^{n-1}=d_{W}(\theta\phi^{n-1})$. Thus we have also $H^{2n}(W(\mathfrak{u}(1)),d_{W})\cong 0$, showing the triviality of the Weil algebra in the simplest case of a circle action. Almost the same direct computation can be carried out in the $l$-dimensional case. Theorem 2.3.1 shows that we are in business: the Weil algebra is exactly an algebraic analog of the universal bundle $EG$. Since the de Rham model for $M$ is just $\Omega(M)$, the product $M\times EG$ can be modeled by the complex $W(\mathfrak{g})\otimes\Omega(M)$, since by the Künneth formula [24] and de Rham’s theorem $H^{*}(EG\times M)=H^{*}(EG)\otimes H^{*}(M)=H^{*}(W(\mathfrak{g}^{*}),d_{W})\otimes H^{*}(\Omega(M),d).$ (2.32) The differential and the $\mathfrak{g}$-actions are extended naturally on this complex as graded derivations, making it into a $\mathfrak{g}$-dg algebra too. Explicitly, $\displaystyle d_{T}$ $\displaystyle:=d_{W}\otimes 1+1\otimes d,$ (2.33) $\displaystyle\iota$ $\displaystyle\equiv\iota\otimes 1+1\otimes\iota,$ $\displaystyle\mathcal{L}$ $\displaystyle\equiv\mathcal{L}\otimes 1+1\otimes\mathcal{L}.$ A model for the homotopy quotient $M_{G}$ can be guessed by the following argument. Since $M_{G}$ is the base of the principal bundle $EG\times G\to M_{G}$, differential forms on $M_{G}$ identify the _basic forms_ on $EG\times M$ (see Appendix A), i.e. those that are both _$G$ -invariant_ and _horizontal_. It is thus reasonable that the homotopy quotient can be modeled by the _basic subcomplex_ of the Weil model. Since the differential closes on the basic subcomplex, we are allowed to take its cohomology, giving the $G$-equivariant cohomology of $M$. This is exactly the content of the _equivariant de Rham’s theorem_. A recent original proof of it can be found in [18]. ###### Theorem 2.3.2 (equivariant de Rham). If $G$ is a connected Lie group, and $M$ is a $G$-manifold, $\boxed{H^{*}_{G}(M)\cong H^{*}\left(\left(W(\mathfrak{g})\otimes\Omega(M)\right)_{bas},d_{T}\right)}.$ The equivariant de Rham’s theorem is telling us that the “correct” differential complex that encodes the topology of the $G$-action on $M$ is _not_ anymore the complex $\Omega(M)$ of differential forms, but a modification of it through the presence of the Weil algebra. Remember always that, via the Weil map, $W(\mathfrak{g})$ can be thought as in correspondence with the presence of a connection and a curvature on some principal $G$-bundle. This means that the right extension of the de Rham complex in presence of a $G$-action embeds somewhat the presence of a connection and a curvature with respect to $G$. We can analyze as an example the simplest case of a $U(1)$-action. The unrestricted Weil model is (from Example 2.3.1) $W(\mathfrak{u}(1))\otimes\Omega(M)=\Omega(M)[\phi]\oplus\Omega(M)[\phi]\theta,$ (2.34) thus any element can be written as $\alpha=\alpha^{(0)}+\alpha^{(1)}\theta,$ (2.35) where $\alpha^{(0)},\alpha^{(1)}\in\Omega(M)[\phi]$ are polynomials in $\phi$ with differential forms as coefficients. The subcomplex of basic forms consists of those elements that satisfy both $\iota_{T}\alpha=0$ and $\mathcal{L}_{T}\alpha=0$, imposing the conditions $\alpha^{(1)}=-\iota_{T}\alpha^{(0)},\qquad\mathcal{L}_{T}\alpha^{(0)}=0.$ (2.36) Thus any basic element can be written as $\alpha=(1-\theta\iota_{T})\sum_{i}(\phi)^{i}\alpha^{(0)}_{i}$, where all the differential forms $\alpha^{(0)}_{i}\in\Omega(M)^{G}$ must be $G$-invariant, and the basic subcomplex can be identified with the polynomials in $\phi$ with invariant differential forms as coefficients. In the next section we will argue that this is not a special case, and that the Weil model can be simplified in general, producing another model for the same equivariant cohomology. ### 2.4 The Cartan model As we said at the beginning of the last section, the cohomology of the Weil complex is not the unique algebraic model for the $G$-equivariant cohomology of the $G$-manifold $M$. Moreover, although very transparent, the Weil model seems overly complicated for differential geometric applications. In fact, the extreme simplicity of the basic subcomplex of the Weil algebra $W(\mathfrak{g})_{bas}$ suggests that a simpler model for equivariant cohomology can be obtained simplifying this one. To see this, we analyze this basic subcomplex first. As we recalled at the end of the last section (for further details see Appendix A), a basic element $\alpha\in W(\mathfrak{g})_{bas}$ is both _horizontal_ and _invariant_ , i.e. $\iota_{X}\alpha=0=\mathcal{L}_{X}\alpha.$ (2.37) The horizontal condition means that we pick only the symmetric algebra inside $W(\mathfrak{g})$, since by definition $\iota_{X}\phi^{a}=0$. Imposing also the $G$-invariance we have $W(\mathfrak{g})_{bas}\cong S(\mathfrak{g}^{*})^{G},$ (2.38) i.e. the basic subcomplex is the algebra of Casimir invariants. It is easy to check that on this subcomplex $d_{W}\cong 0$, so that $H^{*}(W(\mathfrak{g})_{bas},d_{W})=H^{*}(S(\mathfrak{g}^{*})^{G},d_{W})=S(\mathfrak{g}^{*})^{G}\quad\text{in degree 0,}$ (2.39) since every element is closed, and no one element can be exact. Moreover, from the equivariant de Rham’s theorem $H_{G}^{*}(pt)\cong H^{*}(W(\mathfrak{g})_{bas},d_{W})$, so the Casimir invariants are precisely the cohomology of the classifying space, $H^{*}(BG)\cong S(\mathfrak{g}^{*})^{G}.$ (2.40) Motivated by the above simplification, we can turn now to analyze the complete Weil model $(W(\mathfrak{g})\otimes\Omega(M))_{bas}$. Let us see concretely what it means to restrict the attention to a basic element $\alpha\in(W(\mathfrak{g})\otimes\Omega(M))_{bas}$, starting from its expansion on a basis (2.26). Imposing the horizontality condition means, using the multi-index notation $I=(a_{1},\cdots,a_{|I|})$, $0=\iota_{X}\alpha=\iota_{X}\left(\alpha_{0}+\frac{1}{|I|!}\alpha_{I}\theta^{I}\right)\quad\Rightarrow\quad 0=\iota_{X}\alpha_{0}+\frac{1}{|I|!}(\iota_{X}\alpha_{I})\theta^{I}+\frac{1}{|I|!}\alpha_{I}(\iota_{X}\theta^{I}).$ (2.41) Equating the terms of the same degree and taking $X$ to be a basis element, one arrives at the condition on the various components, $\alpha_{a_{1}\cdots a_{|I|}}=(-1)^{|I|}\iota_{a_{1}}\cdots\iota_{a_{|I|}}\alpha_{0}$ (2.42) where we denoted $\iota_{k}:=\iota_{T_{k}}$, meaning that a horizontal element is fully determined by its first component $\alpha_{0}\in S(\mathfrak{g}^{*})\otimes\Omega(M)$, and it can be expressed as $\alpha=\left(\prod_{k=1}^{\dim\mathfrak{g}}(1-\theta^{k}\iota_{k})\right)\alpha_{0}.$ (2.43) This comment, with some more checks (see again [18] for a complete proof), proves the following theorem, and extends the above discussion to the complete Weil model. ###### Theorem 2.4.1 (Mathai-Quillen isomorphism). There is an isomorphism of $\mathfrak{g}$-dg algebras, called the _Mathai- Quillen isomorphism_ [29] (or _Cartan-Weil_ in [18]), $\displaystyle\varphi:(W(\mathfrak{g})\otimes\Omega(M))_{hor}$ $\displaystyle\to S(\mathfrak{g}^{*})\otimes\Omega(M)$ $\displaystyle\alpha=\alpha_{0}+\frac{1}{|I|!}\alpha_{I}\theta^{I}$ $\displaystyle\mapsto\alpha_{0}$ $\displaystyle\left(\prod_{k=1}^{\dim\mathfrak{g}}(1-\theta^{k}\iota_{k})\right)\alpha_{0}$ $\displaystyle\mathrel{\reflectbox{$\mapsto$}}\alpha_{0}.$ The RHS of the isomorphism above inherits the $\mathfrak{g}$-actions and the differential from the Weil model on the LHS, making commutative the following diagram, similarly to (2.28), ${(W(\mathfrak{g})\otimes\Omega(M))_{hor}}$${S(\mathfrak{g}^{*})\otimes\Omega(M)}$${(W(\mathfrak{g})\otimes\Omega(M))_{hor}}$${S(\mathfrak{g}^{*})\otimes\Omega(M).}$$\scriptstyle{d_{T}\ \iota\ \mathcal{L}}$$\scriptstyle{\varphi}$$\scriptstyle{d_{C}\ \iota\ \mathcal{L}}$$\scriptstyle{\varphi}$ (2.44) In particular, the new differential is called _Cartan differential_ , defined such that $d_{C}:=\varphi\circ d_{T}\circ\varphi^{-1}.$ (2.45) It is not difficult to get, directly from this definition, that it can be expressed more simply as $d_{C}=1\otimes d-\phi^{k}\otimes\iota_{k}$ (2.46) where $d$ is the de Rham differential on $\Omega(M)$. The two $\mathfrak{g}$-actions commute with $\varphi$ without modification, so they agree with their behavior in the Weil model, $\mathcal{L}_{X}\phi^{a}=f^{a}_{bc}\phi^{b}X^{c},\qquad\iota_{X}\phi^{a}=0.$ (2.47) Using the nilpotency of $d$ and $\iota$, and Cartan’s magic formula, wee see that the Cartan differential on the horizontal subcomplex squares to a Lie derivative (an infinitesimal symmetry transformation) $d_{C}^{2}=-\phi^{k}\otimes\mathcal{L}_{k},$ (2.48) so when we restrict to the $G$-invariant subspace, $d_{C}^{2}\cong 0$ on $\left(S(\mathfrak{g}^{*})\otimes\Omega(M)\right)^{G}$, as it should. Using the Mathai-Quillen isomorphism and the equivariant de Rham theorem we then have the fundamental result, $\boxed{H_{G}^{*}(M)\cong H^{*}\left(\left(S(\mathfrak{g}^{*})\otimes\Omega(M)\right)^{G},d_{C}\right)}$ (2.49) that simplifies the algebraic model for the $G$-equivariant cohomology of $M$. ###### Definition 2.4.1. The $\mathfrak{g}$-dg algebra $\Omega_{G}(M):=\left(S(\mathfrak{g}^{*})\otimes\Omega(M)\right)^{G}$ with the Cartan differential $d_{C}$ is called the _Cartan model_ for the equivariant cohomology of $M$. Elements of $\Omega_{G}(M)$ are called _equivariant differential forms_ on $M$. The degree of an equivariant form is the total degree with respect to the generators of $\Omega(M)$ (in degree 1), and the generators of $S(\mathfrak{g^{*}})$ (in degree 2). Equivariant forms will be in the next chapters the the principal object of study. In the physical applications we are interested in, we will always search for an interpretation of the space of interest as a Cartan model with respect to the action of a symmetry group $G$. The Cartan differential will be some object that squares to an infinitesimal symmetry, and on the subspace of $G$-invariant forms (or “fields”, in the following) it will define a $G$-equivariant cohomology. Cartan differentials arise in Field Theory as _supersymmetry transformations_ , that we will contextualize in Chapter 4 and relate to equivariant cohomology in Chapter 5. This interpretation will be crucial in treating some of the most important objects in QM and QFT that arise as (infinite-dimensional) path integrals over the space of fields. In fact, we will see in the next chapter that integration of equivariant forms leads to powerful _localization thorems_ , that formally extended to the infinite-dimensional case greatly simplifying those integrals. ###### Example 2.4.1 (Cartan model for $U(1)$-equivariant cohomology). Until Chapter 6, we will actually deal with the equivariant cohomology with respect to a circle action of $G=U(1)$, or at most a torus action of $U(1)^{n}$ for some $n$. As we saw also for the Weil model, this greatly simplifies the problem, so we carry on that example also in the Cartan model for a $U(1)$-action. We recall from Example 2.3.1 that in the Weil model $\iota_{T}\phi=\mathcal{L}_{T}\phi=d_{W}\phi=0,$ i.e. $\phi$ is automatically also $U(1)$-invariant. The equivariant forms are thus $\Omega_{U(1)}(M)=(\mathbb{R}[\phi]\otimes\Omega(M))^{U(1)}\cong\Omega(M)^{U(1)}[\phi],$ so polynomials in $\phi$ with $U(1)$-invariant forms as coefficients. The Cartan differential is, suppressing tensor products, $d_{C}=d-\phi\ \iota_{T}.$ In this 1-dimensional case, the indeterminate $\phi$ is just a spectator, and serves only to properly count the equivariant form-degree. This is important of course, but for many purposes it creates no confusion to suppress its presence. More precisely, we often _localize_ the algebra $\Omega_{G}(M)$, substituting the _indeterminate_ $\phi$ with a _variable_ , and setting it for example to $\phi=-1$,191919This could seem harmless, but it is definitely a non-trivial move. We are really able to do this without spoiling the resulting equivariant cohomology because (algebraic) localization commutes with taking cohomology. More details on this are reported in Appendix B.3. so that $d_{C}=d+\iota_{T}.$ This differential squares to an infinitesimal symmetry generated by $\underline{T}$, $d_{C}^{2}=\mathcal{L}_{T}$. Equivariant differential forms after this localization are just $U(1)$-invariant forms. Often it is useful to generate equivariant forms from invariant differential forms in $\Omega(M)$, for the purpose of integration for example. If $\alpha\in\Omega^{2n}(M)$, an _equivariant extension_ of $\alpha$ is $\tilde{\alpha}\in\Omega_{U(1)}(M)$ such that $\tilde{\alpha}=\alpha+f_{(2n-2)}\phi+f_{(2n-4)}\phi^{2}+\cdots$ where any coefficient is an invariant form in $\Omega(M)$. As an example, we can take the circle acting on the 2-sphere $\mathbb{S}^{2}$, via rotations around a chosen axis. If $\theta$ is the polar coordinate and $\varphi$ is the azimutal coordinate, $(\theta,\varphi)\left(e^{it}\cdot p\right):=(\theta(p),\varphi(p)+t)\qquad\text{for}\ p\in\mathbb{S}^{2},e^{it}\in U(1),$ so that the fundamental vector field is $\underline{T}=\frac{\partial}{\partial\varphi}$.202020Recall that this action has two fixed points, at the North and the South pole. Consider the canonical volume-form $\omega=d\cos{(\theta)}\wedge d\varphi$. It is obviously closed, and also $U(1)$-invariant, since $\mathcal{L}_{T}\omega=0$. Aiming to the extension of $\Omega(M)$ to $\Omega_{U(1)}(M)$, we can find an _equivariantly closed extension_ of the volume form $\tilde{\omega}=\omega+f\phi$, with $f\in C^{\infty}(M)$ such that $d_{C}\tilde{\omega}=0$, so that it is closed in the “correct” complex. This imposes the equation $df=\iota_{T}\omega$,212121$\omega$ is a _symplectic_ form on $\mathbb{S}^{2}$, and the $df=\iota_{T}\omega$ means that $H:=-f$ is the _Hamiltonian function_ with respect to the $U(1)$-action on the sphere. We will deepen this point of view in the next chapter. and so $f=-\cos(\theta)$: $\tilde{\omega}=\omega-\cos(\theta)\phi.$ ### 2.5 The BRST model In this section we mention the last popular model for equivariant cohomology: the so-called _BRST model_ , or sometimes _intermediate model_. It is worth to mention it because we will see in the next chapter that it is (as the first name suggests) intimately related to the BRST method for gauge-fixing in the Hamiltonian formalism. Moreover, its complex is the one that arises naturally in Topological Field Theories (TFT), as we will mention in Chapter 6. It is also important because it provides (as the second name suggests) an “interpolation” between the Weil and the Cartan models that we saw in the last sections, relating the latter more “physical”(or differential geometric) point of view with the former more “topological” one. As an algebra, the (unrestricted) complex of the BRST model is identical to that of the Weil model, $B:=W(\mathfrak{g})\otimes\Omega(M),$ (2.50) but with the new differential (compare to (2.29) and (2.33)) $d_{B}=d_{W}\otimes 1+1\otimes d+\theta^{a}\otimes\mathcal{L}_{a}-\phi^{a}\otimes\iota_{a}$ (2.51) that satisfies $d_{B}^{2}=0$ on $B$, and has the same trivial cohomology of the unrestricted Weil model. The idea that brought to the construction of this model in [30], was essentially to prove along the line we did in the last section the equivalence of the models, but from a slightly different point of view. In fact one can construct $d_{B}$ using an algebra automorphism that carries the Weil model $(B,d_{W})$ into the BRST model $(B,d_{B})$, at the level of the unrestricted algebras. The restriction to the basic subcomplex gives then automatically the Cartan model. The automorphism is given by the map $\varphi:=e^{\theta^{a}\iota_{a}}\equiv\prod_{a}(1+\theta^{a}\otimes\iota_{a}),$ (2.52) that looks very similar to the Mathai-Quillen isomorphism of theorem 2.4.1, but now is applied to the whole algebra and not only on the horizontal part. Analogously to the definition of the Cartan differential, $d_{B}$ is got as (2.51) from the commutativity of the diagram ${B}$${B}$${B}$${B,}$$\scriptstyle{d_{W}}$$\scriptstyle{\varphi}$$\scriptstyle{d_{B}}$$\scriptstyle{\varphi}$ (2.53) so that $d_{B}=\varphi^{-1}\circ d_{W}\circ\varphi$, as well as the two $\mathfrak{g}$-actions. In particular, it results $\displaystyle\iota^{(B)}$ $\displaystyle=\iota\otimes 1\neq\iota^{(W)},$ (2.54) $\displaystyle\mathcal{L}^{(B)}$ $\displaystyle=\mathcal{L}\otimes 1+1\otimes\mathcal{L}=\mathcal{L}^{(W)},$ where we called $\iota^{(W)},\mathcal{L}^{(W)}$ the one defined in (2.33). Thus the BRST differential carries the same information of the Weil differential, giving the same trivial cohomology of the unrestricted Weil model, $H^{*}(B,d_{B})\cong H^{*}(B,d_{W})\cong H_{dR}(M)$ (2.55) where the last equivalence follows from the triviality of the cohomology of the Weil algebra $W(\mathfrak{g}^{*})$. Of course, we have to restrict the the action of $d_{B}$ to the basic subcomplex, i.e. to the intersection with the kernels of $\iota^{(B)}$ and $\mathcal{L}^{(B)}$, to get a meaningful $G$-equivariant cohomology. This reproduces again the Cartan model, as expected. This result shows that there is in fact a whole continuous family of $\mathfrak{g}$-dg algebras that give equivalent models for the $G$-equivariant cohomology of $M$, because we can conjugate the Weil differential through the modified automorphism $\varphi_{t}:=e^{t\theta^{a}\iota_{a}}\qquad\text{with}\ t\in\mathbb{R}.$ (2.56) This produces, by conjugation, the family of differentials and $\mathfrak{g}$-actions on $B$, $\displaystyle d^{(t)}$ $\displaystyle=d_{W}\otimes 1+1\otimes d+t\theta^{a}\otimes\mathcal{L}_{a}-t\phi^{a}\otimes\iota_{a}+\frac{1}{2}t(1-t)f_{ab}^{c}\theta^{a}\theta^{b}\otimes\iota_{c},$ (2.57) $\displaystyle\iota^{(t)}$ $\displaystyle=\iota\otimes 1+(1-t)1\otimes\iota,$ $\displaystyle\mathcal{L}^{(t)}$ $\displaystyle=\mathcal{L}^{(W)}\quad\forall t.$ We see that for $t=0$ we recover the Weil model, while for $t=1$ we get the BRST model, as special cases. When restricted to the basic subcomplex, they all give the same equivariant cohomology. ## Chapter 3 Localization theorems in finite-dimensional geometry In this chapter we are going to introduce one of the most important results of the equivariant cohomology theory: the _Atiyah-Bott-Berline-Vergne (ABBV) localization formula_ for torus actions, discovered independently by Berline and Vergne [7], and by Atiyah and Bott [6]. For the most applications to QM and QFT, we will focus on the case of a circle action, and higher-dimensional generalizations will be postponed to Chapter 6. This formula can be viewed as a generalization of an analogous result of Duistermaat and Heckman [5], that treats the special case in which the torus action is Hamiltonian on a symplectic manifold. We will expand on this point of view in the second part of the chapter, since this is the situation we are more commonly interested in when we treat dynamical systems in physics, at least at the classical level. The formal generalization of these formulas in the infinite-dimensional setting of QFT will be discussed in Chapter 5. Since we are going to deal with integration of equivariant forms, we consider $U(1)$-equivariant cohomologies from the point of view of the Cartan model. The definition and notational conventions for integration of equivariant forms on a smooth $G$-manifold are reported in Appendix B.4, as well as an equivariant version of the Stokes’ theorem, needed for the proof of the localization formulas that are presented in the following. ### 3.1 Equivariant localization principle Let $U(1)$ act (smoothly) on a compact oriented $n$-dimensional manifold $M$ without boundaries,111If not specified a manifold is always “without boundaries” since, strictly speaking, manifolds with boundaries have to be defined in an appropriate separated way. In particular, near points at the boundary the manifold is locally homeomorphic not to an open set in $\mathbb{R}^{n}$, but to an _half-open_ disk in $\mathbb{R}^{n}$. with fixed point set $F\subseteq M$, and consider the integral of a generic $U(1)$-invariant top-form $\int_{M}\alpha\qquad\quad\text{with}\ \alpha\in\Omega^{n}(M)^{U(1)}.$ (3.1) As we saw in Example 2.4.1, in some cases we can find an _equivariantly closed extension_ $\tilde{\alpha}\in\Omega_{U(1)}(M)$ such that $d_{C}\alpha=0$, with $d_{C}=d+\iota_{T}$ (3.2) and $T\cong i$ being the generator of $U(1)$.222Notice that we have localized the Cartan model and set $\phi=-1$, as discussed in Example 2.4.1. This will be our standard convention up to Chapter 6. Then we can deform the integral without changing its value, $I[\tilde{\alpha}]:=\int_{M}\tilde{\alpha}=\int_{M}\alpha$ (3.3) since only the top-degree component $\alpha$ is selected by integration. We are going to argue now that such integration of an equivariantly closed form is completely captured by its values at the fixed point locus $F$, using two different arguments. The first is cleaner, the second less explicit but more common especially in the physics literature. We are going to need in both cases some preliminary facts, that we collect in the following lemma. ###### Lemma 3.1.1. 1. (i) If $G$ is a compact Lie group, any smooth $G$-manifold $M$ admits a $G$-invariant Riemannian metric. In other words, $G$ acts via isometry on $M$, and the fundamental vector field $\underline{T}$ is a Killing vector field,333This follows from two facts: if a $G$-action on $M$ is smooth and _proper_ , then $M$ admits a $G$-invariant Riemannian structure [31]; also, it is easy to prove that any smooth action of a _compact_ Lie group is proper. $\mathcal{L}_{T}g=0.$ 2. (ii) If $G$ is a connected Lie group, then the fixed point locus is the zero locus of all the fundamental vector fields:444This is just reasonable, see [18] for a proof. Connectedness is required because we passed from the action of $G$ to the action of $\mathfrak{g}$ by the exponential map. $F\cong\\{\left.p\in M\right|\underline{A}_{p}=0\quad\forall A\in\mathfrak{g}\\}.$ 3. (iii) For any point $p\in M$, the stabilizer of $p$ under the action of a Lie group $G$ is a closed subgroup of $G$.555By continuity of the action, every sequence inside the stabilizer of $p$ converges inside the stabilizer. ##### $1^{st}$ argument: Poincaré lemma For simplicity, suppose that $F$ contains only isolated fixed points. From lemma (i), we can pick any $U(1)$-invariant metric on $M$, and define through it _open balls_ of radius $\epsilon$ $B(p,\epsilon)$ around any fixed point $p\in F$. Then $U(1)$ acts without fixed points on the complement $\tilde{M}(\epsilon):=M\setminus\bigcup_{p\in F}B(p,\epsilon),$ (3.4) that is a manifold _with boundaries_ , them being the union of the surfaces of the balls at every fixed point (oriented in the opposite direction to the usual one). From lemma (iii), the stabilizer of any point in $\tilde{M}$ is a closed subgroup of $U(1)$, but it cannot be $U(1)$ since we excluded the fixed points, so it is discrete.666The closed subgroups of $U(1)$ are $U(1)$ and the finite cyclic groups $\\{1\\},\mathbb{Z}/n$ with $n\in\mathbb{Z}$. This means that the $U(1)$-action on $\tilde{M}$ is locally free. We would like to find an equivariant version of the Poincaré lemma on $\tilde{M}$, where the action is locally free. This means finding a map $K:\Omega(\tilde{M})^{U(1)}\to\Omega(\tilde{M})^{U(1)}$ of odd-degree such that $[d_{C},K]_{+}=id$. If we are able to find such a map, then any equivariantly closed form $\eta\in\Omega(\tilde{M})^{U(1)}$ is also equivariantly exact, $\eta=(Kd_{C}+d_{C}K)\eta=K(d_{C}\eta)+d_{C}(K\eta)=d_{C}(K\eta).$ (3.5) We can define the map $K$ by multiplication with respect to an equivariant form $\xi\in\Omega(\tilde{M})^{U(1)}$ of pure odd-degree such that $d_{C}\xi=1$, since $[\xi,d_{C}]_{+}=\xi d_{C}+(d_{C}\xi)+(-1)^{\textrm{deg}(\xi)}\xi d_{C}=1.$ (3.6) This form can be defined using again a $U(1)$-invariant metric on $M$, that we call $g$. We define the following 1-form away from the fixed point set, where $\underline{T}=0$, $\beta:=\frac{1}{g(\underline{T},\underline{T})}g(\underline{T},\cdot)$ (3.7) and notice that it is $U(1)$-invariant by invariance of $g$, and $\iota_{T}\beta=1$, so that the action of the Cartan differential on it gives $d_{C}\beta=d\beta+1$. Then the odd-degree form $\xi$ can be defined as $\xi:=\beta(d_{C}\beta)^{-1}=\beta\left(1+d\beta\right)^{-1}=\beta\sum_{i=0}^{n-1}(-1)^{i}(d\beta)^{i}.$ (3.8) The inverse of the form $(1+d\beta)$ can be guessed pretending that $d\beta$ is a number, and using the Taylor expansion $(1+z)^{-1}=\sum_{i=0}^{\infty}(-1)^{i}z^{i}.$ In the case of forms, the sum at the RHS stops at finite order, since by degree reasons $(d\beta)^{i}=0$ for $i>(n/2)$. It is easy to check that $(d_{C}\beta)^{-1}(d_{C}\beta)=1$, $d_{C}\xi=1$, and $\textrm{deg}(\xi)$ is odd. Now we know that any equivariantly closed form in $\Omega(\tilde{M})^{U(1)}$ is also equivariantly exact, so we can simplify the integral $I[\alpha]$ of an equivariantly closed form $\alpha$ using an equivariant version of Stokes’ theorem (see Appendix B.4): $\int_{\tilde{M}}\alpha=\int_{\tilde{M}}d_{C}(\xi\alpha)=\int_{\partial\tilde{M}}\xi\alpha.$ (3.9) Taking the limit $\epsilon\to 0$, the domain of integration on the LHS covers all $M$, and the integral over the boundary on the RHS reduces to a sum of integrals over the boundaries of $n$-spheres centered at each fixed point $p\in F$ (since $\partial M=\emptyset$). Thus the integral of an equivariantly closed form “localizes” as a sum over the fixed points of the $U(1)$-action, $I[\alpha]=\int_{M}\alpha=\lim_{\epsilon\to 0}\int_{\tilde{M}(\epsilon)}\alpha=\sum_{p\in F}\lim_{\epsilon\to 0}\left(-\int_{S^{n-2}_{\epsilon}(p)}(\xi\alpha)\right)=\sum_{p\in F}c_{p}$ (3.10) for some contributions $c_{p}$ at each fixed point. The precise form of these contributions will be discussed in the next section. ##### $2^{nd}$ argument: localization principle The second argument for the localization of the equivariant integral is less explicit, but more direct. Also, it is closer to the approach we will use in the infinite-dimensional context of supersymmetric QFT. Again, we start from the integral $I[\alpha]$ of an equivariantly closed form $\alpha\in\Omega(M)^{U(1)}$. The basic idea is to take advantage of the equivariant cohomological nature of the integral over $M$: this depends really on the cohomology class of the integrand, not on the particular representative. So we can deform the integral staying in the same class in a way that simplifies its evaluation, without changing the final result. To do this, we pick a positive definite $U(1)$-invariant 1-form $\beta$ on $M$, and define the new integral $I_{t}[\alpha]:=\int_{M}\alpha e^{-td_{C}\beta}$ (3.11) with $t\in\mathbb{R}$. It is again an integral of an equivariantly closed form, $d_{C}\left(\alpha e^{-td_{C}\beta}\right)=(d_{C}\alpha)e^{-td_{C}\beta}-t\alpha(d_{C}^{2}\beta)e^{-td_{C}\beta}=0,$ (3.12) since $d_{C}^{2}=\mathcal{L}_{T}$ and $\beta$ is $U(1)$-invariant. To show that this integral is equivalent to $I[\alpha]$, we show that it is independent on the parameter $t$: $\displaystyle\frac{d}{dt}I_{t}[\alpha]$ $\displaystyle=\int_{M}\alpha(-d_{C}\beta)e^{-td_{C}\beta}$ (3.13) $\displaystyle=-\int_{M}d_{C}\left(\alpha\beta e^{-td_{C}\beta}\right)\qquad$ (integration by parts) $\displaystyle=0\qquad$ $\displaystyle\text{(equivariant Stokes' theorem)}.$ Noticing that $I[\alpha]=I_{t=0}[\alpha]$, from the $t$-independence it follows that $I[\alpha]=I_{t}[\alpha]$ for every value of the parameter. We showed that the deformation via the exponential $e^{-td_{C}\beta}$ does not change the equivariant cohomology class of the integrand, so we are free to compute the integral for any value of the parameter. In particular, in the limit $t\to\infty$, we see that the only contributions come from the zero locus of the exponential. This gives the “localization formula” $\int_{M}\alpha=\lim_{t\to\infty}\int_{M}\alpha e^{-td_{C}\beta},$ (3.14) that will be the starting point for all the applications of the equivariant localization principle of the next chapters, also in the infinite-dimensional case in which $M$ describes generically the “space of fields” of a given QFT. The 1-form $\beta$ is usually called “localization 1-form”. Notice that choosing different localization 1-forms produces different practical localization schemes, but at the end of the computation they must all agree on the final result! In particular, by lemma (i) we can pick a $U(1)$-invariant Riemannian metric $g$, and choose the 1-form as $\beta:=g(\underline{T},\cdot).$ (3.15) This makes it positive definite and produces the same localization scheme of the first argument, since its zeros coincide with the zeros of the fundamental vector field $\underline{T}$ and thus with the fixed point locus $F$ of the circle action, by lemma (ii). ### 3.2 The ABBV localization formula for Abelian actions Here we state the celebrated result by Atiyah-Bott and Berline-Vergne, about the localization formulas for circle and torus actions. The rationale of the last section showed that the equivariant cohomology of the manifold $M$ is encoded in the fixed point set $F$ of the symmetry action, but left us with the evaluation of an integral over the fixed point set. We show the result of this integration here, and we are going to give an argument for the proof in the next chapter, with some tools from supergeometry. That proof is different from the original ones in [6, 7], but will introduce a method that can be easily generalized to functional integrals. To warm up, we consider first the simple case of isolated fixed point set $F\subseteq M$, and a $U(1)$-action. Notice that, at any fixed point $p\in F$, the circle action gives a representation of $U(1)$ on the tangent space, since for any $\psi\in U(1)$ $(\psi\cdot)_{*}:T_{p}M\to T_{\psi\cdot p}M\equiv T_{p}M,$ (3.16) so $(\psi\cdot)_{*}\in GL(T_{p}M)$. Since $T_{p}M$ is finite dimensional, it can be decomposed in irreducible representations of $U(1)$, $T_{p}M\cong V_{1}\oplus\cdots\oplus V_{n}.$ (3.17) The circle has to act faithfully on $T_{p}M$, since if there was $v\in T_{p}M$ such that $(\psi\cdot)_{*}v=v$, then the whole curve $\exp(tv)=\exp(t(\psi\cdot)_{*}v)=\psi\cdot\exp(tv)$ would be fixed by $U(1)$, thus $p$ would not be isolated. Recall that the irreducible representations of $U(1)$ are complex 1-dimensional, and are labeled by integers, $\psi=e^{ia}\in U(1),\qquad\rho_{m}(\psi):=e^{ima}\quad\text{with}\ m\in\mathbb{Z}.$ (3.18) This means that the irreducible representations in (3.17) are all non-trivial (of real dimension 2), and that $\dim(M)=2n$. In other words, if a circle action on a manifold $M$ has isolated fixed points, $M$ must be even- dimensional. Excluding the trivial representation with $m=0$, the tangent spaces at the fixed points are thus labeled by a set of integers, $T_{p}M\cong V_{m_{1}}\oplus\cdots\oplus V_{m_{n}}$ (3.19) where $(m_{1},\cdots,m_{n})\in\mathbb{Z}^{n}$ are called the _exponents_ of the circle action at $p\in F$.777In terms of the Lie algebra representation, every exponent $m$ coincide with the _weight_ of the single generator of $U(1)$ in the fundamental representation. They can be regarded as maps $m_{i}:F\to\mathbb{Z}$. We formulate now a simplified version of the localization theorem in term of this local data. The proof of this can be found in [18]. ###### Theorem 3.2.1 (Localization for circle actions). Let $U(1)$ act on a compact oriented manifold $M$ of dimension $\dim(M)=2n$, with isolated fixed point locus $F$. If $m_{1},\cdots,m_{n}:F\to\mathbb{Z}$ are the exponents of the circle action, and $\alpha=\alpha^{(2n)}+\alpha^{(2n-2)}\phi+\alpha^{(2n-4)}\phi^{4}+\cdots+\alpha^{(0)}$ is an equivariant top-form in $\Omega_{U(1)}(M)$ such that $d_{C}\alpha=0$, then $\boxed{\int_{M}\alpha^{(2n)}=\int_{M}\alpha=(2\pi)^{n}\sum_{p\in F}\frac{\alpha^{(0)}(p)}{m_{1}(p)\cdots m_{n}(p)}}$ where the last component $\alpha^{(0)}\in C^{\infty}(M)$. ###### Example 3.2.1 (Localization on the 2-sphere). Let us consider again the case of the height function $H:\mathbb{S}^{2}\to\mathbb{R}$ such that, in spherical coordinates $(\theta,\varphi)$, $H(\theta,\varphi):=\cos(\theta)$. In Example 2.4.1 we related this function to the equivariantly closed extension of the volume form on the 2-sphere, $\tilde{\omega}=\omega+H.$ We can use the last localization theorem to compute integrals involving this “Hamiltonian” function on $\mathbb{S}^{2}$. The 2-sphere has two isolated fixed points at the poles, and only one exponent $m:F\to\mathbb{Z}$. It is not difficult to see that the exponent of the action at the fixed points is $m(N)=1$ at the North pole, and $m(S)=-1$ at the South pole (the sign comes from the orientation of the charts). We can check the theorem with two instructive integrals. The first is simply the area of the sphere, i.e. the integral of $\omega$. Using the theorem we easily get the correct result, $\int_{\mathbb{S}^{2}}\omega=\int_{\mathbb{S}^{2}}(\omega+H)=2\pi\sum_{p\in\\{N,S\\}}\frac{H(p)}{m(p)}=2\pi\left(\frac{\cos(0)}{1}+\frac{\cos(\pi)}{-1}\right)=4\pi.$ The second integral is the “partition function”on the sphere, $Z(t):=\int_{\mathbb{S}^{2}}\omega e^{itH}=\frac{1}{it}\int_{\mathbb{S}^{2}}e^{it(H+\omega)}$ where the second equality comes from degree arguments. This is the integral of an equivariantly closed form, since $d_{C}e^{it(H+\omega)}\propto d_{C}\tilde{\omega}=0$, whose $C^{\infty}(\mathbb{S}^{2})$ component is given by $e^{itH}$. Using the localization theorem we get $Z=\frac{1}{it}2\pi\left(\frac{e^{it\cos(0)}}{1}+\frac{e^{it\cos(\pi)}}{-1}\right)=4\pi\frac{\sin(t)}{t}$ matching the result from the “semiclassical” saddle-point approximation (1.7). We now get to the main theorem, considering a more generic torus action with higher dimensional fixed point locus on $M$. ###### Theorem 3.2.2 (Atiyah-Bott [6], Berline-Vergne [7]). Let the torus $T=U(1)^{l}$ of dimension $l$ act on a compact oriented $d$-dimensional manifold $M$, with fixed point locus $F$. If $\alpha\in\Omega_{T}(M)$ is an equivariantly closed form, i.e. $d_{C}\alpha=0$, and $i:F\hookrightarrow M$ is the inclusion map, then $\boxed{\int_{M}\alpha=\int_{F}\frac{i^{*}\alpha}{\left.e_{T}(R)\right|_{N}}}$ where $\left.e_{T}(R)\right|_{N}$ is the _T-equivariant Euler class_ of the normal bundle of $F$ in $M$. This is the localization formula as originally presented for a torus action and fixed point locus $F$, that is generically an embedded (regular) submanifold of $M$. The _normal bundle_ to $F$ can be regarded as $TN=\faktor{TM}{i_{*}TF},$ (3.20) where the quotient is taken pointwise at any $p\in F$, so that the tangent bundle of $M$ is split as $TM=i_{*}TF\oplus TN$. The finite sum is replaced by an integral over $F$, and the zero-degree component of $\alpha$ is replaced by the component with the correct dimensionality, that matches $\dim(F)$, by pulling-back $\alpha$ on $F$. The product of the exponents at the denominator is represented in general by the equivariant Euler class of the normal bundle, $\left.e_{T}(R)\right|_{N}=\mathrm{Pf}_{N}\left(\frac{R^{T}}{2\pi}\right)=\mathrm{Pf}_{N}\left(\frac{R+\mu}{2\pi}\right),$ (3.21) where the pfaffian is taken over the coordinates that span the normal bundle $TN$, $R$ is the curvature of an invariant Riemannian metric on $M$, $\mu:\mathfrak{t}\to\Omega^{0}(M;\mathfrak{gl}(d))$ is the “moment map” that makes $R^{T}$ an equivariant extension of the Riemannian curvature in the Cartan model (see Appendix B.1). As an example, let us apply the ABBV localization formula in the case of a discrete fixed point set $F$, so that we can recover at least the more readable version of theorem 3.2.1. The normal bundle in this case is the whole tangent bundle and, since $F$ is 0-dimensional, the restriction of the equivariant curvature $R^{T}$ to $F$ makes only its $\Omega^{0}$ component contribute, so $\mathrm{Pf}(R^{T})=\phi^{a}\otimes\mathrm{Pf}(\mu_{a})$. At an isolated fixed point $p$, as we said before, the tangent space $T_{p}$ is a representation space for the torus action. Since the torus is Abelian, analogously to the above discussion this representation can be decomposed as the sum of 2-dimensional _weight spaces_ [32, 22], $T_{p}M\cong\bigoplus_{i=1}^{d/2}V_{v_{i}}.$ (3.22) In Section 4.2 we will see that the moment map at an isolated fixed point encodes exactly these weights, being the representation $\mu(p):\mathfrak{t}\to\mathfrak{gl}(d)\cong\mathrm{End}(T_{p}M)$. The equivariant Euler class computes exactly the product of the weights, $e_{T}(R)_{p}=\frac{1}{(2\pi)^{d/2}}\prod_{i}v_{i}=\frac{1}{(2\pi)^{d/2}}\prod_{i}\phi^{a}\otimes v_{i}(T_{a}),$ (3.23) where $T_{a}$ are the generators of $T$. This recovers the formula for the circle action in theorem 3.2.1, where the exponents play the role of the weights for the single generator of $\mathbb{S}^{1}$. Notice that, as it is clear from the above example, in the generic $l$-dimensional case it is not so convenient to forget about the generators $\\{\phi^{a}\\}$ of $\mathfrak{t}^{*}$, and the ABBV localization formula should be thought as an equivalence of elements in $H^{*}_{T}(pt)=\mathbb{R}[\phi^{1},\cdots,\phi^{l}]$. The LHS is clearly polynomial in $\phi^{a}$, so has to be the RHS. Since in the latter both the numerator and the denominator are polynomials in $\phi^{a}$, some simplification has to occur in the rational expression to give a polynomial as the final answer. ###### Remark. We anticipate that in QFT the pfaffian in the definition of the Euler class is usually realized in terms of a Gaussian integral over Grassmann (anticommuting) variables, as we will see in detail in Section 4.2. These “fermionic” Gaussian integrals arise naturally as “1-loop determinants” from some saddle-point (semi-classical) approximation technique to the partition function of the theory, for example. In general, the differential form $\alpha$ will be an “observable” of the QFT, and the equivariantly closeness condition will be interpreted as it being “supersymmetric”. The localization locus $F$ will be then the fixed point set of a symmetry group that is the “square” of this supersymmetry (as $d_{C}^{2}\propto\mathcal{L}_{T}$ schematically), so a Poincaré symmetry or a gauge symmetry. The integral then localizes onto the “moduli space” of gauge-invariant (or BPS) field configurations. In the context of Hamiltonian mechanics, the gauge symmetry can be one generated by the dynamics of the theory itself, and in this case the path integral localizes onto the classical solutions of the equations of motion. The ABBV formula thus gives a systematic way to understand in which cases the semi-classical approximation results to be exact. We will reexamine this point of view in the next section in the context of finite-dimensional Hamiltonian mechanics, while in Chapter 5 we will describe the infinite- dimensional case of QM and QFT, giving some examples of the ABBV localization formula at work. ### 3.3 Equivariant cohomology on symplectic manifolds As we remarked at the beginning of the chapter, the localization formulas of the last section can be seen as generalizing a similar result showed by Duistermaat and Heckman [5] in the context of Hamiltonian group actions on symplectic manifolds. This special case is of fundamental importance in physics, because this is the context in which classical Hamiltonian mechanics is constructed. In some special cases also the quantum theory can be formally given such a structure, and thus some results from symplectic geometry can be extended to QM and QFT in general. We begin this section by quickly recalling some basic concepts about symplectic and Hamiltonian geometry, then we will describe how this can be seen as a special case of equivariant cohomology theory from the point of view of the localization formulas. #### 3.3.1 Pills of symplectic geometry The notion of _phase space_ can be constructed in a basis-independent way in differential geometry through the definition of _symplectic manifold_. We suggest for example [33, 28, 34] for a complete introduction to the subject. ###### Definition 3.3.1. A _symplectic manifold_ is a pair $(M,\omega)$, where $M$ is a $2n$-dimensional smooth manifold, and $\omega$ is a _symplectic form_ on $M$: 1. (i) $\omega\in\Omega^{2}(M)$; 2. (ii) $d\omega=0$; 3. (iii) $\omega$ is non-degenerate. The fact that $M$ is even-dimensional is not really a requirement but a consequence of its symplectic structure. This is because any skew-symmetric bilinear map on a $d$-dimensional vector space can be represented in a suitable basis by the matrix $\left(\begin{array}[]{c|cc}\mathbf{0}_{k}&\mathbf{0}&\mathbf{0}\\\ \hline\cr\mathbf{0}&\mathbf{0}&-\mathds{1}_{n}\\\ \mathbf{0}&\mathds{1}_{n}&\mathbf{0}\end{array}\right)$ (3.24) with $2n+k=d$. To be non degenerate, it must be $k=0$. The symplectic form is a skew-symmetric bilinear form on $T_{p}M$ at any point $p\in M$, so the even- dimensionality of $M$ follows from its non-degeneracy. On manifolds, a stronger result than the above one holds: the so-called _Darboux theorem_. It states that, for every point $p\in M$, there exists an entire open neighborhood $U_{p}\subseteq M$ and a coordinate system $x:U_{p}\to\mathbb{R}^{2n}$ with respect to which $\omega_{\mu\nu}=\omega(\partial_{\mu},\partial_{\nu})$ has the canonical form (3.24), with $k=0$. The coordinates $x$ are called _Darboux coordinates_.888This means that all symplectic manifolds look locally as the prototype $\mathbb{R}^{2n}$ with $\omega=\sum_{i=1}^{n}dx^{i}\wedge dx^{i+n}$. This is a very strong property, compared for example with the Riemannian case. Notice that from the non-degeneracy of $\omega$ we have a canonical choice for the volume form on $M$, the so-called _Liouville volume form_ $\mbox{vol}:=\frac{\omega^{n}}{n!}=\mathrm{Pf}||\omega^{(x)}_{\mu\nu}||d^{2n}x=dp_{1}\wedge dp_{2}\wedge\cdots dp_{n}\wedge dq^{1}\wedge dq^{2}\wedge\cdots\wedge dq^{n},$ (3.25) where $(q^{\mu},p_{\mu})_{\mu=1,\cdots,n}$ are Darboux coordinates. The closeness of $\omega$ implies that in some cases there can be a 1-form $\theta\in\Omega^{1}(M)$ such that $d\theta=\omega.$ (3.26) Such a 1-form, if it exists, is called _symplectic potential_. In practice, sometimes it is useful to _locally_ define a symplectic potential even if $\omega$ is not globally integrable. Isomorphisms of symplectic manifolds are called _symplectomorphisms_ or _canonical transformations_ , defined as diffeomorphisms that preserve the symplectic structure via pull-back. The standard example of a symplectic manifold is exactly the _phase space_ associated to some $n$-dimensional _configuration space_ $Q$, i.e. its cotangent bundle $M:=T^{*}Q$. A point $p\in Q$ represents the “generalized position” of the system with coordinates $q(p)=(q^{\mu}(p))$ with $\mu=1,\cdots,n$, and a point $p\in T^{*}Q$ represents the “generalized momentum”, with coordinates $\xi(p):=(q^{\mu}\circ\pi(p),\iota_{\mu}(p))\equiv(q^{\mu},p_{\mu})$, where $\pi:TQ\to Q$ is the projection and $\iota_{\mu}\equiv\iota_{\partial/\partial q^{\mu}}$. The cotangent bundle has a canonical integrable symplectic form. In fact, the symplectic potential is the so-called _tautological 1-form_ given by the pull-back of the projection map, $\theta:=\pi^{*}\in\Omega^{1}(T^{*}Q)$. In Darboux coordinates, at a point $p\in T^{*}Q$, $\theta_{p}=\pi^{*}(p)=p_{\mu}dq^{\mu}$ (3.27) where we denoted $dq^{\mu}\equiv d(q\circ\pi)^{\mu}=d\xi^{\mu}$ with $\mu=1\cdots,n$, as 1-forms on the cotangent bundle. The canonical symplectic form is then just $\omega=d\theta$, and in Darboux coordinates $\omega=dp_{\mu}\wedge dq^{\mu}$ (3.28) where again we simplified the notation setting $dp_{\mu}\equiv d\xi^{\mu}$ for $\mu=n+1,\cdots,2n$. Thus the canonical coordinates on the cotangent bundle are Darboux coordinates. One can show that canonical symplectic structures over diffeomorphic manifolds are “canonically compatible”, i.e. if $\phi:Q_{1}\to Q_{2}$ is a diffeomorphism, there is a lift of it as a symplectomorphism between $(T^{*}Q_{1},\omega_{1})$ and $(T^{*}Q_{2},\omega_{2})$. If we take $Q_{1}=Q_{2}$, this means that there is a group homomorphism $\text{Diff}(Q)\to\text{Symp}(T^{*}Q,\omega).$ (3.29) This example showed that symplectic manifolds are the right generalization of the concept of phase space in a fully covariant setting. It is thus common to call functions on a symplectic manifold _observables_. Let us return to a generic symplectic manifold $(M,\omega)$. Giving to it some additional structure, it is possible to define on it _dynamics_ and _symmetries_ in the sense of classical mechanics. Naturally, we call symmetry of $(M,\omega)$ a diffeomorphism $\phi:M\to M$ that preserves the symplectic structure, $\phi^{*}\omega=\omega$, that is a symplectomorphism. At the infinitesimal level, a diffeomorphism can be generated by the flow of a vector field $X\in\Gamma(M)$, and the symmetry condition is rephrased to $\mathcal{L}_{X}\omega=0.$ (3.30) Such a vector field is called _symplectic vector field_. It is easy to realize that a vector field is symplectic if and only if $\iota_{X}\omega=\omega(X,\cdot)$ is closed, by Cartan’s magic formula. More special vector fields are those for which $\iota_{X}\omega$ is exact, so that it exists an observable $f\in C^{\infty}(M)$ such that $df=-\iota_{X}\omega,$ (3.31) where the minus sign is conventional. The vector field $X$ is called _Hamiltonian vector field_ associated to the observable $f$. In components, $\partial_{\mu}f=\omega_{\mu\nu}X^{\nu}\qquad\text{or}\qquad X^{\mu}=\omega^{\mu\nu}\partial_{\nu}f,$ (3.32) where $\omega^{\mu\nu}$ is the “inverse” of the symplectic form. Of course Hamiltonian vector fields are symplectic, and the flow of the Hamiltonian vector field $X$ preserves the value of the Hamiltonian function $f$, since $\mathcal{L}_{X}(f)=X(f)=df(X)=\omega(X,X)=0$. The flow of the Hamiltonian vector field is regarded as the “time-evolution” over the generalized phase space $M$, generated by the observable $f$. ###### Definition 3.3.2. An _Hamiltonian (or dynamical) system_ is a tuple $(M,\omega,H)$, where $(M,\omega)$ is a symplectic manifold and $H\in C^{\infty}(M)$ an observable called _Hamiltonian_. The _time-evolution_ of points $p\in M$ is defined by the flow of the Hamiltonian vector field $X_{H}$ of $H$, $p(t):=\gamma^{H}_{p}(t)$ where $\gamma^{H}_{p}$ is the integral curve of $X_{H}$ with $\gamma^{H}_{p}(0)=p$. In particular, the evolution of an observable $f\in C^{\infty}(M)$ is regulated by the _equation of motion_ $\dot{f}(p):=(f\circ\gamma^{H}_{p})^{\prime}(0)\equiv\left.\mathcal{L}_{X_{H}}(f)\right|_{p}.$ The equation of motion can be rewritten in a more usual way introducing the _Poisson brackets_ $\\{\cdot,\cdot\\}:C^{\infty}(M)\times C^{\infty}(M)\to C^{\infty}(M)$ such that $\\{f,g\\}:=\omega(X_{g},X_{f})$, where $X_{f},X_{g}$ are the Hamiltonian vector fields of $f$ and $g$, respectively. In a chart and with respect to Darboux coordinates $(q^{\mu},p_{\mu})$ on $M$, by the Darboux theorem the Poisson brackets take the usual form $\\{f,g\\}=\frac{\partial f}{\partial q^{\mu}}\frac{\partial g}{\partial p_{\mu}}-\frac{\partial g}{\partial q^{\mu}}\frac{\partial f}{\partial p_{\mu}}.$ (3.33) With this definition we can write $\dot{f}=-\\{H,f\\},\qquad\dot{q}^{\mu}=\frac{\partial H}{\partial p_{\mu}},\qquad\dot{p}_{\mu}=-\frac{\partial H}{\partial q^{\mu}},$ (3.34) recovering the Hamilton’s equations for the Darboux coordinates. The Poisson brackets are anti-symmetric and satisfy the Jacobi identity, so this turns $(C^{\infty}(M),\\{\cdot,\cdot\\})$ into a Lie algebra,999In fact this is a _Poisson algebra_ , i.e. a Lie algebra whose brackets act as a derivation. and one can check that there is a Lie algebra homomorphism $\displaystyle(C^{\infty}(M),\\{\cdot,\cdot\\})$ $\displaystyle\to(\text{Hamiltonian v.f.},[\cdot,\cdot])$ (3.35) $\displaystyle f$ $\displaystyle\mapsto X_{f},$ where we also already used the fact that Hamiltonian vector fields form a Lie subalgebra with respect to the standard commutator on $\Gamma(TM)$. We just reviewed that the concept of symmetry in symplectic geometry is correlated with the concept of dynamics on the symplectic manifold. The next fact that we need is to connect this formalism to the equivariant cohomology one, identifying these symmetries as generated by a _group action_ on $M$. In particular, we would like to identify the Lie subalgebra of Hamiltonian vector fields as the Lie algebra of a Lie group that acts on the symplectic manifold. We can start thus the discussion of symmetry by declaring that $M$ is a $G$-manifold with respect to a Lie group $G$ of Lie algebra $\mathfrak{g}$. Denoting the $G$-action as $\rho$, this is called _symplectic_ if it makes $G$ act by symplectomorphisms on $(M,\omega)$, i.e. $\rho:G\to\text{Symp}(M,\omega).$ (3.36) We can characterize again infinitesimally this action by saying that $\mathfrak{g}$ acts on $\Omega(M)$ via symplectic vector fields: if $A\in\mathfrak{g}$, the corresponding fundamental vector field preserves the symplectic structure, $\mathcal{L}_{A}\omega=0$. We are interested in the special case analogous to the one above, in which not only a fundamental vector field is symplectic, but it is also Hamiltonian. This forces a generalization of the concept of Hamiltonian function, because now there are more than one independent fundamental vector fields to take into account, if $\dim(\mathfrak{g})>1$. ###### Definition 3.3.3. The $G$-action $\rho:G\to\text{Symp}(M,\omega)$ on the symplectic manifold $(M,\omega)$ is said to be an _Hamiltonian action_ if every fundamental vector field is Hamiltonian. In particular, there exists a $\mathfrak{g}^{*}$-valued function $\mu\in\mathfrak{g}^{*}\otimes C^{\infty}(M)$ such that: 1. (i) For every $A\in\mathfrak{g}$, $\mu(A)\equiv\mu_{A}\in C^{\infty}(M)$ is the Hamiltonian function with respect to $\underline{A}$, $d\mu_{A}=-\iota_{A}\omega=\omega(\cdot,\underline{A}).$ 2. (ii) It is $G$-equivariant with respect to the canonical (co)adjoint action of $G$ on $\mathfrak{g}\ (\mathfrak{g}^{*})$,101010If, for every $g\in G$, $Ad_{g}:G\to G$ is the action by conjugation, the adjoint action $Ad_{*g}$ on $\mathfrak{g}$ is the push-forward of $Ad_{g}$, while the coadjoint action $Ad^{*}_{g}$ on $\mathfrak{g}^{*}$ is the pull-back of $Ad_{g^{-1}}$. If $g=\exp(tA)$ for some $A\in\mathfrak{g}$, differentiating one gets the infinitesimal actions of $\mathfrak{g}$ on $\mathfrak{g}$ and $\mathfrak{g}^{*}$, $ad_{A}(B)=[A,B]$ and $ad^{*}_{A}(\eta):=\eta([\cdot,A])$. so for any $g\in G$ $\mu\circ Ad_{*g}=\rho_{g}^{*}\circ\mu\qquad\text{or}\qquad Ad^{*}_{g}\circ\mu=\mu\circ\rho_{g},$ where in the first equation $\mu$ is considered as $\mathfrak{g}\xrightarrow{\mu}C^{\infty}(M)$, in the second one as $M\xrightarrow{\mu}\mathfrak{g}^{*}$. If $G$ is connected, this is equivalent to requiring $\mu:\mathfrak{g}\to C^{\infty}(M)$ to be a Lie algebra anti- homomorphism with respect to the Poisson brackets, $\mu_{[A,B]}=\\{\mu_{B},\mu_{A}\\}\qquad\forall A,B\in\mathfrak{g}.$ The map $\mu$ is called _moment map_ , and $(M,\omega,G,\mu)$ is called _Hamiltonian $G$-space_. In general the job of the moment map is to collect all the “Hamiltonians” with respect to which the system can flow. There are $\dim(G)$ independent of them, one for every generator. In the 1-dimensional case, where $G=U(1)$ (or its non-compact counterpart $G=\mathbb{R}$), the moment map produces only one independent Hamiltonian, $\mu_{T}\equiv H$, and the above definition reduces to the Hamiltonian system $(M,\omega,H)$ of definition 3.3.2. Notice that for any Hamiltonian structure we build on $(M,\omega)$, its flow preserves the symplectic form and thus the canonical Liouville volume form $\omega^{n}/n!$. This is the content of the so-called _Liouville theorem_. #### 3.3.2 Equivariant cohomology for Hamiltonian systems We can first generalize what we noticed in examples 2.4.1 and 3.2.1, in the case of a circle action on a symplectic manifold $(M,\omega)$. In the above examples the manifold was the 2-sphere $\mathbb{S}^{2}$ and the symplectic form was the canonical volume form. Rephrased in terms of symplectic geometry, the existence of an _equivariantly closed extension_ $\tilde{\omega}$ of the symplectic form is the condition of $U(1)$ acting in an Hamiltonian way, since $d_{C}\tilde{\omega}=d_{C}(\omega+H)=\iota_{T}\omega+dH=0$ (3.37) is satisfied if and only if $dH=-\iota_{T}\omega$. This is readily generalizable to the multidimentional case, so that we can describe the Hamiltonian $G$-space $(M,\omega,G,\mu)$ and its classical mechanics in equivariant cohomological terms. In fact, we can always find an equivariantly closed extension of the symplectic form in $\Omega_{G}(M)$, $\tilde{\omega}:=1\otimes\omega-\phi^{a}\otimes\mu_{a}$ (3.38) where $\mu_{a}\equiv\mu_{T_{a}}\in C^{\infty}(M)$ and $T_{a}$ are the dual basis elements with respect to the generators $\phi^{a}$ of $S(\mathfrak{g}^{*})$. It is straightforward to check that $\tilde{\omega}\in\Omega_{G}(M)$ is indeed $G$-invariant, and closed with respect to $d_{C}$ thanks to the Hamiltonian property of the $G$-action, $d\mu_{a}=-\iota_{a}\omega$. In the language of $G$-equivariant bundles (see Appendix B.1), the symplectic structure on $M$ can be seen as the presence of a principal $U(1)$-bundle $P\to M$ whose connection 1-form is the symplectic potential $\theta$ (that has not always a global trivialization on $M$), and whose curvature is the symplectic 2-form $\omega=d\theta$ (that instead transforms covariantly on $M$). $G$ acts symplectically if also $\theta$ is $G$-invariant, $\mathcal{L}_{X}\theta=0\quad\Rightarrow\quad\mathcal{L}_{X}\omega=0\qquad\forall X\in\mathfrak{g},$ (3.39) so that $P\to M$ is a $G$-equivariant bundle. Thus, this equivariant extension to the curvature $\omega$ is the same as in [35, 7]. We return for a moment to the symplectic geometric interpretation, to describe the results of Duistermaat and Heckman related to the localization formulas that we described in the last section. In [5] they proved an important property of the Liouville measure in the presence of an Hamiltonian action by a torus $T$ on $(M,\omega)$. Namely, defining a measure on $\mathfrak{g}^{*}$ as the _push-forward_ of the Liouville measure, $\mu_{*}\left(\frac{\omega^{n}}{n!}\right)(U)=\int_{\mu^{-1}(U)}\frac{\omega^{n}}{n!}\qquad\forall U\subseteq M\ \text{measurable},$ (3.40) they proved that $\mu_{*}\left(\omega^{n}/n!\right)$ is a _piecewise polynomial function_.111111To be more precise, denoting $\mu_{*}\left(\omega^{n}/n!\right)=fd\xi$ with $d\xi$ the standard Lebesgue measure on $\mathfrak{g}^{*}\cong\mathbb{R}^{\dim\mathfrak{g}}$, the function $f$ is piecewise polynomial. This, and an application of the stationary phase approximation showed a localization formula for the oscillatory integral $\int_{M}\frac{\omega^{n}}{n!}\exp{\left(i\mu_{X}\right)}$ (3.41) for every $X\in\mathfrak{t}:=Lie(T)$ with non-null weight at every fixed point of the $T$-action. This can be viewed as the Fourier transform of the Liouville measure, or as the _partition function_ of a 0-dimensional QFT with target space $M$. Let the fixed point locus $F$ be the union of compact connected symplectic manifolds $M_{k}\hookrightarrow M$ of even codimension $2n_{k}$, and denote $(m_{kl})_{l=1,\cdots,n_{k}}$ the weights of the $T$-action at a tangent space of a fixed point $p\in M_{k}$.121212The components of the fixed point set being symplectic is not an assumption, but a consequence of the Hamiltonian action. See [28], proposition IV.1.3. Then the _Duistermaat-Heckman (DH) localization formula_ is $\boxed{\int_{M}\frac{\omega^{n}}{n!}\exp{\left(i\mu_{X}\right)}=\sum_{k}\frac{\mbox{vol}(M_{k})\exp{(i\mu_{X}(M_{k}))}}{\prod_{l}^{n_{k}}(m_{kl}(X)/2\pi)}}$ (3.42) where $\mu_{X}(M_{k})$ denotes the common value of $\mu_{X}$ at every point in $M_{k}$. In equivariant cohomological terms, we can see the above result as a localization formula for the integral of an equivariantly closed form. In fact, if we fix the $U(1)$ symmetry subgroup generated by $X\in\mathfrak{t}$, and consider the Cartan model defined by the differential $d_{C}=d+i\iota_{X}$, the LHS can be rewritten as $\int_{M}\frac{\omega^{n}}{n!}\exp(i\mu_{X})=\int_{M}\exp(\omega+i\mu_{X}),$ (3.43) analogously to what we did in Example 3.2.1, and this is clearly the integral of an equivariantly closed form with respect to the differential $d_{C}$. To see the correspondence with the ABBV formula, let us examine the case of a circle action and discrete fixed point locus $F$. In this case we have only one Hamiltonian function $H:=\mu_{X}$, the weights are just the exponents $m_{1},\cdots,m_{n}$ of the circle action, and the sum over $k$ runs over the isolated fixed points. The DH formula thus recovers exactly the localization formula of theorem 3.2.1: $\int_{M}\exp(\omega+iH)=(2\pi)^{n}\sum_{p\in F}\frac{e^{iH(p)}}{m_{1}(p)\cdots m_{n}(p)}.$ (3.44) As we remarked in (3.23), the denominator can be expressed as the equivariant Euler class of the normal bundle to $F$ (that is just the tangent bundle since $F$ is 0-dimensional), recovering the DH formula as a special case of the ABBV localization formula for torus actions. See also [19] for an explicit correspondence between the two. We wish only to remark again that, especially in the context of Hamiltonian mechanics, this localization formula can be seen as the result of an “exact” saddle-point approximation on the partition function (3.43). This is the point of view we are going to take in the next chapters, when we are going to discuss the generalization of this formula to higher-dimensional QFT, where the integral of the partition function is turned into an infinite-dimensional _path integral_. To see the correspondence with the saddle-point approximation, we recall that the isolated fixed points of the $U(1)$-action are those in which $\underline{X}=0$, so $dH=0$, and thus they are the critical points of the Hamiltonian. We need to assume that the function $H$ is _Morse_ , so that these fixed points are non-degenerate, i.e. the Hessian $\mathrm{Hess}_{p_{0}}(H)_{\mu\nu}=\partial_{\mu}\partial_{\nu}H(p_{0})$ at a given $p_{0}\in F$ has non-null determinant.131313This subject was in fact firstly connected with Morse theory by Witten in [36], where localization is applied in the context of supersymmetric QM to prove Morse inequalities. We do not need to deepen this point of view for what follows, but a discussion about Morse theory and its connection with the DH formula can be found in [19], and references therein. This Hessian can be expressed in terms of the exponents $m_{k}(p_{0})$ via an equivariant version of the Darboux theorem [37, 28]: at any fixed point we can choose Darboux coordinates in which the symplectic form takes its canonical form (3.28), and moreover the action of the fundamental vector field at that tangent space decomposes as in (3.19). The latter can then be expressed as $n$ canonical rotations of the type $\underline{X}=\sum_{\mu=1}^{n}im_{\mu}\left(q^{\mu}\frac{\partial}{\partial p_{\mu}}-p_{\mu}\frac{\partial}{\partial q^{\mu}}\right)$ (3.45) with different weights $m_{k}$. By the general form of the Hamilton’s equations (3.32), this means that the Hamiltonian near the isolated fixed point $p_{0}\in F$ can be expanded as $H(x)=H(p_{0})+\frac{1}{2}\sum_{\mu=1}^{n}im_{\mu}(p_{0})\left(p_{\mu}(x)^{2}+q^{\mu}(x)^{2}\right)+\cdots$ (3.46) Plugging this expansion into the oscillatory integral, we get the saddle-point approximation $\displaystyle\int_{M}d^{n}pd^{n}q\ e^{iH(p,q)}$ $\displaystyle\approx\sum_{p_{0}\in F}e^{iH(p_{0})}\prod_{\mu=1}^{n}\left(\int dp\ e^{-\frac{m_{\mu}(p_{0})}{2}p^{2}}\int dq\ e^{-\frac{m_{\mu}(p_{0})}{2}q^{2}}\right)$ (3.47) $\displaystyle\approx\sum_{p_{0}\in F}e^{iH(p_{0})}\frac{(2\pi)^{n}}{\prod_{\mu}m_{\mu}(p_{0})}$ that, again, is exactly the result of the localization formula above. This motivates in the context of Hamiltonian mechanics, and generalization to infinite-dimensional case, that the denominators appearing in these formulas are exactly the “1-loop determinants” of a would-be semiclassical approximation to the partition function. More aspects of the equivariant theory in contact with symplectic geometry can be found in [22]. ## Chapter 4 Supergeometry and supersymmetry ### 4.1 Gradings and superspaces We give some definitions concerning _graded_ spaces and _super_ -spaces, that are useful for many applications of the localization theorems in physics. In particular, we will see how to translate the problem of integration of differential forms in the context of supergeometry, and how this is useful to prove the ABBV localization formula for circle actions. Also, in the next chapter we will apply this theorem to path integrals in QM and QFT, where the coordinates over which one integrates are those of a “field space” over a given manifold. To construct a suitable Cartan model over this kind of spaces, it is necessary to introduce a graded structure, that physically means to distinguish between _bosonic_ and _fermionic_ fields, and some operation that acts as a “Cartan differential” transforming one type of field into the other. These structures arise in the context of _supersymmetric field theories_ , or in the context of _topological field theories_ , and the differentials here are called _supersymmetry_ transformations or _BRST_ transformations. The precise mathematics behind this is a great subject and we do not seek to be complete here, we just give some of the basic concepts that are necessary to understand what follows. For a more extensive review of the subject, we suggest for example [38]. #### 4.1.1 Definitions To understand the concept of a supermanifold, we need first to recall the linearized case. We already introduced a _graded module_ or _graded algebra_ $V$ over a ring $R$, that is a collection of $R$-modules $\\{V_{n}\\}_{n\in\mathbb{Z}}$ such that $V=\bigoplus_{n\in\mathbb{Z}}V_{n}$. If $V$ is an algebra, it must also satisfy $V_{n}V_{m}\subseteq V_{n+m}$. An element $a\in V_{n}$ for some $n$ is called _homogeneous_ of _degree_ $\mathrm{deg}(a)\equiv|a|:=n$. We now can specialize to the case of _super_ \- vector spaces (or modules) and _super_ -algebras. ###### Definition 4.1.1. A _super vector space_ is a $\mathbb{Z}_{2}$-graded vector space $V=V_{0}\oplus V_{1}$ where $V_{0},V_{1}$ are vector spaces. Its _dimension_ as a super vector space is defined as $\dim{V}:=\left(\dim{V_{0}}|\dim{V_{1}}\right)$. A _superalgebra_ is a super vector space $V$ with the product satisfying $V_{0}V_{0}\subseteq V_{0}\ ;\quad V_{0}V_{1}\subseteq V_{1}\ ;\quad V_{1}V_{1}\subseteq V_{0}.$ A _Lie superalgebra_ is a superalgebra where the product $[\cdot,\cdot]:V\times V\to V$, called _Lie superbracket_ , satisfies also $\displaystyle[a,b]=-(-1)^{|a||b|}[b,a]$ $\displaystyle(\mathrm{supercommutativity}),$ $\displaystyle(-1)^{|a||c|}[a,[b,c]]+(-1)^{|b||c|}[c,[a,b]]+(-1)^{|a||b|}[b,[c,a]]=0$ $\displaystyle(\mathrm{super\ Jacobi\ identity}).$ ###### Definition 4.1.2. The _k-shift_ of a graded vector space $V$ is the graded vector space $V[k]$ such that $(V[k])_{n}=V_{n+k}$ $\forall n\in\mathbb{Z}$. A few remarks are in order. First, it is clear that every graded vector space has naturally also the structure a super vector space, if we split its grading according to “parity”: $V_{even}:=\bigoplus_{n\in 2\mathbb{Z}}V_{n}\ ,\qquad V_{odd}:=\bigoplus_{n\in 2\mathbb{Z}+1}V_{n}.$ (4.1) In physics, the $\mathbb{Z}$-grading occurs on the “field space” as the so- called _ghost number_ , while the $\mathbb{Z}_{2}$-grading with respect to parity distinguish between _bosonic_ and _fermionic_ coordinates. Second, we notice that every vector space $V$ can be considered as a (trivial) super vector space, if we think of it as $V=V\oplus 0$ in even degree or $V[1]=0\oplus V$ in odd degree. Notice that the even/odd parts of a super vector space can be considered as eigenspaces of an automorphism $P:V\to V$ such that $P^{2}=id_{V}$. In this sense, a super vector space is a pair $(V,P)$ made by a vector space and the given automorphism $P$. Morphisms of graded vector spaces are graded linear maps, i.e. grading preserving maps: ###### Definition 4.1.3. A _graded linear map_ $f$ between graded vector spaces $V$ and $W$ is a collection of linear maps $\\{f_{k}:V_{k}\to W_{k}\\}_{k\in\mathbb{Z}}$. A linear map of _k-degree_ is a graded linear map $f:V\to W[k]$. Now we can turn to the non-linear case and consider _supermanifolds_.111Historically two (apparently) different concepts of _supermanifolds_ and _graded manifolds_ were firstly developed. They both aimed to generalizing the mathematics of manifolds to a non-commutative setting, following different approaches. Eventually it was proven in [39] that their definitions are equivalent. Locally, they can be thought as extensions of a manifold via “anticommuting coordinates”: if we take an open set $U\subset\mathbb{R}^{n}$ and a set of coordinates $\\{x^{\mu}:U\to\mathbb{R}\\}_{\mu=1,\cdots,m}$, we can consider a set of additional coordinates $\\{\theta^{i}\\}_{i=1,\cdots,n}$ with the algebraic properties $x^{\mu}\theta^{i}=\theta^{i}x^{\mu}\ ,\qquad\theta^{i}\theta^{j}=-\theta^{j}\theta^{i}.$ (4.2) The anticommuting $\\{\theta^{i}\\}$ can be thought as generators of $\bigwedge(V^{*})$ for some vector space $V$, and the product between them and coordinates of $C^{\infty}(U)$ is then interpreted as a tensor product in $C^{\infty}(U)\otimes\bigwedge(V^{*})=:C^{\infty}(U\times V[1])$.222Being generators of an exterior algebra, $\theta^{i}$ are called “Grassmann-odd” coordinates, while $x^{\mu}$ are called “Grassmann-even” consequently. This terminology is commonly inherited by every graded object (vector fields, forms, etc.) on the supermanifold. If we then patch together different open sets we get globally a manifold structure, with a modified atlas made by a _graded_ ring of local functions $C^{\infty}(U\times V[1])$. More formally, we define: ###### Definition 4.1.4. A (smooth) _supermanifold_ $SM$ of dimension $(m|n)$ is a pair $(M,\mathcal{A})$, where $M$ is a $C^{\infty}$-manifold of dimension $m$, and $\mathcal{A}$ is a sheaf of $\mathbb{R}$-superalgebras that makes $SM$ locally isomorphic to $\left(U,C^{\infty}(U)\otimes\bigwedge(V^{*})\right)$ for some $U\subseteq\mathbb{R}^{m}$ open and some vector space $V$ of finite dimension $\dim{V}=n$. $M$ is called the _body_ of $SM$ and $\mathcal{A}$ is called the _structure sheaf_ (or, sometimes, “soul”) of $SM$.333Note that also a regular d-dimensional smooth manifold can be viewed as a pair $(M,\mathcal{O}_{M})$ composed by a topological space $M$ (Hausdorff and paracompact) with a structure sheaf of local functions $\mathcal{O}_{M}:\ \mathcal{O}_{M}(U)=C^{\infty}(U)$ for every $U\subseteq M$ open, such that locally every $U$ is isomorphic to a subset of $\mathbb{R}^{d}$. We just associated to a real manifold $M$ a graded-commutative algebra $C^{\infty}(SM)$ of functions over $SM$. Locally in a patch $U\subseteq M$, this matches the idea above of having coordinate systems as tuples $(x^{\mu},\theta^{i})_{\stackrel{{\scriptstyle\mu=1,\cdots,m}}{{i=1,\cdots,n}}}$ with the property (4.2). In particular, any local function $\Phi\in\mathcal{A}(U)$ can be trivialized with respect to the graded basis of $\bigwedge(V^{*})$: $\Phi(x,\theta)=\Phi^{(0)}(x)+\Phi_{i}^{(1)}(x)\theta^{i}+\Phi_{ij}^{(2)}\theta^{i}\wedge\theta^{j}+\cdots+\Phi^{(n)}_{i_{1},\cdots,i_{n}}\varepsilon^{i_{1}\cdots i_{n}}\theta^{1}\wedge\cdots\wedge\theta^{n}$ (4.3) where $\Phi^{(l)}_{i_{1}\cdots i_{l}}\in C^{\infty}(U)$ $\forall l\in\\{0,\cdots,n\\}$. The restriction to the zero-th degree $\epsilon:\mathcal{A}\to C^{\infty}_{M}$ such that $\epsilon(\Phi):=\Phi^{(0)}$ is usually called the _evaluation map_. ###### Example 4.1.1. To every super vector space $V=V_{0}\oplus V_{1}$ we can associate the supermanifold $\hat{V}=\left(V_{0},C^{\infty}(V_{0})\otimes\Lambda(V_{1}^{*})\right)\cong\mathbb{R}^{\dim(V_{0})|\dim(V_{1})}.$ More generally, to every vector bundle $E\to M$ with sections $\Gamma(E)$ we can associate the _odd vector bundle_ , denoted $\Pi E$ or $E[1]$, that is the supermanifold with body $M$ and structure sheaf $\mathcal{A}=\Gamma\left(\bigwedge E^{*}\right)$. The _odd tangent bundle_ $\Pi TM$ is the supermanifold with $\mathcal{A}=\Gamma\left(\bigwedge T^{*}M\right)$, i.e. globally the functions here are the differential forms on $M$, $C^{\infty}(\Pi TM)=\Omega(M)$. Coordinates on $\Pi TM$ are just $(x^{\mu},dx^{\mu})_{\mu=1\cdots,m}$, exactly as the coordinates on the tangent bundle $TM$, but now we consider them as generators of a graded algebra. Morphisms of supermanifolds can be given in terms of local morphisms of superalgebras, that respect compatibility between different patches. In particular, a morphism $(f,f^{\\#}):SM\to SN$ is a pair such that $f:M\to N$ is a diffeomorphism, and for every $U\subseteq M$ there is a morphism of superalgebras $f_{U}^{\\#}:\mathcal{A}_{M}(U)\to\mathcal{A}_{N}(f(U))$ that respects $f^{\\#}_{V}\circ\mathrm{res}_{U,V}=\mathrm{res}_{f(U),f(V)}\circ f^{\\#}_{U}$, where $\mathrm{res}_{U,V}$ is the restriction to a subset $V\subseteq U$. In less fancy words, if $\dim{SM}=(m|p)$ and $\dim{SN}=(n|q)$, a local coordinate system $(x,\theta)$ in $SM$ is mapped through $n$ functions $y^{\nu}=y^{\nu}(x,\theta)$ and $q$ functions $\varphi^{j}=\varphi^{j}(x,\theta)$ to a coordinate system $(y,\varphi)$ of $SN$. A vector field $X$ on a supermanifold $SM$, or _supervector field_ , is a derivation on $C^{\infty}(SM)$. Locally, considering $U\subseteq M$ open and $\mathcal{A}(U)=C^{\infty}(U)\otimes\bigwedge(V^{*})$, it can be expressed with respect to a coordinate system $(x,\theta)$ as $X=X^{\mu}(x,\theta)\frac{\partial}{\partial x^{\mu}}+X^{i}(x,\theta)\frac{\partial}{\partial\theta^{i}},$ (4.4) where $(\partial/\partial x^{\mu})$ acts as the corresponding vector field in $\Gamma(TM)$ on the $C^{\infty}(U)$ components and acts trivially on the odd coordinates $\theta^{i}$; $(\partial/\partial\theta^{i})$ acts trivially on $C^{\infty}(U)$, and as an _interior multiplication_ by the dual basis vector $u_{i}\in V$: $\frac{\partial}{\partial\theta^{i}}\theta^{j}:=\theta^{j}(u_{i})=\delta^{j}_{i}$. $X^{\mu},X^{i}$ are local sections in $\mathcal{A}(U)$. Supervectors on $SM$ form the tangent bundle $TSM$. Notice that, in particular $(\partial/\partial x^{\mu})$ preserves the grading of an homogeneous function, i.e. it is a derivation of degree 0, while $(\partial/\partial\theta^{i})$ shifts the grading by -1. ###### Definition 4.1.5. A _graded vector field_ of degree $k$ on $SM$ is a graded linear map $X:C^{\infty}(SM)\to C^{\infty}(SM)[k]$ that satisfies the graded Leibniz rule: $X(\phi\psi)=X(\phi)\psi+(-1)^{k|\phi|}\phi X(\psi)$ for any $\phi,\psi\in\mathcal{A}$ of pure degree. The _graded commutator_ between graded vector fields $X,Y$ is defined as $[X,Y]:=X\circ Y-(-1)^{|X||Y|}Y\circ X.$ From what we said above, partial derivatives $(\partial/\partial x^{\mu})$ with respect to even coordinates commute between each other, while $(\partial/\partial\theta^{i})$ anticommute, being respectively graded vector fields of degree 0 and -1. Then from (4.4) and (4.3) we see that any supervector field can be decomposed with respect to the $\mathbb{Z}_{2}$-grading given by the parity, as the sum $X=X_{(0)}+X_{(1)}$ of an even (_bosonic_) and an odd (_fermionic_) vector field. This makes $\left(\Gamma(TSM),[\cdot,\cdot]\right)$ into a Lie superalgebra. The _value_ at a point $p\in M$ of a supervector field $X\in\Gamma(TSM)$ is defined through the evaluation map: $X_{p}(\Phi):=\epsilon_{p}(X(\Phi))=\left[X^{\mu}(x,\theta)\frac{\partial\Phi}{\partial x^{\mu}}+X^{i}(x,\theta)\frac{\partial\Phi}{\partial\theta^{i}}\right]_{\stackrel{{\scriptstyle x=x(p)}}{{\theta=0}}}.$ (4.5) Clearly a super vector field $X$ is not determined by its values at points, since the evaluation map throws away all the dependence on the Grassmann-odd coordinates $\theta^{i}$ in the coefficient functions $X^{\mu},X^{i}$. This means that at every point $p\in M$, $T_{p}SM$ is a super vector space generated by the symbols $\partial/\partial x^{\mu},\partial/\partial\theta^{i}$ of opposite degrees, with _real_ coefficients. We collect this result in the following proposition. ###### Proposition 4.1.1. Let $SM=(M,\mathcal{A})$ be a supermanifold such that for every chart $U\subseteq M$ $\mathcal{A}=C^{\infty}(U)\otimes\Lambda(V^{*})$. Then at every point $p\in M$, $T_{p}SM\cong T_{p}M\oplus V[1]$ as real super vector spaces. For an odd vector bundle $\Pi E$ this specializes as $T_{p}\Pi E\cong T_{p}M\oplus E_{p}[1]$. Notice that one has always really both a $\mathbb{Z}$ and a $\mathbb{Z}_{2}$ grading of functions and supervector fields, analogously to the remark (4.1). As already mentioned, in field theory and in particular in the BRST formalism, the first one is called _ghost number_ , while the second one is the distinction between _bosonic_ and _fermionic_ degrees of freedom in the theory. The physical (i.e. gauge-invariant) combinations are those of ghost number zero. From the point of view of equivariant cohomology, it is very useful to relate the algebraic models we saw in Chapter 2 to these graded manifold structures. In particular, in field theory we interpret the graded complex of fields as a Cartan model, with a suitable graded equivariant differential given in terms of supersymmetry or BRST transformations. Generically, a _differential_ in supergeometry can be interpreted as a special supervector field on a supermanifold: ###### Definition 4.1.6. A _cohomological vector field_ $Q$ on a supermanifold $SM$ is a graded supervector field of degree +1 satisfying $[Q,Q]=0.$ It is immediate that any cohomological vector field corresponds to a _differential_ on the algebra of functions $C^{\infty}(SM)$, since being it of degree +1, $Q\circ Q=(1/2)[Q,Q]=0$. For example, consider the de Rham differential $d:\Omega(M)\to\Omega(M)$ on a regular smooth manifold $M$. It corresponds to the cohomological vector field on the odd tangent bundle $\Pi TM$ given in local coordinates by $d=\theta^{\mu}\frac{\partial}{\partial x^{\mu}},$ (4.6) where now $\theta^{\mu}\equiv dx^{\mu}$ are odd coordinate functions on $\Pi TM$. Similarly, the _interior multiplication_ $\iota_{X}:\Omega(M)\to\Omega(M)$ with respect to some vector field $X\in\Gamma(TM)$ is a nilpotent supervector field on $\Pi TM$ of degree -1, $\iota_{X}=X^{\mu}\frac{\partial}{\partial\theta^{\mu}}.$ (4.7) #### 4.1.2 Integration In the following we will use this graded machinery to translate the problem of integration of differential forms $\Omega(M)$ on a smooth manifold $M$, to an integration over the related odd tangent bundle $\Pi TM$. For this, we need to consider differential forms on a supermanifold $SM$. If $SM$ has local coordinates $(x^{\mu},\theta^{i})$, we can locally form an algebra generated by the 1-forms $dx^{\mu},d\theta^{i}$, where now $d$ is the de Rham differential on $SM$, acting as a cohomological vector field on $\Pi TSM$: $d=dx^{\mu}\partial_{x^{\mu}}+d\theta^{i}\partial_{\theta^{i}}.$ (4.8) The odd tangent bundle $\Pi TSM$ has thus coordinates $(x^{\mu},\theta^{i},dx^{\mu},d\theta^{i})$, where now $dx^{\mu}$ is odd whereas $d\theta^{i}$ is even.444To be more precise, the algebra of functions locally generated by $(x^{\mu},\theta^{i},dx^{\mu},d\theta^{i})$ on $\Pi TSM$ has _bi-grading_ , _i.e._ it inherits a $\mathbb{Z}_{2}$ grading from the original supermanifold $SM$ and a $\mathbb{Z}$ grading from the action of the de Rham differential (the form-degree). The coordinates have thus bi-degrees $x^{\mu}:(\text{even},0),\qquad\theta^{i}:(\text{odd},0),\qquad dx^{\mu}:(\text{even},1),\qquad d\theta^{i}:(\text{odd},1),$ which result in a total even degree for $d\theta^{i}$ and a total odd degree for $dx^{\mu}$. From this point of view, the de Rham differential acts as a “supersymmetry” transformation: $d:\left\\{\begin{aligned} x^{\mu}&\mapsto dx^{\mu}\\\ \theta^{i}&\mapsto d\theta^{i}\end{aligned}\right.$ (4.9) exchanging bosonic coordinates with fermionic coordinates. Since the odd 1-forms $d\theta^{i}$ are commuting elements, it is not possible to construct, at least in the usual sense, a form of “top degree”on $SM$. We will thus interpret integration over the odd coordinates by the purely algebraic rules of Berezin integration for Grassmann variables: $\int d\theta^{i}\theta^{i}=1,\qquad\int d\theta^{i}1=0,$ (4.10) and such that Fubini’s theorem holds for multidimensional integrals. We see that symbolically $\int d\theta^{i}\leftrightarrow\frac{\partial}{\partial\theta^{i}},$ (4.11) and in particular $\int d\theta^{i}\frac{\partial}{\partial\theta^{i}}\Phi(\theta^{i})=0$ always holds. We will use the important property of the Berezin integral: $\int d^{n}\theta\ e^{-\theta^{i}A_{ij}\theta^{j}}=\mathrm{Pf}(A)$ (4.12) where $d^{n}\theta\equiv d\theta^{1}d\theta^{2}\cdots d\theta^{n}$, to be compared with the usual Gaussian integral for real variables $\int d^{n}x\ e^{-x^{i}A_{ij}x^{j}}=\frac{\pi^{n/2}}{\sqrt{\det(A)}}.$ (4.13) Notice that under an homogeneous change of coordinates $\varphi^{i}=B^{i}_{j}\theta^{j}$, the “measure” shifts as $\int d^{n}\theta\to\det{B}\int d^{n}\varphi$, such that the Gaussian integral (4.12) is invariant under similarity transformations. For a mathematically refined theory of superintegration, we suggest looking at [40]. Concerning the integration of functions on the odd tangent bundle $\Pi TM$, we notice how, making use of the Berezin rules (4.10), this is nothing but a reinterpretation of the usual integrals of differential forms on $M$. If $\dim{M}=d$, the integral of the form $\omega=\sum_{i}\omega^{(i)}$, where $\omega^{(i)}\in\Omega^{i}(M)$ is $\int_{M}\omega=\int_{M}\omega^{(d)}=\int_{M}d^{d}x\ \omega^{(d)}(x),$ (4.14) selecting the top-form of degree $d$. If we consider the same form $\omega\in C^{\infty}(\Pi TM)$, its trivialization in coordinates $(x,\theta)$ is $\omega(x,\theta)=\sum_{i}\omega^{(i)}_{\mu_{1}\cdots\mu_{i}}(x)\theta^{\mu_{1}}\cdots\theta^{\mu_{i}}.$ (4.15) Now the Berezin integration over $d^{d}\theta$ selects just the term with the right number of $\theta$’s, giving $\int_{\Pi TM}d^{d}xd^{d}\theta\ \omega(x,\theta)=\int_{M}d^{d}x\ \omega^{(d)}(x)\int d^{d}\theta\ \theta^{d}\theta^{d-1}\cdots\theta^{1}=\int_{M}d^{d}x\ \omega^{(d)}(x).$ (4.16) ### 4.2 Supergeometric proof of ABBV formula for a circle action We give now a proof of the ABBV integration formula for a $U(1)$-action, starting from the expression (3.14). The “localization 1-form” is chosen as $\beta:=g(\underline{T},\cdot)$ (4.17) where $g$ is a $U(1)$-invariant metric and $\underline{T}$ is the fundamental vector field corresponding to the generator $T\in\mathfrak{u}(1)$. In local coordinates, the action of the Cartan differential $d_{C}=d+\iota_{T}$ on this 1-form can be written as $\begin{split}d_{C}\beta&=B_{\mu\nu}(x)dx^{\mu}dx^{\nu}+g_{\mu\nu}(x)T^{\mu}(x)T^{\nu}(x)\\\ B_{\mu\nu}&=(\nabla_{\mu}T)_{\nu}-(\nabla_{\nu}T)_{\mu}\end{split}$ (4.18) where, again, we suppressed the $S(\mathfrak{u}(1)^{*})$ generator setting $\phi=-1$. We use the result of the last section to rewrite (3.14) as an integral over the odd tangent bundle $\Pi TM$, identifying the odd coordinates $\theta^{\mu}\equiv dx^{\mu}$: $\begin{split}I[\alpha]&=\int_{M}\alpha=\lim_{t\to\infty}\int_{M}\alpha e^{-td_{C}\beta}\\\ &=\lim_{t\to\infty}\int_{\Pi TM}d^{d}xd^{d}\theta\ \alpha(x,\theta)\exp{\left\\{-tB_{\mu\nu}(x)\theta^{\mu}\theta^{\nu}-tg_{\mu\nu}(x)T^{\mu}(x)T^{\nu}(x)\right\\}},\end{split}$ (4.19) where the equivariantly closed form $\alpha$ is the sum of $U(1)$-invariant differential forms in $\Omega(M)^{U(1)}$ suppressing the $S(\mathfrak{u}(1)^{*})$ generator, $\alpha(x,\theta)=\sum_{i}\alpha^{(i)}_{\mu_{1}\cdots\mu_{i}}(x)\theta^{\mu_{1}}\cdots\theta^{\mu_{i}},$ (4.20) such that $d_{C}\alpha=\left(\theta^{\mu}\partial_{x^{\mu}}+T^{\mu}\partial_{\theta^{\mu}}\right)\alpha=0$. Using the Gaussian integrals (4.12) and (4.13), we have the following delta- function representations for Grassmann-even and Grassmann-odd variables $\displaystyle\delta^{(n)}(y)$ $\displaystyle=\lim_{t\to\infty}\left(\frac{t}{\pi}\right)^{n/2}\sqrt{\det{A}}\ e^{-ty^{\mu}A_{\mu\nu}y^{\nu}}$ (4.21) $\displaystyle\delta^{(n)}(\eta)$ $\displaystyle=\lim_{t\to\infty}t^{-n/2}\frac{1}{\mathrm{Pf}A}\ e^{-tA_{\mu\nu}\eta^{\mu}\eta^{\nu}}$ where the limits are understood in the weak sense. Multiplying and dividing by $(t^{n/2})$ in (4.19), and using the delta-representation we rewrite the integral as $I[\alpha]=\pi^{d/2}\int_{\Pi TM}d^{d}xd^{d}\theta\ \alpha(x,\theta)\frac{\mathrm{Pf}B(x)}{\sqrt{\det{g}(x)}}\delta^{(d)}(T(x))\delta^{(d)}(\theta).$ (4.22) The delta function on the odd coordinates simply puts $\theta^{\mu}=0$, that is analogous to selecting the top-degree form in (4.19), so that it remains $\alpha(x,0)\equiv\alpha^{(0)}(x)\in C^{\infty}(M)$. The delta function on the even coordinates instead selects the values at which $T^{\mu}(x)=0$, that corresponds to the fixed point set $F\hookrightarrow M$ of the $U(1)$-action. Suppose this fixed point set to be of dimension 0, i.e. composed by isolated points in $M$. If this is the case, we can simply separate the integral $\int d^{d}x$ in a sum of integrals, each of which domain $\mathcal{D}(p):p\in F$ contains one and only one of those fixed points, and in each of them apply the delta function $\begin{split}I[\alpha]&=\pi^{d/2}\sum_{p\in F}\int_{\mathcal{D}(p)}d^{d}x\ \alpha^{(0)}(x)\frac{\mathrm{Pf}B(x)}{\sqrt{\det{g}(x)}}\delta^{(d)}(T(x))\\\ &=\pi^{d/2}\sum_{p\in F}\frac{\alpha^{(0)}(p)}{|\det{dT}|(p)}\frac{\mathrm{Pf}B(p)}{\sqrt{\det{g}(p)}}.\end{split}$ (4.23) Here the factor $|\det{dT}|(p)$ is the Jacobian from the change of variable $y^{\mu}:=T^{\mu}(x)$. At any point $p\in F$ we have $(\nabla_{\mu}T)_{\nu}(p)=\partial_{\mu}T_{\nu}(p)=\partial_{\mu}T^{\rho}(p)g_{\rho\mu}(p)$ since $T^{\mu}(p)=0$, so the pfaffian in the numerator becomes $\mathrm{Pf}B(p)=\sqrt{2^{d}}\ \mathrm{Pf}{dT}(p)\sqrt{\det{g}(p)},$ (4.24) and we get the result $I[\alpha]=(2\pi)^{d/2}\sum_{p\in F}\frac{\alpha^{(0)}(p)}{\mathrm{Pf}dT(p)}.$ (4.25) Notice that, at $p\in F$, the operator $dT(p)=\partial_{\mu}T^{\nu}\theta^{\mu}\otimes\partial_{\nu}=-[\underline{T},\cdot]$ coincide up to a sign with the infinitesimal action $\mathcal{L}_{T}$ of $T\in\mathfrak{u}(1)$ on the tangent space $T_{p}M$, just because here $T^{\mu}(p)=0$. If we consider a continuous fixed point set, so that $F$ is a regular submanifold of $M$, it is not possible to use the delta functions like in (4.23), but we can consider the decomposition as a disjoint union $M=F\sqcup N$. Points far away from $F$ give a zero contribution to $I[\alpha]$ in the limit $t\to\infty$, so we can consider a neighborhood of $F$ and split here the tangent bundle as $TM\cong i_{*}TF\oplus TN$ where $TN$ is the _normal bundle_ to $F$ in $M$ ($i$ is the inclusion map). Consequently, in this neighborhood we can split the coordinates $(x^{\mu},\theta^{\mu})$ on the odd tangent bundle in tangent and normal to $F$, and rescale the normal components as $1/\sqrt{t}$, $x^{\mu}=x^{\mu}_{0}+\frac{x^{\mu}_{\perp}}{\sqrt{t}},\qquad\theta^{\mu}=\theta_{0}^{\mu}+\frac{\theta_{\perp}^{\mu}}{\sqrt{t}}.$ (4.26) The measure simply splits as $d^{d}xd^{d}\theta=d^{n}x_{0}d^{\hat{n}}x_{\perp}d^{n}\theta_{0}d^{\hat{n}}\theta_{\perp}$, where $n+\hat{n}=d$, thanks to the Berezin integration rules (4.10). Expanding $B_{\mu\nu}(x),g_{\mu\nu}(x),T^{\mu}(x)$ around the $x_{0}$ components, and taking the limit $t\to\infty$, the integral becomes [41] $\displaystyle I[\alpha]=\int d^{n}x_{0}d^{n}\theta_{0}\ \alpha(x_{0},\theta_{0})$ $\displaystyle\int d^{\hat{n}}x_{\perp}d^{\hat{n}}\theta_{\perp}\ $ (4.27) $\displaystyle\exp{\left\\{-B_{\mu\sigma}(x_{0})\left(B_{\nu}^{\sigma}(x_{0})+R^{\sigma}_{\nu\lambda\rho}(x_{0})\theta_{0}^{\lambda}\theta_{0}^{\rho}\right)x_{\perp}^{\mu}x_{\perp}^{\nu}-B_{\mu\nu}(x_{0})\theta_{\perp}^{\mu}\theta_{\perp}^{\nu}\right\\}}$ where $R^{\sigma}_{\nu\lambda\rho}(x_{0})$ is the curvature relative to the metric $g$. The integrals over the normal coordinates are Gaussian, giving the exact “saddle point” contribution $\frac{1}{\mathrm{Pf}_{N}\left(\frac{R+B}{2\pi}\right)(x_{0})}=\frac{1}{\left.e_{T}(R)\right|_{N}},$ (4.28) where $\left.e_{T}(R)\right|_{N}$ is the $U(1)$-_equivariant Euler class_ of the normal bundle to $F$. We notice that this matches the definition in Appendix B.1, since $B:=\nabla T$, seen as an element of the adjoint bundle $\Omega^{0}(M;\mathfrak{gl}(n))$, is a moment map for the Riemannian curvature $R$ satisfying (as can be checked by direct computation) $\nabla B^{\sigma}_{\nu}=-\iota_{T}R^{\sigma}_{\nu}=R^{\sigma}_{\nu}(\cdot,\underline{T}),$ (4.29) where the covariant derivative acts as in the adjoint bundle with respect to the Levi-Civita connection, $\nabla=d+[\Gamma,\cdot]$. As the final piece, the Berezin integration selects the component of $\alpha$ of degree $\dim{F}$ evaluated on $F$, that is just the pull-back $i^{*}\alpha$ along the inclusion map. Summarizing, we are left with $I[\alpha]=\int_{F}\frac{i^{*}\alpha}{\left.e_{T}(R)\right|_{N}}$ (4.30) as presented in Section 3.2, for the case of a circle action. Notice that in Section 3.2 we called the moment map $B^{\sigma}_{\nu}(p)=\mu_{T}(p)^{\sigma}_{\mu}\in\mathrm{End}(T_{p}M)$ at a fixed point $p$. If the fixed point is isolated, the tangent space $T_{p}M$ splits, as in (3.22), as the direct sum of the weight spaces of the $U(1)$-representation, and the vector field $\underline{T}$ acts as a rotation in any subspace. On a suitable coordinate basis $\underline{T}=\sum_{i}v_{i}\left(x^{i}\frac{\partial}{\partial y^{i}}-y^{i}\frac{\partial}{\partial x^{i}}\right),$ (4.31) so that the moment map block-diagonalizes as $B^{\sigma}_{\nu}=\partial_{\nu}T^{\sigma}=\left(\begin{array}[]{ccccc}0&-v_{1}&\cdots&&\\\ v_{1}&0&&&\\\ \cdots&&0&-v_{2}&\\\ &&v_{2}&0&\\\ &&&&\ddots\end{array}\right).$ (4.32) Taking the pfaffian then one gets exactly the product of the weights (or the “exponents”) $v_{i}$ of the circle action, so that the Euler class results $e_{T}(R)=(2\pi)^{-\dim(M)/2}\prod_{i}v_{i}.$ (4.33) This recovers (3.23) for the case of a circle action. See [6] for the result in presence of a torus action. ### 4.3 Introduction to Poincaré-supersymmetry We have seen in the last section how it is useful to translate the integration problem of a differential form on the manifold $M$, into an integration over the supermanifold $\Pi TM$. Here the differential forms $\Omega(M)$ are seen as the _graded_ ring of functions over $\Pi TM$, and the differential $d_{C}=d+\iota_{T}$ being the sum of two graded derivations555Both $d$ and $\iota_{T}$ are nilpotent supervector fields on $\Pi TM$. Without suppressing the degree-2 generator $\phi$ of $S(\mathfrak{u(1)}^{*})$, it is apparent that their sum $d_{C}$ is a cohomological vector field, i.e. a good differential of degree +1, on the subspace of $U(1)$-invariant forms. of degree $\pm 1$, can be viewed as an infinitesimal _supersymmetry_ transformation mapping odd-degree (_fermionic_) forms to even-degree (_bosonic_) forms. In field theory, we already mentioned that the presence of a supersymmetry on the relevant complex of fields often arises in two different ways: * • The differential $d_{C}$ is represented in the physical model by a BRST-like supercharge, introduced because of some gauge freedom. In this case, the complex of fields (analogously to the graded ring of functions $\Omega(M)$) is the BRST complex, and the grading is referred to as _ghost number_. Physical states of the quantum field theory are then created by fields of 0-degree, i.e. the functions on $M$. We could refer to this type of supersymmetry as a “hidden” one, coming from the original internal gauge symmetry of the model. In Hamiltonian systems, as we shall see in the next chapter, this gauge freedom can be simply associated to the Hamiltonian flow. * • The original theory could also be explicitly endowed with a supersymmetry. In this case the base space has a supermanifold structure, and the grading of the field complex follows. This is the case of QFT with Poincaré supersymmetry, where the action of a _supercharge_ as generator of the super-Poincaré algebra can be interpreted as an equivariant differential. In this and the next sections we will review the geometric setup of Poincaré supersymmetry and its generalization to curved spacetimes, interesting case for practical applications of the localization technique in QFT. As QFT (with global Poincaré symmetry) is formulated on Minkowski spacetime666We are really interested in both the Lorentzian and the Euclidean case, so we will express both the Minkowski and Euclidean spaces as $\mathbb{R}^{d}$ without stressing of the signature in the notation. The choice of metric will be clear from the context. $\mathbb{R}^{d}\cong ISO(\mathbb{R}^{d})/O(d),$ (4.34) we can formulate a Poincaré-supersymmetric theory on a super-extension of this space, coming from a given super-extension of the Poincaré group $ISO(\mathbb{R}^{d})$. We then first introduce the super-Poincaré groups starting from the super-extension of their algebras. #### 4.3.1 Super-Poincaré algebra and superspace ###### Definition 4.3.1. A super-Poincaré algebra is the extension of the Poincaré algebra $\mathfrak{iso}(d)\cong\mathbb{R}^{d}\oplus\mathfrak{so}(d)$ as a Lie superalgebra, via a given _real_ spin representation space $S$ of $Spin(d)$ taken in odd-degree: $\mathfrak{siso}_{S}(d)\cong\mathfrak{iso}(d)\oplus S[1].$ (4.35) The super Lie bracket are extended on $S[1]$ through the _symmetric_ and Spin- equivariant bilinear form $\Gamma:S\times S\to\mathbb{R}^{d}$: $[\Psi,\Phi]:=2\Gamma(\Psi,\Phi)=2(\overline{\Psi}\gamma^{\mu}\Phi)P_{\mu}=2(\Psi^{T}C\gamma^{\mu}\Phi)P_{\mu}$ (4.36) where $P_{\mu}$ are generators of the translation algebra $\mathbb{R}^{d}$, $C$ is the charge conjugation matrix, $\overline{\Psi}\in S^{*}$ is the Dirac adjoint of $\Psi$,777In Lorentzian signature $\overline{\Psi}=\Psi^{\dagger}\beta$ with $\beta_{ab}\equiv(\gamma^{0})^{a}_{\ b}$, in Euclidean signature $\overline{\Psi}=\Psi^{\dagger}$. This is due to hermitianity of the generators of the Euclidean algebra, unlike the Lorentzian case. and $\gamma^{\mu}$ are the generators of the Clifford algebra acting on $S$. In a real representation, the Majorana condition $\Psi^{T}C\stackrel{{\scriptstyle!}}{{=}}\overline{\Psi}$ is satisfied. The other brackets involving $S$ are defined by the natural action of $\mathfrak{so}(d)$ on it, and by the trivial action on $\mathbb{R}^{d}$: $\displaystyle\lambda\in\mathfrak{so}(d):\quad$ $\displaystyle[\lambda,\Psi]:=\frac{i}{2}\lambda_{\mu\nu}\Sigma^{\mu\nu}(\Psi),$ (4.37) $\displaystyle[\Psi,\lambda]=-[\lambda,\Psi],$ (4.38) $\displaystyle a\in\mathbb{R}^{d}:\quad$ $\displaystyle[a,\Psi]:=0,$ (4.39) where $\Sigma^{\mu\nu}=\frac{i}{2}\gamma^{[\mu}\gamma^{\nu]}$ are the generators of the rotation algebra in the Spin representation $S$. If the charge conjugation matrix is symmetric in the given representation, we can further enlarge this superalgebra via a “central extension”, considering $\mathfrak{siso}_{S}(d)\oplus\mathbb{R}$ with the extended brackets $\displaystyle\Psi,\Phi\in S[1]:\quad$ $\displaystyle[\Psi,\Phi]:=\Gamma(\Psi,\Phi)+(\Psi^{T}C\Phi),$ (4.40) $\displaystyle x\in\mathbb{R},A\in\mathfrak{siso}_{S}(d):\quad$ $\displaystyle[x,A]:=0.$ (4.41) The super Jacobi identity is satisfied thanks to the Spin-equivariance of the spinor bilinear form: $\Gamma\left(e^{R^{(s)}}\Psi,e^{R^{(s)}}\Phi\right)=e^{R^{(v)}}\Gamma(\Psi,\Phi)$ (4.42) where $R^{(s)},R^{(v)}$ are the same element $R\in\mathfrak{so}(d)$ in the spin and vector representations, respectively. Often the bracket structure of this superalgebra is given in terms of the generators. Regarding the odd part and picking a basis $\left\\{Q_{a}\right\\}$ of $S[1]$, their brackets are888We conventionally raise and lower spinor indices with the charge conjugation matrix: $(\gamma^{\mu})_{ab}:=C_{ac}(\gamma^{\mu})^{c}_{\ b},\qquad(\gamma^{\mu})_{ab}=(\gamma^{\mu})_{ba}.$ Notice that the matrix $C$ represents an inner product on $S$, while the Clifford algebra generators $\\{\gamma^{\mu}\\}$ act on $S$ as endomorphisms, so the index structure follows. A review of classification of Clifford algebras, Spin groups and Majorana spinors can be found in [42]. $[Q_{a},Q_{b}]=2(\gamma^{\mu})_{ab}P_{\mu}+C_{ab}.$ (4.43) The generators $\left\\{Q_{a}\right\\}$ are referred to as _supercharges_. If the real spin representation on $S$ is irreducible as a representation of the corresponding Clifford algebra, we have the minimal amount of supersymmetry and we refer to $\mathfrak{siso}_{S}(d)$ as to an $\mathcal{N}=1$ supersymmetry algebra. If instead the representation is reducible, then $S=\bigoplus_{I=1}^{\mathcal{N}}S^{(I)}$ and we can split the basis of supercharges as $\left\\{Q^{I}_{a}\right\\}$. This case is referred to as _extended_ supersymmetry. In this basis the gamma matrices block-diagonalize as $\gamma^{\mu}\otimes\mathbb{I}$, with $\gamma^{\mu}$ the (minimal) gamma matrices in every $S^{(I)}$, and the central extension part separates as $C\otimes Z$, with $C$ being the (minimal) charge conjugation matrix in every $S^{(I)}$ and $Z$ a matrix of so-called _central charges_. The odd part of the superalgebra then looks like $[Q_{a}^{I},Q_{b}^{J}]=2(\gamma^{\mu})_{ab}\delta^{IJ}P_{\mu}+C_{ab}Z^{IJ}.$ (4.44) The matrix $Z$ must be (anti)symmetric if $C$ is (anti)symmetric. ###### Definition 4.3.2. The subspace $\mathfrak{st}_{S}(d):=\mathbb{R}^{d}\oplus S[1]$ is a Lie superalgebra itself (if there is no central extension), and can be referred to as the _super-translation algebra_. Even if $\mathfrak{st}_{S}(d)$ it is not Abelian, it has the property $[a,[b,c]]=0\qquad\forall a,b,c\in\mathfrak{st}_{S}(d)$ (4.45) that can be easily checked by the definition. This means that the elements of the corresponding _super-translation group_ can be computed exactly using the exponential map and the BHC formula. This space can be identified with the _super-spacetime_. ###### Definition 4.3.3. We can define the full super-Poncaré group as $SISO_{S}(d)=\exp{(\mathfrak{st}_{S}(d))}\rtimes Spin(d),$ (4.46) so that we can identify the superspacetime with respect to the spin representation $S$, analogously to (4.34), as $S\mathbb{R}_{S}^{d}=SISO_{S}(d)/Spin(d)\cong\exp{(\mathfrak{st}_{S}(d))}.$ (4.47) As a supermanifold of dimension $(d|\dim(S))$, $S\mathbb{R}_{S}^{d}$ is characterized by its sheaf of functions, $\mathcal{A}=C^{\infty}(\mathbb{R}^{d})\otimes\bigwedge(S^{*}).$ (4.48) So, $S\mathbb{R}_{S}^{d}$ is the odd vector bundle associated to the spinor bundle over $\mathbb{R}^{d}$ of typical fiber $S$. In particular, on $S\mathbb{R}_{S}^{d}$ we have respectively even and odd coordinates $(x^{\mu},\theta^{a})$, with $\mu=1,\cdots,d$ and $a=1,\cdots,\dim(S)$. As a Lie group, we can get the group operation (the sum by supertranslation) from the exponentiation of its Lie superalgebra. Technically, to use the BHC formula $e^{A}e^{B}=e^{A+B+\frac{1}{2}[A,B]+\cdots}$ (4.49) we would like to deal with a _Lie algebra_ , so we consider the group operation on coordinates functions instead of points on $S\mathbb{R}_{S}^{d}$, taking the space $\left(\mathcal{A}\otimes\mathfrak{st}_{S}(d)\right)_{(0)}=\left(\mathcal{A}_{(0)}\otimes\mathbb{R}^{d}\right)\oplus\left(\mathcal{A}_{(1)}\otimes S[1]\right).$ (4.50) The Lie brackets on this space are inherited from those on $\mathfrak{st}_{S}(d)$ and the (graded) multiplication of functions in $\mathcal{A}$. The only non-zero ones come from couples of elements of $\mathcal{A}_{(1)}\otimes S[1]$: $[f_{1}\otimes\epsilon_{1},f_{2}\otimes\epsilon_{2}]=-2\Gamma(\epsilon_{1},\epsilon_{2})f_{1}f_{2}$ (4.51) where the sign rule has been used since both $f_{1},f_{2}$ and $\epsilon_{1},\epsilon_{2}$ are odd.999We are being a little informal here, but this can be made more rigorous with the help of a construction called _functor of points_. The important thing for us is that in this approach one can work with coordinate functions $x,\theta$ instead of some would-be “points” on the supermanifold (a misleading concept since we know from the last section that a supermanifold is not a set). See [43] for more details. We will use the combination (suppressing tensor products) $[\theta^{a}Q_{a},\varphi^{b}Q_{b}]=-2\Gamma(Q_{a},Q_{b})\theta^{a}\varphi^{b}=-2(\gamma^{\mu})_{ab}\theta^{a}\varphi^{b}P_{\mu}\equiv-2(\theta\gamma^{\mu}\varphi)P_{\mu}.$ This makes $\left(\mathcal{A}\otimes\mathfrak{st}_{S}(d)\right)_{(0)}$ into a Lie algebra, by the antisymmetry of the product between odd functions. Representing via the exponential map $(x,\theta)$ as $\exp\left\\{i(xP+\theta Q)\right\\}$, where we suppressed also index contractions, we can finally use the BHC formula on this algebra to get the group operation on coordinates: $\displaystyle(x,\theta)\,(y,\varphi)$ $\displaystyle=\exp{\left\\{i(xP+\theta Q)\right\\}}\exp{\left\\{i(yP+\varphi Q)\right\\}}$ (4.52) $\displaystyle=\exp{\left\\{i\left(xP+\theta Q+yP+\varphi Q+\frac{i}{2}[\theta Q,\varphi Q]\right)\right\\}}$ $\displaystyle=\exp{\left\\{i\left(x+y+i\theta\gamma\varphi\right)P+i\left(\theta+\varphi\right)Q\right\\}}$ $\displaystyle=\left(x+y-i\theta\gamma\varphi,\theta+\varphi\right).$ If the odd dimension $\dim{(S)}$ is zero, this reduces to a standard translation in $\mathbb{R}^{d}$. The infinitesimal action of the superalgebra $\mathfrak{st}_{S}(d)$ is defined as a Lie derivative with respect to the fundamental vector field representing a given element of the supertranslation algebra. For $\Phi\in\mathcal{A}$, $\epsilon=\epsilon^{a}Q_{a}\in(\mathcal{A}_{(1)}\otimes S[1])$ $\delta_{\epsilon}\Phi(x,\theta):=\mathcal{L}_{\underline{\epsilon}}(\Phi(x,\theta))=\underline{\epsilon}(\Phi)=\epsilon^{a}\underline{Q}_{a}(\Phi(x,\theta)),$ (4.53) where the odd vector field $\underline{Q}_{a}$ is associated to the supercharge $Q_{a}$ through the _left_ translation (4.52): $\displaystyle\underline{Q}_{a}\Phi(x,\theta)$ $\displaystyle=\left.\frac{\partial}{\partial\varphi^{a}}\left(e^{-i\varphi Q}\right)^{*}\Phi(x,\theta)\right|_{\varphi=0}=\left.\frac{\partial}{\partial\varphi^{a}}\Phi\big{(}(0,-\varphi)(x,\theta)\big{)}\right|_{\varphi=0}=$ (4.54) $\displaystyle=\partial_{\mu}\Phi(x,\theta)(i\gamma^{\mu}\theta)_{a}+\partial_{b}\Phi(x,\theta)\delta^{b}_{a}.$ We recognize then $\underline{Q}_{a}=-\frac{\partial}{\partial\theta^{a}}+i(\theta\gamma^{\mu})_{a}\frac{\partial}{\partial x^{\mu}},$ (4.55) and one can check that $[\underline{Q}_{a},\underline{Q}_{b}]=2(\gamma^{\mu})_{ab}\underline{P}_{\mu}$ with $\underline{P}_{\mu}=-i\partial/\partial x^{\mu}$ is associated to the momentum generator. The rest of the super-Poincaré group and its algebra acts naturally on superspace following the same type of arguments. In particular, the spin generators act in the vector and spin representations on the even and odd sectors, respectively. It is useful to introduce also the fundamental vector fields with respect to _right_ supertranslations on $S\mathbb{R}^{d}_{S}$. These are called _superderivatives_ and are easily obtained from the law (4.52) as101010Remember that left and right actions correspond to opposite signs at the exponent in the definition of the fundamental vector fields. $D_{a}:=\frac{\partial}{\partial\theta^{a}}+i(\theta\gamma^{\mu})_{a}\frac{\partial}{\partial x^{\mu}}.$ (4.56) One can check that indeed they satisfy $[\underline{Q}_{a},D_{b}]=0$ and $[D_{a},D_{b}]=-2(\gamma^{\mu})_{ab}P_{\mu}$, since right invariant and left invariant vector fields form anti-isomorphic algebras. #### 4.3.2 Chiral superspace and superfields It is customary to construct supersymmetric field theories starting not from a _real_ spin representation, but from a _complex_ Dirac representation $S$ endowed with a _real structure_ , i.e. an antilinear map $J:S\to S$ which is an involution ($J^{2}=id_{S}$).111111$J$ is the generalization of the “complex conjugation” operation on a $\mathbb{C}$-vector space. In this case, diagonalizing $J$ the representation splits as $S\cong S_{\mathbb{R}}\otimes\mathbb{C}\cong S^{(+)}\oplus S^{(-)}$ where $S_{\mathbb{R}}\cong S^{(+)}\cong iS^{(-)}$ are a real vector spaces. Choosing some basis, the matrix representing the real structure is $J=C(\gamma^{0})^{T}$ in Lorenzian signature, or $J=C$ in Euclidean signature. The _Majorana spinors_ are the elements of $S_{\mathbb{R}}$, that satisfy $J(\Psi)=\Psi$, or in matrix notation $\overline{\Psi}=\Psi^{T}C$. The subspace of Majorana spinors, taken as a real representation, would then give the super-extension of the last paragraph. If instead we are interested in working with the whole complex representation $S$, we are forced to introduce _complexified_ supersymmetry algebra and superspace, and then impose constraints on the resulting objects to properly reduce their degrees of freedom a posteriori. In particular, complex spinors from $S$ generating the supertranslations are taken satisfying the Majorana condition. This is always the case in QFT. As a paradigmatic example, we can take $\mathcal{N}=1$ supersymmetry in (3+1)-dimensions, where $\Psi$ is a Dirac spinor in $S=\mathbb{C}^{4}\cong\mathbb{R}^{4}\otimes\mathbb{C}$. On the complex representation, we can chose a to work in the _chiral basis_ $\\{Q_{a},\tilde{Q}_{\dot{a}}\\}$, splitted in left- and right-handed Weyl spinors. The dotted and undotted indices now run between $1,2$. Here the symmetric pairing $\Gamma:S\times S\to\mathbb{C}^{4}$ is non-zero only on $S^{(L/R)}\times S^{(R/L)}$, and and the restriction to the symmetrized subspace $\Gamma:S^{L}\odot S^{R}\to\mathbb{C}^{4}$ is actually an isomorphism, so the relevant non-zero brackets are $[Q_{a},\tilde{Q}_{\dot{b}}]=2(\gamma^{\mu})_{a\dot{b}}P_{\mu}.$ (4.57) On the chiral basis, $\gamma^{\mu}=\begin{pmatrix}0&\sigma^{\mu}\\\ \bar{\sigma}^{\mu}&0\end{pmatrix},\qquad C=\begin{pmatrix}\varepsilon&0\\\ 0&-\varepsilon\end{pmatrix},\qquad\sigma^{\mu}=(\mathbf{1},\sigma^{i}),\quad\overline{\sigma}^{\mu}=(\mathbf{1},-\sigma^{i}),$ where $(\sigma^{i})_{i=1,2,3}$ are the Pauli matrices and $\varepsilon_{ab}=\varepsilon_{\dot{a}\dot{b}}$ is the totally antisymmetric tensor. The supercharges are represented by the odd vector fields $\underline{Q}_{a}=-\frac{\partial}{\partial\theta^{a}}+i\tilde{\theta}^{\dot{b}}(\gamma^{\mu})_{\dot{b}a}\frac{\partial}{\partial x^{\mu}},\qquad\underline{\tilde{Q}}_{\dot{a}}=-\frac{\partial}{\partial\tilde{\theta}^{\dot{a}}}+i\theta^{b}(\gamma^{\mu})_{b\dot{a}}\frac{\partial}{\partial\overline{x^{\mu}}},$ (4.58) and the supersymmetry action on a superfield is $\delta_{\epsilon}\Phi=(\epsilon^{a}\underline{Q}_{a}+\tilde{\epsilon}^{\dot{a}}\underline{\tilde{Q}}_{\dot{a}})\Phi$ (4.59) where $\overline{\epsilon^{a}}=\tilde{\epsilon}^{\dot{a}}$, being a Majorana spinor. The superderivatives are $D_{a}=\frac{\partial}{\partial\theta^{a}}+i\tilde{\theta}^{\dot{b}}(\gamma^{\mu})_{\dot{b}a}\frac{\partial}{\partial x^{\mu}},\qquad\tilde{D}_{\dot{a}}=\frac{\partial}{\partial\tilde{\theta}^{\dot{a}}}+i\theta^{b}(\gamma^{\mu})_{b\dot{a}}\frac{\partial}{\partial\overline{x^{\mu}}}.$ (4.60) Here we split the odd coordinates $\theta^{a},\tilde{\theta}^{\dot{a}}$ according to the split of the supercharges. The reality constraint on the coordinates then reads $\overline{\theta^{a}}=\tilde{\theta}^{\dot{a}},\qquad\overline{x^{\mu}}=x^{\mu}.$ (4.61) Since we are operating over $\mathbb{C}$, the subspaces of the supertranslation algebra $\mathfrak{st}^{(L/R)}:=\mathbb{C}^{4}\oplus S^{(L/R)}[1]$ (4.62) are both Lie superalgebras over $\mathbb{C}$, and determines the corresponding complex Lie supergroups $S\mathbb{C}^{(L/R)}$. Moreover these subalgebras are _Abelian_ , since $\Gamma$ vanishes on $S^{(L/R)}\times S^{(L/R)}$. ###### Definition 4.3.4. $S\mathbb{C}^{(L)}$ is called _chiral_ superspace and $S\mathbb{C}^{(R)}$ _anti-chiral_ superspace. We can write the complexified superspace as $S\mathbb{C}^{4}_{S}\cong S\mathbb{C}^{(L)}\times_{\mathbb{C}^{4}}S\mathbb{C}^{(R)}$ (4.63) where the $\times_{\mathbb{C}^{4}}$ here denotes the fiber product with respect to the base $\mathbb{C}^{4}$.121212This is analogous to the pull-back bundle, not to be confused with an homotopy quotient. The chiral and anti- chiral superspaces are those identified by the flows of the corresponding superderivatives, since they generates the chiral and anti-chiral part of the supertranslation algebra on $\Gamma(TS\mathbb{C}_{S}^{4})$.131313Here $\Gamma$ denotes the space of sections on the tangent bundle, i.e. the vector fields on $S\mathbb{C}_{S}^{4}$, not the spinor pairing. We can easily find sets of “holomorphic-like” coordinates $y^{\mu}_{(\pm)}:=x^{\mu}\pm i\tilde{\theta}^{\dot{a}}(\gamma^{\mu})_{\dot{a}b}\theta^{b},\qquad\varphi^{a}:=\theta^{a},\qquad\tilde{\varphi}^{\dot{a}}:=\tilde{\theta}^{\dot{a}},$ (4.64) where the superderivatives simplify as $D_{a}=\left\\{\begin{aligned} &\frac{\partial}{\partial\varphi^{a}}\\\ &\frac{\partial}{\partial\varphi^{a}}+2i(\tilde{\varphi}\gamma^{\mu})_{a}\frac{\partial}{\partial y^{\mu}_{(-)}}\end{aligned}\right.,\qquad\tilde{D}_{\dot{a}}=\left\\{\begin{aligned} &\frac{\partial}{\partial\tilde{\varphi}^{\dot{a}}}+2i(\varphi\gamma^{\mu})_{\dot{a}}\frac{\partial}{\partial y^{\mu}_{(+)}}\\\ &\frac{\partial}{\partial\tilde{\varphi}^{\dot{a}}}\end{aligned}\right..$ (4.65) Complexifying the superspace we doubled its real-dimension, and that leads to a sort of “reducibility” of the relevant physical objects, _i.e._ the fields on superspace, or _superfields_. The simplest kind of superfields in the complexified setting are complex _even_ maps $\Phi:S\mathbb{C}^{4}\to\mathbb{C}$, that is sections of a trivial $\mathbb{C}$-line bundle over $S\mathbb{C}^{4}$. We can impose now some constraints on them in order to restore the correct number of degrees of freedom. This is usually done in two different ways: asking the superfields to depend only on the chiral (or anti-chiral) sector of the complexified superspace, or imposing a reality condition. ###### Definition 4.3.5. 1. (i) A _chiral (anti-chiral) superfield_ is a superfield $\Phi$ such that $\tilde{D}_{\dot{a}}\Phi=0\qquad\left(D_{a}\Phi=0\right).$ 2. (ii) A _vector superfield_ is a superfield $V$ such that $V=V^{\dagger}.$ Notice how the first one is a sort of (anti)holomorphicity condition with respect to the chiral/anti-chiral sectors of $S\mathbb{C}^{4}_{S}$, spanned by the coordinates $\varphi^{a},\tilde{\varphi}^{\dot{a}}$. Moreover, we can see that the complex conjugate $\Phi^{\dagger}$ of a chiral superfield $\Phi$ is antichiral. Next we will see how these conditions reflect on the various component fields of the coordinate expansions of $\Phi$ and $V$. It is important to stress that in this particular case of $\mathcal{N}=1$ in (3+1)-dimensions, the complex representation $S\cong\mathbb{C}^{4}$ allows for a real structure and the presence of Majorana spinors, and also for a chiral decomposition into left- and right-handed parts. It does not exist though a common basis for the two decompositions, i.e. $S^{(\pm)}\neq S^{(L/R)}$. In other words, it is not possible to require _both_ the chirality and the Majorana conditions on spinors in 4-dimension, since Majorana spinors contain both left- and right- handed components. This means that in a theory with some supersymmetry in 4-dimensions there will be the same number of left-handed and right-handed degrees of freedom. In $d=2\text{mod}8$ dimensions (in Lorentzian signature), instead, the minimal complexified spin representation $S$ can be decomposed into Majorana-Weyl subrepresentations $S=S_{\mathbb{R}}^{(+)}\oplus S_{\mathbb{R}}^{(-)}$, so that one can choose to work only with real left- handed spinors. In the case of extended supersymmetry, we can thus have in general a different number of left-handed and right-handed real supercharges, and $S=(S_{\mathbb{R}})^{\mathcal{N}_{+}}\oplus(S_{\mathbb{R}})^{\mathcal{N}_{-}}$. This is denoted with $\mathcal{N}=(\mathcal{N}_{+},\mathcal{N}_{-})$. When $\mathcal{N}_{+}$ or $\mathcal{N}_{-}$ is zero, the supersymmetry is called _chiral_. For a review on spinors in different dimensions, Majorana and chirality conditions we refer to [42]. #### 4.3.3 Supersymmetric actions and component field expansion Actions for supersymmetric field theories are constructed integrating over superspace combinations of superfields and their derivatives. For this purpose, it is useful to write the superfields in a so-called “component expansion”, with respect to the generators of $\bigwedge(S^{*})$. In this paragraph we continue with the example of $\mathcal{N}=1$ in (3+1)-dimensions, but the construction is immediatly generalizable to other cases. Consider a complex superfield $\Phi:S\mathbb{C}^{4}\to\mathbb{C}$, its trivialization on the set of coordinates $(x,\theta,\tilde{\theta})$ being $\displaystyle\Phi(x,\theta,\tilde{\theta})=\phi(x)+v_{\mu}(x)(\tilde{\theta}\gamma^{\mu}\theta)+\psi_{a}(x)\theta^{a}+\tilde{\psi}_{\dot{a}}\tilde{\theta}^{\dot{a}}+F(x)\theta^{(2)}+\tilde{F}(x)\tilde{\theta}^{(2)}+$ (4.66) $\displaystyle+\xi_{a}(x)\theta^{a}\tilde{\theta}^{(2)}+\tilde{\xi}_{\dot{a}}(x)\tilde{\theta}^{\dot{a}}\theta^{(2)}+D(x)\theta^{(2)}\tilde{\theta}^{(2)}$ where the the wedge product between the $\theta$’s has been suppressed, and we agree they are anticommuting, $\theta^{(2)}:=\theta^{2}\theta^{1}$ and $\tilde{\theta}^{(2)}:=\tilde{\theta}^{\dot{2}}\tilde{\theta}^{\dot{1}}$. We used the isomorphism $\Gamma:S^{L}\odot S^{R}\to\mathbb{C}^{4}$ to represent the component of degree (1,1) as $v_{a\dot{a}}\mapsto v_{\mu}(\gamma^{\mu})_{a\dot{a}}$, the reason for this will become clear shortly. The expansion stops at top-degree (here 4) for the anticommuting property of the exterior product. In order for $\Phi$ to be an even (scalar) field, we must take the functions $\phi,v_{\mu},F,\tilde{F},D$ to be even (commuting), and the functions $\psi_{a},\tilde{\psi}_{\dot{a}},\xi_{a},\tilde{\xi}_{\dot{a}}$ carrying a spinor index to be odd (anticommuting). These are called _component fields_ of the superfield $\Phi$.141414The possibility of writing down an expansion similar to (4.66) with these properties could again be justified more rigorously thanks to the concept of _functor of point_. If we impose the chirality condition $\tilde{D}_{\dot{a}}\Phi=0$, when expressed in the coordinates $(y_{(-)},\varphi,\tilde{\varphi})$ this simply requires the independence on $\tilde{\varphi}$, so on these coordinates a chiral superfield can be written as $\Phi(y_{(-)},\varphi,\tilde{\varphi})=\phi(y_{(-)})+\psi_{a}(y_{(-)})\varphi^{a}+F(y_{(-)})\varphi^{(2)}.$ (4.67) Taylor-expanding in the old coordinates, this is equivalent to $\Phi(x,\theta,\tilde{\theta})=\phi(x)+\psi_{a}(x)\theta^{a}+F(x)\theta^{(2)}-i\partial_{\mu}\phi(x)(\tilde{\theta}\gamma^{\mu}\theta)+i(\tilde{\theta}\gamma^{\mu}\partial_{\mu}\psi(x))\theta^{(2)}-\partial^{2}\phi(x)\theta^{(2)}\tilde{\theta}^{(2)}$ (4.68) where higher order terms are again automatically zero for degree reasons.151515Here we used $\theta^{a}\theta^{b}=\varepsilon^{ab}\theta^{(2)}$, and $(\tilde{\theta}\gamma^{\mu}\theta)(\tilde{\theta}\gamma^{\mu}\theta)=2\eta^{\mu\nu}\theta^{(2)}\tilde{\theta}^{(2)}$ with a “mostly plus” signature. The “irreducible” chiral superfield has three non-zero field components: a complex scalar field $\phi$, a left-handed Weyl spinor field $\psi$ and another complex scalar field $F$. One can work out the supersymmetry transformations of these component fields from the general rule (4.59), and the result is $\displaystyle\delta_{\epsilon}\phi=\epsilon\psi$ (4.69) $\displaystyle\delta_{\epsilon}\psi=2i(\tilde{\epsilon}\gamma^{\mu})\partial_{\mu}\phi+\epsilon F$ $\displaystyle\delta_{\epsilon}F=-2i\tilde{\epsilon}\gamma^{\mu}\partial_{\mu}\psi$ where spinor contractions are implied, and spinor indices are lowered/raised via the charge conjugation matrix as usual. To construct a minimal Lagrangian density from the superfield $\Phi$, we can look at its mass dimension: this is equal to its lowest component $\phi$, that being a scalar field is $(d-2)/2=1$ in $d=4$ dimensions. Since $\psi$ is a spinor, it has dimension $(d-1)/2=3/2$, so the odd coordinates have always dimension -1/2. The highest component of any superfield has thus dimension two more than the superfield. This means that to construct a Lagrangian we have to take a quadratic expression in $\Phi$. Since the action should be real, the simplest choice is $\Phi^{\dagger}\Phi$. Its top-degree component is $\int d^{2}\theta d^{2}\tilde{\theta}\ \Phi^{\dagger}\Phi=4\left|\partial\phi\right|^{2}+i\overline{\Psi}\not{\partial}\Psi+|F|^{2}+\partial_{\mu}(\cdots)^{\mu}$ (4.70) where $\Psi$ is a Majorana 4-spinor, whose left-handed component is $\psi$. Up to a total derivative, this is the Lagrangian of the free, massless _Wess- Zumino model_ : $S_{WZ,free}[\Phi]=\int_{S\mathbb{R}^{4}_{\mathbb{C}^{4}}}d^{4}xd^{2}\theta d^{2}\tilde{\theta}\ \Phi^{\dagger}\Phi=\int_{\mathbb{R}^{4}}d^{4}x\left(4\left|\partial\phi\right|^{2}+i\overline{\Psi}\not{\partial}\Psi+|F|^{2}\right).$ (4.71) Since the field $F$ appears without derivatives, its equation of motion is an algebraic equation. For this reason, it is called an _auxiliary field_ , and it is customary to substitute its on-shell value in the action. For this simple model, this means putting $F=0$. This procedure makes in general the action to be supersymmetric only if the equations of motion (EoM) are imposed, and is often called _on-shell_ supersymmetry. The physical field content of an $\mathcal{N}=1$ chiral superfield is thus the supersymmetry _doublet_ $(\phi,\psi)$. By CPT invariance, the theory must contain both the chiral field $\Phi$ and its antichiral conjugate $\Phi^{\dagger}$, so that the physical field content of a meaningful theory constructed from it is made by two real scalar fields $\mathrm{Re}(\phi),\mathrm{Im}(\phi)$ and the Majorana spinor $\Psi$. If we now start from a vector superfield $V$, satisfying the reality condition $V=V^{\dagger}$, we reach a different physical field content and supersymmetric Lagrangian. The component expansion in a chart $(x,\theta,\tilde{\theta})$ is the following: $\displaystyle V(x,\theta,\tilde{\theta})=C(x)+\xi_{a}(x)\theta^{a}+\xi^{\dagger}_{\dot{a}}\tilde{\theta}^{\dot{a}}+v_{\mu}(x)(\tilde{\theta}\gamma^{\mu}\theta)+G(x)\theta^{(2)}+G^{\dagger}(x)\tilde{\theta}^{(2)}+$ (4.72) $\displaystyle+\eta_{a}(x)\theta^{a}\tilde{\theta}^{(2)}+\eta^{\dagger}_{\dot{a}}(x)\tilde{\theta}^{\dot{a}}\theta^{(2)}+E(x)\theta^{(2)}\tilde{\theta}^{(2)}$ where $C,v_{\mu},E$ are real fields. It is clear that this is the right type of superfield needed to describe (Abelian) gauge boson fields, represented here by $v_{\mu}$. This is why $V$ is called _vector_ superfield. Notice that the real part of a chiral superfield is a special kind of vector superfield. In particular, its vector component is a derivative: if $\Lambda$ is chiral, from (4.68) $\Lambda+\Lambda^{\dagger}\supset i\partial_{\mu}(\phi-\phi^{\dagger})(\tilde{\theta}\gamma^{\mu}\theta).$ (4.73) This suggest to interpret the transformation $V\mapsto V+(\Lambda+\Lambda^{\dagger})$ (4.74) as the action of a $U(1)$ internal gauge symmetry on superfields. In terms of component fields this gauge transformation reads $\begin{array}[]{lll}C\mapsto C+(\phi+\phi^{\dagger})&\xi_{a}\mapsto\xi_{a}+\psi_{a}&G\mapsto G+F\\\ v_{\mu}\mapsto v_{\mu}-i\partial_{\mu}(\phi-\phi^{\dagger})&\eta_{a}\mapsto\eta_{a}-i(\partial_{\mu}\psi^{\dagger}\gamma^{\mu})_{a}&E\mapsto E-\partial^{2}(\phi+\phi^{\dagger}).\end{array}$ (4.75) We can notice two main things. The first is that the combinations $\displaystyle\lambda_{a}:=\eta_{a}+i(\gamma^{\mu}\partial_{\mu}\xi^{\dagger})_{a}$ (4.76) $\displaystyle D:=E+\partial^{2}C$ are gauge invariants. The second is that, since $C,G,\xi_{a}$ transform as shifts, we can chose a special gauge in which they vanish. This is called the _Wess-Zumino (WZ) gauge_. Chosing a gauge of course breaks explicitly supersymmetry, but it is convenient for most of the calculations. In the WZ gauge, the vector superfield looks like $V=v_{\mu}(\tilde{\theta}\gamma^{\mu}\theta)+\lambda_{a}\theta^{a}\tilde{\theta}^{(2)}+\lambda^{\dagger}_{\dot{a}}\tilde{\theta}^{\dot{a}}\theta^{(2)}+D\theta^{(2)}\tilde{\theta}^{(2)}.$ (4.77) Gauge transformations with immaginary scalar component ($\phi+\phi^{\dagger}=0$) preserve the Wess-Zumino gauge and moreover induce on $v_{\mu}$ the usual $U(1)$ transformation of Abelian vector bosons. Indeed, if $\alpha:=2\mathrm{Im}(\phi)$, $v_{\mu}\mapsto v_{\mu}+\partial_{\mu}\alpha.$ (4.78) As for $F$ in the case of chiral superfields, $D$ is the top component field of the vector superfield $V$. It will have a purely algebraic equation of motion, so it can be considered as an auxiliary field. The physical field content of an $\mathcal{N}=1$ vector superfield is thus the supersymmetry _doublet_ $(v_{\mu},\lambda)$ composed by an Abelian gauge boson and a Majorana spinor, called the _gaugino_. A gauge-invariant supersymmetric action for the Abelian vector superfield can be given in terms of the spinorial superfields defined as $W_{a}:=\frac{1}{2}\tilde{D}^{2}D_{a}V,\qquad\tilde{W}_{\dot{a}}:=\frac{1}{2}D^{2}\tilde{D}_{\dot{a}}V.$ (4.79) $W_{a}$ ($\tilde{W}_{\dot{a}}$) is both chiral (antichiral) and gauge invariant, and moreover it satisfies the “reality” condition $D^{a}W_{a}=\tilde{D}^{\dot{a}}\tilde{W}_{\dot{a}}$. Expanding the component fields in coordinates $(y_{(\pm)},\theta,\tilde{\theta})$, we have $\displaystyle W_{a}=\lambda_{a}+if_{\mu\nu}(\gamma^{\mu}\gamma^{\nu})_{ab}\theta^{b}+D\varepsilon_{ab}\theta^{b}+2i(\gamma^{\mu}\partial_{\mu}\lambda^{\dagger})_{a}\theta^{(2)}$ (4.80) $\displaystyle\tilde{W}_{\dot{a}}=\lambda^{\dagger}_{\dot{a}}-if_{\mu\nu}(\gamma^{\mu}\gamma^{\nu})_{\dot{a}\dot{b}}\tilde{\theta}^{\dot{b}}+D\varepsilon_{\dot{a}\dot{b}}\tilde{\theta}^{\dot{b}}-2i(\gamma^{\mu}\partial_{\mu}\lambda)_{\dot{a}}\tilde{\theta}^{(2)}$ where $f_{\mu\nu}:=(\partial_{\mu}v_{\nu}-\partial_{\nu}v_{\nu})$ is the gauge invariant Abelian field-strength of $v_{\mu}$. A gauge invariant action is then $\displaystyle S[V]$ $\displaystyle=\int d^{4}x\frac{1}{8}\left(\int d^{2}\theta\ W^{a}W_{a}+\int d^{2}\tilde{\theta}\ \tilde{W}^{\dot{a}}\tilde{W}_{\dot{a}}\right)$ (4.81) $\displaystyle=\int d^{4}x\left\\{f_{\mu\nu}f^{\mu\nu}+i\lambda\gamma^{\mu}\partial_{\mu}\lambda^{\dagger}+2D^{2}\right\\}$ that is an $\mathcal{N}=1$ supersymmetric extension of the Abelian Yang-Mills theory in 4-dimensions. The supersymmetry transformations of the component field, under which (4.81) is invariant can be obtained applying the supertranslation on $V$ in WZ gauge. The result will not be in this gauge anymore, but can be translated back in WZ gauge applying an appropriate gauge transformation as (4.75). The result is $\displaystyle\delta_{\epsilon}v_{\mu}=\frac{1}{2}\left(\tilde{\epsilon}\gamma_{\mu}\lambda-\lambda^{\dagger}\gamma_{\mu}\epsilon\right)$ (4.82) $\displaystyle\delta_{\epsilon}\lambda=\frac{i}{2}f_{\mu\nu}\epsilon\gamma^{\mu}\gamma^{\nu}+D\epsilon$ $\displaystyle\delta_{\epsilon}D=-i\left(\tilde{\epsilon}\gamma^{\mu}\partial_{\mu}\lambda+\epsilon\gamma^{\mu}\partial_{\mu}\lambda^{\dagger}\right).$ Notice that in this case the Super Yang-Mills (SYM) action remains supersymmetric even if we impose the EoM on the auxiliary field $D$, setting $D=0$ in both (4.81) and (4.82). This is a special result, that holds in 4, 6 and 10 dimensions [44]. In a generic gauge theory with gauge group $G$, we consider a $G-$valued chiral multiplet $\Phi$ which transforms under a gauge transformations as $\Phi\mapsto e^{\Lambda}\Phi$ (4.83) where $\Lambda$ is a $\mathfrak{g}$-valued chiral superfield. Now the combination $\Phi^{\dagger}\Phi$ is not gauge invariant, so we introduce a $\mathfrak{g}-$valued vector superfield $V$, transforming as $e^{V}\mapsto e^{\Lambda^{\dagger}}e^{V}e^{\Lambda}$ (4.84) that reduces to the previous case (4.74) for Abelian $G=U(1)$. The exponential of a superfield can be defined through its component field expansion, that stops at finite order for degree reasons: $e^{V}=1+v_{\mu}(\tilde{\theta}\gamma^{\mu}\theta)+\lambda_{a}\theta^{a}\tilde{\theta}^{(2)}+\lambda^{\dagger}_{\dot{a}}\tilde{\theta}^{\dot{a}}\theta^{(2)}+(D+2v_{\mu}v^{\mu})\theta^{(2)}\tilde{\theta}^{(2)}.$ (4.85) The kinetic term for the chiral superfield can be rewritten as a gauge invariant combination: $\int d^{4}xd^{2}\theta d^{2}\tilde{\theta}\ \Phi^{\dagger}e^{V}\Phi.$ (4.86) Generalizing the supersymmetric field-strength $W_{a}$ to the non-Abelian case as $W_{a}=\frac{1}{2}\tilde{D}^{2}e^{-V}D_{a}e^{V}$ (4.87) we can write the full matter-coupled gauge theory action: $S[V,\Phi]=\int d^{4}x\left\\{\int d^{2}\theta d^{2}\tilde{\theta}\ \Phi^{\dagger}e^{V}\Phi+\left[\int d^{2}\theta\ \left(\frac{1}{4}\mathrm{Tr}W^{a}W_{a}+W(\Phi)\right)+c.c.\right]\right\\}$ (4.88) where $W(\Phi)$ is a holomorphic function of $\Phi$ called _superpotential_. The expansion in terms of component fields and the supersymmetry variations can be calculated with the same procedure we did in the other cases.161616A more detailed treatment can be found in [45], or [46]. Notice that whenever the center of the Lie algebra $\mathfrak{g}$ is non- trivial, i.e. when there is a $U(1)$ factor in $G$, we could add another supersymmetric and gauge-invariant term to the action (4.88). This is the so- called _Fayet-Iliopoulos term_ : $\int d^{4}xd^{2}\theta d^{2}\tilde{\theta}\ \xi(V)=\int d^{4}x\ \xi_{A}D^{A}$ (4.89) where $\xi=\xi_{A}\tilde{T}^{A}$ is a constant element in the dual of the center of $\mathfrak{g}$. #### 4.3.4 R-symmetry The subgroup of (outer) automorphisms of the supersymmetry group which fixes the underlying Poincaré (Euclidean) group is called _R-symmetry group_. At the level of the algebra, these are linear transformations that act only on the spin representation $S$, leaving the brackets of two spinors unchanged. In the complexified case, when different chiral sectors are present, the R-symmetry acts differently on any sector. For example, in the case of $\mathcal{N}=1$ in (3+1)-dimensions, there is a $U(1)_{R}$ R-symmetry group acting as $Q_{a}\mapsto e^{-i\alpha}Q_{a},\qquad\tilde{Q}_{\dot{a}}\mapsto e^{i\alpha}\tilde{Q}_{\dot{a}},$ (4.90) with $\alpha\in\mathbb{R}$. This clearly leaves the brackets $[Q_{a},\tilde{Q}_{\dot{a}}]$ invariant. The odd coordinates $\theta^{a}$ on superspacetime, being elements of $S^{*}$ transform as $\begin{array}[]{lcl}\theta^{a}\mapsto e^{i\alpha}\theta^{a}&&\tilde{\theta}^{\dot{a}}\mapsto e^{-i\alpha}\tilde{\theta}^{\dot{a}}\\\ d^{2}\theta\mapsto e^{-2i\alpha}d^{2}\theta&&d^{2}\tilde{\theta}\mapsto e^{2i\alpha}d^{2}\tilde{\theta},\end{array}$ (4.91) so that the volume element $d^{4}\theta=d^{2}\theta d^{2}\tilde{\theta}$ is invariant under R-symmetry. This fixes the _R-charge_ of the superpotential $W(\Phi)$ to be 2, if we want the action to be invariant under R-symmetry: $W(\Phi)\mapsto e^{2i\alpha}W(\Phi).$ (4.92) In principle we can chose the chiral superfield $\Phi$ to have any R-charge $r$, since the combination $\Phi^{\dagger}\Phi$ is R-invariant. This, combined with (4.91) means that the different field components in the chiral multiplet transform differently with respect to R-symmetry: $\phi\mapsto e^{ir\alpha}\phi,\quad\psi\mapsto e^{i(r-1)\alpha}\psi,\quad F\mapsto e^{i(r-2)\alpha}F.$ (4.93) The vector superfield, being real is acted upon trivially by $U(1)_{R}$. Its component fields are then forced to transform as $v_{\mu}\mapsto v_{\mu},\quad\lambda\mapsto e^{i\alpha}\lambda,\quad D\mapsto D,$ (4.94) thus the gauge-invariant supersymmetric field-strength $W_{a}$ has R-charge $1$. In general, if the spin representation is reducible and we have extended supersymmetry, the R-group is always compact. For $S=(S_{0})^{\mathcal{N}}$, where $S_{0}$ is a real representation, it is of the type $U(\mathcal{N})$, while for $S=(S^{(+)})^{\mathcal{N}_{+}}\oplus(S^{(-)})^{\mathcal{N}_{-}}$, where $S^{(\pm)}$ are the two real representations of different chirality, it is of the type $U(\mathcal{N}_{+})\times U(\mathcal{N}_{-})$ [43]. Notice the isomorphism $U(n)\cong(SU(n)\times U(1))/\mathbb{Z}_{n}$ (4.95) i.e. $U(n)$ is an n-fold cover of $SU(n)\times U(1)$. In particular, their Lie algebras are isomorphic. In terms of infinitesimal transformations then, the R-symmetry generators can be decomposed in one R-charge plus $\mathcal{N}^{2}-1$ rotation generators. The supercharges are rotated into one another by $Q_{a}^{I}\mapsto e^{-i\alpha}\mathcal{U}^{I}_{J}Q_{a}^{J},\quad\tilde{Q}_{\dot{a}}^{I}\mapsto e^{i\alpha}\mathcal{(U^{\dagger})}^{I}_{J}\tilde{Q}_{\dot{a}}^{J}.$ (4.96) In QFT, R-symmetry may or may not be present as a symmetry of the theory, and in many cases part of this symmetry may be broken by anomaly at quantum level. #### 4.3.5 Supersymmetry multiplets A geometric analysis as the one carried out in the last subsections allows one to find the physical field content of a supersymmetric theory in every dimensions and for any degree of reducibility of the spin representation $S$ that is used to extend the Poincaré algebra. Another systematic way to obtain the same result, from a more algebraic point of view, is to study the representation of the supersymmetry algebra $\mathfrak{siso}_{S}(d)$, in analogy with the Wigner analysis of massive and massless representations of the Poincaré algebra. As the cases encountered above, this study leads to the presence of different _supersymmetry multiplets_ for different choices of spin and $\mathcal{N}$. We will not present this here but refer for example to [47] for a comprehensive review, and list here some results for the multiplets at various $\mathcal{N}$. For $1\leq\mathcal{N}\leq 4$ with spin less or equal to 1, the supersymmetry particle representations simply consists of spin 1 vector particles, spin 1/2 fermions and spin 0 scalars. In the supergeometric approach, these fields are interpreted as components of the same superfield, and thus transform one into another under the supersymmetry algebra. Let $G$ be the gauge group, and $\mathfrak{g}$ its Lie algebra. We are interested mainly in two types of multiplets. The first is the (massless) _vector_ or _gauge multiplet_ , which transforms under the adjoint representation of $\mathfrak{g}$. For $\mathcal{N}=3,4$, this is the only possible multiplet. It turns out that quantum field theories with $\mathcal{N}=3$ supersymmetries coincide with those with $\mathcal{N}=4$ in view of CPT invariance, thus we shall limit our discussion to the $\mathcal{N}=4$ theories.171717To be more precise, it is possible to construct theories with genuine $\mathcal{N}=3$ supersymmetry, but they lack of a Lagrangian description in terms of component fields. For $\mathcal{N}=1,2$, we also have (possibly massive) matter multiplets: for $\mathcal{N}=1$, this is the _chiral multiplet_ , and for $\mathcal{N}=2$ this is the _hypermultiplet_ , both of which may transform under an arbitrary (unitary, and possibly reducible) representation of $G$. In (3+1)-dimensions, the on-shell field content of these multiplets is: * • $\mathcal{N}=1$ _gauge multiplet_ $(A_{\mu},\lambda)$: a gauge boson and a Majorana fermion, the gaugino. * • $\mathcal{N}=1$ _chiral multiplet_ $(\phi,\psi)$: a complex scalar and a left- handed Weyl fermion. * • $\mathcal{N}=2$ _gauge multiplet_ $(A_{\mu},\lambda_{\pm},\phi)$: $\lambda_{\pm}$ form a Dirac spinor, and $\phi$ is a complex _gauge scalar_. Under the $SU(2)_{R}$ symmetry, $A_{\mu}$ and $\phi$ are singlets, while $\lambda_{+},\lambda_{-}$ transform as a doublet. * • $\mathcal{N}=2$ _hypermultiplet_ $(\psi_{+},H,\psi_{-})$: $\psi_{\pm}$ form a Dirac spinor and $H_{\pm}$ are complex scalars. Under the $SU(2)_{R}$ symmetry, $\psi_{+}$ and $\psi_{-}$ transform as singlets, while $H_{+},H_{-}$ transform as a doublet. * • $\mathcal{N}=4$ _gauge multiplet_ $(A_{\mu},\lambda^{i},\Phi_{A})$: $\lambda^{i}$ , $i=1,2,3,4$ are Weyl fermions (equivalents to two Dirac fermions), and $\Phi_{A}$, $A=1,\cdots,6$ are real scalars (equivalents to three complex scalars). Under the $SU(4)_{R}$ symmetry181818The R-symmetry group is actually $SU(2)\times SU(2)\times U(1)$, as we will see in a practical application in the following. the gauge field $A_{\mu}$ is a singlet, the fermions $\lambda^{i}$ transform in the fundamental representation $\mathbf{4}$, the scalars $\Phi_{A}$ transform in the rank-two antisymmetric representation $\mathbf{6}$. Even though in this thesis we do not work explicitly with gravity theories, we will see in the next section that the introduction of _off-shell_ supergravity is necessary in a possible approach to construct globally supersymmetric theories on curved base-spaces. For this purpose, it is useful to remind also the content of massless supersymmetry particle representations with helicity between 1 and 2. These are the _gravitino multiplet_ and the _graviton multiplet_ (or _supergravity multiplet_ , or _metric multiplet_). In general the gravitino multiplet contains degrees of freedom with helicity less or equal than 3/2. Since in a theory without gravity one cannot accept particles with helicity greater than one,191919This comes from the so-called _Weinberg- Witten theorem_ [48]. that multiplet cannot appear in a supersymmetric theory if also a graviton, with helicity 2, does not appear. In (3+1)-dimensions, the field content of the relevant multiplets are: * • $\mathcal{N}=1$ _gravitino multiplet_ $(\Phi_{\mu},B_{\mu})$: a helicity 3/2 fermionic particle and a vector boson. * • $\mathcal{N}=1$ _graviton multiplet_ $(h_{\mu\nu},\Psi_{\mu})$: the _graviton_ , with helicity 2, and its supersymmetric partner the _gravitino_ , of helicity 3/2. * • $\mathcal{N}=2$ _gravitino multiplet_ : a spin 3/2 particle, two vectors and one Weyl fermion. * • $\mathcal{N}=2$ _graviton multiplet_ : graviton, two gravitinos and a vector boson. For $\mathcal{N}>4$ it is not possible to avoid gravity since there do not exist representations with helicity smaller than 3/2. Hence, theories with $\mathcal{N}>4$ are all supergravity theories. #### 4.3.6 Euclidean 3d $\mathcal{N}=2$ supersymmetric gauge theories As an example, which will be used in some applications of the localization principle in the next chapter, we can look at $\mathcal{N}=2$ Euclidean supersymmetry in 3-dimensions. First, notice that the rotation algebra for 3d Euclidean space is $\mathfrak{so}(3)$. The corresponding spin group is thus $SU(2)$, whose fundamental representation $\mathbf{2}$ does not admit a real structure. In fact here the charge conjugation can be taken as the totally antisymmetric symbol $C_{ab}=\varepsilon_{ab}$, and the Majorana condition would be inconsistent: $\psi^{T}C=\psi^{\dagger}\Leftrightarrow\psi=0.$ (4.97) Thus we cannot construct an $\mathcal{N}=1$ Euclidean supersymmetry algebra in 3 dimensions, in the sense of definition (4.3.1). The problem can be cured considering a reducible spin representation $S$, where the spinors and the charge conjugation matrix can be split as $\Psi=(\psi^{a}_{I})^{a=1,2}_{I=1,\cdots,\mathcal{N}},\qquad\mathcal{C}=(\Omega_{IJ}C^{ab})^{a,b=1,2}_{I,J=1,\cdots,\mathcal{N}},$ (4.98) and the same reality condition $\Psi^{\dagger}=\Psi^{T}\mathcal{C}$ now is consistent if also the matrix $\Omega$ squares to $-\mathds{1}$ and is anti- orthogonal: $\Omega=-\Omega^{T}=-\Omega^{-1}.$ (4.99) If we now fix $\mathcal{N}=2$, the resulting spinor representation is analogous to the one of $\mathcal{N}=1$ in 4-dimensions, but now the two Weyl sectors are independent since they generates the two supersymmetries. To see this corrispondence, we can change basis of $S=\mathbf{2}^{(1)}\oplus\mathbf{2}^{(2)}$ from the natural one in terms of the generators $\\{Q^{1}_{a},Q^{2}_{a}\\}$ to $Q_{a}:=\frac{1}{\sqrt{2}}(Q_{a}^{1}+iQ_{a}^{2}),\qquad\tilde{Q}_{a}:=\frac{1}{\sqrt{2}}(Q_{a}^{1}-iQ_{a}^{2}).$ (4.100) In this basis, using (4.44) the super Lie brackets become $\begin{array}[]{lr}\lx@intercol[Q_{a},\tilde{Q}_{b}]=2(\gamma^{\mu})_{ab}P_{\mu}+Z\varepsilon_{ab}\hfil\lx@intercol\\\ \left[Q_{a},Q_{b}\right]=0&[\tilde{Q}_{a},\tilde{Q}_{b}]=0\end{array}$ (4.101) where $Z$ is a constant central charge, and the gamma matrices in this representation can be chosen to be the Pauli matrices $\gamma^{\mu}=\sigma^{\mu}$ for $\mu=1,2,3$.202020There is also another inequivalent representation of the Clifford algebra, as in any odd dimensions, in which $\gamma^{3}=-\sigma^{3}$. We chose the former one. Note that the 4-dimensional $Spin(3,1)$ Lorentz group breaks to $SU(2)\times SU(2)_{R}$, where $Spin(3)\cong SU(2)$ is the 3-dimensional Lorentz group, and the remaining $SU(2)_{R}$ is an R-symmetry acting on the $\mathcal{N}=2$ algebra. The generators $Q_{a}$ and $\tilde{Q}_{a}$ are represented in superspace by odd vector fields whose expressions are formally the same as in (4.58), and the $\mathcal{N}=2$ supersymmetry variation of a superfield $\Phi$ is $\delta_{\epsilon,\eta}\Phi=(\epsilon^{a}Q_{a}+\eta^{a}\tilde{Q}_{a})\Phi$ (4.102) where now, as said before, $\epsilon$ and $\eta$ are two independent complex spinors. If we want to construct a supersymmetric gauge theory in 3-dimensions, we consider the vector superfield, now expressed in WZ gauge as $V(x,\theta,\tilde{\theta})=A_{\mu}(\tilde{\theta}\gamma^{\mu}\theta)+i\sigma\theta\tilde{\theta}+\lambda_{a}\theta^{a}\tilde{\theta}^{(2)}+\lambda^{\dagger}_{a}\tilde{\theta}^{a}\theta^{(2)}+D\theta^{(2)}\tilde{\theta}^{(2)}.$ (4.103) The off-shell $\mathcal{N}=2$ gauge multiplet is then composed by a gauge field $A_{\mu}$, two real scalars $\sigma,D$ and a 2-component complex spinor $\lambda$. Notice that this is just the _dimensional reduction_ of the $\mathcal{N}=1$ multiplet in 4-dimensions, with $\sigma$ coming from the zero- th component of the gauge field in higher dimensions. The only difference with the 4-dimensional vector multiplet is that this zero-th component has been considered purely immaginary, _i.e._ $A_{0}=i\sigma$ with real $\sigma$. This ensures the kinetic term for $\sigma$ to be positive definite and the path integral to converge, matching the would-be dimensional reduction from an Euclidean 4-dimensional theory. If the gauge group is $G$, all fields are valued in its Lie algebra $\mathfrak{g}$. For what we are going to discuss in the next chapter, we now adopt the convention of [49, 50] for the supersymmetry variations of the vector superfield and the supersymmetric actions. Under a proper rescaling of the component fields and of the supercharges, one can work them out in an analogous way to which we did in the last sections, and get $\displaystyle\delta_{\epsilon,\eta}A_{\mu}=\frac{i}{2}(\eta^{\dagger}\gamma_{\mu}\lambda-\lambda^{\dagger}\gamma_{\mu}\epsilon)$ (4.104) $\displaystyle\delta_{\epsilon,\eta}\sigma=\frac{1}{2}(\eta^{\dagger}\lambda-\lambda^{\dagger}\epsilon)$ $\displaystyle\delta_{\epsilon,\eta}D=\frac{i}{2}\left(\eta^{\dagger}\gamma^{\mu}D_{\mu}\lambda-(D_{\mu}\lambda^{\dagger})\gamma^{\mu}\epsilon\right)-\frac{i}{2}\left(\eta^{\dagger}[\lambda,\sigma]-[\lambda^{\dagger},\sigma]\epsilon\right)$ $\displaystyle\delta_{\epsilon,\eta}\lambda=\left(-\frac{1}{2}\gamma^{\mu\nu}F_{\mu\nu}-D+i\gamma^{\mu}D_{\mu}\sigma\right)\epsilon$ $\displaystyle\delta_{\epsilon,\eta}\lambda^{\dagger}=\eta^{\dagger}\left(-\frac{1}{2}\gamma^{\mu\nu}F_{\mu\nu}+D-i\gamma^{\mu}D_{\mu}\sigma\right)$ where $D_{\mu}=\partial_{\mu}+[A_{\mu},\cdot]$ is the gauge-covariant derivative and $\gamma^{\mu\nu}:=\frac{1}{2}[\gamma^{\mu},\gamma^{\nu}]$. Up to some prefactors, they can be seen as a dimensional reduction of (4.82). We can consider two types of gauge supersymmetric actions constructed from the vector multiplet in 3 Euclidean dimensions: the Super Yang-Mills theory, that is a reduction of (4.88), and the Super Chern-Simons (SCS) theory. In superspace, the former one is constructed in the same way as the 4-dimensional case from the spinorial superfield $W_{a}$, while the SCS term is constructed as $S_{CS}=\int d^{3}xd^{2}\theta d^{2}\tilde{\theta}\ \frac{k}{4\pi}\left(\int_{0}^{1}dt\ \mathrm{Tr}\left\\{V\tilde{D}^{a}e^{-tV}D_{a}e^{tV}\right\\}\right).$ (4.105) Integrating out the odd coordinates in superspace, these are given by [49]: $\displaystyle S_{YM}=\int d^{3}x\ \mathrm{Tr}\left\\{\frac{i}{2}\lambda^{\dagger}\gamma^{\mu}D_{\mu}\lambda+\frac{1}{4}F_{\mu\nu}F^{\mu\nu}+\frac{1}{2}D_{\mu}\sigma D^{\mu}\sigma+\frac{i}{2}\lambda^{\dagger}[\sigma,\lambda]+\frac{1}{2}D^{2}\right\\},$ (4.106) $\displaystyle S_{CS}=\frac{k}{4\pi}\int d^{3}x\ \mathrm{Tr}\left\\{\varepsilon^{\mu\nu\rho}\left(A_{\mu}\partial_{\nu}A_{\rho}+\frac{2i}{3}A_{\mu}A_{\nu}A_{\rho}\right)-\lambda^{\dagger}\lambda+2\sigma D\right\\}.$ (4.107) #### 4.3.7 Euclidean 4d $\mathcal{N}=4,2,2^{*}$ supersymmetric gauge theories We describe here another example that will be useful in the next chapter, when we will apply the localization principle to supersymmetric QFT. The $\mathcal{N}=4$ SYM theory on flat space can be derived via dimensional reduction of $\mathcal{N}=1$ SYM in $(9+1)$ dimensions.212121For convergence of the partition function, it would be nicer to start from the $(10,0)$ Euclidean signature. We follow the convention of [11] and start from the $(9,1)$ one, Wick rotating a posteriori the path integral when needed, to match the would-be reduction from the $(10,0)$ theory. The $\mathcal{N}=2$ and $\mathcal{N}=2^{*}$ theories can be derived as modification of the $\mathcal{N}=4$ theory, as we will see later. We start recalling the structure of the 10-dimensional Clifford algebra following the conventions of [11]. This is independent from the choice of the signature, $Cl(9,1)\cong Cl(1,9)\cong\mathrm{Mat}_{32}(\mathbb{R})$, it is real, and generated by the gamma matrices $(\gamma^{M})_{M=0,1,\cdots,9}$ such that $\\{\gamma^{M},\gamma^{N}\\}=2\eta^{MN}$ where $\eta$ is the 10-dimensional Minkowski metric, that we take with signature $(-,+,\cdots,+)$. The fundamental representation of the spin group $Spin(9,1)\hookrightarrow Cl(9,1)$ is then Majorana, and it is moreover reducible under chirality [42] $\gamma^{c}:=-i\gamma^{0}\gamma^{1}\cdots\gamma^{9}$ as $Spin(9,1)=S^{+}\oplus S^{-}\cong\mathrm{Mat}_{16}(\mathbb{R})\oplus\mathrm{Mat}_{16}(\mathbb{R})$. Thus fundamental spinors are Majorana-Weyl, and have 16 real components. In the chiral basis we denote $\displaystyle\gamma^{M}=\left(\begin{array}[]{cc}0&\tilde{\Gamma}^{M}\\\ \Gamma^{M}&0\end{array}\right)\qquad\tilde{\Gamma}^{M},\Gamma^{M}:S^{\pm}\to S^{\mp}$ (4.108) $\displaystyle\gamma^{MN}=\left(\begin{array}[]{cc}\tilde{\Gamma}^{[M}\Gamma^{N]}&0\\\ 0&\Gamma^{[M}\tilde{\Gamma}^{N]}\end{array}\right)=:\left(\begin{array}[]{cc}\Gamma^{MN}&0\\\ 0&\tilde{\Gamma}^{MN}\end{array}\right)$ where $\Gamma^{M},\tilde{\Gamma}^{M}$ act on the Majorana-Weyl subspaces, exchanging chirality, and are taken to be symmetric.222222In the Euclidean signature, we would use $\Gamma^{M}_{E}=\\{\Gamma^{1},\cdots,\Gamma^{9},i\Gamma^{0}\\}$. Let the gauge group $G$ be a compact Lie group, and $\mathfrak{g}$ its Lie algebra. The (on-shell) component field content of the gauge multiplet in 10 dimensions is of a gauge field, locally represented as $A\in\Omega^{1}(\mathbb{R}^{9,1},\mathfrak{g})$, and a gaugino, a Mayorana- Weyl spinor $\Psi:\mathbb{R}^{9,1}\to S^{+}\otimes\mathfrak{g}$ with values in the Lie algebra $\mathfrak{g}$. The field strength of the gauge field is locally represented by $F=dA+[A,A]$, and the associated gauge-covariant derivative on $\mathbb{R}^{9,1}$ is $D_{M}=\partial_{M}+A_{M}$. The supersymmetry variations under the action of the 10-dimensional super-Poincaré algebra are $\displaystyle\delta_{\epsilon}A_{M}$ $\displaystyle=\epsilon\Gamma_{M}\Psi$ (4.109) $\displaystyle\delta_{\epsilon}\Psi$ $\displaystyle=\frac{1}{2}\Gamma^{MN}F_{MN}\epsilon$ where $\epsilon$ is a Majorana-Weyl spinor, analogously to the on-shell version of (4.82) up to the chirality projection and conventional prefactors. The action functional for the $\mathcal{N}=1$ 10-dimensional theory is $S_{10d}=\int d^{10}x\ \mathcal{L}$, with Lagrangian $\mathcal{L}=\frac{1}{g_{YM}^{2}}\mathrm{Tr}\left(\frac{1}{2}F_{MN}F^{MN}-\Psi\Gamma^{M}D_{M}\Psi\right)$ (4.110) where $\mathrm{Tr}$ denotes a symmetric bilinear pairing in $\mathfrak{g}$,232323For semisimple $\mathfrak{g}$, this is the Killing form as usual. and $g_{YM}$ is the Yang-Mills coupling constant. As we remarked in Section 4.3.3, this action is exactly supersymmetric under (4.109) without the addition of auxiliary fields. To get the Euclidean 4-dimensional theory, we perform dimensional reduction along the directions $x^{0},x^{5},\cdots,x^{9}$, assuming independence of the fields on these coordinates. The fields split as $\displaystyle A_{M}$ $\displaystyle\to\left((A_{\mu})_{\mu=1,\cdots,4},(\Phi_{A})_{A=5,\cdots,9,0}\right)$ (4.111) $\displaystyle\Psi$ $\displaystyle\to\left(\psi^{L}\ \chi^{R}\ \psi^{R}\ \chi^{L}\right)^{T}$ where $\psi^{L/R},\chi^{L/R}$ are four-component real chiral spinors. The spacetime symmetry group $Spin(9,1)$ is broken to $Spin(4)\times Spin(5,1)^{\mathcal{R}}\hookrightarrow Spin(9,1)$, where $Spin(4)\cong SU(2)_{L}\times SU(2)_{R}$ acts on the $x^{1},\cdots,x^{4}$ directions, and the R-symmetry group $Spin(5,1)^{\mathcal{R}}$ rotates the other ones. It is often convenient to further break the R-symmetry group to $Spin(4)^{\mathcal{R}}\times SO(1,1)^{\mathcal{R}}\hookrightarrow Spin(5,1)^{\mathcal{R}}$, where the first piece $Spin(4)^{\mathcal{R}}\cong SU(2)_{L}^{\mathcal{R}}\times SU(2)_{R}^{\mathcal{R}}$ rotates the $x^{5},\cdots,x^{8}$ directions, and $SO(1,1)^{\mathcal{R}}$ acts on the $x^{9},x^{0}$ ones. We thus consider the symmetry group $SU(2)_{L}\times SU(2)_{R}\times SU(2)_{L}^{\mathcal{R}}\times SU(2)_{R}^{\mathcal{R}}\times SO(1,1)^{\mathcal{R}}$ (4.112) under which the fields behave as * • $A_{\mu}$: vector of $SU(2)_{L}\times SU(2)_{R}$, scalar under R-symmetry; * • $(\Phi_{I})_{I=4,\cdots,8}$: 4 scalars under $SU(2)_{L}\times SU(2)_{R}$, vector of $SU(2)_{L}^{\mathcal{R}}\times SU(2)_{R}^{\mathcal{R}}$, scalars under $SO(1,1)^{\mathcal{R}}$; * • $\Phi_{9},\Phi_{0}$: scalars under $SU(2)_{L}\times SU(2)_{R}\times SU(2)_{L}^{\mathcal{R}}\times SU(2)_{R}^{\mathcal{R}}$, vector of $SO(1,1)^{\mathcal{R}}$; * • $\psi^{L/R}$: $\left(\frac{1}{2},0\right)/\left(0,\frac{1}{2}\right)$ of $SU(2)_{L}\times SU(2)_{R}$, $\left(\frac{1}{2},0\right)$ of $SU(2)_{L}^{\mathcal{R}}\times SU(2)_{R}^{\mathcal{R}}$, $+/-$ of $SO(1,1)^{\mathcal{R}}$; * • $\chi^{L/R}$: $\left(\frac{1}{2},0\right)/\left(0,\frac{1}{2}\right)$ of $SU(2)_{L}\times SU(2)_{R}$, $\left(0,\frac{1}{2}\right)$ of $SU(2)_{L}^{\mathcal{R}}\times SU(2)_{R}^{\mathcal{R}}$, $-/+$ of $SO(1,1)^{\mathcal{R}}$; here we denoted $+,-$ the inequivalent Majorana-Weyl representations of $SO(1,1)^{\mathcal{R}}$, seen as a subgroup of $Cl(1,1)\cong\mathrm{Mat}_{2}(\mathbb{R})$. The above decomposition of $Spin(9,1)$ into four subrepresentations of $Spin(4)$, rotated into each other by the R-symmetry group, gives the $\mathcal{N}=4$ supersymmetry algebra on $\mathbb{R}^{4}$. The supersymmetry variations of the reduced component fields are given by (4.109), read in terms of the splitting (4.111), $\displaystyle\delta_{\epsilon}A_{\mu}$ $\displaystyle=\epsilon\Gamma_{\mu}\Psi$ (4.113) $\displaystyle\delta_{\epsilon}\Phi_{A}$ $\displaystyle=\epsilon\Gamma_{A}\Psi$ $\displaystyle\delta_{\epsilon}\Psi$ $\displaystyle=\frac{1}{2}\left(\Gamma^{\mu\nu}F_{\mu\nu}+\Gamma^{AB}[\Phi_{A},\Phi_{B}]+\Gamma^{\mu A}D_{\mu}\Phi_{A}\right)\epsilon.$ The action of the 4d $\mathcal{N}=4$ SYM theory is $S^{\mathcal{N}=4}=\int d^{4}x\ \mathcal{L}$ with the Lagrangian obtained by the reduction of (4.110). More explicitly, $S^{\mathcal{N}=4}=\int d^{4}x\frac{1}{g_{YM}^{2}}\mathrm{Tr}\left(\frac{1}{2}F_{\mu\nu}F^{\mu\nu}+(D_{\mu}\Phi_{A})^{2}-\Psi\Gamma^{\mu}D_{\mu}\Psi+\frac{1}{2}[\Phi_{A},\Phi_{B}]^{2}-\Psi\Gamma^{A}[\Phi_{A},\Psi]\right).$ (4.114) Notice that, since contractions of $A,B$ indices are done with a reduced Minkowski metric, upon dimensional reduction from the Lorentzian theory the scalar $\Phi_{0}$ has a negative kinetic term. Analogously to the last section, we consider it to be purely immaginary, i.e. $\Phi_{0}=:i\Phi_{0}^{E}$ with $\Phi_{0}^{E}$ real. This makes the path integral match with the would-be reduction from the Euclidean $(10,0)$-dimensional theory. The $\mathcal{N}=4$ algebra closes _on-shell_. In fact, it can be obtained from (4.113) that $\delta_{\epsilon}^{2}=\frac{1}{2}[\delta_{\epsilon},\delta_{\epsilon}]=-\mathcal{L}_{v}-G_{\Phi}$ (4.115) up to the imposition of the EoM for $\Psi$, $\Gamma^{M}D_{M}\Psi=0$. Here $v^{M}:=\epsilon\Gamma^{M}\epsilon$, $\mathcal{L}_{v}$ is the Lie derivative (the action of the translation algebra) with respect to $v\sim v^{\mu}\partial_{\mu}$, and $G_{\Phi}$ is an infinitesimal gauge transformation with respect to $\Phi:=A_{M}v^{M}$. A famous non- renormalization theorem by Seiberg [51] states that the $\mathcal{N}=4$ theory is actually _superconformal_ , i.e. it has a larger supersymmetry algebra that squares to the _conformal algebra_ , whose generators are the Poincaré generators plus the generators of dilatations and special conformal transformations.242424More precisely, the theorem states that the beta function of $g_{YM}$ is zero non-perturbatively. This means that the theory is fully scale invariant at quantum level. In fact, one can see that $S^{\mathcal{N}=4}$ is classically invariant under supersymmetry variations with respect to the non-constant spinor $\epsilon=\hat{\epsilon}_{s}+x^{\mu}\Gamma_{\mu}\hat{\epsilon}_{c}$ (4.116) where $\hat{\epsilon}_{s},\hat{\epsilon}_{c}$ are constant spinors parametrizing supertranslations and superconformal transformations. This enlarged supersymmetry algebra closes now on the superconformal algebra, $\delta_{\epsilon}^{2}=-\mathcal{L}_{v}-G_{\Phi}-R-\Omega$ (4.117) where $R$ is a $Spin(5,1)^{\mathcal{R}}$ rotation, acting on scalars as $(R\cdot\Phi)_{A}=R_{A}^{B}\Phi_{B}$, and on spinors as $(R\cdot\Psi)=\frac{1}{4}R_{AB}\Gamma^{AB}\Psi$, where $R_{AB}=2\epsilon\tilde{\Gamma}_{AB}\tilde{\epsilon}$. $\Omega$ is an infinitesimal dilatation with respect to the parameter $2(\tilde{\epsilon}\epsilon)$, acting on the gauge field trivially, on scalars as $\Omega\cdot\Phi=-2(\tilde{\epsilon}\epsilon)\Phi$ and on spinors as $\Omega\cdot\Psi=-3(\tilde{\epsilon}\epsilon)\Psi$. This new bosonic transformations are clearly symmetries of $S^{\mathcal{N}=4}$. Now we can restrict the attention to an $\mathcal{N}=2$ subalgebra, considering the variations with respect to Majorana-Weyl spinors of the form $\epsilon=\left(\epsilon^{L}\ 0\ \epsilon^{R}\ 0\right)^{T}$ (4.118) so in the subrepresentation $\left(\left(\frac{1}{2},0\right)\oplus\left(0,\frac{1}{2}\right)\right)\oplus\left(\frac{1}{2},0\right)^{\mathcal{R}}\oplus(+\oplus-)^{\mathcal{R}}$, the eigenspace of $\Gamma^{5678}$ with eigenvalue +1. With respect to these supersymmetry variations, the gauge multiplet further splits in * • $(A_{\mu},\Phi_{9},\Phi_{0},\psi^{L},\psi^{R})$: the $\mathcal{N}=2$ vector multiplet; * • $(\Phi_{I},\chi^{L},\chi^{R})$: the $\mathcal{N}=2$ hypermultiplet, with value in the adjoint representation of $G$. These two multiplets are completely disentangled in the free theory limit $g_{YM}^{2}\to 0$.252525Working out the restricted supersymmetry variations taking into account the splitting of the gaugino, some non-linear term, coupling the fermionic sectors of the two multiplets, survives because of the gauge interaction. In the free theory limit, after the rescaling $A_{M}\mapsto g_{YM}A_{M},\Psi\mapsto g_{YM}\Psi$, these terms go to zero. The same Lagrangian thus equivalently describes an $\mathcal{N}=2$ matter-coupled gauge theory. It is also possible to insert a mass for the hypermultiplet, breaking explicitly the conformal invariance, and obtain the so-called $\mathcal{N}=2^{*}$ theory. Since the fields of the vector multiplet are all scalars under $SU(2)_{R}^{\mathcal{R}}$, and the hypermultiplet fields are all in the $\frac{1}{2}$ representation, these mass terms can at most rotate the hypermultiplet content with an $SU(2)_{R}^{\mathcal{R}}$ transformation. Thus replacing $D_{0}\Phi_{I}\mapsto[\Phi_{0},\Phi_{I}]+M_{I}^{J}\Phi_{J}$ and $D_{0}\Psi\mapsto[\Phi_{0},\Psi]+\frac{1}{4}M_{IJ}\Gamma^{IJ}\Psi$, where $(M^{I}_{J})$ represents an $SU(2)_{R}^{\mathcal{R}}$ rotation in the vector representation, one obtains the mass terms for the $\Phi_{I}$ and $\chi$ fields. Notice that $\delta_{\epsilon}^{2}$ gets a contribution from the Lie derivative with respect to $v^{0}\partial_{0}\cong 0\mapsto v^{0}M$, so that in the $2^{*}$ theory $\displaystyle\delta_{\epsilon}^{2}\Phi_{I}$ $\displaystyle\mapsto(\delta_{\epsilon}\Phi_{I})_{\mathcal{N}=2}-v^{0}M_{I}^{J}\Phi_{J}$ (4.119) $\displaystyle\delta_{\epsilon}^{2}\chi$ $\displaystyle\mapsto(\delta_{\epsilon}\chi)_{\mathcal{N}=2}-\frac{1}{4}v^{0}M_{IJ}\Gamma^{IJ}\chi.$ In the limits of infinite or zero mass, the pure $\mathcal{N}=2$ or $\mathcal{N}=4$ theory is recovered. Notice that, since we argued that $\Phi_{0}$ should be integrated over purely immaginary values for the convergence of the path integral, also $(M_{IJ})$ should be taken purely immaginary. ### 4.4 From flat to curved space Recently, localization theory has been extensively used in the framework of quantum field theories with rigid super-Poincaré symmetry, to compute exactly partition functions or expectation values of certain supersymmetric observables, when the theory is formulated on a _curved_ compact manifold. This cures the corresponding partition functions from infrared divergences making the path integral better defined, and is consistent with the requirement of periodic boundary conditions on the fields, that allows to generalize properly the Cartan model on the infinite dimensional field space. We will come back to this last point in the next chapter, when we will study circle localization of path integrals, while we close this chapter reviewing the idea behind some common approaches used to formulate rigid supersymmetry on curved space. Following the approach of the last section, we would have to understand what does it mean to have supersymmetry on a generic metric manifold $(\mathcal{M},g)$ (Riemannian or pseudo-Riemannian) of dimension $\dim{\mathcal{M}}=d$ from a geometric point of view. The supersymmetry of flat space was constructed as a super-extension of Minkowski (or Euclidean) space $\mathbb{R}^{d}$, starting from a super-extension of the Lie algebra of its isometry group, the Poincaré group. Now in general the Poincaré group is not an isometry group for $\mathcal{M}$, so the super Poincaré algebra $\mathfrak{siso}_{S}(d)$ with respect to some (real or Majorana) spin representation $S$ cannot be fully interpreted as a “supersymmetry” algebra for the space at hand. We can nonetheless associate in some way this algebra to a suitable super-extension of $\mathcal{M}$, and then ask what part of it can be preserved as a supersymmetry of this supermanifold. We follow [52] for this geometric introduction. Since we want to work with spinors, we assume that $\mathcal{M}$ admits a spin-structure. In particular, it exists a (real) spinor bundle $S\to\mathcal{M}$ associated to the spin-structure, with structure sheaf $\mathcal{S}:\mathcal{S}(U)=\Gamma(U,S),\forall U\subset\mathcal{M}$ open. Analogously to the flat superspace of the last section, we make now a super- extension of $\mathcal{M}$ through this spinor bundle considering the _odd spinor bundle_ $S\mathcal{M}_{S}:=\Pi S$, with body $\mathcal{M}$ and structure sheaf $\bigwedge\mathcal{S}^{*}:\bigwedge\mathcal{S}^{*}(U)=C^{\infty}(\mathcal{U})\otimes\bigwedge(S_{0}^{*}),\forall U\subset\mathcal{M}$ open, where $S_{0}$ is the typical fiber of $S$. From proposition 4.1.1, for any $p\in\mathcal{M}$, there is an isomorphism of $\mathbb{Z}_{2}$-graded vector spaces $T_{p}S\mathcal{M}_{S}\cong T_{p}\mathcal{M}\oplus S_{p}[1]$. Now, the vector bundle $V:=T\mathcal{M}\oplus S$ over $\mathcal{M}$ carries the canonical spin-connection induced by the Levi-Civita connection of the manifold $(\mathcal{M},g)$. Assume that we can pick a parallel non-degenerate $Spin(d)$-invariant bilinear form $\beta$ on $S$ with respect to this connection.262626This is always true if $\mathcal{M}$ is simply-connected. We can think of the $Spin(d)$-invariant bilinear form $\tilde{g}=g+\beta$ as a (pseudo-)Riemannian metric on the supermanifold $S\mathcal{M}_{S}$. Moreover, associated to the bilinear form $\beta$ we have the map $\Gamma:S^{2}\to T\mathcal{M}$, that is a point-wise generalization of the usual symmetric and Spin-equivariant bilinear form for a Spin representation $S_{0}\cong S_{p},\forall p\in\mathcal{M}$. This means that we can consider the bundle $\mathfrak{p}(V):=\mathfrak{spin}(d)\oplus V$ (4.120) as a _bundle of super Poincaré algebras_ over $\mathcal{M}$, with the bracket structure extended through $\Gamma$. Having found how to (point-wise) set up the super Poincaré algebra on top of the supermanifold $S\mathcal{M}_{S}$ constructed from $(\mathcal{M},g)$, we wish to establish which section of the super Poincaré bundle $\mathfrak{p}(V)$ produces a suitable generalization of “super-isometry” for $S\mathcal{M}_{S}$. In particular, we pay attention to which sections of $S$, as the odd subbundle of $\mathfrak{p}(V)$, generates “supersymmetries” of the generalized metric $\tilde{g}$. This problem was analyzed in [52], and connected to the problem of finding solution to the so called _Killing spinor equation_ for a section $\psi$ of $S\to\mathcal{M}$. ###### Definition 4.4.1. A section $\psi$ of the spinor bundle $S\to\mathcal{M}$ is called a _twistor spinor_ (or _conformal Killing spinor_) if it exists another section $\phi$ such that, for any vector field $X\in\Gamma(T\mathcal{M})$, $\nabla_{X}\psi=X\cdot\phi$ (4.121) where $X\cdot\phi=X^{\mu}\gamma_{\mu}\phi$ is the _Clifford multiplication_. If in particular $\phi=\lambda\psi$, for some constant $\lambda$, the spinor $\psi$ is called _Killing spinor_. The equation (4.121) is called _twistor_ or _Killing spinor equation_. Note that (4.121) directly implies $\phi=\pm(1/\dim(\mathcal{M}))\not{\nabla}\psi$, where $\not{\nabla}:=\gamma^{\mu}\nabla_{\mu}$ is the Dirac operator, and the sign depends on conventions. The twistor spinor equation is thus equivalently written as $\nabla_{X}\psi=\pm\frac{1}{\dim{(\mathcal{M})}}X\cdot\not{\nabla}\psi.$ (4.122) This characterizes the Killing spinors as those twistor spinors that satisfies also the Dirac equation $\not{\nabla}\psi=m\psi$ for some constant $m$. The main result proved in [52] is stated in the following theorem. ###### Theorem 4.4.1. Consider the supermanifold $S\mathcal{M}_{S}$ with the bilinear form $\tilde{g}=g+\beta$, and a section $\psi$ of $S$. The odd vector field $X_{\psi}$ associated to $\psi$ is a Killing vector field of $(S\mathcal{M}_{S},\tilde{g})$ if and only if $\psi$ is a twistor spinor. Here the Killing vector condition on the supermanifold is a conceptually straightforward generalization of the usual concept of Killing vector fields on a smooth manifold. It can be natually stated in terms of superframe fields. We refer to the above cited article for the details. Notice that, in particular, Killing spinors generate infinitesimal isometries of the supermanifold $S\mathcal{M}_{S}$, and thus are good candidates to describe the “preserved” supersymmetries of the odd part of the super Poincaré algebra, when this is associated to the generic curved manifold $\mathcal{M}$ in the way we saw above. See also [53] for a review on Killing spinors in (pseudo-)Riemannian geometry. From the QFT point of view, it is possible to derive a (generalized) Killing spinor equation, describing the preserved supercharges on the curved space, from a dynamical approach. This idea is based on a procedure also valid in the non-supersymmetric setting, when one aims to deform a certain QFT to redefine it on a generic curved manifold. In this case, one couples the theory to _background_ gravity, letting the metric fluctuate.272727Since the metric can fluctuate and the field theory is defined locally, there is no harm in principle in considering different topologies of the base manifold, like requiring it to be compact. Then the gravitational sector is decoupled from the rest of the theory taking the gravitational constant $G_{N}\to 0$, while the metric is linearized around the chosen _off-shell_ configuration $g=\eta+h$ and the higher order corrections disappear in the limit of weak gravitational interaction. It is important that we do not constrain the gravitational field to satisfy the equation of motion, since it is considered as a background (classical) field. The same idea applies when the theory is defined in the supersymmetric setting: in this case to preserve supersymmetry one has to couple to background _supergravity (SUGRA)_. The resulting field theory will contain then more fields belonging to the so-called _supergravity multiplet_. This time, taking the limit $G_{N}\to 0$ we fix all the background supergravity multiplet to an allowed off-shell configuration. Note that in particular, the auxiliary fields are not eliminated in terms of the other fields using their equations of motion. If then there are supergravity transformations that leave the given background invariant, we say that the corresponding rigid supercharges are _preserved_ on this background. This procedure was systematically introduced in [54], then many cases and classification were made in different dimensions and with different degree of supersymmetriy (see for example [55, 56, 57, 58]). #### 4.4.1 Coupling to background SUGRA Suppose we have a supersymmetric field theory formulated on flat space specified by its Lagrangian $\mathcal{L}^{(0)}$, whose variation under supersymmetry is a total derivative:282828For simplicity, we consider now the formulation on the even space $\mathbb{R}^{d}$ or $\mathcal{M}$, at the level of component fields of the given supersymmetry multiplets. The supersymmetry variation of these component fields are those coming from the action of the odd supertranslations in superspace. $\delta\mathcal{L}^{(0)}=*d*(\cdots)=\partial_{\mu}(\cdots)^{\mu}$ (4.123) We can introduce supergravity by requiring the action of the super-Poincaré group to be _local_ , employing the usual gauge principle and minimal coupling or Noether procedure. In the non-supersymmetric setting, this would mean to introduce a gauge symmetry under local coordinate transformations, realized via diffeomorphisms on $\mathcal{M}$. The Noether current associated to such infinitesimal transformations is the _energy-momentum tensor_ $T^{\mu\nu}$, that we take to be symmetric.292929In general this will not be a symmetric tensor, but there always exists a suitable modification that makes it symmetric, and moreover equivalent to the Hilbert definition of energy momentum tensor as a source of gravitational field. This is the _Belinfante–Rosenfeld tensor_ $\tilde{T}^{\mu\nu}:=T^{\mu\nu}+\frac{1}{2}\nabla_{\lambda}(S^{\mu\nu\lambda}+S^{\nu\mu\lambda}-S^{\lambda\nu\mu})$ (4.124) where $S^{\mu}_{\nu\lambda}$ is the spin part of the Lorentz generators in a given spin representation satisfying $\nabla_{\mu}S^{\mu}_{\nu\lambda}=T_{\nu\mu}-T_{\mu\nu}$, and $\nabla$ is an appropriate torsion-free spin-covariant derivative induced from the metric (see for example [59]). The minimal coupling procedure then requires to modify the Lagrangian, $\mathcal{L}^{\prime}=\mathcal{L}^{(0)}+h_{\mu\nu}T^{\mu\nu}+O(h^{2})$ (4.125) where $h_{\mu\nu}$ is regarded as a variation of the metric from the flat space values $\eta_{\mu\nu}$, and $O(h^{2})$ are seagull non-linear terms that can be fixed requiring the gauge invariance of $\mathcal{L}^{\prime}$. The resulting non-linear coupling is obtainable substituting in the original theory $d\mapsto\nabla,\qquad\eta\mapsto g=\eta+h,$ (4.126) where $\nabla$ is the gauge-covariant derivative with respect to a connection $\Gamma$, that we take as the Levi-Civita connection. The theory is now coupled to a gravitational (classical) _background_. If we want to make the graviton field $h$ dynamical, we can add an Hilbert-Einstein term to $\mathcal{L}^{(0)}$, $\mathcal{L}_{HE}=-\frac{\sqrt{|\det{g}|}}{2\kappa^{2}}Ric_{g}$ (4.127) where $Ric_{g}$ is the Ricci scalar associated to $g$, and $\kappa:=1/M_{p}=\sqrt{8\pi G_{N}}$. In the supersymmetric setting, gauging the super-Poincaré group leads to the introduction of more fields into the theory, since as we know they can be interpreted as components of superfields in superspace, and thus belong to supersymmetry multiplets. In particular, we have to introduce the _graviton multiplet_ composed by the metric $g$, the gravitino $\Psi$ and other (maybe auxiliary) fields. The particular field content depends on the number $\mathcal{N}$ of supersymmetries, the dimensionality of the theory, the presence or absence of an R-symmetry and whether the theory is or not superconformal. Consequently, also the energy-momentum tensor will belong to a multiplet, the so-called _supercurrent multiplet_ , composed by $T$, a _supercurrent_ $J$ associated to the local invariance under (odd) supertranslations, and other fields. We can schematically perform the first steps of the Noether procedure to see how the components of these multiplets arise naturally. We start from the odd part of the super-Poincaré algebra $\mathfrak{iso}(\mathbb{R}^{d})\oplus S[1]$, writing the infinitesimal variation of the Lagrangian in terms of the supercurrent: $\delta_{\epsilon}\mathcal{L}^{(0)}\equiv\epsilon\cdot\mathcal{L}^{(0)}=(\partial_{\mu}J^{\mu})\epsilon$ (4.128) where $\epsilon$ is now a Majorana spinor field, i.e. a section of the spinor bundle with fiber $S$. The supercurrent $J$ is an $S$-valued vector field, and the spinor contraction is done via the usual charge conjugation matrix. We couple this current to a gauge field $\Psi$, to be identified with the gravitino, that is locally an $S$-valued 1-form such that, at linearized level, $\delta_{\epsilon}\Psi_{\mu}=\frac{1}{\kappa}\partial_{\mu}\epsilon$ (4.129) where the constant $\kappa$ is introduced for dimensional reasons.303030If we canonically take mass dimensions of scalars to be $(d-2)/2$ and of spinors to be $(d-1)/2$, and since schematically $\delta_{\epsilon}(boson)=(fermion)\epsilon$ than $[\epsilon]=-1/2$, so $\kappa$ must be a dimensionfull parameter of mass dimension $[\kappa]=(2-d)/d$. We can so identify this constant as the gravitational constant previously defined. Then we add a term to the Lagrangian: $\mathcal{L}^{\prime}=\mathcal{L}^{(0)}+\kappa\Psi_{\mu}J^{\mu}.$ (4.130) Now the variation of $\mathcal{L}^{\prime}$ is proportional to the variation of the current $\delta_{\epsilon}J$. Since the supercurrent is a supersymmetry variation of the original Lagrangian, its variation will be proportional to the action of the translation generators $P_{\mu}$: $\displaystyle\delta_{\epsilon}^{2}\mathcal{L}^{(0)}$ $\displaystyle=\epsilon\cdot(\epsilon\cdot\mathcal{L}^{(0)})=\partial_{\mu}(\delta_{\epsilon}J^{\mu})\epsilon$ (4.131) $\displaystyle=(1/2)[\epsilon,\epsilon]\cdot\mathcal{L}^{(0)}=\overline{\epsilon}\gamma^{\nu}\epsilon(P_{\nu}\cdot\mathcal{L}^{(0)})=(\overline{\epsilon}\gamma_{\nu}\epsilon)\partial_{\mu}T^{\mu\nu}$ $\displaystyle\Rightarrow\delta_{\epsilon}J^{\mu}$ $\displaystyle=\overline{\epsilon}\gamma_{\nu}T^{\nu\mu}$ where in the second line we wrote the variation under the action of the translation generators in terms of the energy-momentum tensor $T$. We try to restore the gauge-invariance of the Lagrangian by minimally coupling this new current to a new gauge field $h$, that we identify as a metric variation, the _graviton_ $\mathcal{L}^{\prime\prime}=\mathcal{L}^{(0)}+\kappa\Psi_{\mu}J^{\mu}+h_{\mu\nu}T^{\mu\nu},$ (4.132) and naturally requiring the supersymmetry variation of the graviton to be $\delta_{\epsilon}h_{\mu\nu}=\kappa\overline{\epsilon}\gamma_{(\mu}\Psi_{\nu)},$ (4.133) making it the superpartner of the gravitino $\Psi$. The Lagrangian $\mathcal{L}^{\prime\prime}$ is again not supersymmetric, since the variation $\delta_{\epsilon}T^{\mu\nu}\neq 0$ in general, so the Noether procedure is not terminated yet. It is not easy to complete this procedure in this way, but in principle repeating this passages we would introduce more linearly coupled currents and gauge fields that, motivated by supersymmetry, we expect to come from the SUGRA supermultiplets mentioned above. To ensure the supersymmetry of the full Lagrangian at non-linear level, as in non supersymmetric gauge theories, non-linear couplings could have to be introduced as well as non-linear terms in the supersymmetry variations. Summarizing, we expect the fully coupled Lagrangian to be schematically of the form $\mathcal{L}=\mathcal{L}^{(0)}+\kappa J^{\mu}\Psi_{\mu}+h_{\mu\nu}T^{\mu\nu}+\sum_{i}\mathcal{B}^{i}\cdot\mathcal{J}^{i}+(\mathrm{seagull\ terms})$ (4.134) where $\mathcal{B}$ is the multiplet of background gauge fields $(h,\Psi,\cdots)$, $\mathcal{J}$ the supercurrent multiplet $(T,J,\cdots)$, and we referred to possible higher-order terms in the background fields as seagull terms. As already said, the particular field content of these multiplets is not unique, so we remain generic for the moment and refer to the next subsections for some examples. We can absorb the terms proportional to $h$ as in the non-supersymmetric case, by making the substitutions $d\mapsto\nabla$ and $\eta\mapsto g=\eta+h$. If we want to have a full gravitational theory, we can add a kinetic term for the source fields, and complete their supersymmetry variations with possible non-linear terms to ensure gauge invariance. Regarding the metric and the gravitino, the kinetic terms are given by the Hilbert-Einstein action (4.127) and the _Rarita- Schwinger_ action $\mathcal{L}_{RS}=-\frac{\sqrt{|\det{g}|}}{2}\overline{\Psi}_{\mu}\gamma^{\mu\nu\rho}(\nabla_{\nu}\Psi)_{\rho},$ (4.135) where $\gamma^{\mu\nu\rho}:=\gamma^{[\mu}\gamma^{\nu}\gamma^{\rho]}$, and $\nabla$ acts on spinors via the spin connection, $(\nabla_{\mu}\Psi)_{\nu}=\partial_{\mu}\Psi_{\nu}+\frac{1}{4}\omega_{\mu}^{ab}\gamma_{ab}\Psi_{\nu}-\Gamma_{\mu\nu}^{\rho}\Psi_{\rho}$. The supersymmetry variations will be generically $\displaystyle\delta_{\epsilon}h_{\mu\nu}=\kappa\overline{\epsilon}\left\\{\gamma_{(\mu}\Psi_{\nu)}+(\cdots)^{F}\right\\}$ (4.136) $\displaystyle\delta_{\epsilon}\Psi_{\mu}=\frac{1}{\kappa}\left\\{\nabla_{\mu}+(\cdots)_{\mu}^{B}\right\\}\epsilon+O(\kappa\Psi^{2}\epsilon)$ where we stressed that non-linear higher-oreder terms for the gravitino are $\kappa$-suppressed, and the ellipses in both cases collect contributions from the other (fermionic or bosonic, respectively) fields of the supergravity multiplet. Notice that also the supersymmetry variations of the field content of the original $\mathcal{L}^{(0)}$ get modified with respect to their flat- space version. Once one has the full supergravity theory, their transformation rules follow from the corresponding formulas in the appropriate matter-coupled off-shell supergravity. We now consider the _rigid limit_ $G_{N}\to 0$ (or $\kappa\to 0$, or $M_{P}\to\infty$) together with the choice of a given background gravitational multiplet $\mathcal{B}$ compatible with the original request $(\mathcal{M},g)$.313131We stress that a rigid supersymmetric background is characterized by a full set of supergravity background fields, i.e. specifying only the metric does not determine the background. In particular, there are distinct backgrounds that have the same metric but lead to different partition functions. Since we think at this classical configuration as a VEV, we require all the fermion fields in the supergravity multiplet to vanish on this background. We also look for those supergravity transformations that leave this background invariant.323232This can be interpreted as a superisometry requirement with respect to the graviton multiplet. These requirements produce the following effects: * • Fermionic gravitational fields as well as the kinetic term for the bosonic gravitational sector do not contribute to the lagrangian: $\mathcal{L}=\left.\mathcal{L}^{(0)}\right|_{\begin{subarray}{c}d\to\nabla\\\ \eta\to g\end{subarray}}+\sum_{i}\mathcal{B}_{B}^{i}\cdot\mathcal{J}_{B}^{i}+(\mathrm{seagull\ terms})_{B}$ (4.137) At the same time, supersymmetry variations of the bosonic gravitational fields automatically vanish. * • Requiring the supersymmetry of the background is then equivalent to $\delta_{\epsilon}B_{F}=0.$ (4.138) In particular, this condition on the gravitino generates the _generalized Killing spinor equation_ $\delta_{\epsilon}\Psi_{\mu}=0\quad\Leftrightarrow\quad\nabla_{\mu}\epsilon=(\cdots)_{\mu}\epsilon$ (4.139) where, again, ellipses stand for terms proportional to bosonic fields in the graviton multiplet. The solutions to this equation determine which sections of the spinor bundle $S\to\mathcal{M}$ generates the preserved supersymmetry transformations on $\mathcal{M}$. #### 4.4.2 Supercurrent multiplets and metric multiplets As we wrote before, the field content of supercurrent multiplets depends on the general properties of the theory at hand. In [60] it was given a definition from basic general requirements starting from superfields in superspace, that allows a classification by specializing to the various particular cases. It is shown that the most general supercurrent is a real superfield $\mathcal{S}_{a\dot{a}}$ satisfying $\begin{array}[]{ll}\lx@intercol\tilde{D}^{\dot{a}}\mathcal{S}_{a\dot{a}}=\chi_{a}+\mathcal{Y}_{a}\hfil\lx@intercol\\\ \tilde{D}^{\dot{a}}\chi_{a}=0&\tilde{D}^{\dot{a}}\chi^{\dagger}_{\dot{a}}=D^{a}\chi_{a}\\\ D_{(a}\mathcal{Y}_{b)}=0&\tilde{D}^{2}\mathcal{Y}_{a}=0.\end{array}$ (4.140) Every supersymmetric theory has such an $\mathcal{S}$-multiplet, containing the stress energy tensor $T$ and the supercurrent $S$. They are the only component fields with spin larger than one, since they couple to the graviton and the gravitino in the metric multiplet, that respectively are the only component fields with spin higher than one in this multiplet. We report here special examples in 4 and 3 dimensions, that can be derived solving the constraints (4.140) in cases where additional conditions on the superfields $\chi_{a}$ and $\mathcal{Y}_{a}$ hold. For $\mathcal{N}=1$ in 4-dimensions we have three possible interesting special cases: 1. 1. The majority of theories admit a reduction of the $\mathcal{S}$-multiplet into the so-called _Ferrara-Zumino (FZ) multiplet_ $\mathcal{J}^{FZ}_{\mu}\to\left(j_{\mu},(S_{\mu})_{a},x,T_{\mu\nu}\right)$ (4.141) where $j_{\mu}$ is a vector field and $x$ a complex scalar field. 2. 2. If the theory has a $U(1)_{R}$ symmetry, the $\mathcal{S}$-multiplet reduces to the so-called _$\mathcal{R}$ -multiplet_, whose lower degree component is the conserved R-current $j_{\mu}^{(R)}$: $\mathcal{R}_{\mu}\to\left(j_{\mu}^{(R)},(S_{\mu})_{a},T_{\mu\nu},C_{\mu\nu}\right)$ (4.142) where $C_{\mu\nu}$ are the components of a conserved 2-form current, the so- called _brane current_.333333In curved space and in presence of topological defects as strings (1-brane) or domain walls (2-brane), the supersymmetry algebra (4.57) can be modified by the presence of _brane charges_ , $\displaystyle\left[Q_{a},\tilde{Q}_{\dot{b}}\right]$ $\displaystyle=2(\gamma^{\mu})_{a\dot{b}}(P_{\mu}+Z_{\mu})$ $\displaystyle[Q_{a},Q_{b}]$ $\displaystyle=(\gamma^{\mu\nu})_{ab}\tilde{Z}_{\mu\nu}$ where $Z_{\mu},\tilde{Z}_{\mu\nu}$ are non-zero for strings and domain walls, respectively. The corresponding tensor currents are the _brane currents_ $C_{\mu\nu},\tilde{C}_{\mu\nu\rho}$, that are topologically conserved. See [61, 62, 60] for more details. 3. 3. For a superconformal theory, the $\mathcal{S}$-multiplet decomposes into the smaller supercurrent $\mathcal{J}_{\mu}\to\left(j_{\mu}^{(R)},(S_{\mu})_{a},T_{\mu\nu}\right)$ (4.143) where $j_{\mu}^{(R)}$ is a conserved superconformal $U(1)_{R}$-current. Both the FZ multiplet and the $\mathcal{R}$-multiplet contain 12+12 real degrees of freedom out of the initial 16+16 of the general $\mathcal{S}$-multiplet, while the superconformal multiplet is reduced to 8+8 real degrees of freedom. The FZ multiplet can be coupled to the so-called “old minimal supergravity multiplet” [63]: $\mathcal{H}_{\mu}\to\left(b_{\mu},(\Psi_{\mu})_{a},M,h_{\mu\nu}\right)$ (4.144) where $b_{\mu}$ is a genuine 1-form field (i.e. non gauge), $M$ is a complex scalar, $(\Psi_{\mu})_{a}$ is the gravitino and $h_{\mu\nu}$ is the graviton. The variation of the gravitino in this case is given by [54] $\delta_{\epsilon}\Psi_{\mu}=-2\nabla_{\mu}\epsilon+\frac{i}{3}\left(M\gamma_{\mu}+2b_{\mu}+2b^{\nu}\gamma_{\mu\nu}\right)\epsilon$ (4.145) that implies a generalized Killing spinor equation of the form $\nabla_{\mu}\epsilon=\frac{i}{6}\left(M\gamma_{\mu}+2b_{\mu}+2b^{\nu}\gamma_{\mu\nu}\right)\epsilon$ (4.146) in the Majorana spinor $\epsilon$, given a background multiplet. In theories with an R-symmetry, one can couple the $\mathcal{R}$-multiplet to the “new minimal supergravity multiplet”[64]: $\mathcal{H}^{(new)}_{\mu}\to\left(A^{(R)}_{\mu},(\Psi_{\mu})_{a},h_{\mu\nu},B_{\mu\nu}\right)$ (4.147) where $A^{(R)}_{\mu}$ is the Abelian gauge field associated to the $U(1)_{R}$ symmetry, and $B_{\mu\nu}$ is a 2-form gauge field that is often treated through its Hodge dual $V^{\mu}:=i(\star B)^{\mu}=(i/2)\varepsilon^{\mu\nu\rho\sigma}\partial_{\nu}B_{\rho\sigma}$. The variation of the gravitino in this case gives rise to the following Killing spinor equation, that in 2-component notation is [54] $\displaystyle\left(\nabla_{\mu}-iA^{(R)}_{\mu}\right)\epsilon_{a}=-iV_{\mu}\epsilon_{a}-iV^{\nu}(\gamma_{\mu\nu}\epsilon)_{a}$ (4.148) $\displaystyle\left(\nabla_{\mu}+iA^{(R)}_{\mu}\right)\tilde{\epsilon}_{\dot{a}}=iV_{\mu}\tilde{\epsilon}_{\dot{a}}+iV^{\nu}(\gamma_{\mu\nu}\tilde{\epsilon})_{\dot{a}}$ where in the parenthesis on the LHS there is a gauge-covariant derivative with respect to the $U(1)_{R}$ symmetry, that acts with opposite charge on the two chiral sector of the spin representation, see (4.90). The $\mathcal{N}=2$ case in 3 Euclidean dimensions can be derived in superspace by dimensional reduction from the four dimensional case: the supercurrent is reduced to a three dimensional $\mathcal{S}$-multiplet with 12+12 real DoF, plus a real scalar superfield $\hat{\mathcal{J}}=\mathcal{S}_{0}\equiv\mathcal{S}_{a\dot{a}}(\sigma_{0})^{a\dot{a}}$, that contains 4+4 real DoF. Again, there are special cases analogue of those above: a FZ multiplet, an $\mathcal{R}$-multiplet, and a superconformal multiplet. For example, the $\mathcal{N}=2$ $\mathcal{R}$-multiplet in 3 dimensions has the field content $\mathcal{R}_{\mu}\to\left(j_{\mu}^{(R)},j_{\mu}^{(Z)},J,(S_{\mu})_{a},(\tilde{S}_{\mu})_{a},T_{\mu\nu}\right)$ (4.149) where $j_{\mu}^{(R)}$ is the conserved R-current, $j_{\mu}^{(Z)}$ is the conserved central charge current and $J$ is a scalar operator, that with the conserved supercurrents and enery-momentum tensor sum up to 8+8 real DoF. This multiplet couples to the tree dimensional $\mathcal{N}=2$ new minimal supergravity multiplet $\mathcal{H}^{(new)}_{\mu}\to\left(A^{(R)}_{\mu},C_{\mu},H,(\psi_{\mu})_{a},(\tilde{\psi}_{\mu})_{a},h_{\mu\nu}\right)$ (4.150) with the graviton, two gravitini, two gauge 1-forms $A^{(R)}$ and $C$, and a scalar $H$. The 1-form $C$ is often treated in terms of the vector field $V^{\mu}:=i(\star dC)^{\mu}=i\varepsilon^{\mu\nu\rho}\partial_{\nu}C_{\rho}$ that is Hodge dual to its field strength. Putting to zero the gravitini and their variations leads to the generalized Killing spinor equations [56] $\displaystyle\left(\nabla_{\mu}-iA^{(R)}_{\mu}\right)\epsilon=-\left(\frac{1}{2}H\gamma_{\mu}+iV_{\mu}+\frac{1}{2}\varepsilon_{\mu\nu\rho}V^{\nu}\gamma^{\rho}\right)\epsilon$ (4.151) $\displaystyle\left(\nabla_{\mu}+iA^{(R)}_{\mu}\right)\tilde{\epsilon}=-\left(\frac{1}{2}H\gamma_{\mu}-iV_{\mu}-\frac{1}{2}\varepsilon_{\mu\nu\rho}V^{\nu}\gamma^{\rho}\right)\tilde{\epsilon}$ where in this case the two spinors $\epsilon,\tilde{\epsilon}$ have to be treated as independent. Notice that both equations (4.148) and (4.151) are linear in the 4 spinor components, so their solutions (if exist) span a vector space of dimension less or equal than 4. #### 4.4.3 $\mathcal{N}=2$ gauge theories on the round 3-sphere It was shown that, in general, solutions of the Killing condition (4.148) in four dimensions exist if $(\mathcal{M},g)$ is an Hermitian manifold, i.e. $\mathcal{M}$ has an integrable complex structure and $g$ is a compatible Hermitian metric. Analogously, the existence of a solution to (4.151) in three dimensions was shown to be equivalent to the manifold admitting a _transversally holomorphic fibration_.343434This is an odd-dimensional analogue to a complex structure. It means, roughly speaking, that $\mathcal{M}$ is locally isomorphic to $\mathbb{R}\times\mathbb{C}$, and its transition functions are holomorphic in the $\mathbb{C}$-sector. If one is interested in the case of maximal number of Killing spinor solutions, a suitable integrability condition (see [56]) gives $\displaystyle H=\text{const},\qquad d(A^{(R)}-V)=0,\qquad g(V,V)=\text{const},$ (4.152) $\displaystyle(\nabla_{\mu}V)_{\nu}=-iH\varepsilon_{\mu\nu\rho}V^{\rho},$ $\displaystyle R_{\mu\nu}=-V_{\mu}V_{\nu}+g_{\mu\nu}(g(V,V)+2H^{2}).$ In particular, if we take $A^{(R)}=V=0$, then $\mathcal{M}$ is of Einstein type and so it has constant sectional curvature. $H^{2}$ is then interpreted as a cosmological constant, and $\mathcal{M}$ can be $\mathbb{S}^{3},\mathbb{T}^{3}$ or $\mathbb{H}^{3}$ if $H$ is purely immaginary, zero or real. All of them are examples of maximally supersymmetric backgrounds in $\mathcal{N}=2$, thus we have 2 solutions for $\epsilon$ and 2 solutions for $\tilde{\epsilon}$ to the Killing equations $\nabla_{\mu}\epsilon=-\frac{H}{2}\gamma_{\mu}\epsilon\ ;\qquad\nabla_{\mu}\tilde{\epsilon}=-\frac{H}{2}\gamma_{\mu}\tilde{\epsilon}.$ (4.153) In particular, if we take $H=-(i/l)$, this solutions are consistent with the $\mathbb{S}^{3}$ round metric $g=l^{2}\left(d\varphi_{1}\otimes d\varphi_{1}+\sin^{2}\varphi_{1}\ d\varphi_{2}\otimes d\varphi_{2}+\sin^{2}\varphi_{1}\sin^{2}\varphi_{2}\ d\varphi_{3}\otimes d\varphi_{3}\right).$ (4.154) The action of the supersymmetry algebra on the curved manifold can be derived by taking the rigid limit of the appropriate algebra of supergravity transformations. In the 3-dimensional case, it can be derived by a “twisted” reduction of the $\mathcal{N}=1$ supergravity in 4 dimensions. The 4-dimensional supersymmetry algebra realizes on the curved manifold as $[\delta_{\epsilon},\delta_{\epsilon}]\phi_{(r)}=[\epsilon,\epsilon]\cdot\phi_{(r)}=2(\overline{\epsilon}\gamma^{\mu}\epsilon)P_{\mu}\cdot\phi_{(r)}$ (4.155) where $\epsilon$ is a Majorana Killing spinor, $\phi_{(r)}$ is a generic field of R-charge $r$, and the local action of the momentum operator is through the fully covariant derivative $P_{\mu}\to-\left(i\nabla_{\mu}+rA^{(R)}_{\mu}\right),$ (4.156) so that (4.155) can be written as $[\delta_{\epsilon},\delta_{\epsilon}]\phi_{(r)}=-2i\left(\mathcal{L}_{v}-irA^{(R)}(v)\right)\phi_{(r)}$ (4.157) where $v^{\mu}:=(\overline{\epsilon}\gamma^{\mu}\epsilon)$ is a Killing vector field thanks to $\epsilon$ being a Killing spinor field. This is reduced to the 3-dimensional case, taking $\epsilon,\tilde{\epsilon}$ now as independent 2-component Killing spinors and $v^{\mu}:=\tilde{\epsilon}\gamma^{\mu}\epsilon$ in 3 dimensions, as (see again [56]) $\displaystyle[\delta_{\tilde{\epsilon}},\delta_{\epsilon}]\phi_{(r,z)}=-2i\left[\mathcal{L}_{v}-iv^{\mu}\left(r(A^{(R)}_{\mu}-\frac{1}{2}V_{\mu})+zC_{\mu}\right)+\tilde{\epsilon}\epsilon(z-rH)\right]\phi_{(r,z)},$ (4.158) $\displaystyle[\delta_{\tilde{\epsilon}},\delta_{\tilde{\epsilon}}]\phi_{(r,z)}=0,\qquad[\delta_{\epsilon},\delta_{\epsilon}]\phi_{(r,z)}=0,$ where $z$ is the charge associated to the action of the central charge $Z$ in (4.101). For the 3-sphere of radius $l=1$, this is simplified to $\displaystyle[\delta_{\tilde{\epsilon}},\delta_{\epsilon}]\phi_{(r,z)}=-2i\left[\mathcal{L}_{v}+\tilde{\epsilon}\epsilon(z+ir)\right]\phi_{(r,z)},$ (4.159) $\displaystyle[\delta_{\tilde{\epsilon}},\delta_{\tilde{\epsilon}}]\phi_{(r,z)}=0,\qquad[\delta_{\epsilon},\delta_{\epsilon}]\phi_{(r,z)}=0.$ We report the resulting supersymmetry variation for the 3-dimensional $\mathcal{N}=2$ vector multiplet $(A_{\mu},\sigma$, $\lambda_{a},\tilde{\lambda}_{a},D)$, with respect to two Killing spinors $\tilde{\epsilon},\epsilon$. This multiplet is uncharged under the action of R-symmetry and of the central charge $Z$. Following conventions of [50] and [49], $\displaystyle\delta A_{\mu}=\frac{i}{2}(\tilde{\epsilon}\gamma_{\mu}\lambda-\tilde{\lambda}\gamma_{\mu}\epsilon)$ (4.160) $\displaystyle\delta\sigma=\frac{1}{2}(\tilde{\epsilon}\lambda-\tilde{\lambda}\epsilon)$ $\displaystyle\delta\lambda=\left(-\frac{1}{2}F_{\mu\nu}\gamma^{\mu\nu}-D+i(D_{\mu}\sigma)\gamma^{\mu}+\frac{2i}{3}\sigma\gamma^{\mu}D_{\mu}\right)\epsilon$ $\displaystyle\delta\tilde{\lambda}=\left(-\frac{1}{2}F_{\mu\nu}\gamma^{\mu\nu}+D-i(D_{\mu}\sigma)\gamma^{\mu}-\frac{2i}{3}\sigma\gamma^{\mu}D_{\mu}\right)\tilde{\epsilon}$ $\displaystyle\delta D=-\frac{i}{2}\left(\tilde{\epsilon}\gamma^{\mu}D_{\mu}\lambda-(D_{\mu}\tilde{\lambda})\gamma^{\mu}\epsilon\right)+\frac{i}{2}\left([\tilde{\epsilon}\lambda,\sigma]-[\tilde{\lambda}\epsilon,\sigma]\right)-\frac{i}{6}\left(\tilde{\lambda}\gamma^{\mu}D_{\mu}\epsilon+(D_{\mu}\tilde{\epsilon})\gamma^{\mu}\lambda\right)$ where now $D_{\mu}=\nabla_{\mu}-iA_{\mu}$ is the gauge-covariant derivative. On the 3-sphere, the actions in (4.106) and (4.107) acquire a factor $\sqrt{g}$ in the measure,353535The pure Chern-Simons term $\left(A\wedge dA+\frac{2i}{3}A^{3}\right)$ is actually unmodified, being already a 3-form. and the Super Yang-Mills Lagrangian gets modified to $\mathcal{L}_{YM}=\mathrm{Tr}\left\\{\frac{i}{2}\tilde{\lambda}\gamma^{\mu}D_{\mu}\lambda+\frac{1}{4}F_{\mu\nu}F^{\mu\nu}+\frac{1}{2}D_{\mu}\sigma D^{\mu}\sigma+\frac{i}{2}\tilde{\lambda}[\sigma,\lambda]+\frac{1}{2}\left(D+\frac{\sigma}{l}\right)^{2}-\frac{1}{4l}\tilde{\lambda}\lambda\right\\}\\\ $ (4.161) where we reinstated the radius $l$, to see that indeed in the limit $l\to\infty$ this becomes the standard Euclidean SYM theory in 3 dimensions. We note an important feature of this Lagrangian, that will be important for the application of the localization principle: this can be written as a supersymmetry variation, i.e. $\tilde{\epsilon}\epsilon\mathcal{L}_{YM}=\delta_{\tilde{\epsilon}}\delta_{\epsilon}\mathrm{Tr}\left\\{\frac{1}{2}\tilde{\lambda}\lambda-2D\sigma\right\\}.$ (4.162) The SCS Lagrangian does not get modified on curved space, since the term depending on the gauge field is topological, and the other ones do not contain derivatives. It is important to remark that, in general, unbroken supersymmetry is consistent only with Anti-de Sitter geometry (or, in Euclidean signature, hyperbolic geometry) [54]. An exception to this is given by those theories that possess a larger group of symmetries, the _superconformal_ group. This is an extension of the super-Poincaré group, to include also conformal transformations of the metric. In this case, supersymmetry can be consistent also on conformally flat backgrounds with positive scalar curvature, of which the $n$-spheres are an example. The $\mathcal{N}=2$ SCS theory of above is an example of superconformal theory. #### 4.4.4 $\mathcal{N}=4,2,2^{*}$ gauge theories on the round 4-sphere We continue also the example of the $\mathcal{N}=4$ 4-dimensional theory, understanding how it can be realized on a different background compatible with $\mathbb{S}^{4}$, and what part of the supersymmetry algebra can be preserved on this background. As in Section 4.3.7, the $\mathcal{N}=2$ and $\mathcal{N}=2^{*}$ cases follow from modifications of the $\mathcal{N}=4$ theory. Using stereographic coordinates $x^{1},\cdots,x^{4}$, such that the North pole is located at $x^{\mu}=0$, the round metric of the 4-sphere of radius $r$ looks explicitly as a conformal transformation of the flat Euclidean metric, $g_{\mu\nu}^{(x)}=\delta_{\mu\nu}e^{2\Omega(x)},\quad\mathrm{where}\ e^{2\Omega(x)}=\frac{1}{\left(1+\frac{x^{2}}{4r^{2}}\right)^{2}}$ (4.163) where $x^{2}=\sum_{\mu=1}^{4}(x^{\mu})^{2}$. As remarked at the end of the last section, the conformal flatness of $\mathbb{S}^{4}$ allows us to deform the superconformal YM theory on it, provided we preserve the conformal symmetry. In order to do this, we modify the kinetic term of the scalars $(\Phi_{A})_{A=5,\cdots,9,0}$ adding a conformal coupling to the curvature: $(\partial\Phi_{A})^{2}\to(\partial\Phi_{A})^{2}+\frac{R}{6}(\Phi_{A})^{2}$ where $R=\frac{12}{r^{2}}$ is the scalar curvature of the metric $g$.363636In $d$-dimensions, the conformal coupling to the curvature scalar is made adding a term $\xi R(\Phi)^{2}$ for the scalar field of canonical mass dimension $\frac{d-2}{2}$, with $\xi=(d-2)/4(d-1)$ (see [65], Appendix D). The scalar curvature of the $d$-sphere of radius $r$ is $R=d(d-1)/r^{2}$. This ensures conformal invariance of the action on the 4-sphere, $S^{\mathcal{N}=4}_{\mathbb{S}^{4}}=\int_{\mathbb{S}^{4}}d^{4}x\sqrt{g}\ \frac{1}{g_{YM}^{2}}\mathrm{Tr}\left(\frac{1}{2}F_{MN}F^{MN}-\Psi\Gamma^{M}D_{M}\Psi+\frac{2}{r^{2}}\Phi_{A}\Phi^{A}\right)$ (4.164) where the derivatives have been promoted to covariant derivatives with respect to the Levi-Civita connection of $g$. Now we have to understand which supersymmetries of the $\mathcal{N}=4$ algebra can be preserved on the new curved background. From theorem 4.4.1, we know that a necessary condition for a section $\epsilon$ of the Majorana-Weyl spinor bundle on $\mathbb{S}^{4}$, to produce a superisometry for the new background, is that it satisfies the _twistor spinor equation_ , or _conformal Killing equation_ $\nabla_{\mu}\epsilon=\tilde{\Gamma}_{\mu}\tilde{\epsilon}$ (4.165) for some other section $\tilde{\epsilon}$. It can be checked that, to ensure supersymmetry of (4.164), $\tilde{\epsilon}$ must be also a twistor spinor satisfying $\nabla_{\mu}\tilde{\epsilon}=-\frac{1}{4r^{2}}\Gamma_{\mu}\epsilon,$ (4.166) and the variations (4.109) have to be modified as the superconformal transformations $\displaystyle\delta_{\epsilon}A_{M}$ $\displaystyle=\epsilon\Gamma_{M}\Psi$ (4.167) $\displaystyle\delta_{\epsilon}\Psi$ $\displaystyle=\frac{1}{2}\Gamma^{MN}F_{MN}\epsilon+\frac{1}{2}\Gamma^{\mu A}\Psi_{A}\nabla_{\mu}\epsilon.$ Since $\mathbb{S}^{4}$ is conformally flat, the number of solutions to (4.165) is maximal and equal to $2\dim{(S^{\pm})}=32$ [53], so the whole $\mathcal{N}=4$ superconformal algebra is preserved.373737The number of generators of the $\mathcal{N}=4$ super-Euclidean algebra is $\dim{(S^{\pm})}=16$. The other 16 are the remaining generators of the superconformal algebra. If one restricts the attention to the $\mathcal{N}=2$ subalgebra, then half of the generators are preserved. If instead the $\mathcal{N}=2^{*}$ theory is considered, the conformal symmetry is broken and only 8 supercharges are preserved on $\mathbb{S}^{4}$. With the above modifications, the $\mathcal{N}=4$ superconformal algebra closes again only on-shell: imposing the EoM for $\Psi$, one gets (see Appendix of [11] for the details of the computation) $\delta_{\epsilon}^{2}=-\mathcal{L}_{v}-G_{\Phi}-R-\Omega$ (4.168) as (4.117). In order to prepare the ground for the exploitation of the localization principle on supersymmetric gauge theories, we remark that, if we want to define correctly an equivariant structure with respect to (at least a $U(1)$ subgroup of) the Poincaré group, we need at least an $\mathcal{N}=1$ supersymmetry subalgebra to close properly (i.e. off-shell). If this is the case, we can use the corresponding variation $\delta_{\epsilon}$ as a Cartan differential with respect to this equivariant cohomology (we are going to justify better this in the next chapter). It is not possible to close off- shell the full $\mathcal{N}=2$ algebra on the hypermultiplet, but fixing a conformal Killing spinor $\epsilon$ satisfying (4.165) and (4.166), it is possible to close the subalgebra generated by $\delta_{\epsilon}$ only. To do this, one has to add auxiliary fields to match the number of off-shell bosonic/fermionic degrees of freedom of the theory [66], analogously to what happens for example to the $\mathcal{N}=1$ vector multiplet in 4-dimensions. In 10-dimensions, we have 16 real fermionic components, and $(10-1)$ real physical bosonic components, so we have to add 7 bosonic (scalar) fields $(K_{i})_{i=1,\cdots,7}$. The modified action $S^{\mathcal{N}=4}_{\mathbb{S}^{4}}=\int_{\mathbb{S}^{4}}d^{4}x\sqrt{g}\ \frac{1}{g_{YM}^{2}}\mathrm{Tr}\left(F_{MN}F^{MN}-\Psi\Gamma^{M}D_{M}\Psi+\frac{2}{r^{2}}\Phi_{A}\Phi^{A}-\sum_{i=1}^{7}K_{i}K_{i}\right)$ (4.169) is supersymmetric under the modified $\mathcal{N}=4$ superconformal transformations $\displaystyle\delta_{\epsilon}A_{M}$ $\displaystyle=\epsilon\Gamma_{M}\Psi$ (4.170) $\displaystyle\delta_{\epsilon}\Psi$ $\displaystyle=\frac{1}{2}\Gamma^{MN}F_{MN}\epsilon+\frac{1}{2}\Gamma^{\mu A}\Psi_{A}\nabla_{\mu}\epsilon+\sum_{i=1}^{7}K_{i}\nu_{i}$ $\displaystyle\delta_{\epsilon}K_{i}$ $\displaystyle=-\nu_{i}\Gamma^{M}D_{M}\Psi.$ Here $\epsilon$ is a fixed conformal Killing spinor, and $(\nu_{i})$ are seven spinors satisfying $\displaystyle\epsilon\Gamma^{M}\nu_{i}=0$ (4.171) $\displaystyle(\epsilon\Gamma_{M}\epsilon)\tilde{\Gamma}^{M}_{ab}=2\left(\sum_{i}(\nu_{i})_{a}(\nu_{i})_{b}+\epsilon_{a}\epsilon_{b}\right)$ $\displaystyle\nu_{i}\Gamma^{M}\nu_{j}=\delta_{ij}\epsilon\Gamma^{M}\epsilon.$ To ensure convergence of the path integral, as we did with the scalar field $\Phi_{0}$, we path integrate the new auxiliary scalars on purely immaginary values, i.e. $K_{j}=:iK_{j}^{E}$ with $K_{j}^{E}$ real. For every fixed non- zero $\epsilon$, there exist seven linearly independent $\nu_{i}$ satisfying these constraints, up to an $SO(7)$ internal rotation, ensuring the closure (4.117) off-shell. Although, if we want $\delta_{\epsilon}$ to describe the equivariant cohomology of a subgroup of the Poincaré group (not the conformal one), we should restrict to those $\epsilon$ that generates only translations and R-symmetries at most (up to unphysical gauge transformations). Thus the dilatation term in (4.117) must vanish, imposing the condition $(\epsilon\tilde{\epsilon})=0$ on the conformal Killing spinors. We describe now which modifications to the above discussion have to be made in order to describe the $\mathcal{N}=2$ and $\mathcal{N}=2^{*}$ theories. The pure $\mathcal{N}=2$ is classically obtained by restricting to the $\mathcal{N}=2$ supersymmetry algebra generated by (4.118) and putting all the fields of the hypermultiplet to zero. At quantum level, this theory breaks in general the conformal invariance, so it is equivalent to consider the $\mathcal{N}=2^{*}$ with hypermultiplet masses introduced as at the end of Section 4.3.7, by $D_{0}\mapsto D_{0}+M$ where $M$ is an $SU(2)_{R}^{\mathcal{R}}$ mass matrix. The mass terms for the fermions break the $SO(1,1)^{\mathcal{R}}$ R-symmetry, so we must restrict the superconformal algebra further to those $\epsilon$ for which the corresponding piece of the R-symmetry in (4.117) vanish. This imposes $(\tilde{\epsilon}\Gamma^{09}\epsilon)=0$. Moreover, this deformed theory is not invariant under the $\mathcal{N}=2$ supersymmetry, because of the non- triviality of the conformal Killing spinor. In fact, using the conformal Killing equations it results that $\delta_{\epsilon}\left(\frac{1}{2}F_{MN}F^{MN}-\Psi\Gamma^{M}D_{M}\Psi+\frac{2}{r^{2}}\Phi_{A}\Phi^{A}\right)=-4\Psi\Gamma^{i}\tilde{\Gamma}^{0}\tilde{\epsilon}M_{i}^{j}\Phi_{j}$ (4.172) where $i,j=5,\cdots,8$, up to a total derivative. If $\epsilon,\tilde{\epsilon}$ are restricted to the the $+1$ eigenspace of $\Gamma^{5678}$, we write $\tilde{\epsilon}=\frac{1}{2r}\Lambda\epsilon$, where $\Lambda$ is a generator of $SU(2)_{L}^{\mathcal{R}}$. Explicitly $\Lambda=\frac{1}{4}\Gamma^{ij}R_{ij}$, with components $(R_{ij})$ normalized such that $R_{ij}R^{ij}=4$. Then, after some gamma matrix technology, (4.172) gives $\delta_{\epsilon}(\cdots)=\frac{1}{2r}(\Psi\Gamma_{i}\epsilon)R^{ik}M_{k}^{j}\Phi_{j}=\frac{1}{2r}(\delta_{\epsilon}\Phi_{i})R^{ik}M_{k}^{j}\Phi_{j}.$ (4.173) Hence, we can modify further the mass-deformed action to get invariance with respect to this subalgebra of the original superconformal algebra on $\mathbb{S}^{4}$, adding the new term $-\frac{1}{4r}(R^{ki}M_{k}^{j})\Phi_{i}\Phi_{j}.$ (4.174) Finally, the action $\displaystyle S^{\mathcal{N}=2^{*}}_{\mathbb{S}^{4}}=\int_{\mathbb{S}^{4}}d^{4}x\sqrt{g}\ \frac{1}{g_{YM}^{2}}\mathrm{Tr}\Biggl{(}F_{MN}F^{MN}-\Psi\Gamma^{M}D_{M}\Psi$ $\displaystyle+\frac{2}{r^{2}}\Phi_{A}\Phi^{A}-$ (4.175) $\displaystyle-\frac{1}{4r}(R^{ki}M_{k}^{j})\Phi_{i}\Phi_{j}-\sum_{i=1}^{7}K_{i}K_{i}\Biggr{)}$ where $D_{0}\Phi_{i}\mapsto[\Phi_{0},\Phi_{i}]+M_{i}^{j}\Phi_{j}$ and $D_{0}\Psi\mapsto[\Phi_{0},\Psi]+\frac{1}{4}M_{ij}\Gamma^{ij}\Psi$, is invariant under the subalgebra generated by a fixed conformal Killing spinor satisfying the conditions $\Gamma^{5678}\epsilon=\epsilon,\qquad\nabla_{\mu}\epsilon=\frac{1}{8r}\Gamma_{\mu}\Gamma^{ij}R_{ij}\epsilon.$ (4.176) #### 4.4.5 Trial and error method Another method that was extensively used in the physics literature to promote a supersymmetric theory on curved spaces is based on a trial and error procedure [67]. Suppose to have a supersymmetric QFT formulated in terms of component fields on flat Minkowski (or Euclidean) space $\mathbb{R}^{d}$, specified by the Lagrangian density $\mathcal{L}^{(0)}$ invariant under the supersymmetry variation $\delta^{(0)}$. The starting point of this approach is to simply “covariantize” the original theory, replacing the flat metric $\eta$ to the desired metric $g$ defined on $\mathcal{M}$ and every derivative $\partial_{\mu}$ with the appropriate Levi-Civita or spin covariant derivative $\nabla_{\mu}$ corresponding to $g$. The problem is that in general this does define the theory on the curved space, but it is not guaranteed that the supersymmetry survives: $\left[\delta^{(0)}\mathcal{L}^{(0)}\right]_{(\eta,d)\to(g,\nabla)}\neq\nabla_{\mu}(\cdots)^{\mu}.$ (4.177) The idea then is to correct the action of the supersymmetry and the Lagrangian with an expansion in powers of $1/r$, where $r$ is a characteristic length of $\mathcal{M}$,383838Being $\mathcal{M}$ compact, we can take it as an embedding in $\mathbb{R}^{n}$ for some $n$, and scale the metric according to some characteristic length $r$. $\displaystyle\delta$ $\displaystyle=\delta^{(0)}+\sum_{i\geq 1}\frac{1}{r^{i}}\delta^{(i)}$ (4.178) $\displaystyle\mathcal{L}$ $\displaystyle=\mathcal{L}^{(0)}+\sum_{i\geq 1}\frac{1}{r^{i}}\mathcal{L}^{(i)}$ requiring order by order the symmetry of the Lagrangian _and_ the closure of the super-algebra. This “trial and error” terminates if one is able to ensure both conditions at some finite order in $1/r$, even though a priori the series contains an infinite number of terms. This procedure has the quality to be simple and operational in principle, but can be very cumbersome in practice to apply. ### 4.5 BRST cohomology and equivariant cohomology In gauge theories, BRST cohomology is a useful device to provide an algebraic description of the path integral quantization procedure, and the renormalizability of non-Abelian Yang-Mills theory in 4 dimensions. This formalism makes use of Lie algebra cohomology, while the BRST model of Section 2.5 describes equivariant cohomology, that is what we use in topological or supersymmetric field theories. It is natural to ask whether there is a relation between these two cohomology theories, and in fact there is. It turns out that equivariant cohomology of a Lie algebra $\mathfrak{g}$ is the same as a “supersymmetrized” Lie algebra cohomology of a corresponding graded Lie algebra $\mathfrak{g[\epsilon]}:=\mathfrak{g}\otimes\bigwedge\epsilon$ [20]. Let us first see how the Weil model $W(\mathfrak{g})=S(\mathfrak{g}^{*})\otimes\bigwedge(\mathfrak{g}^{*})$ (4.179) for the equivariant cohomology of $\mathfrak{g}$ can be seen in more supergeometric terms. Notice that the space $S(\mathfrak{g}^{*})$ may be identified with the (commutative) algebra of functions on the Lie algebra $\mathfrak{g}$, and thus us we can see the Weil algebra $W(\mathfrak{g}^{*})$ as the space of functions on a supermanifold built from the tangent bundle of $\mathfrak{g}$, that is exactly the odd tangent bundle $\Pi T\mathfrak{g}\equiv\Pi\mathfrak{g}$.393939Notice that since $\mathfrak{g}$ is a vector space, $T^{*}\mathfrak{g}\cong\mathfrak{g}^{*}$. Denoting $\\{\tilde{c}^{i}\\}$ and $\\{c^{i}\\}$ the generators of $S(\mathfrak{g}^{*})$ and $\bigwedge(\mathfrak{g}^{*})$ respectively, indeed a function on this superspace is trivialized as $\Phi=\Phi^{(0)}(\tilde{c})+\Phi_{j}^{(1)}(\tilde{c})c^{j}+\Phi_{jk}^{(2)}(\tilde{c})c^{j}c^{k}+\cdots$ (4.180) Introducing generators $\\{\tilde{b}_{i}\\}$ and $\\{b_{i}\\}$ of $\mathfrak{g}[1]$ and $\mathfrak{g}$, such that404040Sometimes the action of this generators is denoted as a graded bracket structure, like $[b_{i},c^{j}]=[\tilde{b}_{i},\tilde{c}^{j}]_{+}=\delta^{j}_{i}$. $b_{i}(c^{j}):=c^{j}(b_{i})=\delta^{j}_{i},\qquad\tilde{b}_{i}(\tilde{c}^{j}):=c^{j}(b_{i})=\delta^{j}_{i},$ (4.181) the Weil differential (2.29) can be written as $d_{W}=\tilde{c}^{i}b_{i}+f^{i}_{jk}c^{j}\tilde{c}^{k}b_{i}-\frac{1}{2}f^{i}_{jk}c^{j}c^{k}\tilde{b}_{i},$ (4.182) that is very reminiscent of the form of a “BRST operator”. In Lie algebra cohomology, the Chevalley-Eilenberg differential on $\bigwedge(\mathfrak{g}^{*})$ is defined on 1-forms $\alpha\in\mathfrak{g}^{*}$ as $\delta\alpha=-\alpha\left([\cdot,\cdot]\right)=\alpha_{i}\left(-\frac{1}{2}f^{i}_{jk}c^{j}c^{k}\right),$ (4.183) and then extended as an antiderivation on the whole complex. If we consider a $\mathfrak{g}$-module $V$, such as the target space of a given field theory of gauge group $G$, with a representation $\rho:\mathfrak{g}\to\mathrm{End}(V)$,414141If $V$ is a field space $C^{\infty}(\mathcal{M})$ over some (super)manifold $\mathcal{M}$, $\mathfrak{g}$ acts as usual as a Lie derivative with respect to the fundamental vector field, $\rho(X)=\mathcal{L}_{X}$. then the CE differential is extended to $\bigwedge(\mathfrak{g}^{*})\otimes V$ as $\displaystyle\delta v(X):=\rho(X)v\qquad\forall v\in V,X\in\mathfrak{g}$ (4.184) $\displaystyle\delta(\alpha\otimes v)=\delta\alpha\otimes v+(-1)^{k}\alpha\otimes\delta v\qquad\forall v\in V,\alpha\in\bigwedge\nolimits^{\\!k}(\mathfrak{g}^{*}).$ This, expressed with respect to a basis $\\{b_{i}\\}$ of $\mathfrak{g}$, coincide with the BRST operator $\delta=c^{i}\rho(b_{i})-\frac{1}{2}f^{i}_{jk}c^{j}c^{k}b_{i}$ (4.185) that satisfies $\delta^{2}=0$. The $c^{i}$ are _ghosts_ , while the $b_{i}$ are _anti-ghosts_. The zero-th cohomology group of the complex $\bigwedge(\mathfrak{g}^{*})\otimes V$ with respect to the differential (4.185) contains those states that have _ghost number_ 0 and are $\mathfrak{g}$-invariant, $H^{0}(\mathfrak{g},V)\cong V^{\mathfrak{g}}$ (4.186) so the interesting “physical” states. We see that there is a difference between the BRST operator (4.185) and the Weil differential (4.182), but we can connect these differentials as follows. To the Lie algebra $\mathfrak{g}$ we can associate the _differential graded Lie algebra_ $\mathfrak{g}[\epsilon]:=\mathfrak{g}\otimes\bigwedge\epsilon$. Here $\epsilon$ is a single generator taken in odd degree, $\mathrm{deg}(\epsilon):=-1$, while $\mathrm{deg}(\mathfrak{g}):=0$. A differential $\partial:\mathfrak{g}[\epsilon]\to\mathfrak{g}[\epsilon]$ is defined as $\partial\epsilon:=1\in\mathfrak{g},\qquad\qquad\partial X:=0\quad\forall X\in\mathfrak{g}.$ (4.187) This superalgebra has generators $b_{i}:=b_{i}\otimes 1$ and $\tilde{b}_{i}:=b_{i}\otimes\epsilon$, and the superbracket structure coming from the Lie brackets on $\mathfrak{g}$ and the (trivial) wedge product on $\bigwedge\epsilon$: $\displaystyle[b_{i},b_{j}]=f^{k}_{ij}b_{k}$ (4.188) $\displaystyle[b_{i},\tilde{b}_{j}]=f^{k}_{ij}\tilde{b}_{k}$ $\displaystyle[\tilde{b}_{i},\tilde{b}_{j}]=0$ making it into a Lie superalgebra. The differential on the generators is rewritten as $\partial b_{i}=0,\qquad\quad\partial\tilde{b}_{i}=b_{i}.$ (4.189) To this “supersymmetrized” algebra we can associate the Lie algebra cohomology with respect to the complex $\bigwedge(\mathfrak{g}[\epsilon]^{*})$, that is generated by $\\{c^{i},\tilde{c}^{i}\\}$ of degrees $\mathrm{deg}(c^{i})=1,\mathrm{deg}(\tilde{c}^{i})=2$, such that $c^{i}(b_{j})=\tilde{c}^{i}(\tilde{b}_{j})=\delta^{i}_{j}.$ (4.190) The BRST differential for the $\mathfrak{g}[\epsilon]$-Lie algebra cohomology is naturally defined analogously to before as $Q\alpha:=-\alpha\left([\cdot,\cdot]\right)$ (4.191) on 1-forms $\alpha\in\mathfrak{g}[\epsilon]^{*}$. But now, because of the Lie algebra extension (4.188), its expression in terms of the generators results $Q=-\frac{1}{2}f^{k}_{ij}c^{i}c^{j}b_{k}+f^{k}_{ij}c^{i}\tilde{c}^{j}\tilde{b}_{k}.$ (4.192) Moreover, the dual $\partial^{*}$ acts as $\partial^{*}=\tilde{c}^{i}b_{i}$ (4.193) with $b_{i}$ acting as in (4.181). Then the total differential on this complex coincides with the Weil differential, $d_{W}\cong\partial^{*}+Q$ (4.194) and we identify an isomorphism of dg algebras $\left(\bigwedge(\mathfrak{g}[\epsilon]^{*}),\partial^{*}+Q\right)\cong\left(W(\mathfrak{g}),d_{W}\right).$ (4.195) If we bring into the game the $\mathfrak{g}$-module $V$ as before, we can work with $\Omega(V)$ as a $\mathfrak{g}[\epsilon]$-dg algebra, defining the $\mathfrak{g}[\epsilon]$ action as424242Again, if $V=C^{\infty}(\mathcal{M})$, then $\Omega(\mathcal{M})$ is the space we considered when we constructed the equivariant cohomology of a $G$-manifold. $(X\otimes 1)\to\mathcal{L}_{X},\qquad(X\otimes\epsilon)\to\iota_{X}.$ (4.196) On the complex $\bigwedge(\mathfrak{g}[\epsilon]^{*})\otimes\Omega(V)$, the total differential inherited from the Weil differential and the $\mathfrak{g}[\epsilon]$-action is $d_{B}:=c^{k}\otimes\mathcal{L}_{k}+\tilde{c}^{k}\otimes\iota_{k}+Q\otimes 1+\partial^{*}\otimes 1+1\otimes d$ (4.197) and it coincides with the one of the BRST model of equivariant cohomology (2.51)! This demonstrates how the BRST quantization formalism and equivariant cohomology are intimately related, and suggests that indeed BRST symmetry operators are good candidates to represent equivariant differentials in QFT, that can be used to employ the localization principle in these kind of physical systems. ## Chapter 5 Localization for circle actions in supersymmetric QFT In this chapter we describe how the equivariant localization principle can be carried out in the infinite dimensional case of path integrals in QM or QFT. In this setting, the first object of interest is the _partition function_ $Z=\int_{\mathcal{F}}D\phi\ e^{iS[\phi]}$ (5.1) where $\mathcal{F}=\Gamma(M,E)$ is the space of fields, i.e. sections of some fiber bundle $E\to M$ with typical fiber (the _target space_) $V$ over the (Lorentzian) $n$-dimensional spacetime $M$, and $S\in C^{\infty}(\mathcal{F})$ is the action functional.111In the (common) case of a trivial bundle, this is equivalent to considering $\mathcal{F}=C^{\infty}(M,V)$, i.e. $V$-valued functions over $M$. Often $V$ is a vector space, otherwise the theory describes a so-called _non-linear $\sigma$-model_. In supersymmetric field theories, $V=\bigwedge(S^{*})$ for some vector space $S$, and the field space acquires a natural graded structure. The fields are supposed to satisfy some prescribed boundary conditions on $\partial M$. In the Riemannian case, the corresponding object has the form $Z=\int_{\mathcal{F}}D\phi\ e^{-S_{E}[\phi]}$ (5.2) where we denoted $S_{E}$ the Euclidean action. If the spacetime has the form $M=\mathbb{R}_{t}\times\Sigma$, this last expression can be reached from the Lorentzian theory via _Wick rotation_ of the time direction $t\mapsto\tau:=it$. If the $\tau$ direction is compactified to a circle of length $T$, we can interpret the Euclidean path integral as the canonical ensemble partition function describing the original QFT at a finite temperature $1/T$. If needed, we are always free to set the length $T$ of the circle to be very large, and find the zero temperature limit when $T\to\infty$. Given an observable $\mathcal{O}\in C^{\infty}(M)$, its _expectation value_ is given by $\left\langle\mathcal{O}\right\rangle=\frac{1}{Z}\int_{\mathcal{F}}D\phi\ \mathcal{O}[\phi]e^{iS[\phi]}\quad\mathrm{or}\quad\left\langle\mathcal{O}\right\rangle_{E}=\frac{1}{Z}\int_{\mathcal{F}}D\phi\ \mathcal{O}[\phi]e^{-S_{E}[\phi]}.$ (5.3) The path integral measure $D\phi$ on the infinite dimensional space $\mathcal{F}$ is not rigorously defined,222In fact, it does not exist in general. but it is usually introduced as $\int D\phi=\mathcal{N}\prod_{x\in M}\int_{V}d\phi(x)$ (5.4) where $\mathcal{N}$ is some (possibly infinite) multiplicative factor, and at every point $x\in M$ we have a standard integral over the fiber $V$. Notice that the infinite factors $\mathcal{N}$ cancel in ratios in the computations of expectation values, so we can still make sense of such objects and formally manipulate them to obtain physical information. Another convergence issue comes with the prescription of boundary conditions in computing the action $S[\phi]$. If $M$ is non compact, this often requires a specific regularization,333For example, one can first assume that the spacetime just extends up to some large but finite typical lenght $r_{0}$, and then send this value to infinity at the end of the calculations. while taking compact spacetimes ensure better convergence properties. Very few quantum systems have an exactly solvable path integral. When this functional integral method was introduced, the only examples where (5.1) could be directly evaluated were the free particle and the harmonic oscillator.444Later on, ad hoc methods for other particular systems were developed, like the solution of the Hydrogen atom by Duru and Kleinert [68], and others. Both these theories are quadratic in the fields and their derivatives, thus the partition function can be computed using the formal functional analog of the classical Gaussian integration formula $\int_{-\infty}^{\infty}d^{n}x\ e^{-\frac{1}{2}x^{k}M_{kl}x^{l}+A_{k}x^{k}}=\frac{(2\pi)^{n/2}}{\sqrt{\det{M}}}e^{\frac{1}{2}A_{i}(M^{-1})^{ij}A_{j}}$ (5.5) where $M=[M_{ij}]$ is an $n\times n$ non singular matrix. The analogue in field theory has $n\to\infty$ and a functional determinant at the denominator, that must be properly regularized in order to be a meaningful convergent quantity (see any standard QFT book, like [69, 70]). In perturbative QFT, one almost never has to explicitly perform such a functional integration. Suppose that the action has the generic form $S=S_{0}+S_{int}$ (5.6) where $S_{0}$ is the free term containing up to quadratic powers of the fields and their derivatives, and the rest is collected in $S_{int}$. Then the expectation value of an observable $\mathcal{O}$, expressible as a combination of local fields, is computed expanding the exponential of the interacting part in Taylor series, and exploiting _Wick’s theorem_ 555Again, see any standard QFT book. for the vacuum expectation values in the free theory: $\left\langle\mathcal{O}\right\rangle=\sum_{k}\frac{1}{k!}\left\langle(iS_{int})^{k}\mathcal{O}\right\rangle_{0}.$ (5.7) Another perturbative approach, especially useful to compute the effective action in a given theory, is the so-called _background field method_ , where the action $S$ is expanded around a classical “background”element $\overline{\phi}\in\mathcal{F}$, $S[\overline{\phi}+\eta]=S[\overline{\phi}]+\int_{M}d^{n}x\left(\frac{\delta S}{\delta\phi(x)}\right)_{\overline{\phi}}\eta(x)+\frac{1}{2}\int_{M}d^{n}xd^{n}y\left(\frac{\delta^{(2)}S}{\delta\phi(x)\delta\phi(y)}\right)_{\overline{\phi}}\eta(x)\eta(y)+\cdots$ (5.8) If we chose $\overline{\phi}$ to be a solution of the classical equation of motion $\left(\frac{\delta S}{\delta\phi(x)}\right)_{\overline{\phi}}=0$, the first order term disappears from the expansion. If we also neglect the terms of order higher than quadratic in $\eta$, and substitute the resulting expression in (5.1) or (5.2), we get the equivalent of the saddle point approximation, or “one-loop approximation” of the partition function $Z\approx e^{-S[\overline{\phi}]}Z_{1-loop}[\overline{\phi}],$ (5.9) where $\displaystyle Z_{1-loop}[\phi]$ $\displaystyle:=\int_{\mathcal{F}}D\eta\ e^{-\frac{1}{2}\eta\cdot\Delta[\phi]\cdot\eta}\equiv\left[\det{\left(\Delta[\phi]\right)}\right]^{-1/2}$ (5.10) $\displaystyle\Delta[\phi](x,y)$ $\displaystyle:=\left(\frac{\delta^{(2)}S}{\delta\phi(x)\delta\phi(y)}\right)_{\phi},$ and we denoted convolution products over $M$ with $(\cdot)$ for brevity. We are interested in those cases in which such a “semiclassical” approximation of the partition function turns out to give the exact result for the path integral in the full quantum theory. This is possible if some symmetry of the field theory, i.e. acting on the space $\mathcal{F}$, allows us to formally employ the equivariant localization principle and reduce the path integration domain from $\mathcal{F}$ to a (possibly finite-dimensional) subspace. In the next part of the chapter we will see some cases in which it is possible to interpret $\mathcal{F}$, or a suitable extension of it, as a Cartan model with some (super)symmetry operator acting as the Cartan differential. As we already anticipated, this is possible if $\mathcal{F}$ has a graded structure that can both arise from the supersymmetry of the underlying spacetime, or can be introduced via an extension analogous to what happens in the BRST formalism. We will first describe the application of the Duistermaat-Heckman theorem in the case of Hamiltonian QM, where the equivariant structure can be constructed from the symplectic structure of the underlying theory. Then we will be concerned with more general applications of the localization principle in supersymmetric QFT, where the super-Poincaré group action allows for an equivariant cohomological interpretation. In both frameworks, we present examples of localization under the action of a single supersymmetry, whose “square” generates a bosonic $U(1)$ symmetry. Supersymmetric localization was recently applied to many cases of QFT on curved spacetimes, so we must consider those curved background that preserve at least one supersymmetry, as discussed in Section 4.4. ### 5.1 Localization principle in Hamiltonian QM We consider now, following [19] and refernces therein, the path integral quantization of an Hamiltonian system $(M,\omega,H)$, of the $2n$-dimensional phase space $M$ with symplectic form $\omega$, and an Hamiltonian function $H\in C^{\infty}(M)$. This is simply QM viewed as a (0+1)-dimensional QFT over the base “spacetime” $\mathbb{R}$ or $\mathbb{S}^{1}$, that now is only “time”, and with target space $M$, that physically represents the phase space of the system. The fields of the theory are the paths $\gamma:\mathbb{R}(\mathbb{S}^{1})\to M$, that means we consider a trivial total space $E=\mathbb{R}(\mathbb{S}^{1})\times M$. In principle the time axis can be chosen to be the real line (or an interval with some prescribed boundary conditions), or the circle (that corresponds to periodic boundary conditions), but we will soon see that it is much convenient technically to chose the latter possibility, so consider the “loop space” $\mathcal{F}=C^{\infty}(\mathbb{S}^{1},M)$. A field for us is so a closed curve $\gamma:[0,T]\to M$ such that $\gamma(0)=\gamma(T)$. Since we make the periodicity explicit in $t$, we interpret this parameter as an “Euclidean” Wick-rotated time, so that $T$ is the inverse temperature of the canonical ensemble. We set up now some differential geometric concept on the loop space that we are going to use in the following. If $\\{x^{\mu}\\}$ are coordinates on $M$, on the loop space we can choose an infinite set of coordinates $\\{x^{\mu}(t)\\}$ for $\mu\in{1,\cdots,2n}$ and $t\in[0,T]$, such that for any $\gamma\in\mathcal{F}$ $x^{\mu}(t)[\gamma]:=x^{\mu}(\gamma(t)).$ Using the standard rules of functional derivation, a vector field in $\Gamma(T\mathcal{F})$ can be thus expressed locally with respect to these coordinates as $X=\int_{0}^{T}dt\ X^{\mu}(t)\frac{\delta}{\delta x^{\mu}(t)}$ (5.11) where $X^{\mu}(t)$ are functions over $\mathcal{F}$, and $(\delta/\delta x^{\mu}(t))_{\gamma}$ is a basis element of the tangent space $T_{\gamma}\mathcal{F}$ at $\gamma$. Many other geometric objects can be lifted from $M$ to $\mathcal{F}$ following this philosophy. For example, for any function in $C^{\infty}(M)$ as the Hamiltonian $H$, we can define $H(t)\in C^{\infty}(\mathcal{F})$ at a given time $t$, such that $H(t)[\gamma]:=H(\gamma(t))$. The action functional instead is the function over the loop space such that $\displaystyle S[\gamma]$ $\displaystyle=\int_{0}^{T}dt\ \left[\dot{q}^{a}(t)p_{a}(t)-H(p(t),q(t))\right]$ (5.12) $\displaystyle=\int_{0}^{T}dt\ \left[\theta_{\gamma(t)}(\dot{\gamma})-H(\gamma(t))\right]$ where in the first line we expressed $\gamma$ through its trivialization in Darboux coordinates $\\{q^{a},p_{a}\\}$ with $a\in\\{1,\cdots,n\\}$, and in the second line we expressed the same thing more covariantly using the (local) symplectic potential $\theta$ of $\omega$ and the velocity vector field $\dot{\gamma}$ of the curve. Concerning differential forms, if we consider the basis set $\\{\eta^{\mu}(t):=dx^{\mu}(t)\\}$, a $k$-degree element of $\Omega(\mathcal{F})$ can be expressed locally as $\alpha=\int_{0}^{T}dt_{1}\cdots\int_{0}^{T}dt_{k}\ \frac{1}{k!}\alpha_{\mu_{1}\cdots\mu_{k}}(t_{1},\cdots,t_{k})\eta^{\mu_{1}}(t_{1})\wedge\cdots\wedge\eta^{\mu_{k}}(t_{k})$ (5.13) and we recall that $\Omega(\mathcal{F})$ can be considered as the space of functions over the _super-loop space_ $\Pi T\mathcal{F}$, of coordinates $\\{x^{\mu}(t),\eta^{\mu}(t)\\}$. The de Rham differential on the loop space can be expressed as the cohomological vector field on $\Pi T\mathcal{F}$ $d_{\mathcal{F}}=\int_{0}^{T}dt\ \eta^{\mu}(t)\frac{\delta}{\delta x^{\mu}(t)}.$ (5.14) Finally, it is natural to lift on the loop space the symplectic structure of $M$, as well as a choice of Riemannian metric, as $\displaystyle\Omega$ $\displaystyle=\int_{0}^{T}dt\ \frac{1}{2}\omega_{\mu\nu}(t)\eta^{\mu}(t)\wedge\eta^{\nu}(t)$ (5.15) $\displaystyle G$ $\displaystyle=\int_{0}^{T}dt\ g_{\mu\nu}(t)\eta^{\mu}(t)\otimes\eta^{\nu}(t),$ i.e. $\Omega_{\mu\nu}(t,t^{\prime}):=\omega_{\mu\nu}(t)\delta(t-t^{\prime})$ and $G_{\mu\nu}:=g_{\mu\nu}(t)\delta(t-t^{\prime})$. $\Omega$ is closed under the loop space differential $d_{\mathcal{F}}$.666Strictly speaking, the 2-form $\Omega$ should be called “pre-symplectic”, since although it is certainly closed, it is not necessarily non-degenerate on the loop space. Considering the standard Liouville measure on $M$ $\frac{\omega^{n}}{n!}=d^{2n}x\ \mathrm{Pf}(\omega(x))=d^{n}qd^{n}p,$ (5.16) we can write now the path integral measure for QM on the loop space as an infinite product of the Liouville one for any time $t\in[0,T]$, and get $\int_{\mathcal{F}}\frac{\Omega^{n}}{n!}=\int_{\mathcal{F}}D^{2n}x\ \mathrm{Pf}(\Omega[x])=\int_{\Pi T\mathcal{F}}D^{2n}xD^{2n}\eta\ \frac{\Omega^{n}}{n!}.$ (5.17) Here in the last equality we rewrote the integral over $\mathcal{F}$ as an integral over the super-loop space, analogously to (4.16). The path integral for the quantum partition function is thus $\displaystyle Z(T)$ $\displaystyle=\int_{\mathcal{F}}D^{2n}x\ \mathrm{Pf}(\Omega[x])e^{-S[x]}$ (5.18) $\displaystyle=\int_{\Pi T\mathcal{F}}D^{2n}xD^{2n}\eta\ \frac{\Omega^{n}[x,\eta]}{n!}e^{-S[x]}$ $\displaystyle=\int_{\Pi T\mathcal{F}}D^{2n}xD^{2n}\eta\ e^{-(S[x]+\Omega[x,\eta])}$ where in the last line we exponentiated the loop space symplectic form, making explicit the formal analogy with the Duistermaat-Heckman setup. In particular, we associated to the “classical” Hamiltonian system $(M,\omega,H)$ an Hamiltonian system $(\mathcal{F},\Omega,S)$ on the loop space. Here the loop space Hamiltonian function $S$ generates an Hamiltonian $U(1)$-action on $\mathcal{F}$, that can be used to formally apply the same equivariant localization principle as in the finite-dimensional case. As for the proof of the ABBV formula, we introduced a graded structure on field space formally rewriting the path integration on the super-loop space $\Pi T\mathcal{F}$. This graded structure is now simply given by the form-degree on the extended field space $\Omega(\mathcal{F})$. Let now $X_{S}$ be the Hamiltonian vector field associated to $S\in C^{\infty}(\mathcal{F})$, such that $d_{\mathcal{F}}S=-\iota_{X_{S}}\Omega$, or equivalently $X_{S}=\Omega(\cdot,d_{\mathcal{F}}S)$. Explicitly, in coordinates $\\{x^{\mu}(t)\\}$ $\begin{split}X_{S}^{\mu}(t)&=\int_{0}^{T}dt^{\prime}\ \Omega^{\mu\nu}(t,t^{\prime})\frac{\delta S}{\delta x^{\nu}(t)}\\\ &=\omega^{\mu\nu}(x(t))\bigl{(}\omega_{\nu\rho}(x(t))\dot{x}^{\rho}(t)-\partial_{\nu}H(x(t))\bigr{)}\\\ &=\dot{x}^{\mu}(t)-X_{H}^{\mu}(x(t))\end{split}$ (5.19) where $\dot{x}(t)$ is the vector field on $\mathcal{F}$ with components such that $\dot{x}^{\mu}(t)[\gamma]:=(x^{\mu}\circ\gamma)^{\prime}(t)\equiv\dot{\gamma}^{\mu}(t)$. The flow of $X_{S}$ defines the Hamiltonian $U(1)$-action on $\mathcal{F}$ and the infinitesimal action of the Lie algebra $\mathfrak{u}(1)$ on $\Omega(\mathcal{F})$ through the Lie derivative $\mathcal{L}_{X_{S}}$. The Cartan model for the $U(1)$-equivariant cohomology of $\mathcal{F}$ is then defined by the space of equivariant differential forms $\Omega_{S}(\mathcal{F}):=\left(\mathbb{R}[\phi]\otimes\Omega(\mathcal{F})\right)^{U(1)}\cong\Omega(\mathcal{F})^{U(1)}[\phi]$ (5.20) and the equivariant differential $\displaystyle Q_{S}$ $\displaystyle:=\mathds{1}\otimes d_{\mathcal{F}}-\phi\otimes\iota_{X_{S}}\equiv d_{\mathcal{F}}+\iota_{X_{S}}$ (5.21) $\displaystyle=\int_{0}^{T}dt\bigl{(}\eta^{\mu}(t)+\dot{x}^{\mu}(t)-X^{\mu}_{H}(x(t))\bigr{)}\frac{\delta}{\delta x^{\mu}(t)},$ where as usual we localized algebraically setting $\phi=-1$ to ease the notation. The square of this operator gives, after some simplifications $Q_{S}^{2}=\int_{0}^{T}dt\left(\frac{d}{dt}-\left.\mathcal{L}_{X_{H}}\right|_{x(t)}\right)$ (5.22) where the second term is the Lie derivative on $M$ with respect to $X_{H}$, lifted on $\mathcal{F}$ at every value of $t$. The first term, when evaluated on a field, gives only contributions from the values at $t=0,T$, and so it vanishes thanks to the fact that we chose periodic boundary conditions! This means that, with this choice, the Cartan model on field space is completely determined by the lift of the $U(1)$-invariant forms on $M$, for which $\mathcal{L}_{X_{H}}\alpha=0$. Consistently, if we restrict to this subspace of $\Omega(M)$ where the energy is preserved, $Q_{S}\equiv Q_{\dot{x}}=d_{\mathcal{F}}+\iota_{\dot{x}}$ acts as the _supersymmetry_ operator generating time-translations on the base $\mathbb{S}^{1}$: $Q_{\dot{x}}^{2}=\frac{1}{2}[Q_{\dot{x}},Q_{\dot{x}}]=\int_{0}^{T}dt\frac{d}{dt}$ (5.23) resembling the supersymmetry algebra (4.44) for $\mathcal{N}=1$ and 1-dimensional spacetime. We will see in the next section that this is not just a coincidence, but we can relate this model to a supersymmetric version of QM. This restricted differential acts on coordinates of the super-loop space as $Q_{\dot{x}}x^{\mu}(t)=\eta^{\mu}(t),\qquad Q_{\dot{x}}\eta^{\mu}(t)=\dot{x}^{\mu}(t),$ (5.24) while the full equivariant differential acts as $Q_{S}x^{\mu}(t)=\eta^{\mu}(t),\qquad Q_{S}\eta^{\mu}(t)=X_{S}^{\mu}(t),$ (5.25) both exchanging “bosonic” with “fermionic” degrees of freedom. We remark that we started from a standard (non supersymmetric) Hamiltonian theory on $\mathcal{F}$, and from this we constructed a supersymmetric theory on $\Pi T\mathcal{F}$, whose supersymmetry is encoded in the Hamiltonian symmetry (so in the symplectic structure) of the original theory. This “hidden” supersymmetry is thus interpretable, in the spirit of Topological Field Theory, as a BRST symmetry, and the differential $Q_{S}$ as a “BRST charge” under which the augmented action $(S+\Omega)$ is supersymmetric: $Q_{S}(S+\Omega)=d_{\mathcal{F}}S+d_{\mathcal{F}}\Omega+\iota_{X_{S}}\Omega=d_{\mathcal{F}}S+0-d_{\mathcal{F}}S=0.$ (5.26) In other words, $(S+\Omega)$ is an _equivariantly closed extension_ of the symplectic 2-form $\Omega$, analogously to the finite-dimensional Hamiltonian geometry discussed in Chapter 3.3. The same argument cannot be straightforwardly applied to any QFT, since in general we do not have a symplectic structure on the field space,777A symplectic structure can be induced from the action principle on the subspace of solutions of the classical EoM, but it does not lift on the whole field space in general. but in the presence of a gauge symmetry we know that a BRST supersymmetry can be used to define the equivariant cohomology on the field space and exploit the localization principle. We will expand this a bit in the next chapter. It is now possible to mimic the procedure of Section 4.2 to localize the supersymmetric path integral $Z(T)=\int_{\Pi T\mathcal{F}}D^{2n}xD^{2n}\eta\ e^{-(S[x]+\Omega[x,\eta])}$ (5.27) seen as an integral of an equivariantly closed form. We modify the integral introducing the “localizing action” $S_{loc}[x,\eta]:=Q_{S}\Psi[x,\eta]$, with localization 1-form $\Psi\in\Omega^{1}(\mathcal{F})^{U(1)}$, the so-called “gauge-fixing fermion”: $Z_{T}(\lambda)=\int_{\Pi T\mathcal{F}}D^{2n}xD^{2n}\eta\ e^{-(S[x]+\Omega[x,\eta]-\lambda Q_{S}\Psi[x,\eta])}$ (5.28) where $\lambda\in\mathbb{R}$ is a parameter. The resulting integrand is again explicitly equivariantly closed, and we can check that this path integral is formally independent on the parameter $\lambda$. Indeed, after a shift $\lambda\mapsto\lambda+\delta\lambda$, (5.28) becomes $Z_{T}(\lambda+\delta\lambda)=\int_{\Pi T\mathcal{F}}D^{2n}xD^{2n}\eta\ e^{-(S[x]+\Omega[x,\eta]-\lambda Q_{S}\Psi[x,\eta]-\delta\lambda Q_{S}\Psi[x,\eta])},$ (5.29) and we can make a change of integration variables to absorb the resulting shift at the exponential. Since the exponential is supersymmetric, we change variables using a supersymmetry transformation: $\begin{split}&x^{\mu}(t)\mapsto x^{\prime\mu}(t):=x^{\mu}(t)+\delta\lambda\Psi Q_{S}x^{\mu}(t)=x^{\mu}(t)+\delta\lambda\eta^{\mu}(t)\\\ &\eta^{\mu}(t)\mapsto\eta^{\prime\mu}(t):=\eta^{\mu}(t)+\delta\lambda\Psi Q_{S}\eta^{\mu}(t)=\eta^{\mu}(t)+\delta\lambda X_{S}^{\mu}(t)\end{split}$ (5.30) so that the only change in the integral comes from the integration measure, $\displaystyle D^{2n}xD^{2n}\eta\mapsto D^{2n}x^{\prime}D^{2n}\eta^{\prime}$ $\displaystyle=\mathrm{Sdet}\left[\begin{array}[]{cc}\partial x^{\prime}/\partial x&\partial x^{\prime}/\partial\eta\\\ \partial\eta^{\prime}/\partial x&\partial\eta^{\prime}/\partial\eta\end{array}\right]D^{2n}xD^{2n}\eta$ (5.31) $\displaystyle=e^{-\delta\lambda Q_{S}\Psi}D^{2n}xD^{2n}\eta.$ Putting all together, $\displaystyle Z_{T}(\lambda+\delta\lambda)$ $\displaystyle=\int_{\Pi T\mathcal{F}}D^{2n}x^{\prime}D^{2n}\eta^{\prime}\ e^{-(S[x^{\prime}]+\Omega[x^{\prime},\eta^{\prime}]-\lambda Q_{S}\Psi[x^{\prime},\eta^{\prime}]-\delta\lambda Q_{S}\Psi[x^{\prime},\eta^{\prime}])}$ (5.32) $\displaystyle=\int_{\Pi T\mathcal{F}}D^{2n}xD^{2n}\eta\ e^{-(S[x]+\Omega[x,\eta]-\lambda Q_{S}\Psi[x,\eta])}=Z_{T}(\lambda).$ The same property can be seen less rigorously by exploiting some sort of (arguable) infinite-dimensional version of Stokes’ theorem: $\displaystyle\frac{d}{d\lambda}Z_{T}(\lambda)$ $\displaystyle=\int_{\Pi T\mathcal{F}}D^{2n}xD^{2n}\eta\ (Q_{S}\Psi)e^{-(S[x]+\Omega[x,\eta]-\lambda Q_{S}\Psi[x,\eta])}$ (5.33) $\displaystyle=\int_{\Pi T\mathcal{F}}D^{2n}xD^{2n}\eta\left(Q_{S}\Psi e^{-(S[x]+\Omega[x,\eta]-\lambda Q_{S}\Psi[x,\eta])}\right)=0,$ that holds if we assume the path integration measure to be non-anomalous under $Q_{S}$. Since the path integral (5.28) is independent on the parameter, we can take the limit $\lambda\to\infty$ and obtain the localization formula $Z(T)=\lim_{\lambda\to\infty}\int_{\Pi T\mathcal{F}}D^{2n}xD^{2n}\eta\ e^{-(S[x]+\Omega[x,\eta]-\lambda Q_{S}\Psi[x,\eta])}$ (5.34) that “localizes” $Z(T)$ onto the zero locus of $S_{loc}$. Of course different choices of gauge-fixing fermion induce different final localization formulas for the path integral, but at the end they should all give the same result. We now present two different localization formulas derived from (5.34) with different choices of localizing term. The fist canonical choice of gauge fixing fermion we can make mimics the same procedure we used in the finite-dimensional case. Under the same assumptions we made in Section 3.1, we pick a $U(1)_{H}$-invariant metric $g$ on $M$, and lift it to $\mathcal{F}$ using (5.15), so that the resulting $G$ is $U(1)_{S}$-invariant: $\mathcal{L}_{S}G=Q_{S}^{2}G=0$. Then the localization 1-form is taken to be $\Psi[x,\eta]:=G(X_{S},\cdot)=\int_{0}^{T}dt\ g_{\mu\nu}(x(t))\Bigl{(}\dot{x}^{\mu}(t)-X^{\mu}_{H}(x(t))\Bigr{)}\eta^{\nu}(t),$ (5.35) so that the localization locus is the subspace where $X_{S}=0$, i.e. the moduli space of classical trajectories [41]: $\mathcal{F}_{S}=\left\\{\gamma\in\mathcal{F}:\left(\frac{\delta S}{\delta x^{\mu}(t)}\right)_{\gamma}=0\right\\}.$ (5.36) If this space consists of isolated, non-degenerate trajectories, we can apply the non-degenerate version of the ABBV formula for a circle action, and get $Z(T)=\sum_{\gamma\in\mathcal{F_{S}}}\frac{\mathrm{Pf}\left[\omega(\gamma(t))\right]}{\sqrt{\det{[dX_{S}[\gamma]/2\pi]}}}e^{-S[\gamma]}$ (5.37) where the pfaffian and the determinant are understood in the functional sense, spanning both the phase space indices $\mu\in\\{1,\cdots,2n\\}$ and the time continuous index $t\in[0,T]$. In general, for non isolated classical trajectories we have to decompose any $\gamma\in\mathcal{F}$ near to the fixed point set, splitting the classical component and normal fluctuations, as we did in Section 4.2. Then, rescaling the normal fluctuations by $\sqrt{\lambda}$ and thanks to the Berezin integration rules on the super-loop space, the same argument of the finite-dimensional case leads to $Z(T)=\int_{\mathcal{F_{S}}}D^{2n}x\ \frac{\mathrm{Pf}\left[\omega(x(t))\right]}{\left.\mathrm{Pf}\left[\delta^{\mu}_{\nu}\partial_{t}-(B+R)^{\mu}_{\nu}(x(t))\right]\right|_{N\mathcal{F}_{S}}}e^{-S[x]}$ (5.38) where $B^{\mu}_{\nu}=g^{\mu\rho}(\nabla_{[\rho}X_{H})_{\nu]}$, while $\nabla$ and $R$ are the connection and curvature of the metric $g$ on $M$, evaluated on $\mathcal{F}$ time-wise as usual. Notice that this localization scheme makes the contribution from the classical configurations explicit, resembling the exactness of the saddle point approximation (5.9), with 1-loop determinant given by the pfaffian at the denominator. However, even if the integration domain has been reduced, one must still perform a difficult infinite- dimensional path integration to get the final answer, whose $T$-dependence for example looks definitely non-trivial from (5.38). We can consider another choice of localizing term to simplify the final result, setting the gauge-fixing fermion to $\Psi[x,\eta]:=G(\dot{x},\cdot)=\int_{0}^{T}dt\ g_{\mu\nu}(x(t))\dot{x}^{\mu}(t)\eta^{\nu}(t).$ (5.39) With this choice, the gauge-fixed action reads $\displaystyle S$ $\displaystyle[x]+\Omega[x,\eta]+\lambda Q_{S}\Psi[x,\eta]=$ (5.40) $\displaystyle=\int_{0}^{T}dt\biggl{(}\dot{x}^{\mu}\theta_{\mu}-H+\frac{1}{2}\omega_{\mu\nu}\eta^{\mu}\eta^{n}u+\lambda\left(g_{\mu\nu,\sigma}\dot{x}^{\mu}\eta^{\sigma}\eta^{\nu}+\eta^{\mu}\partial_{t}(g_{\mu\nu}\eta^{\nu})+g_{\mu\nu}\dot{x}^{\mu}\dot{x}^{\nu}-g_{\mu\nu}\dot{x}^{\mu}X_{H}^{\nu}\right)\biggr{)}$ $\displaystyle=\int_{0}^{T}dt\biggl{(}\lambda g_{\mu\nu}\dot{x}^{\mu}\dot{x}^{\nu}+\lambda\eta^{\mu}\nabla_{t}\eta^{\nu}+\dot{x}^{\mu}\theta_{\mu}+\frac{1}{2}\omega_{\mu\nu}\eta^{\mu}\eta^{\nu}-H-\lambda g_{\mu\nu}\dot{x}^{\mu}X_{H}^{\nu}\biggr{)}$ where in the second line the time-covariant derivative acts as $\nabla_{t}\eta^{\nu}=\partial_{t}\eta^{\nu}+\Gamma^{\nu}_{\rho\sigma}\dot{x}^{\rho}\eta^{\sigma}$, and the localization locus is the subset of constant loops $\mathcal{F}_{0}:=\left\\{\gamma\in\mathcal{F}:\dot{\gamma}=0\right\\}\cong M$ (5.41) that is, points in $M$. Splitting again the loops near this subspace in constant modes plus fluctuations, and rescaling the latter as we did before, the path integral is reduced to a finite-dimensional integral over $M$, the Niemi-Tirkkonen localization formula [71] $Z(T)=\int_{\Pi TM}\sqrt{g}d^{2n}xd^{2n}\eta\ \frac{e^{-T(H(x)-\omega(x,\eta))}}{\sqrt{\det^{\prime}{\left[g_{\mu\nu}\partial_{t}-(B_{\mu\nu}+R_{\mu\nu})\right]}}}$ (5.42) where the prime on the determinant means it is taken over the normal fluctuation modes, excluding the constant ones, giving exactly the equivariant Euler form of the normal bundle to $\mathcal{F}_{0}$. This formula is much more appealing since it contains no refernce to $T$-dependent submanifolds of $\mathcal{F}$, and the evaluation of the action on constant modes simply gives the Hamiltonian at those points multiplied by $T$. The functional determinant at the denominator requires a specific regularization: using the $\zeta$-function method, it can be simplified to $\frac{1}{\sqrt{\det^{\prime}{\left[g_{\mu\nu}\partial_{t}-(B_{\mu\nu}+R_{\mu\nu})\right]}}}=\sqrt{\det{\left[\frac{\frac{T}{2}(B+R)_{\mu\nu}}{\sinh{\left(\frac{T}{2}(B+R)_{\mu\nu}\right)}}\right]}}=\hat{A}_{H}(TR)$ (5.43) where in the last equality we rewrote, by definition, the determinant as the _$U(1)_{H}$ -equivariant Dirac $\hat{A}$-genus_ of the curvature $R$ up to a constant $T$, that is the Dirac $\hat{A}$-genus of the equivariant extension of the curvature, $R+B$ [72]. Note that the exponential can be rewritten as the $U(1)_{H}$-equivariant Chern character of the symplectic form, $e^{-H+\omega}=\mathrm{ch}_{H}(\omega)$ (5.44) and so the partition function can be nicely rewritten as $Z(T)=\int_{M}\mathrm{ch}_{H}(T\omega)\wedge\hat{A}_{H}(TR)$ (5.45) in terms of equivariant characteristic classes of the phase space $M$ with respect to the $U(1)_{H}$ Hamiltonian group action, that are determined by the initial classical system. The only remnant of the quantum theory is in the, now very explicit, dependence on the inverse temperature $T$. This form of the partition function emphasizes the fact that if we put $H=0$, we end up with a _topological_ theory. In this case there are no propagating physical degrees of freedom, and the partition function only describes topological properties of the underlying phase space. In the next section we will report a non- trivial example of this kind. ### 5.2 Localization and index theorems A famous and important application of the localization principle to loop space path integrals is the supersymmetric derivation of the _Atiyah-Singer index theorem_ [73] [74] [75]. The theorem relates the _analytical index_ of an elliptic differential operator on a compact manifold to a topological invariant, connecting the local data associated to solutions of partial differential equations to global properties of the manifold. Supersymmetric localization allowed to prove this statement, in a new way with respect to the original proof, for different examples of classical differential operators. We describe here the application to the index of the Dirac operator acting on the spinor bundle $S$ on an even-dimensional compact manifold $M$, in presence of a gravitational and electromagnetic background, specified by the metric $g$ and a $U(1)$ connection 1-form $A$ on a $\mathbb{C}$-line bundle $L_{\mathbb{C}}$ over $M$.888This can be extended to non-Abelian gauge groups as well, but for simplicity we report the Abelian case. The Dirac operator is defined as the fiber-wise Clifford action of the covariant derivative on the twisted spinor bundle $TM\otimes S\otimes L_{\mathbb{C}}$ over $M$: $i\not{\nabla}=i\gamma^{\mu}\left(\partial_{\mu}+\frac{1}{8}\omega_{\mu ij}[\gamma^{i},\gamma^{j}]+iA_{\mu}\right)$ (5.46) where $\\{\gamma^{\mu}\\}$ are the gamma-matrices generating the Clifford algebra in the given spin representation, satisfying the anticommutation relation $\\{\gamma^{\mu},\gamma^{\nu}\\}=2g^{\mu\nu},$ (5.47) and $\omega$ is the spin connection related to the metric.999See Appendix A.2. The “curved” and “flat” indices are related through the vielbein $e^{i}_{\mu}(x)$, $g_{\mu\nu}(x)=e^{i}_{\mu}(x)e^{j}_{\nu}(x)\eta_{ij},\qquad\gamma^{i}=e^{i}_{\mu}(x)\gamma^{\mu}(x),$ (5.48) with $\eta$ the flat metric. The analytical index of the Dirac operator is defined as [76] $\mbox{index}(i\not{\nabla}):=\dim\mathrm{Ker}(i\not{\nabla})-\dim\mathrm{coKer}(i\not{\nabla})=\dim\mathrm{Ker}(i\not{\nabla})-\dim\mathrm{Ker}(i\not{\nabla}^{\dagger}).$ (5.49) We have thus to find the number of zero-energy solutions of the Dirac equation $i\not{\nabla}\Psi=E\Psi,$ (5.50) where $\Psi$ is a Dirac spinor. In even dimensions, we can decompose the problem in the chiral basis of the spin representation, where $\gamma^{i}=\left(\begin{array}[]{cc}0&\sigma^{i}\\\ \overline{\sigma}^{i}&0\end{array}\right),\quad\gamma^{c}=\left(\begin{array}[]{cc}\mathds{1}&0\\\ 0&-\mathds{1}\end{array}\right),\quad i\not{\nabla}=\left(\begin{array}[]{cc}0&D\\\ D^{\dagger}&0\end{array}\right),\quad\Psi=\left(\begin{array}[]{c}\psi_{-}\\\ \psi_{+}\end{array}\right),$ (5.51) and the chirality matrix is denoted by $\gamma^{c}$. In this representation, we see that the index counts the number of zero-energy modes with positive chirality minus the number of zero-energy modes of negative chirality, $\mbox{index}(i\not{\nabla})=\dim\mathrm{Ker}(D)-\dim\mathrm{Ker}(D^{\dagger}).$ (5.52) It is possible to give a path integral representation of this index, a key ingredient to apply the localization principle. In order to do that, we first prove that it can be rewritten as a _Witten index_ , $\mbox{index}(i\not{\nabla})=\mathrm{Tr}_{\mathcal{H}}\left(\gamma^{c}e^{-T\Delta}\right)$ (5.53) where $\Delta:=(i\not{\nabla})^{2}$ is the Shröedinger operator (the covariant Laplacian) and the parameter $T>0$ is a regulator for the operator trace, taken over the space $\mathcal{H}$ of Dirac spinors, sections of the twisted spinor bundle over $M$. ###### Proof of (5.53). First, we notice that $(i\not{\nabla})$ is symmetric and elliptic, and since $M$ is compact it is also essentially self-adjoint [77]. Thus, it has a well defined spectrum $\\{\Psi^{E}\\}$ that forms a basis of the function space at hand. The same modes diagonalize also the Schrödinger operator, $\Delta\Psi^{E}=E^{2}\Psi^{E}$ (5.54) so we can shift the attention to solutions of the Schrödinger equation with eigenvalue satisfying $E^{2}=0$. It is easy to see that the Dirac operator anticommutes with the chirality matrix $\gamma^{c}$, and so the Schrödinger operator commutes with it. Thus we can split the field space in complementary subspaces $\mathcal{S}^{E}_{\pm}:=\left\\{\Psi\in\mathcal{H}:\Delta\Psi=E^{2}\Psi,\gamma^{c}\Psi=\pm\Psi\right\\}$, for every eigenvalue $E^{2}$ and chirality $(\pm)$. For every non-zero energy, we establish an isomorphism $\mathcal{S}^{E}_{+}\cong\mathcal{S}^{E}_{-}$: using the fact that $(i\not{\nabla})$ and $\gamma^{c}$ anticommute, starting from a solution with eigenvalue $E^{2}$ and definite chirality $\Psi_{\pm}$ we can construct another one with opposite chirality $(i\not{\nabla}\Psi_{\pm})$, $\displaystyle\Delta(i\not{\nabla}\Psi_{\pm})=i\not{\nabla}i\not{\nabla}i\not{\nabla}\Psi_{\pm}=E^{2}(i\not{\nabla}\Psi_{\pm})$ (5.55) $\displaystyle\gamma^{c}(i\not{\nabla}\Psi_{\pm})=-i\not{\nabla}\gamma^{c}\Psi_{\pm}=-(\pm)(i\not{\nabla}\Psi_{\pm}).$ Thus the maps $\phi_{\pm}:\mathcal{S}^{E}_{\pm}\to\mathcal{S}^{E}_{\mp}$ such that $\phi_{\pm}(\Psi):=\frac{i\not{\nabla}}{|E|}\Psi$ (5.56) are both well defined and are right and left inverse of each other, giving the bijective correspondence for every $|E|\neq 0$. This is the well known fact that particles and antiparticles are created in pairs with opposite energy and chirality. If we take the trace over the field space $\mathcal{H}$, this is splitted into the sum of the traces over every subspace $\mathcal{S}^{E}$ of definite energy squared. In every one of them the restricted Witten index gives $\mathrm{Tr}_{\mathcal{S}^{E}}\left(\gamma^{c}e^{-T\Delta}\right)=e^{-TE^{2}}\left(\mathrm{Tr}_{\mathcal{S}^{E}_{+}}(\mathds{1})-\mathrm{Tr}_{\mathcal{S}^{E}_{-}}(\mathds{1})\right)=0$ (5.57) for every $E\neq 0$. So the whole trace is formally independent on $T$, resolving on the subspace of zero-energy, where the number of chirality + and - eigenstates is different in general: $\displaystyle\mathrm{Tr}_{\mathcal{H}}\left(\gamma^{c}e^{-T\Delta}\right)$ $\displaystyle=\mathrm{Tr}_{E=0}(\gamma^{c})$ (5.58) $\displaystyle=\\#^{E=0}(\mathrm{chirality\ (+)\ modes})-\\#^{E=0}(\mathrm{chirality\ (-)\ modes}).$ ∎ The Witten index representation (5.53) and the chirality decomposition of the operators of interest (5.51) permit to see the current problem as a $\mathcal{N}=1$ supersymmetric QM on the manifold $M$, identifying chirality +(-) spinors with bosonic(fermionic) states. Here, the supersymmetry algebra (4.57) is simply101010Often this is called $\mathcal{N}=1/2$ supersymmetry, leaving the name $\mathcal{N}=1$ for the complexified algebra with generators $Q,\tilde{Q}$, and imposition of Majorana condition. (choosing an appropriate normalization) $[Q,Q]=2H,$ (5.59) and corresponds to the Schrödinger operator above, if we make the following identifications: $\displaystyle i\not{\nabla}$ $\displaystyle\leftrightarrow\quad Q$ (5.60) $\displaystyle\Delta=(i\not{\nabla})^{2}$ $\displaystyle\leftrightarrow\quad H=Q^{2}.$ The chirality matrix $\gamma^{c}$ is identified with the operator $(-1)^{F}$, where $F$ is the fermion number operator, that assigns eigenvalue $+1$ to bosonic states and $-1$ to fermionic states. The Witten index representation in the quantum system is thus $\displaystyle\mbox{index}(i\not{\nabla})$ $\displaystyle=\mathrm{Tr}\left((-1)^{F}e^{-TH}\right)$ (5.61) $\displaystyle=n^{E=0}(\mathrm{bosons})-n^{E=0}(\mathrm{fermions}).$ The proof above, translated in terms of the quantum system, shows that in supersymmetric QM eigenstates of the Hamiltonian have non-negative energy, and are present in fermion-boson pairs for every non-zero energy. Since $Q^{2}$ is a positive-definite Hermitian operator, the zero modes $|0\rangle$ of $H$ are supersymmetric, $Q|0\rangle=0$ (they do not have supersymmetric partners). Thus the non-vanishing of the Witten index (5.61) is a sufficient condition to ensure that there is at least one supersymmetric vacuum state available, whereas its vanishing is a necessary condition for spontaneous braking of supersymmetry by the vacuum. The Witten index has a super-loop space path integral representation [78], so that we can rewrite (5.61) as $\mbox{index}(i\not{\nabla})=\int D^{2n}\phi D^{2n}\psi\ e^{-TS[\phi,\psi]}$ (5.62) where $S$ is the Euclidean action corresponding to the Hamiltonian $H$ and the fields are defined on the unit circle. The appropriate supersymmetric theory which describes a spinning particle on a gravitational background is the 1-dimensional _supersymmetric non-linear $\sigma$ model_. The superspace formulation of this model considers the base space as a super extension of the 1-dimensional spacetime with coordinates $(t,\theta)$, and $M$ as the target space with covariant derivative given by the Dirac operator. A superfield, trivialized with respect to coordinates $(t,\theta)$ and $(x^{\mu})$ on $M$ is then $\Phi^{\mu}(t,\theta)\equiv(x^{\mu}\circ\Phi)(t,\theta)=\phi^{\mu}(t)+\psi^{\mu}(t)\theta,$ (5.63) and supersymmetry transformations are given by the action of the odd vector field $\underline{Q}=\partial_{\theta}+\theta\partial_{t}$, that in terms of component fields reads $\delta\phi^{\mu}=\psi^{\mu},\qquad\delta\psi^{\mu}=\dot{\phi}^{\mu}.$ (5.64) Denoting the superderivative as $D=-\partial_{\theta}+\theta\partial_{t}$, the action of the non-linear $\sigma$ model coupled to the gauge field $A$ can be given as $S[\Phi]=\int dt\int d\theta\ \frac{1}{2}g_{\Phi(t,\theta)}\left(D\Phi,\dot{\Phi}\right)+A(D\Phi)$ (5.65) where $D\Phi$ and $\dot{\Phi}$ are thought as (super)vector fields on $M$ such that, for any function $f\in C^{\infty}(M)$, $D\Phi(f):=D(f\circ\Phi)$ and $\dot{\Phi}(f):=\partial_{t}(f\circ\Phi)$. Inserting the trivialization for the metric components $g_{\mu\nu}(\Phi(t,\theta))=g_{\mu\nu}(\phi)+\theta\psi^{\sigma}(t)g_{\mu\nu,\sigma}(\phi)$, and the component expansion for $\Phi$, the action is simplified to $S[\phi,\psi]=\int dt\left(\frac{1}{2}g_{\mu\nu}(\phi)\dot{\phi}^{\mu}\dot{\phi}^{\nu}+\frac{1}{2}g_{\mu\nu}(\phi)\psi^{\mu}(\nabla_{t}\psi)^{\nu}+A_{\mu}(\phi)\dot{\phi}^{\mu}-\frac{1}{2}\psi^{\mu}F_{\mu\nu}\psi^{\nu}\right)$ (5.66) where we suppressed the $t$-dependence, $F_{\mu\nu}=\partial_{[\mu}A_{\nu]}$ are the components of the electromagnetic field strength, and the time- covariant derivative $\nabla_{t}$ acts as $(\nabla_{t}V)^{\sigma}(\phi(t))=\dot{V}^{\sigma}(\phi(t))+\Gamma^{\sigma}_{\mu\nu}\dot{\phi}^{\mu}(t)V^{\nu}(\phi(t)).$ The action (5.66) is formally equivalent to the model-independent action (5.40) of the last section, with $T$ behaving like the localization parameter $\lambda$, if we identify $\theta$ with the electromagnetic potential $A$ and $\omega$ with the field strength $F$, and if we set the Hamiltonian and its associated vector field $H,X_{H}$ to zero. This means that we can give to it an equivariant cohomological interpretation on the super-loop space $\Pi T\mathcal{F}$ over $M$, with coordinates identified with $(\phi^{\mu},\psi^{\mu})$, as was pointed out first by Atiyah and Witten [8]. Moreover, since the Hamiltonian vanishes, this action describes no propagating physical degrees of freedom, and thus the model is topological. Indeed, its value has to give the index of the Dirac operator, expected to be a topological quantity. To emphasize the equivariant cohomological nature of the model, we notice that the action functional can be split in the loop space (pre-)symplectic 2-form $\Omega[\phi,\psi]:=\int dt\ \frac{1}{2}\psi^{\mu}\left(g_{\mu\nu}\nabla_{t}-F_{\mu\nu}\right)\psi^{\nu}$ (5.67) and the loop space Hamiltonian $\mathcal{H}=\int dt\ \left(\frac{1}{2}g_{\mu\nu}\dot{\phi}^{\mu}\dot{\phi}^{\nu}+A_{\mu}\dot{\phi}^{\mu}\right).$ (5.68) They satisfy $d_{\mathcal{F}}\mathcal{H}=-\iota_{\dot{\phi}}\Omega$, so the supersymmetry transformation (5.64) is rewritten in terms of the Cartan differential $Q_{\dot{\phi}}=d_{\mathcal{F}}+\iota_{\dot{\phi}}$, and $Q_{\dot{\phi}}S=Q_{\dot{\phi}}(\mathcal{H}+\Omega)=0$. Moreover, we can find a loop space symplectic potential $\Sigma$ such that $S$ is equivariantly exact: $\displaystyle S[\phi,\psi]$ $\displaystyle=Q_{\dot{x}}\Sigma[\phi,\psi]$ (5.69) $\displaystyle\mathrm{where}\quad\Sigma[\phi,\psi]$ $\displaystyle:=\int dt\left(g_{\mu\nu}(\phi)\dot{\phi}^{\mu}+A_{\nu}(\phi)\right)\psi^{\nu}.$ Notice that, since the Hamiltonian $H$ vanishes, the localizing $U(1)$ symmetry is the one generated by time-translation with respect to the base space $\mathbb{S}^{1}$, an intrinsic property of the geometric structure that underlies the model. We can now apply the Niemi-Tirkkonen formula, localizing the path integral (5.62) into the moduli space of constant loops, i.e. as an integral over $M$. The result is the same as equation (5.45), but now since the Hamiltonian vanishes, the Chern class and the Dirac $\hat{A}$-genus (see Appendix B.1) are not equivariantly extended by the presence of an Hamiltonian vector field, giving the topological formula $\mbox{index}(i\not{\nabla})=\int_{M}\mbox{ch}(F)\wedge\hat{A}(R).$ (5.70) This is the result of the Atiyah-Singer index theorem for the Dirac operator on the twisted spinor bundle over $M$. Similar applications of the localization principle to variations of the non-linear $\sigma$ model give correct results for other classical complexes as well (de Rham, Dolbeault for example), in terms of different topological invariants [74]. ### 5.3 Equivariant structure of supersymmetric QFT and supersymmetric localization principle In the last section we saw how to give an equivariant cohomological interpretation to a model exhibiting Poincaré-supersymmetry, in terms of the super-loop space symplectic structure introduced before. In the following we are interested in applying the same kind of supersymmetric localization principle to higher dimensional QFT on a, possibly curved, $n$-dimensional spacetime $M$, where there is some preserved supersymmetry. In [9][10] it is argued that any generic quantum field theory with at least an $\mathcal{N}=1$ Poincaré supersymmetry admits a field space Hamiltonian (symplectic) structure and a corresponding $U(1)$-equivariant cohomology responsible for localization of the supersymmetric path integral. The key feature is an appropriate off-shell component field redefinition which defines a splitting of the fields into loop space “coordinates” and their associated “differentials”. In general, unlike the simplest case of the last section where bosonic fields were identified with coordinates and fermionic fields with 1-forms, loop space coordinates and 1-forms involve both bosonic and fermionic fields. It is proven that, within this field redefinition on the super-loop space, any supersymmetry charge $Q$ can be identified with a Cartan differential $Q=d_{\mathcal{F}}+\iota_{X_{+}}$ (5.71) analogously to (5.21), whose square generates translations in a given “light- cone” direction $x_{+}$, $Q^{2}=\mathcal{L}_{X_{+}}\sim\int_{M}\frac{\partial}{\partial x^{+}}$ (5.72) that corresponds to the $U(1)$ symmetry that can be used to exploit the localization principle. Taking the base spacetime to be compact in the light- cone direction, the periodic boundary conditions ensure $Q^{2}=0$, analogously to the loop space assumption of the one dimensional case. Also, it is argued that the supersymmetric action can be generally split into the sum of a loop space scalar function and a (pre-)symplectic 2-form, $S_{susy}=\mathcal{H}+\Omega$ (5.73) related by $d_{\mathcal{F}}\mathcal{H}=-\iota_{X_{+}}\Omega$. Thus, the supersymmetry of the action can be seen in general as the $U(1)$-equivariant closeness required to the application of the localization principle, and the path integral localizes onto the locus of constant loops, i.e. zero-modes of the fields. Even without entering in the details of this construction in terms of auxiliary fields redefinition, we feel now allowed to translate in full generality the circle localization principle in the framework of Poincaré- supersymmetric QFT. In the component-field description, we consider a rigid supersymmetric background over the given compact111111This ensures the loop space interpretation of above. spacetime $M$ and a graded field space $\mathcal{F}$ that plays the role of the super-loop space over $M$, whose even-degree forms are bosonic fields and the odd-degree forms are fermionic fields. The (infinitesimal) supersymmetry action of a preserved supercharge $Q$ plays the role of the Cartan differential $d_{\mathcal{F}}$, squaring to a bosonic symmetry that corresponds to the (infinitesimal) action of a $U(1)$ symmetry group, $Q^{2}\sim\mathcal{L}_{X}$ with $X$ an _even_ vector field. The Cartan model for the $U(1)$-equivariant cohomology of $\mathcal{F}$ is defined by the subcomplex of $Q$-invariant (or supersymmetric, or “BPS”) observables, where the supercharge squares to zero. We consider a supersymmetric model specified by the (Euclidean) action functional $S\in C^{\infty}(\mathcal{F})$ such that $\delta_{Q}S=0$, and a BPS observable $\mathcal{O}$ such that $\delta_{Q}\mathcal{O}=0$. Now the partition function (5.2) and the expectation value (5.3) are seen as integrations of equivariantly closed forms with respect to the differential $Q$. The supersymmetric localization principle then tells us that we can modify the respective integrals adding an equivariantly exact localizing term to the action, $\lambda S_{loc}[\Phi]:=\lambda\delta_{Q}\mathcal{V}[\Phi]$ (5.74) where $\mathcal{V}\in\Omega^{1}(\mathcal{F})^{U(1)}$ is a $Q^{2}$-invariant fermionic functional, the “gauge-fixing fermion” of Section 5.1, and $\lambda\in\mathbb{R}$ is a parameter. Assuming the supersymmetry $\delta_{Q}$ to be non anomalous, the partition function and the expectation value are not changed by this modification, i.e. the the new integrand lies in the same equivariant cohomology class, $\displaystyle\frac{d}{d\lambda}Z(\lambda)=\int_{\mathcal{F}}D\Phi\ (-\delta_{Q}\mathcal{V}[\Phi])e^{-(S+\lambda\delta_{Q}\mathcal{V})[\Phi]}=-\int_{\mathcal{F}}D\Phi\ \delta_{Q}\left(\mathcal{V}[\Phi]e^{-(S+\lambda\delta_{Q}\mathcal{V})[\Phi]}\right)=0$ (5.75) $\displaystyle\frac{d}{d\lambda}\langle\mathcal{O}\rangle_{\lambda}=-\frac{1}{Z(\lambda)}\left(\frac{d}{d\lambda}Z(\lambda)\right)\langle\mathcal{O}\rangle_{\lambda}+\frac{1}{Z(\lambda)}\int_{\mathcal{F}}D\Phi\ \delta_{Q}\left(\mathcal{V}[\Phi]\mathcal{O}[\Phi]e^{-(S+\lambda\delta_{Q}\mathcal{V})[\Phi]}\right)=0$ by the same argument of Section 5.1. Assuming the bosonic part of $\delta_{Q}\mathcal{V}$ to be positive-semidefinite, and using the $\lambda$-independence of the path integral, we can evaluate the partition function or the expectation value in the limit $\lambda\to+\infty$, getting the localization formulas $Z=\lim_{\lambda\to\infty}\int_{\mathcal{F}}D\Phi\ e^{-(S+\lambda S_{loc})[\Phi]},\qquad\langle\mathcal{O}\rangle=\frac{1}{Z}\lim_{\lambda\to\infty}\int_{\mathcal{F}}D\Phi\ \mathcal{O}[\Phi]e^{-(S+\lambda S_{loc})[\Phi]}.$ (5.76) The path integrals localize then onto the locus $\mathcal{F}_{0}$ of saddle points of $S_{loc}$. Following again the same argument of Section 4.2 we can in fact expand the fields about these saddle point configurations, rescale the normal fluctuations as $\Phi=\Phi_{0}+\frac{1}{\sqrt{\lambda}}\tilde{\Phi},$ (5.77) and the augmented action functional as $(S+\lambda S_{loc})[\Phi]=S[\Phi_{0}]+\frac{1}{2}\int_{M}d^{n}x\int_{M}d^{n}y\left(\frac{\delta^{2}S_{loc}}{\delta\Phi(x)\delta\Phi(y)}\right)_{\Phi_{0}}\tilde{\Phi}(x)\tilde{\Phi}(y)+O(\lambda^{-1/2}).$ (5.78) The functional measure on the normal sector $D\tilde{\Phi}$ is not affected by the rescaling, since the supersymmetric model contains the same number of bosonic and fermionic physical component fields,121212Note that this has to be true _off-shell_ , i.e. without imposing any EoM. and the corresponding Jacobians cancel by Berezin integration rules. The integral over this fluctuations is Gaussian and can be performed, giving the “1-loop determinant” analogous to the equivariant Euler class that appeared in Theorem 3.2.2. The leftover integral corresponds to the saddle point formula (5.9), but as an exact equality: $\displaystyle Z$ $\displaystyle=\int_{\mathcal{F}_{0}}D\Phi_{0}\ e^{-S[\Phi_{0}]}Z_{1-loop}[\Phi_{0}]$ (5.79) $\displaystyle\langle\mathcal{O}\rangle$ $\displaystyle=\frac{1}{Z}\int_{\mathcal{F}_{0}}D\Phi_{0}\ \mathcal{O}[\Phi_{0}]e^{-S[\Phi_{0}]}Z_{1-loop}[\Phi_{0}]$ where $Z_{1-loop}[\Phi_{0}]:=\left(\mathrm{Sdet}\left[\frac{\delta^{2}S_{loc}}{\delta\Phi(x)\delta\Phi(y)}[\Phi_{0}]\right]\right)^{-1}$ (5.80) and the super-determinant denotes collectively the result of Gaussian intergrations over bosonic or fermionic fields. Although the choice of localizing term $\mathcal{V}$ is arbitrary, and different choices give in principle different localization loci, the final result must be the same for every choice. At the end of Section 4.4 we remarked that the $\mathcal{N}=2$ supersymmetric Yang-Mills Lagrangian on a 3-dimensional maximally supersymmetric background is $Q$-exact, and thus can be used as a localizing term for supersymmetric gauge theories on this type of 3-dimensional spacetimes. It turns out that also the $\mathcal{N}=2$ matter (chiral) Lagrangian is $Q$-exact in three dimensions [79]. A canonical choice of localizing action can be, schematically [67] $S_{loc}[\Phi]:=\int_{M}\delta_{Q}\sum_{f}\left((\delta_{Q}\Phi_{f})^{\dagger}\Phi_{f}+\Phi_{f}^{\dagger}(\delta_{Q}\Phi_{f}^{\dagger})^{\dagger}\right)$ (5.81) where the sum runs over the fermionic fields of the theory. Its bosonic part is $\left.S_{loc}[\Phi]\right|_{bos}=\sum_{f}\left(|\delta_{Q}\Phi_{f}|^{2}+|\delta_{Q}\Phi_{f}^{\dagger}|^{2}\right),$ (5.82) that is indeed positive semidefinite. The corresponding localization locus is the subcomplex of BPS configurations, $[\mathrm{fermions}]=0,\qquad\delta_{Q}[\mathrm{fermions}]=0.$ (5.83) Concluding this general and schematic discussion, there are a couple of remarks we wish to point out. Firstly, in the above formulas we always considered generic BPS observables that are expressed through (local or non- local) combinations of the fields. In other words, their quantum expectation values are defined as insertions in the path integral of corresponding (classical) functionals on the field space. Examples of local objects of this kind are correlation functions of fundamental fields. A famous class of non- local quantum operators that are expressible as classical functionals are the so-called _Wilson loops_. In gauge theory with gauge group $G$ and local gauge field $A$, a Wilson loop in the representation $R$ of $Lie(G)$ over the closed curve $C:\mathbb{S}^{1}\to M$ is defined by $W_{R}(C):=\frac{1}{\dim R}\mathrm{Tr}_{R}\left(\mathcal{P}\exp{i\oint_{\mathbb{S}^{1}}C^{*}(A)}\right)$ (5.84) where the trace $\mathrm{Tr}_{R}$ is taken in the given representation.131313In the adjoint representation, this denotes an invariant inner product in $\mathfrak{g}$, for example the Killing form for a semisimple Lie algebra. This is gauge invariant, and represents physically the phase acquired by a charged probe particle in the representation $R$ after a tour on the curve $C$, in presence of the gauge potential $A$. Mathematically, if $R$ is the adjoint representation, the Wilson loop represents the parallel transport map between the fibers of the principal $G$-bundle defining the gauge theory. These operators have many interesting applications in physics: depending on the chosen curve $C$ their expectation value can be interpreted as an order parameter for the confinement/deconfinement phase transitions in QCD or the Bremsstrahlung function for an accelerated particle [80, 81, 82]. In the case of 3-dimensional Chern-Simons theory, they can be used to study topological invariants in knot theory [83]. In supersymmetric theories, they are particularly relevant for tests of the AdS/CFT correspondence [84]. In the next sections we will review some interesting cases in which expectation values of this type of operators can be evaluated exactly using the supersymmetric localization principle. There exists another class of interesting operators in the quantum theory, that cannot be expressed as classical functionals on the field space. These are the so-called _disorder operators_ , and their expectation values are defined by a restriction of the path integral to those field configurations which have prescribed boundary conditions around some artificial singularity introduced in spacetime. An example of these are the _’t Hooft operators_ , which introduce a Dirac monopole singularity along a path in a 4-dimensional space [85]. These kind of operators can be also studied non-perturbatively with the help of localization techniques [86]. For a great review of different examples of localization computations in supersymmetric QFT, see [12]. The second remark we wish to make is that, in presence of a gauge symmetry, the action functionals in the above formulas have to be understood as the quantum (i.e. gauge-fixed) action in order to give meaning to the corresponding partition function. That is, one has to introduce Faddeev-Popov ghost fields in the theory and the associated BRST transformations $\delta_{BRST}$. We have seen in Section 4.5 that it is always possible to see the BRST complex in terms of equivariant cohomology on the field space, so this supersymmetry transformation have to be incorporated in the equivariant structure of the supersymmetric theory. In this case, the field space acquires a $\mathbb{Z}$-grading corresponding to the ghost number, on top of the $\mathbb{Z}_{2}$ one from supersymmetry, and the appropriate Cartan differential with respect to which the equivariant cohomological structure is defined is then the total supersymmetry variation $Q=\delta_{susy}+\delta_{BRST}$. ### 5.4 Localization of $\mathcal{N}=4,2,2^{*}$ gauge theory on the 4-sphere In this section we review, following the seminal work of Pestun [11], how to exploit the supersymmetric localization principle in $\mathcal{N}=4$ Euclidean Super Yang-Mills theory on the four-sphere $\mathbb{S}^{4}$. The $\mathcal{N}=2$ and $\mathcal{N}=2^{*}$ theories can be also treated with the same technique. In particular, it was possible to solve exactly the partition function of the theory and the expectation value of the Wilson loop defined by $W_{R}(C):=\frac{1}{\dim R}\mathrm{Tr}_{R}\left(\mathcal{P}\exp{i\oint_{C(\mathbb{S}^{1})}(A_{\mu}\dot{C}^{\mu}+|\dot{C}|\Phi_{0})dt}\right)$ (5.85) where $C$ is a closed equatorial curve on $\mathbb{S}^{4}$ of tangent vector $\dot{C}$, and the scalar field $\Phi_{0}$ is required by supersymmetry, as will become clear later. The localization procedure makes the path integral reduce to a finite-dimensional integral over the Lie algebra of the gauge group, a so-called “matrix model”. #### 5.4.1 The action and the supersymmetric Wilson loop We will consider the theories revisited in Sections 4.3.7 and 4.4.4. We report the action of the $\mathcal{N}=2^{*}$ theory on the 4-sphere, $\displaystyle S^{\mathcal{N}=2^{*}}_{\mathbb{S}^{4}}=\int_{\mathbb{S}^{4}}d^{4}x\sqrt{g}\ \frac{1}{g_{YM}^{2}}\mathrm{Tr}\left(F_{MN}F^{MN}-\Psi\Gamma^{M}D_{M}\Psi\right.$ $\displaystyle+\frac{2}{r^{2}}\Phi_{A}\Phi^{A}-$ (5.86) $\displaystyle\left.-\frac{1}{4r}(R^{ki}M_{k}^{j})\Phi_{i}\Phi_{j}-\sum_{i=1}^{7}K_{i}K_{i}\right)$ where $D_{0}\Phi_{i}\mapsto[\Phi_{0},\Phi_{i}]+M_{i}^{j}\Phi_{j}$ and $D_{0}\Psi\mapsto[\Phi_{0},\Psi]+\frac{1}{4}M_{ij}\Gamma^{ij}\Psi$, $i,j=5,\cdots,8$. In the limit of zero mass $M$ we get the $\mathcal{N}=4$ YM theory, while in the limit of infinite mass the $\mathcal{N}=2$ hypermultiplet decouples and the pure $\mathcal{N}=2$ YM theory is recovered. This is invariant under the superconformal transformations (4.170) that we report here, $\displaystyle\delta_{\epsilon}A_{M}$ $\displaystyle=\epsilon\Gamma_{M}\Psi$ (5.87) $\displaystyle\delta_{\epsilon}\Psi$ $\displaystyle=\frac{1}{2}\Gamma^{MN}F_{MN}\epsilon+\frac{1}{2}\Gamma^{\mu A}\Psi_{A}\nabla_{\mu}\epsilon+\sum_{i=1}^{7}K_{i}\nu_{i}$ $\displaystyle\delta_{\epsilon}K_{i}$ $\displaystyle=-\nu_{i}\Gamma^{M}D_{M}\Psi$ with $(\nu_{i})_{i=1,\cdots,7}$ satisfying (4.171), and $\epsilon$ being a conformal Killing spinor satisfying (4.165) and (4.166). When the mass is non- zero, the Killing condition is restricted to (4.176), or equivalently $\tilde{\epsilon}=\frac{1}{2r}\Lambda\epsilon$ (5.88) where $\Lambda$ is an $SU(2)_{L}^{R}$ generator. The superconformal algebra closes schematically as $\delta_{\epsilon}^{2}=-\mathcal{L}_{v}-G_{\Phi}-(R+M)-\Omega.$ (5.89) To obtain a Poincaré-equivariant differential interpretation of this variation, we want $\delta_{\epsilon}$ to generate rigid supersymmetry, i.e. square only to the Poincaré algebra (plus R-symmetry, up to gauge transformations). Thus, to eliminate the dilatation contribution, we impose also the condition $\epsilon\tilde{\epsilon}=0.$ (5.90) If the mass is non-zero, the $SU(1,1)^{\mathcal{R}}$ is broken, so also its contribution should be eliminated, imposing further the condition $\tilde{\epsilon}\Gamma^{09}\epsilon=0.$ (5.91) Solutions to (4.165) and (4.166) are easy to compute in the flat space limit $r\to\infty$: here $\nabla_{\mu}=\partial_{\mu}$, and $\partial_{\mu}\tilde{\epsilon}=0$ imposes $\tilde{\epsilon}(x)=\hat{\epsilon}_{c}$ constant. Thus, the conformal Killing spinor in flat space is just the one considered in (4.116), $\epsilon(x)=\hat{\epsilon}_{s}+x^{\mu}\Gamma_{\mu}\hat{\epsilon}_{c}$ (5.92) where the first constant term generates supertranslations, while the term linear in $x$ generates superconformal transformations. The constant spinors $\hat{\epsilon}_{s},\hat{\epsilon}_{c}$ parametrize in general the space of solutions of the conformal Killing spinor equation. For a finite radius $r$, using stereographic coordinates and the round metric (4.163), the covariant derivative acts as $\nabla_{\mu}\epsilon=\left(\partial_{\mu}+\frac{1}{4}\omega_{ij\mu}\Gamma^{ij}\right)\epsilon$, where $\omega$ is the spin connection $\omega^{i}_{j\mu}=\left(e^{i}_{\mu}e^{\nu}_{j}-e_{j\mu}e^{i\nu}\right)\partial_{\nu}\Omega$ (5.93) and $e$ is the vielbein corresponding to the metric.141414Here we use latin indices as “flat” indices and greek indices as “curved” indices, so that as $g_{\mu\nu}=e^{i}_{\mu}e^{j}_{\nu}\delta_{ij}$. The general solution in this coordinate system is $\epsilon(x)=\frac{1}{\sqrt{1+\frac{x^{2}}{4r^{2}}}}\left(\hat{\epsilon}_{s}+x^{\mu}\Gamma_{\mu}\hat{\epsilon}_{c}\right)\qquad\tilde{\epsilon}(x)=\frac{1}{\sqrt{1+\frac{x^{2}}{4r^{2}}}}\left(\hat{\epsilon}_{c}-\frac{x^{\mu}\Gamma_{\mu}}{4r^{2}}\hat{\epsilon}_{s}\right)$ (5.94) that indeed simplifies to (5.92) in the limit of infinite radius. The conditions (5.90), (5.91), (5.88) are rewritten in terms of the constant spinors as $\hat{\epsilon}_{s}\hat{\epsilon}_{c}=\hat{\epsilon}_{s}\Gamma^{09}\hat{\epsilon}_{c}=0\qquad\hat{\epsilon}_{s}\Gamma^{\mu}\hat{\epsilon}_{s}=\frac{1}{4r^{2}}\hat{\epsilon}_{c}\Gamma^{M}\hat{\epsilon}_{c}\qquad\hat{\epsilon}_{c}=\frac{1}{2r}\Lambda\hat{\epsilon}_{s}.$ (5.95) The second condition is solved if the two constant spinors are taken to be chiral with respect to the 4-dimensional chirality operator $\Gamma^{1234}$, so that both terms vanish automatically. In Pestun’s conventions, they are chosen to have the same definite chirality and orthogonal to each other (to satisfy the first condition), so that $\epsilon$ is chiral only at the North and South poles, where $x^{2}=0,\infty$.151515There cannot be chiral spinor fields on $\mathbb{S}^{4}$ without zeros, because a chiral spinor defines an almost complex structure at each point, but $\mathbb{S}^{4}$ has no almost complex structure. Since $\mathbb{S}^{4}$ has constant scalar curvature, it can be proved that the conformal Killing condition on $\epsilon$ actually implies that $\epsilon$ is also a Killing spinor, $\nabla_{\mu}\epsilon=\mu\Gamma_{\mu}\epsilon$ for some constant $\mu$. This condition implies that the spinor is never zero, since it has constant norm. Thus $\epsilon$ cannot be chiral. [55] The Wilson loop under consideration is of the type considered in [87, 88], $W_{R}(C):=\frac{1}{\dim R}\mathrm{Tr}_{R}\left(\mathcal{P}\exp{i\oint_{C}dt(A_{\mu}\dot{C}^{\mu}+|\dot{C}|\Phi_{0})}\right)$ (5.96) where $C:[0,1]\to\mathbb{S}^{4}$ is an equatorial closed curve, parametrized in stereographic coordinates as $(x\circ C)(t)=2r(\cos{(t)},\sin{(t)},0,0)$, spanning a great circle of radius $r$. Its tangent vector is $\dot{C}(t)=2r(-\sin(t),\cos(t),0,0)$, and the normalization $|\dot{C}|=2r$ in front of $\Phi_{0}$ is needed for the reparametrization invariance of the line integral. We argue now that this Wilson loop preserves some supersymmetry under the action of $\delta_{\epsilon}$. In fact, its variation is proportional to $\delta_{\epsilon}W_{R}(C)\propto\epsilon\left(\Gamma_{\mu}\dot{C}^{\mu}+2r\Gamma_{0}\right)\Psi$ (5.97) and for this to vanish for every value of the gaugino $\Psi$, it must be that $\displaystyle 0=$ $\displaystyle\epsilon\left(\Gamma_{\mu}\dot{C}^{\mu}+2r\Gamma_{0}\right)\propto\left(\hat{\epsilon}_{s}+C^{\mu}\Gamma_{\mu}\hat{\epsilon}_{c}\right)\left(\Gamma_{\mu}\dot{C}^{\mu}+2r\Gamma_{0}\right)$ (5.98) $\displaystyle\Leftrightarrow 0=$ $\displaystyle\sin(t)\left(-\hat{\epsilon}_{s}\Gamma_{1}+2r\hat{\epsilon}_{c}\Gamma_{2}\Gamma_{0}\right)+\cos(t)\left(\hat{\epsilon}_{s}\Gamma_{2}+2r\hat{\epsilon}_{c}\Gamma_{1}\Gamma_{0}\right)+\left(2r\hat{\epsilon}_{c}\Gamma_{1}\Gamma_{2}+\hat{\epsilon}_{s}\Gamma_{0}\right)$ where we inserted the values for $C^{\mu},\dot{C}^{\mu}$ and simplified some trivial terms. For this to vanish at all $t$, the three parentheses have to vanish separately, giving the condition $\hat{\epsilon}_{c}=\frac{1}{2r}\Gamma_{0}\Gamma_{1}\Gamma_{2}\hat{\epsilon}_{s}.$ (5.99) This condition halves the number of spinors that preserve the Wilson loop under supersymmetry, so this is called a _1/2-BPS_ operator.161616This is just common terminology, that does not refer to any BPS condition between mass and central charges in the supersymmetry algebra (see [47]). It only means that the observable under consideration preserves half of the supercharges. In the $\mathcal{N}=4$ case, it preserves 16 supercharges. If the mass of the hypermultiplet is turned on, the third condition in (5.95) has a non-zero solution for $\hat{\epsilon}_{s}$ if $\det(\Lambda-\Gamma_{0}\Gamma_{1}\Gamma_{2})=0$, that fixes $\Lambda$ up to a sign. #### 5.4.2 Quick localization argument Without considering the unphysical redundancy in field space given by the gauge symmetry of the theory, we can give a quick argument for the localization of the $\mathcal{N}=4$ SYM, using the procedure outlined in Section 5.3. We consider the $U(1)$-equivariant cohomology generated by the action of a fixed supersymmetry $\delta_{\epsilon}$.171717 $\delta_{\epsilon}$ squares to the Poincaré algebra up to an gauge transformation, so we are really considering an $(U(1)\rtimes G)$-equivariant cohomology, because $\delta_{\epsilon}$-closed equivariant forms are supersymmetric and gauge invariant observables. Ignoring the gauge fixing procedure, we are really not considering the complete field space, since the BRST procedure teaches us that in presence of a gauge symmetry this is automatically extended to include ghosts, that may contribute to the localization locus. It turns out that ghosts contribution is trivial, so the rough argument already gives the correct localization locus. Here we continue with this simplified procedure, and in the next section we are going to argue the above claim. Since $\delta_{\epsilon}S^{\mathcal{N}=4}_{\mathbb{S}^{4}}=0$ is equivariantly closed (off-shell) with respect to the variations (5.87), we can perform the usual trick and add the localizing term $\displaystyle\lambda S_{loc}$ $\displaystyle:=\lambda\delta_{\epsilon}\mathcal{V}$ (5.100) $\displaystyle\mathrm{where}\quad\mathcal{V}$ $\displaystyle:=\mathrm{Tr}\left(\Psi\overline{\delta_{\epsilon}\Psi}\right)$ where $\overline{\delta_{\epsilon}\Psi}$ is defined by complex conjugation in the Euclidean signature, $\overline{\delta_{\epsilon}\Psi}=\frac{1}{2}\tilde{\Gamma}^{MN}F_{MN}\epsilon+\frac{1}{2}\Gamma^{\mu A}\Psi_{A}\nabla_{\mu}\epsilon-\sum_{i=1}^{7}K_{i}\nu_{i}.$ (5.101) The bosonic part of the localizing action is $\left.S_{loc}\right|_{bos}=\mathrm{Tr}\left(\delta_{\epsilon}\Psi\overline{\delta_{\epsilon}\Psi}\right)$ (5.102) that is positive semi-definite. The localization locus is then the subspace of fields such that $[\mathrm{fermions}]=0;\qquad\delta_{\epsilon}[\mathrm{fermions}]\ s.t.\ \left.S_{loc}\right|_{bos}=0.$ (5.103) The solution to (5.103) is found by inserting the relevant supersymmetry variation in $\left.S_{loc}\right|_{bos}$, collecting the terms as a sum of positive semi-definite contributions and requiring them to vanish separately. Under the assumption of smooth gauge field, this is given by the field configurations such that, up to a gauge transformation (see [11] for the details) $\left\\{\begin{array}[]{ll}A_{\mu}=0&\mu=1,\cdots,4\\\ \Phi_{i}=0&i=5,\cdots,9\\\ \Phi^{E}_{0}=a\in\mathfrak{g}&\mathrm{constant}\\\ K_{i}^{E}=-2(\nu_{i}\tilde{\epsilon})a&i=5,6,7\\\ K_{I}=0&I=1,\cdots,4\end{array}\right..$ (5.104) So the physical sector of the theory localizes onto the zero-modes of $\Phi_{0}$. If also singular gauge field configurations are allowed, (5.103) receives contributions from instanton solutions, where $F_{\mu\nu}=0$ everywhere except from the North or the South pole. These configurations can contribute non-trivially to the partition function. Computing the action on the smooth solutions one gets $S^{\mathcal{N}=4}_{\mathbb{S}^{4}}[a]=\frac{1}{g_{YM}^{2}}\int_{\mathbb{S}^{4}}d^{4}x\sqrt{g}\mathrm{Tr}\left(\frac{2}{r^{2}}(\Phi^{E}_{0})^{2}+(K_{i}^{E})^{2}\right)=\frac{1}{g_{YM}^{2}}\mbox{vol}(\mathbb{S}^{4})\frac{3}{r^{2}}\mathrm{Tr}\left(a^{2}\right)=\frac{8\pi^{2}r^{2}}{g_{YM}^{2}}\mathrm{Tr}(a^{2})$ (5.105) where we used $\mbox{vol}(\mathbb{S}^{4})=\frac{8}{3}\pi^{2}r^{4}$ and $(\nu_{i}\tilde{\epsilon})^{2}=\frac{1}{4r^{2}}$. This last equation can be derived from the conditions (4.171) and the form of the conformal Killing spinor $\epsilon$. The action is given by constant field contributions, so the path integral is expected to be reduced to a finite-dimensional integral over the Lie algebra $\mathfrak{g}$, of the form $Z\sim\int_{\mathfrak{g}}da\ e^{-\frac{8\pi^{2}r^{2}}{g_{YM}^{2}}\mathrm{Tr}(a^{2})}|Z_{inst}[a]|^{2}Z_{1-loop}[a].$ (5.106) Here $Z_{inst}$ is the instanton partition function, coming from the singular gauge field contributions of above. Since $G$ is often considered to be a matrix group, this partition function is said to describe a _matrix model_. The Wilson loop (5.96), evaluated on this locus is given by $W_{R}(C)=\frac{1}{\dim R}\mathrm{Tr}_{R}e^{2\pi ra}.$ (5.107) This type of matrix models can be approached by reducing the integration over $\mathfrak{g}$ to an integration over its Cartan subalgebra $\mathfrak{h}$ (we will discuss this better later, in Section 5.4.5).181818For finite-dimensional semi-simple complex Lie algebras, this is the maximal Abelian subalgebra. In general it is the maximal Lie subalgebra such that there exists a basis extension $\mathfrak{h}\oplus span(e_{\alpha})\cong\mathfrak{g}$, and it holds the eigenvalue equation $[h,e_{\alpha}]=\rho_{\alpha}(h)e_{\alpha}$ for any $h\in\mathfrak{h}$ and a certain eigenvalue $\rho_{\alpha}(h)$. $\rho_{\alpha}:\mathfrak{h}\to\mathbb{C}$ are called _roots_ of $\mathfrak{g}$. Assuming the zero-mode $a\in\mathfrak{h}$, we can conveniently rewrite the trace in the representation $R$ as the sum of all the _weights_ $\rho(a)$ of $a$ in $R$,191919Analogously to the definition of root in the adjoint representation of $\mathfrak{g}$, if the Lie algebra acts on the representation $R$, the _weight space_ $R_{\rho}$ of weight $\rho:\mathfrak{h}\to\mathbb{C}$ is defined as the subspace of elements $A\in R$ such that $h\cdot A=\rho(h)A$. $W_{R}(C)=\frac{1}{\dim R}\sum_{\rho\in\Omega(R)}n(\rho)e^{2\pi r\rho(a)},$ (5.108) where $n(\rho)$ is the multiplicity of the weight $\rho$, and $\Omega(R)$ is the set of all the weights in the representation $R$. We finally notice that the same result for the localization locus works also for the $\mathcal{N}=2$ and the $\mathcal{N}=2^{*}$ theories, and both theories localize to the same matrix model. If the mass term for the hypermultiplet is considered, this can of course give a non-trivial contribution to the 1-loop determinant. #### 5.4.3 The equivariant model As remarked at the end of Section 5.3, in presence of a gauge symmetry the path integral has to be defined with respect to a gauge-fixed action. To do so, one has to enlarge the field space to include the appropriate Faddeev- Popov ghosts in a BRST complex with the differential $\delta_{B}$. We consider then the total differential $Q=\delta_{\epsilon}+\delta_{B}$ (5.109) where $\epsilon$ is a fixed conformal Killing spinor that closes off-shell the superconformal algebra, so that the gauge-invariant SYM action is $Q$-closed. From the equivariant cohomology point of view, this operator is an equivariant differential with respect to the $U(1)_{\epsilon}\rtimes G$ symmetry group acting on the enlarged field space. To gauge fix the path integral, following the BRST procedure with respect to the differential $Q$, the action has to be extended as $S_{phys}[A,\Psi,K,ghosts]=S_{SYM}[A,\Psi,K]+Q\mathcal{O}_{g.f.}[A,ghosts]$ (5.110) with a gauge-fixing fermion $\mathcal{O}[A,ghosts]$. Upon path integration over ghosts, this new term has to give the gauge-fixing action and Fadee-Popov determinant. The localization principle is then exploited augmenting again the action with a $Q$-exact term, $\lambda Q\mathcal{V}$ with again $\mathcal{V}:=\mathrm{Tr}(\Psi\overline{\delta_{\epsilon}\Psi}).$ (5.111) This effectively gives the same localization term of the previous paragraph, since $\mathcal{V}$ is gauge-invariant. The BRST-like complex considered in [11] is given by the following ghost and auxiliary field extension. The ghost $c$, anti-ghost $\tilde{c}$ and standard Lagrange multiplier for the $R_{\xi}$-gauges $b$ (“Nakanishi-Lautrup” field) are introduced, respectively odd, odd and even with respect to Grassmann parity. Since the path integral is expected to localize on zero-modes, constant fields $c_{0},\tilde{c}_{0}$ (odd) and $a_{0},\tilde{a}_{0},b_{0}$ (even) are also introduced. On the original fields of the SYM theory, the BRST differential acts as a gauge transformation parametrized by $c$. On the gauge field $A_{\mu}$ $\delta_{B}A_{\mu}=-[c,D_{\mu}].$ (5.112) On ghosts and zero-modes the BRST transformation is defined by $\begin{array}[]{llll}\delta_{B}c=-a_{0}-\frac{1}{2}[c,c]&\delta_{B}\tilde{c}=b&\delta_{B}\tilde{a}_{0}=\tilde{c}_{0}&\delta_{B}b_{0}=c_{0}\\\ \delta_{B}a_{0}=0&\delta_{B}b=[a_{0},\tilde{c}]&\delta_{B}\tilde{c}_{0}=[a_{0},\tilde{a}_{0}]&\delta_{B}c_{0}=[a_{0},b_{0}]\end{array}$ (5.113) and its square generates a gauge transformation with respect to the (bosonic) constant field $a_{0}$, $\delta_{B}^{2}=[a_{0},\cdot].$ (5.114) The supersymmetry complex, constituted by the original fields, is reparametrized with respect to the basis $\\{\Gamma^{M}\epsilon,\nu^{i}\\}$, with $M=1,\cdots,9;i=1,\cdots,7$, of the 10-dimensional Majorana-Weyl bundle over $\mathbb{S}^{4}$. Expanding $\Psi$ over such a basis we have $\Psi=\sum_{M=1}^{9}\Psi_{M}(\Gamma^{M}\epsilon)+\sum_{i=1}^{7}\Upsilon_{i}\nu^{i}$ (5.115) and the superconformal transformations are rewritten as $\displaystyle\delta_{\epsilon}A_{M}=\Psi_{M}$ (5.116) $\displaystyle\delta_{\epsilon}\Psi_{M}=-(\mathcal{L}_{v}+R+G_{\Phi})A_{M}$ $\displaystyle\delta_{\epsilon}\Upsilon_{i}=H_{i}$ $\displaystyle\delta_{\epsilon}H_{i}=-(\mathcal{L}_{v}+R+G_{\Phi})\Upsilon_{i},$ where $H_{i}:=K_{i}+2(\nu_{i}\tilde{\epsilon})\Phi_{0}+\frac{1}{2}F_{MN}\nu_{i}\Gamma^{MN}\epsilon+\frac{1}{2}\Phi_{A}\nu_{i}\Gamma^{\mu A}\nabla_{\mu}\epsilon.$ (5.117) With this field redefinition, we see that the supersymmetry transformations can be schematized in the form $\delta_{\epsilon}X=X^{\prime}\qquad\delta_{\epsilon}X^{\prime}=[\phi+\epsilon,X]\qquad\left(\delta_{\epsilon}\phi=0\right)$ (5.118) where $\phi:=-\Phi=v^{M}A_{M}$, $[\phi,X^{\prime}]:=-G_{\Phi}X^{\prime}$ denotes a gauge transformation, $[\epsilon,X^{\prime}]:=-(\mathcal{L}_{v}+R)X^{\prime}$ denotes a Lorentz transformation. Here $X=(A_{M}(x),\Upsilon_{i}(x))$, $X^{\prime}=(\Psi_{M}(x),H_{i}(x))$ are the coordinates in the super-loop space interpretation of the supersymmetric model, of opposite statistics. As we pointed out in Section 5.3, we espect every Poincaré-supersymmetric theory to have such super-loop equivariant structure, and this is an example of the fact that in higher dimensional QFT the reparametrization of the fields necessary to make this apparent can be non-trivial. Indeed, the loop space coordinates $X$ and the corresponding 1-forms $X^{\prime}$ mix the bosonic/fermionic field components of the original parametrization! Combining the two complexes, and giving supersymmetry transformation properties to the ghost sector, the equivariant differential $Q$ is taken to act as $\begin{array}[]{lll}QX=X^{\prime}-[c,X]&Qc=\phi- a_{0}-\frac{1}{2}[c,c]&Q\tilde{c}=b\\\ QX^{\prime}=[\phi+\epsilon,X]-[c,X^{\prime}]&Q\phi=-[c,\phi+\epsilon]&Qb=[a_{0}+\epsilon,\tilde{c}]\\\ Q\tilde{a}_{0}=\tilde{c}_{0}&Qb_{0}=c_{0}&\\\ Q\tilde{c}_{0}=[a_{0},\tilde{c}_{0}]&Qc_{0}=[a_{0},b_{0}].&\end{array}$ (5.119) Moreover, $Qa_{0}=Q\epsilon=0$. This differential squares to a constant gauge transformation generated by $a_{0}$ and the Lorentz transformation generated by $\epsilon$, $Q^{2}=[a_{0}+\epsilon,\cdot].$ (5.120) Notice that to make explicit the super-loop structure when the combined complex is taken into account, one needs another non-trivial reparametrization of the fields, $\tilde{X}^{\prime}:=X^{\prime}-[c,X]\qquad\tilde{\phi}:=\phi- a_{0}-\frac{1}{2}[c,c].$ (5.121) This makes the tranformations look like $Q(\mbox{field})=\mbox{field}^{\prime}\qquad Q(\mbox{field}^{\prime})=[a_{0}+\epsilon,\mbox{field}]$ (5.122) and the new pairs of coordinate/1-form in the extended super-loop space are $(c,\tilde{\phi})$, $(\tilde{c},b)$, $(\tilde{a}_{0},\tilde{c}_{0})$, $(b_{0},c_{0})$. The gauge-fixing term considered for the extended quantum action is, schematically $S_{g.f.}=\int_{\mathbb{S}^{4}}Q\left(\tilde{c}\left(\nabla^{\mu}A_{\mu}+\frac{\xi_{1}}{2}b+b_{0}\right)-c\left(\tilde{a}_{0}-\frac{\xi_{2}}{2}a_{0}\right)\right)$ (5.123) where the bilinear product in $\mathfrak{g}$ is suppressed in the notation, assuming contraction of Lie algebra indices. Upon integration of the auxiliary field, this term produces the usual gauge fixing term for the Lorentz gauge $\nabla^{\mu}A_{\mu}=0$, and the ghost term of the action. Moreover, the path integral is independent of the parameters $\xi_{1},\xi_{2}$ (we refer to [11] for the proof). We finally claim that the localization principle for the gauge-fixed theory remains the same, with the additional condition of vanishing ghosts in the localization locus, and identifying the zero-mode of $\Phi_{0}$ with $a_{0}$. In fact from the gauge-fixing term, $S_{g.f.}\supset-\int_{\mathbb{S}^{4}}\left(\phi- a_{0}-\frac{1}{2}[c,c]\right)\tilde{a}_{0}$ (5.124) and integrating over $\tilde{a}_{0}$, we have the condition $\phi=a_{0}+\frac{1}{2}[c,c]$, that in the localization locus where $c=0$ and $\phi=-v^{M}A_{M}=\Phi_{0}$, becomes precisely $\Phi_{0}=a_{0}$. #### 5.4.4 Localization formulas We stated that the path integral localizes (apart from instanton corrections) to the zero-modes of the bosonic constant $a_{0}\in\mathfrak{g}$, that correspond to the zero-modes of $\Phi_{0}$. For the same reason of the scalar field corresponding to the reduced time-direction of the (9,1)-theory, we integrate over immaginary $a_{0}=ia_{0}^{E}$, where $a_{0}^{E}$ is real. The application of the localization principle is now straightforward in principle, although very cumbersome in practice. In particular, integrating out the Gaussian fluctuations around the localization locus, the arising one-loop determinant in the partition function results of the form $Z_{1-loop}=\left(\frac{\det{K_{f}}}{\det{K_{b}}}\right)^{1/2}$ (5.125) where $K_{f},K_{b}$ are the kinetic operators acting on the fermionic and bosonic fluctuation modes after the usual expansion of $Q\mathcal{V}$. This factor requires in general a regularization, and it has been computed for the $\mathcal{N}=2,\mathcal{N}=2^{*}$ and $\mathcal{N}=4$ theory, using an appropriate generalization of the Atiyah-Singer theorem seen in Section 5.2 applied to transversally elliptic operators. The instanton partition functions have been also simplified for the theories under consideration. We refer to [89, 11, 90] for the explicit form of instanton contributions in the cases of $\mathcal{N}=2,2^{*}$. For the maximally supersymmetric $\mathcal{N}=4$ SYM theory, the results for the 1-loop determinant and the instanton partition function are of the very simple form $Z_{1-loop}^{\mathcal{N}=4}=1,\qquad Z_{inst}^{\mathcal{N}=4}=1,$ (5.126) so that the resulting localization formulas for the partition function and the expectation value of the supersymmetric Wilson loop presented before become $\displaystyle Z_{\mathbb{S}^{4}}$ $\displaystyle=\frac{1}{\mbox{vol}(G)}\int_{\mathfrak{g}}da\ e^{-\frac{8\pi^{2}r^{2}}{g_{YM}^{2}}\mathrm{Tr}(a^{2})},$ (5.127) $\displaystyle\langle W_{R}(C)\rangle$ $\displaystyle=\frac{1}{\dim R}\frac{1}{Z\ \mbox{vol}(G)}\int_{\mathfrak{g}}da\ e^{-\frac{8\pi^{2}r^{2}}{g_{YM}^{2}}\mathrm{Tr}(a^{2})}\mathrm{Tr}_{R}\left(e^{2\pi ra}\right).$ This result proved a previous conjecture, based on a perturbative analysis by Erickson-Semenoff-Zarembo [91]. Their calculation for $\langle W_{R}(C)\rangle$ with $G=U(N)$ showed that the Feynman diagrams with internal vertices cancel up to order $g^{4}N^{2}$, and that the sum of all ladder diagrams (planar diagrams with no internal vertices) exponentiate to a matrix model. The result of this exponentiation gives an expectation value that coincides with the strong-coupling prediction of the AdS/CFT correspondence for $\mathcal{N}=4$ SYM,202020This “correspondence” conjectures a duality between the $\mathcal{N}=4$ SYM in 4 dimensions and type IIB superstring theory in an $AdS_{5}\times\mathbb{S}^{5}$ background. In particular, when the parameters of the gauge theory are taken to be such that $N\to\infty$ and $g_{YM}^{2}N\to\infty$ (namely, in the planar and strong ’t Hooft coupling limit), $\mathcal{N}=4$ SYM is dual to classical type IIB supergravity on $AdS_{5}\times\mathbb{S}^{5}$ and the computation of the Wilson loop in this limit is mapped to the evaluation of a minimal surface in this space [80, 92]. thus they conjectured that the diagrams with vertices have to vanish at all orders. Later this conjecture was supported by Drukker-Gross [93], and finally proven with the exact localization technique described above. We quote now the results for the 1-loop determinants in the $\mathcal{N}=2,2^{*}$ theories. For this, it is useful to introduce the notation $\displaystyle\mbox{det}_{R}f(a)$ $\displaystyle:=\prod_{\rho}f(\rho(a))$ (5.128) $\displaystyle H(z)$ $\displaystyle:=e^{-(1+\gamma)z^{2}}\prod_{n=1}^{\infty}\left(1-\frac{z^{2}}{n^{2}}\right)^{n}\prod_{n=1}^{\infty}e^{z^{2}/n}$ with $\rho$ running over the weights of $R$ (if $R=Ad$, the weights are the roots of $\mathfrak{g}$), $\gamma$ being the Euler-Mascheroni constant. Let also be $m^{2}:=\frac{1}{4}M_{ij}M^{ij}$, and recall that $m$ (as well as $a_{0}$) should take immaginary values. From [11] we have $\displaystyle Z_{1-loop}^{\mathcal{N}=2^{*}}[a_{0};M]$ $\displaystyle=\exp{\left(-r^{2}m^{2}\left((1+\gamma)-\sum_{n=1}^{\infty}\frac{1}{n}\right)\right)}\mbox{det}_{Ad}\left[\frac{H(ra_{0})}{\left[H(r(a_{0}+m))H(r(a_{0}-m))\right]^{-1/2}}\right],$ (5.129) $\displaystyle Z_{1-loop}^{\mathcal{N}=2,pure}[a_{0}]$ $\displaystyle=\mbox{det}_{Ad}H(ra_{0}),$ (5.130) $\displaystyle Z_{1-loop}^{\mathcal{N}=2,W}[a_{0}]$ $\displaystyle=\frac{\det_{Ad}H(ra_{0})}{\det_{W}H(ra_{0})},$ (5.131) where the first result is for the massive $\mathcal{N}=2^{*}$ theory, the second one is derived putting $m=0$ in the first line, and describes the pure $\mathcal{N}=2$ SYM, the third one is for the matter-coupled theory to a massles hypermultiplet in the representation $W$. Notice that the exponential prefactor in the first line diverges, but is independent of $a_{0}$, and thus simplifies in ratios during the computation of expectation values. Also, the third line holds literally if the ($a_{0}$-independent) divergent factors are the same for the vector and the hypermultiplet. #### 5.4.5 The Matrix Model for $\mathcal{N}=4$ SYM As an example, we include here an explicit computation for the Gaussian matrix model (5.127) in the case of $\mathcal{N}=4$ SYM [85, 94, 93]. We will take in particular the case of the compact matrix group $G=U(N)$ with the Wilson loop in the fundamental representation $R=\mathbf{N}$, but first we analyze generically how to simplify such an integration over the Lie algebra $\mathfrak{g}$. We normalize the invariant volume element $da$ on $\mathfrak{g}$ such that $\int_{\mathfrak{g}}da\ e^{-\frac{2}{\xi^{2}}\mathrm{Tr}(a^{2})}=\left(\frac{\xi^{2}\pi}{2}\right)^{\dim(G)/2}$ (5.132) for any parameter $\xi$. In the $U(N)$ case, $\mathfrak{g}=\mathfrak{u}(N)=\\{\text{Hermitian}\ N\times N\ \text{matrices}\\}$, so this means taking $da=2^{N(N-1)/2}\prod_{i=1}^{N}da_{ii}\prod_{1\leq j<i\leq N}d\text{Re}(a_{ij})d\text{Im}(a_{ij}).$ (5.133) Setting $\xi^{2}:=g_{YM}^{2}/(4\pi^{2}r^{2})$ we have $Z_{\mathbb{S}^{4}}=\frac{1}{\mbox{vol}(G)}\left(\frac{g_{YM}^{2}}{8\pi r^{2}}\right)^{\dim(G)/2}.$ (5.134) To simplify the integration of the Wilson loop expectation value, we notice that the matrix model has a leftover gauge symmetry under constant gauge transformations, since both the measure and the traces are invariant under the adjoint action of $G$. Thus we can “gauge-fix” the integrand to depend only on the Cartan subalgebra $\mathfrak{h}\subset\mathfrak{g}$, setting $a=Ad_{g*}(X)$ (5.135) for some $g\in G/T$ and $X\in\mathfrak{h}$, with $T$ being the maximal torus in $G$ generated by $\mathfrak{h}$. There is more than one $X$ related to $a$ by conjugation, but they are related via the action of the _Weyl group_ of $G$, that we call $\mathcal{W}$. Taking this into account, after the gauge- fixing we can perform the integral over the orbits obtaining a volume factor $\frac{\mbox{vol}(G/T)}{|\mathcal{W}|}.$ (5.136) The gauge-fixing can be done with the usual Faddeev-Popov (FP) procedure, that is inserting the unity decomposition $1=\int dg\ \Delta^{2}(X)\delta(F(a^{(g)})),$ (5.137) where the delta-function fixes the condition (5.135), and the FP determinant is given by $\Delta(X)^{2}=\prod_{\alpha}|\alpha(X)|=\prod_{\alpha>0}\alpha(X),$ (5.138) where $\alpha:\mathfrak{h}\to\mathbb{C}$ are the roots of $\mathfrak{g}$, and in the second equality we used that roots come in pairs $(\alpha,-\alpha)$. We can rewrite the expectation value of the circular Wilson loop as $\displaystyle\langle W_{R}(C)\rangle$ $\displaystyle=\left(\frac{\xi^{2}\pi}{2}\right)^{\dim(G)/2}\frac{\mbox{vol}(G/T)}{|\mathcal{W}|\dim R}\int_{\mathfrak{h}}dX\ \Delta(X)^{2}e^{-\frac{2}{\xi^{2}}\mathrm{Tr}(X^{2})}\mathrm{Tr}_{R}\left(e^{2\pi rX}\right)$ (5.139) $\displaystyle=\left(\frac{\xi^{2}\pi}{2}\right)^{\dim(G)/2}\frac{\mbox{vol}(G/T)}{|\mathcal{W}|\dim R}\sum_{\rho\in\Omega(R)}n(\rho)\int_{\mathfrak{h}}dX\ \Delta(X)^{2}e^{-\frac{2}{\xi^{2}}\mathrm{Tr}(X^{2})}e^{2\pi r\rho(X)}$ where in the second line we used (5.108). We specialize now to the case $G=U(N)$, $\mathfrak{h}=\\{X=\mathrm{diag}(\lambda_{1},\cdots,\lambda_{N})|\lambda_{i}\in\mathbb{R}\\}$, and we take $r=1$. The adjoint action of $U(N)$ is the conjugation $a\mapsto gag^{\dagger}$, so the FP determinant is defined by $1=\int dg\ \Delta(X)^{2}\prod_{ij}\delta((gXg^{\dagger})_{ij}),$ (5.140) that imposes the off-diagonal terms to vanish in the given gauge. Expressing $g=e^{M}$ with $M\in\mathfrak{u}(N)$, $\Delta(X)^{2}=\prod_{ij}\det_{kl}\left|\frac{\delta(e^{M}Xe^{-M})_{ij}}{\delta M_{kl}}\right|=\prod_{ij}\det_{kl}\left|\delta_{ki}\delta_{lj}(\lambda_{j}-\lambda_{i})\right|=\prod_{i>j}(\lambda_{i}-\lambda_{j})^{2}$ (5.141) is the so called _Vandermonde determinant_ , that can be related to the following matrix $\Delta(\lambda)=\det||\lambda_{i}^{j-1}||=\det\left(\begin{array}[]{ccccc}1&\lambda_{1}&\lambda_{1}^{2}&\cdots&\lambda_{1}^{N-1}\\\ 1&\lambda_{2}&\lambda_{2}^{2}&\cdots&\lambda_{2}^{N-1}\\\ &&\vdots&&\\\ 1&\lambda_{N}&\lambda_{N}^{2}&\cdots&\lambda_{N}^{N-1}\\\ \end{array}\right).$ (5.142) The partition function can thus be expressed as $Z_{\mathbb{S}^{4}}=\frac{1}{N!}\frac{1}{(2\pi)^{N}}\int\left(\prod_{i}d\lambda_{i}\right)\left(\prod_{i>j}(\lambda_{i}-\lambda_{j})^{2}\right)e^{-\frac{2}{\xi^{2}}\sum_{i}\lambda_{i}^{2}},$ (5.143) where $N!$ is the order of the Weyl group $\mathcal{W}=S_{N}$ and $(2\pi)^{N}$ is the volume of the $N$-torus $U(1)^{N}$, while the Wilson loop in the fundamental representation inserts in the path integral a factor $\frac{1}{N}\sum_{j=1}^{N}e^{2\pi\lambda_{j}}.$ (5.144) There are two main approaches to the evaluation of this matrix model and the computation of the Wilson loop expectation value, at least in the limit $N\to\infty$. ##### $1^{st}$ method: saddle-point The first method that we present is based on a suitable saddle-point approximation in the large-$N$ limit. To see the possibility for this interpretation, we rewrite the partition function as $\displaystyle Z_{\mathbb{S}^{4}}$ $\displaystyle=\frac{1}{N!}\int\prod_{i}\frac{d\lambda_{i}}{2\pi}\ e^{-N^{2}S_{eff}(\lambda)}$ (5.145) $\displaystyle\text{with}\quad S_{eff}(\lambda)$ $\displaystyle:=\frac{8\pi^{2}}{tN}\sum_{i=1}^{N}\lambda_{i}^{2}-\frac{2}{N^{2}}\sum_{i>j}\log|\lambda_{i}-\lambda_{j}|,$ where $t:=g_{YM}^{2}N$ is the ’t Hooft coupling constant. This can be viewed as an effective action of a zero-dimensional QFT describing $N$ sites (the eigenvalues $\lambda_{i}$), where the first piece is a “one-body” harmonic potential, and the second one is a repulsive “two-body” interaction. Notice that every sum is roughly of order $\sim N$, so $S_{eff}\sim O(1)$ in $N$. The limit $N\to\infty$, with $t$ fixed, can be regarded as a semi-classical approximation (we could compare it to “$1/\sqrt{\hbar}\to\infty$”), and in that limit we can solve the integral using a saddle-point approximation. The saddle points are those values of $\lambda_{i}$ that solve the classical EoM $0=\frac{\delta S_{eff}}{\delta\lambda_{i}}\qquad\Rightarrow\qquad 0=\frac{16\pi^{2}}{tN}\lambda_{i}-\frac{2}{N^{2}}\sum_{j\neq i}\frac{1}{\lambda_{i}-\lambda_{j}}.$ (5.146) In the large-$N$ limit we can study this equation in the continuum approximation, assuming the eigenvalues $\lambda_{i}$ to take values in a compact interval $I=[a,b]$, so that the (normalized) _eigenvalue distribution_ $\rho(\lambda)=\frac{1}{N}\sum_{i=1}^{N}\delta(\lambda-\lambda_{i})$ (5.147) is regarded as a continuous function of compact support on $I$. Then every sum can be replaced by an integration over the reals, $\frac{1}{N}\sum_{i=1}^{N}f(\lambda_{i})\to\int d\lambda\ f(\lambda)\rho(\lambda),$ (5.148) and (5.146) becomes $\frac{8\pi^{2}}{t}\lambda=\mathcal{P}\int\frac{\rho(\lambda^{\prime})d\lambda^{\prime}}{\lambda-\lambda^{\prime}},$ (5.149) where we took the principal value of the integral to avoid the pole at $\lambda_{i}=\lambda_{j}$. This is an integral equation in $\rho(\lambda)$, whose solution gives the distribution of the eigenvalues at the saddle-point locus of the partition function. It is useful to introduce an auxiliary function on the complex plane, the “resolvent” $\omega(z):=\int\frac{\rho(\lambda)d\lambda}{z-\lambda},$ (5.150) that has three important properties for our purposes: 1. (i) it is analytic on $\mathbb{C}\setminus I$, since there are poles for $z=\lambda$ when $z\in I$; 2. (ii) thanks to the normalization of $\rho$, asymptotically for $|z|\to\infty$ it goes as $\omega(z)\sim\frac{1}{z}$; 3. (iii) using the residue theorem and the delta-function representation $\frac{\epsilon}{z^{2}+\epsilon^{2}}\xrightarrow{\epsilon\to 0^{+}}\pi\delta(z),$ (5.151) it relates to the eigenvalue distribution by the discontinuity equation $\rho(\lambda)=-\frac{1}{2\pi i}\lim_{\epsilon\to 0^{+}}\left[\omega(\lambda+i\epsilon)-\omega(\lambda-i\epsilon)\right].$ (5.152) Knowing the resolvent we can easily compute the eigenvalue distribution by this last property, so we rewrite the saddle-point equation in terms of it. To compute $\omega$, we can start again from (5.149), multiply by $1/(\lambda-z)$ and integrate over $\lambda$ with the usual measure $\rho(\lambda)d\lambda$: $\frac{8\pi^{2}}{t}\int d\lambda\ \rho(\lambda)\frac{\lambda}{\lambda-z}=\int d\lambda\frac{\rho(\lambda)}{\lambda-z}\ \mathcal{P}\int d\lambda^{\prime}\frac{\rho(\lambda^{\prime})}{\lambda-\lambda^{\prime}}.$ (5.153) We can add $\pm z$ at the numerator of the LHS, and use the formula (_Sokhotski–Plemelj theorem_) $\mathcal{P}\int\frac{f(z)}{z}dz=\lim_{\epsilon\to 0^{+}}\frac{1}{2}\left(\int\frac{f(z)}{z+i\epsilon}dz+\int\frac{f(z)}{z-i\epsilon}dz\right)$ (5.154) to break the principal value on the RHS. Inserting the definition of the resolvent and using the residue theorem, this gives $\frac{8\pi^{2}}{t}-\frac{8\pi^{2}}{t}\lambda\omega(\lambda)=-\frac{1}{2}\omega(\lambda)^{2},$ (5.155) that is solved for $\omega(\lambda)=\frac{8\pi^{2}}{t}\left(\lambda\pm\sqrt{\lambda^{2}-\frac{t}{4\pi^{2}}}\right).$ (5.156) In order to match the right asymptotic behavior $\omega(z\to\infty)\sim 1/z$, we have to chose the minus sign. With this choice, we can compute the saddle- point eigenvalue distribution using the discontinuity equation (5.152), $\displaystyle\rho(\lambda)$ $\displaystyle=-\frac{1}{2\pi i}\frac{8\pi^{2}}{t}\lim_{\epsilon\to 0^{+}}\left[\omega(\lambda+i\epsilon)-\omega(\lambda-i\epsilon)\right]$ (5.157) $\displaystyle=\frac{4\pi}{it}\lim_{\epsilon\to 0^{+}}\left[\sqrt{\lambda^{2}-\frac{t}{4\pi^{2}}+2i\epsilon\lambda}-\sqrt{\lambda^{2}-\frac{t}{4\pi^{2}}-2i\epsilon\lambda}\right]$ $\displaystyle=\frac{4\pi}{it}\left(2\sqrt{\lambda^{2}-\frac{t}{4\pi^{2}}}\right)$ $\displaystyle=\frac{8\pi}{t}\sqrt{\frac{t}{4\pi^{2}}-\lambda^{2}}$ where we used that the principal square root has a branch cut on the real line. This function is called _Wigner semi-circle distribution_ , it has support on the interval $I=[-\sqrt{t}/2\pi,\sqrt{t}/2\pi]$, and here it is correctly normalized to 1. Now that we have the saddle-point locus in terms of the eigenvalue distribution, we can compute the expectation value for the circular Wilson loop in the fundamental representation. Since the exponential factor (5.144) is of order $\sim N^{0}$, this does not contribute to the saddle-point equation in the $N\to\infty$ limit. We can thus still use the Wigner distribution at zero-order in $1/N^{2}$, and insert in the path integral the trace in the continuum limit, $\displaystyle\langle W_{\mathbf{N}}(C)\rangle$ $\displaystyle=\int d\lambda\ \langle\rho(\lambda)\rangle e^{2\pi\lambda}$ (5.158) $\displaystyle=\frac{8\pi}{t}\int_{-\sqrt{t}/2\pi}^{\sqrt{t}/2\pi}d\lambda\ e^{2\pi\lambda}\sqrt{\frac{t}{4\pi^{2}}-\lambda^{2}}+O\left(1/N^{2}\right)$ $\displaystyle=\frac{2}{\sqrt{t}}I_{1}\left(\sqrt{t}\right)+O\left(1/N^{2}\right)$ where $I_{1}(z)$ is a modified Bessel function of the first kind. In the weak and strong coupling limits $t\gg,\ll 1$ the expectation value gives $\displaystyle t\ll 1:\qquad\langle W_{\mathbf{N}}(C)\rangle$ $\displaystyle\sim 1+\frac{t^{2}}{8}+\frac{t^{4}}{192}+\cdots$ (5.159) $\displaystyle t\gg 1:\qquad\langle W_{\mathbf{N}}(C)\rangle$ $\displaystyle\sim\sqrt{\frac{2}{\pi}}t^{-3/4}e^{\sqrt{t}},$ (5.160) so it explodes in the strong coupling limit, with an _essential_ singularity.212121Interestingly, the strong coupling limit can be checked independently using holography, where Wilson loops are given by minimal surfaces in AdS [80, 92]. ##### $2^{nd}$ method: orthogonal polynomials Another technique to solve matrix models involve the use of orthogonal polynomials [93]. Our starting point is again the partition function, $Z=\frac{1}{N!}\int\prod_{i=1}^{N}\left(\frac{d\lambda_{i}}{2\pi}e^{-\frac{8\pi^{2}N}{t}\lambda_{i}^{2}}\right)\Delta(\lambda)^{2}.$ (5.161) Introducing the $L^{2}(\mathbb{R})$ measure $d\mu(x):=dx\ e^{-\frac{8\pi^{2}N}{t}x^{2}},$ (5.162) we can write the partition function as $Z=\frac{1}{N!}\int\prod_{i=1}^{N}d\mu(\lambda_{i})\Delta(\lambda)^{2}.$ (5.163) Recalling that the Vandermonde determinant is evaluated from the matrix (5.142), expressed in terms of the polynomials $\\{1,x,x^{2},\cdots\\}$, we notice that we can equivalently express it in terms of another set of _monic_ polynomials, $p_{k}(x)=x^{k}+\sum_{j=0}^{k-1}a_{j}^{(k)}x^{j}$ (5.164) since by elementary row operations $\Delta(\lambda)=\det||\lambda_{i}^{j-1}||=\det||p_{j-1}(\lambda_{i})||.$ (5.165) It is useful to chose the set $\\{p_{k}\\}_{k\geq 0}$ to be _orthogonal_ with respect to the matrix model measure, $\int d\mu(\lambda)p_{n}(\lambda)p_{m}(\lambda)=h_{n}\delta_{nm}$ (5.166) since the knowledge of this set, and in particular of the normalization constants $h_{n}$, allows to compute the partition function. Writing the determinant as $\Delta(\lambda)=\sum_{\sigma\in S_{N}}(-1)^{\mbox{sign}(\sigma)}\prod_{k=1}^{N}p_{\sigma(k)-1}(\lambda_{k}),$ then (5.163) reduces to $Z=\prod_{k=0}^{N-1}h_{k}.$ (5.167) In our case the matrix model is Gaussian, and the corresponding set of orthogonal polynomials are the _Hermite polynomials_ , $H_{n}(x):=e^{x^{2}}\left(-\frac{d}{dx}\right)^{n}e^{-x^{2}},\qquad\int_{-\infty}^{+\infty}dx\ e^{-x^{2}}H_{n}(x)H_{m}(x)=\delta_{nm}2^{n}n!\sqrt{\pi}$ (5.168) so, normalizing $h_{n}=1$ and inserting the correct prefactors, we consider the set of _orthonormal_ polynomials with respect to the measure $d\mu(\lambda)$ $P_{n}(\lambda):=\sqrt{\sqrt{\frac{8\pi N}{t}}\frac{1}{2^{n}n!}}\ H_{n}\left(\frac{\sqrt{8\pi^{2}N}}{t}\lambda\right).$ (5.169) The expectation value of any observable of the type $\mathrm{Tr}(f(X))=\sum_{k}f(\lambda_{k})$ can be simplified as $\displaystyle\langle\mathrm{Tr}f(X)\rangle$ $\displaystyle=\frac{1}{N!Z}\int\left(\prod_{i=1}^{N}d\mu(\lambda_{i})\right)\Delta(\lambda)^{2}\sum_{k=1}^{N}f(\lambda_{k})$ (5.170) $\displaystyle\begin{aligned} =\frac{1}{N!}\sum_{k}\sum_{\sigma\in S_{N}}\int d\mu(\lambda_{1})\ P_{\sigma(1)-1}(\lambda_{1})^{2}\cdots\int d\mu(\lambda_{k})\ P_{\sigma(k)-1}(\lambda_{k})^{2}f(\lambda_{k})\cdots&\\\ \cdots\int d\mu(\lambda_{N})\ P_{\sigma(N)-1}(\lambda_{N})^{2}&\end{aligned}$ $\displaystyle=\sum_{j=0}^{N-1}\int d\mu(\lambda)\ P_{j}(\lambda)^{2}f(\lambda).$ Applying this formula to the expectation value of the circular Wilson loop in the fundamental representation we have $\displaystyle\langle W_{\mathbf{N}}(C)\rangle$ $\displaystyle=\frac{1}{N}\left\langle\mathrm{Tr}\exp(2\pi X)\right\rangle$ (5.171) $\displaystyle=\frac{1}{N}\sum_{j=0}^{N-1}\int d\lambda\ P_{j}(\lambda)^{2}e^{-\frac{8\pi^{2}N}{t}\lambda^{2}+2\pi\lambda}.$ A useful formula to simplify this integral is $\int_{-\infty}^{+\infty}dx\ H_{n}(x)^{2}e^{-(x-c)^{2}}=2^{n}n!\sqrt{\pi}L_{n}(-2c^{2})$ (5.172) where $c$ is a constant and $L_{n}(x)$ are the _Laguerre polynomials_ , satisfying the properties $\displaystyle L_{n}^{(m)}(x)$ $\displaystyle=\frac{1}{n!}e^{x}x^{m}\left(\frac{d}{dx}\right)^{n}\left(e^{-x}x^{n+m}\right),$ (5.173) $\displaystyle L_{n}(x)$ $\displaystyle\equiv L_{n}^{(0)}(x),$ (5.174) $\displaystyle L_{n}^{(m+1)}(x)$ $\displaystyle=\sum_{j=0}^{n}L_{j}^{(m)}(x),$ (5.175) $\displaystyle L_{n}^{(m)}(x)$ $\displaystyle=\sum_{k=0}^{n}\binom{n+m}{n-k}\frac{(-x)^{k}}{k!}.$ (5.176) Substituting (5.172) in (5.171), and expanding in series we have $\displaystyle\langle W_{\mathbf{N}}(C)\rangle$ $\displaystyle=\frac{1}{N}e^{c^{2}}L_{N-1}^{(1)}(-2c^{2})\qquad\text{with}\ c:=\sqrt{\frac{t}{8N}}$ (5.177) $\displaystyle\begin{aligned} =&\frac{1}{N}\sum_{k=0}^{\infty}\frac{1}{k!}\left(\frac{t}{8N}\right)^{k}\sum_{j=0}^{N-1}\frac{N!}{(j+1)!(N-1-j)!}\frac{1}{j!}\left(\frac{t}{4N}\right)^{j}-\\\ &-\frac{1}{N}\sum_{j=1}^{N-1}\frac{1}{j!(j+1)!}\left(\frac{t}{4}\right)^{j}\frac{j(j+1)}{2}+\frac{1}{N}\sum_{j=0}^{N-1}\frac{1}{j!(j+1)!}\left(\frac{t}{4}\right)^{j}\frac{1}{2}+O(1/N^{2})\end{aligned}$ $\displaystyle=\sum_{j=0}^{N-1}\frac{1}{j!(j+1)!}\left(\frac{t}{4}\right)^{j}+O\left(1/N^{2}\right)$ where we expanded the first terms with respect to powers of $1/N$, and already noticed that for $N\gg 1$ the odd-power terms cancel. We can thus examine the large-$N$ limit, and inserting the definition of the modified Bessel function $I_{n}(2x)=\sum_{k=0}^{\infty}\frac{x^{n+2k}}{k!(n+k)!}$ the expectation value gives $\langle W_{\mathbf{N}}(C)\rangle=\frac{2}{\sqrt{t}}I_{1}\left(\sqrt{t}\right)+O(1/N^{2}),$ (5.178) matching the result obtained with the saddle point technique in (5.158). In general, the expansion is in powers of $1/N^{2}$ rather than $1/N$, as expected from the analogy “$N^{2}\leftrightarrow 1/\hbar$” that we noticed in (5.145). Solutions to the matrix model for higher representations have also been found, see [85, 95, 96]. ### 5.5 Localization of $\mathcal{N}=2$ Chern-Simons theory on the 3-sphere In this section we review another example of supersymmetric localization applied to the computation of Wilson loop expectation values, in an $\mathcal{N}=2$ matter-coupled Euclidean Super Chern-Simons (SCS) theory on the 3-sphere $\mathbb{S}^{3}$. We follow the derivation of Kapustin-Willet- Yaakov [49], and Mariño [50], inspired in part by the work discussed in the previous section. We consider a generic compact Lie group $G$ as the gauge group, with Lie algebra $\mathfrak{g}$. #### 5.5.1 Matter-coupled $\mathcal{N}=2$ Euclidean SCS theory on $\mathbb{S}^{3}$ The case of $\mathcal{N}=2$ Euclidean supersymmetry on $\mathbb{S}^{3}$ was discussed as an example in Sections 4.3.6 and 4.4.3 for the gauge sector. We report the action for the SCS theory $S_{CS}=\frac{k}{4\pi}\int_{\mathbb{S}^{3}}d^{3}x\sqrt{g}\ \mathrm{Tr}\left\\{\frac{\varepsilon^{\mu\nu\rho}}{\sqrt{g}}\left(A_{\mu}\partial_{\nu}A_{\rho}+\frac{2i}{3}A_{\mu}A_{\nu}A_{\rho}\right)-\tilde{\lambda}\lambda+2\sigma D\right\\}$ (5.179) and the supersymmetry variations, already considered in curved space $\displaystyle\delta A_{\mu}=\frac{i}{2}(\tilde{\epsilon}\gamma_{\mu}\lambda-\tilde{\lambda}\gamma_{\mu}\epsilon)$ (5.180) $\displaystyle\delta\sigma=\frac{1}{2}(\tilde{\epsilon}\lambda-\tilde{\lambda}\epsilon)$ $\displaystyle\delta\lambda=\left(-\frac{1}{2}F_{\mu\nu}\gamma^{\mu\nu}-D+i(D_{\mu}\sigma)\gamma^{\mu}+\frac{2i}{3}\sigma\gamma^{\mu}D_{\mu}\right)\epsilon$ $\displaystyle\delta\tilde{\lambda}=\left(-\frac{1}{2}F_{\mu\nu}\gamma^{\mu\nu}+D-i(D_{\mu}\sigma)\gamma^{\mu}-\frac{2i}{3}\sigma\gamma^{\mu}D_{\mu}\right)\tilde{\epsilon}$ $\displaystyle\delta D=-\frac{i}{2}\left(\tilde{\epsilon}\gamma^{\mu}D_{\mu}\lambda-(D_{\mu}\tilde{\lambda})\gamma^{\mu}\epsilon\right)+\frac{i}{2}\left([\tilde{\epsilon}\lambda,\sigma]-[\tilde{\lambda}\epsilon,\sigma]\right)-\frac{i}{6}\left(\tilde{\lambda}\gamma^{\mu}D_{\mu}\epsilon+(D_{\mu}\tilde{\epsilon})\gamma^{\mu}\lambda\right).$ where the $D_{\mu}$ are gauge-covariant derivatives with respect to the metric and spin connection induced by the round metric (4.154), that in stereographic coordinates $x^{\mu=1,2,3}$ is given by $g_{\mu\nu}=e^{2\Omega(x)}\delta_{\mu\nu}\qquad e^{2\Omega(x)}=\left(1+\frac{x^{2}}{4r^{2}}\right)^{-2}$ (5.181) with $r$ being the radius of the embedding $\mathbb{S}^{3}\hookrightarrow\mathbb{R}^{4}$. We remark again that this supersymmetric action is actually superconformal, thus can preserve supersymmetry on this conformally flat background, even with positive scalar curvature. The new background preserves all the original $\mathcal{N}=2$ algebra, generated by conformal Killing spinors $\epsilon,\tilde{\epsilon}$, taken to satisfy $\nabla_{\mu}\epsilon=\frac{i}{2r}\gamma_{\mu}\epsilon,\qquad\nabla_{\mu}\tilde{\epsilon}=\frac{i}{2r}\gamma_{\mu},\tilde{\epsilon}$ (5.182) where every equation has two possible solutions. We consider also coupling the theory to matter fields, adding them in chiral multiplets in a representation $R$ of the gauge group, to preserve supersymmetry. The 3-dimensional $\mathcal{N}=2$ chiral multiplet (or hypermultiplet) is, as for the gauge multiplet, given by dimensional reduction of the $\mathcal{N}=1$ chiral multiplet in 4 dimensions: a complex scalar $\phi$, a 2-component Dirac spinor222222Recall that $Spin(3)=SU(2)$ has no Majorana spinors. We consider the reduced 4-dimensional Majorana spinor $\psi$ as a 3-dimensional Dirac (complex) spinor, since they have the same number of real components. $\psi$ and an auxiliary complex scalar $F$. Every field comes with its complex conjugate from the corresponding anti-chiral multiplet. The supersymmetric action for the matter multiplet coupled to the gauge multiplet is given by $S_{m}=\int_{\mathbb{S}^{3}}d^{3}x\sqrt{g}\ \left(D_{\mu}\tilde{\phi}D^{\mu}\phi+\frac{3}{4r^{2}}\tilde{\phi}\phi+i\tilde{\psi}\not{D}\psi+\tilde{F}F+\tilde{\phi}\sigma^{2}\phi+i\tilde{\phi}D\phi+i\tilde{\psi}\sigma\psi+i\tilde{\phi}\tilde{\lambda}\psi-i\tilde{\psi}\lambda\phi\right)$ (5.183) where the $\mathfrak{g}$-valued fields in the gauge multiplets act on the chiral multiplet in the representation $R$. This is the “covariantization” of the flat space action for the matter multiplet (see for example [97]), with the addition of the conformal coupling of the scalar field to the curvature, $\frac{3}{4r^{2}}\tilde{\phi}\phi$. The supersymmetry transformations for the chiral multiplet, with respect to the conformal Killing spinors $\epsilon,\tilde{\epsilon}$, are $\displaystyle\delta\phi$ $\displaystyle=\tilde{\epsilon}\psi\qquad\delta\tilde{\phi}=\tilde{\psi}\epsilon$ (5.184) $\displaystyle\delta\psi$ $\displaystyle=(-i\gamma^{\mu}D_{\mu}\phi-i\sigma\phi)\epsilon-\frac{i}{3}\gamma^{\mu}(\nabla_{\mu}\epsilon)\phi+\tilde{\epsilon}F$ $\displaystyle\delta\tilde{\psi}$ $\displaystyle=\tilde{\epsilon}(i\gamma^{\mu}D_{\mu}\tilde{\phi}+i\sigma\tilde{\phi})+\frac{i}{3}(\nabla_{\mu}\tilde{\epsilon})\gamma^{\mu}\tilde{\phi}+\epsilon\tilde{F}$ $\displaystyle\delta F$ $\displaystyle=\epsilon(-i\gamma^{\mu}D_{\mu}\psi+i\lambda\phi+i\sigma\psi)$ $\displaystyle\delta\tilde{F}$ $\displaystyle=(iD_{\mu}\tilde{\psi}\gamma^{\mu}-i\tilde{\lambda}\tilde{\phi}+i\sigma\tilde{\psi})\tilde{\epsilon}.$ The above variations generates a superconformal algebra that closes off-shell: $[\delta_{\epsilon},\delta_{\tilde{\epsilon}}]=-i(\mathcal{L}_{v}+G_{\Lambda}+R_{\alpha}+\Omega_{f})$ (5.185) where $\mathcal{L}_{v}$ is the Lie derivative (translation) along the Killing vector field $v=(\tilde{\epsilon}\gamma^{\mu}\epsilon)\partial_{\mu}$, acting on one forms as $\mathcal{L}_{v}(A)_{\mu}=v^{\nu}\partial_{\nu}A_{\mu}+A_{\nu}\partial_{\mu}v^{\nu}$, and on spinors as $\mathcal{L}_{v}\psi=\nabla_{\nu}\psi-\frac{1}{4}(\nabla_{\mu}v_{\nu})\gamma^{\mu\nu}\psi$. $G_{\Lambda}$ is a gauge transformation with respect to the parameter $\Lambda:=A(v)+\sigma(\tilde{\epsilon}\epsilon)$. $R_{\alpha}$ is a $U(1)^{\mathcal{R}}$ R-symmetry transformation, and $\Omega_{f}$ is a dilatation [50]. The matter coupled action $S_{CS}+S_{m}$ is known to be superconformal at quantum level, but one could also add a superpotential for the matter multiplet. This choice is restricted by the condition of unbroken superconformal symmetry both at classical and at quantum level, since the localization principle works only if the supersymmetry algebra closes off- shell. It turns out that the localization locus is at trivial configurations of the matter sector, thus the precise choice of superpotential does not influence the computation. #### 5.5.2 The supersymmetric Wilson loop The Wilson loop under consideration, in the representation $R$ of the gauge group, is defined as [97] $W_{R}(C)=\frac{1}{\dim{R}}\mathrm{Tr}_{R}\left(\mathcal{P}\exp{\oint_{C}dt\ (iA_{\mu}\dot{C}^{\mu}+\sigma)}\right)$ (5.186) with $C:\mathbb{S}^{1}\to\mathbb{S}^{3}$ a closed curve of tangent vector $\dot{C}$, normalized such that $|\dot{C}|=1$. In order to localize its expectation value, we have to consider those curves such that this operator preserves some supersymmetry on the 3-sphere. Its variation under (5.180) is proportional to $\delta W_{R}(C)\propto-\tilde{\epsilon}(\gamma_{\mu}\dot{C}^{\mu}+1)\lambda+\tilde{\lambda}(\gamma_{\mu}\dot{C}^{\mu}-1)\epsilon.$ (5.187) Imposing the vanishing of this expression for all gauginos, we get the following conditions on the conformal Killing spinors, $\tilde{\epsilon}(\gamma_{\mu}\dot{C}^{\mu}+1)=0,\qquad(\gamma_{\mu}\dot{C}^{\mu}-1)\epsilon=0.$ (5.188) We have two more conditions on the conformal Killing spinors, thus the maximum number of solutions is reduced by half. The Wilson loop can at most be invariant under two of the four possible supersymmetry variations, and for that it is called _1/2-BPS_. We can find explicitly one family of supersymmetric Wilson loops and one supersymmetry variation with respect to which we are going to perform the localization procedure. In order to solve the conformal Killing equations and the conditions (5.188), we chose explicitly an orthonormal basis and a corresponding vielbein on $\mathbb{S}^{3}$. Since as a manifold $\mathbb{S}^{3}\cong SU(2)$, we can use Lie theory to describe the geometry on the 3-sphere. In particular, the vielbein can be chosen proportional to the Maureer-Cartan form $\Theta\in T^{*}(SU(2))\otimes\mathfrak{su}(2)$,232323Again, we use Roman letters as “flat” indices, and Greek letters as “curved” indices. $e^{i}_{\mu}:=\frac{r}{2}e^{i}(\Theta(\partial_{\mu}))$ (5.189) where $\\{e^{i}\\}$ is a basis of $\mathfrak{su}(2)^{*}$, dual to a basis $\\{T_{i}\\}$ of $\mathfrak{su}(2)$.242424Say, the standard basis given by the Pauli matrices, $T_{i}:=\sigma_{i}/\sqrt{2}$. One can check that this vielbein is consistent with the round metric, giving $g_{\mu\nu}=e_{\mu}^{i}e_{\nu}^{j}\delta_{ij}$ (see [50]). Using this orthonormal basis, the spin connection components are $(\omega_{\mu})_{ij}=\frac{1}{r}e^{k}_{\mu}\varepsilon_{ijk}$ (5.190) where $\varepsilon_{ijk}$ is the Levi-Civita symbol. In this basis the conformal Killing spinor equation for $\epsilon$ looks particularly simple, $\left(\partial_{\mu}+\frac{1}{8}(\omega_{\mu})_{ij}[\gamma^{i},\gamma^{j}]\right)\epsilon=\frac{i}{2r}\gamma_{\mu}\epsilon\quad\Leftrightarrow\quad\partial_{\mu}\epsilon=0$ (5.191) where we used the commutator $[\gamma^{i},\gamma^{j}]=2i\left.\varepsilon^{ij}\right._{k}\gamma^{k}$. We see that the components of $\epsilon$ are constants. The corresponding condition for the supersymmetry of the Wilson loop then requires $\gamma_{\mu}\dot{C}^{\mu}$ to be constant too, as the components of the vector field $\dot{C}^{i}$ in the orthonormal frame. This means that the Wilson loop has to describe grat circles on $\mathbb{S}^{3}$. Following [49], we take $\dot{C}$ parallel to one of the $e^{i}$, say $e^{3}$, and the conformal Killing spinor to satisfy $(\gamma_{3}-1)\epsilon=0.$ (5.192) We will consider the one dimensional subalgebra generated by such restricted spinor, and put $\tilde{\epsilon}=0$. #### 5.5.3 Localization: gauge sector We focus now on the localization of the Chern-Simons path integral, without coupling to the matter multiplet. Ignoring the issue of gauge fixing, we would add to the action the localizing term $tS_{loc}=t\delta\mathcal{V}$, with $t\in\mathbb{R}^{+}$ a parameter, $\delta$ being the supersymmetry transformation generated by the conformal Killing spinor $\epsilon$ described in the last section, and $\mathcal{V}$ some fermionic functional whose bosonic part is positive semi-definite. At the end of Section 4.4, we pointed out that the Super Yang-Mills Lagrangian is an example of $\delta$-exact term, so we put $\displaystyle S_{loc}:=2S_{YM}=\int_{\mathbb{S}^{3}}d^{3}x\sqrt{g}\ \mathrm{Tr}\left(i\tilde{\lambda}\gamma^{\mu}D_{\mu}\lambda+\frac{1}{2}F_{\mu\nu}F^{\mu\nu}+D_{\mu}\sigma D^{\mu}\sigma+i\tilde{\lambda}[\sigma,\lambda]+\right.$ (5.193) $\displaystyle+\left.\left(D+\frac{\sigma}{r}\right)^{2}-\frac{1}{2r}\tilde{\lambda}\lambda\right)$ whose bosonic part is indeed positive semi-definite. This localizing term can be derived also from the functional [49] $\mathcal{V}=\int_{\mathbb{S}^{3}}d^{3}x\sqrt{g}\ \mathrm{Tr}\left((\delta\tilde{\lambda})\lambda\right)$ (5.194) analogously to the one used in the previous chapter for the gauge multiplet. $S_{YM}$ being supersymmetric means that $\delta^{2}=0$ on $\mathcal{V}$, making the localization principle applicable. As usual, the limit $t\to\infty$ localizes the path integral on the configurations that make this term vanish: the terms involving bosonic fields are separately non-negative, while the gaugino and its conjugate have to vanish identically. Summarizing, the localization locus is given by $\left\\{\begin{aligned} &\lambda=\tilde{\lambda}=0\\\ &F=0\Rightarrow A=0\ \text{(up to a gauge transformation)}\\\ &\sigma=a\in\mathfrak{g}\ (\mathrm{constant})\\\ &D=-\frac{1}{r}a\end{aligned}\right.$ (5.195) Keeping into account the gauge-fixing procedure (as we should), the ghost $c$, anti-ghost $\tilde{c}$ and Lagrange multiplier $b$ are added to the theory, taking value in the Lie algebra $\mathfrak{g}$, together with the BRST differential $\delta_{B}$ that acts as $\delta_{B}X=-[c,X]\qquad\delta_{B}c=-\frac{1}{2}[c,c]\qquad\delta_{B}\tilde{c}=b\qquad\delta_{B}b=0$ (5.196) where $X$ is any field in the original theory, acted by a gauge transformation parametrized by $c$. The BRST differential is nilpotent, $\delta_{B}^{2}=0$. The total differential $Q:=\delta_{\epsilon}+\delta_{B}$ (5.197) acts now as the equivariant differential for the $(U(1)\rtimes G)$-equivariant cohomology in the BRST-augmented field space. The original CS action is automatically $Q$-closed since it is gauge invariant, so we can combine the localization principle with the gauge-fixing procedure adding to the Lagrangian the term $Q\left((\delta\tilde{\lambda})\lambda-\tilde{c}\left(\frac{\xi}{2}b-\nabla^{\mu}A_{\mu}\right)\right)$ (5.198) where we suppressed the Lie algebra bilinear $\mathrm{Tr}$ for notational convenience. Since the first term is gauge invariant, $\delta_{B}\left((\delta\tilde{\lambda})\lambda\right)=0$, this gives the same localization term as before. If $\delta[\mbox{ghosts}]=0$ on the gauge-fixing subcomplex, the second term gives $Q\left(\tilde{c}\left(\frac{\xi}{2}b-\nabla^{\mu}A_{\mu}\right)\right)=\frac{\xi}{2}b^{2}-b\nabla^{\mu}A_{\mu}+\tilde{c}\nabla^{\mu}D_{\mu}c+\tilde{c}\nabla^{\mu}\delta A_{\mu}.$ (5.199) The first two terms give, upon path integration over $b$, the usual gauge- fixing Lagrangian in the $R_{\xi}$-gauge; the third term is the ghost Lagrangian. The fourth term $\propto\left(\tilde{c}\nabla^{\mu}\tilde{\lambda}\gamma_{\mu}\right)$ does not change the partition function: if we see this term as a perturbation of the gauge-fixed action, all diagrams with insertion of $\left(\tilde{c}\nabla^{\mu}\tilde{\lambda}\gamma_{\mu}\right)$ will vanish, since $\tilde{c}$ is coupled only to $c$ via the propagator but there are no vertices containing $c$. In other words, the fermionic determinant arising from the path integration over ghosts is not changed by this term. The modified localizing term (5.199) is $Q$-closed: the old localizing term because of gauge invariance and supersymmetry, while the gauge-fixing and ghost terms follows by $Q^{2}A_{\mu}=0$ that is easy to check. After path integration over the auxiliary $b$ the limit $t\to\infty$ finally localizes the theory to the same locus (5.195), with ghosts put to zero. Evaluating the classical action at the saddle point configuration, we get $S_{CS}[a]=\frac{k}{4\pi}\int_{\mathbb{S}^{3}}d^{3}x\sqrt{g}\ \mathrm{Tr}\left(-\frac{2}{r}a^{2}\right)=-k\pi r^{2}\mathrm{Tr}(a^{2})$ (5.200) where we used $\mbox{vol}(\mathbb{S}^{3})=2\pi^{2}r^{3}$. The supersymmetric Wilson loop observable (5.186) localizes to $W_{R}(C)=\frac{1}{\dim{R}}\mathrm{Tr}_{R}\left(e^{2\pi ra}\right)$ (5.201) since the curve $C$ is a great circle of radius $r$. Integrating as usual the rescaled fluctuations above the localization configuration, and taking the limit $t\to\infty$ as in (5.76), the partition function and the Wilson loop expectation value are thus given by a finite-dimensional integral over $\mathfrak{g}$ with Gaussian measure, the “matrix model” $\displaystyle Z$ $\displaystyle=\int_{\mathfrak{g}}da\ e^{-k\pi r^{2}\mathrm{Tr}(a^{2})}Z_{1-loop}^{g}[a]$ (5.202) $\displaystyle\langle W_{R}(C)\rangle$ $\displaystyle=\frac{1}{Z\dim{R}}\int_{\mathfrak{g}}da\ e^{-k\pi r^{2}\mathrm{Tr}(a^{2})}Z_{1-loop}^{g}[a]\mathrm{Tr}_{R}\left(e^{2\pi ra}\right).$ As we pointed out in the last section, the integration over the Lie algebra $\mathfrak{g}$ can be reduced over its Cartan subalgebra $\mathfrak{h}$, exploiting the gauge invariance of the matrix model under the adjoint action of $\mathfrak{g}$ itself. This for example means, in the case of a matrix gauge group, that we integrate over the diagonalized matrices “fixing the gauge” of the matrix model. The corresponding Faddeev-Popov determinant is also called _Vandermonde determinant_ , $\prod_{\alpha}\left(\rho_{\alpha}(a)\right)$ (5.203) where the product runs over the roots of $\mathfrak{g}$. There is left an overcounting given by the possible permutations of the roots, the action of the _Weyl group_ $\mathcal{W}$ of $\mathfrak{g}$, cured dividing by its order $|\mathcal{W}|$. The path integrals are thus rewritten as $\displaystyle Z$ $\displaystyle=\frac{1}{|\mathcal{W}|}\int_{\mathfrak{\mathfrak{h}}}da\ \prod_{\alpha}\left(\rho_{\alpha}(a)\right)e^{-k\pi r^{2}\mathrm{Tr}(a^{2})}Z_{1-loop}^{g}[a]$ (5.204) $\displaystyle\langle W_{R}(C)\rangle$ $\displaystyle=\frac{1}{Z|\mathcal{W}|\dim{R}}\int_{\mathfrak{h}}da\ \prod_{\alpha}\left(\rho_{\alpha}(a)\right)e^{-k\pi r^{2}\mathrm{Tr}(a^{2})}Z_{1-loop}^{g}[a]\mathrm{Tr}_{R}\left(e^{2\pi ra}\right).$ Here we summarize the computation of the 1-loop determinant from [49]. For convenience, we put $r=1$ and $\xi=1$. Inserting the contribution of ghosts, the Lagrangian for the localizing term is given by (suppressing the $\mathrm{Tr}$) $\mathcal{L}_{loc}=\frac{1}{2}F_{\mu\nu}F^{\mu\nu}+D_{\mu}\sigma D^{\mu}\sigma+\left(D+\sigma\right)^{2}+i\tilde{\lambda}\not{D}\lambda+i[\tilde{\lambda},\sigma]\lambda-\frac{1}{2}\tilde{\lambda}\lambda+\partial_{\mu}\tilde{c}D^{\mu}c-\frac{1}{2}b^{2}+b\nabla^{\mu}A_{\mu}.$ (5.205) Considering the limit $t\to\infty$, we rescale as usual the fields around the configuration (5.195): $\sigma=a+\sigma^{\prime}/\sqrt{t},\qquad D=-a+D^{\prime}/\sqrt{t},\qquad X=X^{\prime}/\sqrt{t},$ (5.206) where $X$ are all the fields without zero modes, and then rename $\sigma^{\prime}\to\sigma$, $D^{\prime}\to D$, $X^{\prime}\to X$. In the limit, only quadratic terms in the fluctuations survive, $\mathcal{L}_{loc}\sim\frac{1}{2}\partial_{[\mu}A_{\nu]}\partial^{[\mu}A^{\nu]}-[A_{\mu},a]^{2}+(\partial\sigma)^{2}+(D+\sigma)^{2}+i\tilde{\lambda}\not{\nabla}\lambda+i[\tilde{\lambda},a]\lambda-\frac{1}{2}\tilde{\lambda}\lambda+|\partial\tilde{c}|^{2}-\frac{1}{2}b^{2}+b\nabla^{\mu}A_{\mu}.$ (5.207) The resulting theory is free, and we can integrate it giving the corresponding 1-loop determinant. We will neglect all overall normalization constant from the Gaussian integrations. The integral over the auxiliary field $b$ gives the gauge fixing term $-\frac{1}{2}(\nabla^{\mu}A_{\mu})^{2}$. The contribution from $D$ is purely Gaussian and can be integrated out removing the corresponding term. The integration over $\sigma$ gives a determinant $\det{(\nabla^{2})}^{-1/2}$, and the (Grassman) integral over the ghosts gives $\det{(\nabla^{2})}$. It is useful to separate the gauge field as (Helmolz- Hodge decomposition) $A_{\mu}=B_{\mu}+\partial_{\mu}\phi$ with $\phi$ scalar and $B_{\mu}$ divergenceless, $\nabla^{\mu}B_{\mu}=0$. With this decomposition, the Lorentz gauge condition becomes $\nabla^{2}\phi=0$, and we can integrate $\phi$ giving a determinant $\det{(\nabla^{2})}^{-1/2}$, that cancels the above two other contributions. We are left with $-B_{\mu}\Delta B^{\mu}-[a,B_{\mu}]^{2}+i\tilde{\lambda}\not{\nabla}\lambda+i[\tilde{\lambda},a]\lambda-\frac{1}{2}\tilde{\lambda}\lambda$ (5.208) where $\Delta$ is the vector Laplacian. Now we use the fact that the path integral can be reduced over the Cartan subalgebra of $\mathfrak{g}$, considering $a\in\mathfrak{h}$, and $B_{\mu}=B_{\mu}^{(\mathfrak{h})}+B_{\mu}^{\alpha}e_{\alpha}$ (5.209) where $B_{\mu}^{(\mathfrak{h})}$ is the component of $B_{\mu}$ along $\mathfrak{h}$, and similarly for the gaugino. This component does not enter in the Lie brackets with $a$, so its contribution to the path integral is independent of $a$, and we drop it. The remaining interesting terms are $\sum_{\alpha}\left(B_{\mu}^{-\alpha}(-\Delta+\rho_{\alpha}(a)^{2})B_{\mu}^{\alpha}+\tilde{\lambda}^{-\alpha}\left(i\not{\nabla}+i\rho_{\alpha}(a)-\frac{1}{2}\right)\lambda^{\alpha}\right)$ (5.210) where the $a$-dependent kinetic terms are clearly identified, and the component fields appearing are real or complex valued scalars and spinors. The Gaussian integration over these fields lead to the determinant factors $Z^{g}_{1-loop}[a]=\prod_{\alpha}\frac{\det\left(i\not{\nabla}+i\rho_{\alpha}(a)-\frac{1}{2}\right)}{\det{\left(-\Delta+\rho_{\alpha}(a)^{2}\right)}^{1/2}}.$ (5.211) Now, using the fact that the eigenvalues of the Laplacian on divergenceless vectors are $(l+1)^{2}$ with degeneracy $2l(l+2)$, and the eigenvalues of $i\not{\nabla}$ are $\pm\left(l+\frac{1}{2}\right)$ with degeneracy $l(l+1)$, where $l\in\mathbb{Z}^{+}$, the corresponding determinants can be written as infinite products $\prod_{\alpha}\prod_{l=1}^{\infty}\frac{(l+i\rho_{\alpha}(a))^{l(l+1)}(-l-1+i\rho_{\alpha}(a))^{l(l+1)}}{((l+1)^{2}+\rho_{\alpha}(a)^{2})^{l(l+2)}}=\prod_{\alpha}\prod_{l=1}^{\infty}\frac{(l+i\rho_{\alpha}(a))^{(l+1)}}{(l-i\rho_{\alpha}(a))^{(l-1)}}$ (5.212) where the equality follows after some simplifications. Since roots come in pairs $(\rho_{\alpha},-\rho_{\alpha})$, taking the square of this one gets $\left(Z^{g}_{1-loop}[a]\right)^{2}=\prod_{\alpha}\prod_{l=1}^{\infty}\frac{(l^{2}+\rho_{\alpha}(a)^{2})^{(l+1)}}{(l^{2}+\rho_{\alpha}(a)^{2})^{(l-1)}}=\prod_{\alpha}\prod_{l=1}^{\infty}\left(l^{2}+\rho_{\alpha}(a)^{2}\right)^{2}.$ (5.213) Collecting a factor $l^{4}$ the product splits in the factorization formula for the hyperbolic sine, $\frac{\sinh(\pi z)}{\pi z}=\prod_{l=1}^{\infty}\left(1+\frac{z^{2}}{l^{2}}\right)$ (5.214) and an $a$-independent divergent part that can be regularized with the zeta- function method, $\prod_{l=1}^{\infty}l^{4}=e^{4\sum_{l=1}^{\infty}\log(l)}=e^{-4\zeta^{\prime}(0)}=e^{2\log(2\pi)}.$ (5.215) Up to an overall normalization constant, the $a$-dependence of the 1-loop determinant is finally given by $Z_{1-loop}^{g}[a]=\prod_{\alpha}\left(\frac{2\sinh(\pi\rho_{\alpha}(a))}{\pi\rho_{\alpha}(a)}\right)$ (5.216) where we see cancellation between the denominator and the Vandermonde determinant (5.203). Collecting the above results, the localization formulas for the partition function and the expectation value of the supersymmetric Wilson loop in the pure CS theory are $\displaystyle Z$ $\displaystyle\sim\frac{1}{|\mathcal{W}|}\int_{\mathfrak{\mathfrak{h}}}da\ e^{-k\pi\mathrm{Tr}(a^{2})}\prod_{\alpha}\left(2\sinh(\pi\rho_{\alpha}(a))\right)$ (5.217) $\displaystyle\langle W_{R}(C)\rangle$ $\displaystyle=\frac{1}{Z|\mathcal{W}|\dim{R}}\int_{\mathfrak{h}}da\ e^{-k\pi\mathrm{Tr}(a^{2})}\mathrm{Tr}_{R}\left(e^{2\pi a}\right)\prod_{\alpha}\left(2\sinh(\pi\rho_{\alpha}(a))\right)$ $\displaystyle=\frac{1}{\dim{R}}\frac{\int_{\mathfrak{h}}da\ e^{-k\pi\mathrm{Tr}(a^{2})}\mathrm{Tr}_{R}\left(e^{2\pi a}\right)\prod_{\alpha}\left(2\sinh(\pi\rho_{\alpha}(a))\right)}{\int_{\mathfrak{\mathfrak{h}}}da\ e^{-k\pi\mathrm{Tr}(a^{2})}\prod_{\alpha}\left(2\sinh(\pi\rho_{\alpha}(a))\right)}.$ These general localization formulas can be tested comparing their results for specific choices of $G$ to perturbative calculations, for example. In the case of $U(N)$ gauge group, the integral over the Cartan subalgebra is an integral over diagonal matrices $a=\mathrm{diag}(\lambda_{1},\cdots,\lambda_{N})$, and the roots are given by $\rho_{ij}(a)=\lambda_{i}-\lambda_{j}$ for $i\neq j$. The Weyl group is $S_{N}$, thus $|\mathcal{W}|=N!$. If we take the Wilson loop in the fundamental representation, from (5.217) we get $\displaystyle Z$ $\displaystyle\sim\frac{1}{N!}\int\left(\prod_{i}d\lambda_{i}\ e^{-k\pi\lambda_{i}^{2}}\right)\prod_{i\neq j}2\sinh(\pi(\lambda_{i}-\lambda_{j})),$ (5.218) $\displaystyle\langle W_{\mathbf{N}}(C)\rangle$ $\displaystyle=\frac{1}{ZN!N}\int\left(\prod_{i}d\lambda_{i}\ e^{-k\pi\lambda_{i}^{2}}\right)\left(e^{2\pi\lambda_{1}}+\cdots+e^{2\pi\lambda_{N}}\right)\prod_{i\neq j}2\sinh(\pi(\lambda_{i}-\lambda_{j})),$ that are sums of Gaussian integrals, and can be computed exactly. The result for the Wilson loop expectation value is $\langle W_{\mathbf{N}}(C)\rangle=\frac{1}{N}e^{-Ni\pi/k}\frac{\sin\left(\frac{\pi N}{k}\right)}{\sin\left(\frac{\pi}{k}\right)},$ (5.219) which is known as the exact result [83], up to the overall phase factor $e^{-Ni\pi/k}$. This kind of phase factors arise in perturbative calculations in the so-called _framing_ of the Wilson loop. A perturbative calculation of the Wilson loop involves computations of correlators of the type $\langle A_{\mu_{1}}(x_{1})A_{\mu_{2}}(x_{2})\cdots\rangle$, where $x_{1},x_{2},\cdots$ are coordinates of points on the image of the curve $C$. This contribution diverges when $x_{1}=x_{2}$, so it is necessary to choose some regularization scheme to perform the computations. For example, considering the 2-point function $\langle A_{\mu_{1}}(x)A_{\mu_{2}}(y)\rangle$, this clashing of points can be avoided requiring that $y$ is integrated over a shifted curve $C_{f}$ such that $C^{\mu}_{f}(\tau)=C^{\mu}(\tau)+\alpha\ n^{\mu}(\tau)$ (5.220) where $n$ is orthogonal to $\dot{C}$. The choice of such an orthogonal component (frame) at every point on the curve is called framing. Even if at the end of the calculation one takes $\alpha\to 0$, this procedure leaves a deformation-dependent term, that in pure $U(N)$ CS is $e^{\frac{i\pi N}{k}\chi(C,C_{f})}$ (5.221) where $\chi(C,C_{f})$ is a topological invariant that takes integer values corresponding to the number of times the path $C_{f}$ winds around $C$. We see that localization produces an expectation value at framing -1 (see also [98] for a detailed discussion about framing). #### 5.5.4 Localization: matter sector We turn now to the result for the localization of the matter-coupled theory. This is of course gauge invariant, so the equivariant differential acts effectively as $Q\sim\delta$, since the ghost sector has been already considered in the previous paragraph. This means that, following the localization principle, we have to extend the matter action with a $\delta$-exact term. We are free to consider the canonical choice (5.81) as in [49], or using the fact that [50] the matter action (5.183) is actually given by a supersymmetry variation, as the case of the YM action. This means that we can consider the localizing terms $tS_{m}+tS_{YM+ghosts}$ or, schematically $t\int\delta\left((\delta\psi)^{\dagger}\psi+\tilde{\psi}(\delta\tilde{\psi})^{\dagger}\right)+tS_{YM+ghosts}$ that have positive semi-definite bosonic parts. The second term in both choices is the one analyzed in the previous paragraph, and gives the same localization locus for the gauge and ghost sector, while both the first terms vanishes for the field configurations $\psi=0,\qquad\phi=0,\qquad F=0.$ (5.222) This means that the classical action of the matter sector does not contribute to the partition function, but only in the 1-loop determinant. Expanding the fields around this configuration and scaling the fluctuations with the usual $1/\sqrt{t}$ factor, we see that there are no couplings to the gauge sector fluctuations that survive in the $t\to\infty$ limit, but only to the zero mode $a$ of $\sigma$. Thus the determinant factorizes as $Z_{1-loop}[a]=Z_{1-loop}^{g}[a]Z_{1-loop}^{m}[a].$ (5.223) If matter is present in different copies of chiral multiplets, in maybe different representations of the gauge group, the determinant factorizes in the same way for each multiplet. The determinant for the matter sector can be computed diagonalizing the the kinetic operators acting on the scalar and the fermion field, after having integrated out the auxiliary $F$, and considering the path integration over the Cartan subalgebra with $a\in\mathfrak{h}$. In particular, the relevant kinetic operators that have to be diagonalized are $K_{b}^{(\rho)}=\left(-\nabla^{2}+\rho(a)^{2}-i\rho(a)+\frac{3}{4}\right),\qquad K_{f}^{(\rho)}=\left(i\not{\nabla}+i\rho(a)\right),$ (5.224) for the (complex) bosonic and fermionic parts, where $a$ is regarded as acting on the representation $R$ with weights $\\{\rho\\}$. The eigenvalues of $-\nabla^{2}$ are $4j(j+1)$ with $j=0,\frac{1}{2},\cdots$ with degeneracy $(2j+1)^{2}$, that we can rewrite as $l(l+2)$ with degeneracy $(l+1)^{2}$ and $l=0,1,\cdots$. The eigenvalues of $i\not{\nabla}$ are $\pm\left(l+\frac{1}{2}\right)$ with degeneracy $l(l+1)$, with $l=1,2,\cdots$. Thus the one loop determinant results, after a change of dummy index and some simplifications $\displaystyle Z_{1-loop}^{m}[a]$ $\displaystyle=\prod_{\rho}\frac{\det(K_{f})}{\det(K_{b})}$ (5.225) $\displaystyle=\prod_{\rho}\prod_{l=1}^{\infty}\frac{\left(l+\frac{1}{2}+i\rho(a)\right)^{l(l+1)}\left(l+\frac{1}{2}-i\rho(a)\right)^{l(l+1)}}{\left(l+\frac{1}{2}+i\rho(a)\right)^{l^{2}}\left(l-\frac{1}{2}-i\rho(a)\right)^{l^{2}}}$ $\displaystyle=\prod_{\rho}\prod_{l=1}^{\infty}\left(\frac{l+\frac{1}{2}+i\rho(a)}{l-\frac{1}{2}-i\rho(a)}\right)^{l}$ This product can be regularized using the zeta-function. We refer to [50] for the details of the computation, and report here the result in the case the fields take value in a self-conjugate representation $R$ of the gauge group:252525For example, if $R=S\oplus S^{*}$. $Z_{1-loop}^{m}[a]=\prod_{\rho}\left(2\cosh{(\pi\rho(a))}\right)^{-1/2}$ (5.226) where now $a\in R(\mathfrak{h})$ and $\rho(a)$ is the weight of the Cartan element in the representation $R$. Summarizing, we have seen that the application of the supersymmetric localization principle to the matter-coupled SCS theory on $\mathbb{S}^{3}$ reduces the path integral to a finite-dimensional integral describing a matrix model over the Lie algebra of the theory. Using the notation (5.128), the localization formulas for the partition function and the supersymmetric Wilson loop expectation value, with matter multiplets coming in self-conjugate representations $R_{1}\oplus R_{1}^{*},R_{2}\oplus R_{2}^{*},\cdots$ are $\displaystyle Z$ $\displaystyle=\frac{1}{|\mathcal{W}|}\int_{\mathfrak{\mathfrak{h}}}da\ e^{-k\pi\mathrm{Tr}(a^{2})}\frac{\det_{ad}2\sinh(\pi a)}{\left(\det_{R_{1}}2\cosh{(\pi a)}\right)\left(\det_{R_{2}}2\cosh{(\pi a)}\right)\cdots}$ (5.227) $\displaystyle\langle W_{R}(C)\rangle$ $\displaystyle=\frac{1}{Z|\mathcal{W}|\dim{R}}\int_{\mathfrak{h}}da\ e^{-k\pi\mathrm{Tr}(a^{2})}\mathrm{Tr}_{R}\left(e^{2\pi a}\right)\frac{\det_{ad}2\sinh(\pi a)}{\left(\det_{R_{1}}2\cosh{(\pi a)}\right)\left(\det_{R_{2}}2\cosh{(\pi a)}\right)\cdots}$ #### 5.5.5 The ABJM matrix model ABJM theory is a special type of matter-coupled SCS theory in 3-dimensions constructed in [99], that has the interesting property to be dual under the AdS/CFT conjecture to a certain orbifold background in M-theory. It consists of two copies of $\mathcal{N}=2$ SCS theory, each one with gauge group $U(N)$, and opposite levels $k,-k$. In addition, the are four matter (chiral and anti- chiral) supermultiplets $\Phi_{i},\tilde{\Phi}_{i}$, with $i=1,2$, in the bi- fundamental representation of $U(N)\times U(N)$, $(\mathbf{N},\bar{\mathbf{N}})$ and $(\bar{\mathbf{N}},\mathbf{N})$. This field content can be represented as the _quiver_ in Fig. 5.1. Figure 5.1: The quiver for ABJM theory. The two nodes represent the gauge multiplets, with the convention of specifying the level of the CS term. The oriented links represent the matter multiplets in the bi-fundamental and anti- bi-fundamental representations. The superpotential for the matter part is given by $W=\frac{4\pi}{k}(\Phi_{1}\tilde{\Phi}_{1}\Phi_{2}\tilde{\Phi}_{2}-\Phi_{1}\tilde{\Phi}_{2}\Phi_{2}\tilde{\Phi}_{1}),$ (5.228) and this structure actually enhance the supersymmetry of the resulting theory to $\mathcal{N}=6$.262626This is not apparent from the original action, but can be realized noticing that the superpotential has an $SU(2)\times SU(2)$ symmetry that rotates separately the $\Phi_{i}$ and the $\tilde{\Phi}_{i}$. This, combined with the original $SU(2)^{\mathcal{R}}$ symmetry of the theory, gives an $SU(4)\cong Spin(6)$ symmetry that acts non-trivially on the supercharges. Thus the final theory has to have an enhanced $\mathcal{N}=6$ supersymmetry. If now $a=\mathrm{diag}(\lambda_{1},\cdots,\lambda_{N},\hat{\lambda}_{1},\cdots,\hat{\lambda}_{N})$, the weights in the bi-fundamental representations are $\rho_{i,j}^{(N,\bar{N})}(a)=\lambda_{i}-\hat{\lambda}_{j},\qquad\rho_{i,j}^{(\bar{N},N)}(a)=\hat{\lambda}_{j}-\lambda_{i}.$ (5.229) Plugging this information into (5.227), the partition function in this case localizes to the following matrix model, $Z\sim\frac{1}{N!N!}\int\left(\prod_{i}d\lambda_{i}d\hat{\lambda}_{i}\ e^{-k\pi(\lambda_{i}^{2}-\hat{\lambda}_{i}^{2})}\right)\frac{\prod_{i\neq j}\left(2\sinh(\pi(\lambda_{i}-\lambda_{j}))2\sinh(\pi(\hat{\lambda}_{i}-\hat{\lambda}_{j}))\right)}{\prod_{i,j}\left(2\cosh(\pi(\lambda_{i}-\hat{\lambda}_{j}))\right)}.$ (5.230) The circular Wilson loop under consideration can be called now _1/6 BPS_ with respect to the enhanced supersymmetry of the model. Its expectation value in the fundamental representation is obtained by plugging a factor $(1/N)\sum_{i}e^{2\pi\lambda_{i}}$ as before. This matrix model cannot be solved exactly as in the case of the pure CS discussed above, but can be studied in the $N\to\infty$ limit with the saddle-point technique showed in Section 5.4.5 [50, 94]. We also mention that, in this particular theory with enhanced $\mathcal{N}=6$ supersymmetry, it was possible to construct a _1/2 BPS_ Wilson loop (so invariant under half of the $\mathcal{N}=6$ supersymmetry algebra). The latter can be solved applying the same localization scheme that brings to the matrix model describing the 1/6 BPS Wilson loop presented above [100, 101]. A compact review introducing the state of the art on recent results about supersymmetric Wilson loops in ABJM and related theories can be found in [84]. ## Chapter 6 Non-Abelian localization and 2d YM theory In this chapter we are going to summarize the result obtained mainly in [15] by Witten. This was the first attempt in the physics literature of extending the equivariant localization formalism to possibly non-Abelian group actions. In that work, a modified definition of equivariant integration was defined, and this allowed for an extension of the same procedure discussed in Chapter 3 to show the localization property of integrals computed over spaces with generic symmetry group $G$. This new formalism was applied to the study of 2-dimensional Yang-Mills (YM) theory over a Riemann surface, a relatively simple model from the physical point of view, but with a very rich underlying mathematical structure. In the following, we are going first to review the geometry of this special model, in connection with the symplectic geometry introduced in Section 3.3, as a motivation for the more mathematical discussion about the Witten’s equivariant integration and non-Abelian localization principle that will follow. Next, we will review the ideas underlying the application of this new localization principle to the YM theory, and how this application results in a “mapping” between this model and a suitable topological theory, establishing the topological nature of the YM theory in the weak coupling limit. In the final section, we will summarize the interpretation given by the localization framework to the already existing solution for the partition function of this model. As we pointed out in the Introduction, other generalizations of the Duistermaat-Heckman theorem to non-Abelian Hamiltonian systems also appeared in the mathematical literature, as the result obtained by Jeffrey and Kirwan in [16]. Other applications of this extended formalism followed, and Witten’s approach was used for example more recently to describe Chern-Simons theories over a special class of 3-manifolds in [102]. ### 6.1 Prelude: moment maps and YM theory In the next section we are going to review Witten’s extension of the equivariant localization principle to possibly non-Abelian group actions, and a generalization of the DH formula in this direction. In [15] this was applied to reinterpret the weak coupling limit of pure YM theory on a Riemann surface. This theory is exactly solvable, in the sense that its partition function can be expressed in closed form, and its zero-coupling limit is known to describe a topological field theory. These features make 2-dimensional YM theory very appealing from the mathematical structure it carries, and make it possible to compare results or interpretations obtained via this “new” localization method with already existing solutions of the problem. We are going to discuss more about the topological interpretation of 2d YM theory later, while in this section we review some results introduced by Atiyah and Bott [103] about the symplectic structure underlying this special QFT. This can be useful to contextualize the generic discussion of the next section, and it prepares the ground for the formal application of the non- Abelian localization principle. We start by considering the partition function of YM theory on a compact orientable Riemannian manifold $\Sigma$ of arbitrary dimension, $\displaystyle Z(\epsilon)$ $\displaystyle=\frac{1}{\mathrm{vol}(\mathcal{G}(P))}\left(\frac{1}{2\pi\epsilon}\right)^{\dim(\mathcal{G})/2}\int_{\mathcal{A}(P)}DA\ e^{-S[A]},$ (6.1) $\displaystyle S[A]$ $\displaystyle=-\frac{1}{2\epsilon}\int_{\Sigma}\mathrm{Tr}(F^{A}\wedge\star F^{A}).$ Here $\epsilon:=g_{YM}^{2}$ is the square of the YM coupling constant. To describe the rest of the ingredients, let us recall the geometry underlying the gauge theory (to fill some of the details, see Appendix A.1). The dynamical field here is the _connection_ $A\in\Omega(P;\mathfrak{g})$ on a principal $G$-bundle $P\xrightarrow{\pi}\Sigma$, where $G$ is a compact connected Lie group with Lie algebra $\mathfrak{g}$. The path integral is thus taken over the space $\mathcal{A}(P)$ of $G$-equivariant vertical 1-forms with values in $\mathfrak{g}$, that is naturally an _affine space_ modeled on the infinite-dimensional vector space $\mathfrak{a}$ of $G$-equivariant horizontal 1-forms with values in $\mathfrak{g}$. This gives to $\mathcal{A}(P)$ the structure of an infinite-dimensional manifold, whose tangent spaces are $T_{A}\mathcal{A}(P)\cong\mathfrak{a}\cong\Omega^{1}(\Sigma;\mathrm{ad}(P))$, where we identified horizontal forms over $P$ with forms over the base $\Sigma$.111Recall that horizontality means essentially to have components only in the “directions” of the base space, and the $G$-equivariance ensures the right transformation behavior as forms valued in the _adjoint bundle_ $\mathrm{ad}(P)$, the associated bundle to $P$ that has $\mathfrak{g}$ as typical fiber. Thus $\mathfrak{a}\cong\Omega^{1}(\Sigma;\mathrm{ad}(P))$. In other words, any vector field $\alpha\in\Gamma(T\mathcal{A}(P))$ can be expanded locally as $\alpha=\alpha_{\mu}^{a}T_{a}\otimes dx^{\mu},\qquad\alpha_{\mu}^{a}\in C^{\infty}(\Sigma\times\mathcal{A}(P)),$ (6.2) with coefficients that depend on the point $A\in\mathcal{A}(P)$ and $p\in\Sigma$. The curvature $F^{A}=dA+\frac{1}{2}[A\stackrel{{\scriptstyle\wedge}}{{,}}A]$ of the connection $A$ is a horizontal 2-form over $P$, so we can identify it as a 2-form on the adjoint bundle without loss of information, $F^{A}\in\Omega^{2}(\Sigma;\mathrm{ad}(P))$. As such, it can be integrated as a differential form over $\Sigma$. In the action $S[A]$, “$\mathrm{Tr}$” represents a (negative definite) invariant inner product on $\mathfrak{g}$, and $\star$ is the Hodge dual operation, that is identified by the presence of a metric on $\Sigma$.222The definition of the Hodge star is, implicitly, $\alpha\wedge\star\beta=g^{-1}(\alpha,\beta)\omega$ for any $\alpha,\beta\in\Omega^{k}(\Sigma)$. Here $g^{-1}$ is the “inverse” metric on $\Sigma$, that extends multi-linearly its action on every tangent space as $g^{-1}(\alpha,\beta)=g^{\mu_{1}\nu_{1}}\cdots g^{\mu_{k}\nu_{k}}\alpha_{\mu_{1}\cdots\mu_{k}}\beta_{\nu_{1}\cdots\nu_{k}}$. $\omega$ is a volume form (that can be induced by $g$, for example). The Hodge star satisfies the property $\star^{2}\alpha=(-1)^{k(\dim(\Sigma)-k)}\alpha$. $\mathcal{G}(P)\cong\Omega^{0}(\Sigma;\mathrm{Ad}(P)$ is the group of gauge transformations, that is locally equivalent to the space of $G$-valued functions over $\Sigma$, and acts naturally on $\mathcal{A}(P)$. If $\phi\in Lie(\mathcal{G}(P))\cong\Omega^{0}(\Sigma;\mathrm{ad}(P))$ is an element of the Lie algebra of infinitesimal gauge transformations, its associated fundamental vector field at the point $A\in\mathcal{A}(P)$ is $\underline{\phi}_{A}\equiv\delta_{\phi}A=\nabla^{A}\phi=d\phi+[A,\phi].$ (6.3) The path integral measure $DA$ can be defined formally as the Riemannian measure induced by a metric on the affine space $\mathcal{A}(P)$. The latter can be induced by the metrics on $\Sigma$ and on $\mathfrak{g}$, and defined pointwise in $\mathcal{A}(P)$ as $(\alpha,\beta)_{A}:=-\int_{\Sigma}\mathrm{Tr}(\alpha^{A}\wedge\star\beta^{A})$ (6.4) for every $\alpha^{A},\beta^{A}\in\Omega^{1}(\Sigma;\mathrm{ad}(P))$. With this definition, the YM action can be rewritten as $S[A]=\frac{1}{2\epsilon}(F,F)_{A}.$ (6.5) We can now specialize the discussion to the case in which $\dim(\Sigma)=2$, i.e. the base space is a Riemann surface. It is a well-known fact in geometry that any Riemann surface is a _Kähler manifold_ : it admits a Riemannian metric $g$, a symplectic form $\omega$ (that can be a choice of volume form), and a complex structure $J$ such that the compatibility condition $g(\cdot,\cdot)=\omega(\cdot,J(\cdot))$ is satisfied.333A complex structure on a vector space $V$ is an isomorphism $J:V\to V$ such that $J^{2}=-id_{V}$. It intuitively plays the role of “multiplication by $i$” when one considers the complexified $V^{\mathbb{C}}:=V\otimes\mathbb{C}$, allowing for a decomposition of $V^{\mathbb{C}}$ in a holomorphic subspace (generated by the eigenvectors with eigenvalue $+i$) and anti-holomorphic subspace (generated by the eigenvectors with eigenvalue $-i$). A manifold $M$ has _almost complex structure_ if there is a tensor $J\in\Gamma(T^{1}_{1}M)$ that acts as a complex structure in every tangent space. If the holomorphic decomposition can be extended on an entire neighborhood of every point by a suitable choice of coordinates, $M$ has _complex structure_ , and admits an atlas of holomorphic coordinates. Riemann surfaces can thus be thought as 2-dimensional real manifolds, or 1-dimensional complex manifolds. This special property holds also for $\mathcal{A}(P)$, since in addition to the metric (6.4) we can define the symplectic form $\Omega\in\Omega^{2}(\mathcal{A}(P))$ such that $\Omega_{A}(\alpha,\beta):=-\int_{\Sigma}\mathrm{Tr}(\alpha^{A}\wedge\beta^{A}),$ (6.6) and the complex structure on $T\mathcal{A}(P)$ is provided by the Hodge duality, $\star:\Omega^{1}(\Sigma;\mathrm{ad}(P))\to\Omega^{1}(\Sigma;\mathrm{ad}(P))$ such that $\star^{2}=-1$. Then the compatibility condition is immediately satisfied, since $(\cdot,\cdot)=\Omega(\cdot,\star(\cdot))$. The fact that $\Omega$ is symplectic can be seen by noticing that, in any basis, it has constant components (i.e. independent from $A\in\mathcal{A}(P)$): $\Omega_{ab}^{\mu\nu}(A)=\Omega_{A}(T_{a}\otimes dx^{\mu},T_{b}\otimes dx^{\nu})=-\mathrm{Tr}(T_{a}T_{b})\varepsilon^{\mu\nu}\left(\int_{\Sigma}dx^{1}dx^{2}\right)\quad\in\mathbb{R}.$ (6.7) The non-degeneracy follows from the non-degeneracy of $\mathrm{Tr}$ and of $\int_{\Sigma}$, and the skew-symmetry is obvious from the definition. Thus $\mathcal{A}(P)$ is Kähler. For our applications, we focus on the fact that $\mathcal{A}(P)$ has now a canonical symplectic structure. It is natural to wonder if it possible to extend all the machinery that we introduced in Section 3.3 also to this case, and in particular if the $\mathcal{G}(P)$-action on $\mathcal{A}(P)$ results to be symplectic or Hamiltonian with respect to $\Omega$. The answer was given by in [103], and we state it in the following theorem. ###### Theorem 6.1.1 (Atiyah-Bott). In 2-dimensions, the group $\mathcal{G}(P)$ of gauge transformations acts in an Hamiltonian way on $\mathcal{A}(P)$, with a moment map identified by the curvature $F$. ###### Proof. To see this, let us introduce the moment map as $\mu:Lie(\mathcal{G}(P))\to C^{\infty}(\mathcal{A}(P))$ such that $\mu_{\phi}(A):=\langle F^{A},\phi\rangle=-\int_{\Sigma}\mathrm{Tr}(F^{A}\phi),$ (6.8) and check that the Hamiltonian property is satisfied. For every $\alpha\in\Gamma(T\mathcal{A}(P))$ and $\phi\in Lie(\mathcal{G}(P))$, we compute $\displaystyle(\iota_{\phi}\Omega_{A})(\alpha)$ $\displaystyle=\Omega_{A}(\underline{\phi},\alpha)=-\int_{\Sigma}\mathrm{Tr}(\nabla^{A}\phi\wedge\alpha^{A})=\int_{\Sigma}\mathrm{Tr}(\phi\nabla^{A}\alpha^{A}),$ (6.9) $\displaystyle\left.\delta\mu_{\phi}\right|_{A}(\alpha)$ $\displaystyle=-\int_{\Sigma}\mathrm{Tr}\left(F^{A+\alpha}\phi-F^{A}\phi\right)=-\int_{\Sigma}\mathrm{Tr}\left(\phi\nabla^{A}\alpha^{A}\right),$ (6.10) where $\delta$ is the de Rham differential on $\mathcal{A}(P)$, that acts in the usual sense of variational calculus. We see that $\iota_{\phi}\Omega=-\delta\mu_{\phi}$, thus $\mu$ provides a correct moment map for the $\mathcal{G}(P)$-action. If we identify $Lie(\mathcal{G}(P))$ with $Lie(\mathcal{G}(P))^{*}$ through the pairing $\langle\cdot,\cdot\rangle$ introduced above, and regard the curvature $F:\mathcal{A}(P)\to\Omega^{2}(\Sigma;\mathrm{ad}(P))$ as an element of $C^{\infty}(\mathcal{A}(P))\otimes Lie(\mathcal{G}(P))^{*}$, we can simply write that $\mu\equiv F$. ∎ Another corollary of $\mathcal{A}(P)$ being Kähler is that the path integral measure $DA$ is formally equivalent to the Liouville measure induced from $\Omega$, since by compatibility of the structures the latter is equivalent to the Riemannian measure induced by $(\cdot,\cdot)$. Since we are working on an infinite-dimensional space, we can write this measure formally as $DA=\exp(\Omega),$ (6.11) as we did in (3.43) but with $n=\infty$. With this identification, we see that the path integral of the 2-dimensional YM theory acquires the very suggestive form $Z(\epsilon)\propto\int_{\mathcal{A}(P)}\exp\left(\Omega-\frac{1}{2\epsilon}(\mu,\mu)\right).$ (6.12) This path integral resembles very much an infinite-dimensional version of the type of integrals we treated when discussing the Duistermaat-Heckman localization formula in Section 3.3, but with the fundamental difference that now the exponent of the integrand is not the moment map, but its square. We will return to this point in the next section. Here we notice that in the weak coupling limit $\epsilon\to 0$ the path integral will receive contributions from the saddle points of the action $S=\frac{1}{2\epsilon}(\mu,\mu)$, that is the space of solutions of the classical equations of motion $\nabla^{A}\star F^{A}=0$. Every one of these contributions brings roughly a term that decays as $\sim\exp\left(-1/\epsilon\right)$ to the partition function, the main one being determined by the absolute minimum at $\mu=0$, the subspace of flat connections $\mu^{-1}(0)\subset\mathcal{A}(P)$. Eliminating the redundancy from the gauge freedom of the theory, the most interesting piece of the _physical_ field space, especially in the weak coupling limit, is thus determined by the quotient $\mathcal{A}_{0}:=\faktor{\mu^{-1}(0)}{\mathcal{G}(P)},$ (6.13) or in other words, when computing the path integral one is interested in the $\mathcal{G}(P)$-equivariant cohomology of $\mu^{-1}(0)$, $H_{\mathcal{G}}^{*}(\mu^{-1}(0))\cong H^{*}(\mathcal{A}_{0})$. The quotient $\mathcal{A}_{0}$ is the _moduli space of flat connections_. It turns out that this space has a nice interpretation in symplectic geometry in terms of _symplectic reduction_. A theorem by Marsden-Weinstein-Meyer (MWM) [104, 105, 34] in fact states that, in a generic Hamiltonian $G$-space $(M,\omega,G,\mu)$, if the zero-section of the moment map $\mu^{-1}(0)\subset M$ is acted on _freely_ by $G$, then the base space $M_{0}:=\mu^{-1}(0)/G$ of the principal $G$-bundle $\mu^{-1}(0)\xrightarrow{\pi}M_{0}$ is a symplectic manifold, with symplectic form $\omega_{0}\in\Omega^{2}(M_{0})$ satisfying $\left.\omega\right|_{\mu^{-1}(0)}=\pi^{*}(\omega_{0}).$ (6.14) In other words, the restriction of $\omega$ to $\mu^{-1}(0)$ is a basic form, completely determined by a symplectic form $\omega_{0}$ on the base space. The space $(M_{0},\omega_{0})$ is called Marsden–Weinstein quotient, symplectic quotient or symplectic reduction of $M$ by $G$.444Symplectic reduction in classical mechanics on $M=\mathbb{R}^{2n}$ occurs when one of the momenta is an integral of motion, $0=\dot{p}_{n}=-\partial_{n}H$. In that case, one can solve the system in the reduced coordinates $(q^{1},\cdots,q^{n-1},p_{1},\cdots,p_{n-1})$ and then solve for the $n^{th}$ coordinate separately. The MWM theorem essentially generalizes this process in a fully covariant setting. Returning to the case of YM theory, this means that in the limit $\epsilon\to 0$, when the path integral is reduced to $\mathcal{A}_{0}$ by gauge fixing, the symplectic form can be reduced without loss of information on this base space. In the next section we will see that, applying localization, this is extended to the whole exponential. ### 6.2 A localization formula for non-Abelian actions In the last section we found an Hamiltonian interpretation of the system $(\mathcal{A}(P),\Omega,\mathcal{G}(P),\mu\equiv F)$ for YM theory on a 2-dimensional Riemann surface $\Sigma$. Here we would like to make contact with the DH formula, that we described for analogous systems in finite- dimensional geometry. We notice that the main differences with the case treated in Section 3.3 are essentially two: $\mathcal{G}(P)$ is non-Abelian in general for non-Abelian gauge groups $G$, and the path integral is not in the form of an oscillatory integral of the DH type. Indeed, schematically we have $\text{DH}\to\int\exp{\left(\omega+i\mu\right)},\qquad\quad\text{YM}\to\int\exp{\left(\Omega-\frac{1}{2}|\mu|^{2}\right)}.$ (6.15) In the following, we will describe the solution proposed in [15] to generalize the DH formula to the non-Abelian case starting from the first integral in (6.15), and how this procedure can be used to recover the second one, of the YM type. We consider a generic Hamiltonian system $(M,\omega,G,\mu)$ with compact semisimple Lie group $G$ of dimension $\dim(G)=s$, and the associated Cartan model defined by the space $\Omega_{G}(M)=(S(\mathfrak{g}^{*})\otimes\Omega(M))^{G}$ of equivariant forms, on which we defined the action of the extended operators555We adopt Witten’s conventions and substitute $\phi^{a}\mapsto i\phi^{a}$ in the definition of the Cartan differential, analogously to the DH case of Section 3.3. $\displaystyle d_{C}$ $\displaystyle=1\otimes d-i\phi^{a}\otimes\iota_{a},$ (6.16) $\displaystyle\mathcal{L}_{a}$ $\displaystyle\equiv 1\otimes\mathcal{L}_{a}+\mathcal{L}_{a}\otimes 1\qquad\text{with}\quad\mathcal{L}_{a}\phi^{b}=f^{b}_{ac}\phi^{c},$ $\displaystyle\iota_{a}$ $\displaystyle\equiv 1\otimes\iota_{a}.$ An element $\alpha\in\Omega_{G}(M)$ is an invariant polynomial in the generators $\phi^{a}$ of $S(\mathfrak{g}^{*})$, with differential forms on $M$ as coefficients. This means that integration over $M$ provides a map in equivariant cohomology of the type $\int_{M}:H^{*}_{G}(M)\to S(\mathfrak{g}^{*})^{G},$ (6.17) or in other words that the integral of an equivariant form is in general a polynomial in the $\phi^{a}$. This is not quite satisfactory, as we would like an integration that generalizes the standard de Rham case, giving a map $H^{*}_{G}(M)\to\mathbb{R}$ (or $\mathbb{C}$). In the case of $G=U(1)$ we often solved this problem by setting the unique generator $\phi=-1$ (or $\phi=i$ in this conventions), thus constructing a map in the localized cohomology, $\int_{M}:H_{U(1)}^{*}(M)_{\phi}\to\mathbb{R}$. Here in the non- Abelian case, the trick of algebraic localization is not so trivial in practice, and we avoid it. An alternative and fruitful idea to saturate the $\phi$-dependence is to make them dynamical variables, and integrate over them too. Since $\phi^{a}$ can be regarded as an Euclidean coordinate over $\mathfrak{g}$, this means defining an integration over $M\times\mathfrak{g}$. As a vector space, the Lie algebra has a natural measure $d^{s}\phi$ that is unique up to a multiplicative factor. We fix that factor by choosing a (positive-definite) inner product $(\cdot,\cdot)$ on $\mathfrak{g}^{*}$ and setting666The inner product on $\mathfrak{g}^{*}$ is induced from an inner product on $\mathfrak{g}$, and when we write $(\phi,\phi)$ we really mean $\sum_{a}(\phi^{a},\phi^{a})$. This can be stated more formally defining $\phi:=\phi^{a}\otimes T_{a}\in\mathfrak{g}^{*}\otimes\mathfrak{g}$, and then letting act the inner- product on the $T_{a}$’s, normalized in order to produce a Kronecker delta. We avoid this cumbersome notation, since the action of the various inner products is always clear from the context. $\int_{\mathfrak{g}}d^{s}\phi\ e^{-\frac{\epsilon}{2}(\phi,\phi)}=\left(\frac{2\pi}{\epsilon}\right)^{s/2},$ (6.18) essentially as we did in (5.132). Since our goal is to integrate equivariant forms that have polynomial dependence on the $\phi^{a}$, or at most expressions of the form $\exp(\phi^{a}\otimes\mu_{a})$ that have exponential dependence, integrating over $\mathfrak{g}$ with the bare measure $d^{s}\phi$ would produce possible divergences. To ensure convergence of these class of functions, the _equivariant integration_ is defined as [15] $\boxed{\int_{M\times\mathfrak{g}}\alpha:=\frac{1}{\mathrm{vol}(G)}\int_{M}\int_{\mathfrak{g}}\frac{d^{s}\phi}{(2\pi)^{s}}e^{-\frac{\epsilon}{2}(\phi,\phi)}\alpha}$ (6.19) where $\epsilon$ is inserted as a regulator. Notice that in general the limit $\epsilon\to 0$ is not well-defined, for what we said above. With this enhanced definition of equivariant integration of elements of $\Omega_{G}(M)$, we can apply the equivariant localization principle to the present case. Let $\alpha\in\Omega_{G}(M)$ be an equivariantly closed form, so that $d_{C}\alpha=0$, and choose an equivariant 1-form $\beta\in\Omega_{G}^{1}(M)=\Omega^{1}(M)^{G}$. The latter is independent on $\phi^{a}$, and plays the role of the localization 1-form. By the same arguments of Section 3.1, $\alpha$ and $\alpha e^{d_{C}\beta}$ are representatives of the same equivariant cohomology class in $H^{*}_{G}(M)$, and we can deform the integral of $\alpha$ as $I[\alpha;\epsilon]:=\int_{M\times\mathfrak{g}}\alpha=\int_{M\times\mathfrak{g}}\alpha e^{td_{C}\beta}\qquad\forall t\in\mathbb{R}.$ (6.20) In particular, taking the limit $t\to\infty$, this integral localizes on the critical point set of the localization 1-form $\beta$. This can be seen simply by expanding the definition of equivariant integration from (6.20), $\displaystyle I[\alpha;\epsilon]$ $\displaystyle=\frac{1}{\mathrm{vol}(G)}\int_{M}\int_{\mathfrak{g}}\frac{d^{s}\phi}{(2\pi)^{s}}\ \alpha\exp\left(-\frac{\epsilon}{2}(\phi,\phi)-it\phi^{a}(\iota_{a}\beta)+td\beta\right)$ (6.21) $\displaystyle=\frac{1}{\mathrm{vol}(G)}\int_{M}\int_{\mathfrak{g}}\frac{d^{s}\phi}{(2\pi)^{s}}\ \alpha\exp\left(-\frac{\epsilon}{2}(\phi,\phi)-\frac{t^{2}}{2\epsilon}\sum_{a}(\iota_{a}\beta)^{2}+td\beta\right),$ where in the second line we completed the square and shifted variable in the $\phi$-integral. Since the term $td\beta$ gives a polynomial dependence on $t$ (by degree reasons, it is expanded up to a finite order), the limit $t\to\infty$ converges and makes the integral localize on the critical points of $\iota_{a}\beta$. This shows the localization property of equivariant integrals in the non-Abelian setting. If for example we suppose $\alpha$ to be independent on the $\phi^{a}$, we can perform the Gaussian integration to further simplify $I[\alpha;\epsilon]$, $I[\alpha;\epsilon]=\frac{1}{\mathrm{vol}(G)(2\pi\epsilon)^{s/2}}\int_{M}\alpha\exp\left(-\frac{t^{2}}{2\epsilon}\sum_{a}(\iota_{a}\beta)^{2}+td\beta\right).$ (6.22) We now apply the above non-Abelian localization principle to the special case in which $\alpha=\exp(\omega-i\phi^{a}\otimes\mu_{a})$, i.e. generalizing the DH formula of Section 3.3. We will suppress tensor products in the following, for notational convenience. First of all, it is straightforward to see that this form is equivariantly closed, $d_{C}e^{\omega-i\phi^{a}\mu_{a}}\propto d_{C}(\omega-i\phi^{a}\mu_{a})=-i\phi^{a}\iota_{a}\omega-i\phi^{a}d\mu_{a}=-i\phi^{a}\iota_{a}\omega+i\phi^{a}\iota_{a}\omega=0.$ (6.23) The DH oscillatory integral becomes, following the same steps of (6.20) and (6.21), $\displaystyle Z(\epsilon)$ $\displaystyle=\int_{M\times\mathfrak{g}}\frac{\omega^{n}}{n!}e^{-i\phi^{a}\mu_{a}}$ (6.24) $\displaystyle=\frac{1}{\mathrm{vol}(G)}\int_{M}\int_{\mathfrak{g}}\frac{d^{s}\phi}{(2\pi)^{s}}\exp\left(\omega-i\phi^{a}\mu_{a}-\frac{\epsilon}{2}(\phi,\phi)+t(d\beta-i\phi^{a}(\iota_{a}\beta))\right)$ $\displaystyle=\frac{1}{\mathrm{vol}(G)(2\pi\epsilon)^{s/2}}\int_{M}\exp\left(\omega-\frac{1}{2\epsilon}(\mu,\mu)-\frac{t^{2}}{2\epsilon}\sum_{a}(\iota_{a}\beta)^{2}+\frac{it}{\epsilon}\sum_{a}\mu_{a}(\iota_{a}\beta)\right),$ and it is independent of $t$. Specializing to the case $t=0$, we get $\boxed{\int_{M\times\mathfrak{g}}\frac{\omega^{n}}{n!}e^{-i\phi^{a}\mu_{a}}=\frac{1}{\mathrm{vol}(G)(2\pi\epsilon)^{s/2}}\int_{M}\exp\left(\omega-\frac{1}{2\epsilon}(\mu,\mu)\right)},$ (6.25) that shows the equivalence of the YM type partition function and the equivariant integral of the DH type! If instead we take the limit $t\to\infty$, we see that the integral localizes on the critical points of $(\iota_{a}\beta)$. With a smart choice of localization 1-form, we can show that this localization locus coincides with the critical point set of the function $S:=\frac{1}{2}(\mu,\mu)$. ###### Proof. Since $(M,\omega)$ is symplectic, it admits an almost complex structure $J\in\Gamma(T^{1}_{1}M)$ and a Riemannian metric $G\in\Gamma(T^{0}_{2}M)$ such that $\omega(\cdot,J(\cdot))=G(\cdot,\cdot)$ (see [34], proposition 12.6).777In the case of 2-dimensional YM theory, we recall that $J\equiv\star$ is the Hodge duality operator. We pick the localization 1-form $\beta:=dS\circ J=J^{\sigma}_{\nu}\partial_{\sigma}S\ dx^{\nu}=J^{\sigma}_{\nu}\sum_{a}\mu_{a}(\partial_{\sigma}\mu_{a})\ dx^{\nu},$ and the localization condition $\iota_{a}\beta=0$. Now we use the compatible metric $G$, that has components $G_{\mu\nu}=\omega(\partial_{\mu},J(\partial_{n}u))=\omega_{\mu\sigma}J^{\sigma}_{\nu}$. We consider its “inverse” $G^{-1}$ acting on $T^{*}M$ with components $G^{\mu\nu}=J^{\mu}_{\sigma}\omega^{\nu\sigma}$, where $\omega^{\nu\sigma}$ are the components of the “inverse” symplectic form, and compute the norm of the 1-form $dS$, $G^{-1}(dS,dS)=(\partial_{\mu}SJ^{\mu}_{\sigma})\omega^{\nu\sigma}\partial_{\nu}S=\beta_{\sigma}\sum_{a}\mu_{a}(\omega^{\nu\sigma}\partial_{\nu}\mu_{a})=-\sum_{a}\mu_{a}\beta_{\sigma}(T_{a})^{\sigma}=-\sum_{a}\mu_{a}(\iota_{a}\beta)=0,$ where we used the Hamiltonian equation $d\mu_{a}=-\iota_{a}\omega$ and the localization condition $\iota_{a}\beta=0$. By the non-degeneracy of $G$, this condition is equivalent to $dS=0$, that precisely identifies the critical points of $S$. ∎ Rephrasing the above result in the language of the last section, we just showed in general terms that the 2-dimensional YM partition function localizes on the moduli space of solutions of the EoM, meaning that this theory is essentially classical. We remark again that this localization locus consists of two qualitatively different types of points: those that minimize absolutely $S$, that is $\mu^{-1}(0)\subset M$, and the higher extrema with $\mu\neq 0$. The former ones in the gauge theory are the flat connections, and they give the dominant contribution to the partition function. The latter ones decays exponentially in the limit $\epsilon\to 0$ as $\sim\exp(-S/\epsilon)$. In general thus the partition function can be written as a sum of terms coming from all these disconnected regions of $M$, $Z(\epsilon)=\sum_{n}Z_{n}(\epsilon).$ (6.26) Let us consider the dominant piece $Z_{0}(\epsilon)$ coming from $\mu^{-1}(0)$, that we interpret in the gauge theory as the rough answer in the weak coupling limit, and that we can select by restricting the integration over $M$ to a suitable neighborhood $N$ of $\mu^{-1}(0)$, $Z_{0}(\epsilon)=\frac{1}{\mathrm{vol}(G)}\int_{N}\int_{\mathfrak{g}}\frac{d^{s}\phi}{(2\pi)^{s}}\exp\left(\omega-i\phi^{a}\mu_{a}-\frac{\epsilon}{2}(\phi,\phi)+td_{C}(dS\circ J)\right),$ (6.27) where we inserted the localization 1-form such that the $t\to\infty$ limit identifies the critical locus $\mu^{-1}(0)$. Cohomological arguments show that, if $G$ acts freely on $\mu^{-1}(0)$, this integral retracts on the symplectic quotient $M_{0}:=\mu^{-1}(0)/G$, giving $Z_{0}(\epsilon)=\int_{M_{0}}\exp\left(\omega_{0}+\epsilon\Theta\right)$ (6.28) for some 4-form $\Theta\in\Omega^{4}(M_{0})$. In particular, we see that in the weak coupling limit $\epsilon\to 0$, $Z_{0}(0)$ gives the volume of the symplectic quotient $M_{0}$. ###### Argument for (6.28). The precise proof is technical and it can be found in [15], we only sketch the main instructive ideas here. The neighborhood $N$ is chosen small enough to be preserved by the $G$-action, and represents the split with respect to the normal bundle we used in Section 4.2. Thus it retracts equivariantly onto $\mu^{-1}(0)$, meaning that it is homotopic to $\mu^{-1}(0)$ and that the homotopy commutes with the $G$-action. First of all we recall what we noticed at the end of the last section: if $G$ acts freely on $\mu^{-1}(0)$ the MWM theorem tells us that the symplectic form retracts on the symplectic quotient $M_{0}:=\mu^{-1}(0)/G$, so it does not contribute to the integration over the “normal directions” to $M_{0}$ in $N$. Here we are not considering a simple integration over $N$, but an equivariant integration, that provides a map $H^{*}_{G}(N)\to\mathbb{C}$. So in this case we consider the equivariantly closed extension $\tilde{\omega}=\omega-i\phi^{a}\mu_{a}$ as representative of a cohomology class $[\tilde{\omega}]\in H^{2}_{G}(M)$. When restricted over $N$, this class is the pull-back of a cohomology class $[\omega_{0}]\in H^{2}(M_{0})$, since $H^{*}_{G}(N)\cong H^{*}_{G}(\mu^{-1}(0))\cong H^{*}(M_{0})$, the first equivalence following from the retraction of $N$ onto $\mu^{-1}(0)$ and the second from the fact that the $G$-action is free on $\mu^{-1}(0)$ (these properties were explained in Chapter 2). Thus we can substitute in the integral $\tilde{\omega}\mapsto\omega_{0}$ without changing the final result. The same kind of argument works for the term $\frac{1}{2}(\phi,\phi)$. It is easy to check that this is both $G$-invariant and equivariantly closed, so it represents an element $\left[\frac{1}{2}(\phi,\phi)\right]\in H^{4}_{G}(M)$. When we restrict it to $N$, as above, this class is the pull-back of some class $[\Theta]\in H^{4}(M_{0})$, and we can make the substitution $\frac{1}{2}(\phi,\phi)\mapsto\Theta$ in the integral without changing the final result. Since both $\omega_{0}$ and $\Theta$ are standard differential forms over $M_{0}$ and thus independent of $\phi$, the integration over $\mathfrak{g}$ goes along only with the remaining term $d_{C}(dS\circ J)$. We already know that on $\mu^{-1}(0)\subset N$ this term is zero, so one has to show that its integral over the normal directions to $\mu^{-1}(0)$ in $N$ produces a trivial factor of 1. In [15] it is proven that $\frac{1}{\mathrm{vol}(G)}\int_{F}\int_{\mathfrak{g}}\frac{d^{s}\phi}{(2\pi)^{s}}\exp\left(td_{C}(dS\circ J)\right)=1,$ where $F$ is any fiber of the normal bundle to $\mu^{-1}(0)$ in $N$. From this (6.28) follows. ∎ ###### Example 6.2.1 (The height function on the 2-sphere, again). Beside the main application of Witten’s localization principle to non-Abelian gauge theories, we try now to apply this new formalism to the old and simple example of the height function on the 2-sphere, to compare it with the results obtained in Chapter 3. Setting $G=U(1)$, $M=\mathbb{S}^{2}$, $\mu=\cos(\theta)$ and $\omega=d\cos(\theta)\wedge d\varphi$, the equivariant integration (6.19) of the DH oscillatory integral gives $\displaystyle Z(\epsilon)$ $\displaystyle=\int_{\mathbb{S}^{2}\times\mathfrak{g}}\omega e^{-i\phi\mu}=\frac{1}{2\pi}\int_{0}^{2\pi}d\varphi\int_{-1}^{+1}d\cos\theta\int_{-\infty}^{+\infty}\frac{d\phi}{2\pi}\exp\left(-i\phi\cos\theta-\frac{\epsilon}{2}\phi^{2}\right)$ $\displaystyle=\frac{1}{\sqrt{2\pi\epsilon}}\int_{-1}^{+1}dx\ e^{-\frac{x^{2}}{2\epsilon}}=1-2I(\epsilon),$ where $I(\epsilon):=\int_{1}^{\infty}\frac{dx}{\sqrt{2\pi\epsilon}}\exp(-x^{2}/2\epsilon)$ is a trascendental error function. The three terms in the final result for $Z(\epsilon)$ (two of which are equal to $-I(\epsilon)$) correspond to the contributions of the extrema of $(\cos\theta)^{2}$: the two maxima at $\theta=0,\pi$ contribute with $-I(\epsilon)$ and the minimum at $\theta=\pi/2$ contributes with $1$. The latter is the dominant piece when $\epsilon\to 0$, since $I(0)=0$. We see that in general the modified equivariant integration of this new formalism gives an incredibly complicated answer, when compared to the simple result of Example 3.2.1 obtained via the usual equivariant localization principle. ###### Remark. We notice that we could have expressed equivalently the whole dissertation above in supergeometric language, since we discussed in Section 4.1 that integration over $M$ is equivalent to integration over $\Pi TM$. In these terms, maybe more common in QFT, we can introduce coordinates $(x^{\mu},\psi^{\mu},\phi^{a})$ over $\Pi TM\times\mathfrak{g}$, where $\psi^{\mu}:=dx^{\mu}$ are Grassmann-odd, and interpret elements of $\Omega_{G}(M)$ as elements of $C^{\infty}(\Pi TM\times\mathfrak{g})^{G}$. An equivariant form $\alpha$ is thus (locally) a $G$-invariant function of $(x,\psi,\phi)$. For example, the Cartan differential and the definition of equivariant integration become $\displaystyle d_{C}$ $\displaystyle=\psi^{\mu}\frac{\partial}{\partial x^{\mu}}-i\phi^{a}T_{a}^{\mu}\frac{\partial}{\partial\psi^{\mu}},$ (6.29) $\displaystyle\int_{M\times\mathfrak{g}}\alpha$ $\displaystyle:=\frac{1}{\mathrm{vol}(G)}\int d^{2n}xd^{2n}\psi\frac{d^{s}\phi}{(2\pi)^{s}}\ \alpha(x,\psi,\phi)e^{-\frac{\epsilon}{2}(\phi,\phi)}.$ ### 6.3 “Cohomological” and “physical” YM theory In this section we are going to review the relation between 2-dimensional YM theory that we described in Section 6.1 and a topological field theory (TFT) that can be viewed as its “cohomological” counterpart. We can translate almost verbatim the general principles that we discussed in the last section, setting $M\mapsto\mathcal{A}(P),\quad\omega\mapsto\Omega,\quad G\mapsto\mathcal{G}(P),\quad\mu\mapsto F,$ (6.30) while we regard $G$ as a compact connected Lie group that acts as the gauge group on the principal bundle $P\to\Sigma$ over a Riemann surface $\Sigma$. The moment map is formally equivalent to the curvature $F$ if we identify $Lie(\mathcal{G}(P))\cong Lie(\mathcal{G}(P))^{*}$ through an inner product on $\mathfrak{g}$, as in (6.8). The non-Abelian localization principle of the last section, if used in reverse, already showed that an equivalent way to express the standard YM theory is through a “first-order formulation” $S[A,\phi]=-\int_{\Sigma}\mathrm{Tr}\left(i\phi F^{A}+\frac{\epsilon}{2}\phi\star\phi\right)$ (6.31) where we consider $\phi\in\Omega^{0}(\Sigma;\mathrm{ad}(P))\cong Lie(\mathcal{G}(P))$, and $F^{A}\in\Omega^{2}(\Sigma;\mathrm{ad}(P))$ is the curvature of $A$.888More precisely, we should say that $A^{a}_{\mu}$ and $\phi^{a}$ are _coordinates_ functions on $\mathcal{A}(P)\times Lie(\mathcal{G}(P))$, so they effectively are elements of $C^{\infty}(\mathcal{A}(P))\otimes Lie(\mathcal{G}(P))^{*}$. This caveat will be logically important in the following, and it goes along with the _functor of points_ approach we used in the supergeometric discussion of Chapter 4. This is essentially what is written in (6.25), where on the LHS we have the first-order action (the $\epsilon$ dependence is contained in the equivariant integration), and on the RHS we have the standard YM action $S[A]=\frac{1}{2\epsilon}(F,F)_{A}$. The first-order formulation has the quality of showing very clearly the weak coupling limit behavior when $\epsilon\to 0$, that is less obvious in the standard formulation. In this limit, the theory becomes topological, in the sense that the action does not depend on the metric anymore (the metric appears in the Hodge duality $\star$), $S_{\epsilon\to 0}[A,\phi]=-\int_{\Sigma}\mathrm{Tr}(i\phi F^{A}).$ (6.32) This theory is called “BF model”, and it is the prototype of a TFT of Schwarz- type. The YM theory can thus be seen as a “regulated version” of a truly topological field theory.999Notice that, although YM theory is clearly dependent on the metric of $\Sigma$, in 2 dimensions it shows a very “weak” dependence to it. In fact, in this dimensionality the action can be simplified as (suppressing the constants and the Lie algebra inner product) $\displaystyle\int F\wedge\star F$ $\displaystyle=\int g^{\mu\rho}g^{\nu\sigma}F_{\mu\nu}F_{\rho\sigma}\sqrt{|g|}d^{2}x=\int(F_{01})^{2}g^{\mu\rho}g^{\nu\sigma}\varepsilon_{\mu\nu}\varepsilon_{\rho\sigma}\sqrt{|g|}d^{2}x=$ $\displaystyle=\int(F_{01})^{2}|g^{-1}|\sqrt{|g|}d^{2}x=\int(F_{01})^{2}|g|^{-1/2}d^{2}x,$ so the metric does not appear through its components, but only in the invariant quantity $\det(g)$. At least classically, it is intuitive from the EoM with respect to $\phi$ that the only contribution to the classical solutions comes from the moduli space of flat connections, where $F^{A}=0$ up to gauge transformations. It is not trivial, though, to infer that this is all the theory has to offer also at the quantum level, that is essentially the result we showed in general terms in the last section, via the localization principle applied to the path integral $Z(\epsilon\to 0)$. In this section we discuss, following [15], how the localization principle can be translated in the language of TFT, in order to give a more physical interpretation of the abstract mathematical results that we discussed in finite dimensions. In particular, we will see that the BF model partition function can be recovered as an expectation value in a TFT, and that this ensures its localization properties onto the moduli space of flat connections. The regulated version at $\epsilon\neq 0$ will not follow precisely this behavior, as we already know that higher extrema of the YM action contribute to the partition function $Z(\epsilon)$, but these will have a nice interpretation in terms of the moduli space. #### Intermezzo: TFT This is a good moment to explain briefly in more general terms what one usually means by TFT, and how localization enters in this subject. Traditionally, TFT borrows the language of BRST formalism for quantization of gauge theories, as many examples of topological theories arise from that context. Recall that the standard BRST quantization procedure is based on the definition of a differential, the “BRST charge” $Q$, that acts on an extended graded field space, whose grading counts the “ghost number” (in other words, $Q$ is an operator of degree $\mathrm{gh}(Q)=+1$). The BRST charge represents an infinitesimal supersymmetry transformation, that squares to zero in the gauge-fixed theory. On the Hilbert space, physical states are those of ghost number zero, and that are annihilated by the BRST charge, $Q|\text{phys}\rangle=0,\quad\mathrm{gh}|\text{phys}\rangle=0.$ (6.33) The action of $Q$ on the field space is often denoted with a Poisson bracket- like notation, $\Phi\mapsto\delta_{Q}\Phi:=-\\{Q,\Phi\\}$ for any field $\Phi$. By gauge invariance of the vacuum, $Q|0\rangle=0$, and so for any operator $\mathcal{O}$ one has that $\langle 0|\\{Q,\mathcal{O}\\}|0\rangle=0$. For a QFT being “topological”, in physics one usually means that all its quantum properties are independent from a choice of a metric on the base space $M$.101010Here the word “topological” is somewhat overused. Mathematically, a topological space consists of a set $M$ and a _topology_ $\mathcal{O}_{M}$, that is roughly the set of all “open neighborhoods” in $M$. In QFT one almost always works on base spaces that have the structure of a _manifold_ of some kind (smooth, complex, …), so that it allows for the presence of an _atlas_ $\mathcal{A}_{M}$ of charts that identifies it locally as $\mathbb{R}^{n}$ for some (constant) $n$. A manifold is thus a triple $(M,\mathcal{O}_{M},\mathcal{A}_{M})$, and the choice of a metric is only on top of this structure. So metric-independence does not generically mean that the QFT describes only the topology of $M$, but it can depend on the choice of a smooth (or complex, …) structure on $M$. This is rephrased in the requirement that the partition function of the theory should be metric- independent. Assuming the path integral measure to be metric-independent and $Q$-invariant (so that the BRST symmetry is non anomalous), the variation with respect to the metric of the partition function is $\delta_{g}Z\propto\int_{\mathcal{F}}[D\Phi]e^{-S[\Phi]}\delta_{g}S[\Phi]=\int_{\mathcal{F}}[D\Phi]e^{-S[\Phi]}\left(\int_{M}d^{n}x\sqrt{|g|}\delta g^{\mu\nu}T_{\mu\nu}\right)\propto\langle 0|T_{\mu\nu}|0\rangle,$ (6.34) so a suitable definition of TFT is the one that requires the energy-momentum tensor $T_{\mu\nu}$ to be a BRST variation, $T_{\mu\nu}=\\{Q,V_{\mu\nu}\\}$ for some operator $V_{\mu\nu}$. This would ensure $\delta_{g}Z=0$ for what we said above. Collecting the above remarks, we can give the following “working definition” [106]. A _Topological Field Theory_ is a QFT defined over a $\mathbb{Z}$-graded field space $\mathcal{F}$, with a nilpotent operator $Q$ (i.e. a _cohomological vector field_ on $\mathcal{F}$), and a $Q$-exact energy-momentum tensor $T_{\mu\nu}=\\{Q,V_{\mu\nu}\\}$, for some $V_{\mu\nu}\in C^{\infty}(\mathcal{F})$. _Physical states_ are defined to be elements of the $Q$-cohomology of $\mathcal{F}$ in degree zero, $|\text{phys}\rangle\in H^{0}(\mathcal{F},Q)$.111111We stress that, if we see $\mathcal{F}$ as a graded extension of an original field space $\mathcal{F}_{0}$ acted upon by gauge transformations, the $Q$-cohomology of $\mathcal{F}$ is exactly the analogous of the gauge-equivariant cohomology $H_{gauge}^{*}(\mathcal{F}_{0})$, computed in the Cartan model with Cartan differential $Q$. ###### Remark. * • $Q$ is called “BRST charge” or “operator”, but in general it can be every supersymmetry charge (as it is a cohomological vector field). We saw examples in the last chapter where $Q=\delta_{susy}+\delta_{BRST}$, where $\delta_{BRST}$ is the actual gauge-supersymmetry, and $\delta_{susy}$ is a Poincaré-supersymmetry. * • The one above is a good “working definition” for most examples, but it is not completely adequate in all cases. Indeed, there are examples of QFT where $T_{\mu\nu}$ fails to be BRST-exact, but nonetheless one can still establish the topological nature of the model. We do not need to treat any example of this kind, so we refer to [106] for more details. * • We already encountered an example of TFT in Section 5.2, i.e. supersymmetric QM. There, we called it topological because its partition function was determined by topological invariants of the base space, here we point out that it indeed fits into this general discussion. In fact, in (5.69) we saw that its action can be expressed as $S=\\{Q,\Sigma\\}$, where $Q\equiv Q_{\dot{x}}$ was the $U(1)$-Cartan differential on the loop space, and $\Sigma$ some other loop space functional. The $Q$-exactness of the energy-momentum tensor follows from this. * • If the energy-momentum tensor is $Q$-exact, in particular the Hamiltonian satisfies $\langle\text{phys}|H|\text{phys}\rangle=\langle\text{phys}|T_{00}|\text{phys}\rangle=0.$ (6.35) This means that the energy of any physical state is zero, and thus the TFT does not contain propagating degrees of freedom. It can only describe “topological” properties of the base space. TFTs fall in two main broad categories. The first one is constituted by the so-called TFT of _Witten-type_ (or also _cohomological TFT_). Their defining property is to have a $Q$-exact quantum action, $S_{q}=\\{Q,V\\}$ (6.36) for some operator $V$. As the case of supersymmetric QM, the energy-momentum tensor is automatically $Q$-exact too, $T_{\mu\nu}=\frac{2}{\sqrt{|g|}}\left\\{Q,\delta V/\delta g_{\mu\nu}\right\\}.$ (6.37) From the equivariant point of view, these theories have a very simple localization property. The action is representative of the trivial $Q$-equivariant cohomology class, so the partition function can be in fact written as $Z\propto\int_{\mathcal{F}}[D\Phi]e^{-tS_{q}[\Phi]}$ (6.38) for any $t\in\mathbb{R}$, by the standard argument of supersymmetric localization. In this case $\mathcal{F}$ is non-compact and so simply taking $t=0$ is not really allowed, but the limit $t\to\infty$ is perfectly defined, producing the path integral localization onto the space of solutions of the classical EoM of $S_{q}$. This means that all TFT of Witten-type are completely determined by their semiclassical approximation! The second main class of TFT is called of _Schwarz-type_ (or also _quantum TFT_). In this case one starts with a classical action $S$ that is metric- independent, so that the classical energy-momentum tensor is zero. The usual BRST quantization of this action produces a quantum action of the type $S_{q}=S+\\{Q,V\\}$ for some $V$, and again $T_{\mu\nu}=\frac{2}{\sqrt{|g|}}\left\\{Q,\delta V/\delta g_{\mu\nu}\right\\},$ (6.39) since the classical piece does not contribute. Chern-Simons theory and BF theory are of this type. Regarding the second remark above, we mention that some Schwarz-type TFTs fail to respect the $Q$-exactness property of $T_{\mu\nu}$, for example in the case of non-Abelian BF theories in dimension $d>3$ (this happens essentially because one can ensure nilpotency of the BRST charge only on-shell, and moreover $Q$ has to be defined in a metric-dependent way). As we anticipated before we are interested in BF theories, an example of Schwarz-type TFT, and in particular in their 2-dimensional realization. On a base space $\Sigma$ of generic dimension $d$, the classical action of the BF theory is $S_{BF}[A,B]:=-\int_{\Sigma}\mathrm{Tr}(B\wedge F^{A}),$ (6.40) where $F^{A}$ is the curvature 2-form of a connection $A$, and $B$ is a $(d-2)$-form with values in the adjoint bundle. The BRST charge is the usual gauge-supersymmetry. In the 2-dimensional case, $B\equiv i\phi\in\Omega^{0}(\Sigma;\mathrm{ad}(P))$, and the naive BRST quantization of this theory has no problems (as we mentioned, in more than 3 dimensions the definition of $Q$ have to be modified and things complicate). The quantum action is $\displaystyle S_{q}[A,B,\pi,b,c]$ $\displaystyle=-\int_{\Sigma}\mathrm{Tr}\left(BF^{A}+\pi(\nabla^{A}\star A)+b(\nabla^{A}\star\nabla^{A})c\right)$ (6.41) $\displaystyle=S_{BF}[A,B]+\\{Q,V\\},$ $\displaystyle\text{with}\quad V$ $\displaystyle:=-\int_{\Sigma}\mathrm{Tr}\left(b(\nabla^{A}\star A)\right),$ where we introduced a ghost $c$, an anti-ghost $b$ and an auxiliary field $\pi$ with the BRST transformations properties $\delta_{Q}A=\nabla^{A}c,\quad\delta_{Q}c=-\frac{1}{2}[c,c],\quad\delta_{Q}b=\pi,\quad\delta_{Q}\pi=0,\quad\delta_{Q}B=-[B,c].$ (6.42) Even though this theory is not of Witten-type, so it has no direct localization onto the subspace of classical solutions, it turns out that this is still the case. Traditionally, this is shown finding a suitable redefinition of coordinates in field space that trivializes the bosonic sector of the action (meaning that there are no derivatives acting on bosonic fields), and whose Jacobian cancels in the path integral with the 1-loop determinant obtained by integrating out the fermions (in this case, the ghosts fields). This is called _Nicolai map_ , and for the 2-dimensional BF model is given by the redefinitions [106, 107] $\xi(A):=F^{A},\qquad\eta(A):=\nabla^{A_{c}}\star A_{q},$ (6.43) where $A:=A_{c}+A_{q}$ is the expansion of the gauge field around a classical (on-shell) solution $A_{c}$. Assuming the fermions being integrated out, the path integrals over $B$ and $\pi$ exactly identify the space of zeros of $\xi,\eta$, that is the moduli space of solutions to $F^{A}=0$ up to gauge transformations, $\mathcal{A}_{0}$. We do not pursue this direction further, but summarize the approach taken in [15], more related to localization. #### Cohomological approach Thinking in equivariant cohomological terms, one can expect to show the localization of the BF model (and of its “regulated” version, _i.e._ YM theory) finding a suitable “localization 1-form” $V^{\prime}$. This has to be such that the deformation of the action given by $t\\{Q,V^{\prime}\\}$ induces the path integral to localize in this subspace when $t\to\infty$, analogously to what we did for example in Section 5.1, and also to the discussion of the last section in the finite-dimensional case. This is exactly what is shown in [15], where the partition function of the model of interest is found as an expectation value inside a cohomological TFT, proving automatically its localization behavior. This cohomological TFT is constructed in a way such that the action of the BRST operator $Q$ coincides with the Cartan differential $d_{C}$ arising in the symplectic formulation of 2-dimensional YM theory. Using the more common supergeometric language of Sections 5.1, 5.2 and recalled in (6.29) (in finite dimensions), we move from the field space $\mathcal{A}(P)\times Lie(\mathcal{G}(P))$ to $\mathcal{F}:=\Pi T\mathcal{A}(P)\times Lie(\mathcal{G}(P))$, introducing the graded coordinates $(A_{\mu}^{a},\psi_{\mu}^{a},\phi^{a})$ and regarding $d_{C}$ as a supersymmetry (BRST) transformation, $d_{C}\equiv-\\{Q,\cdot\\}=\int_{\Sigma}d\Sigma\left(\psi_{\mu}^{a}\frac{\delta}{\delta A^{a}_{\mu}}-i(\underline{\phi}^{a})^{\mu}\frac{\delta}{\delta\psi_{\mu}^{a}}\right),$ (6.44) where we recall from (6.3) that the fundamental vector field associated to the action of a Lie algebra element $\phi\in Lie(\mathcal{G}(P))$ is $\underline{\phi}=\nabla\phi$. The BRST transformation of every field follows the rule $\delta_{Q}\Phi=-\\{Q,\Phi\\}\equiv d_{C}\Phi$, and on the coordinates we have $\delta_{Q}A=\psi,\qquad\delta_{Q}\psi=-i\nabla\phi,\qquad\delta_{Q}\phi=0.$ (6.45) The ghost numbers of the elementary fields are $\mathrm{gh}(A,\psi,\phi)=(0,1,2)$. Other multiplets could be added as well, but this is the basic one we start with. Before describing in more detail the particular cohomological theory and its relation with the physical YM theory, we summarize the strategy that has been followed. One starts with a suitable cohomological TFT with action $S_{c}=\\{Q,V\\},$ (6.46) for some operator $V$, that ensures the localization onto the moduli space $\mathcal{A}_{0}$ of flat connections. Then the mapping to the physical theory is done by the common equivariant localization procedure. The TFT is deformed adding a cohomologically trivial localizing action, $S(t)=S_{c}+t\\{Q,V^{\prime}\\}=\\{Q,V+tV^{\prime}\\},$ (6.47) for some gauge invariant operator $V^{\prime}$ that forces only the interesting YM multiplet $(A,\psi,\phi)$ to survive. Since the field space here is non-compact, some additional care must be taken in claiming the $t$-independence of the deformed theory. In particular, the new term must not introduce new fixed points of the $Q$-symmetry, that would contribute to the localization locus of the resulting theory. These fixed points, if presents in the theory at $t\neq 0$, can be interpreted as “flowing from infinity” in the moduli space (since this is, as just remarked, non-compact).121212This will be exactly the case in going to the YM theory with $\epsilon\neq 0$, where the new fixed points are just the higher extrema of the action $S=\frac{1}{2}(F,F)$. If this is not the case, one can infer properties of the “physical” theory at $t=\infty$ by making computations in the cohomological one at $t=0$.131313Notice that this is logically the opposite of what one does usually in using localization. Here the “easy” theory is the cohomological one at $t=0$, while the more difficult (but more interesting) is the one at $t\neq 0$. ##### The cohomological theory The cohomological theory considered in [15] makes use of the additional multiplets $(\lambda,\eta)$ and $(\chi,H)$ with the transformation properties $\delta_{Q}\lambda=\eta,\qquad\delta_{Q}\eta=-i[\phi,\lambda],\qquad\delta_{Q}\chi=-iH,\qquad\delta_{Q}H=[\phi,\chi].$ (6.48) The extended field space and ghost numbers are $\begin{array}[]{rccccc}\mathcal{F}=&\underbrace{\Pi T\mathcal{A}(P)\times Lie(\mathcal{G}(P))}&\times&\underbrace{\left(\Omega^{0}(\Sigma;\mathrm{ad}(P))\right)^{2}}&\times&\underbrace{\left(\Omega^{0}(\Sigma;\mathrm{ad}(P))\right)^{2}}\\\ &(A,\psi,\phi)&&(\lambda,\eta)&&(\chi,H)\\\ \mathrm{gh}=&(0,1,2)&&(-2,-1)&&(-1,0).\end{array}$ (6.49) Of course the definition of the BRST operator will be extended from (6.44), but we do not need it explicitly. The operator $V$ defining the TFT is chosen to be $V=\frac{1}{h^{2}}\int_{\Sigma}d\Sigma\ \mathrm{Tr}\left(\frac{1}{2}\chi(H-2(\star F^{A}))+g^{\mu\nu}(\nabla^{A}_{\mu}\lambda)\psi_{\nu}\right),$ (6.50) where $h\in\mathbb{R}$ is a parameter from which the theory is completely independent (it is analogous to $1/t$ appearing in (6.38)), that can be interpreted as the “coupling constant” of the TFT. Computing the cohomological action $S_{c}$ one sees that the $H$ field plays an auxiliary role, and can be eliminated setting $H=\star F$. Analyzing then the theory in the $h\to 0$ limit (its semiclassical “exact” approximation), the localization locus is identified by the BRST-fixed point (we refer to [15] for the details) $\delta_{Q}\chi=0\quad(\Rightarrow F=0),\qquad\delta_{Q}\psi=0\quad(\Rightarrow\nabla\phi=0).$ (6.51) This is analogous to the usual situation in Poincaré-supersymmetric theories, as discussed in Chapter 5, where the localization locus was always identified by the subcomplex of BPS configurations, given by the vanishing of the variation of the fermions. Here the “fermions” are the fields with odd ghost number. This means that the final moduli space contains $\mathcal{A}_{0}$, plus maybe some contributions from the zero-modes of the other bosonic (even ghost number) fields, $\lambda$ and $\phi$. ##### $t\neq 0$ deformation The deformation (6.47) can be made in order to reduce effectively the field content of the theory to the YM multiplet only, $(A,\psi,\phi)$. In particular, to eliminate the non-trivial presence of the field $\lambda$ from the contributions to the localization locus one can consider $V^{\prime}=-\frac{1}{h^{2}}\int_{\Sigma}d\Sigma\ \mathrm{Tr}\left(\chi\lambda\right).$ (6.52) Computing the deformed action $S(t)$, one sees that for $t\neq 0$ all the additional fields $H,\chi,\lambda,\eta$ can be integrated out (again, some more details on the technical passages can be found in [15]). In particular, the EoM for $H$ and $\lambda$ are $H=0,\qquad\lambda=-\frac{1}{t}(\star F).$ (6.53) This is already the sign that the localizing term $V^{\prime}$ qualitatively changed the localization property of the theory. In fact, we see that for $t=0$ we do not have an algebraic equation for $\lambda$, but (6.53) reduces to the solution $F=0,H=0$ of the cohomological theory. The theories defined by $S(t)$ and $S_{c}$ may be thus different, but the failure of their equivalence can only come from new components of the moduli space that flow in from infinity for $t\neq 0$; the contribution of the “old” component must be independent of $t$. Taking the limit $t\gg 1$, the dominant contribution to the deformed action is (suppressing the $A$-dependence) $\displaystyle S(t\gg 1)$ $\displaystyle=-\frac{1}{t}\left\\{Q,\int_{\Sigma}d\Sigma\ \mathrm{Tr}\left(\psi^{\mu}\nabla_{\mu}(\star F)\right)\right\\}$ (6.54) $\displaystyle=\frac{1}{t}\int_{\Sigma}d\Sigma\ \mathrm{Tr}\Bigl{(}i\nabla_{\mu}\phi\nabla^{\mu}(\star F)-(\star F)[\psi_{\mu},\psi^{\mu}]+\nabla_{\mu}\psi^{\mu}\epsilon^{\nu\sigma}\nabla_{\nu}\psi_{\sigma}\Bigr{)}.$ The main point is that the $\phi$-EoM is actually equivalent to the YM equation $\nabla\star F=0$. This means that the moduli space of the deformed theory contains the moduli space of the standard YM theory, and indeed includes all the higher extrema corresponding to non-flat connections. These solutions with $F\neq 0$ have $\lambda\sim-1/t$, and thus their contribution to the path integral goes roughly as $\exp(-1/t)$, as expected. When $t=0$, the cohomological theory is recovered and the only contribution to the moduli space is given by the flat connections. ##### Connection with 2-dimensional YM theory We already argued that the deformed TFT gained all the YM spectrum “flowing from infinity in the moduli space”, but it remains to see how one can get practically the YM (and BF) partition function from the theory defined by $S(t)$. This is simply obtained by another deformation of the exponential in the partition function: we notice that the YM action is gauge invariant, so it is meaningful to compute the expectation value of $e^{S_{YM}}$ in the TFT. Thus we consider an exponential operator of the form $\begin{gathered}\exp\left(\omega_{0}+\epsilon\Theta\right)\\\ \text{with}\quad\omega_{0}:=\int_{\Sigma}\mathrm{Tr}\left(i\phi F+\frac{1}{2}\psi\wedge\psi\right),\qquad\Theta:=\frac{1}{2}\int_{\Sigma}\mathrm{Tr}(\phi\star\phi).\end{gathered}$ (6.55) Since the quantity $\left\langle\exp\left(\omega_{0}+\epsilon\Theta\right)\right\rangle_{t}\propto\int_{\Pi T\mathcal{A}(P)\times Lie(\mathcal{G}(P))}DAD\psi D\phi\ \exp\left(\omega_{0}+\epsilon\Theta-S(t)\right)$ (6.56) is well defined for $t\to\infty$, we can actually take $t=\infty$ and drop $S(\infty)=0$ (recall that the path integral is independent on the actual value of $t$), getting exactly the YM partition function $\left\langle\exp\left(\omega_{0}+\epsilon\Theta\right)\right\rangle_{t}\propto\int_{\Pi T\mathcal{A}(P)\times Lie(\mathcal{G}(P))}DAD\psi D\phi\ \exp\left(\omega_{0}+\epsilon\Theta\right)\propto Z(\epsilon),$ (6.57) up to some normalization constant. In the limit $\epsilon\to 0$, when only the BF model survives, the path integral over $\phi$ produces the constraint $\delta(F)$, so localizing the expectation value onto the space of flat connections. This means that, although we started from different theories $S_{c}\cong S(0)$ and $S(t)$, this particular expectation value satisfies $\left\langle\exp\left(\omega_{0}\right)\right\rangle_{t}=\left\langle\exp\left(\omega_{0}\right)\right\rangle_{t=0}$ (6.58) and the BF model is recovered as an expectation value in the cohomological theory. This gives another interpretation to the topological behavior of the BF model, and a measure of the failure of 2-dimensional YM theory in being topological. Concluding, we only point out that the the operators $\omega_{0}$ and $\Theta$ are precisely the infinite-dimensional realization in this example of the general expressions in (6.28). In fact, the symplectic 2-form on $\mathcal{A}(P)$ $\Omega=\int_{\Sigma}\mathrm{Tr}(\psi\wedge\psi)$ (6.59) only serves to have a formal interpretation of the measure $DAD\psi e^{\Omega}$, since the field $\psi$ is really a spectator in the action $S[A,\psi,\phi]=-\int_{\Sigma}\mathrm{Tr}\left(i\phi F+\frac{\epsilon}{2}\phi\star\phi+\frac{1}{2}\psi\wedge\psi\right).$ (6.60) ### 6.4 Localization of 2-dimensional YM theory As we said at the beginning of the chapter, 2-dimensional YM theory is an exactly solvable theory, whose partition function can be expressed in closed form, for example by group characters expansion methods [20, 15]. This makes it possible to compare results from the localization formalism, and obtain a new geometric interpretation of the already present solution of the theory. In general, its partition function on a Riemann surface $\Sigma$ of genus $g$, with a simply-connected gauge group $G$, is given as $Z(\epsilon)=(\mathrm{vol}(G))^{2g-2}\sum_{R}\frac{1}{\dim(R)^{2g-2}}e^{-\epsilon\tilde{C}_{2}(R)},$ (6.61) where the sum runs over the representations $R$ of $G$, and $\tilde{C}_{2}(R)$ is related to the quadratic Casimir $C_{2}(R):=\sum_{a}\mathrm{Tr}_{R}(T_{a}T_{a})$ of the representation $R$ by some normalization constant. For not simply-connected $G$ this formula has to be slightly modified (see [15]).141414Simply-connectedness implies that the principal $G$-bundle $P\to\Sigma$ has to be trivial. When one drops this condition, the triviality is not ensured and contributions to the formula appear due to singular points in $\Sigma$ for the connection. We are not interested in reviewing the proof of (6.61) in general, but we present a quick argument for a simple example, that already contains the logic behind it. ###### Example 6.4.1 (YM theory with genus $g=1$). We quickly motivate the result for the YM partition function on a genus 1 surface. We can think of this surface as a disk with boundary, that is homeomorphic to a sphere with one hole. The radial direction is identified with the interval $[0,T]$, and the angular coordinate as $[0,L]$ with the edges identified. It is more natural to compute the partition function of the theory in the Hamiltonian formulation, $Z=\mathrm{Tr}_{\mathcal{H}}\mathcal{P}\exp\left(-\int_{0}^{T}dt\ H(t)\right)$ (6.62) up to possible normalization factors, where $\mathcal{H}$ is the Hilbert space of the system. To this end, let us consider the canonical quantization of the YM action. To make sense of the partition function we must fix a gauge, and we do this setting $A_{t}=0$ (temporal gauge). In this gauge the action simplifies as $S[A]=-\frac{1}{2\epsilon}\int dxdt\ \mathrm{Tr}(F_{01}^{2})=\frac{1}{2\epsilon}\int dxdt\ (\partial_{t}A^{a}_{x})^{2},$ (6.63) where we expanded $A_{\mu}=A^{a}_{\mu}T_{a}$ with respect to the generators of $\mathfrak{g}$, and suppressed the inner product implicitly summing over the Lie algebra indices. We see that the only non-zero canonical momentum is $\Pi^{a}_{x}=(1/\epsilon)\partial_{t}A^{a}_{x}$, acting on the Hilbert space as $\Pi^{a}_{x}(t,x)\mapsto\frac{\delta}{\delta A_{x}^{a}(t,x)}$. The canonical Hamiltonian in temporal gauge is thus $H(t)=\epsilon\int_{0}^{L}dx\ (\Pi^{a}_{x}(t,x))^{2}\mapsto\epsilon\int_{0}^{L}dx\ \frac{\delta}{\delta A_{x}^{a}(t,x)}\frac{\delta}{\delta A_{x}^{a}(t,x)}.$ (6.64) The Hilbert space $\mathcal{H}$ can be considered to consist of gauge invariant functions $\Psi(A)$. The only gauge invariant data obtained from the gauge field at any point $p\in\Sigma$ is its holonomy, $U_{p}[A]:=\mathcal{P}\exp\left(\oint_{C(p)}A\right)\quad\in G,$ (6.65) where $C(p)$ is a loop about $p$. In the case of the one holed-sphere, all loops are homotopic to the one on the boundary, so $\Psi\in\mathcal{H}$ must be an invariant function of $U=\mathcal{P}\exp\int_{0}^{L}dxA_{1},$ (6.66) and independent of $t\in[0,T]$. Any invariant function must be expandible in characters of representations of $G$, so $\Psi(U)=\sum_{R}\Psi_{R}\mathrm{Tr}_{R}(U)$, where $\Psi_{R}\in\mathbb{C}$ and $\mathrm{Tr}_{R}(U)$ is the Wilson loop in the representation $R$ of $G$. We notice that the basis functions $\chi_{R}(U):=\mathrm{Tr}_{R}(U)$ diagonalize the Hamiltonian, since $H\chi_{R}(U)=\epsilon\mathrm{Tr}_{R}\int_{0}^{1}dx\ T_{a}T_{a}\ \mathcal{P}\exp\int_{0}^{L}dxA_{1}=\epsilon LC_{2}(R)\chi_{R}(U),$ (6.67) where $C_{2}(R):=\mathrm{Tr}_{R}(T_{a}T_{a})$ is the quadratic Casimir in the representation $R$, a time-independent eigenvalue of $H$. Via this diagonalization the partition function is easily computed, $Z=\sum_{R}e^{-TL\epsilon c(R)},$ (6.68) matching (6.61) for $g=1$. All the geometric information about $\Sigma$ that enters in $Z$ is its total area $TL$, and any other local property. Notice that in the topological limit $\epsilon\to 0$, the Hamiltonian vanishes (as the theory has no propagating degrees of freedom) and the partition function simplifies further. From the localization formalism discussed in the last sections, we expect the partition function to be of the type $Z(\epsilon)=Z_{0}(\epsilon)+\sum_{n}Z_{n}(\epsilon),$ (6.69) with $Z_{0}(\epsilon)$ representing the contribution from the moduli space $\mathcal{A}_{0}$ of flat connections, such that $Z_{0}(0)\sim\mathrm{vol}(\mathcal{A}_{0})$, and the other $Z_{n}(\epsilon)$ coming from contributions of the higher extrema of the YM action, such that $Z_{n}(\epsilon)\sim\exp(-1/\epsilon)$ in the weak coupling limit $\epsilon\ll 1$. Using cohomological arguments, in [15] (also nicely reviewed in [102]) it was shown how to recover the general features of (6.61), and in particular how to interpret it in terms of an $\epsilon$-expansion at weak coupling, in relation to the expected form (6.69). The detailed derivation is cumbersome and requires some more technical background, so we refer to the article for it, but the logic is essentially the same as for the discussion at the end of Section 6.2. The strategy is the following. Any solution to the YM EoM identifies a disconnected region $\mathcal{S}_{n}\subset\mathcal{A}(P)$. For every such region, one fixes a small neighborhood $N_{n}$ around $\mathcal{S}_{n}$ that equivariantly retracts onto it. The technically difficult passage is to perform the integral over the “normal directions” to $\mathcal{S}_{n}$ in $N_{n}$, and then reduce it on the moduli space $\mathcal{A}_{n}:=\mathcal{S}_{n}/\mathcal{G}(P)$. The main difficulty is that in general the MWM theorem (or its equivariant counterpart) does not work, since the action of $\mathcal{G}(P)$ is not generally free on $\mathcal{S}_{n}$ (also for $n=0$), as we assumed in writing down (6.28) for the $\mathcal{S}_{0}\equiv\mu^{-1}(0)$ component. For the higher extrema, this is readily seen by the fact that the equation $\nabla(\star F)=0\qquad\text{with}\ F\neq 0$ (6.70) identifies a vacuum $f:=(\star F)$ as a preferred element of $\mathfrak{g}$ (being it covariantly constant over $\Sigma$), and thus the gauge group is spontaneously broken to a subgroup $G_{f}\subseteq G$. The action of the whole gauge group thus cannot be free on this subspace, and the quotient $\mathcal{S}_{n}/\mathcal{G}(P)$ is singular. Via a suitable choice of localization 1-form one is still able to extract information by this integral over the normal directions, and in particular to compute the $\epsilon$-dependence of the higher extrema contributions. We limit ourselves now to the comparison of the exact result (6.61) applied to the case $G=SU(2)$, with the expectation (6.69) obtained by cohomological arguments. For this gauge group, the character expansion of the partition function results $Z(\epsilon)=\frac{1}{(2\pi^{2})^{g-1}}\sum_{n=1}^{\infty}\frac{\exp(-\epsilon\pi^{2}n^{2})}{n^{2g-2}}.$ (6.71) Simply taking $\epsilon=0$, we see that this is finite and proportional to a Riemann zeta-function, but to explore better the $\epsilon$-dependence it is convenient to consider $\frac{\partial^{g-1}Z}{\partial\epsilon^{g-1}}=\left(-\frac{1}{2}\right)^{g-1}\sum_{n=1}^{\infty}\exp\left(-\epsilon\pi^{2}n^{2}\right)=\frac{(-1)^{g-1}}{2^{g}}\left(-1+\sum_{n\in\mathbb{Z}}\exp\left(-\epsilon\pi^{2}n^{2}\right)\right).$ (6.72) This is not quite in the expected form, since the exponentials in the sum go to zero for $\epsilon\to 0$ but not as $\exp(-1/\epsilon)$. We can bring this expression closer to the desired result using the Poisson summation formula $\sum_{n\in\mathbb{Z}}f(n)=\sum_{k\in\mathbb{Z}}\hat{f}(k)\qquad\text{with}\ \hat{f}(k):=\int_{-\infty}^{+\infty}f(x)e^{-2\pi ikx},$ (6.73) where $f$ is a function and $\hat{f}$ its Fourier transform, and rewriting the sum of exponentials in (6.72) as $\frac{\partial^{g-1}Z}{\partial\epsilon^{g-1}}=\frac{(-1)^{g-1}}{2^{g}}\left(-1+\sqrt{\frac{1}{\pi\epsilon}}\sum_{k\in\mathbb{Z}}\exp\left(-\frac{k^{2}}{\epsilon}\right)\right).$ (6.74) This is exactly the result that could be obtained via integration over normal coordinates in the localization framework (see [102], eq. (4.102)), but fundamentally differs from our expectation, since for $\epsilon\to 0$ the contribution from the flat connections (with $k=0$) is singular for the presence of the square root. This means that the partition function is not really a polynomial in $\epsilon$ for small couplings, but an expression of the form $Z(\epsilon)=\sum_{m=0}^{g-2}a_{m}\epsilon^{m}+a_{g-3/2}\epsilon^{g-3/2}+\text{exponentially small terms}.$ (6.75) The singularity in $Z(\epsilon\to 0)$ arises because, for gauge group $SU(2)$, the subspace $\mu^{-1}(0)$ is singular and the MWM theorem does not apply. A simpler situation would occur considering the gauge group $SO(3)$ (which is not simply connected) and a non-trivial principal bundle over $\Sigma$. In this case, the character expansion of the partition function requires some modifications with respect to (6.61), the result being $Z(\epsilon)=\frac{1}{(8\pi^{2})^{g-1}}\sum_{n=1}^{\infty}\frac{(-1)^{n+1}\exp(-\pi^{2}\epsilon n^{2})}{n^{2g-2}}.$ (6.76) Following the same idea as above, we look at the $(g-1)^{th}$ derivative $\frac{\partial^{g-1}Z}{\partial\epsilon^{g-1}}=\frac{(-1)^{g}}{8^{g-1}}\sum_{n=1}^{\infty}(-1)^{n}\exp(-\pi^{2}\epsilon n^{2})=\frac{(-1)^{g}}{8^{g-1}}\frac{1}{2}\left(-1+\sum_{n\in\mathbb{Z}}(-1)^{n}\exp(-\pi^{2}\epsilon n^{2})\right),$ (6.77) and we rewrite the sum using the Poisson summation formula, getting $\frac{\partial^{g-1}Z}{\partial\epsilon^{g-1}}=\frac{(-1)^{g}}{2\cdot 8^{g-1}}\left(-1+\sqrt{\frac{1}{\pi\epsilon}}\sum_{k\in\mathbb{Z}}\exp\left(-\frac{\left(k+1/2\right)^{2}}{\epsilon}\right)\right).$ (6.78) This time we see that the contribution for $k=0$ from the moduli space of flat connections is finite for $\epsilon\to 0$, and the whole $\partial^{g-1}Z/\partial\epsilon^{g-1}$ is constant up to exponentially small terms. This means that the partition function at weak coupling $Z(\epsilon\to 0)$ is a regular polynomial of degree $g-1$ in $\epsilon$, up to exponentially decaying terms, $Z(\epsilon)=\sum_{m=0}^{g-2}a_{m}\epsilon^{m}+O(\epsilon^{g-1}),$ (6.79) and it reflects the fact that, for a non-trivial $SO(3)$-bundle, $\mu^{-1}(0)$ is smooth and acted on freely by $G$. These two quick examples capture the way this localization framework can give a very geometric interpretation to the $\epsilon$-expansion of the partition function, and its dependence on the classical geometry of the moduli space. Finally, we point out that an analogous treatment was done more recently in [102] to analyze in this cohomological framework Chern-Simons theory on 3-dimensional _Seifert manifolds_. A Seifert manifold is a smooth object that can be described as an $\mathbb{S}^{1}$-bundle over a 2-dimensional orbifold, and this feature makes it possible to dimensionally reduce the Chern-Simons theory along the direction of the circle $\mathbb{S}^{1}$ to a 2-dimensional YM theory over a singular base space. It turns out that the localization locus of the resulting theory receives contributions only from the flat connections over the total space. This is in accordance with the fact that Chern-Simons theories are by themselves TFT (of Schwarz-type). 2-dimensional YM theories have been studied extensively in the past years, and many interesting results were obtained thanks to their non-perturbative solvability. For example, exact results for Wilson loops expectation values and their relation with higher dimensional supersymmetric theories were studied in [108, 109, 110]. A relation with certain topological string theories and supersymmetric black hole entropy computations were analyzed in [111]. A duality between higher- dimensional supersymmetric gauge theories and deformations of 2-dimensional YM theory was revisited in [112, 113]. Localization techniques play an important role in all those cases. ## Chapter 7 Conclusion In this thesis we reviewed and summarized the main features of the formalism of equivariant cohomology, the powerful localization theorems first introduced by Atiyah-Bott and Berline-Vergne, and the principles that allow to formally apply these integration formulas to QFT. From the physical point of view, the equivariant (or supersymmetric) localization principle gives a systematic approach to understand when the “semiclassical” approximation of the path integral, describing the partition function or an expectation value in QFT, can give an exact result for the full quantum dynamics. We discussed the applicability of these techniques in the context of _supersymmetric_ theories. These are characterized by a space of fields that is endowed with a graded structure and the presence of some symmetry operator whose “square” gives a standard “bosonic” symmetry of the action functional. This supersymmetry operator is interpreted as a differential acting on the subspace of symmetric configurations in field space, and its cohomology describes the field theoretical analog of the $G$-equivariant cohomology of a $G$-manifold. After having introduced the general features of the mathematical theory of equivariant cohomology and equivariant localization, we reviewed the concepts in supergeometry that allow for the construction of supersymmetric QFT, and that constitute the correct framework to translate the mathematical theory in the common physical language. Since many recent applications of the localization principle aimed at the computations of path integrals in supersymmetric QFT on curved spaces, we included a discussion of the main tools needed to define supersymmetry in such instances. Then we collected some examples from the literature of application of the supersymmetric localization principle to path integrals in QFT of diverse dimensions. The common feature of these examples is that, via a suitable “cohomological” deformation of the action functional, it is possible to reduce the infinite-dimensional path integral to a finite-dimensional one that represents its semiclassical limit, as stressed above. We described cases in which this reduction relates the partition function to topological invariants of the geometric structure underlying the theory, namely the cases of supersymmetric QM (a 1-dimensional QFT) and the weak coupling limit of 2-dimensional Yang-Mills theory (its “topological” limit). We also reviewed the more recent applications to the computations of the expectation values of supersymmetric Wilson loops in 3- and 4-dimensional gauge theories, namely Supersymmetric Chern-Simons theory and Supersymmetric Yang-Mills theory defined on the 3- and the 4-sphere. In these cases, the path integral results to be equivalently described by some 0-dimensional QFT with a Lie algebra as target space, called “matrix model”. In the last few decades, the literature concerning the applications of supersymmetric localization has grown exponentially, and many other advanced examples of its use in the physics context have been found. From the point of view of supersymmetric QFT, a consistent slice of the state-of-the-art on the subject can be found in [12], including computations analogue to the one we showed for Wilson loop expectation values or topological invariants over more complicated geometries. From the point of view of Quantum (Super)Gravity, these techniques have found applications in the computations of the Black Hole quantum entropy [114, 115]. In many circumstances, localization allows for the analysis of properties of QFT at strong coupling, an otherwise prohibited region of study with conventional perturbative techniques. This feature can be used also to test a class of conjectural dualities between some types of gauge theories and string theories, the so-called AdS/CFT correspondences [96]. Concerning the subject of Wilson loops in 3-dimensional Chern-Simons theories and their relations to matrix models and holography, for which localization has played an important role, a recent reference that concisely reviews the state-of-the-art is [84]. ## Appendix A Some differential geometry ### A.1 Principal bundles, basic forms and connections Here we recall some notions about principal bundles that can be useful to follow the discussion, especially of the first chapters of this thesis. Principal bundles are the geometric construction behind the concepts of _covariant derivatives_ and _connections_ in gauge theory or General Relativity, for example. If $G$ is a Lie group, a principal $G$-bundle is a smooth bundle $P\xrightarrow{\pi}M$ such that 1. (i) $P$ is a (right) $G$-manifold; 2. (ii) the $G$-action on $P$ is _free_ ; 3. (iii) as a bundle, $P\to M$ is isomorphic to $P\to P/G$, where the projection map is canonically defined as $p\mapsto[p]$. Notice that since the $G$-action is free, a principal $G$-bundle is a fiber bundle with typical fiber $G$, and by the third property it is at least _locally trivial_ , i.e. over every open set $U\subseteq M$ it looks like $G\times U$. Morphisms of principal bundles are naturally defined as maps between bundles that preserve the $G$-structure, so $G$-equivariant maps. A principal bundle is _trivial_ if it is isomorphic through a principal bundle isomorphism to the trivial product bundle $G\times M$. A useful fact is that the triviality of a principal bundle is completely captured by the existence of a _global_ section $\sigma:M\to P$ such that $\pi\circ\sigma=id_{M}$. Since every principal bundle is locally trivial, than local sections can always be chosen and they constitute a so-called _local trivialization of the bundle_. The main example of principal bundle that occurs in the geometric construction of spacetime is the _frame bundle_ $LM$ over some $n$-dimensional smooth manifold $M$. At every point $p\in M$, the elements of the fiber $L_{p}M$ are the _frames_ at $p$, i.e. all the possible bases $e=(e_{1},\cdots,e_{n})$ for the tangent space $T_{p}M$. $LM$ has a natural $GL(n,\mathbb{R})$ right action that corresponds to the rotation of the basis, $e\cdot g:=(e_{k}g^{k}_{1},\cdots,e_{k}g^{k}_{n})$ for $g\in GL(n,\mathbb{R})$. In gauge theories, the _structure group_ (or sometimes _gauge group_) $G$ of the theory is the Lie group acting on the right on a principal $G$-bundle. Since the fibers of the principal bundle are essentially the Lie group $G$, tangent vectors on $P$ can come from its Lie algebra $\mathfrak{g}$. This leads to the following definition. ###### Definition A.1.1. The _vertical sub-bundle_ $VP$ of the tangent bundle $TP$ is the disjoint union $VP:=\bigsqcup_{p\in P}V_{p}P,\qquad\text{with}\ V_{p}P:=\mathrm{Ker}(\pi_{*p})=\\{X\in T_{p}P|\pi_{*}(X)=0\\}\subset T_{p}P.$ Analogously but for differential forms, the _basic forms_ inside $\Omega(P)$ are those forms $\omega\in\mathrm{Im}(\pi^{*})$, so that it exists an $\alpha\in\Omega(M)$ such that $\omega=\pi^{*}\alpha$. The space of basic forms is denoted $\Omega(P)_{bas}$. The vertical vectors in every $V_{p}P$ are in one to one correspondence with the Lie algebra elements in $\mathfrak{g}$, through the Lie algebra homomorphism $X\in\mathfrak{g}\mapsto\underline{X}:=\left.\frac{d}{dt}\right|_{t=0}\left(\cdot e^{tX}\right)^{*}$ (A.1) that maps Lie algebra elements to the corresponding _fundamental vector fields_.111The choice of the sign at the exponential differs from the one in (2.19) because here we are considering a right action. Fundamental vector fields satisfy the following properties: 1. (i) $[\underline{X},\underline{Y}]=\underline{[X,Y]}$; 2. (ii) the integral curve of $\underline{X}$ through $p\in M$ is $\displaystyle\gamma_{p}:\mathbb{R}$ $\displaystyle\to M$ (A.2) $\displaystyle t$ $\displaystyle\mapsto\gamma_{p}(t)=p\cdot e^{tX};$ 3. (iii) denoting with $r_{g}$ the right action of $g\in G$, $\left(r_{g}\right)_{*}(\underline{X}_{p})=\left(\underline{Ad_{g^{-1}*}(X)}\right)_{r_{g}(p)}.$ (A.3) As for the vertical vector fields being encoded in the Lie algebra $\mathfrak{g}$, also the basic forms can be characterized in terms on the (infinitesimal) action of $\mathfrak{g}$ on $\Omega(P)$. This can be seen introducing the following definitions. ###### Definition A.1.2. A differential form $\omega\in\Omega(P)$ is said to be _$G$ -invariant_ if it is preserved by the $G$ action: $\omega=(r_{g})^{*}\omega\qquad\forall g\in G.$ The space of $G$-invariant forms is commonly denoted $\Omega(P)^{G}$. A differential form is called _horizontal_ if it is annihilated by vertical vector fields, $\iota_{X}\omega=0\qquad\forall X\in\Gamma(VP).$ The properties of being invariant and horizontal can be also stated infinitesimally with respect to the action of the Lie algebra $\mathfrak{g}$. If we define $\mathcal{L}_{X}:=\mathcal{L}_{\underline{X}},\qquad\iota_{X}:=\iota_{\underline{X}},\qquad\forall X\in\mathfrak{g},$ (A.4) then an invariant form is characterized by $\mathcal{L}_{X}\omega=0$ for every $X\in\mathfrak{g}$, and a horizontal form by $\iota_{X}\omega=0$ for every $X\in\mathfrak{g}$. This makes the concepts of invariant and horizontal elements independent from the principal bundle structure, so that they can be defined by this characterization for every $\mathfrak{g}$-dg algebra, as in Section 2.3. Also basic forms can be defined for every $\mathfrak{g}$-dg algebra, combining the definitions of invariant and horizontal forms, thanks to the following theorem: ###### Theorem A.1.1 (Characterization of basic forms). Schematically, $\text{invariant}+\text{horizontal}\Leftrightarrow\text{basic}.$ ###### Proof. For notational convenience only, let us consider 1-forms. * $(\Leftarrow)$ If $\omega=\pi^{*}\alpha$ is basic, then $(r_{g})^{*}\omega=(r_{g})^{*}\pi^{*}\alpha=(\pi\circ r_{g})^{*}\alpha=\omega,$ since the principal bundle is locally trivial. So $\omega$ is also invariant. For a vertical vector $X\in VP$, $\iota_{X}\omega=(\pi^{*}\alpha)(X)=\alpha(\pi_{*}(X))=0.$ So $\omega$ is also horizontal. * $(\Rightarrow)$ Let $\omega\in\Omega^{1}(P)$ be horizontal and invariant. Since $\pi$ is surjective, for every vector $X\in T_{p}P$ there exists $Y=\pi_{*}(X)\in T_{\pi(p)}M$. We can define $\alpha\in\Omega^{1}(M)$ such that, at every $x\in M$ $\alpha_{x}(Y):=\omega_{p}(X)\qquad\text{for}\ p\in\pi^{-1}(x),$ and thanks to the horizontality and invariance of $\omega$ we can check that this form is well defined, i.e. independent from the choice of point $p$ in the fiber $\pi^{-1}(x)$ and from the choice of vector $X$ such that $\pi_{*}(X)=Y$. In fact, if $X^{\prime}\in T_{p}P$ is another vector such that $\pi_{*}(X^{\prime})=Y$, then $\pi_{*}(X-X^{\prime})=0$ so $(X-X^{\prime})\in V_{p}P$. By horizontality, $\omega(X-X^{\prime})=0\Rightarrow\omega(X)=\omega(X^{\prime})$, so $\alpha$ is independent from the choice of vector. Moreover, if $p^{\prime}\in\pi^{-1}(x)$ is another point in the fiber, there exists a $g\in G$ such that $r_{g}(p)=p^{\prime}$, so by $G$-invariance $\omega_{p^{\prime}}=\omega_{p}$ and thus $\alpha$ is independent from the choice of point in the fiber. ∎ ###### Proposition A.1.1. The differential $d$ closes on the subspace $\Omega(P)_{bas}$ of basic forms, defining a proper subcomplex. This extends to any $\mathfrak{g}$-dg algebra. ###### Proof. Consider $\omega\in\Omega(P)_{bas}$, and its differential $d\omega$. We characterize basic forms by being horizontal and $G$-invariant. By Cartan’s magic formula the Lie derivative commutes with the differential, so for every $X\in\mathfrak{g}$, $\mathcal{L}_{X}d\omega=d(\mathcal{L}_{X}\omega)=0$. Thus $d\omega$ is still $G$-invariant. Also, $\iota_{X}d\omega=\mathcal{L}_{X}\omega-d\iota_{X}\omega=0$. Thus $d\omega$ is still horizontal, and so basic. ∎ We recall now the definition of connection and curvature on principal bundles, from which one inherits covariant derivatives on associated vector bundles. ###### Definition A.1.3. An _(Ehresmann) connection_ on a principal $G$-bundle $P\to M$ is an _horizontal distribution_ $HP$, i.e. a smooth choice at every point $p\in P$ of vector subspaces $H_{p}P\subset T_{p}P$ such that 1. (i) $T_{p}P=V_{p}P\oplus H_{p}P$; 2. (ii) $(r_{g})_{*}(H_{p}P)=H_{p\cdot g}P$ ($G$-equivariance of the horizontal projection). Given a horizontal distribution $HP$, every vector $X\in T_{p}P$ decomposes into an _horizontal_ and a _vertical_ part, $X=hor(X)+ver(X).$ A _connection 1-form_ on $P\to M$ is Lie algebra-valued 1-form $A\in\Omega^{1}(P)\otimes\mathfrak{g})$ such that 1. (i) for any $X\in\mathfrak{g}$, $\iota_{X}A=A(\underline{X})=X$ (_vertical_ 1-form); 2. (ii) for any $g\in G$, $(r_{g})^{*}A=(Ad_{g^{-1}*}\circ A)$ ($G$-equivariance). The choice of a horizontal distribution is equivalent to the choice of a connection 1-form on $P$, since at every $p\in P$ one can use $A$ as a projection onto the vertical subspace $V_{p}P\cong\mathfrak{g}$, and $\pi_{*}$ as a projection onto the horizontal subspace, identifying $V_{p}P:=\mathrm{Ker}(\pi_{*p})$ and $H_{p}P:=\mathrm{Ker}(A_{p})$. This choice is smooth and $G$-equivariant since $A$ is, by definition. Notice that the splitting $HP\oplus VP$ induces a splitting $\Omega^{1}(P)=\Omega^{1}_{hor}(P)\oplus\Omega^{1}_{ver}(P)$, and that we can identify the “space of connection 1-forms” as $\mathcal{A}(P):=\\{A\in(\Omega^{1}_{ver}(P)\otimes\mathfrak{g})|A\ \text{is }G\text{-equivariant}\\}.$ (A.5) It is easy to see that for every $A,A^{\prime}\in\mathcal{A}(P)$, their difference is not a connection, and in fact it is an _horizontal_ $\mathfrak{g}$-valued 1-form, $(A-A^{\prime})\in\mathfrak{a}:=\\{a\in(\Omega^{1}_{hor}(P)\otimes\mathfrak{g})|a\ \text{is }G\text{-equivariant}\\}.$ (A.6) This means that every connection $A$ can be written as another connection $A^{\prime}$ plus a horizontal form, or in other words that $(\mathcal{A}(P),\mathfrak{a})$ can be seen as a natural _affine space_ , modeled on the infinite-dimensional vector space $\mathfrak{a}$. As for any affine space, one can think of the space of connections as an infinite- dimensional smooth manifold, with tangent spaces $T_{A}\mathcal{A}(P)\cong\mathfrak{a}$ at every $A\in\mathcal{A}(P)$. ###### Definition A.1.4. The _covariant exterior derivative_ on $\Omega(P)$ is $D:=d\circ hor^{*}$. The _curvature_ of a connection 1-form $A$ is $F:=DA=dA(hor(\cdot),hor(\cdot))\in\Omega^{2}(P)\otimes\mathfrak{g}.$ The curvature $F$ satisfies the following properties: 1. (i) by definition, $F$ is horizontal: $\iota_{X}F=0$ for every $X\in\mathfrak{g}$; 2. (ii) by $G$-equivariance of $A$, $F$ is $G$-equivariant too; 3. (iii) it obeys the structural equation $F=dA+\frac{1}{2}[A\stackrel{{\scriptstyle\wedge}}{{,}}A]$ (A.7) where $[A\stackrel{{\scriptstyle\wedge}}{{,}}A]=f^{a}_{bc}A^{b}\wedge A^{c}\otimes T_{a}$ with respect to a basis $\\{T_{a}\\}$ of $\mathfrak{g}$ and the structure constants $f^{a}_{bc}$; 4. (iv) it obeys the second Bianchi identity, $DF=0\qquad\text{or}\qquad dF=[F\stackrel{{\scriptstyle\wedge}}{{,}}A].$ (A.8) One can consider the very trivial construction of a principal $G$-bundle as $G\to pt$, where $P=G\times pt\cong G$. Here the right $G$-action is simply the diagonal action (trivial on $pt$, induced by the natural action on $G$). On this bundle there is a canonical choice of connection 1-form, the _Maureer- Cartan (MC) form_ $\Theta\in\Omega^{1}(G)\otimes\mathfrak{g}$. For every vector $X\in T_{g}G$ at some $g\in G$, there is a Lie algebra element $A\in T_{e}G\cong\mathfrak{g}$ such that $X=l_{g*}(A)$, and the MC form is defined by $\Theta_{g}(X):=l_{g^{-1}*}(X)=A.$ (A.9) One can check that this form is indeed $G$-equivariant, and it is obviously vertical, giving a connection 1-form. Moreover it satisfies the _Maurer-Cartan equation_ $d\Theta+\frac{1}{2}[\Theta\stackrel{{\scriptstyle\wedge}}{{,}}\Theta]=0,$ (A.10) so that by (A.7) we see that its curvature is zero. On the trivial principal $G$-bundle $G\times M\to M$ one can always define a connection 1-form by pulling back the MC connection along the projection $\pi_{1}:G\times M\to G$. In the general case, the principal bundle $P$ is locally trivial, so in any local patch $G\times U_{\alpha}\to U_{\alpha}$ one can pull back the MC connection and use a suitable partition of unity to glue together the local pieces to a global connection 1-form on $P$. This shows that any principal bundle allows for a connection. The curvature $F$ of the chosen connection $A$ measures, in a sense, the deviation of $A$ from being the Maurer-Cartan connection. As said before, a connection on a principal $G$-bundle allows for the definition of a covariant derivative on _associated vector bundles_. An associated vector bundle to the principal $G$-bundle $P\xrightarrow{\pi}M$ is a vector bundle constructed over $M$ with some typical fiber $V$ (a vector space) that has a (left) $G$-action compatible with the one on $P$. Precisely, the associated bundle is $P_{V}\xrightarrow{\pi_{V}}M$, where $\displaystyle P_{V}:=\faktor{(P\times V)}{\sim_{G}}\qquad\text{with}\ (p,v)\sim_{G}(p\cdot g,g^{-1}\cdot v)\ \forall(p,v)\in P\times V,g\in G,$ (A.11) $\displaystyle\pi_{V}([p,v]):=\pi(p),$ and it has indeed typical fiber $V$. In the case of the frame bundle $P=LM$, one can construct the tangent bundle $TM$, the cotangent bundle $T^{*}M$ and all the tensor bundles as associated to $LM$. In fact, for the tangent bundle for example, the typical fiber is $V:=\mathbb{R}^{n}$ and the $GL(n,\mathbb{R})$-action is $(g\cdot v)^{k}=g^{k}_{j}v^{j}$. This encodes the change of basis rule if we see vectors as elements $[e,v]\in LM_{\mathbb{R}^{n}}$, $[e,v]\equiv e_{k}v^{k}\sim_{GL}[e\cdot g,g^{-1}\cdot v]\equiv e_{j}g^{j}_{k}(g^{-1})^{k}_{j}v^{j}=e_{k}v^{k}.$ (A.12) On the associated vector bundle, a _field_ (in physics terms) is a (local, at least) section $\phi:M\to P_{V}$, that can be always seen locally as a $V$-valued function on every $U\subseteq M$, $\tilde{\phi}:U\to V$, so that $\phi(x)=[p,\tilde{\phi}(x)]$ for some chosen $p\in\pi^{-1}(x),x\in U$. Another example of this concept that came up in Chapter 2 is the _homotopy quotient_ $M_{G}:=(M\times EG)/G$ of a $G$-manifold $M$. This is precisely the associated bundle with fiber $V=M$ to the principal $G$-bundle $EG\to BG$ (however, this is not an associated _vector_ bundle, since $M$ is not a vector space in general). As we said at the beginning of this appendix, every principal bundle is locally trivial, so that it exists a set of _local trivializations_ $\\{U_{\alpha},\varphi_{\alpha}:\pi^{-1}(U_{\alpha})\to U_{\alpha}\times G\\}$, where $\\{U_{\alpha}\\}$ covers $M$, and $\varphi_{\alpha}$ is $G$-equivariant. This means that $\varphi_{\alpha}(p)=(\pi(p),g_{\alpha}(p))$ for some $G$-equivariant map $g_{\alpha}:\pi^{-1}(U_{\alpha})\to G$,222In this case $G$-equivariance means $g_{\alpha}(p\cdot h)=g_{\alpha}(p)\cdot h$. that makes every fiber diffeomorphic to $G$. To this local trivialization, one can canonically associate a family of local sections $\\{\sigma_{\alpha}:U_{\alpha}\to\pi^{-1}(U_{\alpha})\\}$, determined by the maps $\varphi_{\alpha}$ so that for every $m\in U_{\alpha}$, $\varphi_{\alpha}(\sigma_{\alpha}(m))=(m,e)$, where $e\in G$ is the identity element. In other words, $g_{\alpha}\circ\sigma_{\alpha}:U_{\alpha}\to G$ is the constant function over the local patch $U_{\alpha}\subseteq M$ that maps every point to the identity. Conversely, a local section $\sigma_{\alpha}$ allows us to identify the fiber over $m$ with $G$. Indeed, given any $p\in\pi^{-1}(m)$, there is a unique group element $g_{\alpha}(p)\in G$ such that $p=\sigma_{\alpha}(m)\cdot g_{\alpha}(p)$. Using these canonical local data, the connection $A$ and the curvature $F$ can be pulled back on $M$ giving the local _gauge field_ $A^{(\alpha)}:=\sigma_{\alpha}^{*}(A)$ and _field strength_ $F^{(\alpha)}:=\sigma_{\alpha}^{*}(F)$. The _covariant derivative_ along the tangent vector $X\in TM$ of a local $V$-valued function $\tilde{\phi}:U_{\alpha}\to V$ is defined as $\nabla_{X}\tilde{\phi}:=d\tilde{\phi}(X)+A^{(\alpha)}(X)\cdot\tilde{\phi},$ (A.13) where the second term denotes the action of the Lie algebra on $V$, that for matrix groups coincides with the action of $G$. We denote schematically the covariant derivative as $\nabla=d+A$ on a generic associated vector bundle. When $V=\mathfrak{g}$ we have the so-called _adjoint bundle_ , often denoted $\text{ad}(P)$, that is in one-to-one correspondence with the space $\mathfrak{a}$ above, of horizontal and $G$-equivariant Lie algebra-valued forms on $P$. On this special associated bundle, the covariant derivative acts with the infinitesimal adjoint action of $\mathfrak{g}$, $\nabla=d+[A,\cdot].$ (A.14) By the horizontal property of the curvature, we see that $F$ can be regarded as a 2-form on $M$ with values in $\text{ad}(P)$.333Being horizontal means that pulling it back on the base space, we do not lose information on the 2-form. In fact, the local representation of the curvature $F^{(\alpha)}=\sigma_{\alpha}^{*}(F)$ still transforms covariantly also as a $\mathfrak{g}$-valued 2-form over $M$. Strictly speaking, the covariant derivative on the adjoint bundle acts on this local representative. Notice that the gauge field $A^{(\alpha)}$ instead looks only locally as an element of the adjoint bundle, but globally it does not respect the “right” transformation property, and indeed it comes from a global 1-form on $P$ that is not horizontal, but vertical. Then the Bianchi identity can be rewritten in terms of the covariant derivative, $\nabla F=dF+[A,F]=[F,A]+[A,F]=0.$ (A.15) As the last piece of information, we recall the meaning of _gauge transformations_ from the perspective of the principal bundle. Locally, we can think of them as local actions of the gauge group $G$, so that a gauge transformation is a map that associates to every point $x\in M$ an element $g(x)\in G$, acting on the local field strength in the adjoint representation. At the level of the principal bundle, this can be viewed more formally defining the group $\mathcal{G}(P)\subset\mathrm{Diff}(P)$ of principal bundle maps of the type444As a principal bundle map it is by definition $G$-equivariant, $\Psi(p\cdot g)=\Psi(p)\cdot g$, and it commutes with the projection, $\pi(\Psi(p))=\pi(p)$, for every $p\in P$. ${P}$${P}$${M.}$$\scriptstyle{\Psi}$$\scriptstyle{\pi}$$\scriptstyle{\pi}$ (A.16) We notice right-away that, from the local point of view, this can indeed be identified with the space of sections $\Omega^{0}\left(M;\mathrm{Ad}(P)\right)$ of the bundle $\mathrm{Ad}(P)$, associated to $P$ with typical fiber $G$ and $G$-action defined by conjugation (the adjoint representation of $G$ on itself).555Notice that this looks like $P\to\Sigma$ as a fiber bundle, since both are locally trivial with fiber $G$. It is only the $G$ action that distinguishes them. On $P$ we have a right action, on $\mathrm{Ad}(P)$ we have the left action on the fibers $g\cdot f:=gfg^{-1}$. ###### Proof of $\mathcal{G}(P)\cong\Omega^{0}(M;\mathrm{Ad}(P))$. We can see that associated to every element $\Psi\in\mathcal{G}(P)$ there is a unique class of local sections $\\{\psi_{\alpha}:U_{\alpha}\to G\\}$ that transforms in the adjoint representation, and vice versa. * $(\Rightarrow)$ In every local patch $U_{\alpha}$, let us define the map $\tilde{\psi}_{\alpha}:\pi^{-1}(U_{\alpha})\to G$ such that $\tilde{\psi}_{\alpha}(p):=g_{\alpha}(\Psi(p))\ g_{\alpha}(p)^{-1},$ where $g_{\alpha}$ is the trivialization map inside $U_{\alpha}$. By equivariance of $\Psi$ and $g_{\alpha}$, $\tilde{\psi}_{\alpha}$ is $G$-invariant, so it depends only on the base point $\pi(p)$. Thus we can define $\psi_{\alpha}:U_{\alpha}\to G$ such that $\psi_{\alpha}(x):=\tilde{\psi}_{\alpha}(p)\qquad\text{for some}\ p\in\pi^{-1}(x).$ In changing local patch, this transforms in the adjoint representation. In fact, if $x\in U_{\alpha}\cap U_{\beta}$ $\displaystyle\psi_{\beta}(x)$ $\displaystyle=g_{\beta}(\Psi(p))\ g_{\beta}(p)^{-1}$ $\displaystyle=\left[g_{\beta}(\Psi(p))g_{\alpha}(\Psi(p))^{-1}\right]g_{\alpha}(\Psi(p))g_{\alpha}(p)\left[g_{\alpha}(p)^{-1}g_{\beta}(p)\right]$ $\displaystyle=g_{\alpha\beta}(x)\psi_{\alpha}(x)g_{\alpha\beta}(x)^{-1},$ where in the last passage we recognized that $g_{\beta}(p)g_{\alpha}(p)^{-1}$ is $G$-invariant and thus can be written as a map $g_{\alpha\beta}:U_{\alpha}\cap U_{\beta}\to G$ that depends only on the base point $\pi(p)$, and we used that $\pi\circ\Psi=\pi$. * $(\Leftarrow)$ Starting from a class of local sections $\\{\psi_{\alpha}\\}$, we define the $G$-invariant maps $\tilde{\psi}_{\alpha}:=\psi_{\alpha}\circ\pi$. Then we can obtain $\Psi$ by “inverting” the above definition in every patch and gluing them together, $\Psi(p):=\sigma_{\alpha}(p)\cdot\left(\tilde{\psi}_{\alpha}(p)g_{\alpha}(p)\right).$ ∎ The group of gauge transformations $\mathcal{G}(P)$ acts naturally on the space $\mathcal{A}(P)$ via pull-back, $\Psi\cdot A:=\Psi^{*}(A)\qquad\forall A\in\mathcal{A}(P),\Psi\in\mathcal{G}(P).$ (A.17) If we consider the local gauge field $A^{(\alpha)}$, one can prove that the trivialization of the gauge-transformed connection follows the usual rule $(\Psi\cdot A)^{(\alpha)}=\psi_{\alpha}^{-1}A^{(\alpha)}\psi_{\alpha}+\psi^{-1}_{\alpha}(d\psi_{\alpha}).$ (A.18) From this local expression, it is easy to find the representation of $Lie(\mathcal{G}(P))$ on $T\mathcal{A}(P)$. In fact, writing $\Psi$ as $\exp(X)$ for some $X\in Lie(\mathcal{G}(P))\cong\Omega^{0}(M;\mathrm{ad}(P))$, we can recognize the associated fundamental vector field as $\underline{X}_{A}=\left.\frac{d}{dt}\right|_{t=0}e^{-tX}\cdot A=dX+[A,X]=\nabla^{A}X.$ (A.19) This make us see the usual “infintesimal variation” $\delta_{X}A$ as a tangent vector $\delta_{X}A\equiv\underline{X}_{A}\in T_{A}\mathcal{A}(P)$ at the point $A\in\mathcal{A}(P)$. ### A.2 Spinors in curved spacetime In QFT, _fermionic_ particles are described geometrically by _spinors_ , i.e. fields that transform under the Lorentz algebra in representations whose angular momentum is _half-integer_. At the level of Lie groups, they transform thus in representations of the double-cover of the rotation group of spacetime, $SO(1,d-1)$ (or $SO(d)$ in the Euclidean case), $SO(1,d-1)\cong\faktor{Spin(1,d-1)}{\mathbb{Z}_{2}}.$ (A.20) For simplicity, let us denote the dimension by $d$ for the rest of the section, since the discussion is valid both for the Euclidean and the Lorentzian signature. In Minkowski spacetime $(\mathbb{R}^{d},\eta)$, there exists a preferred class of _global_ coordinate systems, the global “inertial frames”, where the metric is diagonal $\eta_{\mu\nu}=diag(-1,+1,\cdots,+1)$ (A.21) and that are preserved by the Lorentz transformations. Working only with such special type of coordinate systems, one can introduce and work with spinors as living in double-valued representations of the Lorentz algebra, and transforming as $\displaystyle\text{vector fields}:\qquad V^{\mu}\mapsto\Lambda^{\mu}_{\nu}V^{\nu},$ (A.22) $\displaystyle\text{spinor fields}:\qquad\Psi_{\alpha}\mapsto S(\Lambda)^{\beta}_{\alpha}\Psi_{\beta},$ where, if $\Lambda=\exp(i\omega_{\mu\nu}M^{\mu\nu})$, $S(\Lambda)=\exp(i\omega_{\mu\nu}\Sigma^{\mu\nu})$. When we move on to the description of a generically curved spacetime $M$, there is a priori no such choice of “preferred” coordinate systems, and a general coordinate transformation (GCT) is generated by a diffeomorphism $M\to M$, reflecting on the tangent spaces as $GL(d,\mathbb{R})$ basis changes. $SO(d)$ injects as a subgroup of the General Linear group, but $Spin(d)$ does not, since it is a double cover, so it is not clear a priori how GCTs act on spinor fields. Tensor fields are naturally present in the fully covariant formalism as fields over the manifold $M$, but to define spinors one has to introduce further structure. The solution to this puzzle is really to (try to) mimic the same idea applied the the Minkowski case, and employ the presence of a (pseudo-)Riemannian metric $g$ on $M$. On the metric manifold $(M,g)$ all the tangent bundles arise as associated bundles to the _frame bundle_ $LM$, that is a principal $GL(d,\mathbb{R})$-bundle over $M$. Using the presence of a metric on $M$, one can restrict the frame bundle to a principal $SO(d)$-bundle, by considering only those frames $e=(e_{1},\cdots,e_{d})$ such that, at a given point $g(e_{i},e_{j})=\eta_{ij},$ (A.23) where $\eta$ is the “flat” Minkowski (or Euclidean) metric. This reduction defines the so called _orthonormal frame bundle_ $LM^{(SO)}\xrightarrow{\pi}M$. A section of this bundle is an orthonormal frame, or _tetrad_. It is customary to denote with Latin indices the expansion of every vector field with respect to an orthonormal frame, and with Greek indices the expansion with respect to a generic (for example chart-induced) frame:666Latin indices are sometimes called “flat”, and Greek ones “curved”. If one needs to raise and lower indices, flat indices are understood to be multiplied by the diagonalized metric $\eta_{ij}$, curved indices by $g_{\mu\nu}$. $V=V^{i}e_{i}=V^{\mu}\frac{\partial}{\partial x^{\mu}}\qquad\text{for}\ V\in\Gamma(TM).$ (A.24) The choice of an orthonormal frame is encoded in the choice of a _vielbein_ , or _solder form_ on $M$, that is a linear identification of the tangent bundle with the typical fiber $\mathbb{R}^{d}$: $\displaystyle E:TM$ $\displaystyle\to\mathbb{R}^{d}$ (A.25) $\displaystyle V$ $\displaystyle\mapsto E(V):=(\tilde{e}^{i}(V))_{i=1,\cdots,d},$ where $(\tilde{e}^{i})$ is the dual frame to a chosen orthonormal frame $(e_{i})$. Notice that the choice of a metric is in one to one correspondence with the choice of a vielbein, since $g(\cdot,\cdot)=\langle E(\cdot),E(\cdot)\rangle,$ (A.26) where $\langle\cdot,\cdot\rangle$ is the canonical inner product on $\mathbb{R}^{d}$ with the chosen signature. In a chart-induced basis, $g_{\mu\nu}=e^{i}_{\mu}e^{j}_{\nu}\eta_{ij}$, where we denoted the components of the vielbein $(E(\partial_{\mu}))^{i}\equiv e^{i}_{\mu}$. The “inverse vielbein” at any point is the matrix $e^{\mu}_{i}$ such that $e^{i}_{\mu}e^{\mu}_{j}=\delta^{i}_{j}$. Once this orthonormal reduction is made, one can define spinor bundles as associated bundles to a principal $Spin(d)$-bundle, that must be compatible with the orthonormal frame bundle. This is made precise by defining the presence of a spin-structure on $M$. ###### Definition A.2.1. A _spin-structure_ on $(M,g)$ is a principal $Spin(d)$-bundle $Spin(M)\xrightarrow{\pi_{S}}M$, together with a principal bundle map777Recall that a principal bundle map by definition commutes with the projections, $\pi(\Phi(p))=\pi_{S}(p)$. ${Spin(M)}$${LM^{(SO)}}$${M}$$\scriptstyle{\Phi}$$\scriptstyle{\pi_{S}}$$\scriptstyle{\pi}$ with respect to the double-cover map $\varphi:Spin(d)\to SO(d)$. This means that the equivariance condition is $\Phi(s\cdot g)=\Phi(s)\cdot\varphi(g)\qquad\forall s\in Spin(M),g\in Spin(d).$ A section of $Spin(M)\to M$ is called _spin-frame_. We notice that the equivariance condition in this definition is just the formal requirement that spinors and tensors transform all together with compatible rotations by the action of the respective groups. Although the above restriction of the frame bundle to the orthonormal frame bundle can always be done in presence of a metric on $M$, a spin-structure does not necessarily exist, and if it does it is not necessarily unique. There can be topological obstructions to this process that can be characterized in terms of the cohomology of $M$.888In particular, it turns out that a spin-structure exists if and only if the second Stiefel–Whitney class of $M$ vanishes [27]. By this construction, and from the canonical Levi-Civita covariant derivative $\nabla$ on $(M,g)$, we can induce a connection 1-form on the orthonormal frame bundle and on the spin-frame bundle, and thus have a compatible covariant derivative on associated spinor bundles. Let us recall that the Levi-Civita connection on $(M,g)$ is the unique metric-compatible and torsion free connection, i.e. $\begin{array}[]{lcl}\nabla_{X}g=0&\Leftrightarrow&X(g(Y,Z))=g(\nabla_{X}Y,Z)+g(Y,\nabla_{X}Z),\\\ T=0&\Leftrightarrow&\nabla_{X}Y-\nabla_{Y}X=[X,Y].\end{array}$ (A.27) This covariant derivative is associated to the gauge field $\Gamma\in\Omega^{1}(M)\otimes\mathfrak{gl}(n,\mathbb{R})$ such that $\Gamma^{\rho}_{\mu\nu}:=(\nabla_{\mu}(\partial_{\nu}))^{\rho}$. Simply restricting to orthonormal frames, one can induce a connection 1-form on $LM^{(SO)}$, $\omega\in\Omega^{1}(LM^{(SO)})\otimes\mathfrak{so}(d)$ such that in any trivialization induced by a local frame $(U\subset M,e:U\to LM^{(SO)})$ the gauge field has components $\omega(X)^{i}_{j}:=(\nabla_{X}(e_{i}))^{j}=X^{\mu}e^{j}_{\nu}(\nabla_{\mu}e_{i})^{\nu}=X^{\mu}e^{j}_{\nu}\left(\partial_{\mu}e^{\nu}_{i}+\Gamma^{\nu}_{\mu\sigma}e^{\sigma}_{i}\right)\quad\text{or}\quad\omega(X)_{ij}=g(\nabla_{X}e_{i},e_{j}),$ (A.28) and it can be written as $\omega^{(U)}:=e^{*}\omega=\frac{1}{2}\omega_{ij}M^{ij}$, where $M^{ij}$ are the generators of $\mathfrak{so}(d)$. Given a spin-structure as in the above definition, we can induce a _compatible spin-connection_ $\tilde{\omega}\in\Omega^{1}(Spin(M))\otimes\mathfrak{so}(d)$ by pulling back $\omega$, $\tilde{\omega}:=\Phi^{*}\omega$.999Notice that $Lie(Spin(d))\cong Lie(SO(d))\cong\mathfrak{so}(d)$. In a given patch $U\subset M$, if $s:U\to Spin(M)$ is a local spin-frame and $e:=\Phi\circ s$ is the associated tangent frame, the local gauge fields representing the spin-connection and the Levi- Civita connection coincide, $\tilde{\omega}^{(U)}:=s^{*}\tilde{\omega}=(\Phi\circ s)^{*}\omega=\omega^{(U)},$ (A.29) so in particular the local components of the compatible spin-connection are defined as $\tilde{\omega}(X)^{i}_{j}=(\nabla_{X}(e_{i}))^{j}.$ (A.30) The covariant derivative on an associated spinor bundle is defined as usual. Let $V$ be the typical fiber, acted upon by the representation $\rho:Spin(d)\to GL(V)$. Then for every local $V$-valued function $\psi:U\to V$, $\nabla_{X}\psi=d\psi(X)+\frac{1}{2}\omega(X)_{ij}\rho(M^{ij})\cdot\psi.$ (A.31) If in particular we take the fundamental representation of $Spin(d)$, i.e. $\psi$ is a _Dirac spinor_ , the generators are $\rho(M^{ij})=\Sigma^{ij}:=\frac{1}{4}[\gamma^{i},\gamma^{j}]$, where $\gamma^{i}$ are the Dirac matrices. Thus, $\nabla_{\mu}\psi=\partial_{\mu}\psi+\frac{1}{8}\omega_{\mu ij}[\gamma^{i},\gamma^{j}]\cdot\psi.$ (A.32) We quote the fact that, in general, one is not forced to consider a spin- connection that is compatible with the Levi-Civita connection.101010For example in SUGRA it is sometimes convenient to work with torsion-full spin connections. However, in this work we always implicitly define covariant derivatives on spinors via a compatible spin-connections. A discussion about spinors in curved spacetime can be found also in [65]. ## Appendix B Mathematical background on equivariant cohomology ### B.1 Equivariant vector bundles and equivariant characteristic classes We recall the definitions of characteristic classes on principal bundles [116] and then their equivariant version when the bundle supports a $G$-action for some Lie group $G$. Consider a principal $H$-bundle $P\xrightarrow{\pi}M$ with connection 1-form $A$, and curvature $F$. Both are forms on $P$ with values in the Lie algebra $\mathfrak{h}$. A _polynomial_ on $\mathfrak{h}$ is an element $f\in S(\mathfrak{h}^{*})$, and it is called _invariant polynomial_ if it is invariant with respect to the adjoint action of $H$ on $\mathfrak{h}$, $f(Ad_{*h}X)=f(X)\qquad\forall X\in\mathfrak{h},h\in H.$ (B.1) For example, if $H$ is a matrix group, the adjoint action is simply $Ad_{*h}X=hXh^{-1}$. If $f$ is an invariant polynomial of degree $k$, then $f(F)$ is an element of $\Omega^{2k}(P)$. Explicitly, with respect to a basis $(T_{a})_{a=1,\cdots,\dim\mathfrak{h}}$ of $\mathfrak{h}$ and the dual basis $(\alpha^{a})_{a=1,\cdots,\dim\mathfrak{h}}$ of $\mathfrak{h}^{*}$, if $F=F^{a}T_{a}$ and $f=f_{a_{1}\cdots a_{k}}\alpha^{a_{1}}\cdots\alpha^{a_{k}}$, then $f(F)=f_{a_{1}\cdots a_{k}}F^{a_{1}}\wedge\cdots\wedge F^{a_{k}}.$ (B.2) The above form has three remarkable properties: 1. (i) $f(F)$ is a basic form on $P$, i.e. it exists a $2k$-form $\Lambda\in\Omega^{2k}(M)$ such that $f(F)=\pi^{*}\Lambda$; 2. (ii) $d\Lambda=0$, or equivalently $df(F)=0$; 3. (iii) the cohomology class $[\Lambda]\in H^{2k}(M)$ is independent on the connection $F$. The cohomology class $[\Lambda]$ on $M$ is called _characteristic class_ of $P$ associated to the invariant polynomial $f$. Denoting $\text{Inv}(\mathfrak{h})\subseteq S(\mathfrak{h}^{*})$ the algebra of invariant polynomials on $\mathfrak{h}$, the map $\displaystyle w:\text{Inv}(\mathfrak{h})$ $\displaystyle\to H^{*}(M)$ (B.3) $\displaystyle f$ $\displaystyle\mapsto[\Lambda]$ is called _Chern-Weil homomorphism_. If one is considering a vector bundle $E\to M$ associated to the principal $H$-bundle $P\to M$, here the connection 1-form $A$ and the curvature $F$ are represented only locally via $\mathfrak{h}$-valued forms on $M$. Under a change of trivialization the local connection does not transform covariantly, but the local curvature does (by conjugation), so the invariant polynomial $f(F)$ is independent on the frame and it defines a global form on $M$. The definition of characteristic classes could be thus given in terms of the local curvature of a vector bundle, without changing the result. We need mainly three examples of characteristic classes, associated to the invariant polynomials $\mathrm{Tr},\det$ and $\mathrm{Pf}$, that corresponds for matrix groups to the standard trace, determinant and pfaffian. These are the _Chern character_ $\mbox{ch}(F):=\mathrm{Tr}\left(e^{F}\right),$ (B.4) the _Euler class_ $e(F):=\mathrm{Pf}\left(\frac{F}{2\pi}\right),$ (B.5) and the _Dirac $\hat{A}$-genus_ $\hat{A}(F):=\sqrt{\det{\left[\frac{\frac{1}{2}F}{\sinh\left(\frac{1}{2}F\right)}\right]}}.$ (B.6) Now we turn the discussion to the case of $G$-equivariant bundles [35, 7, 19]. ###### Definition B.1.1. A _$G$ -equivariant vector bundle_ is a vector bundle $E\xrightarrow{\pi}M$, such that: 1. (i) both $E$ and $M$ are $G$-spaces and $\pi$ is $G$-equivariant; 2. (ii) $G$ acts linearly on the fibers. A principal $H$-bundle $P\xrightarrow{\pi}M$ is $G$-equivariant if 1. (i) both $E$ and $M$ are $G$-spaces and $\pi$ is $G$-equivariant; 2. (ii) the $G$-action commutes with the $H$-action on $P$. Usually, a connection $A$ on a $G$-equivariant principal bundle is required to be _$G$ -invariant_, that is $\mathcal{L}_{X}A=0$ for every $X\in\mathfrak{g}$. If $G$ is compact, this choice is always possible by averaging any connection over $G$ to obtain a $G$-invariant one [35]. Since the $G$\- and the $H$-actions commute, a principal $H$-bundle $P\xrightarrow{\pi}M$ induces another principal $H$-bundle $P_{G}\xrightarrow{\pi_{G}}M_{G}$ over the homotopy quotient $M_{G}$. Topologically, the _equivariant characteristic classes_ of $P\xrightarrow{\pi}M$ are the ordinary characteristic classes of $P_{G}\xrightarrow{\pi_{G}}M_{G}$, thus defining elements in the $G$-equivariant cohomology $H_{G}^{*}(M)$. From the differential geometric point of view, they can be derived as _equivariantly closed extensions_ of the ordinary characteristic classes in the Cartan model. In particular, in [35, 7] it was shown that the equivariant characteristic class associated to an invariant polynomial $f$ is represented by $f(F^{\mathfrak{g}})$, where $F^{\mathfrak{g}}=1\otimes F+\phi^{a}\otimes\mu_{a}$ (B.7) is the equivariant extension of the curvature $F$ on the principal $H$-bundle. $\phi^{a=1,\cdots,\dim\mathfrak{g}}$ are the generators of $S(\mathfrak{g}^{*})$ in the Cartan model, and the map $\mu:\mathfrak{g}\to\Omega(P;\mathfrak{h})$ such that $\mu_{X}:=-\iota_{X}A=-A(\underline{X})$ (B.8) is called _moment map_ , with analogy to the symplectic case. We denoted $\mu_{a}\equiv\mu_{T_{a}}$ with $T_{a=1,\cdots,\dim\mathfrak{g}}$ the basis of $\mathfrak{g}$ dual to $\phi^{a}$. If we define $\nabla=d+A$ the covariant derivative, that in the adjoint bundle acts as $\nabla\omega=d\omega+[A\stackrel{{\scriptstyle\wedge}}{{,}}\omega]$, we notice that we can obtain the above equivariant curvature in the Cartan model from the _equivariant covariant derivative_ $\nabla^{\mathfrak{g}}:=1\otimes\nabla-\phi^{a}\otimes\iota_{a}$ (B.9) that is completely analogous to the definition of the Cartan differential (2.46). With this definition, the equivariant curvature can be expressed as $F^{\mathfrak{g}}=(\nabla^{\mathfrak{g}})^{2}+\phi^{a}\otimes\mathcal{L}_{a},$ (B.10) where the last piece takes care of the non-nilpotency of the Cartan differential on generic differential forms, and moreover it satisfies an equivariant variation of the Bianchi identity $(\nabla^{\mathfrak{g}}F^{\mathfrak{g}})=0.$ (B.11) Notice that, if we assume the connection $A$ to be $G$-invariant, the moment map $\mu$ indeed satisfies a moment map equation with respect to the curvature $F$ (see Section 3.3.2), $\nabla\mu_{X}=-\iota_{X}F\qquad\forall X\in\mathfrak{g}.$ (B.12) Once a suitable equivariant extension of the curvature $F^{\mathfrak{g}}$ is known, the particular equivariant characteristic classes are simply a modification of the old ones, so the equivariant version of the above Chern character, Euler class and Dirac $\hat{A}$-genus are given by $\mathrm{ch}_{G}(F):=\mathrm{Tr}\left(e^{F^{\mathfrak{g}}}\right),\quad e_{G}(F):=\mathrm{Pf}\left(\frac{F^{\mathfrak{g}}}{2\pi}\right),\quad\hat{A}_{G}(F):=\sqrt{\det{\left[\frac{\frac{1}{2}F^{\mathfrak{g}}}{\sinh\left(\frac{1}{2}F^{\mathfrak{g}}\right)}\right]}},$ (B.13) respectively. ### B.2 Universal bundles and equivariant cohomology In this section we motivate the well-definiteness of equivariant cohomology of Section 2.2, starting from the definition of the space $EG$. Proofs for the various propositions we are going to state informally and/or without proof can be found for example in [18, 24, 21]. We should mention that the mathematically correct approach to this subject works considering only _CW complexes_. These are special types of topological spaces that can be constructed by “attaching deformed disks” to each other [24]. We only quote that any smooth manifold can be given the structure of a CW complex, so that in the smooth setting we do not need to bother with this subtlety.111This is a result of Morse theory, see [18] and references therein. ###### Definition B.2.1. A principal $G$-bundle $\pi:EG\rightarrow BG$ is called _universal G-bundle_ if: 1. (i) for any principal $G$-bundle $P\rightarrow X$, there exists a map $h:X\rightarrow BG$ such that $P\cong h^{*}(EG)$ (the pull-back bundle of $EG$ through $h$); 2. (ii) if $h_{0},h_{1}:X\rightarrow BG$ are such that $h_{0}^{*}(EG)\cong h_{1}^{*}(EG)$, then the two maps are homotopic. The base space $BG$ is called _classifying space_. The classifying property (i) required of $EG$ means that for every principal $G$-bundle there is a copy of it sitting inside $EG\rightarrow BG$. The important fact is that existence can be proven for a large class of interesting cases, the argument going as follows. First recall that _homotopic maps pull back to isomorhic bundles_ , i.e. if $E\rightarrow B$ is a vector bundle, $X$ a paracompact space, then $g,h:X\rightarrow B\ \text{homotopic maps}\ \Rightarrow g^{*}(E)\cong h^{*}(E).$ (B.14) Then the property (ii) in the definition above states that if $E\rightarrow B$ is a universal bundle, $\Rightarrow$ is replaced by $\Leftrightarrow$. Now let, for any paracompact space $X$, $P_{G}(X):=\left\\{\text{isomorphism classes of principal G-bundles over X}\right\\}$ (B.15) and for some space $BG$ (to be identified with the classifying space), $[X,BG]:=\left\\{\text{homotopy classes of maps }X\rightarrow BG\right\\}.$ (B.16) Notice that the definition of $P_{G}(X)$ is totally independent from the notion of universal $G$-bundle. Considering then the map $\displaystyle\phi:[X,BG]$ $\displaystyle\rightarrow P_{G}(X)$ (B.17) $\displaystyle[h:X\rightarrow BG]$ $\displaystyle\mapsto h^{*}(EG),$ by (B.14) we have that it is well-defined (independent from the representatives). The conditions (i) and (ii) are equivalent to surjectivity and injectivity of $\phi$, so finally $P_{G}(X)\cong[X,BG]$. Since $P_{G}(X)$ exists, this proves the existence of the classifying space $BG$ and of the universal bundle $EG\rightarrow BG$.222In the language of category theory, we could say that $P_{G}(\cdot)$ is a contravariant functor, _representable_ through $[\cdot,BG]$. We can now motivate the well-definiteness of the Borel construction for equivariant cohomology. A fundamental result for this is that _a principal $G$-bundle is a universal bundle if and only if its total space is (weakly) contractible_.333A _weakly contractible_ space is a topological space whose homotopy groups are all trivial. Clearly any contractible space is weakly contractible. It is a fact that every CW complex that is weakly contractible is also contractible [24], so for our purposes the two concepts coincide. The contractibility of $EG$ makes its cohomology trivial, so that, since $(M\times EG)\sim M$, we have $H^{*}(M\times EG)\cong H^{*}(M)$. When we take the homotopy quotient, the product by $EG$ acts as a “regulator” of the resulting cohomology. In fact, if the action of $G$ on $M$ is free, such that $M\to M/G$ is a principal $G$-bundle, one can prove that for any (weakly) contractible $G$-space $E$ $\faktor{(M\times E)}{G}\sim\faktor{M}{G},$ (B.18) where $\sim$ here stands for “weakly homotopic”.444Two spaces are _weakly homotopic_ if they have the same homotopy groups. Again, homotopy equivalence implies weak homotopy equivalence, and for CW complexes these two concepts coincide. In general, even if the $G$-action is not free, two homotopy quotients with respect to different (weakly) contractible $G$-spaces $E$ and $E^{\prime}$ are (weakly) homotopy equivalent, $\faktor{(M\times E)}{G}\sim\faktor{(M\times E^{\prime})}{G}.$ (B.19) Another known fact is that _weakly homotopic spaces have the same (co)homology groups, for all coefficients_ , generalizing (2.9). Putting together these properties, we have that $H^{*}\left(\faktor{(M\times E)}{G}\right)\cong H^{*}\left(\faktor{(M\times E^{\prime})}{G}\right),$ (B.20) so that the resulting cohomology is independent of the choice of contractible principal $G$-bundle. The homotopy quotient thus well-defines the $G$-equivariant cohomology of $M$, producing an “homotopically correct” version of its orbit space. As pointed out in Section 2.2, when the $G$-action is free on $M$ this reproduces the naive definition of cohomology of the quotient space $M/G$. #### Every compact Lie group has a universal bundle In Section 2.2 we gave the example of the universal bundle for the circle, $EU(1)=\mathbb{S}^{\infty}$ and $BU(1)=\mathbb{C}P^{\infty}$. One can generalize this construction to concretely define a universal bundle for any compact Lie group $G$. This is because any such Lie group embeds into $U(n)$ or $O(n)$ (the maximal compact subgroups of $GL(n,\mathbb{C})$ and $GL(n,\mathbb{R})$), for some $n$, and for them one can construct universal bundles explicitly. As a subgroup, $G$ will act freely on the given universal bundle. Then one can take this to be its universal bundle too. A class of principal $O(n)$\- or $U(n)$-bundles is given by the so-called _Stiefel manifolds_. A Stiefel manifold $V_{k}(\mathbb{F}^{n})$ is the set of all _orthonormal $k$-frames_ in $\mathbb{F}^{n}$, where $\mathbb{F}=\mathbb{R},\mathbb{C}$, and the orthonormality is defined with respect to the canonical Euclidean or sesquilinear inner products. A _$k$ -frame_ is an ordered set $(v_{1},\cdots,v_{k})$ of $k$ linearly independent vectors in $\mathbb{F}^{n}$. Notice that when $k=1$, $V_{1}(\mathbb{C}^{n})$ is the set of all unit vectors in $\mathbb{C}^{n}\cong\mathbb{R}^{2n}$, i.e. the $(2n-1)$-sphere. The latter is acted freely by the group $U(1)$ by diagonal multiplication, and analogously the Stiefel manifold $V_{k}(\mathbb{C}^{n})$ is acted freely by $U(k)$, that essentially rotates the vectors of the $k$-frames. Analogously, $V_{k}(\mathbb{R}^{n})$ is acted freely by $O(k)$. Thus we have the generalization of the sequence of principal $U(k)$\- and $O(k)$-bundles, that in the limit $n\to\infty$ produces the contractible universal bundles $EU(k)=V_{k}(\mathbb{C}^{\infty})$ and $EO(k)=V_{k}(\mathbb{R}^{\infty})$. The Stiefel manifolds can thus be seen as a “higher dimensional versions” of the spheres, in the sense of the following consideration: $V_{k}(\mathbb{R}^{n})\cong\faktor{O(n)}{O(n-k)}.$ (B.21) Comparing with Example 2.2.1, where we remarked that $\mathbb{S}^{n-1}\cong O(n)/O(n-1)$, we see that indeed $\mathbb{S}^{n-1}\equiv V_{1}(\mathbb{R}^{n})$. The $(n-1)$-sphere is just the set of unit vectors in $\mathbb{R}^{n}$, so orthonormal 1-frames. The base spaces $G_{k}(\mathbb{C}^{n})=V_{k}(\mathbb{C}^{n})/U(k)$ and $G_{k}(\mathbb{R}^{n})=V_{k}(\mathbb{R}^{n})/O(k)$ are the sets of equivalence classes of $k$-frames, that identify $k$_-hyperplanes_ through the origin inside $\mathbb{C}^{n}$ or $\mathbb{R}^{n}$. These manifolds are called _Grassmannians_. The _infinite Stiefel manifold_ $V_{k}(\mathbb{F}^{\infty})$ and the _infinite Grassmannian_ $G_{k}(\mathbb{F}^{\infty})$ are thus the total space of the universal bundle and the classifying space for the unitary and orthogonal groups $U(k)$ and $O(k)$, and generalize the universal bundle $\mathbb{S}^{\infty}\to\mathbb{C}P^{\infty}$ of the circle. As recalled above, any compact Lie group $G$ can be embedded as a closed subgroup of an orthogonal group (or a unitary group). This means that $G$ also acts freely on $V_{k}(\mathbb{F}^{\infty})$ for some $k$, and in turn $V_{k}(\mathbb{F}^{\infty})\to V_{k}(\mathbb{F}^{\infty})/G$ is a principal $G$-bundle, whose total space is a contractible space. This gives the universal bundle for any compact Lie group $G$. #### Module structure of equivariant cohomology We end this section with a more algebraic comment about the construction of equivariant cohomology. Notice first that $pt_{G}=\faktor{(pt\times EG)}{G}\cong BG\quad\Rightarrow\quad H_{G}^{*}(pt)\cong H^{*}(BG),$ (B.22) so the equivariant cohomology of a point is the standard cohomology of the classifying space $BG$, generalizing Example 2.2.2. Thus the equivariant cohomology $H_{G}^{*}(\cdot)$ inherits analogous functorial properties to the standard (singular) cohomology of the last section, with respect to the ring $H^{*}(BG)$ instead of the coefficient ring $A\cong H^{*}(pt;A)$. To see this, let us first notice that a $G$-equivariant function $f:M\to N$ between the two $G$-spaces $M,N$ induces a well-defined map between the two homotopy quotients, $\displaystyle f_{G}:M_{G}$ $\displaystyle\rightarrow N_{G}$ (B.23) $\displaystyle[m,e]$ $\displaystyle\mapsto[f(m),e].$ This induced map inherits many properties from $f$: 1. (i) if $f$ is injective (surjective), then $f_{G}$ is injective (surjective); 2. (ii) if $id:M\rightarrow M$ is the identity, then $id_{G}:M_{G}\rightarrow M_{G}$ is the identity; 3. (iii) $(h\circ f)_{G}=h_{G}\circ f_{G}$; 4. (iv) if $f:M\rightarrow N$ is a fiber bundle with fiber $F$, then $f_{G}:M_{G}\rightarrow N_{G}$ is also a fiber bundle with fiber $F$. As pointed out in Section 2.1, a map between two topological spaces induces a map (in the opposite direction) between the associated singular cohomologies, so $f_{G}^{*}:\left(H^{*}(N_{G})\equiv H^{*}_{G}(N)\right)\to\left(H^{*}(M_{G})\equiv H^{*}_{G}(M)\right).$ (B.24) Defining thus a trivial map $\phi:M\to pt$, we see from (B.22) that the induced homomorphism $\phi^{*}_{G}:H^{*}(BG)\to H^{*}_{G}(M)$ makes the equivariant cohomology $H^{*}_{G}(M)$ naturally into a $H^{*}(BG)$-module!555Recall that singular cohomology has a ring structure. Also, in general $f_{G}^{*}:H^{*}_{G}(N)\to H^{*}_{G}(M)$ is a $H^{*}(BG)$-module homomorphism.666In category theory terminology, we could say that the Borel construction $(\cdot)_{G}$ is a covariant functor from the category of $G$-spaces to Top (or Man), and $H_{G}^{*}(\cdot)$ is a contravariant functor between Top (or Man) and the category of $H^{*}(BG)$-modules. Notice that the cohomology of the classifying space $BG$ is usually very simple, as we pointed out in Section 2.4 via its associated Weil model. There is a curious difference between standard cohomology and equivariant cohomology regarding the associated coefficient rings. In the former case, it is clear from the various examples in Section 2.1 that the coefficient ring $\mathbb{R}\cong H^{*}(pt)$ always embeds into the cohomology $H^{*}(M)$ (also for other commutative rings). In the case of equivariant cohomology, on the other hand, the coefficient ring $H_{G}^{*}(pt)=H^{*}(BG)=S(\mathfrak{g}^{*})^{G}$ does not, since the map $\phi_{G}^{*}$ above is not injective in general, as it is clear also from the example of $H^{*}_{U(1)}(\mathbb{S}^{1})=\mathbb{R}$. It turns out that the condition for $H^{*}(BG)$ to embed in $H^{*}_{G}(M)$ is that $G$ acts on $M$ _with fixed points_. We can argue briefly why this is the case. Let $p\in M$ be a fixed point. The inclusion $i:\\{p\\}\to M$ is $G$-equivariant since the action on $p$ is trivial, so there is a well-defined map $i_{G}:pt_{G}=BG\to M_{G}$. This is easily checked to be a section of the bundle $M_{G}\xrightarrow{\pi}BG$, with respect to the projection map $\pi([m,e]):=[e]\in BG$. The identity $\pi\circ i_{G}=id_{M_{G}}$ lifts to the pull-backs in the opposite direction: $i_{G}^{*}\circ\pi^{*}=id$ on $H_{G}^{*}(pt)=H^{*}(BG)$. This means that the map $\pi^{*}:H^{*}(BG)\to H_{G}^{*}(M)$ has a left-inverse, and thus it is injective. This property can be seen in the example of the $U(1)$-equivariant cohomology of the 2-sphere. In this case there are two fixed points, and indeed $H^{*}(BU(1))=\mathbb{R}[\phi]$ embeds in $H^{*}_{U(1)}(\mathbb{S}^{2})=\mathbb{R}[\phi]\oplus\mathbb{R}[\phi]y$, where $y$ can be identified in the Cartan model with the equivariantly closed extension of the volume form, $y\equiv[\tilde{\omega}]$. ### B.3 Fixed point sets and Borel localization We now spend a few words about a procedure that we used many times without many worries, that is to “algebraically localize” the space of equivariant differential forms $\Omega(M)^{U(1)}[\phi]$ with respect to the indeterminate $\phi$, setting it to $\phi=-1$. This localization was useful to simplify the notation in many occasions, but it really has a non-trivial deeper meaning. In fact, it allows to show in a more algebraic way that the $G$-equivariant cohomology of the smooth $G$-manifold $M$ is encoded in the fixed point set $F$ of the $G$-action, at least when $G$ is a torus. The fundamental theorem concerning this point is the so-called _Borel localization theorem_ , that sometimes allows to obtain the ring structure of the equivariant cohomology of the manifold from that of its fixed point set. We consider the case of a circle action here. First, let us recall what localization in algebra means. If $R$ is a commutative ring, the _localization_ of $R$ with respect to a closed subset $S\subseteq R$ is a way to formally introduce a multiplicative inverse for every element of $S$ in $R$, so to introduce _fractions_ in $R$, analogously to what one does in the construction of the rational numbers $\mathbb{Q}$ from the integers $\mathbb{Z}$. This procedure makes the former commutative ring into a field (in the algebraic sense). Since we are interested in $U(1)$-equivariant cohomologies, let us consider an $\mathbb{R}[\phi]$-module $N$, and practically define the _localization of N with respect to $\phi$_ as $N_{\phi}\cong\left\\{\left.\frac{x}{\phi^{n}}\right|x\in N,n\in\mathbb{N}\right\\},$ (B.25) identifying elements in $N_{\phi}$ as $\frac{x}{\phi^{n}}\sim\frac{y}{\phi^{m}}\Leftrightarrow\exists k\in\mathbb{N}:\phi^{k}(\phi^{m}x-\phi^{n}y)=0\ \text{in}\ N.$ (B.26) The simplest example of such a localized module is just $\mathbb{R}[\phi]_{\phi}\cong\mathbb{R}[\phi^{-1},\phi]$, i.e. the Laurent polynomials in $\phi$. Notice that there is always an $\mathbb{R}[\phi]$-module homomorphism that makes $N$ inject into $N_{\phi}$, $i:N\to N_{\phi}$ such that $i(x):=x/\phi^{0}$. If $f:N\to M$ is an $\mathbb{R}[\phi]$-module homomorphism, then there is a well-defined induced homomorphism between the localized modules $f_{\phi}:N_{\phi}\to M_{\phi}$ such that $f(x/\phi^{n}):=f(x)/\phi^{n}$. The important algebraic property of localization for what concerns this discussion is that _it commutes with cohomology_ : if $(A,d)$ is a differential complex, $A^{(0)}\xrightarrow{d}A^{(1)}\xrightarrow{d}\cdots,\qquad d^{2}=0,$ (B.27) where $A^{(i)}$ are $\mathbb{R}[\phi]$-modules, then also $(A_{\phi},d_{\phi})$ is a differential complex, and $H^{*}(A,d)_{\phi}\cong H^{*}(A_{\phi},d_{\phi}).$ (B.28) Quite analogously, from Example 2.4.1 onward we substitute the _indeterminate_ $\phi\in S(\mathfrak{u}(1)^{*})$ with a _variable_ , and then set it to the value $\phi=-1$ for notational convenience. Stated more formally, we start from the Cartan model of $U(1)$-equivariant differential forms $\Omega(M)^{U(1)}[\phi]$, that has clearly an $\mathbb{R}[\phi]$-module structure, and localize it to $\Omega(M)^{U(1)}[\phi]_{\phi}$, so introducing $\phi$ also at the denominator. This puts $\phi$ on the same footing as a real variable, so that we are allowed to fix it to some value, for convenience only. Notice that operations like (3.8), where we “invert” an equivariant form, are allowed only in the localized module $\Omega(M)^{U(1)}[\phi]_{\phi}$, where the division by $\phi$ is meaningful. From the result (B.28), we understand that this localization of the Cartan model does not spoil the resulting equivariant cohomology $H_{U(1)}^{*}(M)$, because the two operations commute.777Notice that $H_{U(1)}^{*}(M)$ has generically an $\mathbb{R}[\phi]$-module structure, by the discussion in Appendix B.2 and the application of the Weil model (see Section 2.4) $H^{*}(BG)\cong S(\mathfrak{u}(1)^{*})\cong\mathbb{R}[\phi]$. The Borel localization theorem relates really the localized equivariant cohomologies of the $U(1)$-manifold $M$ and of its fixed point locus $F$. To understand what this has to say about the actual equivariant cohomology of $M$, we recall first some other algebraic facts. A _torsion_ element in a module $N$ over a ring $R$, is an element $x\in N$ such that $\exists r\neq 0\in R:rx=0$. If $N$ is an $\mathbb{R}[\phi]$ module, the element $x$ is said to be $\phi$-_torsion_ if it exists some power of $\phi$ that annihilates it: $\phi^{k}x=0$ for some $k\in\mathbb{N}$. The module $N$ is $\phi$-torsion if every one of its elements is $\phi$-torsion. It is easy to see that888Just consider that in the localized module $x\sim\frac{\phi^{k}}{\phi^{k}}x$, so if $x$ is $\phi$-torsion it is equivalent to 0 in $N_{\phi}$. $N\ \text{is }\phi\text{-torsion}\quad\Leftrightarrow\quad N_{\phi}=0.$ (B.29) Applying this to the case of $N=H^{*}_{U(1)}(M)$, we can see that the equivariant cohomology in the case of a free $U(1)$-action on $M$ is $\phi$-torsion. In fact, if the action is free, we can easily compute $H_{U(1)}^{*}(M)=H^{*}(M/U(1))$, so that $H_{U(1)}^{k}(M)=0$ in some degree $k>\dim(M/U(1))$. This means that $\phi^{k}\cdot H_{U(1)}^{*}(M)=0$ for some $k$ high enough. The first argument in Section 3.1 in fact is the proof that more is true: $H^{*}_{U(1)}(M)$ is $\phi$-torsion if the $U(1)$-action is _locally free_ on $M$, since we found essentially $(H^{*}_{U(1)}(M))_{\phi}=0$ as the Poincaré lemma, after having introduced $\phi$ at the denominator.999In Section 3.1 $M$ is the manifold without its fixed point set, there called $\tilde{M}$. This motivates the following theorem, that states that, _up to torsion_ , the $U(1)$-equivariant cohomology of $M$ is concentrated on its fixed point set. A proof can be found in [22, 18]. ###### Theorem B.3.1 (Borel localization). Let $U(1)$ act smoothly on the manifold $M$, with compact fixed point set $F$. The inclusion $i:F\hookrightarrow M$ induces an isomorphism of algebras over $\mathbb{R}[\phi]$, $i^{*}_{\phi}:H^{*}_{U(1)}(M)_{\phi}\to H^{*}_{U(1)}(F)_{\phi}.$ This theorem is an “abstract version” of the localization theorems described in Chapter 3, and intuitively gives another way to see that they have to be true, without having to travel through all the smooth algebraic models and the integration theory that we described in due time. It shows that localization is something present at a very low level of structure, originating just from the topological nature of equivariant cohomology. ### B.4 Equivariant integration and Stokes’ theorem In this section we define what it means to integrate a $G$-equivariant differential form $\omega\in\Omega_{G}(M)$ over a smooth, oriented $G$-manifold $M$ of dimension $dim(M)=n$ and we report an extended version of Stokes’ theorem that applies in the equivariant setup. Let $G$ be a connected Lie group acting (smoothly) on the left on $M$, being $\\{\phi^{a}\\}_{a=1,\cdots,dim(\mathfrak{g})}$ a basis for $\mathfrak{g}^{*}:=Lie(G)^{*}$. If the equivariant form $\omega$ is of degree $k$, we can express it as $\omega=\omega^{(k)}+\omega^{(k-2)}_{a}\phi^{a}+\omega^{(k-4)}_{ab}\phi^{a}\phi^{b}+\cdots=\sum_{p\geq 0}\omega^{(k-2p)}_{a_{1}\cdots a_{p}}\phi^{a_{1}}\cdots\phi^{a_{p}}$ (B.30) where the coefficients are differential forms on $M$, tensor products have been suppressed and we require $\omega$ to be $G$-invariant. The natural way to define integration of such objects is obtained just making the integral $\int_{M}$ act on the coefficients $\omega^{(k-2p)}_{a_{1}\cdots a_{p}}$ of the $\phi$-expansion of $\omega$. In this way, one obtains a map $\int_{M}:\Omega_{G}(M)\to S(\mathfrak{g}^{*})\equiv\mathbb{R}[\phi^{a}].$ (B.31) Thanks to the equivariant Stokes’ theorem (to be stated later), this descends also in equivariant cohomology, $\int_{M}:H^{*}_{G}(M)\to S(\mathfrak{g}^{*})$, analogously to the standard (non-equivariant) case. ###### Definition B.4.1. The integral on $M$ of the $G$-equivariant form $\omega$ of $\textrm{deg}(\omega)=k$ is defined as $\int_{M}\omega:=\sum_{p\geq 0}\left(\int_{M}\omega^{(k-2p)}_{a_{1}\cdots a_{p}}\right)\phi^{a_{1}}\cdots\phi^{a_{p}}.$ Notice that if $n$ and $k$ are of different parity, the integral is automatically zero. If instead $k=n+2m$ for some $m\in\mathbb{Z}$, then $\int_{M}\omega=\begin{cases}\left(\int_{M}\omega^{(n)}_{a_{1}\cdots a_{m}}\right)\phi^{a_{1}}\cdots\phi^{a_{m}}&k\geqslant n\\\ 0&k<n.\end{cases}$ (B.32) In particular, if we have a top form $\omega\in\Omega(M)$ on M and $\tilde{\omega}$ is any equivariant extension of $\omega$ in $\Omega_{G}(M)$, then we can deform the integral $\int_{M}\omega=\int_{M}\tilde{\omega}$ (B.33) without changing its value. We can then prove the equivariant version of the Stokes’ theorem. ###### Theorem B.4.1. Let $G$ be a connected Lie group acting (smoothly) on the left on a smooth manifold $M$ with boundary $\partial M$. If $\omega\in\Omega_{G}(M)$ of $\textrm{deg}(\omega)=k$, then $\int_{M}d_{C}\omega=\int_{\partial M}\omega$ where $d_{C}=1\otimes d+\phi^{a}\otimes\iota_{a}$ is the Cartan differential and $\iota_{a}\equiv\iota_{T_{a}}$, with $\\{T_{a}\\}$ a basis of $\mathfrak{g}^{*}$ dual to $\\{\phi^{a}\\}$. ###### Proof. The proof follows from the direct evaluation and the standard Stokes’ theorem. If the integral is not zero, selecting the component of top-degree, $\left.(d_{C}\omega)\right|_{(n)}=d\omega^{(n-1)}_{I}\phi^{I}+(\iota_{\\{a}\omega^{(n+1)}_{J\\}})\phi^{a}\phi^{J},$ where $I=(a_{1},\cdots,a_{(k-n)/2})$ and $J=(a_{1},\cdots,a_{(k-n+1)/2})$. The second term vanishes since $\omega^{(n+1)}=0$ by dimensionality. So the integral of $d_{C}\omega$ is the integral of the first term, on which we can use the standard version of Stokes’ theorem, and getting the statement of the theorem. ∎ ## References * [1] H. Cartan “Notions d’algèbre différentielle; application aux groupes de Lie et aux variétés où opère un groupe de Lie” In _Colloque de topologie (espaces fibrés)_ , 1951 * [2] H. Cartan “La transgression dans un groupe de Lie et dans un espace fibré principal” In _Colloque de topologie (espaces fibrés)_ , 1951 * [3] A. Borel “Sur la cohomologie des espaces fibrés principaux et des espace homogènes des groupes de Lie compacts” In _Annals of Mathematics_ 57, 1953 * [4] A. Borel “Seminar on Transformation Groups”, Annals of Mathematics Studies Princeton University Press, 1960 URL: http://www.jstor.org/stable/j.ctt1bd6jxd * [5] J.J. Duistermaat and G.J. Heckman “On the variation in the cohomology of the symplectic form of the reduced phase space” In _Invent. Math._ 69, 1982 * [6] M.F. Atiyah and R. Bott “The moment map and equivariant cohomology” In _Topology_ 23, 1984 DOI: https://doi.org/10.1016/0040-9383(84)90021-1 * [7] N. Berline and M. Vergne “Zeros d’un champ de vecteurs et classes caracteristiques equivariantes” In _Duke Mathematical Journal_ 50 Duke University Press, 1983 DOI: 10.1215/S0012-7094-83-05024-X * [8] M.F. Atiyah “Circular symmetry and stationary-phase approximation” In _Colloque en l’honneur de Laurent Schwartz - Volume 1_ , Astérisque Société mathématique de France, 1985 URL: http://www.numdam.org/item/AST_1985__131__43_0 * [9] A.Y. Morozov, A.J. Niemi and K. Palo “Supersymplectic geometry of supersymmetric quantum field theories” In _Nuclear Physics B_ 377, 1992 DOI: https://doi.org/10.1016/0550-3213(92)90026-8 * [10] K. Palo “Symplectic geometry of supersymmetry and nonlinear sigma model” In _Physics Letters B_ 321, 1994 DOI: 10.1016/0370-2693(94)90327-1 * [11] V. Pestun “Localization of Gauge Theory on a Four-Sphere and Supersymmetric Wilson Loops” In _Communications in Mathematical Physics_ 313, 2012 DOI: 10.1007/s00220-012-1485-0 * [12] V. Pestun et al. “Localization techniques in quantum field theories” In _Journal of Physics A: Mathematical and Theoretical_ 50, 2017 DOI: 10.1088/1751-8121/aa63c1 * [13] J. Maldacena “The Large N Limit of Superconformal Field Theories and Supergravity” In _International Journal of Theoretical Physics_ 38, 1999 DOI: 10.1023/a:1026654312961 * [14] V. Guillemin and E. Prato “Heckman, Kostant, and Steinberg formulas for symplectic manifolds” In _Advances in Mathematics_ 82, 1990 DOI: 10.1016/0001-8708(90)90087-4 * [15] E. Witten “Two dimensional gauge theories revisited” In _Journal of Geometry and Physics_ 9, 1992 DOI: 10.1016/0393-0440(92)90034-x * [16] L.C. Jeffrey and F.C. Kirwan “Localization for nonabelian group actions”, 1993 arXiv:alg-geom/9307001 * [17] R. Bott and L.W. Tu “Differential Forms in Algebraic Topology” Springer-Verlag, 1982 DOI: 10.1007/978-1-4757-3951-0 * [18] L.W. Tu “Introductory Lectures on Equivariant Cohomology”, Annals of Mathematics Studies Princeton University Press, 2020 * [19] J.R. Szabo “Equivariant Localization of Path Integrals”, 1996 arXiv:hep-th/9608068 * [20] S. Cordes, G. Moore and S. Ramgoolam “Lectures on 2D yang-mills theory, equivariant cohomology and topological field theories” In _Nuclear Physics B - Proceedings Supplements_ 41, 1995 DOI: 10.1016/0920-5632(95)00434-b * [21] R. Bott “An Introduction to Equivariant Cohomology” In _NATO Sci. Ser. C_ 530, 1999 DOI: 10.1007/978-94-011-4542-8“˙3 * [22] V.W. Guillemin and S. Sternberg “Supersymmetry and equivariant de Rham theory” Springer, 1999 * [23] M. Nakahara “Geometry, topology and physics” Institute of Physics Publishing, 2003 * [24] A. Hatcher “Algebraic topology” Cambridge University Press, 2002 * [25] L.W. Tu “An introduction to manifolds” Springer, 2010 * [26] J.J. Rotman “Advanced modern algebra” Prentice Hall, 2003 * [27] A. Marsh “Mathematics for Physics” World Scientific Publishing, 2018 DOI: 10.1142/10816 * [28] M. Audin “Torus Actions on Symplectic Manifolds” Springer Basel, 2004 DOI: 10.1007/978-3-0348-7960-6 * [29] V. Mathai and D. Quillen “Superconnections, Thom classes, and equivariant differential forms” In _Topology_ 25, 1986 DOI: 10.1016/0040-9383(86)90007-8 * [30] J. Kalkman “BRST model for equivariant cohomology and representatives for the equivariant Thom class” In _Communications in Mathematical Physics_ 153, 1993 DOI: 10.1007/BF02096949 * [31] M. Kankaanrinta “Proper smooth G-manifolds have complete G-invariant Riemannian metrics” In _Topology and its Applications_ 153, 2005 DOI: 10.1016/j.topol.2005.01.034 * [32] T. Bröcker and T. Dieck “Representations of Compact Lie Groups” Springer-Verlag, 1985 DOI: 10.1007/978-3-662-12918-0 * [33] J.E. Marsden and T.S. Ratiu “Introduction to Mechanics and Symmetry: A Basic Exposition of Classical Mechanical Systems” Springer Publishing Company, Incorporated, 1998 DOI: 10.1007/978-0-387-21792-5 * [34] A.C. Da Silva “Lectures on Symplectic Geometry” Springer-Verlag, 2008 DOI: 10.1007/978-3-540-45330-7 * [35] R. Bott and L.W. Tu “Equivariant characteristic classes in the Cartan model”, 2001 arXiv:math/0102001 * [36] E. Witten “Supersymmetry and Morse theory” In _Journal of Differential Geometry_ 17, 1982 DOI: 10.4310/jdg/1214437492 * [37] V. Guillemin and S. Sternberg “Symplectic techniques in physics” Cambridge University Press, 1990 * [38] A.S. Cattaneo and F. Schätz “Introduction to supergeometry” In _Reviews in Mathematical Physics_ 23, 2011 DOI: 10.1142/s0129055x11004400 * [39] M. Batchelor “Two approaches to supermanifolds” In _Trans. Amer. Math. Soc._ , 1980 DOI: 10.1090/S0002-9947-1980-0554332-9 * [40] E. Witten “Notes On Supermanifolds and Integration”, 2012 arXiv:1209.2199 * [41] A.. Niemi and K. Palo “Equivariant Morse theory and quantum integrability”, 1994 arXiv:hep-th/9406068 * [42] J. Figueroa-O’Farrill “Majorana Spinors”, Notes available at https://www.maths.ed.ac.uk/~jmf/Teaching/Notes.html * [43] V. Varadarajan “Supersymmetry for mathematicians: an introduction” Courant Lecture Notes in Mathematics, American Mathematical Society, 2004 DOI: 10.1090/cln/011 * [44] L. Brink, J.H. Schwarz and J. Scherk “Supersymmetric Yang-Mills theories” In _Nuclear Physics B_ 121, 1977 DOI: 10.1016/0550-3213(77)90328-5 * [45] J. Wess and J. Bagger “Supersymmetry and supergravity” Princeton University Press, 1992 * [46] J. Figueroa-O’Farrill “BUSSTEPP Lectures on Supersymmetry”, 2001 arXiv:hep-th/0109172 * [47] E. D’Hoker and D.Z. Freedman “Supersymmetric gauge theories and the AdS/CFT correspondence”, 2002 arXiv:hep-th/0201253 * [48] S. Weinberg and E. Witten “Limits on massless particles” In _Physics Letters B_ 96, 1980 DOI: 10.1016/0370-2693(80)90212-9 * [49] A. Kapustin, B. Willett and I. Yaakov “Exact results for Wilson loops in superconformal Chern-Simons theories with matter” In _Journal of High Energy Physics_ 2010, 2010 DOI: 10.1007/jhep03(2010)089 * [50] M. Mariño “Lectures on localization and matrix models in supersymmetric Chern–Simons-matter theories” In _Journal of Physics A: Mathematical and Theoretical_ 44, 2011 DOI: 10.1088/1751-8113/44/46/463001 * [51] N. Seiberg “Supersymmetry and Nonperturbative beta Functions” In _Physics Letters B_ 206, 1988 DOI: 10.1016/0370-2693(88)91265-8 * [52] D.V. Alekseevsky, V. Cortés, C. Devchand and U. Semmelmann “Killing spinors are Killing vector fields in Riemannian supergeometry” In _Journal of Geometry and Physics_ 26, 1998 DOI: 10.1016/s0393-0440(97)00036-3 * [53] H. Baum “Conformal Killing spinors and special geometric structures in Lorentzian geometry: A Survey”, 2002 arXiv:math/0202008 * [54] G. Festuccia and N. Seiberg “Rigid supersymmetric theories in curved superspace” In _Journal of High Energy Physics_ 2011, 2011 DOI: 10.1007/jhep06(2011)114 * [55] C. Klare, A. Tomasiello and A. Zaffaroni “Supersymmetry on curved spaces and holography” In _Journal of High Energy Physics_ 2012, 2012 DOI: 10.1007/jhep08(2012)061 * [56] C. Closset, T.T. Dumitrescu, G. Festuccia and Z. Komargodski “Supersymmetric field theories on three-manifolds” In _Journal of High Energy Physics_ 2013, 2013 DOI: 10.1007/jhep05(2013)017 * [57] T.T. Dumitrescu, G. Festuccia and N. Seiberg “Exploring curved superspace” In _Journal of High Energy Physics_ 2012, 2012 DOI: 10.1007/jhep08(2012)141 * [58] A. Kehagias and J.G. Russo “Global supersymmetry on curved spaces in various dimensions” In _Nuclear Physics B_ 873, 2013 DOI: 10.1016/j.nuclphysb.2013.04.010 * [59] P. Di Francesco, P. Mathieu and D. Sénéchal “Conformal Field Theory” Springer-Verlag, 1997 DOI: 10.1007/978-1-4612-2256-9 * [60] T.T. Dumitrescu and N. Seiberg “Supercurrents and brane currents in diverse dimensions” In _Journal of High Energy Physics_ 2011, 2011 DOI: 10.1007/jhep07(2011)095 * [61] S. Ferrara and M. Porrati “Central extensions of supersymmetry in four and three dimensions” In _Physics Letters B_ 423, 1998 DOI: 10.1016/s0370-2693(97)01586-4 * [62] A. Gorsky and M. Shifman “More on the tensorial central charges in N=1 supersymmetric gauge theories: BPS wall junctions and strings” In _Physical Review D_ 61, 2000 DOI: 10.1103/physrevd.61.085001 * [63] K.S. Stelle and P.C. West “Minimal auxiliary fields for supergravity” In _Physics Letters B_ 74, 1978 DOI: 10.1016/0370-2693(78)90669-X * [64] M.F. Sohnius and P.C. West “An alternative minimal off-shell version of N=1 supergravity” In _Physics Letters B_ 105, 1981 DOI: 10.1016/0370-2693(81)90778-4 * [65] R.M. Wald “General Relativity” Chicago University Press, 1984 DOI: 10.7208/chicago/9780226870373.001.0001 * [66] N. Berkovits “A ten-dimensional superYang-Mills action with off-shell supersymmetry” In _Phys. Lett. B_ 318, 1993 DOI: 10.1016/0370-2693(93)91791-K * [67] S. Cremonesi “An introduction to localization and supersymmetry in curved space” In _PoS_ Modave2013, 2013 * [68] I.H. Duru and H. Kleinert “Solution of the path integral for the H-atom” In _Physics Letters B_ 84, 1979 DOI: 10.1016/0370-2693(79)90280-6 * [69] M.E. Peskin and D.V. Schroeder “An Introduction to quantum field theory” Addison-Wesley, 1995 * [70] M. Srednicki “Quantum Field Theory” Cambridge University Press, 2007 URL: https://web.physics.ucsb.edu/~mark/qft.html * [71] A.J. Niemi and O. Tirkkonen “Cohomological partition functions for a class of bosonic theories” In _Physics Letters B_ 293, 1992 DOI: 10.1016/0370-2693(92)90893-9 * [72] N. Berline, E. Getzler and M. Vergne “Heat Kernels and Dirac Operators” Springer, 2004 * [73] D. Friedan and P. Windey “Supersymmetric derivation of the Atiyah-Singer index and the chiral anomaly” In _Nuclear Physics B_ 235, 1984 DOI: 10.1016/0550-3213(84)90506-6 * [74] L. Alvarez-Gaume “Supersymmetry and the Atiyah-Singer Index Theorem” In _Communications in Mathematical Physics_ 90, 1983 DOI: 10.1007/BF01205500 * [75] A. Hietamaki, A.Y. Morozov, A.J. Niemi and K. Palo “Geometry of N=1/2 supersymmetry and the Atiyah-Singer index theorem” In _Physics Letters B_ 263, 1991 DOI: 10.1016/0370-2693(91)90481-5 * [76] M.. Atiyah and I.. Singer “The index of elliptic operators on compact manifolds” In _Bulletin of the American Mathematical Society_ 69, 1963 URL: https://projecteuclid.org:443/euclid.bams/1183525276 * [77] J.A. Wolf and S.S. Chern “Essential Self Adjointness for the Dirac Operator and Its Square” In _Indiana University Mathematics Journal_ 22, 1973 URL: http://www.jstor.org/stable/24890502 * [78] S. Cecotti and L. Girardello “Functional Measure, Topology and Dynamical Supersymmetry Breaking” In _Physics Letters B_ 110, 1982 DOI: 10.1016/0370-2693(82)90947-9 * [79] B. Willett “Localization on three-dimensional manifolds” In _Journal of Physics A: Mathematical and Theoretical_ 50, 2017 DOI: 10.1088/1751-8121/aa612f * [80] J. Maldacena “Wilson Loops in Large-N Field Theories” In _Physical Review Letters_ 80, 1998 DOI: 10.1103/physrevlett.80.4859 * [81] D. Correa, J. Henn, J. Maldacena and A. Sever “An exact formula for the radiation of a moving quark in N=4 super Yang Mills” In _Journal of High Energy Physics_ 2012, 2012 DOI: 10.1007/jhep06(2012)048 * [82] O. Aharony et al. “The deconfinement and Hagedorn phase transitions in weakly coupled large N gauge theories” In _Comptes Rendus Physique_ 5, 2004 DOI: 10.1016/j.crhy.2004.09.012 * [83] E. Witten “Quantum Field Theory and the Jones polynomial” In _Communications in Mathematical Physics_ 121, 1989 DOI: 10.1007/BF01217730 * [84] N. Drukker et al. “Roadmap on Wilson loops in 3d Chern–Simons-matter theories” In _Journal of Physics A: Mathematical and Theoretical_ 53, 2020 DOI: 10.1088/1751-8121/ab5d50 * [85] J. Gomis, T. Okuda and D. Trancanelli “Quantum ’t Hooft operators and S-duality in N=4 super Yang-Mills”, 2009 arXiv:0904.4486 * [86] J. Gomis, T. Okuda and V. Pestun “Exact results for ’t Hooft loops in Gauge theories on $S^{4}$” In _Journal of High Energy Physics_ 2012, 2012 DOI: 10.1007/jhep05(2012)141 * [87] N. Drukker, S. Giombi, R. Ricci and D. Trancanelli “More supersymmetric Wilson loops” In _Physical Review D_ 76, 2007 DOI: 10.1103/physrevd.76.107703 * [88] N. Drukker, S. Giombi, R. Ricci and D. Trancanelli “Wilson loops: From 4D supersymmetric Yang-Mills theory to 2D Yang-Mills theory” In _Physical Review D_ 77, 2008 DOI: 10.1103/physrevd.77.047901 * [89] N.A. Nekrasov “Seiberg-Witten Prepotential From Instanton Counting”, 2002 arXiv:hep-th/0206161 * [90] T. Okuda and V. Pestun “On the instantons and the hypermultiplet mass of N=2* super Yang-Mills on $S^{4}$” In _Journal of High Energy Physics_ 03, 2012 DOI: 10.1007/JHEP03(2012)017 * [91] J.K. Erickson, G.W. Semenoff and K. Zarembo “Wilson loops in supersymmetric Yang–Mills theory” In _Nuclear Physics B_ 582, 2000 DOI: 10.1016/s0550-3213(00)00300-x * [92] S.-J. Rey and J.-T. Yee “Macroscopic strings as heavy quarks: Large-N gauge theory and anti-de Sitter supergravity” In _The European Physical Journal C_ 22 Springer ScienceBusiness Media LLC, 2001 DOI: 10.1007/s100520100799 * [93] N. Drukker and D.J. Gross “An exact prediction of N=4 supersymmetric Yang–Mills theory for string theory” In _Journal of Mathematical Physics_ 42, 2001 DOI: 10.1063/1.1372177 * [94] M. Mariño “Les Houches lectures on matrix models and topological strings”, 2004 arXiv:hep-th/0410165 * [95] T. Okuda and D. Trancanelli “Spectral curves, emergent geometry, and bubbling solutions for Wilson loops” In _Journal of High Energy Physics_ 09, 2008 DOI: 10.1088/1126-6708/2008/09/050 * [96] K. Zarembo “Localization and AdS/CFT correspondence” In _Journal of Physics A: Mathematical and Theoretical_ 50, 2017 DOI: 10.1088/1751-8121/aa585b * [97] D. Gaiotto and X. Yin “Notes on superconformal Chern-Simons-Matter theories” In _Journal of High Energy Physics_ 2007, 2007 DOI: 10.1088/1126-6708/2007/08/056 * [98] M.S. Bianchi et al. “Framing and localization in Chern-Simons theories with matter” In _Journal of High Energy Physics_ 2016, 2016 DOI: 10.1007/jhep06(2016)133 * [99] O. Aharony, O. Bergman, D. Jafferis and J. Maldacena “N = 6 superconformal Chern-Simons-matter theories, M2-branes and their gravity duals” In _Journal of High Energy Physics_ 2008, 2008 DOI: 10.1088/1126-6708/2008/10/091 * [100] N. Drukker and D. Trancanelli “A Supermatrix model for N=6 super Chern-Simons-matter theory” In _Journal of High Energy Physics_ 02, 2010 DOI: 10.1007/JHEP02(2010)058 * [101] M. Mariño and P. Putrov “Exact Results in ABJM theory from Topological Strings” In _Journal of High Energy Physics_ 06, 2010 DOI: 10.1007/JHEP06(2010)011 * [102] C. Beasley and E. Witten “Non-Abelian Localization For Chern-Simons Theory”, 2005 arXiv:hep-th/0503126 * [103] M.. Atiyah and R. Bott “The Yang-Mills Equations over Riemann Surfaces” In _Philosophical Transactions of the Royal Society of London_ 308, A, 1983 * [104] J. Marsden and A. Weinstein “Reduction of symplectic manifolds with symmetry” In _Reports on Mathematical Physics_ 5, 1974 DOI: 10.1016/0034-4877(74)90021-4 * [105] K. Meyer “Symmetries and Integrals in Mechanics” In _Dynamical Systems_ Academic Press, 1973 DOI: 10.1016/B978-0-12-550350-1.50025-4 * [106] D. Birmingham, M. Blau, M. Rakowski and G. Thompson “Topological field theory” In _Physics Reports_ 209, 1991 DOI: 10.1016/0370-1573(91)90117-5 * [107] M. Blau and G. Thompson “Topological gauge theories of antisymmetric tensor fields” In _Annals of Physics_ 205, 1991 DOI: 10.1016/0003-4916(91)90240-9 * [108] A. Bassetto and L. Griguolo “Two-dimensional QCD, instanton contributions and the perturbative Wu-Mandelstam-Leibbrandt prescription” In _Physics Letters B_ 443, 1998 DOI: 10.1016/S0370-2693(98)01319-7 * [109] A. Bassetto et al. “Correlators of supersymmetric Wilson-loops, protected operators and matrix models in N=4 SYM” In _Journal of High Energy Physics_ 08, 2009 DOI: 10.1088/1126-6708/2009/08/061 * [110] A. Bassetto and S. Thambyahpillai “Quantum ’t Hooft Loops of SYM N=4 as instantons of YM2 in Dual Groups SU(N) and SU(N)/ZN” In _Letters in Mathematical Physics_ 98, 2011 DOI: 10.1007/s11005-011-0480-2 * [111] N. Caporaso et al. “Topological Strings, Two-Dimensional Yang-Mills Theory and Chern-Simons Theory on Torus Bundles”, 2006 arXiv:hep-th/0609129 * [112] R.J. Szabo and M. Tierz “q-deformations of two-dimensional Yang-Mills theory: classification, categorification and refinement” In _Nuclear Physics B_ 876, 2013 DOI: 10.1016/j.nuclphysb.2013.08.001 * [113] L. Santilli, R.J. Szabo and M. Tierz “Five-dimensional cohomological localization and squashed q-deformations of two-dimensional Yang-Mills theory” In _Journal of High Energy Physics_ 2020, 2020 DOI: 10.1007/jhep06(2020)036 * [114] A. Dabholkar, J. Gomes and S. Murthy “Quantum black holes, localization, and the topological string” In _Journal of High Energy Physics_ 2011, 2011 DOI: 10.1007/jhep06(2011)019 * [115] A. Zaffaroni “Lectures on AdS Black Holes, Holography and Localization” In _Living Reviews in Relativity_ 23, 2019 arXiv:1902.07176 * [116] L.W. Tu “Differential Geometry” Springer, 2017 DOI: 10.1007/978-3-319-55084-8
# Some applications of transversality for infinite dimensional manifolds Kaveh Eftekharinasab Topology lab. Institute of Mathematics of National Academy of Sciences of Ukraine, Tereshchenkivska st. 3, Kyiv, 01601 Ukraine <EMAIL_ADDRESS> ###### Abstract. We present some transversality results for a category of Fréchet manifolds, the so-called $MC^{k}$-Fréchet manifolds. In this context, we apply the obtained transversality results to construct the degree of nonlinear Fredholm mappings by virtue of which we prove a rank theorem, an invariance of domain theorem and a Bursuk-Ulam type theorem. ###### Key words and phrases: Transversality, degree of nonlinear Fredholm mappings, Fréchet manifolds ###### 2020 Mathematics Subject Classification: 57N75, 58B15, 47H11. This paper is devoted to the development of transversality and its applications to degree theory of nonlinear Fredholm mappings for non- Banachable Fréchet manifolds. The elaboration is mostly, but not entirely, routine; we shall discuss the related issues. In attempting to develop transversality to Fréchet manifolds we face the following drawbacks which are related to lack of a suitable topology on a space of continuous linear maps: 1. (1) In general, the set of isomorphisms between Fréchet spaces is not open in the space of continuous linear mappings. 2. (2) In general, the set of Fredholm operators between Fréchet spaces is not open in the space of continuous linear mappings. Also, a key point in the proof of an infinite dimensional version of Sard’s theorem is that a Fredholm mapping $\varphi$ near origin has a local representation of the form $\varphi(u,v)=(u,\eta(u,v))$ for some smooth mapping $\eta$; indeed, this is a consequence of an inverse function theorem. To obtain a version of Sard’s theorem for Fréchet manifolds, based on the ideas of Müller [4], it was proposed by the author ([1]) to consider Fredholm operators which are Lipschitz on their domains. There is an appropriate metrizable topology on a space of Lipschitz linear mappings so that if we employ this space instead of a space of continuous linear mappings, the mentioned openness issues and the problem of stability of Fredholm mappings under small perturbation can be resolved. Furthermore, for mappings belong to a class of differentiability, bounded or $MC^{k}$-differentiability which is introduced in [4], a suitable version of an inverse function theorem is available, [4, Theorem 4.7]. An example of Lipschitz-Fredholm mapping of class $MC^{k}$ can be found in [3], where the Sard’s theorem [1, Theorem 4.3] is applied to classify all the holomorphic functions locally definable; this gives the additional motivation to study further applications of Sard’s theorem. In this paper, first we improve the transversality theorem [2, Theorem 4.2] by considering all mappings of class $MC^{k}$, then use it to prove the parametric transversality theorem. Then, for Lipschitz-Fredholm mappings of class $MC^{k}$ we apply the transversality theorem to construct the degree (due to Cacciappoli, Shvarts and Smale), which is defined as the group of non- oriented cobordism class of $\varphi^{-1}(q)$ for some regular value $q$. We then prove a rank theorem for Lipschitz-Fredholm mappings of class $MC^{k}$ , and use it to prove an invariance of domain theorem and a Fredholm alternative theorem. Also, using the parametric transversality theorem we obtain a Bursuk-Ulam type theorem. ## 1\. Bounded Fréchet manifolds In this section, we shall briefly recall the basics of $MC^{k}$-Fréchet manifolds for the convenience of readers, which also allows us to establish our notations for the rest of the paper. For more studies, we refer to [1, 2]. Throughout the paper we assume that $E,F$ are Fréchet spaces and $CL(E,F)$ is the space of all continuous linear mappings from $E$ to $F$ topologized by the compact-open topology. If $T$ is a topological space by $U\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}T$ we mean $U$ is open in $T$. Let $\varphi:U\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}E\to F$ be a continuous map. If the directional (Gâteaux) derivatives $\operatorname{D}\varphi(x)h=\lim_{t\to 0}\dfrac{\varphi(x+th)-\varphi(x)}{t}$ exist for all $x\in U$ and all $h\in E$, and the induced map $\operatorname{D}\varphi(x):U\rightarrow CL(E,F)$ is continuous for all $x\in U$, then we say that $\varphi$ is a Keller’s differentiable map of class $C^{1}_{c}$. The higher directional derivatives and $C^{k}_{c}$-mappings, $k\geq 2$, are defined in the obvious inductive fashion. To define bounded or $MC^{k}$-differentiability, we endow a Fréchet space $F$ with a translation invariant metric $\varrho$ defining its topology, and then introduce the metric concepts which strongly depend on the choice of $\varrho$. We consider only metrics of the following form $\varrho(x,y)=\sup_{n\in\mathbb{N}}\dfrac{1}{2^{n}}\dfrac{\left\lVert x-y\right\rVert_{F,n}}{1+\left\lVert x-y\right\rVert_{F,n}},$ where $\left\lVert\cdot\right\rVert_{F,n}$ is a collection of seminorms generating the topology of $F$. Let $\sigma$ be a metric that defines the topology of a Fréchet space $E$. Let $\mathbb{L}_{\sigma,\varrho}(E,F)$ be the set of all linear mappings $L:E\rightarrow F$ which are (globally) Lipschitz continuous as mappings between metric spaces $E$ and $F$, that is $\mathpzc{Lip}(L)\,\coloneq\displaystyle\sup_{x\in E\setminus\\{0_{F}\\}}\dfrac{\varrho(L(x),0_{F})}{\sigma(x,0_{F})}<\infty,$ where $\mathpzc{Lip}(L)$ is the (minimal) Lipschitz constant of $L$. The translation invariant metric $\mathbbm{d}_{\sigma,\varrho}:\mathbb{L}_{\sigma,\varrho}(E,F)\times\mathbb{L}_{\sigma,\varrho}(E,F)\longrightarrow[0,\infty),\,\,(L,H)\mapsto\mathpzc{Lip}(L-H)_{\sigma,\varrho}\,,$ (1.1) on $\mathbb{L}_{\sigma,\varrho}(E,F)$ turns it into an Abelian topological group. We always topologize the space $\mathbb{L}_{\sigma,\varrho}(E,F)$ by the metric (1.1). Let $\varphi:U\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}E\rightarrow F$ be a continuous map. If $\varphi$ is Keller’s differentiable, $\operatorname{D}\varphi(x)\in\mathbb{L}_{\sigma,\varrho}(E,F)$ for all $x\in U$ and the induced map $\operatorname{D}\varphi(x):U\rightarrow\mathbb{L}_{\sigma,\varrho}(E,F)$ is continuous, then $\varphi$ is called bounded differentiable or $MC^{1}$ and we write $\varphi^{(1)}=\varphi^{\prime}$. We define for $k>1$ mappings of class $MC^{k}$, recursively. An $MC^{k}$-Fréchet manifold is a Hausdorff second countable topological space modeled on a Fréchet space with an atlas of coordinate charts such that the coordinate transition functions are all $MC^{k}$-mappings. We define $MC^{k}$-mappings between Fréchet manifolds as usual. Henceforth, we assume that $M$ and $N$ are connected $MC^{k}$-Fréchet manifolds modeled on Fréchet spaces $(F,\varrho)$ and $(E,\sigma)$, respectively. A mapping $\varphi\in\mathbb{L}_{\sigma,\varrho}(E,F)$ is called Lipschitz- Fredholm operator if its kernel has finite dimension and its image has finite co-dimension. The index of $\varphi$ is defined by $\operatorname{Ind}\varphi=\dim\ker\varphi-\operatorname{codim}\operatorname{Img}\varphi.$ We denote by $\mathcal{LF}(E,F)$ the set of all Lipschitz-Fredholm operators, and by $\mathcal{LF}_{l}(E,F)$ the subset of $\mathcal{LF}(E,F)$ consisting of those operators of index $l$. An $MC^{k}$-Lipschitz-Fredholm mapping $\varphi:M\rightarrow N,\,k\geq 1$, is a mapping such that for each $x\in M$ the derivative $\operatorname{D}\varphi(x):T_{x}M\longrightarrow T_{f(x)}N$ is a Lipschitz- Fredholm operator. The index of $\varphi$, denoted by $\operatorname{Ind}{\varphi}$, is defined to be the index of $\operatorname{D}\varphi(x)$ for some $x$ which does not depend on the choice of $x$, see [1, Definition 3.2 ]. Let $\varphi:M\rightarrow N$ $(k\geq 1)$ be an $MC^{k}$-mapping. We denote by $T_{x}\varphi:T_{x}M\rightarrow T_{\varphi(x)}N$ the tangent map of $f$ at $x\in M$ from the tangent space $T_{x}M$ to the tangent space $T_{\varphi(x)}N$. We say that $\varphi$ is an immersion (resp. submersion) provided $T_{x}\varphi$ is injective (resp. surjective) and the range $\operatorname{Img}(T_{x}\varphi)$ (resp. the kernel $\ker(T_{x}\varphi)$) splits in $T_{\varphi(x)}N$ (resp. $T_{x}M$) for any $x\in M$. An injective immersion $f:M\rightarrow N$ which gives an isomorphism onto a submanifold of $N$ is called an embedding. A point $x\in M$ is called a regular point if $\operatorname{D}f(x):T_{x}M\longrightarrow T_{f(x)}N$ is surjective. The corresponding value $f(x)$ is a regular value. Points and values other than regular are called critical points and values, respectively. Let $\varphi:M\to N$ be an $MC^{k}$-mapping, $k\geq 1$. We say that $\varphi$ is transversal to a submanifold $S\subseteq N$ and write $\varphi\pitchfork S$ if either $\varphi^{-1}(S)=\emptyset$, or if for each $x\in\varphi^{-1}(S)$ 1. (1) $(T_{x}\varphi)(T_{x}M)+T_{\varphi(x)}S=T_{\varphi(x)}N$, and 2. (2) $(T_{x}\varphi)^{-1}(T_{\varphi(x)}S)$ splits in $T_{x}M$. In terms of charts, $\varphi\pitchfork S$ when $x\in\varphi^{-1}(S)$ there exist charts $(\phi,\mathcal{U})$ around $x$ and $(\psi,\mathcal{V})$ around $\varphi(x)$ such that $\psi:\mathcal{V}\to\mathcal{V}_{1}\times\mathcal{V}_{2}$ is an $MC^{k}$-isomorphism on a product, with $\psi(\varphi(x))=(0_{E},0_{E})\,\quad\varphi(S\cap\mathcal{V})=\mathcal{V}_{1}\times\left\\{0_{E}\right\\}.$ Then the composite mapping $\mathcal{U}\xrightarrow{\varphi}\mathcal{V}\xrightarrow{\psi}\mathcal{V}_{1}\times\mathcal{V}_{2}\xrightarrow{\mathrm{Pr}_{V_{2}}}\mathcal{V}_{2}.$ is an $MC^{k}$-submersion, where $\mathrm{Pr}_{V_{2}}$ is the projection onto $\mathcal{V}_{2}$. ## 2\. Transversality theorems We generalize [2, Theorem 4.2] and [2, Corollary 4.1] for not necessarily Lipschitz-Fredholm mappings and finite dimensional submanifolds. We shall need the following version of the inverse function theorem for $MC^{k}$-mappings. ###### Theorem 2.1. [4, Theorem 4.7] Let $\mathcal{U}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}E$, $u_{0}\in\mathcal{U}$ and $\varphi:\mathcal{U}\rightarrow E$ an $MC^{k}$-mapping, $k\geq 1$. If $\varphi^{\prime}(u_{0})\in\operatorname{Aut}{(E)}$, then there exists $\mathcal{V}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}\mathcal{U}$ of $u_{0}$ such that $\varphi(\mathcal{V})$ is open in $E$ and $\varphi|_{\mathcal{V}}:\mathcal{V}\to\varphi(\mathcal{V})$ is an $MC^{k}$\- diffeomorphism. ###### Proposition 2.1. Let $\varphi:M\to N$ be an $MC^{k}$-mapping, $S\subset N$ an $MC^{k}$-submanifold and $x\in\varphi^{-1}(S)$. Then $\varphi\pitchfork S$ if and only if there are charts $(\mathcal{U},\phi)$ around $x$ with $\phi(x)=0_{E}$ and $(\mathcal{V},\psi)$ around $y=\varphi(x)$ in $S$ with $\psi(y)=0_{F}$ such that the following hold: 1. (1) There are subspaces $\bf E_{1}$ and $\bf E_{2}$ of $E$, and $\bf F_{1}$ and $\bf F_{2}$ of $F$ such that $E=\bf E_{1}\oplus\bf E_{2}$ and $F=\bf F_{1}\oplus F_{2}$. Moreover, $\psi(S\cap\mathcal{V})=F_{1}$ and $\displaystyle\phi(\mathcal{U})=E_{1}+E_{2}\subseteq{\bf E_{1}}\oplus{\bf E_{2}}$ $\displaystyle\psi(\mathcal{V})=F_{1}+F_{2}\subseteq{\bf F_{1}}\oplus{\bf F_{2}},$ where $0_{E}\in E_{i}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}{\bf E_{i}}$ and $0_{F}\in F_{i}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}{\bf F_{i}}$, $i=1,2$. 2. (2) In the charts the local representative of $\varphi$ has the form $\varphi_{\phi\psi}=\overline{\varphi}+\hat{\varphi}\circ\mathrm{Pr}_{E_{2}},$ (2.1) where $\overline{\varphi}:E_{1}+E_{2}\to F_{1}$ is an $MC^{k}$-mapping, $\hat{\varphi}$ is an $MC^{k}$-isomorphism of $\bf E_{2}$ onto $\bf F_{2}$ and $\mathrm{Pr}_{E_{2}}:E\to{\bf E_{2}}$ is the projection. ###### Proof. Sufficiency: Let $(\mathcal{U},\phi)$ and $(\mathcal{V},\psi)$ be charts that satisfy the assumptions we will prove $\varphi\pitchfork S.$ In the charts, by using the identifications $T_{x}M\simeq E,\,T_{y}\simeq F$, the tangent map $T_{x}\varphi:T_{x}M\to T_{y}N$ has the representation $T_{x}\varphi=\varphi^{\prime}_{\phi\psi}(0_{E}):E\to F.$ Also, we have the identification $T_{y}S\simeq{\bf F_{1}}$. Let $\mathrm{Pr}_{F_{2}}:F\to{\bf F_{2}}$ be the projection onto ${\bf F_{2}}$. Since $\varphi^{\prime}_{\phi\psi}(0_{E})=(\overline{\varphi})^{\prime}+\hat{\varphi}\circ\mathrm{Pr}_{E_{2}}$ and $(\overline{\varphi})^{\prime}(0):E\to{\bf F_{1}}$, it follows that for all $e\in E,\,e=e_{1}+e_{2}\in{\bf E_{1}}\oplus{\bf E_{2}}$ $\varphi^{\prime}_{\phi\psi}(0_{E})e=(\overline{\varphi})^{\prime}(0_{E})e+\hat{\varphi}\circ\mathrm{Pr}_{E_{2}}.$ Thus, $\mathrm{Pr}_{F_{2}}\circ\varphi^{\prime}_{\phi\psi}(0_{E})e=\mathrm{Pr}_{F_{2}}\circ\hat{\varphi}\circ\,\mathrm{Pr}_{F_{2}}(e)$ which means $\mathrm{Pr}_{F_{2}}\circ\varphi^{\prime}_{\phi\psi}(0_{E})=\mathrm{Pr}_{F_{2}}\circ\hat{\varphi}\circ\,\mathrm{Pr}_{F_{2}}$ (2.2) it is a surjective mapping of $E$ onto ${\bf F_{2}}$. Moreover, we have $\displaystyle\ker(\mathrm{Pr}_{F_{2}}\circ\varphi^{\prime}_{\phi\psi}(0_{E}))$ $\displaystyle=\varphi^{\prime}_{\phi\psi}(0_{E})^{-1}(\mathrm{Pr}_{F_{2}}(0_{E}))=\varphi^{\prime}_{\phi\psi}(0_{E})^{-1}(F_{1})$ $\displaystyle=\left\\{e=e_{1}+e_{2}\in{\bf E_{1}}\oplus{\bf E_{2}}\mid(\overline{\varphi})^{\prime}(0_{E})e+\hat{\varphi}(e_{2})\in\mathbf{F}_{1}\right\\}$ $\displaystyle=\left\\{e=e_{1}+e_{2}\in{\bf E_{1}}\oplus{\bf E_{2}}\mid\hat{\varphi}(e_{2})=0_{F}\right\\}$ $\displaystyle=\left\\{e=e_{1}+e_{2}\in{\bf E_{1}}\oplus{\bf E_{2}}\mid e_{2}=0_{E}\right\\}$ $\displaystyle=\mathbf{E}_{1}.$ Which is an $MC^{k}$\- splitting in $E$ with a component $E_{2}$. From (3) it follows that $\mathrm{Pr}_{F_{2}}\circ\varphi^{\prime}_{\phi\psi}(0_{E})=\hat{\varphi}:\mathrm{Pr}_{E_{2}}\to\mathrm{Pr}_{F_{2}}$ which is an $MC^{k}$-isomorphism. Necessity: Suppose $\varphi\pitchfork S$. Since $S$ is an $MC^{k}$-submanifold of $N$ and $y=\varphi(x)\in S$, there is a chart $(\mathcal{W},\mathbbmss{w})$ around $y$ having the submanifold property for $S$ in $N$: $\displaystyle\mathbbmss{w}(\mathcal{W})=W_{1}+W_{2}\subset\mathbf{F_{1}}\oplus\mathbf{F_{2}}=F,$ $\displaystyle\mathbbmss{w}(S\cap\mathcal{W})=W_{1}\subset F,\quad\mathbbmss{w}(y)=0_{E}.$ Also, there is a chart $(\mathcal{X},\mathbbmss{x})$ around $x$ such that $\mathbbmss{x}(x)=0_{F},\,\varphi(\mathcal{X})\subset\mathcal{W}$ and $\varphi_{\mathbbmss{x}\mathbbmss{w}}:\mathbbmss{x}(\mathcal{X})\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}E\to\mathbbmss{w}(\mathcal{W})\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}F$ is of class $MC^{k}$. It follows that $\varphi_{\mathbbmss{x}\mathbbmss{w}}(0_{E})\circ\mathrm{Pr}_{F_{2}}:E\to\mathbf{F_{2}}$ is an $MC^{k}$-submersion as $\varphi\pitchfork S$. That is, $\varphi_{\mathbbmss{x}\mathbbmss{w}}\circ\mathrm{Pr}_{F_{2}}$ and $\mathbf{E_{1}}\coloneq\varphi_{\mathbbmss{x}\mathbbmss{w}}^{\prime}(0_{E})^{-1}(\mathbf{F_{1}})$ splits in $E$ with the complement $\mathbf{E_{2}}$ such that $\eta\coloneq\mathrm{Pr}_{F_{2}}\circ\varphi_{\mathbbmss{x}\mathbbmss{w}}(0_{E})\mid_{\mathbf{E_{2}}}:\mathbf{E_{2}}\to\mathbf{F_{2}}$ is an $MC^{k}$-isomorphism. Set $\tau\coloneq(\mathrm{Pr}_{E_{1}}+\eta^{-1}\circ\mathrm{Pr}_{F_{2}}\circ\varphi_{\mathbbmss{x}\mathbbmss{w}}):\mathbbmss{x}(\mathcal{X})\to\mathbbmss{w}(\mathcal{W})$, then $\tau$ is an $MC^{k}$-mapping and $\tau(0_{E})=0_{E}$ and $\tau^{\prime}(0_{E})=(\mathrm{Pr}_{E_{1}}+\eta^{-1}\circ\mathrm{Pr}_{F_{2}}\circ\varphi_{\mathbbmss{x}\mathbbmss{w}})^{\prime}(0_{E})=\mathrm{Pr}_{E_{1}}+\mathrm{Pr}_{E_{2}}=\mathrm{Id}_{E}$. Because, for all $e=e_{1}+e_{2}\in\mathbf{E_{1}}\oplus\mathbf{E_{2}}$ we have $(\varphi_{\mathbbmss{x}\mathbbmss{w}})^{\prime}(0_{E})e=(\varphi_{\mathbbmss{x}\mathbbmss{w}})^{\prime}(0_{E})e_{1}+(\varphi_{\mathbbmss{x}\mathbbmss{w}})^{\prime}(0_{E})e_{2}$. Whence, $\mathrm{Pr}_{F_{2}}\circ(\varphi_{\mathbbmss{x}\mathbbmss{w}})^{\prime}(0_{E})e=\mathrm{Pr}_{F_{2}}\circ(\varphi_{\mathbbmss{x}\mathbbmss{w}})^{\prime}(0_{E})e_{2}=\tau(e_{2})$, hence, $\tau^{-1}\circ\mathrm{Pr}_{F_{2}}\circ(\varphi_{\mathbbmss{x}\mathbbmss{w}})^{\prime}(0_{E})e=\tau^{-1}(\tau(e_{2}))=\mathrm{Pr}_{E_{2}}.$ By the inverse mapping theorem 2.1, $\tau$ is a local $MC^{k}$-diffeomorphism. Assume $0_{E}\in\mathcal{X}_{1}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}\mathbbmss{x}(\mathcal{X})$ is small enough. Let $\mathbbmss{x}_{1}:\mathcal{X}_{1}\to 0_{E}\in\mathcal{X}_{2}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}E$ be an $MC^{k}$-diffeomorphism such that $\tau\circ\mathbbmss{x}_{1}^{-1}=\mathrm{Id}_{F}.$ (2.3) Thus, $\mathrm{Pr}_{F_{2}}\circ\tau\circ\varphi_{\mathbbmss{x}\mathbbmss{w}}=\eta\circ\mathrm{Pr}_{E_{2}}.$ (2.4) If, $e=e_{1}+e_{2}\in\mathbbmss{x}_{1}(\mathcal{X}_{1})$ and $\mathbbmss{x}_{1}^{-1}(e)=\bar{e_{1}}+\bar{e_{2}}$, then by (2.3) and (2.4) we obtain $\displaystyle\tau\circ\mathbbmss{x}_{1}^{-1}(e)$ $\displaystyle=\tau(\bar{e_{1}}+\bar{e_{2}})$ $\displaystyle=\bar{e_{1}}+\eta^{-1}\circ\mathrm{Pr}_{F_{2}}\circ\varphi_{\mathbbmss{x}\mathbbmss{w}}(\bar{e_{1}}+\bar{e_{2}})$ $\displaystyle=e_{1}+e_{2}.$ Therefore, $\bar{e_{1}}=e_{1}$ and $\tau^{-1}\circ\mathrm{Pr}_{F_{2}}\circ\varphi_{\mathbbmss{x}\mathbbmss{w}}=e_{2}$ and so $\mathrm{Pr}_{F_{2}}\circ\varphi_{\mathbbmss{x}\mathbbmss{w}}(\bar{e_{1}}+\bar{e_{2}})=\tau(e_{2})=\tau\circ\mathrm{Pr}_{E_{2}}(e_{1}+e_{2}).$ This means, $\mathrm{Pr}_{E_{2}}\circ\varphi_{\mathbbmss{x}\mathbbmss{w}}\circ\mathbbmss{x}_{1}^{-1}(e)=\eta\circ\mathrm{Pr}_{E_{2}}(e)$ for all $e\in\mathbbmss{x}_{1}(\mathcal{X}_{1})$. Now, define $\displaystyle\phi\coloneq\mathbbmss{x}_{1}\circ\mathbbmss{x},\quad\mathcal{U}\coloneq\mathbbmss{x}^{-1}(\mathcal{X}),$ $\displaystyle\psi\coloneq\mathbbmss{w},\quad\mathcal{V}\coloneq\mathrm{small\,enough\,neighborhood\,of}\,y\,\mathrm{in}\mathcal{W}.$ Then, $(\phi,\mathcal{U})$ and $(\psi,\mathcal{V})$ are the desired charts. Indeed, $\displaystyle\varphi_{\phi\psi}$ $\displaystyle=\mathbbmss{w}\circ\varphi\circ(\mathbbmss{x}_{1}\circ\mathbbmss{x})$ $\displaystyle=\mathbbmss{w}\circ\varphi\circ\mathbbmss{x}^{-1}\circ\mathbbmss{x}^{-1}=\varphi_{\mathbbmss{x}\mathbbmss{w}}\circ\mathbbmss{x}_{1}^{-1}.$ Thus, $\displaystyle\varphi_{\phi\psi}$ $\displaystyle=\mathrm{Pr}_{F_{1}}\circ\varphi_{\mathbbmss{x}\mathbbmss{w}}\circ\mathbbmss{x}_{1}^{-1}+\mathrm{Pr}_{F_{2}}\circ\varphi_{\mathbbmss{x}\mathbbmss{w}}\circ\mathbbmss{x}_{1}^{-1}$ $\displaystyle=\overline{\varphi}+\hat{\varphi}\circ\mathrm{Pr}_{E_{2}},$ if we set $\hat{\varphi}\coloneq\eta$ and $\overline{\varphi}\coloneq\mathrm{Pr}_{E_{1}}\circ\varphi_{\mathbbmss{x}\mathbbmss{w}}\circ\mathbbmss{x}_{1}^{-1}$. ∎ ###### Theorem 2.2 (Transversality Theorem). Let $\varphi:M\to N$ be an $MC^{k}$-mapping, $k\geq 1$, $S\subset N$ an $MC^{k}$-submanifold and $\varphi\pitchfork S$. Then, $\varphi^{-1}(S)$ is either empty of $MC^{k}$-submanifold of $M$ with $(T_{x}\varphi)^{-1}(T_{y}S)=T_{x}(\varphi^{-1}(S)),\,x\in\varphi^{-1}(S),\,y=\varphi(x).$ If $S$ has finite co-dimension in $N$, then $\operatorname{codim}(\varphi^{-1}(S))=\operatorname{codim}S$. Moreover, if $\dim S=m<\infty$ and $\varphi$ is an $MC^{k}$-Lipschitz-Fredholm mapping of index $l$, then $\dim\varphi^{-1}(S)=l+m$. ###### Proof. Let $x\in\varphi^{-1}(S)$, then by Proposition 2.1 there are chart $(\phi,\mathcal{U})$ around $x$ and $(\psi,\mathcal{V})$ around $y=\varphi(x)$ such that $\displaystyle\phi(\mathcal{U})=E_{1}+E_{2}\subseteq{\bf E_{1}}\oplus{\bf E_{2}},$ $\displaystyle\psi(\mathcal{V})=F_{1}+F_{2}\subseteq{\bf F_{1}}\oplus{\bf F_{2}},$ $\displaystyle\varphi_{\phi\psi}=\overline{\varphi}+\hat{\varphi}\circ\mathrm{Pr}_{E_{2}},$ (2.5) where $\overline{\varphi}:E_{1}+E_{2}\to F_{1}$ is an $MC^{k}$-mapping, $\hat{\varphi}$ is an $MC^{k}$-isomorphism of $\bf E_{2}$ onto $\bf F_{2}$ and $\mathrm{Pr}_{E_{2}}:E\to{\bf E_{2}}$ is the projection. Let $\hat{e}\in\varphi^{-1}(S)\cap\mathcal{U}$, then $\hat{f}=\varphi(\hat{e})\in S\cap\mathcal{V}$ and $\psi(\varphi(\hat{e}))\in F_{1}\subset\mathbf{F_{1}}$. By (2), if $\phi(\hat{e})=e_{1}+e_{2}\in E_{1}+E_{2}$ we have $\displaystyle\varphi_{\phi\psi}(\phi(\hat{e}))$ $\displaystyle=\varphi_{\phi\psi}(e_{1}+e_{2})$ $\displaystyle=\overline{\varphi}(e_{1}+e_{2})+\hat{\varphi}(e_{2})\in F_{1}\subset\mathbf{F_{1}}.$ It follows $\hat{\varphi}(e_{2})=0_{E},e_{2}=0_{E},$ since $\hat{\varphi}(e_{2})\in\mathbf{F_{2}}$ and $\mathbf{F_{1}}\cap\mathbf{F_{2}}=\left\\{0_{F}\right\\}$. Thus, $\phi(\hat{e})\in F_{1}$ for all $\hat{e}\in\varphi^{-1}(S)\cap\mathcal{U}$. Therefore, $E_{1}\subset\phi(\varphi^{-1}(S)\cap\mathcal{U})$, since for each $e_{1}\in E_{1}$ we have $\varphi_{\phi\psi}(e_{1})=\overline{\varphi}(e_{1})+\hat{\varphi}\circ\mathrm{Pr}_{E_{2}}(e_{1})=\overline{\varphi}(e_{1})\in F_{1}.$ Hence, $\psi\circ\varphi\circ\phi^{-1}(e_{1})\in F_{1}$ implies that $\varphi\circ\phi^{-1}(e_{1})\in\psi^{-1}(F_{1})=S\cap\mathcal{V}$ and so $\varphi\circ\phi^{-1}(e_{1})$ which means $\phi^{-1}(e_{1})\in\varphi(S)\cap\mathcal{V}$ that yields $e_{1}\in\psi(\varphi^{-1}(S)\cap\mathcal{V})$. Therefore, for $x\in\varphi^{-1}(S)$ there is a chart $(\phi,\mathcal{U})$ with $\phi(\mathcal{U})=E_{1}+E_{2}\subset\mathbf{E_{1}}\oplus\mathbf{E_{2}}$ and $\phi(x)=0_{E},\,\phi(\varphi^{-1}(S)\cap\mathcal{V})=E_{1},$ which means $\varphi^{-1}(S)$ is an $MC^{k}$-submanifold in $M$. In the charts, we have $T_{x}\simeq E,\,T_{y}N\simeq F,\,T_{x}(\varphi^{-1}(S))\simeq\mathbf{E_{1}}$ and $T_{y}S\simeq\mathbf{F_{1}}$. From the proof of Proposition 2.1 we have $\varphi_{\phi\psi}^{\prime}(0_{E})^{-1}(\mathbf{F_{1}})=\mathbf{E_{1}}$ which yields $(T_{x}\varphi)^{-1}(T_{y}S)=T_{x}(\varphi^{-1}(S))$. If $S$ has finite co-dimension then $\mathbf{F_{2}}$ has finite dimension and thus by Proposition 2.1, $\operatorname{codim}(\varphi^{-1}(S))=\operatorname{codim}\varphi^{-1}(S\cap\mathcal{V})=\dim(\mathbf{F_{2}})=\operatorname{codim}(S).$ The proof of the last statement is standard. ∎ As an immediate consequence we have: ###### Corollary 2.1. Let $\varphi:M\rightarrow N$ be an $MC^{k}$-mapping, $k\geq 1$. If $q$ is a regular value of $\varphi$, then the level set $\varphi^{-1}(q)$ is a submanifold of $M$ and its tangent space at $p=\varphi(q)$ is $\ker T_{p}\varphi$. Moreover, if $q$ is a regular value of $\varphi$ and $\varphi$ is an $MC^{k}$-Lipschitz-Fredholm mapping of index $l$, then $\dim\varphi^{-1}(S)=l$. To prove the parametric transversality theorem we apply the following Sard’s theorem. ###### Theorem 2.3. [2, Theorem 3.2] If $\varphi:M\rightarrow N$ is an $MC^{k}$-Lipschitz-Fredholm map with $k>\max\\{{\operatorname{Ind}\varphi,0}\\}$. Then, the set of regular values of $\varphi$ is residual in $N$. ###### Theorem 2.4 (The Parametric Transversality Theorem). Let $A$ be a manifold of dimension $n$, $S\subset N$ a submanifold of finite co-dimension $m$. Let $\varphi:M\times A\to N$ be an $MC^{k}$-mapping, $k\geq\left\\{1,n-m\right\\}$. If $\varphi\pitchfork S$, then the set of all points $x\in M$ such that the mappings $\varphi_{x}:A\to N,\,(\varphi_{x}(\cdot)\coloneq\varphi(x,\cdot))$ are transversal to $S$, is residual $M$. ###### Proof. Let $\mathbf{S}=\varphi^{-1}(S)$, $\mathrm{Pr}_{M}:M\times A\to M$ the projection onto $M$ and $\mathrm{Pr}_{\mathbf{S}}$ be its restriction to $\mathbf{S}$. First, we prove that $\mathrm{Pr}_{\mathbf{S}}$ is an $MC^{k}$-Fredholm-Lipschitz mapping of index $n-m$, i.e., $T_{(m,a)}\mathrm{Pr}_{\mathbf{S}}:T_{(m,a)}\mathbf{S}\to T_{m}M$ is a Lipschitz-Fredholm operator of index $n-m$. By Theorem 2.2 the inverse image $\mathbf{S}$ is an $MC^{k}$-submanifold of $M\times A$, with model space $\mathbb{S}$, so that $\mathrm{Pr}_{\mathbf{S}}$ is an $MC^{k}$-mapping. Let $\pi_{M}$ and $\pi_{\mathbf{S}}$ be the local representatives of $\mathrm{Pr}_{M}$ and $\mathrm{Pr}_{\mathbf{S}}$, respectively. We show that $\pi_{M}$ and consequently $\pi_{\mathbf{S}}$ are Lipschitz-Fredholm operators of index $n-m$. Finite dimensionality of $\mathbb{R}^{n}$ and closedness of $\mathbb{S}$ implies that $K\coloneq\mathbb{S}+(\\{0\\}\times\mathbb{R}^{n})$ is closed in $E\times\mathbb{R}^{n}$. Also, $\operatorname{codim}K$ is finite because it contains the finite co-dimensional subspace $\mathbb{S}$. Therefore $K$ has a finite-dimensional complement $K_{1}\subset E\times\left\\{0\right\\}$, that is $E\times\mathbb{R}^{n}=K\oplus K_{1}$. Let $K_{2}\coloneq\mathbb{S}\cap\left\\{0\right\\}\times\mathbb{R}^{n}$. Since $K_{2}\subset\mathbb{R}^{n}$ we can choose closed subspaces $\mathbb{S}_{1}\subset\mathbb{R}^{n}$ and $\mathbb{R}_{0}\subset\left\\{0\right\\}\times\mathbb{R}^{n}$ such that $\mathbb{S}=\mathbb{S}_{1}\oplus K_{1}$ and $\left\\{0\right\\}\times\mathbb{R}^{n}=K_{1}\oplus\mathbb{R}_{0}$. Whence, $K=\mathbb{S}_{1}\oplus K_{1}\oplus\mathbb{R}_{0}$ and $E\times\mathbb{R}^{n}=\mathbb{S}_{1}\oplus K_{1}\oplus\mathbb{R}_{0}\oplus K_{2}$. The mapping $\pi_{\mathbf{S}}\mid_{\mathbb{S}_{1}\oplus K_{2}}:\mathbb{S}_{1}\oplus K_{2}\to E$ is an isomorphism, $K_{1}=\ker\pi_{\mathbf{S}}$, and $\pi_{M}(K_{2})$ is a finite dimensional complement to $\pi_{M}(\mathbb{S})$ in $\mathbb{R}^{n}$. Thus, $\pi_{M}$ is a Lipschitz-Fredholm operator and we have $\displaystyle\operatorname{Ind}\pi_{M}$ $\displaystyle=\dim K_{1}-\dim K_{2}$ $\displaystyle=\dim(K_{1}\oplus\mathbb{R}_{0})-\dim(\mathbb{R}_{0}\oplus K_{2}).$ Since, $K_{1}\oplus\mathbb{R}_{0}=\left\\{0\right\\}\times\mathbb{R}^{n}$ and $\mathbb{R}_{0}\oplus K_{2}$ is a complement to $\mathbb{S}$ in $E\times\mathbb{R}^{n}$ and therefore its dimension is $n$, so the index of $\pi_{M}$ is $n-m$. Now, we prove that if $x$ is a regular value of $\mathrm{Pr}_{\mathbf{S}}$ if and only if $\varphi_{x}\pitchfork S$. From the definition of $\varphi\pitchfork$ we have $\forall(x,a)\in\mathbf{S}$ $(T_{(x,a)}\varphi)(T_{x}M\times T_{a}A)+T_{\varphi(x,a)}S=T_{\varphi(x,a)}N,$ (2.6) and $(T_{(x,a)}\varphi)^{-1}(T_{\varphi_{(x,a)}}S)\,\mathrm{splits\,in}\,T_{x}M\times T_{a}A.$ (2.7) Since $A$ has finite dimension, it follows that the mapping $a\in A\mapsto\varphi{(x,a)}$ for a fixed $x\in M$ is transversal to $S$ if and only if $\forall(x,a)\in\mathbf{S},T_{a}\varphi_{x}(T_{a}A)+T_{\varphi(x,a)}S=T_{\varphi(x,a)}S.$ (2.8) Since $\mathrm{Pr}_{\mathbf{S}}$ is a Lipschitz-Fredholm mapping, $\ker T\mathrm{Pr}_{\mathbf{S}}$ splits at any point as its dimension is finite. Then $x$ is a regular value of $\mathrm{Pr}_{\mathbf{S}}$ if and only if $\forall(x,a)\in\mathbf{S},\forall v\in T_{x}M,\exists u\in T_{a}A\colon T_{(v,u)}\varphi(v,u)\in T_{(x,a)}S.$ (2.9) Pick $x\in M$ and $a\in A$ such that $(x,a)\in\mathbf{S}$ and let $w\in T_{(x,a)}S$. By (2.6) and (2.7) we obtain that there exist $v\in T_{a}A,\,x_{1}\in T_{x}M,\,y_{1}\in T_{(x,a)}S$ such that $T_{(x,a)}\varphi(v,x_{1})+y_{1}=w.$ (2.10) Then, there exists $x_{2}\in T_{x}M$ such that $T_{(x,a)}\varphi(v,x_{2})\in T_{\varphi_{(x,a)}}S$. Hence, $\displaystyle w$ $\displaystyle=T_{(x,a)}\varphi(v,x_{1})-T_{(x,a)}\varphi(v,x_{2})+T_{(x,a)}\varphi(v,x_{2})+y_{1}$ $\displaystyle=T_{(x,a)}\varphi(0,x_{1}-x_{2})+T_{(x,a)}\varphi(v,x_{2})+y_{1}$ $\displaystyle=T_{(x,a)}\varphi(0,u)+T_{\varphi(x,a)}S+y_{2}\in T_{a}\varphi_{x}(T_{a}A),$ where $u=x_{1}-x_{2}$ and $y_{2}=T_{(x,a)}\varphi(v,x_{2})+y_{1}\in T_{\varphi_{(x,a)}}S$. Thus, (2.8) holds. Now we show that (2.8) implies (2.9). Pick $a\in A,\,x\in M$ such that $(x,a)\in\mathbf{S}$. Let $v\in T_{x}M,\,a_{1}\in T_{a}A,\,y_{1}\in T_{\varphi_{(x,a)}}S$ and set $w\coloneq T_{(x,a)}\varphi_{(v,x_{1})}+y_{1}$. By (2.8) there exist $a_{2}\in T_{a}A$ and $y_{2}\in T_{\varphi_{(x,a)}}S$ such that $w=T_{a}\varphi_{x}(a_{2})+y_{2}$. Then, $0_{E}=T_{(x,a)}\varphi(v,a_{1})-T_{a}\varphi_{x}(a_{2})+y_{1}-y_{2}=T_{(x,a)}\varphi(v,a_{1}-a_{2})+y_{1}-y_{2},$ so $T_{(x,a)}\varphi(v,a_{1}-a_{2})=y_{2}-y_{1}\in T_{\varphi_{(x,a)}}S$ so (2.9) holds. Thus, we showed that if $x$ is a regular value of $\mathrm{Pr}_{\mathbf{S}}$ if and only if $\varphi_{x}\pitchfork S$. Since $\mathrm{Pr}_{\mathbf{S}}:\mathbf{S}\to M$ is a Lipschitz-Fredholm of class $MC^{k}$with the index $n-m$ and $\operatorname{codim}\mathbf{S}=\operatorname{codim}S=m$ and $k>\left\\{0,n-m\right\\}$, the Sard’s theorem 2.3 concludes the theorem. ∎ ## 3\. The degree of Lipschitz-Fredholm mappings In this section we construct the degree of $MC^{k}$-Lipschitz-Fredholm mappings and apply it to prove an invariance of domain theorem, a rank theorem and a Bursuk-Ulam type theorem. The construction of the degree relies on the following transversality result. ###### Theorem 3.1. [2, Theorem 3.3] Let $\varphi:M\to N$ be an $MC^{k}$-Lipschitz-Fredholm mapping, $k\geq 1$. Let $\imath:\mathcal{A}\to N$ be an $MC^{1}$-embedding of a finite dimension manifold $\mathcal{A}$ with $k>\max\\{\operatorname{Ind}\varphi+\dim\mathcal{A},0\\}$. Then there exists an $MC^{1}$ fine approximation $\mathbf{g}$ of $\imath$ such that $\mathbf{g}$ is embedding and $\varphi\pitchfork\mathbf{g}$. Moreover, suppose $S$ is a closed subset of $\mathcal{A}$ and $\varphi\pitchfork\imath(S)$, then $\mathbf{g}$ can be chosen so that $\imath=\mathbf{g}$ on $S$. We shall need the following theorem that gives the connection between proper and closed mappings. ###### Theorem 3.2. [5, Theorem 1.1] Let $A,B$ be Hausdorff manifolds, where $A$ is a connected infinite dimensional Fréchet manifold, and $B$ satisfies the first countability axiom, and let $\varphi:A\to B$ be a continuous closed non- constant map. Then $\varphi$ is proper. Let $\varphi:M\to N$ be a non-constant closed Lipschitz-Fredholm mapping with index $l\geq 0$ of class $MC^{k}$such that $k>l+1$. If $q$ is a regular value of $\varphi$, then by Theorem 3.2 and Corollary 2.1 the preimage $\varphi^{-1}(q)$ is a compact submanifold of dimension $l.$ Let $\imath:[0,1]\hookrightarrow N$ be an $MC^{1}$-embedding that connects two distinct regular values $q_{1}$ and $q_{2}$. By Theorem 3.1 we may suppose $\imath$ is transversal to $\varphi$. Thus, by Theorem 2.2 the preimage ${\bf M}\coloneq\varphi^{-1}(\imath([0,1]))$ is a compact $(l+1)$-dimensional submanifold of $M$ such that its boundary, $\partial{\bf M}$, is the disjoint union of $\varphi^{-1}(q_{1})$ and $\varphi^{-1}(q_{2})$, $\partial{\bf M}=\varphi^{-1}(q_{1})\amalg\varphi^{-1}(q_{2})$. Therefore, $\varphi^{-1}(q_{1})$ and $\varphi^{-1}(q_{2})$ are non-oriented cobordant which gives the invariance of the mapping. Following Smale [6] we associate to $\varphi$ a degree, denoted by $\deg\varphi$, defined as the non-oriented cobordism class of $\varphi^{-1}(q)$ for some regular value $q$. If $l=0$, then $\deg\varphi\in\mathbb{Z}_{2}$ is the number modulo 2 of preimage of a regular value. Let $\mathcal{O}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}M$. Suppose $\varphi:\overline{\mathcal{O}}\to N$ is a non- constant closed continuous mapping such that its restriction to $\mathcal{O}$ is an $MC^{k+1}$-Lipschitz-Fredholm mapping of index $k$, $k\geq 0$. Let $p\in N\setminus\varphi(\partial\overline{\mathcal{O}})$ and let $\mathbf{p}$ a regular value of $\varphi$ in the connected component of $N\setminus\varphi(\partial\overline{\mathcal{O}})$ containing $p$, the existence of such regular value follows from Sard’s theorem 2.3. Again, we associate to $\varphi$ a degree, $\deg(\varphi,p)$, defined as non-oriented class of $k$-dimensional compact manifold $\varphi^{-1}(\mathbf{p})$. This degree does not depend on the choice of $\mathbf{p}$. The following theorem which presents the local representation of $MC^{k}$-mappings is crucial for the rest of the paper. ###### Theorem 3.3. [1, Theorem 4.2] Let $\varphi:\mathcal{U}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}E\rightarrow F$ be an $MC^{k}$-mapping, $k\geq 1$, $u_{0}\in\mathcal{U}$. Suppose that $\operatorname{D}\varphi(u_{0})$ has closed split image $\mathbf{F_{1}}$ with closed topological complement $\mathbf{F_{2}}$ and split kernel $\mathbf{E_{2}}$ with closed topological complement $\mathbf{E_{1}}$. Then, there are two open sets $\mathcal{U}_{1}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}\mathcal{U}$ and $\mathcal{V}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}\mathbf{F_{1}}\oplus\mathbf{E_{2}}$ and an $MC^{k}$-diffeomorphism $\Psi:\mathcal{V}\rightarrow\mathcal{U}_{1}$, such that $(\varphi\circ\Psi)(f,e)=(f,\eta(f,e))$ for all $(f,e)\in\mathcal{V}$, where $\eta:\mathcal{V}\to\mathbf{E_{2}}$ is an $MC^{k}$\- mapping. ###### Theorem 3.4 (Rank theorem for $MC^{k}$-mappings). Let $\varphi:\mathcal{U}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}E\to F$ be an $MC^{k}$-mapping, $k\geq 1$. Suppose $u_{0}\in\mathcal{U}$ and $\operatorname{D}\varphi(u_{0})$ has closed split image $\mathbf{F_{1}}$ with closed complement $\mathbf{F_{2}}$ and split kernel $\mathbf{E_{2}}$ with closed complement $\mathbf{E_{1}}$. Also, assume $\operatorname{D}\varphi(\mathcal{U})(E)$ is closed in F and $\operatorname{D}\varphi(u)|_{\mathbf{E_{1}}}:\mathbf{E_{1}}\to\operatorname{D}\varphi(u)(E)$ is an $MC^{k}$-isomorphism for each $u\in\mathcal{U}$. Then, there exist open sets $\mathcal{U}_{1}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}\mathbf{F_{1}}\oplus\mathbf{E_{2}},\,\mathcal{U}_{2}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}E,\,\mathcal{V}_{1}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}F$, and $\mathcal{V}_{2}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}F$ and there are $MC^{k}$-diffeomorphisms $\phi:\mathcal{V}_{1}\to\mathcal{V}_{2}$ and $\psi:\mathcal{U}_{1}\to\mathcal{U}_{2}$ such that $(\phi\circ\varphi\circ\psi)(f,e)=(f,0),\quad\forall(f,e)\in\mathcal{U}_{1}.$ ###### Proof. By Theorem 3.3 there exits an $MC^{k}$-diffeomorphism $\psi:\mathcal{U}_{1}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}\mathbf{F_{1}}\oplus\mathbf{E_{2}}\to\mathcal{U}_{2}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}E$ such that $\varphi(f,e)=(\varphi\circ\psi)(f,e)=(f,\eta(f,e)),$ where $\eta:\mathcal{V}\to\mathbf{E_{2}}$ is an $MC^{k}$\- mapping. Let $\mathrm{Pr}_{1}:F\to\mathbf{F_{1}}$ be the projection. We obtain $\mathrm{Pr}_{1}\circ\operatorname{D}\varphi(f,e)(w,v)=(w,0)$, for $w\in\mathbf{F_{1}}$ and $v\in\mathbf{E_{2}}$ because $\operatorname{D}\varphi(f,e)(w,v)=(w,\operatorname{D}\eta(f,e)(w,v)).$ Hence, $\mathrm{Pr}_{1}\circ\operatorname{D}\varphi(f,e)|_{\mathbf{F_{1}}\times\left\\{0\right\\}}$ is the identity mapping, $\mathrm{Id}_{\mathbf{F_{1}}}$, on $\mathbf{F_{1}}$. Thereby, $\operatorname{D}\varphi(f,e)|_{\mathbf{F_{1}}\times\left\\{0\right\\}}:\mathbf{F_{1}}\times\left\\{0\right\\}\to\operatorname{D}\varphi(f,e)(\mathbf{F_{1}}\oplus\mathbf{E_{2}})$ is one-to-one and therefore by our assumption $\operatorname{D}\varphi(f,e)\circ\mathrm{Pr}_{1}|_{\operatorname{D}\varphi(f,e)(\mathbf{F_{1}}\oplus\mathbf{E_{2}})}$ is the identity mapping. Suppose $(w,\operatorname{D}\eta(f,e)(w,v))\in\operatorname{D}\varphi(f,e)(\mathbf{F_{1}}\oplus\mathbf{E_{2}})$, we obtain $\operatorname{D}\eta(f,e)v=0$ for all $v\in\mathbf{E_{2}}$, which means $\operatorname{D}_{2}\eta(f,e)=0$, since $\displaystyle(\operatorname{D}\varphi(f,e)\circ\mathrm{Pr}_{1})(w,\operatorname{D}\eta(f,e)(w,v))=$ $\displaystyle\operatorname{D}\varphi(f,e)(w,0)$ $\displaystyle=$ $\displaystyle(w,\operatorname{D}\eta(f,e)(w,0))$ $\displaystyle=$ $\displaystyle(w,\operatorname{D}_{1}(f,e)w).$ We have $\operatorname{D}^{2}\varphi(f,e)v=(0,\operatorname{D}_{2}\eta(f,e)v)$, i.e., $\operatorname{D}^{2}\varphi(f,e)=0$ which means $\varphi$ does not depend on the variable $y\in\mathbf{E_{2}}$. Let $\mathrm{Pr}_{2}:\mathbf{F_{1}}\oplus\mathbf{E_{2}}\to\mathbf{F_{1}}$ be the projection and $\varphi_{f}\coloneq\varphi(f,e)=(\varphi\circ\mathrm{Pr}_{2})(f,e),$ so that $\varphi_{f}:\mathrm{Pr}_{2}(\mathcal{U}_{1})\subset\mathbf{F_{1}}\to F$. Let $\mathbf{v_{0}}\coloneq(\mathrm{Pr}_{2}\circ\psi^{-1})(u_{0})$ and $\mathcal{V}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}\mathcal{U}$ be an open neighborhood of $\mathbf{v_{0}}$. Define the mapping $\begin{array}[]{cccc}\Phi:\mathcal{V}\times\mathbf{F_{2}}\to\mathbf{F_{1}}\oplus\mathbf{F_{2}}\\\ \Phi(f,e)=\varphi(f)+(0,e).\end{array}$ By the open mapping theorem $\operatorname{D}\Phi(\mathbf{v_{0}},0)=(\operatorname{D}\varphi(\mathbf{v_{0}}),\mathrm{Id}_{\mathbf{F_{2}}}):E\oplus\mathbf{F_{2}}\to F$ is a linear $MC^{k}$-isomorphism, where $\mathrm{Id}_{\mathbf{F_{2}}}$ is the identity mapping of $\mathbf{F_{2}}$. Now $\Phi$ satisfies the inverse function theorem 2.1 at $\mathbf{v_{0}}$, therefore, there exist $\mathcal{V}_{1},\mathcal{V}_{2}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}F$ such that $(\mathbf{v_{0}},0)\in\mathcal{V}_{2}$ and $\Phi(\mathbf{v_{0}},0)=\varphi_{\mathbf{v_{0}}}(\mathbf{v_{0}})\in\mathcal{V}_{1}$ and an $MC^{k}$-diffeomorphism $\phi:\mathcal{V}_{1}\to\mathcal{V}_{2}$ such that $\phi^{-1}=\Phi|_{\mathcal{V}_{1}}$. Thus, for $(f,0)\in\mathcal{V}_{2}$ we have $(\phi\circ\varphi)(f)=(\varphi\circ\Phi)(f,0)=(f,0),$ and therefore, $(\phi\circ\varphi\circ\psi)(f,e)=(f,0),\quad\forall(f,e)\in\mathcal{U}_{1}.$ ∎ As an immediate consequence we have the following: ###### Corollary 3.1 (Rank theorem for Lipschitz-Fredholm mappings). Let $\varphi:M\to N$ be an $MC^{\infty}$-Lipschitz-Fredholm mapping of index $k$ and $\dim\ker\operatorname{D}\varphi(x)=m$, $\forall x\in M$. Let $\complement_{1},\complement_{2}$ be topological complements of $\mathbb{R}^{m}$ in $E$ and $\mathbb{R}^{m-k}$ in $F$, respectively. Then, there exist charts $\phi:\mathcal{U}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}M\to E=\mathbb{R}^{m}\oplus\complement_{1}$ with $\phi(x)=0_{E}$ and $\psi:\mathcal{V}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}N\to F=\mathbb{R}^{m-k}\oplus\complement_{2}$ with $\psi(x)=0_{F}$ such that $\psi\circ\varphi\circ\phi^{-1}(f,0)=(f,0).$ The following theorem gives the openness property of the set of Lipschitz- Fredholm mappings. ###### Theorem 3.5. [1, Theorem 3.2] The set $\mathcal{LF}(E,F)$ is open in $\mathcal{L}_{d,g}(E,F)$ with respect to the topology defined by the metric (1.1). Furthermore, the function $T\rightarrow\operatorname{Ind}T$ is continuous on $\mathcal{LF}(E,F)$, hence constant on connected components of $\mathcal{LF}(E,F)$. The proof of the following theorem is a minor modification of [7, Theorem 2]. ###### Theorem 3.6. Let $\varphi:M\to N$ be a Lipschitz-Fredholm mapping of class $MC^{k}$, $k\geq 1$. Then, the set $\mathsf{Sing}(\varphi)\coloneq\left\\{m\mid\operatorname{D}\varphi(m)\mathrm{is\,not\,injective}\right\\}$ is nowhere dense in $M$. ###### Proof. This is a local problem so we assume $M$ is an open set in $E$ and $N$ is an open set in $F$. Let $s\in\mathsf{Sing}(\varphi)$ be arbitrary and $\mathcal{U}$ an open neighborhood of $s$ in $\mathsf{Sing}(\varphi)$. For each $n\in\mathbb{N}\cup\left\\{0\right\\}$ define $S_{n}\coloneq\left\\{m\in M\mid\dim\operatorname{D}\varphi(m)\geq n\right\\}.$ Then, $M=M_{0}\supset M_{1}\supset\cdots,$ therefore, is a unique $n_{0}$ such that $M=M_{n_{0}}\neq M_{n_{0}+1}$. Let $m_{0}\in M_{n_{0}}\setminus M_{n_{0}+1}$ such that $\dim\ker\operatorname{D}\varphi(m_{0})=n_{0}$. By Theorem 3.5, there exists an open neighborhood $\mathcal{V}$ of $m_{0}$ in $\mathcal{U}$ such that for all $v\in\mathcal{V}$ we have $\dim\ker\operatorname{D}\varphi(v)\leq n_{0}$ and hence $\dim\operatorname{D}\varphi(v)=n_{0}\geq 1$. By Corollary 3.1, there is a local representative $\varphi$ around zero such that $\psi\circ\varphi\circ\phi^{-1}(f,e)=(f,0)$ for $(f,e)\in\complement_{1}\oplus\mathbb{R}^{n_{0}}$ which contradicts the injectivity of $\varphi$, therefore, $\mathsf{Sing}(\varphi)$ contains a nonempty open set. The closedness of $\mathsf{Sing}(\varphi)$ is obvious in virtue of Theorem 3.5. ∎ ###### Theorem 3.7 (Invariance of domain for Lipschitz-Fredholm mappings). Let $\varphi:M\to N$ be an $MC^{k}$-Lipschitz-Fredholm mapping of index zero, $k>1$. If $\varphi$ is locally injective, then $\varphi$ is open. ###### Proof. Let $p\in U\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}M$ and $q=\varphi(p)$. The point $p$ has a connected open neighborhood $\,\mathcal{U}\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}M$ such that $\varphi\mid_{\overline{\mathcal{U}}}:\overline{\mathcal{U}}\to N$ is proper and injective. Whence $q\notin\varphi(\partial\mathcal{U})$ and $\varphi(\partial\mathcal{U})$ is closed in $N$. Let $\mathcal{V}$ be a connected component of $N\setminus\varphi(\partial\mathcal{U})$ containing $q$ which is its open neighborhood. Since $\mathcal{U}$ is connected it implies that $\varphi(\mathcal{U})\subset\mathcal{V}$. It follows from $\varphi(\partial\mathcal{U})\cap\mathcal{V}=\emptyset$ that $\overline{\mathcal{U}}\cap\varphi^{-1}(\mathcal{V})=\mathcal{U}$ and so $\varphi\mid_{\mathcal{U}}:\mathcal{U}\to N$ is proper and injective. By Theorem 3.6 there is a point $x\in M$ such that the tangent map $T_{x}\varphi$ is injective and since $\operatorname{Ind}\varphi=0$ it is surjective too. Therefor, $y=\varphi(x)$ is a regular value with $\varphi^{-1}(y)=\left\\{x\right\\}$ and so $\deg\varphi=1$. It follows that $\varphi$ is surjective, because if it is not , then any point in $N\setminus\varphi(M)$ is regular and $\deg\varphi=0$ which is contradiction. Then, $\mathcal{V}=\varphi(\mathcal{U})$ is the open neighborhood of $q$. ∎ ###### Corollary 3.2 (Nonlinear Fredholm alternative). Let $\varphi:M\to N$ be an $MC^{k}$-Lipschitz-Fredholm mapping of index zero, $k>1$. If $N$ is connected and $\varphi$ is locally injective, then $\varphi$ is surjective and finite covering mapping. If $M$ is connected and $N$ is simply connected, then $\varphi$ is a homeomorphism. The following theorem is a generalization of the Bursuk-Ulam theorem, the proof is slight modification of the Banach case. ###### Theorem 3.8. Let $\varphi:\overline{\mathcal{U}}\to F$ be a non-constant closed Lipschitz- Fredholom mapping of class $MC^{2}$ with index zero, where $U\mathrel{\mathchoice{\ooalign{$\displaystyle\subseteq$\cr\raise 1.67915pt\hbox{$\displaystyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\textstyle\subseteq$\cr\raise 1.67915pt\hbox{$\textstyle{\scriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptstyle\subseteq$\cr\raise 1.18399pt\hbox{$\scriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}{\ooalign{$\scriptscriptstyle\subseteq$\cr\raise 0.6458pt\hbox{$\scriptscriptstyle{\scriptscriptstyle\circ}\mkern 2.0mu$}\cr}}}F$ is symmetric. If $\varphi$ is odd and for $u_{0}\in\overline{U}$ we have $u_{0}\notin\varphi(\partial\overline{\mathcal{U}})$. Then $\deg(\varphi,u_{0})\equiv 1\mod 2.$ ###### Proof. Since $\operatorname{D}\varphi(u_{0})$ is a Lipschitz-Fredholm mapping with index zero $F=\mathbf{F_{1}}\oplus\ker\varphi=\mathbf{F_{2}}\oplus\operatorname{Img}\varphi$ and $\dim\mathbf{F_{2}}=\dim\ker\varphi$. The image $\varphi(\partial\mathcal{U})$ is closed as $\varphi$ is closed, hence $\mathbf{a}=\varrho(\varphi(\mathcal{U}),u_{0})>0$ because $u_{0}\notin\varphi(\partial\mathcal{U})$. Let $\phi:F\to F$ be a global Lipschitz-compact linear operator with $\mathpzc{Lip}(\phi)<\mathbf{b}$ for some $\mathbf{b}>0$. Define the mapping $\Phi_{\phi}:\overline{U}\to F$ by $\Phi_{\phi}(u)=\varphi(u)+\phi(u)$. Then $\Phi_{\phi}$ is a Lipschitz-Fredholm mapping of index zero. Suppose $\mathbf{b}<\mathbf{a}/\mathbf{k}$ for some $\mathbf{k}>1$, then $\varrho(\Phi_{\phi}(u),u_{0})\geq\varrho(\varphi(e),u_{0})-\mathpzc{Lip}(\phi)\varrho(U,u_{0})>\mathbf{a}-\mathbf{bk}>0,\quad\forall u\in\partial\mathcal{U}.$ Therefore, $u_{0}\notin\Phi_{\phi}(\partial\mathcal{U})$. We obtain $\deg(\varphi,u_{0})=\deg(\Phi_{\phi},u_{0})$ as the mapping $\psi:[0,1]\times\overline{U}\to F$ defined by $(t,u)\to\varphi(u)+t\phi(u)$ is proper and $u_{0}\notin\psi(\partial U)$ for all $t$. Considering the fact that $\psi(-u)=-\psi(u)$, we may use the perturbation by compact operators to find the degree of $\varphi$. Let $\mathsf{C}$ be a set of global Lipschitz-compact linear operators $\phi:F\to F$ with $\mathpzc{Lip}(\phi)<\mathbf{b}<\mathbf{a}/\mathbf{k}$. Let $\widehat{\phi}\in\mathsf{C}$ be such that its restriction to $\mathbf{F_{1}}$ equals $u_{0}$ and $\widehat{\phi}\mid_{\ker\operatorname{D}\varphi(u_{0})}:\ker\operatorname{D}\varphi(u_{0})\to\mathbf{F_{2}}$ is an $MC^{1}$-isomorphism. Therefore, $\operatorname{D}\varphi(u_{0})+\widehat{\phi}$ and consequently $\operatorname{D}\varphi(u_{0})$ is an $MC^{1}$-isomorphism. Now define the mapping $\Psi:\mathcal{U}\times\mathsf{C}\to F$ by $(u,\phi)=\Phi_{\phi}(u)$. For sufficiently small $\mathbf{b}$ the differential $\operatorname{D}\Psi(u,\phi)(v,\psi)=(\operatorname{D}\varphi(u)+\phi)v+\psi(u)$ is surjective at $u_{0}$ as $\operatorname{D}\varphi(u_{0})$ is an $MC^{1}$-isomorphism. Also, it is clear that it is surjective at the other points. Then, the mapping $\Psi$ satisfies the assumption of Theorem 2.4, therefore, $\Psi^{-1}(u_{0})$ is a submanifold and the mapping $\Pi:\Psi^{-1}(u_{0})\to\mathsf{C}$ induced by the projection onto the second order is Lipschitz-Fredholm of index zero. By employing the local version of Sard’s theorem we may find a regular point $\overline{\phi}$ of $\Pi$, and from the proof of the Theorem 2.4 it follows that $u_{0}$ is a regular value of $\Phi_{\overline{\phi}}$ and consequently $u_{0}$ is a regular value of $\varphi$. Thus, properness and $\varphi(-u)=-\varphi(u)$ imply that $\varphi^{-1}(u_{0})=\left\\{u_{0},f_{1},-f_{1},\cdots f_{m},-f_{m}\right\\}$ and therefore $\deg(\varphi,u_{0})\equiv 1\mod 2$. ∎ ## References * [1] Eftekharinasab, K., Sard’s theorem for mappings between Fréchet manifolds, Ukr. Math. J., Vol. 62, No. 11 (2011) 1896–1905. doi: 10.1007/s11253-011-0478-z. * [2] Eftekharinasab, K., Transversality and Lipschitz-Fredholm maps, Zb. Pr. Inst. Mat. NAN Ukr. Vol. 12, No. 6 (2015), 89–104. * [3] Jones, G., Kirby, J., Le Gal, O. and Servi, T., On Local definability of holomorphic functions, The Quarterly J. Math., Vol. 70, No. 4 (2019) 1305–1326. doi: 10.1093/qmath/haz015 * [4] O. Müller, A metric approach to Fréchet geometry, Journal of Geometry and physics, Vol. 58, No.11 (2008), 1477-1500. doi: 10.1016/j.geomphys.2008. * [5] R. S. Sadyrkhanov, On infinite dimensional features of proper and closed mappings, Proceedings of the AMS, Vol. 98, No. 4 (1986) 643–648. doi:10.2307/2045743. * [6] Smale S. An infinite dimensional version of Sard’s theorem, Amr. J. Math., Vol. 87, No. 4 (1965) 861–866. doi: 10.2307/2373250. * [7] Tromba, A. J., Some theorems on Fredholm maps, Proceedings of AMS., Vol. 34, No. 2 (1972) 578–585. doi:10.2307/2038410.
# QCD Critical Point and High Baryon Density Matter ††thanks: Presented at workshop on ”Criticality in QCD and the Hadron Resonance Gas”, Wroclaw (online), July 29-31, 2020 B. Mohanty1,2 and N. Xu2,3 1School of Physical Sciences, National Institute of Science Education and Research, HBNI, Jatni 752050, India, 2Institute of Modern Physics, 509 Nanchang Road, Lanzhou 730000, China and 3Nuclear Science Division, Lawrence Berkeley National Laboratory, Berkeley, CA 94720, USA ###### Abstract We report the latest results on the search for the QCD critical point in the QCD phase diagram through high energy heavy-ion collisions. The measurements discussed are based on the higher moments of the net-proton multiplicity distributions in heavy-ion collisions. A non-monotonic variation in the product of kurtosis times the variance of the net-proton distribution is observed as a function of the collision energy with 3$\sigma$ significance. We also discuss the results of the thermal model in explaining the measured particle yield ratios in heavy-ion collisions and comparison of the different variants of hardon resonance gas model calculation to the data on higher moments of net-proton distributions. We end with a note that the upcoming programs in high baryon density regime at various experimental facilities will complete the search for the QCD critical point through heavy-ion collisions. 25.75.-q,25.75.Nq, 12.38.Mh, 12.38.-t,25.75.Gz ## 1 Introduction Figure 1: Conjectured QCD phase diagram of temperature ($T$) versus baryonic chemical potential ($\mu_{\mathrm{B}}$). See text for details. Relativistic heavy-ion collisions at varying center of mass energy ($\sqrt{s_{NN}}$) allows for the study of the phase diagram of nuclear matter [1]. The underlying theory is the one that governs the strong interactions - Quantum Chromodynamics (QCD). The conjectured phase diagram of QCD is shown in Fig. 1. The current status of the phase diagram is as follows. There are two distinct phases in the phase structure: de-confined state of quarks and gluons called the quark gluon plasma (QGP) and the confined state of gas of hadrons and resonances (HRG). The phase boundary (shown as a solid line in Fig. 1) between the hadronic gas phase and the high-temperature quark-gluon phase is a first-order phase transition line, which begins at large baryon chemical potential ($\mu_{B}$) and small temperature ($T$) and curves towards smaller $\mu_{B}$ and larger $T$. This line ends at the QCD critical point whose conjectured position, indicated by a square, is uncertain both theoretically and experimentally. At smaller $\mu_{B}$ there is a cross over indicated by a dashed line. The region of $\mu_{\mathrm{B}}$/$T$ $\leq$ 2 is shown as dot- dashed line. A comparison between RHIC data and lattice QCD (LQCD) calculations disfavours the possible QCD critical point being located at $\mu_{\mathrm{B}}$/$T$ $\leq$ 2 [2, 3]. The red-yellow dotted line corresponds to the chemical freeze-out obtained from the fits of particle yields in heavy- ion collisions using a thermal model. The liquid-gas transition region features a second order critical point (red-circle) and a first-order transition line (yellow line) that connect the critical point to the ground state of nuclear matter ($T$ $\sim$ 0 and $\mu_{\mathrm{B}}$ $\sim$ 925 MeV) [4]. The regions of the phase diagram accessed by past (AGS and SPS), ongoing (LHC, RHIC, SPS and RHIC operating in fixed target mode), and future (FAIR and NICA) experimental facilities are also indicated. In this proceeding, we discuss the success and tests of the hadron resonance gas model using the particle ratios and fluctuations in net-proton number produced in heavy-ion collisions. We also discuss the status of the search for the QCD critical point and future experimental directions in this connection at the upcoming facilities. ## 2 Particle ratio and thermal model Thermal models, assuming approximate local thermal equilibrium, have been successfully applied to matter produced in heavy-ion collisions. Most popular variant of such a model employs Grand Canonical Ensemble (GCE), hence uses chemical potentials to account for conservation of quantum numbers on an average [5]. For systems created via elementary collisions (small system) or via low energy heavy-ion collisions, the Canonical Ensemble (CE) approach is used. In the large volume limit, the GCE and the CE formalisms should be equivalent. In heavy-ion collisions at energies spanning from few GeV to few TeV it may be worthwhile to ask at what collision energy a transition from GCE to CE occurs [6] ? ### 2.1 Success of thermal model Figure 2: (1) Ratio of yields of kaon to pion ($K^{+}/\pi^{+}$ (circles) and $K^{-}/\pi^{-}$ (triangles) produced in central heavy-ion collisions at mid- rapidity as a function of $\sqrt{s_{NN}}$. Thermal fits are also shown as bands (yellow band for $K^{+}$/$\pi^{+}$ and green band for $K^{-}$/$\pi^{-}$) in the plot. Dot-dashed line represents the net-baryon density at the chemical freeze-out. The dot-dashed line represents the net-baryon density at the Chemical Freeze-out as a function of collision energy, calculated from the thermal model [13]. (2) Ratio of yields of $\phi$-meson to kaon ($\phi/K^{-}$) produced in central heavy-ion collisions at mid-rapidity as a function of $\sqrt{s_{NN}}$. The various bands shows the thermal model expectation from grand canonical ensemble (GCE) and canonical ensemble (CE) formulations in the HRG model. Figure 2 (1) in the upper panel shows the energy dependence of $K$/$\pi$ particle yield ratio produced in heavy-ion collisions at AGS [7, 8, 9], SPS [10, 11] and RHIC [12]. The thermal model calculation explains the $K$/$\pi$ ratios that reflect the strangeness content relative to entropy of the system formed in heavy-ion collisions. This can be treated as a success of the application of thermal model to heavy-ion collisions. A peak in the energy dependence of $K^{+}$/$\pi^{+}$ could be due to associated production dominance at lower energies as the baryon stopping is large. The peak is consistent with the calculated net baryon density reaching a maximum [13] has been suggested to be a signature of a change in degrees of freedom (baryon to meson [14] or hadrons to QGP [15]) while going from lower to higher energies. The $K^{-}$/$\pi^{-}$ ratio seems unaffected by the changes in the net-baryon density with collision energy and shows a smooth increasing trend. ### 2.2 Transition from grand canonical to canonical ensemble Figure 2 (2) in the lower panel shows the energy dependence of $\phi$/$K^{-}$ yield ratio measured in heavy-ion collisions [16, 17, 18]. As one moves from higher to lower collision energy, the $\phi$/$K^{-}$ ratio changes rapidly from a constant value to larger values. The transition happens below the collision energy where the freeze-out net-baryon density peaks (see upper panel). Thermal model calculations with GCE explains the measurements up to collision energy of $5$ GeV. At lower energies the GCE model expectation is that the $\phi$/$K^{-}$ ratio should decrease in contrast to that observed in experiments. On the other hand, the increase in $\phi$/$K^{-}$ at lower energies is explained by thermal model with CE framework for strangeness production. The results are also sensitive to the choice of the additional control parameter, $r_{\rm sc}$, in CE framework, which decides the typical spatial size of $s\bar{s}$ correlations. Hence, we find that a high statistics and systematic measurement of $\phi$/$K^{-}$ yield ratio can be used to test the transition of GCE to CE in thermal models. As the size of the $s\bar{s}$ correlations depends on the medium properties, such studies will provide valuable data for estimation of the volume in which open strangeness is produced. ## 3 Net-proton number fluctuations and QCD critical point The QCD critical point is a landmark on the QCD phase diagram. Experimental signatures for critical point is enhanced fluctuations coupled to the critical modes. In this respect the baryon number fluctuations are sensitive to the criticality [19]. At the critical point, generally, the correlation length takes large values, and that leads to non-Gaussian fluctuations [20]. Higher- order fluctuations are more sensitive to the criticality, the third order ($S\sigma$) and the fourth order ($\kappa\sigma^{2}$) are common measures for the QCD critical point search, where $\sigma$, $S$ and $\kappa$ are called the standard deviation, skewness and the kurtosis of the distribution, respectively. Experimentally, net-proton distribution is considered as a proxy for net-baryon distributions. ### 3.1 Net-proton number fluctuations Figure 3: (1) $S\sigma$ and (2) $\kappa\sigma^{2}$ of net-proton distributions for 70-80% peripheral (open squares) and 0-5% central (filled-circles) Au+Au collisions as a function of $\sqrt{s_{NN}}$ [21]. Projected statistical uncertainty for the second phase of the RHIC BES program is shown by the green-band and the blue arrow shows the region of $\sqrt{s_{NN}}$ to be covered by the STAR experiments fixed-target program. Results of calculations are shown for different variants (Ideal GCE [23], excluded volume [24] and CE [25]) of HRG model and transport model (UrQMD). The solid red and the dashed blue line in (2) is a schematic representation of expectation from a QCD based model calculation in presence of a critical point. Figure 3 shows the most relevant measurements over the widest range in $\mu_{B}$ ($20-450$ MeV) to date for the critical point search [21]. As we go from observables involving lower order moments ($S\sigma$) to higher order moments ($\kappa\sigma^{2}$), deviations between central and peripheral collisions for the measured values increases. Central collisions $\kappa\sigma^{2}$ data show a non-monotonic variation with collision energy with respect to the statistical baseline of $\kappa\sigma^{2}$ = 1 at a significance of $\sim$ 3$\sigma$ [21]. The deviations of $\kappa\sigma^{2}$ below the baseline are qualitatively consistent with theoretical considerations including a critical point [22]. In addition, experimental data show deviation from heavy-ion collision models without a critical point. This can be seen from the table 1 which shows values of a $\chi^{2}$ test between the experimental data and various models. In all cases, within 7.7 $<$ $\sqrt{s_{NN}}$ (GeV) $<$ 27, the $\chi^{2}$ tests return $p$-values that are less than 0.05. This implies that the monotonic energy dependence from all of the models are statistically inconsistent with the data. Although a non- monotonic variation of the experimental data with collision energy looks promising for the QCD critical point search, a more robust conclusion can be derived when the uncertainties get reduced and significance above $5\sigma$ is reached. This is the plan for the RHIC Beam Energy Scan Phase-II program. Table 1: The $p$ values of a $\chi^{2}$ test between data and various models for the $\sqrt{s_{NN}}$ dependence of $\it{S}\sigma$ and $\kappa\sigma^{2}$ values of net-proton distributions in 0-5% central Au+Au collisions. The results are for the $\sqrt{s_{NN}}$ range 7.7 to 27 GeV [21] which is the relevant region for the physics analysis presented here. Moments | HRG GCE | HRG EV | HRG CE | UrQMD ---|---|---|---|--- | | (r = 0.5 fm) | | $\it{S}\sigma$ | $<$ 0.001 | $<$ 0.001 | 0.0754 | $<$ 0.001 $\kappa\sigma^{2}$ | 0.00553 | 0.0145 | 0.0450 | 0.0221 ### 3.2 Comparison to Lattice QCD inspired fits In the previous sub-section we have seen that the data deviates from the expectations based on UrQMD and HRG models. Figures 4 and 5 show that several features of the data are qualitatively consistent with LQCD calculations of net baryon-number fluctuations up to NLO in $\mu_{B}/T$ [2]. Specifically, (a) $M/\sigma^{2}>S\sigma$, where $M$ is the mean of the net-proton distribution; $C_{3}/C_{1}$ is smaller than unity and tending to decrease with increasing $M/\sigma^{2}$; and with increasing $M/\sigma^{2}$, the cumulant ratio $C_{4}/C_{2}$ departs further away from unity than the ratio $C_{3}/C_{1}$ for $\sqrt{s_{{}_{NN}}}\geq 19.6$ GeV. The LQCD inspired fits are of the form: $C_{3}/C_{1}$ = $p_{0}$ \+ $p_{1}$ $(C_{1}/C_{2})^{2}$; $C_{4}/C_{2}$ = $p_{2}$ \+ $p_{3}$ $(C_{1}/C_{2})^{2}$ and $C_{3}/C_{2}$ = $p_{0}$ $C_{1}/C_{2}$\+ $p_{1}$ $(C_{1}/C_{2})^{3}$. Where $p_{0}$, $p_{1}$, $p_{2}$, and $p_{3}$ are fit parameters and we have used the equivalence between product of the moments and ratios of cumulants as $C_{1}/C_{2}$ = $M/\sigma^{2}$; $C_{3}/C_{1}$ = $S\sigma^{3}/M$ and $C_{4}/C_{2}$ = $\kappa\sigma^{2}$. The good agreement between data and LQCD inspired fits for $\sqrt{s_{NN}}$ range between 200 to 19.6 GeV, suggests that the heavy-ion collisions have produced a strongly interacting QCD matter. Figure 4: Net-proton cumulant ratios as a function of $M/\sigma^{2}$. Also shown are the expectations from different variants of HRG model (lines), UrQMD (yellow band) and LQCD inspired fits (green bands) [2]. Figure 5: $S\sigma$ versus the $M/\sigma^{2}$ of net-proton distribution in high energy heavy-ion collisions. Also shown are the expectation from HRG, UrQMD and LQCD inspired fits [2]. ## 4 Experimental programs for high baryon density Figure 6: Interaction rates (in Hz) for high-energy nuclear collision facilities as a function of $\sqrt{s_{NN}}$ [26]. Accelerators in collider mode are shown by blue symbols (ALICE, sPHENIX, RHIC BES-II and NICA) and those operating in fixed target mode by red symbols (STAR fixed traget (FXT), FAIR (CBM, SIS), HADES, and HIAF). As seen from the measurements discussed in previous section, to complete the critical point search program a high statistics phase - II of the beam energy scan program at RHIC is needed. In addition, future new experiments, which are all designed with high rates, large acceptance, and the state-of-the-art particle identification, at the energy region where baryon density is high, i.e., 500 MeV $<\mu_{B}<$ 800 MeV, see Fig. 6, will be needed. The new facilities for studying high baryon density matter includes (a) Nuclotron- based Ion Collider fAcility (NICA) at the Joint Institute for Nuclear Research (JINR), Dubna, Russia [27], (b) Compressed Baryonic Matter (CBM) at Facility for Antiproton and Ion Research (FAIR), Darmstadt, Germany [28], and (c) CSR External-target Experiment (CEE) at High Intensity heavy-ion Accelerator Facility (HIAF), Huizhou, China [29]. ## 5 Summary and Outlook The workshop dealt with two topics: Criticality and hadron resonance gas models. Criticality: A robust and vibrant research program is now established both experimentally (several facilities) and theoretically to study the QCD phase structure [30] and seeking for the QCD critical point in the phase diagram. The observables are well established and the results from a first systematic measurements are promising. Thermal models: Another success story has been use of hadron resonance gas models to extract freeze-out dynamics, provide evidences for local thermalisation in heavy-ion collisions and act as baseline for several measurements in heavy-ion collisions. This can be extended further to test the details of the model, like GCE vs. CE, and applications to higher order fluctuations to probe true thermal nature of the system formed in heavy-ion collisions [31]. High baryon density: Gradual shift of attention of the heavy-ion community is expected towards a return to the low energy collisions, where state-of-art accelerator facility with large luminosity and much advances detector systems with excellent particle identification will allow us to unravel the physics of a rotating high baryon density QCD matter subjected to magnetic field, similar to the neutron stars. A̱cknowledgments F. Karsch, V. Koch, A. Pandav, and K. Redlich for exciting discussions. We also thank the colleagues from STAR and ALICE collaborations. B.M. was supported in part by the Chinese Academy of Sciences President’s International Fellowship Initiative and J C Bose Fellowship from Department of Science of Technology, Government of India. N.X. was supported in part by the Chinese NSF grant No.11927901 and the US DOE grant No.KB0201022. ## References * [1] STAR Internal Note - SN0493, 2009. * [2] Bazavov, A._et al._ , _Phys. Rev._ D96, 074510 (2017). * [3] Bazavov, A._et al._ , _Phys. Rev._ D95, 054504 (2017). * [4] Fukushima, K. and Hatsuda, T., _Rept. Prog. Phys._ 74, 014001 (2011). * [5] A. Andronic, P. Braun-Munzinger, K. Redlich and J. Stachel, Nature 561 (2018) no.7723, 321-330. * [6] R. Hagedorn and K. Redlich, Z. Phys. C 27 (1985), 541. * [7] L. Ahle et al. [E866 and E917], Phys. Lett. B 476 (2000), 1-8. * [8] L. Ahle et al. [E-802 and E-866], Phys. Rev. C 60 (1999), 044904. * [9] L. Ahle et al. [E866 and E917], Phys. Lett. B 490 (2000), 53-60. * [10] S. V. Afanasiev et al. [NA49], Phys. Rev. C 66 (2002), 054902. * [11] C. Alt et al. [NA49], Phys. Rev. C 77 (2008), 024903. * [12] L. Adamczyk et al. [STAR], Phys. Rev. C 96 (2017) no.4, 044904. * [13] J. Randrup and J. Cleymans, Phys. Rev. C 74 (2006), 047901. * [14] J. Cleymans, H. Oeschler, K. Redlich and S. Wheaton, Phys. Lett. B 615 (2005), 50-54. * [15] M. Gazdzicki and M. I. Gorenstein, Acta Phys. Polon. B 30 (1999), 2705. * [16] B. I. Abelev et al. [STAR], Phys. Lett. B 673 (2009), 183-191. * [17] J. Adam et al. [STAR], Phys. Rev. C 102 (2020) no.3, 034909. * [18] G. Agakishiev et al. [HADES], Phys. Rev. C 80 (2009), 025209. * [19] Y. Hatta and M. A. Stephanov, Phys. Rev. Lett. 91 (2003), 102003 [erratum: Phys. Rev. Lett. 91 (2003), 129901]. * [20] M. A. Stephanov, Phys. Rev. Lett. 102 (2009), 032301. * [21] J. Adam et al. [STAR], [arXiv:2001.02852 [nucl-ex]]. * [22] M. A. Stephanov, Phys. Rev. Lett. 107 (2011), 052301. * [23] P. Garg, D. K. Mishra, P. K. Netrakanti, B. Mohanty, A. K. Mohanty, B. K. Singh and N. Xu, Phys. Lett. B 726 (2013), 691-696. * [24] A. Bhattacharyya, S. Das, S. K. Ghosh, R. Ray and S. Samanta, Phys. Rev. C 90 (2014) no.3, 034909. * [25] P. Braun-Munzinger, B. Friman, K. Redlich, A. Rustamov and J. Stachel, [arXiv:2007.02463 [nucl-th]]. * [26] K. Fukushima, B. Mohanty and N. Xu, [arXiv:2009.03006 [hep-ph]]. * [27] N. S. Geraksiev [NICA/MPD], J. Phys. Conf. Ser. 1390 (2019) no.1, 012121. * [28] T. Ablyazimov et al. [CBM], Eur. Phys. J. A 53 (2017) no.3, 60. * [29] S. Ruan, J. Yang, J. Zhang, G. Shen, H. Ren, J. Liu, J. Shangguan, X. Zhang, J. Zhang and L. Mao, et al. Nucl. Instrum. Meth. A 892 (2018), 53-58. * [30] S. Gupta, X. Luo, B. Mohanty, H. G. Ritter and N. Xu, Science 332 (2011), 1525-1528. * [31] S. Gupta, D. Mallick, D. K. Mishra, B. Mohanty and N. Xu, [arXiv:2004.04681 [hep-ph]].
# On a “Wonderful" Bruhat-Tits group scheme Vikraman Balaji Chennai Mathematical Institute, Plot number H1, Sipcot IT Park, Siruseri, Chennai, India<EMAIL_ADDRESS>and Yashonidhi Pandey Indian Institute of Science Education and Research, Mohali Knowledge city, Sector 81, SAS Nagar, Manauli PO 140306, India<EMAIL_ADDRESS><EMAIL_ADDRESS> ###### Abstract. In this paper we make a universal construction of Bruhat-Tits group scheme on wonderful embeddings of adjoint groups in the absolute and relative settings. We make a similar construction for the wonderful embeddings of adjoint Kac- Moody groups. These have natural classifying properties reflecting the orbit structure on the wonderful embeddings. ###### Key words and phrases: Bruhat-Tits group scheme, parahoric group, Loop groups, Wonderful compactification ###### 2000 Mathematics Subject Classification: 14L15,14M27,14D20 The support of Science and Engineering Research Board under Mathematical Research Impact Centric Support File number: MTR/2017/000229 is gratefully acknowledged. ###### Contents 1. 1 Introduction 1. 1.1 Group embeddings and buildings 2. 1.2 Statement of main results 1. 1.2.1 The case of Tits building 2. 1.2.2 The relative case of the Bruhat-Tits buiding 2. 2 Preliminaries 1. 2.0.1 Lie-data of $G/k$ 2. 2.0.2 Apartment data 3. 2.0.3 Loop groups and their parahoric subgroups 4. 2.0.4 Bruhat-Tits group scheme 5. 2.0.5 Standard Parahoric subgroups 3. 3 A Bruhat-Tits group scheme on the wonderful compactification 1. 3.0.1 The structure of ${\bf X}$ 2. 3.0.2 The local toric structure of ${\bf X}$ 3. 3.0.3 Construction of a Lie-algebra bundle $\mathcal{R}$ on ${\bf Y}_{{}_{0}}$ 4. 3.0.4 Construction of a Lie-algebra bundle $\mathcal{R}$ on ${\bf Y}$ 5. 3.0.5 Weil restrictions and Lie algebras 6. 3.0.6 Restriction of $\mathcal{R}$ to infinitesimal standard curves $U_{I}$ 7. 3.0.7 The Lie algebra bundle $\mathcal{R}$ on $\bf X$ 8. 3.0.8 Bruhat-Tits group scheme ${\mathcal{G}}_{{}_{\bf X}}^{{}^{\varpi}}$ on $\bf X$ 4. 4 The Weyl alcove and apartment case 1. 4.0.1 On the torus-embedding ${\bf Y}_{{}_{0}}^{{}^{\text{aff}}}$ 2. 4.0.2 Construction of a finite-dimensional Lie algebra bundle $J$ on ${\bf Y}_{{}_{0}}^{{}^{\text{aff}}}$ together with parabolic structures 3. 4.0.3 The parahoric group scheme on the torus embedding ${{\bf Y}^{{}^{\text{aff}}}}$ 5. 5 The Bruhat-Tits group scheme on ${\bf X}^{{}^{{aff}}}$ 6. 6 The Bruhat-Tits group scheme on ${\bf X}^{{}^{{poly}}}$ 7. 7 An analogue of a construction of Mumford 8. 8 Appendix on parabolic and equivariant bundles 1. 8.0.1 The group scheme situation 2. 8.0.2 Kawamata Coverings ## 1\. Introduction Let $G$ be an almost simple, simply-connected group over an algebraically closed field $k$ of characteristic zero and let $G_{{}_{{}_{\text{ad}}}}:=G/Z(G)$. The aim of this note is to construct certain universal group schemes on * • the De Concini-Procesi wonderful compactification [7] $\bf X$ of $G_{{}_{{}_{\text{ad}}}}$, * • the loop “wonderful embedding" ${\bf X}^{{}^{\text{aff}}}$ of the adjoint affine Kac-Moody group $G^{{}^{\text{aff}}}_{{}_{\text{ad}}}$, constructed by Solis [16], * • certain toroidal embeddings $\bar{G}_{{}_{{}_{\text{ad,A}}}}$ of the relative group scheme $G_{{}_{{}_{\text{ad,A}}}}$ modeled after [13]. The group schemes we construct are sufficiently universal to be called “wonderful". A new point of view, playing a central role in this work, is that parabolic vector bundle on logarithmic schemes can be used as a tool to make geometric constructions. Hitherto, they have been objects of study occuring as points in certain moduli spaces. Let us briefly motivate the theme of this note. ### 1.1. Group embeddings and buildings Let us recall that Tits’ buildings are basically of two types. The first one is the “absolute" Tits building or spherical building which is attached to a semi-simple group over a general field. This simplicial complex is built from simplices which correspond to parabolic subgroups. The apartments of the building correspond to parabolic subgroups containing a fixed maximal torus. This is built up out of Euclidean spaces decomposed by the usual Weyl chambers. The second one is the Bruhat-Tits building which is the “relative" building attached to a semi-simple group over a complete-valued field. This is based on its parahoric subgroups and built up out of Euclidean spaces decomposed into affine Weyl chambers. The two types of buildings can also be seen from an algebro-geometric perspective. In the absolute case, we work with a semisimple group $G_{{}_{\text{ad}}}$ of adjoint type. In this setting one has the wonderful embedding $G_{{}_{\text{ad}}}\subset{\bf X}$ where $G_{{}_{\text{ad}}}$ sits as an open dense subset of $\bf X$. The complement ${\bf X}\setminus G_{{}_{\text{ad}}}$ is stratified by subsets $Y$ and there is a bijection $Y\mapsto\\{P_{{}_{Y}}|B\subset P_{{}_{Y}}\\}$ from these strata to parabolic subgroups $P_{{}_{Y}}\subset G$ containing a fixed Borel subgroup $B$. Furthermore, this bijection extends to an isomorphism between the Tits building and the canonical complex associated with the toroidal embedding $G_{{}_{\text{ad}}}\subset{\bf X}$ (see Mumford [13, Page 178]). A second perspective is when the ground field is endowed with a complete non- archimedean discrete valuation. Let $A=k\llbracket z\rrbracket$ be a complete discrete valuation ring, $K=k(\\!(z)\\!)$ its quotient field. In this setting our basic model was constructed by Mumford [13]. He constructs a toroidal embedding $G_{{}_{\text{ad, A}}}\subset\bar{G}_{{}_{\text{ad, A}}}$ of the split group scheme $G_{{}_{\text{ad, A}}}=G_{{}_{\text{ad, A}}}\times{\rm Spec}\,A$. The strata of $\bar{G}_{{}_{\text{ad, A}}}\setminus G_{{}_{\text{ad, A}}}$ correspond bijectively to parahoric subgroups of $G(K)$ in a way that naturally extends to an isomorphism of the graph of the embedding $G_{{}_{\text{ad, A}}}\subset\bar{G}_{{}_{\text{ad, A}}}$ with the Bruhat-Tits building of $G\times{\rm Spec}\,A$ over $A$. ### 1.2. Statement of main results Classical Bruhat-Tits theory associates, to each facet $\Sigma$ of the Bruhat- Tits building, a smooth group scheme $\mathcal{G}_{{}_{\Sigma}}$ on ${\rm Spec}\,~{}A$, with connected fibres and whose generic fibre is $G\times_{{}_{{\rm Spec}\,k}}{\rm Spec}\,~{}K$. We call $\mathcal{G}_{{}_{\Sigma}}$ a Bruhat-Tits group scheme on ${\rm Spec}\,A$. The $A$-valued points $\mathcal{G}_{{}_{\Sigma}}(A)\subset G(K)$ are precisely the parahoric subgroups of $G(K)$. In this paper we construct universal analogues of the Bruhat-Tits group scheme. #### 1.2.1. The case of Tits building In the first setting, namely in the case of the Tits building, we construct an affine group scheme $\mathcal{G}_{{}_{\bf X}}$ over $\bf X$ whose restriction along each curve transversal to a strata of ${\bf X}\setminus G_{{}_{\text{ad}}}$ corresponds to the Bruhat-Tits group scheme associated to the parabolic subgroup under the bijection $Y\mapsto\\{P_{{}_{Y}}|B\subset P_{{}_{Y}}\\}$ mentioned above. To state our theorem, we introduce some relevant notations and notions. Let ${\bf X}:=\overline{G_{{}_{{}_{\text{ad}}}}}$ be the wonderful compactification of $G_{{}_{{}_{\text{ad}}}}$. We construct a locally free sheaf of Lie algebras on $\bf X$. This construction is essentially toric in the sense that it is first constructed on the toric varieties based on the negative Weyl chambers and then the ones on the bigger spaces is deduced from this construction. We endow the locally free sheaf $\mathcal{R}$ with a canonical parabolic structure at the generic points of the divisor ${\bf X}\setminus G_{{}_{{}_{\text{ad}}}}$ together with a compatible loop Lie algebra structure. We fix data $(T,B,G)$ of $G$. Let $S$ denote the set of simple roots of $G$ and $\mathbb{S}=S\cup\\{\alpha_{0}\\}$ denote the set of affine simple roots. For $\emptyset\neq\mathbb{I}\subset\mathbb{S}$ let $\mathcal{G}_{\mathbb{I}}$ denote the associated Bruhat-Tits group scheme on a dvr and for any $I\subset S$ let $\mathcal{G}^{{}^{st}}_{I}=\mathcal{G}_{\mathbb{I}}$ where $\mathbb{I}=I\cup\\{\alpha_{0}\\}$. For ${I}\subset S$, let $Z_{I}$ denote the corresponding strata of $\bf X$ and for any point $z_{{}_{I}}\in Z_{{}_{I}}$, let $C_{{}_{I}}\subset{\bf X}$ be a smooth curve with generic point in $G_{{}_{\text{ad}}}$ and closed point $z_{{}_{I}}$. Let $U_{{}_{z}}\subset C_{{}_{I}}$ be a formal neighbourhood of $z_{{}_{I}}$. We these notations, Theorem 3.10 is as follows. ###### Theorem 1.1. There exists an affine “wonderful" Bruhat-Tits group scheme ${\mathcal{G}}_{{}_{\bf X}}^{{}^{\varpi}}$ on $\bf X$ satisfying the following classifying properties. 1. (1) There is an identification of the Lie-algebra bundles $\text{Lie}({\mathcal{G}}_{{}_{\bf X}}^{{}^{\varpi}})\simeq\mathcal{R}$. 2. (2) For $\emptyset\neq I\subset S$ the restriction of ${\mathcal{G}}_{{}_{{\bf X}}}^{{}^{\varpi}}$ to the formal neighbourhood $U_{{}_{z_{{}_{I}}}}$ of $z_{{}_{I}}$ in $C_{{}_{I}}$ §(3.0.3) is isomorphic to the standard Bruhat- Tits group scheme $\mathcal{G}^{{}^{st}}_{{}_{I}}$ §(2.0.5). In (3.11), we have indicated generalizations to fields of positive characteristics. #### 1.2.2. The relative case of the Bruhat-Tits buiding In the second scenario we work in the setting of loop groups and construct an affine group scheme over a “wonderful" embedding ${\bf X}^{{}^{\text{aff}}}$ constructed by Solis [16]. In the relative case the group scheme is obtained by “integrating" a locally free sheaf of Lie algebras $\bf R$ on ${\bf X}^{{}^{\text{aff}}}$. Its construction is achieved by constructing a locally free sheaf of Lie-algebras $J$ on a finite dimensional scheme ${\bf Y}^{{}^{\text{aff}}}$ which is the closure of a torus-embedding and whose translates build up the ind-scheme ${\bf X}^{{}^{\text{aff}}}$. The sheaf $J$ comes equipped with a canonical parabolic structure on a normal crossing divisor. The sheaf $J$ plays the role analogous to $\mathcal{R}$ (on $\bf X$) once we view ${\bf Y}^{{}^{\text{aff}}}$ as built out of affine Weyl group translates of the affine space $\mathbb{A}^{\ell+1}$ whose standard coordinate hyperplanes play the role of strata of largest dimension of $\bf X$. Using this perspective, we then briefly consider a toroidal embedding $\bar{G}_{{}_{ad,A}}$ the structure of which is modeled after Mumford’s construction in [13]. We then define an affine group scheme over $\bar{G}_{{}_{ad,A}}$ which has properties analogous to those of $\mathcal{G}_{{}_{\bf X}}$. We first introduce some notation to state our theorem more precisely. Let $LG$ denote the loop group of $G$. Let $L^{\ltimes}G=\mathbb{G}_{m}\ltimes LG$ where the rotational torus $\mathbb{G}_{m}$ acts on $LG$ by acting on the uniformizer (cf. §4). Let $G^{{}^{\text{aff}}}$ denote the Kac-Moody group associated to the affine Dynkin diagram of $G$. Recall that $G^{{}^{\text{aff}}}$ is given by a central extension of $L^{\ltimes}G$ by $\mathbb{G}_{m}$. Let $T_{{}_{\text{ad}}}:=T/Z(G)$ and let us denote by $T^{\ltimes}_{{}_{\text{ad}}}$ the torus $\mathbb{G}_{m}\times T_{{}_{\text{ad}}}\subset G^{{}^{\text{aff}}}_{{}_{\text{ad}}}$. In ${\bf X}^{{}^{\text{aff}}}$, the closure ${\bf Y}^{{}^{\text{aff}}}:=\overline{T^{\ltimes}_{{}_{\text{ad}}}}$ gives a torus- embedding. The complement $Z:={\bf Y}^{{}^{\text{aff}}}\setminus T^{\ltimes}_{{}_{\text{ad}}}$ is a union $\cup_{{}_{\alpha\in\mathbb{S}}}H_{{}_{\alpha}}$ of $\ell+1$ smooth divisors meeting at normal crossings. For $\alpha\in\mathbb{S}$, let $\xi_{{}_{\alpha}}\in H_{{}_{\alpha}}$ denote the generic points of the divisors $H_{{}_{\alpha}}$’s. Let $A_{{}_{\alpha}}=\mathcal{O}_{{}_{{\bf Y}^{{}^{\text{aff}}},\xi_{{}_{\alpha}}}}$ be the dvr’s obtained by localizing at the height $1$-primes given by the $\xi_{{}_{\alpha}}$’s. Let $Y_{{}_{\alpha}}:={\rm Spec}\,(A_{{}_{\alpha}})$. In Theorem 4.1 we show the existence of a finite dimensional Lie-algebra bundle $J$ on ${\bf Y}^{{}^{\text{aff}}}$ which extends the trivial bundle with fiber $\mathfrak{g}$ on the open dense subset ${T^{\ltimes}_{{}_{\text{ad}}}}\cap{\bf Y}^{{}^{\text{aff}}}\subset{\bf Y}^{{}^{\text{aff}}}$ and for each $\alpha\in\mathbb{S}$ we have $L^{+}(J_{{}_{Y_{{}_{\alpha}}}})\simeq L^{+}({\text{Lie}}(\mathcal{G}_{{}_{\alpha}}))$. Then in Proposition 5.1 we show the existence of a finite dimensional Lie-algebra bundle $\bf R$ on ${\bf X}^{{}^{\text{aff}}}$ which extends the trivial Lie algebra bundle $G_{{}_{\text{ad}}}^{{}^{\text{aff}}}\times\mathfrak{g}$ on the open dense subset $G_{{}_{\text{ad}}}^{{}^{\text{aff}}}\subset{\bf X}^{{}^{\text{aff}}}$ and whose restriction to ${\bf Y}^{{}^{\text{aff}}}$ is $J$. The ind-scheme ${\bf X}^{{}^{\text{aff}}}$ has divisors $D_{\alpha}$ for $\alpha\in\mathbb{S}$ such that the complement of their union is ${\bf X}^{{}^{\text{aff}}}\setminus G_{ad}^{aff}$. With these notations, let us state Theorem 5.2. ###### Theorem 1.2. There exists an affine “wonderful" Bruhat-Tits group scheme ${\mathcal{G}}_{{}_{{\bf X}^{{}^{\text{aff}}}}}^{{}^{\varpi}}$ on ${\bf X}^{{}^{\text{aff}}}$ together with a canonical isomorphism $\text{Lie}({\mathcal{G}}_{{}_{\bf X^{aff}}}^{{}^{\varpi}})\simeq{\bf R}$. It further satisfies the following classifying property: For $h\in{\bf X}^{{}^{\text{aff}}}\setminus G_{ad}^{aff}$, let $\mathbb{I}\subset\mathbb{S}$ be the subset such that $h\in\cap_{\alpha\in I}D_{\alpha}$. Let $C_{{}_{\mathbb{I}}}\subset{\bf X}^{{}^{\text{aff}}}$ be a smooth curve with generic point in $G^{aff}_{{}_{\text{ad}}}$ and closed point $h$. Let $U_{{}_{h}}\subset C_{{}_{\mathbb{I}}}$ be a formal neighbourhood of the closed point $h$. Then, the restriction ${\mathcal{G}}_{{}_{{\bf X}^{{}^{\text{aff}}}}}^{{}^{\varpi}}|_{{}_{U_{{}_{h}}}}$ is isomorphic to the Bruhat-Tits group scheme $\mathcal{G}_{{}_{\mathbb{I}}}$ §(2.0.4) on $U_{{}_{h}}$. ###### Acknowledgments​​ . We thank J. Heinloth and M.Brion for their questions and comments. They have helped us very much in expressing our results with greater precision. ## 2\. Preliminaries #### 2.0.1. Lie-data of $G/k$ Let $G$ be an almost simple, simply-connected group over $k$ (see (3.11)) with the data $(T,B,G)$. Let $X(T)\,=\,\text{Hom}(T,\mathbb{G}_{{}_{m}})$ be the group of characters of $T$ and $Y(T)\,=\,\text{Hom}(\mathbb{G}_{m},T)$ be the group of all one–parameter subgroups of $T$. Let $G_{{}_{{}_{\text{ad}}}}:=G/Z(G)$ and let $\mathfrak{g}$ denote the Lie- algebra. We denote $\Phi^{+},\Phi^{-}\subset\Phi$ the set of positive and negative roots with respect to $B$. Let $S=\\{\alpha_{{}_{1}},\ldots,\alpha_{{}_{\ell}}\\}$ denote the set of simple roots of $G$, where $\ell$ is the rank of $G$. Let $\alpha^{\vee}$ denote the coroot corresponding to $\alpha\in S$ . #### 2.0.2. Apartment data Let $\mathbb{S}=S\cup\\{\alpha_{0}\\}$ denote the set of affine simple roots. Let $\mathcal{A}_{T}$ denote the affine apartment corresponding to $T$. It can be identified with the affine space ${\mathbb{E}}=Y(T)\otimes_{\mathbb{Z}}\mathbb{R}$ together with its origin $0$. Let $\mathbf{a}_{0}$ be the unique Weyl alcove of $G$ whose closure contains $0$ and which is contained in the dominant Weyl chamber corresponding to $B$. Under the natural pairing between $Y(T)\otimes_{\mathbb{Z}}\mathbb{Q}$ and $X(T)\otimes_{\mathbb{Z}}\mathbb{Q}$, the integral basis elements dual to $S$ are called the fundamental co-weights $\\{\omega^{\vee}_{\alpha}|\alpha\in S\\}$. Let $c_{\alpha}$ be the coefficient of $\alpha$ in the highest root. The vertices of the Weyl alcove $\mathbf{a}_{0}$ are indexed by $0$ and $\displaystyle\theta_{{}_{\alpha}}:={{\omega^{\vee}_{\alpha}}\over{c_{\alpha}}},\alpha\in S.$ (2.0.1) Indeed, any rational point $\theta\in\mathbf{a}_{0}$ can be expressed as $\theta_{{}_{\lambda}}$ so that there is a unique pair $(d,\lambda)\in\mathbb{N}\times Y(T)$ defined by the condition that $d$ is the least positive integer such that $\lambda=d.\theta_{{}_{\lambda}}\in Y(T).$ (2.0.2) Thus, if $e_{\alpha}$ is the order of $\omega^{\vee}_{\alpha}$ in the quotient of the co-weight lattice by $Y(T)$, it follows that for $\theta_{{}_{\alpha}}$, the number $d$ is: $d_{{}_{\alpha}}:=e_{\alpha}.c_{\alpha}.$ (2.0.3) An affine simple root $\alpha\in\mathbb{S}$ may be viewed as an affine functional on $\mathcal{A}_{T}$. Any non-empty subset $\mathbb{I}\subset\mathbb{S}$ defines the facet $\Sigma_{\mathbb{I}}\subset\overline{\mathbf{a}_{0}}$ where exactly the $\alpha$’s not lying in $\mathbb{I}$ vanish. So $\mathbb{S}$ corresponds to the interior of the alcove and the vertices of the alcove correspond to $\alpha\in\mathbb{S}$. Conversely any facet $\Sigma\subset\overline{\mathbf{a}_{0}}$ defines non-empty subset $\mathbb{I}\subset\mathbb{S}$. For $\emptyset\neq\mathbb{I}\subset\mathbb{S}$, the barycenter of $\Sigma_{\mathbb{I}}$ is given by $\theta_{\mathbb{I}}:=\frac{1}{|\mathbb{I}|}\sum_{\alpha\in\mathbb{I}}\theta_{{}_{\alpha}}.$ (2.0.4) #### 2.0.3. Loop groups and their parahoric subgroups Let $k=\mathbb{C}$, $A=k\llbracket z\rrbracket$ and let $R$ denote an arbitrary $k$-algebra. We denote the field of Laurent polynomials by $K=k(\\!(z)\\!)=k\llbracket z\rrbracket[z^{-1}]$. The loop group $LG$ on a $k$-algebra $R$ is given by $G(R(\\!(t)\\!))$. Similarly the loop Lie algebra $L\mathfrak{g}$ is given by $L\mathfrak{g}(R)=\mathfrak{g}(R(\\!(t)\\!))$. We can similarly define the positive loops (also called jet groups in the literature) $L^{+}(G)$ to be the subfunctor of $LG$ defined by $L^{+}(G)(R):=G(R\llbracket z\rrbracket)$. The positive loop construction extends more generally to any group scheme $\mathcal{G}\rightarrow{\rm Spec}\,(A)$ and any vector bundle $\mathfrak{P}\rightarrow{\rm Spec}\,(A)$ whose sheaf of sections carries a Lie-bracket. Let $L^{\ltimes}G=\mathbb{G}_{m}\ltimes LG$ where the rotational torus $\mathbb{G}_{m}$ acts on $LG$ by acting on the uniformizer via the loop rotation action as follows: $u\in\mathbb{G}_{m}(R)$ acts on $\gamma(z)\in LG(R)=G(R(\\!(t)\\!))$ by $u\gamma(z)u^{-1}=\gamma(uz)$. A maximal torus of $L^{\ltimes}G$ is $T^{\ltimes}=\mathbb{G}_{m}\times T$. A $\eta\in Hom(\mathbb{G}_{m},T^{\ltimes})\otimes_{\mathbb{Z}}\mathbb{Q}$ over $R(s)$ can be viewed as a rational $1$-PS, i.e., a $1$-PS $\mathbb{G}_{m}\rightarrow T^{\ltimes}$ over $R(w)$ where $w^{n}=s$ for some $n\geq 1$. In this case for $\gamma(z)\in L^{\ltimes}G(R)$ we will view $\eta(s)\gamma(z)\eta(s)^{-1}$ as an element in $L^{\ltimes}G(R(w))$. Hence, by the condition on $\gamma(z)\in L^{\ltimes}G(R)$ “$\lim_{{}_{s\rightarrow 0}}\eta(s)\gamma(z)\eta(s)^{-1}$ exists in $L^{\ltimes}G(R)$", we mean that there exists a $n\geq 1$ such that for $w^{n}=s$, we have $\eta(s)\gamma(z)\eta(s)^{-1}\in L^{\ltimes}G(R\llbracket w\rrbracket).$ (2.0.5) If $\eta=(a,\theta)$ for $a\in\mathbb{Q}$ and $\theta$ a rational $1$-PS of $T$, then we have $\eta(s)\gamma(z)\eta(s)^{-1}=\theta(s)\gamma(s^{a}z)\theta(s)^{-1}.$ (2.0.6) We note further that for any $0<d\in\mathbb{N}$, setting $\eta=(\frac{a}{d},\frac{\theta}{d})$ we have $\eta(s)\gamma(z)\eta(s)^{-1}=\theta(s^{\frac{1}{d}})\gamma(s^{\frac{a}{d}}z)\theta(s^{\frac{1}{d}})^{-1}.$ (2.0.7) So for $\eta=\frac{1}{d}(1,\theta)$ observe that the statement "$\lim_{s\rightarrow 0}\eta(s)\gamma(z)\eta(s)^{-1}$ exists " is a condition which is equivalent to $\theta(s)\gamma(s)\theta(s)^{-1}\in L^{\ltimes}G(R\llbracket w\rrbracket).$ (2.0.8) In other words, the condition of existence of limits is independent of $d$, and we may further set $z=s$ in $\gamma(z)$. More generally, if $a>0$ the condition for $(a,\theta)$ and $(1,\frac{\theta}{a})$ are equivalent. We may write this condition on $\gamma(s)\in L^{\ltimes}G(R)$ or $LG(R)$ as $"\lim_{s\rightarrow 0}\theta(s)\gamma(s)\theta(s)^{-1}\quad\text{ exists in}\quad L^{\ltimes}G(R)\quad\text{or}\quad LG(R)\quad\text{if}\quad\gamma(z)\in LG(R)"$ (2.0.9) For $r\in\Phi$, let $u_{r}:\mathbb{G}_{a}\rightarrow G$ denote the root subgroup. If for some $b\in\mathbb{Z}$ and $t(z)\in R\llbracket z\rrbracket$ we take $\gamma(z):=u_{r}(z^{b}t(z))\in Lu_{r}(R)$ , and $\eta:=(1,\theta)$, then $\eta(s)u_{r}(z^{b}t(z))\eta(s)^{-1}=\theta(s)u_{r}((sz)^{b}t(sz))\theta(s)^{-1}=u_{r}(s^{r(\theta)}(sz)^{b}t(sz)).$ (2.0.10) So for $\eta=(1,\theta)$ the condition that the limit exists is equivalent to $r(\theta)+b\geq 0\iff\lfloor r(\theta)+b\rfloor=\lfloor r(\theta)\rfloor+b\geq 0\iff b\geq-\lfloor r(\theta)\rfloor.$ (2.0.11) We note the independence of the above implications on the number $d$ occurring in the equation $\eta:=\frac{1}{d}(1,\theta)$. Let $\pi_{{}_{1}}:T^{\ltimes}\rightarrow\mathbb{G}_{m}$ be the first projection. For any rational $1$-PS $\eta:\mathbb{G}_{m}\rightarrow T^{\ltimes}$ we say $\eta$ is positive if $\pi_{{}_{1}}\circ\eta>0$, negative if $\pi_{{}_{1}}\circ\eta<0$ and non-zero if it is either positive or negative. Any non-zero $\eta=(a,\theta)$ defines the following positive-loop functors from the category of $k$-algebras to the category of groups and Lie-algebras: $\displaystyle{\mathcal{P}}_{{}_{\eta}}^{\ltimes}(R):=\\{\gamma\in L^{\ltimes}G(R)|\quad\lim_{s\rightarrow 0}\eta(s)\gamma(z)\eta(s)^{-1}\quad\text{exists in}\quad L^{\ltimes}G(R)\\},$ (2.0.12) $\displaystyle{\mathfrak{P}}_{{}_{\eta}}^{\ltimes}(R):=\\{h\in L^{\ltimes}\mathfrak{g}(R)|\quad\lim_{s\rightarrow 0}{Ad}(\eta(s))(h(z))\quad\text{exists in}\quad L^{\ltimes}\mathfrak{g}(R)\\},$ (2.0.13) $\displaystyle{\mathcal{P}}_{{}_{\eta}}(R):=\\{\gamma\in LG(R)|\quad\lim_{s\rightarrow 0}\eta(s)\gamma(z)\eta(s)^{-1}\quad\text{exists in}\quad LG(R)\\},$ (2.0.14) $\displaystyle{\mathfrak{P}}_{{}_{\eta}}(R)=\\{h\in L\mathfrak{g}(R)|\quad\lim_{s\rightarrow 0}Ad(\eta(s))(h(z))\quad\text{exists in}\quad L\mathfrak{g}(R)\\},$ (2.0.15) $\displaystyle\text{Thus},~{}~{}\mathcal{P}_{{}_{\eta}}:=\mathcal{P}_{{}_{\eta}}^{\ltimes}\cap(1\times LG)\quad\text{and}\quad{\mathfrak{P}}_{{}_{\eta}}:={\mathfrak{P}}_{{}_{\eta}}^{\ltimes}\cap(0\oplus L\mathfrak{g}).$ (2.0.16) A parahoric subgroup of $L^{\ltimes}G$ (resp. $LG$) is a subgroup that is conjugate to $\mathcal{P}^{\ltimes}_{{}_{\eta}}$ (resp. $\mathcal{P}_{{}_{\eta}}$) for some $\eta$. In this paper, we will mostly be using only the case when $\eta=\frac{1}{d}(1,\theta)$. In this case, letting $L\mathfrak{g}(R)=\mathfrak{g}(R(\\!(s)\\!))$ as in (2.0.9), we may reformulate (2.0.16) as ${\mathfrak{P}}_{{}_{\eta}}(R)=\\{h\in L\mathfrak{g}(R)|\quad\lim_{s\rightarrow 0}Ad(\theta(s))(h(s))\quad\text{exists in}\quad L\mathfrak{g}(R)\\}.$ (2.0.17) By the conditions (2.0.12) and (2.0.16), using (2.0.11), we may express $\mathcal{P}_{{}_{\eta}}$ in terms of generators as follows: $\mathcal{P}_{{}_{\eta}}(R)=<T(R\llbracket z\rrbracket),u_{r}(z^{-\lfloor(r,\theta)\rfloor}R\llbracket z\rrbracket),r\in\Phi>,$ (2.0.18) Let $\mathfrak{t}$ denote the Cartan sub-algebra of $T$ and $\mathfrak{u}_{r}$ denote the root algebras associated to $u_{r}$. Then the Lie-algebra functor of $\mathcal{P}_{{}_{\eta}}$ in terms of generators is given by $Lie(\mathcal{P}_{{}_{\eta}})(R)=<\mathfrak{t}(R\llbracket z\rrbracket),\mathfrak{u}_{r}(z^{-\lfloor(r,\theta)\rfloor}R\llbracket z\rrbracket),r\in\Phi>.$ (2.0.19) Thus, for any $\eta$ of the type $(1,\theta)$, we get the equality $\text{Lie}(\mathcal{P}_{{}_{\eta}})={\mathfrak{P}}_{{}_{\eta}}.$ (2.0.20) This can be seen by (2.0.13) and (2.0.16) and then replacing the conjugation in (2.0.6) by $Ad(\eta(s))$. For a rational $1$-PS $\theta$ of $T$, with $\eta=(1,\theta)$ we will have the notation: ${\mathfrak{P}}_{{}_{\theta}}:={\mathfrak{P}}_{{}_{\eta}}.$ (2.0.21) In particular, for $r\in\Phi$ let $\mathfrak{g}_{{}_{r}}\subset\mathfrak{g}$ be the root subspace. Let $m_{{}_{r}}(\theta):=-\lfloor{(r,\theta)}\rfloor$. The parahoric Lie algebra ${\mathfrak{P}}_{{}_{\theta}}(k)\subset\mathfrak{g}(K)$ has $T$-weight space decomposition as: $\displaystyle{\mathfrak{P}}_{{}_{\theta}}(k)=\mathfrak{t}(A)\bigoplus z^{{}^{m_{{}_{r}}(\theta)}}\mathfrak{g}_{{}_{r}}(A).$ (2.0.22) #### 2.0.4. Bruhat-Tits group scheme To each facet $\Sigma_{\mathbb{I}}\subset\overline{\mathbf{a}_{0}}$, Bruhat- Tits theory associates a smooth group scheme $\mathcal{G}_{{}_{\mathbb{I}}}$ on ${\rm Spec}\,~{}A$, with connected fibres and whose generic fibre is $G\times_{{}_{{\rm Spec}\,k}}{\rm Spec}\,~{}K$. We call $\mathcal{G}_{{}_{\mathbb{I}}}$ a Bruhat-Tits group scheme on ${\rm Spec}\,A$. To $\mathcal{G}_{{}_{\mathbb{I}}}$, we can associate a pro-algebraic group $L^{+}(\mathcal{G}_{{}_{\mathbb{I}}})$ over $k$ as follows: $L^{+}(\mathcal{G}_{{}_{\mathbb{I}}})(R):=\mathcal{G}_{{}_{\mathbb{I}}}(R\llbracket z\rrbracket).$ (2.0.23) The $L^{+}(\mathcal{G}_{{}_{\mathbb{I}}})$ also characterise $\mathcal{G}_{{}_{\mathbb{I}}}$. We thus have the identifications: $L^{+}(\mathcal{G}_{{}_{\mathbb{I}}})=\mathcal{P}_{{}_{\eta}}\quad\text{and}\quad\text{Lie}(L^{+}(\mathcal{G}_{{}_{\mathbb{I}}}))={\mathfrak{P}}_{{}_{\eta}}$ (2.0.24) for $\eta=(1,\theta)$ where $\theta$ is any rational $1$-PS lying in $\Sigma_{\mathbb{I}}$. #### 2.0.5. Standard Parahoric subgroups The standard parahoric subgroups of $G(K)$ are parahoric subgroups of the distinguished hyperspecial parahoric subgroup $G(A)$. These are realized as inverse images under the evaluation map $ev:G(A)\to G(k)$ of standard parabolic subgroups of $G$. In particular, the standard Iwahori subgroup ${\mathfrak{I}}$ is a standard parahoric and indeed, ${\mathfrak{I}}=ev^{-1}(B)$. Denoting by $Q_{{}_{I}}\subset G$ the parabolic subgroup associated to the subset $I\subset S$ we will denote by $\mathcal{G}^{{}^{st}}_{{}_{{I}}}$ the standard parahoric subgroups of $G(A)$ determined by $L^{+}(\mathcal{G}^{{}^{st}}_{{}_{{I}}})(k):=ev^{-1}(Q_{{}_{I}}).$ (2.0.25) Thus, for ${I}\subset S$, setting $\mathbb{I}:=I\cup\\{\alpha_{0}\\}$ we have $L^{+}(\mathcal{G}^{{}^{st}}_{{}_{{I}}})(k)=L^{+}(\mathcal{G}_{{}_{\mathbb{I}}})(k)\quad\text{and}\quad L^{+}(\mathfrak{P}_{{}_{I}}^{{}^{st}})=L^{+}(\mathfrak{P}_{{}_{\mathbb{I}}}).$ (2.0.26) ## 3\. A Bruhat-Tits group scheme on the wonderful compactification #### 3.0.1. The structure of ${\bf X}$ Let ${\bf X}:=\overline{G_{{}_{{}_{\text{ad}}}}}$ be the wonderful compactification of $G_{{}_{{}_{\text{ad}}}}$. Let $\\{D_{\alpha}|\alpha\in S\\}$ denote the irreducible smooth divisors of ${\bf X}$. Let $D:=\cup_{{\alpha\in S}}D_{\alpha}$. Then $X\setminus G_{{}_{{}_{\text{ad}}}}=D$. The pair $({\bf X},D)$ is the primary example of a $(G_{{}_{{}_{\text{ad}}}}\times G_{{}_{{}_{\text{ad}}}})$-homogenous pair [5, Brion]. Let $P_{{}_{I}}$ be the standard parabolic subgroup defined by subsets $I\subset S$ , the notation being such that the Levi subgroup $L_{{}_{I,ad}}$ containing $T_{{}_{\text{ad}}}$ has root system with basis $S\setminus I$. Recall that the $(G_{{}_{{}_{\text{ad}}}}\times G_{{}_{{}_{\text{ad}}}})$-orbits in $\bf X$ are indexed by subsets $I\subset S$ and have the following description: $Z_{{}_{I}}=(G_{{}_{{}_{\text{ad}}}}\times G_{{}_{{}_{\text{ad}}}})\times_{{}_{P_{{}_{I}}\times P^{-}_{{}_{I}}}}L_{{}_{I,ad}}.$ Then by [4, Proposition A1], each $Z_{{}_{I}}$ contains a unique base point $z_{{}_{I}}$ such that $(B\times B^{-}).z_{{}_{I}}$ is dense in $Z_{{}_{I}}$ and there is a $1$-PS $\lambda$ of $T$ satisfying $P_{{}_{I}}=P(\lambda)$ and $\lim_{{}_{t\to 0}}\lambda(t)=z_{{}_{I}}$. The closures of these $\lambda$ define curves $C_{{}_{I}}\subset{\bf X}$ which meet the strata $Z_{{}_{I}}$ transversally at $z_{{}_{I}}$. In particular, if $I=\\{\alpha\\}$ a singleton, then the divisor $D_{{}_{\alpha}}$ is the orbit closure $\bar{Z}_{{}_{I}}$ and the $1$-PS can be taken to be the fundamental co-weight $\omega^{\vee}_{\alpha}$. The closure of the $1$-PS $\omega^{\vee}_{\alpha}:\mathbb{G}_{{}_{m}}\to G_{{}_{\text{ad}}}$ defines the curve $C_{{}_{\alpha}}\subset\bf X$ transversal to the divisor $D_{{}_{\alpha}}$ at the point $z_{{}_{\alpha}}$. #### 3.0.2. The local toric structure of ${\bf X}$ Let ${\bf Y}:=\overline{T_{{}_{\text{ad}}}}$ be the closure of $T_{{}_{\text{ad}}}$ in ${\bf X}$. Recall that ${\bf Y}$ is a projective toric variety associated to the fan of Weyl chambers. In what follows, we will mostly work with the affine toric embedding ${\bf Y}_{{}_{0}}:=\overline{T_{{}_{ad,0}}}\simeq{\mathbb{A}}^{{}^{\ell}}$ which is the toric variety associated to the negative Weyl chamber. Let $U$ (resp $U^{{}^{-}}$) be the unipotent radical (resp. its opposite) of the Borel subgroup $B\subset G$. We also recall that one may identify $U\times U^{{}^{-}}\times{\bf Y}_{{}_{0}}$ with an open subset of $\bf X$ and moreover the $G\times G$-translates of $U\times U^{{}^{-}}\times{\bf Y}_{{}_{0}}$ covers the whole of $\bf X$. #### 3.0.3. Construction of a Lie-algebra bundle $\mathcal{R}$ on ${\bf Y}_{{}_{0}}$ The aim of this section is to construct a Lie algebra bundle on ${\bf Y}_{{}_{0}}$ (see (3.3)), which may be termed “parahoric". We close by making a similar construction on $\bf Y$ and $\bf X$. The construction is motivated by the structure of the kernel of the Tits fibration in [5]. We begin with a couple of elementary lemmas which should be well known. ###### Lemma 3.1. Let $E$ be a locally free sheaf on an irreducible smooth scheme $X$. Let $\xi\in X$ be the generic point and let $W\subset E_{{}_{\xi}}$ be an $\mathcal{O}_{{}_{\xi}}$-submodule. Then there exists a unique coherent subsheaf $F\subset E$ such that $F_{{}_{\xi}}=W$ and $Q:=Coker(F\hookrightarrow E)$ is torsion-free. Moreover $F$ is a reflexive sheaf. ###### Proof. Define $\tilde{F}$ on the affine open $U\subset X$ by $\tilde{F}(U):=E(U)\cap W$. Then it is easily seen that $\tilde{F}$ defines a coherent subsheaf $F\subset E$ and it is the maximal coherent subsheaf of $E$ whose fiber over $\xi$ is $W$. To check $Q$ is torsion-free, let $T$ be the torsion submodule of $Q$. Let $K:=Ker(E\to Q/T)$. Then since $\xi\notin Supp(T)$, so $K_{{}_{\xi}}=W$ and hence, by the maximality of $F$ we have $K=F$. Since $E$ is locally free, we have $F^{\vee\vee}/F\hookrightarrow E/F$. But since $F^{\vee\vee}/F$ is only torsion, and $E/F=Q$ is torsion-free, it follows that $F$ is automatically a reflexive sheaf.∎ ###### Lemma 3.2. Let $X$ be as above and $i:U\hookrightarrow X$ an open subset such that $X\setminus U$ has codimension $\geq 2$ in $X$. Let $F_{{}_{U}}$ be a reflexive sheaf on $U$. Then $i_{{}_{*}}(F_{{}_{U}})$ is a reflexive sheaf on $X$ which extends $F_{{}_{U}}$. ###### Proof. . By [11, Corollaire 9.4.8], there exists a coherent $\mathcal{O}_{{}_{X}}$-module $F_{1}$ such that $F_{1}|_{{}_{U}}\simeq F_{{}_{U}}$. Set $F:=F_{1}^{{}^{**}}$ to be the double dual. Then $F$ is reflexive and also since $F_{{}_{U}}$ is reflexive, $F|_{{}_{U}}\simeq F_{{}_{U}}$. Hence, we have $i_{{}_{*}}(F_{{}_{U}})=F$ [12, Proposition 1.6]. ∎ Let $\lambda=\sum_{\alpha\in S}k_{{}_{\alpha}}\omega_{{}_{\alpha}}^{{\vee}}$ be a dominant $1$-PS of $T_{{}_{\text{ad}}}$. It defines the curve $C_{{}_{\lambda}}\subset\bf Y$. When the $k_{{}_{\alpha}}$ are constrained to be in $\\{0,1\\}$, then these curves are the standard curves $C_{{}_{I}},I\subset S$ considered in §(3.0.1), which meet the strata transversally at the points $z_{{}_{I}}$. We call these dominant $\lambda$’s as standard. In particular, the curve defined by $\omega_{{}_{\alpha}}^{{\vee}}$ meets the divisor $H_{{}_{\alpha}}$ transversally at $z_{{}_{\alpha}}$. The formal neighbourhood of the closed point $z_{{}_{\lambda}}$ in $C_{{}_{\lambda}}$ identifies with the spectrum $U_{{}_{\lambda}}:={\rm Spec}\,(A_{{}_{\lambda}})$ of $A_{{}_{\lambda}}=k\llbracket s\rrbracket$ with quotient field $K_{{}_{\lambda}}=k(\\!(s)\\!)$. We set $\theta_{\lambda}:=\sum_{\alpha\in S}k_{{}_{\alpha}}\frac{\theta_{{}_{\alpha}}}{\ell+1}.$ (3.0.1) When $k_{\alpha}\in\\{0,1\\}$, then $\theta_{\lambda}$ does not lie on the far wall of $\mathbf{a}_{0}$. So $\mathfrak{P}^{{}^{st}}_{{}_{\theta_{\lambda}}}=\mathfrak{P}_{{}_{\theta_{\lambda}}}$ §(2.0.5). ###### Theorem 3.3. There exists a canonical Lie-algebra bundle $\mathcal{R}$ on ${\bf Y}_{{}_{0}}$ which extends the trivial bundle with fiber $\mathfrak{g}$ on ${T_{{}_{\text{ad}}}}\subset{\bf Y}_{{}_{0}}$ and such that for $K_{{}_{\alpha}}\in\\{0,1\\}$ we have the identification of functors from the category of $k$-algebras to $k$-Lie-algebras: $L^{+}(\mathfrak{P}^{{}^{st}}_{{}_{\theta_{\lambda}}})=L^{+}(\mathcal{R}\mid_{{}_{U_{{}_{\lambda}}}}).$ (3.0.2) ###### Proof. With notations as in (2.0.3), consider the inclusion of lattices: $\bigoplus_{{}_{\alpha\in S}}{\mathbb{Z}}.\omega_{{}_{\alpha}}^{{\vee}}\hookrightarrow\bigoplus_{{}_{\alpha\in S}}{\mathbb{Z}}.\frac{{\theta_{{}_{\alpha}}}}{e_{{}_{\alpha}}.(\ell+1)}.$ (3.0.3) This induces an inclusion of $k$-algebras $B_{{}_{0}}\subset B$, where $B:=k[y_{\alpha},y_{\alpha}^{-1}]_{{}_{\alpha\in S}},~{}~{}\\\ B_{{}_{0}}:=k[x_{\alpha},x_{\alpha}^{-1}]_{{}_{\alpha\in S}}$ (3.0.4) and $B_{{}_{0}}\subset B$ is given by the equations: $\\{{y_{{}_{\alpha}}^{{}^{(\ell+1)d_{{}_{\alpha}}}}=x_{{}_{\alpha}}}\\}_{{}_{\alpha\in S}}$. Let ${\bf T}:={\rm Spec}\,(B),\text{and}T_{{}_{\text{ad}}}={\rm Spec}\,(B_{{}_{0}})$ and let $p:{\bf T}\to T_{{}_{\text{ad}}}$ be the natural morphism. We define the “roots" map: $\displaystyle\mathfrak{r}:{\bf T}\to T_{{}_{\text{ad}}},~{}~{}~{}\text{as}$ (3.0.5) $\displaystyle{\mathfrak{r}}^{\\#}\big{(}\big{(}x_{\alpha}\big{)}\big{)}:=\big{(}y_{\alpha}\big{)}.$ (3.0.6) Note that as a map between tori, $\mathfrak{r}$ is an isomorphism. We consider the map $\displaystyle Ad\circ\mathfrak{r}:{\bf T}\times\mathfrak{g}\rightarrow{\bf T}\times\mathfrak{g}$ (3.0.7) $\displaystyle({\bf t},x)\mapsto\big{(}{{\bf t}},Ad\big{(}\mathfrak{r}({\bf t})\big{)}(x)\big{)}.$ (3.0.8) We define the embedding of modules $j:\mathfrak{g}(B_{{}_{0}})\hookrightarrow\mathfrak{g}(B)\stackrel{{\scriptstyle Ad\circ\mathfrak{r}}}{{\rightarrow}}\mathfrak{g}(B)$ (3.0.9) where the second map is the one induced by (3.0.7) on sections. When $\mathfrak{g}(B_{0})$ is viewed as sections of the trivial bundle on $T_{{}_{\text{ad}}}$ with fibers $\mathfrak{g}$, then it also has a $T_{{}_{\text{ad}}}$-weight space decomposition. Let $B^{+}=k[y_{{}_{\alpha}}]_{{}_{\alpha\in S}}$ and $B_{{}_{0}}^{+}:=k[x_{{}_{\alpha}}]_{{}_{\alpha\in S}}$. Taking intersection as Lie submodules of $\mathfrak{g}(B)$ we define the following $B_{{}_{0}}^{+}$-module $\mathcal{R}(B_{{}_{0}}^{+}):=j(\mathfrak{g}(B_{{}_{0}}))\cap\mathfrak{g}(B^{+}).$ (3.0.10) Observe that $\mathcal{R}$ is a reflexive sheaf (3.1) on the affine embedding $T_{{}_{\text{ad}}}\hookrightarrow{\bf Y}_{{}_{0}}:={\rm Spec}\,(B_{{}_{0}}^{+})(\simeq{\mathbb{A}}^{{}^{\ell}})$. Further, this is a sheaf of Lie algebras with its Lie bracket induced from $\mathfrak{g}(B)$. We now check that $\mathcal{R}$ is locally-free on ${\bf Y}_{{}_{0}}$. Observe that the intersection (3.0.10) also respects $T_{{}_{\text{ad}}}$-weight space decomposition on the sections. More precisely, for a root $r\in\Phi$, since by definition we have the identification $\mathcal{R}(B_{{}_{0}}^{+})_{r}=j(\mathfrak{g}_{r}(B_{{}_{0}}))\cap\mathfrak{g}_{r}(B^{+})$, we get the following equalities: $j(\mathfrak{g}(B_{{}_{0}}))\cap\mathfrak{g}_{r}(B^{+})=j(\mathfrak{g}_{r}(B_{{}_{0}}))\cap\mathfrak{g}(B^{+})=\mathcal{R}(B_{{}_{0}}^{+})_{r}.$ (3.0.11) Since $\mathcal{R}$ is reflexive, so there is an open subset $U\subset{\bf Y}_{{}_{0}}$, whose complement is of codimension at least two such that the restriction $\mathcal{R}^{\prime}:=\mathcal{R}|_{{}_{U}}$ is locally free. Clearly $U$ contains $T_{{}_{\text{ad}}}$ and the generic points $\zeta_{{}_{\alpha}}$ of the divisors $H_{{}_{\alpha}}$. This gives a decomposition on $\mathcal{R}^{\prime}$ obtained by restriction from $\mathcal{R}$. The locally free sheaf $\mathcal{R}^{\prime}$ is a direct sum of the trivial bundle $\text{Lie}(T_{{}_{\text{ad}}})\times U$ (coming from the $0$-weight space) and the invertible sheaves coming from the root decomposition. Now since invertible sheaves extend across codimension $\geq 2$, the reflexivity of $\mathcal{R}$ implies that this direct sum decomposition of $\mathcal{R}^{\prime}$ extends to ${\bf Y}_{{}_{0}}$ (see [12, Proposition 1.6, page 126]). Whence $\mathcal{R}$ is locally free. Let $\lambda:\mathbb{G}_{{}_{m}}\to T_{{}_{\text{ad}}}$ be a $1$-PS. This defines a rational $1$-PS $\theta_{{}_{\lambda}}:\mathbb{G}_{{}_{m}}\to T_{{}_{\text{ad}}}$. The map $\mathfrak{r}:{\bf T}\to T_{{}_{\text{ad}}}$ is abstractly an isomorphism of tori and we can therefore consider the rational $1$-PS $\boldsymbol{\theta_{{}_{\lambda}}}:\mathbb{G}_{{}_{m}}\to{\bf T}$ defined by $\boldsymbol{\theta_{{}_{\lambda}}}:=\mathfrak{r}^{{}^{-1}}\circ\theta_{{}_{\lambda}}$. Let $p:{\bf T}\to T_{{}_{\text{ad}}}$ be the canonical map induced by $B_{{}_{0}}\subset B$. We observe the following: 1. (1) $p\circ\boldsymbol{\theta_{{}_{\lambda}}}=\lambda$, 2. (2) $\mathfrak{r}\circ\boldsymbol{\theta_{{}_{\lambda}}}=\theta_{{}_{\lambda}}.$ Let $\lambda=\sum k_{{}_{\alpha}}.\omega_{{}_{\alpha}}^{{\vee}}$ be a standard dominant $1$-PS of $T_{{}_{\text{ad}}}$. For example, the case $\lambda=\omega_{{}_{\alpha}}^{{\vee}}$, viewed as a $1$-PS, may be expressed in $\ell$-many coordinates as follows: $\omega_{{}_{\alpha}}^{{\vee}}(s)=(1,\ldots,s,\ldots 1)$ (3.0.12) with $s$ at the coordinate corresponding to $\alpha\in S$. Thus, we may express $\lambda$ as: $\lambda(s)=\prod_{\alpha\in S}\omega_{{}_{\alpha}}^{{\vee}}(s^{{}^{k_{{}_{\alpha}}}}).$ (3.0.13) Set $f_{{}_{\alpha}}:=\frac{k_{{}_{\alpha}}}{d_{{}_{\alpha}}.(\ell+1)}$. In this case we have the expression: $\boldsymbol{\theta_{{}_{\lambda}}}(s)=\prod_{\alpha\in S}\omega_{{}_{\alpha}}^{{\vee}}(s^{{}^{f_{{}_{\alpha}}}}).$ (3.0.14) We now check the isomorphism $L^{+}({\mathcal{R}}|_{{}_{U_{{}_{\lambda}}}})\simeq L^{+}(\mathfrak{P}_{{}_{\theta_{\lambda}}}^{{}^{st}})$ by first evaluating at $k$-valued points. By (3.0.10), a section s of $\mathcal{R}$ over $U_{{}_{\lambda}}={\rm Spec}\,A_{{}_{\lambda}}$, is firstly given by a local section $\text{\cursive s}_{{}_{K}}$ over the generic point of $A_{{}_{\lambda}}=k\llbracket s\rrbracket$. This is firstly an element in $j(\mathfrak{g}(K))$. Observe that this element may be written as $\text{\cursive s}_{{}_{K}}=Ad\big{(}\mathfrak{r}({\boldsymbol{\theta_{{}_{\lambda}}}(s)}\big{)}(x_{{}_{K}}))=Ad\big{(}\theta_{{}_{\lambda}}(s)\big{)}(x_{{}_{K}})).$ Let $d$ be any positive integer such that $d.\theta_{\lambda}$ becomes a $1$-PS of $T_{{}_{\text{ad}}}$. Let $(\tilde{B},w,L)$ be defined by taking a $d$-th root of the uniformizer $s$ of $A_{{}_{\lambda}}$ and let $\tilde{U}_{{}_{\lambda}}={\rm Spec}\,(\tilde{B})$. By (3.0.14), we may express $\boldsymbol{\theta_{{}_{\lambda}}}$ in terms of $w$ as $\boldsymbol{\theta_{{}_{\lambda}}}(s)=\prod_{\alpha\in S}\omega_{{}_{\alpha}}^{{\vee}}(w^{{}^{d.f_{{}_{\alpha}}}}).$ In other words, $\theta_{\lambda}(s)=\theta_{\lambda}(w^{d})=(d\theta_{{}_{\lambda}})(w)$ has become integral in $w$. Therefore, the membership of s in $\mathfrak{g}(B^{+})$ (3.0.10) gets interpreted as follows: $Ad(\theta_{\lambda}(s))(x_{K})\in L\mathfrak{g}(\tilde{B}).$ (3.0.15) But this is exactly the Lie-algebra version of the condition "$\lim_{s\rightarrow 0}$" condition for $\eta=(1,\theta)$ in the observation (2.0.8). In our situation we have assumed $R=k$ and so (3.0.15) is equivalent to $\text{\cursive s}\in\mathfrak{P}_{{}_{\theta_{{}_{\lambda}}}}(k).$ (3.0.16) But by (2.0.26), $\mathfrak{P}_{{}_{\theta_{{}_{\lambda}}}}(k)=\mathfrak{P}_{{}_{\theta_{{}_{\lambda}}}}^{{}^{st}}(k)$ because $\theta_{{}_{\lambda}}$ lies in the alcove $\bf a_{{}_{0}}$ but not on the far wall. Thus, we get the equality $\mathcal{R}|_{{}_{U_{{}_{\lambda}}}}(A_{{}_{\lambda}})=\mathfrak{P}_{{}_{\theta_{{}_{\lambda}}}}^{{}^{st}}(k).$ (3.0.17) This proves the assertion for $k$-valued points. The above proof goes through for all $k$-algebras because the Lie-bracket is already defined on $k$, and so is the underlying module structure. ∎ ###### Remark 3.4. Let $\lambda$ be an arbitrary dominant $1$-PS of $T_{{}_{\text{ad}}}$. In general, the coefficients $k_{{}_{\alpha}}$ in its expression in terms of the $\omega_{{}_{\alpha}}^{\vee}$ need not be such that $\sum k_{{}_{\alpha}}/(\ell+1)<1$. Then we get the identification $L^{+}(\mathfrak{P}_{{}_{\theta_{\lambda}}})=L^{+}(\mathcal{R}\mid_{{}_{U_{{}_{\lambda}}}})$ without the standardness of the parahoric. ###### Remark 3.5. In the setting of Theorem 3.3, we may view $p:{\rm Spec}\,(B^{{}^{+}})\to{\rm Spec}\,(B_{{}_{0}}^{{}^{+}})$ as a ramified covering space of affine toric varieties induced by the inclusion (3.0.3) of lattices. The Galois group for this covering is the dual of the quotient of lattices. The computation in 3.3 can be seen in the light of [1], in the sense that via an "invariant direct image" process, one is able to recover the complete data of all standard parahoric Lie algebras from this explicit Kawamata cover. #### 3.0.4. Construction of a Lie-algebra bundle $\mathcal{R}$ on ${\bf Y}$ In this subsection, we deduce the existence of Lie algebra bundles on the projective non-singular toric variety $\bf Y$ with the classifying properties. The variety $\bf Y$ is the toric variety for ${T_{{}_{\text{ad}}}}$ with fan consisting of the Weyl chambers. The complement ${\bf Y}\setminus{T_{{}_{\text{ad}}}}$ is a union of translates of the hyperplanes $H_{{}_{\alpha}}\subset{\bf Y}_{{}_{0}}$ by the Weyl group $W$. Thus, for each $w\in W$, we can take the locally free sheaf $\mathcal{R}$ of Lie-algebras on ${\bf Y}_{{}_{0}}$ and its $w$-translate $w.\mathcal{R}$ on ${\bf Y}_{{}_{w}}:=w{\bf Y}_{{}_{0}}$ in $\bf Y$. Let $Y^{\prime\prime}\subset{\bf Y}_{{}_{0}}$ be the open subset consisting of the open orbit $T_{{}_{\text{ad}}}$ and of the orbits of codimension 1. In other words, $Y^{\prime\prime}$ is obtained from ${\bf Y}_{{}_{0}}$ by removing all the $T_{{}_{\text{ad}}}$-orbits of codimension at least 2. The data above defines a locally free sheaf of Lie algebras on $\cup_{{}_{w\in W}}w.Y^{\prime\prime}$ because $w.Y^{\prime\prime}\cap Y^{\prime\prime}$ is either $Y^{\prime\prime}$ or $T_{{}_{\text{ad}}}$. Using (3.2), we get a reflexive sheaf which by abuse of notation can be called $\mathcal{R}$ on the whole of ${\bf Y}$. That $\mathcal{R}$ on $\bf Y$ is locally free is immediate since its restrictions to the open cover given by ${\bf Y}_{{}_{0}}$ and its $W$-translates is locally free. The Lie algebra structure also extends since it is already there on an open subset with complement of codimension $\geq 2$. Thus, we conclude: ###### Corollary 3.6. There exists a canonical Lie-algebra bundle $\mathcal{R}$ on ${\bf Y}$ which extends the trivial bundle with fiber $\mathfrak{g}$ on ${T_{{}_{\text{ad}}}}\subset{\bf Y}$ with properties as in (3.3) at the translates of the hyperplanes $H_{{}_{\alpha}}$. #### 3.0.5. Weil restrictions and Lie algebras Let $X$ be an arbitrary $k$-scheme. For an affine (or possibly ind-affine) group scheme $\mathcal{H}\rightarrow X$, we denote $\text{Lie}(\mathcal{H})$ the sheaf of Lie-algebras on $X$ whose sections on $U\rightarrow X$ are given by $\text{Lie}(\mathcal{H})(U)=\text{ker}(\mathcal{H}(U\times k[\epsilon])\rightarrow\mathcal{H}(U)).$ (3.0.18) ###### Lemma 3.7. Let $p:\tilde{X}\rightarrow X$ be a finite flat map of noetherian schemes. Let $Res_{\tilde{X}/X}$ denote the “Weil restriction of scalars" functor. Let $\mathcal{H}\rightarrow\tilde{X}$ be an affine group scheme. Then we have a natural isomorphism ${\text{Lie}}(Res_{\tilde{X}/X}\mathcal{H})\simeq Res_{\tilde{X}/X}{\text{Lie}}(\mathcal{H}).$ (3.0.19) When $p$ is also Galois with Galois group $\Gamma$, then in characteristic $0$, we have ${\text{Lie}}((Res_{\tilde{X}/X}\mathcal{H})^{\Gamma})\simeq(Res_{\tilde{X}/X}{\text{Lie}}(\mathcal{H}))^{\Gamma}.$ (3.0.20) ###### Proof. See [9, Page 533] and [10, 3.1, page 293] respectively. ∎ #### 3.0.6. Restriction of $\mathcal{R}$ to infinitesimal standard curves $U_{I}$ Let $U_{{}_{I}}$ be the formal neighbourhood of the standard curve $C_{{}_{I}}$ §(3.0.3) at its closed point. We call the corresponding dominant $1$-PS $\lambda$ of $T_{{}_{\text{ad}}}$ to be standard. Let $\theta_{{}_{\lambda}}$ be the point in $\mathbf{a}_{0}$ (3.0.1). Recall that by [1, Proposition 5.1.2] for each $\theta_{{}_{\lambda}}$ there exists a ramified cover $q_{{}_{\lambda}}:U^{\prime}_{{}_{\lambda}}\to U_{{}_{\lambda}}$ (of ramification index $d$ as in (2.0.2)) together with a $\Gamma_{{}_{\lambda}}$-equivariant $G$-torsor $E_{{}_{\lambda}}$ such that the adjoint group scheme $\mathcal{H}_{{}_{\lambda}}=E_{{}_{\lambda}}(G)$ has simply-connected fibers isomorphic to $G$ and we have the identification of $U_{{}_{\lambda}}$-group schemes: $\displaystyle\big{(}\text{Res}_{{}_{U^{\prime}/U}}(\mathcal{H}_{{}_{\lambda}})\big{)}^{{}^{\Gamma_{{}_{\lambda}}}}\simeq\mathcal{G}^{{}^{st}}_{{}_{\theta_{\lambda}}}.$ (3.0.21) We therefore get the following useful corollary to (3.3). ###### Corollary 3.8. For any standard dominant $1$-PS $\lambda$ of $T_{{}_{ad}}$, with notations as above we have an isomorphism as sheaves of Lie algebras: $\mathcal{R}|_{{}_{U_{{}_{\lambda}}}}\simeq q^{{}^{\Gamma}}_{{}_{\lambda,*}}(E_{{}_{\lambda}}(\mathfrak{g})).$ (3.0.22) ###### Proof. This is an immediate consequence of (3.3) and [1, Proposition 5.1.2]. ∎ All results in this subsection generalize suitably to non-standard curves corresponding to dominant $1$-PS in $T_{{}_{\text{ad}}}$ by (3.4). #### 3.0.7. The Lie algebra bundle $\mathcal{R}$ on $\bf X$ Let $\\{D_{\alpha}|\alpha\in S\\}$ denote the irreducible smooth boundary divisors of ${\bf X}$. Set $H_{{}_{\alpha}}:=D_{{}_{\alpha}}\cap{\bf Y}_{{}_{0}}$ for each $\alpha$. Recall that $Z:={\bf Y}_{{}_{0}}\setminus T_{{}_{\text{ad}}}$ is a union $\cup_{{}_{\alpha\in S}}H_{{}_{\alpha}}$ which are $\ell$ smooth hyperplanes meeting at simple normal crossings. For $\alpha\in S$, let $\zeta_{{}_{\alpha}}$’s denote the generic points of the divisors $H_{{}_{\alpha}}$’s. Let $A_{{}_{\alpha}}=\mathcal{O}_{{}_{{{\bf Y}_{{}_{0}}},\zeta_{{}_{\alpha}}}}$ (3.0.23) be the dvr’s obtained by localizing at the height $1$-primes given by the $\zeta_{{}_{\alpha}}$’s and let $Y_{{}_{\alpha}}:={\rm Spec}\,(A_{{}_{\alpha}})$. Base changing by the local morphism $Y_{{}_{\alpha}}\to\bf Y_{{}_{0}}$, we have a Lie algebra bundle $\mathcal{R}|_{{}_{Y_{{}_{\alpha}}}}$ for each $\alpha$. Moreover, the Lie algebra bundle $\mathcal{R}$ on ${\bf Y}_{{}_{0}}$ gives canonical gluing data to glue $\mathcal{R}|_{{}_{Y_{{}_{\alpha}}}}$ with the trivial bundle $\mathfrak{g}\times T_{{}_{\text{ad}}}$. Let $\xi_{{}_{\alpha}}\in D_{{}_{\alpha}},\alpha\in S$ denote the generic points of the divisors $D_{{}_{\alpha}}$’s and let $B_{{}_{\alpha}}:=\mathcal{O}_{{}_{{\bf X},\xi_{{}_{\alpha}}}}$ (3.0.24) be the dvr’s obtained by localizing at the height $1$-primes given by the $\xi_{{}_{\alpha}}$’s. By a transport of structures, for each $\alpha$ the gluing datum of $\mathcal{R}|_{{}_{Y_{{}_{\alpha}}\cap T_{{}_{\text{ad}}}}}$, gives a Lie-algebra bundle gluing datum on ${\rm Spec}\,(B_{{}_{\alpha}})\cap G_{{}_{\text{ad}}}$. Thus, the gluing data of the bundle $\mathcal{R}$ on ${\bf Y}_{{}_{0}}$ at the the $\zeta_{{}_{\alpha}}$’s can be now used to extend the trivial bundle $\mathfrak{g}\times G_{{}_{\text{ad}}}$ to the $\xi_{{}_{\alpha}}$’s as a locally free sheaf of Lie algebras. The rest of the proof is as for $\bf Y$, together with the observation that the $G\times G$-translates of the open subset $U\times U^{{}^{-}}\times{\bf Y}_{{}_{0}}$ cover $\bf X$, where the bundle is simply the pull-back from ${\bf Y}_{{}_{0}}$. Thus we have: ###### Corollary 3.9. There exists a canonical Lie-algebra bundle $\mathcal{R}_{{}_{\bf X}}$ on ${\bf X}$ which extends the trivial bundle with fiber $\mathfrak{g}$ on ${G_{{}_{\text{ad}}}}\subset{\bf X}$ with properties as in (3.3). #### 3.0.8. Bruhat-Tits group scheme ${\mathcal{G}}_{{}_{\bf X}}^{{}^{\varpi}}$ on $\bf X$ Let $L_{{}_{\alpha}}$ be the quotient field of $B_{{}_{\alpha}}$. Let $X_{{}_{\alpha}}:={\rm Spec}\,(B_{{}_{\alpha}})$. By (3.8), for each $\alpha$ there exist a $\Gamma_{{}_{\alpha}}$-equivariant $G$-torsor $E_{\alpha}$ on a ramified cover $q_{{}_{\alpha}}:X^{\prime}_{{}_{\alpha}}\to X_{{}_{\alpha}}$ such that the adjoint group scheme $\mathcal{H}_{{}_{\alpha}}=E_{\alpha}(G)$ has simply-connected fibers isomorphic to $G$ and we have the identification of $B_{{}_{\alpha}}$-group schemes: $\displaystyle\big{(}\text{Res}_{{}_{X^{\prime}_{{}_{\alpha}}/X_{{}_{\alpha}}}}(\mathcal{H}_{{}_{\alpha}})\big{)}^{{}^{\Gamma_{{}_{\alpha}}}}\simeq\mathcal{G}^{{}^{st}}_{{}_{\theta{{}_{{}_{\alpha}}}}}.$ (3.0.25) Before stating the main result of this section, we make a few remarks which might help the reader. The basic underlying principle in these constructions is that the combinatorial data encoded in the triple consisting of the Weyl chamber, the fan of Weyl chambers, and the Tits building, is geometrically replicated by the inclusion ${\bf Y_{{}_{0}}}\subset{\bf Y}\subset\bf X$. The "wonderful" Bruhat-Tits group scheme which arises on $\bf X$ has its local Weyl-chamber model on the affine toric variety $\bf Y_{{}_{0}}$. Indeed, in this case, the Kawamata cover is even explicit (3.5). In particular, the theorem below, can be executed for ${\bf Y}_{{}_{0}}$ but this will give the group scheme associated to the data coming from the Weyl chamber alone. ###### Theorem 3.10. There exists an affine “wonderful" Bruhat-Tits group scheme ${\mathcal{G}}_{{}_{\bf X}}^{{}^{\varpi}}$ on $\bf X$ satisfying the following classifying properties. 1. (1) There is an identification of the Lie-algebra bundles $\text{Lie}({\mathcal{G}}_{{}_{\bf X}}^{{}^{\varpi}})\simeq\mathcal{R}_{{}_{\bf X}}$. 2. (2) For $\emptyset\neq I\subset S$ the restriction of ${\mathcal{G}}_{{}_{{\bf X}}}^{{}^{\varpi}}$ to the formal neighbourhood $U_{{}_{z_{{}_{I}}}}$ of $z_{{}_{I}}$ in $C_{{}_{I}}$ §(3.0.3) is isomorphic to the standard Bruhat- Tits group scheme $\mathcal{G}^{{}^{st}}_{{}_{I}}$ §(2.0.5). ###### Proof. Let $X_{{}_{\alpha}}$ be as above. Note that we can identify the open subset ${\rm Spec}\,(L_{{}_{\alpha}})$ with ${G_{{}_{\text{ad}}}}\cap X_{{}_{\alpha}}$. Set $X^{\prime}:=G_{{}_{\text{ad}}}\cup_{\alpha\in S}X_{{}_{\alpha}}.$ (3.0.26) Hence ${\bf X}\setminus X^{\prime}$ is a colimit of closed subschemes of $\bf X$ of codimension at least $2$. Consider the fpqc morphism: $G_{{}_{{}_{\text{ad}}}}\bigsqcup_{\alpha\in S}X_{{}_{\alpha}}\to{X^{\prime}}.$ (3.0.27) We restrict $\mathcal{R}=\mathcal{R}_{{}_{\bf X}}$ further to $X^{\prime}$. By (3.9) , over the open subset $G_{{}_{\text{ad}}}\subset X^{\prime}$, we have $\mathcal{R}\simeq G_{ad}\times\mathfrak{g}$. Further, the transition functions of $\mathcal{R}$ on the intersections ${\rm Spec}\,(L_{{}_{\alpha}}):=G_{{}_{{}_{\text{ad}}}}\cap{\rm Spec}\,(B_{{}_{\alpha}})$ take values in $\text{Aut}(\mathfrak{g})$. Since $\text{Aut}(\mathfrak{g})=\text{Aut}(G)$, it follows that the effective descent datum provided by pulling back $\mathcal{R}$ to $G_{{}_{{}_{\text{ad}}}}\bigsqcup_{\alpha\in S}X_{{}_{\alpha}}$ gives a descent datum to glue the trivial group scheme $G\times G_{{}_{{}_{\text{ad}}}}$ on $G_{{}_{{}_{\text{ad}}}}$ with $\mathcal{G}^{{}^{st}}_{{}_{\alpha}}$ on ${\rm Spec}\,(B_{{}_{\alpha}})$ along ${\rm Spec}\,(L_{{}_{\alpha}})$. Since the group schemes are affine, this descent datum is also effective. In other words, we get the group scheme $\mathcal{G}^{\circ}\rightarrow X^{\prime}.$ (3.0.28) This can also be seen without “descent" theory as follows. Since the group schemes $\mathcal{G}^{{}^{st}}_{{}_{\alpha}}$ on ${\rm Spec}\,(B_{{}_{\alpha}})$ are of finite type, they can be extended to an affine subscheme $\bf X_{{}_{f}}$. By a further shrinking of this neighbourhood of $\xi_{{}_{\alpha}}$ one can glue it to $G\times G_{{}_{{}_{\text{ad}}}}$ along the intersection. So we can think of $X^{\prime}$ as an honest open subset of $X$ with complement of codimension at least $2$. By (3.8), the Lie algebra bundle $\mathcal{R}$ gets canonical parabolic structures at the generic points $\xi_{{}_{\alpha}}$ of the divisors $\\{D_{{}_{\alpha}}\\}_{{}_{\alpha\in S}}$. For a parabolic vector bundle with prescribed rational weights such as $\mathcal{R}$, by [2] (see §8) we get the following data: * • a global Kawamata cover (8.0.2) $p:Z\rightarrow{\bf X}$ ramified over $D$ with ramification prescribed by the weights $\\{d_{{}_{\alpha}}\\}$ (2.0.3), with Galois group $\Gamma$ which “realizes the local ramified covers $q_{{}_{\alpha}}$ at the points $\xi_{{}_{\alpha}}$" i.e. the isotropy subgroup of $\Gamma$ at $\xi_{{}_{\alpha}}$ is $\Gamma_{{}_{\alpha}}$ and * • an equivariant vector bundle $V$ on $Z$ such that $p_{{}_{*}}^{\Gamma}(V)\simeq\mathcal{R}.$ (3.0.29) Let $Z^{\prime}:=p^{-1}(X^{\prime})$ and $V^{\prime}:=V|_{{}_{Z^{\prime}}}$. Gluing the trivial $G$-torsor with the $\\{E_{\alpha}\\}$ by the transition functions of $V^{\prime}$, we make a $\Gamma$-equivariant principal $\text{Aut}(G)$-torsor $\mathcal{E}^{\circ}$. Its associated group scheme $\mathcal{E}^{\circ}\times^{{}^{\text{Aut}(G)}}G=\mathcal{H}^{\circ}$ is a now group scheme on $Z^{\prime}$ with simply-connected fibers $G$, albeit with transition functions in $\text{Aut}(G)$. Moreover, we have $\big{(}\text{Res}_{{}_{Z^{\prime}/X^{\prime}}}(\mathcal{H}^{\circ})\big{)}^{{}^{\Gamma}}\simeq\mathcal{G}^{\circ}.$ (3.0.30) Further, as locally-free sheaves, we also have $\displaystyle\text{Lie}(\mathcal{H}^{\circ})=V^{\prime}.$ (3.0.31) We transport the structure of Lie-bracket from $\text{Lie}(\mathcal{H}^{\circ})$ to $V^{\prime}$ on $Z^{\prime}$. This is non-degenerate everywhere on $Z^{\prime}$ since $\mathcal{H}^{\circ}$ is the adjoint group scheme of $\mathcal{E}^{\circ}$. Hence $V^{\prime}$ has fiber type $\mathfrak{g}$. Since $\text{codim}(Z\setminus Z^{\prime})\geq 2$, the Lie bracket $[.,.]$ on $V^{\prime}$ extends to a Lie bracket on the locally free sheaf $V$ with the Killing form being non-degenerate on the whole of $Z$. In other words $V$ is now a locally free sheaf of Lie algebras on the whole of $Z$ with semisimple fibres; these fibres are isomorphic to the Lie algebra $\mathfrak{g}$ by the rigidity of semisimple Lie algebras. We now wish to extend $\mathcal{E}^{\circ}$ as an equivariant $\text{Aut}(G)$-torsor $\mathcal{E}$ on the whole of $Z$ such that its associated Lie-algebra bundle $\mathcal{E}(\mathfrak{g}):=\mathcal{E}\times^{{}^{\text{Aut}(G)}}\mathfrak{g}$ becomes isomorphic $V$. Since $G$ is simply-connected, making the identification $\text{Aut}(\mathfrak{g})=\text{Aut}(G)$, we see that the transition functions of $V$ give the gluing for defining $\mathcal{E}$ on $Z$ satisfying our desired requirements. Let $\mathcal{H}:=\mathcal{E}\times^{{}^{\text{Aut}(G)}}G$ denote the group scheme associated to the adjoint group scheme of $\mathcal{E}$. This is an equivariant group scheme on $Z$ and we define $\displaystyle{\mathcal{G}}_{{}_{\bf X}}^{{}^{\varpi}}:=\big{(}\text{Res}_{{}_{Z/{\bf X}}}(\mathcal{H})\big{)}^{{}^{\Gamma}}.$ (3.0.32) Then ${\mathcal{G}}_{{}_{\bf X}}^{{}^{\varpi}}|_{{X^{\prime}}}=\mathcal{G}^{\circ}$. Since by Lemma (3.7), the functor “invariant direct image" commutes with taking Lie algebras, we moreover get isomorphisms of locally-free sheaves of Lie algebras $\text{Lie}({\mathcal{G}}_{{}_{\bf X}}^{{}^{\varpi}})\simeq p_{{}_{*}}^{\Gamma}(\mathcal{E}(\mathfrak{g}))\simeq\mathcal{R}.$ (3.0.33) This proves the first claim in the theorem. Finally, let us verify the classifying property of the group scheme $\mathcal{H}$. For $\alpha\in S$, at the closed points $z_{{}_{\alpha}}$ of the curves ${\rm Spec}\,(A_{{}_{\alpha}})$ we have $\mathcal{H}|{{}_{{\rm Spec}\,(A_{{}_{\alpha}})}}\simeq\mathcal{H}_{{}_{\alpha}}$. The classifying property is tautologically valid here because this was designed expressly in (3.0.25). For the closed points $z_{{}_{\lambda}}$ of $C_{{}_{\lambda}}$ corresponding to strata of lower dimension, we proceed as follows. Consider the base change of the Kawamata cover $p:Z\rightarrow{\bf X}$ to the curve $C_{{}_{\lambda}}\subset{\bf X}$ and further to the formal neighbourhood $U_{{}_{z_{{}_{\lambda}}}}\subset C_{{}_{\lambda}}$ of $z_{{}_{\lambda}}$. Let $p_{{}_{z}}:W_{{}_{z}}\to U_{{}_{z}}$ be the restriction of $p$ to a connected component of $Z\times_{{}_{U_{{}_{z}}}}{\bf X}$. Then, $p_{{}_{z}}$ gives a Galois cover with Galois group some cyclic group $\mu_{{}_{d}}\subset\Gamma$ of order $d$. By (3.0.32), the restriction ${\mathcal{G}}_{{}_{{\bf Y}}}^{{}^{\varpi}}|_{{}_{U_{{}_{z}}}}$ is the “invariant direct image" of $\mathcal{H}|_{{}_{W_{{}_{z}}}}$. Also, we have the isomorphism $\mathcal{R}|_{{}_{U_{{}_{z}}}}\simeq p_{{}_{*}}^{\mu_{{}_{d}}}(\mathcal{E}(\mathfrak{g})_{{}_{W_{{}_{z}}}})$. Further, by (3.8), $L^{+}(\mathcal{R}|_{{}_{U_{{}_{z_{{}_{\lambda}}}}}})\simeq L^{+}(\mathfrak{P}^{{}^{std}}_{{}_{\theta_{{}_{\lambda}}}})$. By (3.8), we have another ramified cover $q_{{}_{\lambda}}:U^{\prime}_{{}_{z}}\to U_{{}_{z}}$, but now with Galois group $\Gamma_{{}_{\lambda}}$ and an equivariant $(\Gamma_{{}_{\lambda}},G)$-torsor $E_{{}_{\lambda}}$ on $U^{\prime}_{{}_{z}}$, such that $q^{{}^{\Gamma_{{}_{\lambda}}}}_{{}_{*}}(E_{{}_{\lambda}}(G))\simeq\mathcal{G}^{{}^{st}}_{{}_{\theta_{{}_{\lambda}}}}$ and $q^{{}^{\Gamma_{{}_{\lambda}}}}_{{}_{*}}(E_{{}_{\lambda}}(\mathfrak{g}))\simeq\mathcal{R}|_{{}_{U_{{}_{z_{{}_{\lambda}}}}}}$. To finish the proof, we need to show the following isomorphism of group schemes: ${\mathcal{G}}_{{}_{{\bf X}}}^{{}^{\varpi}}|_{{}_{U_{{}_{z}}}}\simeq\mathcal{G}^{{}^{st}}_{{}_{\theta_{{}_{\lambda}}}}.$ (3.0.34) Over a common cover of $U^{\prime}_{{}_{z}}$ and $W_{{}_{z}}$ (which will continue to be a Kawamata cover of $U_{{}_{z}}$), we can identify the pull- back of Lie sheaves $E_{{}_{\lambda}}(\mathfrak{g})$ and $\mathcal{E}(\mathfrak{g})|_{{}_{W_{{}_{z}}}}$ as equivariant Lie sheaves since both give invariant direct images isomorphic to $\mathcal{R}|_{{}_{U_{{}_{z_{{}_{\lambda}}}}}}$. Therefore, on the same common cover we have an identification of pull-backs of $E_{{}_{\lambda}}\times^{G}\mathfrak{g}$ with $\mathcal{E}(\mathfrak{g})|_{{}_{W_{{}_{z}}}}=\mathcal{E}\times^{Aut(G)}\mathfrak{g}|_{{}_{W_{{}_{z}}}}$. Hence we have an identification of pull-backs of $E_{{}_{\lambda}}\times^{G}G$ with $\mathcal{E}\times^{Aut(G)}G|_{{}_{W_{{}_{z}}}}$ as equivariant group schemes with simply-connected fibers. Now the invariant direct image of the first group scheme is a standard parahoric group scheme because the curve $C_{{}_{\lambda}}$ is standard. On the other hand, the second group scheme is $\mathcal{H}|_{{}_{W_{{}_{z}}}}$ whose invariant direct image by construction is ${\mathcal{G}}_{{}_{{\bf X}}}^{{}^{\varpi}}|_{{}_{U_{{}_{z}}}}$. Thus, we have proven (3.0.34). ∎ ###### Remark 3.11. The above construction goes through without change over an algebraically closed field $k$ of characteristic $p$ coprime to the $d_{{}_{\alpha}}$’s (2.0.3). The existence of $\bf X$ is known from the works of Strickland [17] and De Concini-Springer [8] and Kawamata covering works under the above conditions on characteristics. ## 4\. The Weyl alcove and apartment case We continue to use the notations as in previous sections. Let $G^{{}^{\text{aff}}}$ denote the Kac-Moody group associated to the affine Dynkin diagram of $G$. Recall that $G^{{}^{\text{aff}}}$ is given by a central extension of $L^{\ltimes}G$ by $\mathbb{G}_{m}$. Analogous to the wonderful compactification of $G_{{}_{{}_{\text{ad}}}}$, P. Solis in [16] has constructed a wonderful embedding ${\bf X}^{{}^{\text{aff}}}$ for $G^{{}^{\text{aff}}}_{{}_{\text{ad}}}:=G^{{}^{\text{aff}}}/Z(G^{{}^{\text{aff}}})=\mathbb{G}_{m}\ltimes LG/Z(G)$. It is an ind-scheme containing $G^{{}^{\text{aff}}}_{{}_{\text{ad}}}$ as a dense open ind-scheme and carrying an equivariant action of $L^{\ltimes}G\times L^{\ltimes}G$. Let $T_{{}_{\text{ad}}}:=T/Z(G)$, $T^{\ltimes}_{{}_{\text{ad}}}:=\mathbb{G}_{m}\times T_{{}_{\text{ad}}}\subset G^{{}^{\text{aff}}}_{{}_{\text{ad}}}$, where $\mathbb{G}_{m}$ is the rotational torus. In ${\bf X}^{{}^{\text{aff}}}$, the closure ${\bf Y}^{{}^{\text{aff}}}:=\overline{T^{\ltimes}_{{}_{\text{ad}}}}$ gives a torus- embedding. It is covered by the affine Weyl group $W^{{}^{\text{aff}}}$-translates of the affine torus embedding ${\bf Y}_{{}_{0}}^{{}^{\text{aff}}}:=\overline{T^{\ltimes}_{{}_{\text{ad},0}}}\simeq\mathbb{A}^{\ell+1}$ given by the negative Weyl alcove. #### 4.0.1. On the torus-embedding ${\bf Y}_{{}_{0}}^{{}^{\text{aff}}}$ Recall that $Z:={\bf Y}_{{}_{0}}^{{}^{\text{aff}}}\setminus T^{\ltimes}_{{}_{\text{ad}}}$ is a union $\cup_{{}_{\alpha\in\mathbb{S}}}H_{{}_{\alpha}}$ of $\ell+1$ many standard coordinate hyperplanes meeting at normal crossings. For $\alpha\in\mathbb{S}$, let the $\zeta_{{}_{\alpha}}$’s denote the generic points of the divisors $H_{{}_{\alpha}}$’s. Let $A_{{}_{\alpha}}=\mathcal{O}_{{}_{{\bf Y}_{{}_{0}}^{{}^{\text{aff}}},\zeta_{{}_{\alpha}}}}$ (4.0.1) be the dvr’s obtained by localizing at the height $1$-primes given by the $\zeta_{{}_{\alpha}}$’s. Let $K_{{}_{\alpha}}$ be the quotient field of $A_{{}_{\alpha}}$. Let $Y_{{}_{\alpha}}:={\rm Spec}\,(A_{{}_{\alpha}})$. Note that we can identify the open subset ${\rm Spec}\,(K_{{}_{\alpha}})$ with ${T^{\ltimes}_{{}_{\text{ad}}}}\cap Y_{{}_{\alpha}}$. Let $Y^{\prime}:={T^{\ltimes}_{{}_{\text{ad}}}}\cup_{\alpha}Y_{{}_{\alpha}}.$ (4.0.2) The complement ${\bf Y}_{{}_{0}}^{{}^{\text{aff}}}\setminus Y^{\prime}$ can again be realized as a colimit of open subsets of ${\bf Y}_{{}_{0}}^{{}^{\text{aff}}}$ whose codimension is at least $2$ in ${\bf Y}_{{}_{0}}^{{}^{\text{aff}}}$. #### 4.0.2. Construction of a finite-dimensional Lie algebra bundle $J$ on ${\bf Y}_{{}_{0}}^{{}^{\text{aff}}}$ together with parabolic structures This construction is exactly analogous to the construction of $\mathcal{R}$ on $\bf Y$ in §3.0.3. We let $T^{\ltimes}_{{}_{\text{ad}}}$ and $\mathbb{S}$ play the role of $T_{{}_{\text{ad}}}$ and $S$. More precisely, let $\lambda=\sum_{\alpha\in\mathbb{S}}k_{{}_{\alpha}}\omega_{{}_{\alpha}}^{{\vee}}$ be a non-zero dominant $1$-PS of $T_{{}_{\text{ad}}}^{\ltimes}$. We set $\eta_{\lambda}:=\sum_{\alpha\in\mathbb{S}}k_{{}_{\alpha}}\frac{(1,\theta_{{}_{\alpha}})}{\ell+1}\quad\text{and}\quad\theta_{\lambda}:=\sum_{\alpha\in\mathbb{S}}\frac{k_{{}_{\alpha}}}{\sum_{{}_{\alpha\in\mathbb{S}}k_{{}_{\alpha}}}}\theta_{{}_{\alpha}}$ (4.0.3) and $C_{{}_{\lambda}}$ the curve in ${\bf Y}^{{}^{\text{aff}}}$ defined by $\eta_{\lambda}$ and $U_{{}_{\lambda}}$ the formal neighbourhood of the closed point of $z_{{}_{\lambda}}$ of $C_{{}_{\lambda}}$. So if we substitute the rational number $a$ of §2.0.3 by the number $\frac{\sum_{\alpha\in\mathbb{S}}k_{{}_{\alpha}}}{\ell+1}$, we have the parahoric group identification $\mathfrak{P}_{{}_{\eta_{\lambda}}}=\mathfrak{P}_{{}_{\theta_{\lambda}}}.$ (4.0.4) We may prove the following theorem exactly like (3.3). There we viewed the standard alcove as a cone over the far wall. Note here that unlike the situation in (3.3), we do not expect to get standard parahoric structures down all the strata. However, for strata contained in the divisor associated to $\alpha_{{}_{0}}\in\mathbb{S}$, the parahorics which occur will be standard as before. ###### Theorem 4.1. There exists a canonical Lie-algebra bundle $J$ on ${\bf Y}_{{}_{0}}^{{}^{\text{aff}}}$ which extends the trivial bundle with fiber $\mathfrak{g}$ on ${T^{\ltimes}_{{}_{\text{ad}}}}\subset{\bf Y}^{{}^{\text{aff}}}$ and such that for $k_{{}_{\alpha}}\in\\{0,1\\}$ we have the identification of functors from the category of $k$-algebras to $k$-Lie- algebras: $L^{+}(\mathfrak{P}_{{}_{\eta_{{}_{\lambda}}}})=L^{+}(\mathcal{R}\mid_{{}_{U_{{}_{\lambda}}}}).$ (4.0.5) ###### Corollary 4.2. The Lie algebra bundle $J_{{}_{{\bf Y}^{{}^{\text{aff}}}}}$ gets canonical parabolic structures (§8) at the generic points $\xi_{{}_{\alpha}}$ of the $W^{{}^{\text{aff}}}$-translates of the divisors $H_{{}_{\alpha}}\subset{\bf Y}_{{}_{0}}^{{}^{\text{aff}}}$, $\alpha\in\mathbb{S}$. ###### Proof. We prescribe ramification indices $d_{\alpha}$ (2.0.3) on the divisor $H_{{}_{\alpha}}$. Then as in the proof of (3.6), the identification (3.0.22) of the Lie algebra structures of $J_{{}_{{\bf Y}^{{}^{\text{aff}}}}}$ and the parahoric Lie algebra structures on the localizations of the generic points of $H_{{}_{\alpha}}$ allows us to endow parabolic structures at the generic points of the divisors.∎ #### 4.0.3. The parahoric group scheme on the torus embedding ${{\bf Y}^{{}^{\text{aff}}}}$ ###### Theorem 4.3. There exists an affine “wonderful" Bruhat-Tits group scheme ${\mathcal{G}}_{{}_{{\bf Y}^{{}^{\text{aff}}}}}^{{}^{\varpi}}$ on $\bf Y$ together with a canonical isomorphism $\text{Lie}({\mathcal{G}}_{{}_{{\bf Y}^{{}^{\text{aff}}}}}^{{}^{\varpi}})\simeq J$. It further satisfies the following classifying property: For any point $h\in{{\bf Y}^{{}^{\text{aff}}}}\setminus T^{\ltimes}_{ad}$, let $\mathbb{I}\subset\mathbb{S}$ be a subset such that $h\in\cap_{\alpha\in\mathbb{I}}H_{{}_{\alpha}}$. Let $C_{{}_{\mathbb{I}}}\subset{{\bf Y}^{{}^{\text{aff}}}}$ be a smooth curve with generic point in $T^{\ltimes}_{{}_{\text{ad}}}$ and closed point $h$. Let $U_{{}_{h}}\subset C_{{}_{\mathbb{I}}}$ be a formal neighbourhood of the closed point $h$. Then, the restriction ${\mathcal{G}}_{{}_{{\bf Y}^{{}^{\text{aff}}}}}^{{}^{\varpi}}|_{{}_{U_{{}_{h}}}}$ is isomorphic to the Bruhat-Tits group scheme $\mathcal{G}_{{}_{\mathbb{I}}}$ on $U_{{}_{h}}$. ###### Proof. Recall that ${{\bf Y}^{{}^{\text{aff}}}}$ is covered by affine spaces $Y_{{}_{w}}\simeq\mathbb{A}^{\ell+1}$ parametrized by the affine Weyl group $W^{{}^{\text{aff}}}$. Each $Y_{{}_{w}}$ is the translate of ${{\bf Y}_{{}_{0}}^{{}^{\text{aff}}}}$. The translates of the divisors $H_{{}_{\alpha}}$ meet each $Y_{{}_{w}}$ in the standard hyperplanes on $\mathbb{A}^{\ell+1}$ and thus we can prescribe the same ramification data at the hyperplanes on each of the $Y_{{}_{w}}$’s. On the other hand, although we have simple normal crossing singularities, we do not have an analogue of the Kawamata covering lemma for schemes such as ${{\bf Y}^{{}^{\text{aff}}}}$. The lemma is known only in the setting of quasi-projective schemes. So to construct the group scheme, we employ a different approach. We observe firstly that, the formalism of Kawamata coverings applies in the setting of the affine spaces $Y_{{}_{w}}\subset{{\bf Y}^{{}^{\text{aff}}}}$. Let $p_{w}:Z_{{}_{w}}\rightarrow Y_{{}_{w}}$ be the associated Kawamata cover (8.0.2) with Galois group $\Gamma_{w}$. In Corollary 4.2 we observed that the Lie-algebra bundle $J$ has a canonical parabolic structure like $\mathcal{R}$ in the proof of Theorem 3.10. Letting $Y_{{}_{w}}$ play the role of ${\bf Y_{{}_{0}}}$ and using all arguments in the proof of Theorem 3.10 we obtain $\mathcal{H}_{{}_{w}}\rightarrow Z_{{}_{w}}$ which is $\Gamma_{w}$-group scheme with fibers isomorphic to $G$ whose invariant direct image is a group scheme $\mathcal{G}_{w}$ such that $\text{Lie}(\mathcal{G}_{w})=J|_{Y_{{}_{w}}}$. The induced parabolic structure on $J|_{Y_{{}_{w}}}$ is the restriction of the one on $J$. Indeed, by (4.2) these parabolic structures are essentially given at the local rings at the generic points $\xi_{\alpha}$ of the divisors $H_{{}_{\alpha}}$ and hence these parabolic structures on $J$ agree on the intersections $Y_{{{}_{uv}}}:=Y_{{}_{u}}\cap Y_{{}_{v}}$. Let $Z_{{}_{uv}}:=p_{u}^{-1}(Y_{{}_{uv}})$. Let $\tilde{Z}_{{}_{uv}}$ be the normalization of a component of $Z_{{}_{uv}}\times_{Y_{{}_{uv}}}Z_{{}_{vu}}$. Then $\tilde{Z}_{{}_{uv}}$ serves as Kawamata cover (8.0.2) of $Y_{{}_{uv}}$ (see [18, Corollary 2.6, page 56]). We consider the morphisms $\tilde{Z}_{{}_{uv}}\to Z_{{}_{u}}$ (resp.$\tilde{Z}_{{}_{uv}}\to Z_{{}_{v}}$) and let $\mathcal{H}_{{}_{u,\tilde{Z}}}$ (resp. $\mathcal{H}_{{}_{v,\tilde{Z}}}$) denote the pull-backs of $\mathcal{H}_{{}_{u}}$ (resp. $\mathcal{H}_{{}_{v}}$) to $\tilde{Z}_{{}_{uv}}$. Let $\Gamma$ denote the Galois group for $\tilde{Z}_{{}_{uv}}\rightarrow Y_{{}_{uv}}$. Then by Lemma 3.7, the invariant direct images of the equivariant Lie algebra bundles $\text{Lie}(\mathcal{H}_{{}_{u,\tilde{Z}}})$ and $\text{Lie}(\mathcal{H}_{{}_{v,\tilde{Z}}})$ coincide with the Lie algebra structure on $J$ restricted to the ${Y_{{}_{uv}}}$ and also as isomorphic parabolic bundles. Therefore we have a natural isomorphism of equivariant Lie- algebra bundles $\text{Lie}(\mathcal{H}_{u,\tilde{Z}})\simeq\text{Lie}(\mathcal{H}_{v,\tilde{Z}}).$ (4.0.6) As in the proof of Theorem 3.10, this gives a canonical identification of the equivariant group schemes $\mathcal{H}_{{}_{u,\tilde{Z}}}$ and $\mathcal{H}_{{}_{v,\tilde{Z}}}$ on $\tilde{Z}_{{}_{uv}}$. Since the invariant direct image of both the group schemes $\mathcal{H}_{{}_{u,\tilde{Z}}}$ and $\mathcal{H}_{{}_{v,\tilde{Z}}}$ are the restrictions $\mathcal{G}_{{}_{u,{Y_{{}_{uv}}}}}$ and $\mathcal{G}_{{}_{v,{Y_{{}_{uv}}}}}$, it follows that on $Y_{{}_{uv}}=Y_{{}_{u}}\cap Y_{{}_{v}}$ we get a canonical identification of group schemes: $\displaystyle\mathcal{G}_{{}_{u,{Y_{{}_{uv}}}}}\simeq\mathcal{G}_{{}_{v,{Y_{{}_{uv}}}}}.$ (4.0.7) These identifications are canonically induced from the gluing data of the Lie algebra bundle $J$ for the cover $Y_{{}_{w}}$’s. Therefore the cocycle conditions are clearly satisfied and the identifications (4.0.7) glue to give the group scheme ${\mathcal{G}}_{{}_{{\bf Y}^{{}^{\text{aff}}}}}^{{}^{\varpi}}$ on ${\bf Y}^{{}^{\text{aff}}}$. The verification of the classifying property follows exactly as in the proof of Theorem 3.10. ∎ ## 5\. The Bruhat-Tits group scheme on ${\bf X}^{{}^{{aff}}}$ Let ${\bf X}^{{}^{{aff}}}$ as in §4. We begin with a generality. Let $\mathbb{X}$ be an ind-scheme. By an open-subscheme $i:\mathbb{U}\hookrightarrow\mathbb{X}$ we mean an ind-scheme such that for any $f:{\rm Spec}\,(A)\rightarrow\mathbb{X}$, the natural morphism $\mathbb{U}\times_{\mathbb{X}}{\rm Spec}\,(A)\rightarrow{\rm Spec}\,(A)$ is an open immersion. For a sheaf $\mathbb{F}$ on $\mathbb{U}$ by $i_{{}_{*}}(\mathbb{F})$ we mean the sheaf associated to the pre-sheaf on the “big site" of $\mathbb{X}$, whose sections on $f:{\rm Spec}\,(A)\rightarrow\mathbb{X}$ are given by $\mathbb{F}(\mathbb{U}\times_{\mathbb{X}}{\rm Spec}\,(A))$. The ind-scheme ${\bf X}^{{}^{\text{aff}}}$ has a certain open subset ${\bf X}_{{}_{0}}$ whose precise definition is somewhat technical [16, Page 705]. Let us mention the properties relevant for us. Recall that ${\bf X}^{{}^{\text{aff}}}=(G_{{}_{\text{ad}}}^{{}^{\text{aff}}}\times G_{{}_{\text{ad}}}^{{}^{\text{aff}}})~{}{\bf X}_{{}_{0}}$ and in fact ${\bf X}^{{}^{\text{aff}}}=(G_{{}_{\text{ad}}}^{{}^{\text{aff}}}\times G_{{}_{\text{ad}}}^{{}^{\text{aff}}})~{}{{\bf Y}^{{}^{\text{aff}}}}$. Further, the torus-embedding ${\bf Y}^{{}^{\text{aff}}}$ is covered by ${\bf Y}^{{}^{\text{aff}}}_{{}_{w}}\simeq\mathbb{A}^{{}^{\ell+1}}$ which are $W^{{}^{\text{aff}}}$-translates of ${\bf Y}^{{}^{\text{aff}}}_{{}_{0}}$, where ${\bf Y}^{{}^{\text{aff}}}_{{}_{0}}={{\bf Y}^{{}^{\text{aff}}}}\cap{\bf X}_{{}_{0}}$. We remark that, analogous to the case of ${\bf Y}_{{}_{0}}\subset{\bf Y}\subset{\bf X}$, just as the toric variety ${\bf Y}_{{}_{0}}$ was associated to the negative Weyl chamber, the toric variety ${\bf Y}^{{}^{\text{aff}}}_{{}_{0}}$ is associated to the negative Weyl alcove. So let us denote by $0$ the neutral element of $W^{{}^{\text{aff}}}$ also. Let $U^{\pm}\subset B^{\pm}$ be the unipotent subgroups. Let $\mathcal{U}^{\pm}:=ev^{-1}(U^{\pm})$ where $ev:G(A)\rightarrow G(k)$ is the evalution map. Further by [16, Prop 5.3] ${\bf X}_{{}_{0}}=\mathcal{U}\times{\bf Y}^{{}^{\text{aff}}}_{{}_{0}}\times\mathcal{U}^{-}.$ (5.0.1) Let ${\bf X}_{{}_{w}}:=\mathcal{U}\times{\bf Y}^{{}^{\text{aff}}}_{{}_{w}}\times\mathcal{U}^{-}$. These cover $\mathcal{U}\times{{\bf Y}^{{}^{\text{aff}}}}\times\mathcal{U}^{-}$. For $g\in G_{{}_{\text{ad}}}^{{}^{\text{aff}}}\times G_{{}_{\text{ad}}}^{{}^{\text{aff}}}$, let ${\bf X}_{g}:=g{\bf X}_{0}\quad{\bf X}_{{}_{g,w}}:=g{\bf X}_{{}_{w}}\quad{\bf Y}^{{}^{\text{aff}}}_{{}_{g,w}}:=g{\bf Y}^{{}^{\text{aff}}}_{{}_{w}}.$ (5.0.2) Note that we have the projection ${\bf X}_{{}_{g,w}}\to{\bf Y}^{{}^{\text{aff}}}_{{}_{g,w}}$ which is a $\mathcal{U}\times\mathcal{U}^{-}$-bundle. By [16, Theorem 5.1] the ind-scheme ${\bf X}^{{}^{\text{aff}}}$ has divisors $D_{\alpha}$ for $\alpha\in\mathbb{S}$ such that the complement of their union is ${\bf X}^{{}^{\text{aff}}}\setminus G_{{}_{\text{ad}}}^{{}^{\text{aff}}}$. The next proposition shows the existence of a finite dimensional Lie algebra bundle on ${\bf X}^{{}^{\text{aff}}}$ which is analogous to the bundle $\mathcal{R}$ over ${\bf X}$. ###### Proposition 5.1. There is a finite dimensional Lie algebra bundle $\bf R$ on ${\bf X}^{{}^{\text{aff}}}$ which extends the trivial Lie algebra bundle $G_{{}_{\text{ad}}}^{{}^{\text{aff}}}\times\mathfrak{g}$ on the open dense subset $G_{{}_{\text{ad}}}^{{}^{\text{aff}}}\subset{\bf X}^{{}^{\text{aff}}}$ and whose restriction to ${\bf Y}^{{}^{\text{aff}}}$ is $J$. ###### Proof. Following the arguments for (3.6) and (3.9), let $J^{\prime}$ be the locally free sheaf obtained on the open subset ${\bf X^{\prime}}$ obtained by the union of $G^{{}^{\text{aff}}}$ and the height one prime ideals of ${\bf X}^{{}^{\text{aff}}}$. Let ${\bf R}$ denote its push-forward. To check that the push-forward is locally free, without loss of generality we may consider its restriction to ${\bf X}_{0}$. But on ${\bf X}_{0}$, the pushforward of $J^{\prime}$ restricts to the pull-back of a Lie-algebra bundle $J$ on ${\bf Y}^{{}^{\text{aff}}}$ constructed in (4.1), which completes the argument. The rest of the properties follows immediately. ∎ ###### Theorem 5.2. There exists an affine “wonderful" Bruhat-Tits group scheme ${\mathcal{G}}_{{}_{{\bf X}^{{}^{\text{aff}}}}}^{{}^{\varpi}}$ on ${\bf X}^{{}^{\text{aff}}}$ together with a canonical isomorphism $\text{Lie}({\mathcal{G}}_{{}_{\bf X^{aff}}}^{{}^{\varpi}})\simeq{\bf R}$. It further satisfies the following classifying property: For any point $h\in{\bf X}^{{}^{\text{aff}}}\setminus G_{{}_{\text{ad}}}^{{}^{\text{aff}}}$, let $\mathbb{I}\subset\mathbb{S}$ be defined by the condition $h\in\cap_{\alpha\in\mathbb{I}}D_{\alpha}$. Let $C_{{}_{\mathbb{I}}}\subset{\bf X}$ be a smooth curve with generic point in $G_{{}_{\text{ad}}}^{{}^{\text{aff}}}$ with closed point $h$. Let $U_{{}_{h}}\subset C_{{}_{\mathbb{I}}}$ be a formal neighbourhood of $h$. Then, the restriction ${\mathcal{G}}_{{}_{{\bf X}^{{}^{\text{aff}}}}}^{{}^{\varpi}}|_{{}_{U_{{}_{h}}}}$ is isomorphic to the Bruhat-Tits group scheme $\mathcal{G}_{{}_{\mathbb{I}}}$ on ${\rm Spec}\,(A)$. ###### Proof. We begin by observing that ${\bf X}_{{}_{g,w}}$ and ${\bf R}\rightarrow{\bf X}^{{}^{\text{aff}}}$ play the role of ${{\bf Y}_{{}_{g,w}}^{{}^{\text{aff}}}}$ and $J\rightarrow{{\bf Y}^{{}^{\text{aff}}}}$ in the proof of Theorem 4.3. Therefore the group scheme $\mathcal{G}_{{}_{g,w}}$ glue together and we obtain the global group scheme ${\mathcal{G}}_{{}_{{\bf X}^{{}^{\text{aff}}}}}^{{}^{\varpi}}$. The verification of the classifying property follows exactly as in the proof of Theorem 3.10. ∎ ## 6\. The Bruhat-Tits group scheme on ${\bf X}^{{}^{{poly}}}$ In [16], apart from the loop group case Solis also discusses the polynomial loop group situation which in fact is relatively simpler. Consequently, the analysis as well as the results in the loop group situation go through with some simplifications. Let $L_{{}_{\text{poly}}}G$ be defined by $L_{{}_{\text{poly}}}G(R^{\prime})=G(R^{\prime}[z^{{}^{\pm}}])$ (6.0.1) and we define $L_{{}_{\text{poly}}}^{\ltimes}G$ similarly as $L_{{}_{\text{poly}}}^{\ltimes}G:=\mathbb{G}_{m}\ltimes L_{{}_{\text{poly}}}G$. Solis constructs a “wonderful" embedding of the polynomial loop group $L_{{}_{\text{poly}}}^{\ltimes}G/Z(G)$. We denote this space by ${\bf X}^{{}^{\text{poly}}}$. It has an open subset ${\bf X}^{{}^{\text{poly}}}_{{}_{0}}$ whose definition is somewhat technical [16, §5.3], but, which plays the role completely analogus to ${\bf X}_{{}_{0}}$ for ${\bf X}^{{}^{\text{aff}}}$. We will continue to denote by $\bf Y:=\overline{T^{\ltimes}_{{}_{\text{ad}}}}$ the toral embedding of $T^{\ltimes}_{{}_{\text{ad}}}$ obtained by taking its closure in ${\bf X}^{{}^{\text{poly}}}$. As in the loop group case, in the polynomial case too $\bf Y$ is covered by infinitely many affine spaces $\mathbb{A}^{\ell+1}$ parametrized by the affine Weyl group $W^{{}^{\text{aff}}}$. In fact, if $t^{-\alpha_{i}}$ are the regular functions on $T^{\ltimes}_{{}_{\text{ad}}}$ given by the character $-\alpha_{i}$, then $\overline{T^{\ltimes}_{{}_{ad,0}}}:=\overline{T^{\ltimes}_{{}_{\text{ad}}}}\cap{\bf X}_{{}_{0}}^{{}^{poly}}\simeq{\rm Spec}\,\mathbb{C}[t^{-\alpha_{0}},\ldots,t^{-\alpha_{\ell}}].$ We index these affine spaces by the affine Weyl group $W^{{}^{\text{aff}}}$ and denote them as before by $Y_{{}_{w}}$’s, $w\in W^{{}^{\text{aff}}}$. As before, we may construct a finite-dimensional Lie-algebra bundle $J$ on ${{\bf Y}^{{}^{\text{aff}}}}$, then ${\bf R}$ on ${\bf X}^{{}^{\text{poly}}}$ and then construct a group scheme ${\mathcal{G}}_{{}_{{\bf X}^{{}^{\text{poly}}}}}^{{}^{\varpi}}$. Since the inductive structure of ${\bf X}^{{}^{\text{poly}}}$ is identical to ${\bf X}^{{}^{\text{aff}}}$, the proofs are also identical. Therefore we omit them. Now we outline a construction in the polynomial loop situation which will be generalized in the next section. For any affine space $Y_{{}_{w}}$, we have an affine embeddings $\displaystyle i_{w}:T_{{}_{\text{ad},w}}\times\mathbb{G}_{m}=T^{\ltimes}_{{}_{\text{ad},w}}\hookrightarrow Y_{{}_{w}}$ (6.0.2) Furthermore, the projection $p:T_{{}_{\text{ad},w}}\times\mathbb{G}_{m}\to\mathbb{G}_{m}$ extends to a morphism $p_{w}:Y_{{}_{w}}\to\mathbb{A}^{1}$ such that $p^{{}^{-1}}(0)$ is the union of the hyperplanes. At the level of coordinates, this extension is simply the product of the coordinate functions. Further $i_{w}$ and $p_{w}$ glue to give (6.0.7) ## 7\. An analogue of a construction of Mumford In [13] towards the very end Mumford gives a beautiful construction of the geometric realization of the relative case of buildings via toroidal embeddings. He deals with the general situation of an arbitrary discrete valuation ring $R$ with algebraically closed residue field $k=R/\mathfrak{m}$. Our aim in this section is limited to the case when $k=\mathbb{C}$, $\mathfrak{m}=(z)\subset\mathbb{C}[z]$ and $R=\mathbb{C}[z]_{{}_{\mathfrak{m}}}$. Let $K=\mathbb{C}(z)$ and let $S={\rm Spec}\,R$ and $\eta$ be the generic point and $o\in S$ the closed point. For schemes $X$ over $S$, $X_{{}_{\eta}}$ will be the fibre over $\eta$ and $X_{{}_{o}}$ the fibre over $o$. We will briefly sketch an analogue of Mumford’s construction in our setting and outline of the relationship between this construction to Solis’s approach and as a consequence construct a “wonderful" Bruhat-Tits group scheme on the toroidal embedding of $G_{{}_{ad,}}\times S$. For the purposes of this section alone so as to remain consistent with the one in [13], we will have the following set of notations. $\displaystyle H:=G_{{}_{\text{ad}}}$ (7.0.1) $\displaystyle H_{{}_{S}}:=G\times S$ (7.0.2) $\displaystyle T_{{}_{S}}\subset H_{{}_{S}}$ (7.0.3) $\displaystyle T_{{}_{S}}:=T_{{}_{\text{ad}}}\times S$ (7.0.4) $\displaystyle T_{{}_{K}}:=T_{{}_{\text{ad}}}\times{\rm Spec}\,K$ (7.0.5) i.e. $H_{{}_{S}}$ is a split adjoint semisimple group scheme over $S$, $T_{{}_{S}}\subset H_{{}_{S}}$ a fixed split maximal torus and $G$ will as before stand for the simply connected group. We will denote $U^{\pm}_{S}:=U^{\pm}\times S$. Base-changing (6.0.7) by ${\rm Spec}\,R\rightarrow\mathbb{A}^{1}$, by the definition of $\bf Y$ we obtain: (7.0.10) We continue to denote by $Y_{{}_{w}}$ the base-change. The orbit space $H_{{}_{S}}/T_{{}_{S}}$ exists as an affine scheme over $S$ and $H_{{}_{S}}$ via the quotient map $\pi:H_{{}_{S}}\to H_{{}_{S}}/T_{{}_{S}}$ is a locally free principal $T_{{}_{S}}$-bundle over $H_{{}_{S}}/T_{{}_{S}}$. We now consider the associated fibre bundle: (7.0.15) Note that $T_{\eta}=Y_{w,\eta}$ and hence $H_{\eta}=(H_{S}\times^{{}^{T_{S}}}Y_{{}_{w}})_{{}_{\eta}}$. Let us denote by $\displaystyle\overline{Z}_{w}:=H_{S}\times^{{}^{T_{S}}}Y_{{}_{w}}$ (7.0.16) and observe that $U_{{}_{S}}^{-}\times Y_{{}_{w}}\times U_{{}_{S}}=Z_{{}_{w}}$ is an open subset of $\overline{Z}_{w}$. In fact, over $U_{{}_{S}}^{-}\times U_{{}_{S}}\subset H_{{}_{S}}/T_{{}_{S}}$, the quotient map $\pi:H_{S}\rightarrow H_{S}/T_{S}$ is trivial. Analogous to Mumford’s definition [13, page 206], we define $\displaystyle\overline{H_{{}_{S}}}:=\bigcup_{{}_{x\in H(K)}}(H_{S}\times^{{}^{T_{S}}}Y_{{}_{w}}).x$ (7.0.17) where the generic fibre $H_{\eta}$ is identified in each copy $(H_{S}\times^{{}^{T_{S}}}Y_{{}_{w}}).x$ by a translate of $H_{\eta}$ by right multiplication by $x$ in the Iwahori subgroup. It takes a bit to prove that there is such an action. Note that we have the identification: $\displaystyle H_{\eta}=\overline{H_{{}_{\eta}}}$ (7.0.18) Consider the inclusion $G_{{}_{\text{ad}}}\times\mathbb{G}_{m}\hookrightarrow L_{{}_{\text{poly}}}^{\ltimes}G/Z(G)$. Let ${\bf M}=\overline{G_{{}_{\text{ad}}}\times\mathbb{G}_{m}}$ in ${\bf X}^{{}^{poly}}$ as a scheme over $\mathbb{C}$. It is locally of finite type. ###### Proposition 7.1. The $S$-scheme $\overline{H_{{}_{S}}}$ as a $\mathbb{C}$-scheme is isomorphic to $\bf M$. In particular, $\overline{H_{{}_{S}}}$ is a regular scheme over $\mathbb{C}$ such that the closed fibre $\overline{H_{o}}$ is a complete scheme over $\mathbb{C}$ which is a union of smooth components meeting at normal crossings. The embedding $H_{{}_{S}}\subset\overline{H_{{}_{S}}}$ is a toroidal embedding without self-intersection. Finally, the strata of $\overline{H_{{}_{S}}}-H_{{}_{S}}$ are precisely the parahoric subgroups of $H(K)$. This bijection between the strata and the parahoric subgroups extends to an isomorphism of the graph of the embedding $H_{{}_{S}}\subset\overline{H_{{}_{S}}}$ with the Bruhat-Tits building of $H$ over $S$. By the methods in this note we immediately deduce the existence of a “wonderful" Bruhat-Tits group scheme $\mathcal{G}_{{}_{\bf M}}^{{}^{\varpi}}$ on $\bf M$ which behaves naturally with respect to the strata. ###### Remark 7.2. A difference between Mumford’s construction and ours is that in Mumford’s construction the fibre over the closed point in $S$ is a reducible divisor but not necessarily with simple normal crossing singularities. In fact, the components associated to the parahorics which are not hyperspecial come with multiplicity being the coefficient $c_{{}_{\alpha}}$ of the associated simple root in the expression of the highest root. So a Kawamata lemma is not immediately applicable in that situation. ###### Remark 7.3. Mumford’s construction works over algebraically closed fields of any characteristic while we assumed the characteristic to be $0$. We did this to make the construction in the loop group situation following [16]. On the other hand, it seems that with a bit more work the whole construction of $\mathcal{G}_{{}_{\bf M}}^{{}^{\varpi}}$ will go through for characteristics $p$ such that $p$ is coprime to the $d_{{}_{\alpha}}$’s in (2.0.3). ## 8\. Appendix on parabolic and equivariant bundles In this section we recall and summarize results on parabolic bundles and equivariant bundles on Kawamata covers. These play a central role in the constructions of the Bruhat-Tits group schemes made above. Consider a pair $(X,D)$, where $X$ is a smooth quasi-projective variety and $D=\sum_{{}_{j=0}}^{{}^{\ell}}D_{{}_{j}}$ is a reduced normal crossing divisor with non-singular components $D_{{}_{j}}$ intersecting each other transversely. The basic examples we have in mind in the note are discrete valuation rings with its closed point, the wonderful compactification $\bf X$ with its boundary divisors or affine toric varieties. Let $E$ be a locally free sheaf on $X$. Let $n_{{}_{j}},j=0\ldots,\ell$ be positive integers attached to the components $D_{{}_{j}}$. Let $\xi$ be a generic point of $D$. Let $E_{{}_{\xi}}:=E\otimes_{{}_{\mathcal{O}_{{}_{X}}}}\mathcal{O}_{{}_{X,\xi}}$ and $\bar{E}_{{}_{\xi}}:=E_{{}_{\xi}}/\mathfrak{m}_{{}_{\xi}}E_{{}_{\xi}}$, $\mathfrak{m}_{{}_{\xi}}$ the maximal ideal of $\mathcal{O}_{{}_{X,\xi}}$. ###### Definition 8.1. A parabolic structure on $E$ consists of the following data: * • A flag $\bar{E}_{{}_{\xi}}=F^{{}^{1}}\bar{E}_{{}_{\xi}}\supset\ldots\supset F^{r{{}_{{}_{j}}}}\bar{E}_{{}_{\xi}}$, at the generic point $\xi$ of each of the components $D_{{}_{j}}$ of $D$, * • Weights $d_{{}_{s}}/n_{{}_{j}}$, with $0\leq d_{{}_{s}}<n_{{}_{j}}$ attached to $F^{{}^{s}}\bar{E}_{{}_{\xi}}$ such that $d_{1}<\cdots<d_{r_{j}}$ By saturating the flag datum on each of the divisors we get for each component $D_{{}_{j}}$ a filtration: $\displaystyle E_{{}_{D_{{}_{j}}}}=F^{{}^{1}}_{{}_{j}}\supset\ldots\supset F^{r{{}_{{}_{j}}}}_{{}_{j}}$ (8.0.1) of sub-sheaves on $D_{{}_{j}}$. Define the coherent subsheaf $\mathcal{F}^{s}_{{}_{j}}$, where $0\leq j\leq\ell$ and $1\leq s\leq r_{{}_{j}}$, of $E$ by: $\displaystyle 0\to\mathcal{F}^{s}_{{}_{j}}\to E\to E_{{}_{D_{{}_{j}}}}/F^{s}_{{}_{j}}\to 0$ (8.0.2) the last map is by restriction to the divisors. In [2] it is shown that there is a Kawamata covering of $p:(Y,\tilde{D})\to(X,D)$ with suitable ramification data and Galois group $\Gamma$. It is also shown that there is a $\Gamma$-equivariant vector bundle $V$ on $Y$ such that the invariant direct image $p^{\Gamma}_{{}_{*}}(V)=E$ which also recovers the filtrations $\mathcal{F}^{s}_{{}_{j}}$. For the sake of completeness, we give a self-contained ad hoc argument for this construction which is more in the spirit of the present note. The data given in (8.1) is as in [14] which deals with points on curves. Note that in our setting we have rational weights. Under these conditions, the aim is to set up an equivalence $\displaystyle p^{\Gamma}_{{}_{*}}:\\{\Gamma-\text{equivariant~{}bundles~{}on}~{}Y\\}\to\\{\text{parabolic~{}bundles~{}on}~{}X\\}$ (8.0.3) and as the notation suggests, this is achieved by taking invariant direct images. Since $p:Y\to X$ is finite and flat, if $V$ is locally free on $Y$ then so is $p^{\Gamma}_{{}_{*}}(V)$. An equivariant bundle $V$ on $Y$ is defined in an analytic neighbourhood of a generic point $\zeta$ of a component by a representation of the isotropy group $\Gamma_{{}_{\zeta}}$ and locally (in the analytic sense) the action of $\Gamma_{{}_{\zeta}}$ is the product action. This is called the local type in [15] or [14]. By appealing to the $1$-dimensional case, we get canonical parabolic structures on $E$ at the generic points $\xi$ of the components $D_{{}_{j}}$ in $X$. This construction is an equivalence. Given a parabolic bundle $E$ on $(X,D)$, to get the $\Gamma$-equivariant bundle $V$ on $Y$, such that $p^{\Gamma}_{{}_{*}}(V)=E$, we proceed as follows. If we did know the existence of such a $V$ then we can consider the inclusion $p^{*}(E)\subset V$. Taking its dual (and since $V/p^{*}(E)$ is torsion, taking duals is an inclusion): $\displaystyle V^{*}\hookrightarrow(p^{*}(E))^{*}=p^{*}(E^{*})$ (8.0.4) By the $1$-dimensional case (obtained by restricting to the height $1$-primes at the generic points) we see that the quotient $T_{{}_{\zeta}}:=p^{*}(E^{*})_{{}_{\zeta}}/V_{{}_{\zeta}}^{*}$ is a torsion $\mathcal{O}_{{}_{Y,\zeta}}$-module and $T_{{}_{\zeta}}$ is completely determined by the parabolic structure on $E$. The parabolic structure on $E$ therefore determines canonical quotients: $\displaystyle p^{*}(E^{*})_{{}_{\zeta}}\to T_{{}_{\zeta}}$ (8.0.5) for each generic point $\zeta$ of components in $Y$ above the components $D_{{}_{j}}$ in $X$. The discussion above suggest how one would construct such a $V$; we begin with these quotients (8.0.5). Then we observe that there is a maximal coherent subsheaf $V^{\prime}\subset p^{*}(E^{*})$ such that $\displaystyle p^{*}(E^{*})_{{}_{\zeta}}/V_{{}_{\zeta}}^{\prime}=T_{{}_{\zeta}}.$ (8.0.6) Away from $p^{{}^{-1}}(D)$ in $Y$, the inclusion $V^{\prime}\hookrightarrow p^{*}(E^{*})$ is an isomorphism. Dualizing again, we get an inclusion $p^{*}(E)\hookrightarrow(V^{\prime})^{{}^{*}}$ which is an isomorphism away from $p^{{}^{-1}}(D)$. Set $W=(V^{\prime})^{{}^{*}}$. Since $W$ is the dual of a coherent sheaf, it is reflexive. It is not hard to check that $W$ coincides with the vector bundle $V$ that Biswas constructs in codimension $2$ and hence everywhere by [12, Proposition 1.6, page 126]. This gives us the desired equivalence (8.0.3). #### 8.0.1. The group scheme situation Let $(X,D)$ be as above. Let $\xi$ be the generic point of a component $D_{{}_{j}}$ of $D$ and let $A:=\mathcal{O}_{{}_{X,\xi}}$ and $K$ be the quotient field of $A$. Let $\mathcal{G}_{{}_{\theta}}$ be a Bruhat-Tits group scheme on ${\rm Spec}\,(A)$ associated to a vertex $\theta$ of the Weyl alcove. We always assume that these group scheme are generically split. Let $B=\mathcal{O}_{{}_{Y,\zeta}}$ where $\zeta$ is the generic point of a component of $Y$ above $D_{{}_{j}}$ and let $L$ be the quotient field of $B$. We assume that the local ramification data for the Kawamata covering has numbers $d_{{}_{\alpha}}$ (2.0.3). Let $\Gamma_{{}_{\zeta}}$ be the stabilizer of $\Gamma$ at $\zeta\in Y$. The results of [1, Proposition 5.1.2 and Remark 2.3.3] show that there exists an equivariant group scheme $\mathcal{H}_{{}_{B}}$ on ${\rm Spec}\,(B)$ with fiber isomorphic to the simple connected group $G$ and such that $\text{Res}_{{}_{B/A}}(\mathcal{H}_{{}_{B}})^{{}^{\Gamma_{{}_{\zeta}}}}\simeq\mathcal{G}_{{}_{\theta}}$. Suppose that we have a group scheme $\mathcal{G}_{{}_{X^{\prime}}}$ on an open $X^{\prime}\subset X$ which includes all the height $1$ primes coming from the divisors $D_{{}_{j}}$’s with the following properties: * • Away from the divisor $D\subset X$, $\mathcal{G}$ is the constant group scheme with fiber $G$. * • The restrictions $\mathcal{G}\mid_{{}_{{\rm Spec}\,(A)}}$ at the generic points $\xi$ are isomorphic to the Bruhat-Tits group scheme $\mathcal{G}_{{}_{\theta}}$ for varying $\xi$ and $\theta$. In other words, $\mathcal{G}_{{}_{X^{\prime}}}$ is obtained by a gluing of the constant group schemes with $\mathcal{G}_{{}_{\theta}}$’s along ${\rm Spec}\,(K)$ by an automorphism of constant group scheme $G_{{}_{K}}$. Now consider the inverse image of the constant group scheme $p^{{}^{*}}(G_{{}_{X-D}})\simeq G\times p^{*}(X-D)$. Then using the gluing on $X^{\prime}$, we can glue the constant group scheme $p^{{}^{*}}(G_{{}_{X-D}})$ with the local group schemes $\mathcal{H}_{{}_{B}}$ for each generic point $\zeta$ to obtain a group scheme $\mathcal{H}_{{}_{Y^{\prime}}}$ on $Y^{\prime}=p^{{}^{-1}}(X^{\prime})$ such that: $\displaystyle\text{Res}_{{}_{Y^{\prime}/X^{\prime}}}(\mathcal{H}_{{}_{Y^{\prime}}})^{{}^{\Gamma}}\simeq\mathcal{G}_{{}_{X^{\prime}}}.$ (8.0.7) #### 8.0.2. Kawamata Coverings Let $X$ be a smooth quasi-projective variety and let $D=\sum_{{}_{i=0}}^{{}^{\ell}}D_{i}$ be the decomposition of the simple or reduced normal crossing divisor $D$ into its smooth components (intersecting transversally). The “Covering Lemma” of Y. Kawamata (see [18, Lemma 2.5, page 56]) says that, given positive integers $n_{{}_{0}},\ldots,n_{{}_{\ell}}$, there is a connected smooth quasi-projective variety $Z$ over $\mathbb{C}$ and a Galois covering morphism $\displaystyle\kappa:Z\to X$ (8.0.8) such that the reduced divisor $\kappa^{{}^{*}}{D}:=\,({\kappa}^{{}^{*}}D)_{{}_{\text{red}}}$ is a normal crossing divisor on $Z$ and furthermore, ${\kappa}^{{}^{*}}D_{{}_{i}}=n_{{}_{i}}.({\kappa}^{{}^{*}}D_{i})_{{}_{\text{red}}}$. Let $\Gamma$ denote the Galois group for the covering map $\kappa$. The isotropy group of any point $z\in Z$, for the action of $\Gamma$ on $Z$, will be denoted by ${\Gamma}_{{}_{z}}$. It is easy to see that the stabilizer at generic points of the irreducible components of $(\kappa^{{}^{*}}D_{i})_{{}_{\text{red}}}$ are cyclic of order $n_{{}_{i}}$. ## References * [1] V.Balaji and C.S. Seshadri, Moduli of parahoric $\mathcal{G}$–torsors on a compact Riemann surface, J. Algebraic Geometry, 24, (2015),1-49. * [2] I. Biswas, Chern classes for parabolic bundles. J. Math. Kyoto Univ. 37 (1997), no. 4, 597–613. doi:10.1215/kjm/1250518206. * [3] S. Bosch, W.Lutkebohmert and M.Raynaud, Neron Models, Ergebnisse 21, Springer Verlag, (1990). * [4] M. Brion, The behaviour at infinity of the Bruhat decomposition, Comm. Math. Helv. 73 (1998)137-174. * [5] M. Brion, Log homogeneous varieties, Actas del XVI Coloquio Latinoamericano de Álgebra, 1–39, Biblioteca de la Revista Matematica Iberoamericana, Madrid, 2007. * [6] F.Bruhat and J.Tits, Groupes réductifs sur un crops local II: Schémas en groupes. Existence d’une donnée radicielle valuée, Publications Mathématiques de l’IHÉS 60 (1984) 5-184. * [7] C. De Concini and C. Procesi, Complete symmetric varieties, in Invariant theory (Montecatini, 1982), pp. 1–44, Lecture Notes in Math. 996, Springer, 1983. * [8] C. De Concini and T. Springer, Compactification of Symmetric Varieties, Transform. Groups 4 (1999), no. 2-3, pp. 273-300. * [9] B. Conrad, O.Gabber and Gopal Prasad, Pseudo-reductive groups, Cambridge University Press, New Mathematical Monographs, No 17.(2010), 2nd Edition. * [10] B. Edixhoven, Neron models and tame ramification, Compositio Mathematica, 81, (1992), 291-306. * [11] A. Grothendieck, Élements de Ǵeométrie Algébrique, Math. Publ. IHES 4, (1960), 5-228. * [12] R. Hartshorne, Stable Reflexive Sheaves, Mathematische Annalen 254 (1980), 121-176. * [13] G. Kempf, F. Knudsen, D. Mumford, and B. Saint-Donat, Toroidal Embeddings I, Springer Lecture Notes 339, 1973. * [14] V. Mehta and C.S. Seshadri, Moduli of vector bundles on curves with parabolic structures, Math. Ann, 248, (1980), 205-239. * [15] C.S. Seshadri, Moduli of $\pi$–vector bundles over an algebraic curve, Questions On algebraic Varieties, C.I.M.E, Varenna, (1969), 139-261. * [16] P. Solis, A wonderful embedding of the loop group. Advances in Mathematics 313, , (2017), 689-717. * [17] E. Strickland, A vanishing Theorem for Group Compactifications, Math. Ann. 277 (1987), pp. 165-171. * [18] E. Vieweg, Projective Moduli for Polarized Manifolds, , Ergebnisse, Springer(1995).
# Graphical Models for Financial Time Series and Portfolio Selection Ni Zhan<EMAIL_ADDRESS>, Yijia Sun<EMAIL_ADDRESS>, Aman Jakhar and He Liu Carnegie Mellon University (2020) ###### Abstract. We examine a variety of graphical models to construct optimal portfolios. Graphical models such as PCA-KMeans, autoencoders, dynamic clustering, and structural learning can capture the time varying patterns in the covariance matrix and allow the creation of an optimal and robust portfolio. We compared the resulting portfolios from the different models with baseline methods. In many cases our graphical strategies generated steadily increasing returns with low risk and outgrew the S&P 500 index. This work suggests that graphical models can effectively learn the temporal dependencies in time series data and are proved useful in asset management. dynamic clustering, portfolio selection, autoencoders ††journalyear: 2020††copyright: rightsretained††conference: ACM International Conference on AI in Finance; October 15–16, 2020; New York, NY, USA††booktitle: ACM International Conference on AI in Finance (ICAIF ’20), October 15–16, 2020, New York, NY, USA††doi: 10.1145/3383455.3422566††isbn: 978-1-4503-7584-9/20/10 ## 1\. Introduction Portfolio selection is a common problem in finance. In general, investors wish to maximize returns while minimizing risk. Markowitz theory suggests that portfolio diversification minimizes risk and the optimal portfolio takes into account correlated movements across assets. Existing works on porfolio selection use the historical covariance matrix of returns. However, factors such as market index, sector, industry and other stocks and commodities that may affect asset cash flows can result in a high degree of correlation among equities. This causes the historical covariance matrix to be ill conditioned and the optimal portfolio highly sensitive to small changes. Expanding the universe of assets also requires a greater amount of data to estimate the covariance matrix. Furthermore regime changes and the non-stationary nature of financial markets discourage the use of static covariance matrices. From a graph viewpoint, estimating the covariance using historic returns models a fully connected graph between all assets. The fully connected graph appears to be a poor model in reality, and substantially adds to the computational burden and instability of the problem. The goal of this work is to develop graphical models that can capture the time varying patterns in the covariance matrix and reflect the cross-series dynamics at multiple time indices. Using graph inference algorithms and thresholding, we plan to discover and incorporate the factor dependencies in a partially connected graph. Therefore, we leverage graphical models that are able to reflect temporal changes among stocks thus addressing the issue of correlations changing over time. Within a time period, a graph allows selection of independent assets for the portfolio, which should improve robustness of our solution. In order to learn the overall time series features, we use principle component analysis (PCA) and autoencoders to capture the latent space distribution. We employ variational autoencoders with Gaussian and Cauchy priors to model temporal dependencies and reflect multi-scale dynamics in the latent space. Additionally, we construct a sequence of graphical models using dynamic clustering techniques and structural learning. We associate a graphical model to each time interval and update the graph when moving to the next time point. We use price data of US equities from the S$\&$P 500 index to construct graphical models to create portfolios, simulate returns, and compare with benchmarks. ## Related Work Various graph methods have been used for the portfolio selection problem. The literature includes many examples of variance-covariance networks that analyze complicated interactions and market structure of financial assets (Ledoit and Wolf, 2003; Buser, 1977; Polson and Tew, 2000). Liu et. al. used elliptical- copula graphical models on stock returns, and used the graphs to select independent stocks (Liu et al., 2015). The paper chose elliptical-copula graph model over Gaussian graphical models because elliptical-copula models better model heavy tail distributions common in finance. The paper shows that graphs modeling asset “independence” can be learned from historical returns and used for effective portfolio strategies, however it did not consider dynamic graphs. Previously, time varying behavior was modeled as dynamic networks whose topology changes with time. Talih and Hengartner proposed a graphical model for sequences of Gaussian random vectors when changes in the underlying graph occur at random times, mimicking the time varying relationships among collection of portfolios (Talih and Hengartner, 2005). Time series data is separated into a pre-determined number of blocks. The sample precision matrix estimated within each block serves as the foundation to construct the time- varying sequences of graphs which arises in blocks as shown in Fig. 1. Figure 1. Blocks of Gaussian graph sequences Blocks of Gaussian graph sequences. The graphs $G_{b}$ and $G_{b+1}$ in successive blocks differ only by the addition or deletion of at most one edge. Markov Chain Monte Carlo is used to recover the segmented time varying Gaussian graphical sequences. Experiments using this model were done to estimate portfolio return in five U.S. industries. However, this paper assumed that the total number of distinct networks are known a priori and the network evolution is restricted to changing at most a single edge at a time. Additionally, dimensionality increases when temporal resolution is small, imposing significant computational burden. As an extension to dynamic dependence networks, Isogai (Isogai, 2017) proposed a novel approach to analyze a dynamic correlation network of highly volatile financial asset returns by using network clustering algorithms to mitigate high dimensionality. Two types of network clustering algorithms (Isogai, 2014, 2016) were employed to transform correlation network of individual portfolio returns into a correlation network of group based returns. Groups of correlation networks were further clustered into only three representative networks by clustering along the time axis to summarize information on the intertemporal differences in the correlation structure. A case study was conducted on Japanese stock dataset. Other studies used autoencoders to dimensionally reduce stock returns (Gu et al., 2019; Heaton et al., 2016). ## Methods The primary objective of this work is to exploit a collection of graphical models to analyze dynamic dependencies among stocks and aid trading strategies. We would like to have a better understanding of the stock features by learning the latent space of stock time series. With a good latent space representation, stocks with similar features will fall into the same cluster. To capture the time-varying correlation among stocks, we divide the overall time horizon into multiple time intervals, zoom in to each interval and construct a local graphical model, and update the local connectivities as we move on to the next temporal period. In each time interval, we aim to identify a suitable number of stock clusters based on the graphical model and develop a portfolio selection strategy. After selecting our portfolios, we test their performance using a backtest simulation and market data. We employ a few different clustering and structural learning techniques, including PCA, autoencoders, agglomerative clustering, affinity propagation clustering, and graphical LASSO, to create portfolios that maximize return and minimize risk. #### Dataset We used price data of US equities from tech, financials, and energy sectors of the S$\&$P 500 index. The data matrix has available stocks as data observations and daily returns as features, i.e. the data matrix has shape [number of stocks, number of daily returns]. Data was obtained on a closing basis at a daily frequency for January 1st, 2012 until January 1st, 2020. We excluded stocks which had missing data for this time period. We used adjusted close to account for splits, dividends etc. The daily closing price data for the stocks and other financial variables was obtained using yahoo finance for python. #### PCA We used empirical returns of one year prior as training data for the simulated year. We used PCA to dimensionally reduce the data matrix of the stocks to three components. We inverse transformed stock representations in reduced space to full data, and calculated the L-2 norm of difference between actual data and recovered data for each stock. The ten stocks with the largest difference in L-2 norm were selected for ”Max Difference” portfolio while ten stocks with the smallest difference were selected for ”Min Difference” portfolio. PCA extracts information about the stock returns, and stocks with large difference between recovered and actual data indicate unexpected or ”difficult to capture” behavior. #### Autoencoders To extend the dimension reduction method and capture more complex interactions, we tested two autoencoders. The observed variables $\mathbf{x}$ for the autoencoders are empirical returns of the selected stocks. The latent space $\mathbf{z}$ has two dimensions. The variational distribution $q_{\phi}(\mathbf{z}|\mathbf{x})$ approximating $p(\mathbf{z}|\mathbf{x})$ is assumed Normal for both autoencoders. The likelihood $p_{\theta}(\mathbf{x}|\mathbf{z})$ is assumed Normal for one autoencoder, and Cauchy for the other. Specifically, for the Normal autoencoder, $p_{\theta}(\mathbf{x}|\mathbf{z})=N(\mathbf{x};\mu_{\theta}(\mathbf{z}),\Sigma^{2}_{\theta}(\mathbf{z}))$, and for the Cauchy autoencoder, $p_{\theta}(\mathbf{x}|\mathbf{z})=\text{Cauchy}(\mathbf{x};x_{0,\theta}(\mathbf{z}),\gamma_{\theta}(\mathbf{z}))$. For both autoencoders, the recognition and generative networks are parameterized by neural nets with one hidden layer with 100 reLU neurons and latent space of two dimensions. We chose a Cauchy distribution because stock return distributions usually have heavy tails, and therefore expect the Cauchy autoencoder to have more reliable results. To select portfolios, we found the latent space representation of each stock, and calculated the L-2 norm of difference between real data and generated data (from latent space) for each stock. The ten stocks with max and min L-2 norms were selected for Max and Min Difference portfolios, respectively, similar to the PCA strategy. #### Dynamic clustering and graphical sequences To better capture how stocks move in relation to one another throughout the time period, we utilize clustering and structural learning techniques to create dynamic graph structures. Even though the dataset contains stocks from three sectors, within the same sector, some stocks are more correlated than the others, and stocks from different sectors can also have non-negligible correlations. The goal here is to identify the most suitable number of clusters to assign stocks from three sectors into. Two clustering techniques are adopted here: agglomerative clustering and affinity propagation clustering. Agglomerative clustering divides stocks into a number of clusters according to pair-wise Euclidian distance. It requires the number of clusters to be pre- determined. We used hierarchical clustering with Ward’s minimum variance criterion to produce a dendrogram which in turn is used to determine the number of clusters by drawing a horizontal line and counting the number of vertical lines it intercepts. A total of 15 clusters were used for agglomerative clustering. Affinity propagation (Frey and Dueck, 2007) is a clustering technique that does not require an input number of clusters. It relies on similarity calculation between pairs of data points to determine a subset of representative examples in the dataset. The similarity between two points satisfies that $s(x_{i},x_{j})>s(x_{i},x_{k})$ if and only if $x_{i}$ is more similar to $x_{j}$ than to $x_{k}$. A responsibility matrix $\mathcal{R}$ and availability matrix $\mathcal{A}$ serve as message exchanging paths between data points. Clusters are updated by alternating between the responsibility matrix update and availability matrix update. $\displaystyle r(i,k)$ $\displaystyle=$ $\displaystyle s(i,k)-\max_{k^{\prime}\neq k}\\{a(i,k^{\prime})+s(i,k^{\prime})\\}$ $\displaystyle a(i,k)$ $\displaystyle=$ $\displaystyle\min\big{(}0,r(k,k)+\sum_{i^{\prime}\notin\\{i,k\\}}\max(0,r(i^{\prime},k))\big{)}\;i\neq k$ $\displaystyle a(k,k)$ $\displaystyle=$ $\displaystyle\sum_{i^{\prime}\neq k}\max(0,r(i^{\prime},k))$ This approach is able to identify a high quality set of exemplars and the corresponding clusters with much lower error and lower computational burdens compared to agglomerative clustering. The number of clusters is flexible and updated throughout the time horizon. It is suitable to identify clusters when the data size is large. The data matrix contains daily returns of each stock, and has shape [number of stocks, number of daily returns]. We used spectral embedding on the daily returns to transform the stocks to a 2D plane and reduce dimensionality. An example of the 2D embedding of stocks is shown in Fig. 2. Edge connectivities were created using graphical LASSO with thicker edge indicating stronger correlation. At the beginning of each quarter, we rely on daily returns from the previous quarter, construct a lower-dimensional embedding space, and generate clusters using the before-mentioned clustering techniques. The edge connectivities from graphical LASSO are connectivity input for agglomerative clustering. A new clustering is created at the beginning of each quarter and therefore updated throughout the time horizon quarterly. For agglomerative clustering and affinity propagation portfolios, the portfolio selection strategy is as follows. For each cluster, the top ten stocks with minimum Euclidean distance (in the embedding space) to cluster centers were added to the portfolio. If a cluster had fewer than ten stocks, no stocks were added to the portfolio from that cluster. Because the clusters were created quarterly, the portfolios were also selected quarterly. To benchmark the model performance, we also constructed portfolios using PCA followed by KMeans clustering. KMeans with static clusters as well as quarterly updated clusters were both used as benchmarks. In all cases with KMeans, we used a fixed number of clusters (ten clusters), and the stock with the minimum Euclidean distance to each cluster center was part of the portfolio. Therefore KMeans portfolios had ten total stocks. Figure 2. An example of connected graphs from dynamic clustering An example of connected graphs from dynamic clustering. ### Testing Frameworks We used portfolio rebalancing strategy to test the performance of stock selection. This method rebalances the portfolio at some frequency (we used monthly). The weights across the portfolio’s stocks are either equal weights or determined from mean-variance optimization (described below). The portfolio is initialized with weights of the selected stocks. At each rebalance time point, shares are bought or sold to renormalize the dollar amounts by weight across the stocks. We compared the PCA and autoencoder portfolio selection strategy with rebalancing and buying and holding the S&P 500. Metrics to evaluate strategies included total returns, daily return standard deviations, and Sharpe ratios across simulation time-periods. Note that high Sharpe is desirable and indicates high returns with lower risk. We followed the equal weight rebalancing strategy to test the performance of dynamic clustering, comparing affinity propagation, agglomerative clustering, PCA KMeans portfolios and the S&P 500. At the end of each quarter, stock clusters were updated by selling the existing portfolio and buying stocks from the new cluster with equal dollar amounts. #### Efficient frontier weights Efficient frontier weights were determined using PyPortfolioOpt, with expected means and covariance calculated from empirical returns of the year prior to simulation. The solver optimized for maximum Sharpe, and stock weights were unconstrained between 0 to 1. ## Results ### PCA, Autoencoder Portfolios The two autoencoders were trained with Auto-encoding Variational Bayes (AEVB) that optimizes a stochastic estimate of evidence lower bound objective (ELBO) (Kingma and Welling, 2013). The autoencoders were trained for over 200 epochs, and the lower bound of log-likelihood converged. The training for each year was repeated ten times as the training is stochastic. The max and min difference portfolios across the ten trainings were aggregated per year, and the ten stocks which appeared most frequently in a year were used for simulating the following year. The PCA and autoencoders portfolio selection and monthly rebalance strategy was implemented for five simulation years, 2014-2018 inclusive. We rebalanced using both equal weights and efficient frontier weights. To compare against simpler methods, we constructed additional portfolios: volatility (Vol) and ”average return over volatility” (AvgRet/Vol). The max and min Vol portfolios had ten stocks with highest or lowest standard deviation of returns, respectively, in year prior to simulation. A stock selected for Max PCA or autoencoder portfolio has a large error in its model representation, and may have high volatility. Therefore we wanted to compare volatility alone with PCA and autoencoder portfolios. The ”average return over volatility” represents a proxy for Sharpe ratio, and we wanted to test if individual stocks with high Sharpe combined would have good portfolio performance. Table 1 shows the simulation results: yearly returns (%) and standard deviation of daily returns (%). Daily return standard deviations are higher for max portfolios than min portfolios for PCA, autoencoder, and Vol, which is expected based on their construction. Table 1 also includes the average return and Sharpe ratio over the five simulation years. Risk free return in Sharpe ratio was one-year Treasury yield averaged for that year. Results for 2019 are reported as a forward-test (the data was completely withheld prior to reporting), and holding S&P 500 is shown for comparison. From Table 1, there are several observations which show the PCA and autoencoder strategies are useful in portfolio selection and able to create portfolios with higher return at lower risk, more so than volatility or AvgRet/Vol alone. The Max PCA and Max autoencoder portfolios perform better or on-par with Max Vol in terms of yearly return and Sharpe, from 2014-2018. For PCA and autoencoder, Max outperforms Min in both Sharpe and average returns, which is not the case with Vol. Specifically Min Vol has higher Sharpe but lower average returns than Max Vol. The Max AvgRet/Vol portfolio has higher Sharpe but lower average return than the other portfolios, and is not always able to obtain an optimal weights solution. Therefore, the model-based strategies aid in choosing portfolios which have better returns, Sharpe, and weights optimization capability, over Vol or AvgRet/Vol alone. Other observations are that PCA seems to have lower risk than autoencoder. This is likely because PCA is a simpler model and stochasticity was introduced in autoencoder training. Improvements to the autoencoder model can be considered for future work. For some returns such as 309% and 142%, the efficient frontier allocated all assets into one stock when given ten stocks. Picking one outperforming stock can give extraordinarily high returns compared with selecting multiple stocks for a portfolio. In the cases with 309% and 142% returns, the efficient frontier optimizer selected only one stock because portfolio weight allocations were unconstrained between 0 and 1. Future work can include constrained weight allocation for the efficient frontier portfolios. In 2018, the general market trended down, and Min Vol portfolio performed the best, while years 2014 through 2017 were bull markets. The best portfolio selection strategy is likely different depending on the overall market trend. It would be interesting to consider optimal strategies for different market trends and transitions. Table 1 included Normal autoencoder results only because the Cauchy autoencoder had similar results. We examined the latent space representations of the stocks. Between PCA and the two autoencoders, the Cauchy autoencoder had the best performance in separating stocks by sector, shown in Fig 3. This shows the model learned relevant information about the stocks. We find it quite remarkable that the autoencoder clustered the sectors considering the only training data was daily stock return for a year. It may show that stock returns are quite correlated within sectors. The strategies commonly selected some stocks while others were more specific to certain models. For example, for the Min portfolios, BRK-B, USB were common across Vol, PCA, and autoencoders. AFL was more often selected by Cauchy autoencoder, MMC by Vol, BLK by Normal autoencoder. For the Max portfolios, AMD, MU were common across Vol, PCA, and autoencoders, while AKAM was more specifically selected by PCA. The fact that some stocks were common across all models shows that models learned relevant information, while some stocks specifically selected by certain models shows that models learned distinct information. In addition, Max portfolio stocks exhibited ”outlier” behavior. Table 1. PCA, Autoencoder, Volatility Portfolio returns and standard deviations (%) by simulation year | | | 2014 | 2015 | 2016 | 2017 | 2018 | Avg Yr Ret | Daily Ret Std | Sharpe | 2019 Test ---|---|---|---|---|---|---|---|---|---|---|--- Equal Weights | PCA | Max | 47.7 | 10.4 | 65.4 | 34.2 | -3.1 | 30.9 | 1.39 | 1.19 | 36.5 | Min | 10.7 | 1.69 | 25.7 | 19.6 | -16.8 | 8.2 | 0.94 | 0.48 | 32.3 AutoEnc | Max | 29.8 | 3.85 | 75.8 | 25.3 | -5.02 | 26.0 | 1.5 | 0.88 | 28.2 | Min | 12.9 | -1.66 | 23.5 | 16.0 | -20.7 | 6.0 | 1.05 | 0.31 | 33.4 Vol | Max | 46.0 | -0.86 | 74.4 | 18.9 | -5.52 | 26.6 | 1.57 | 0.84 | 37.4 | Min | 5.54 | 2.64 | 21.7 | 22.8 | 3.54 | 11.2 | 0.79 | 1.13 | 31.2 AvgRet/ Vol | Max | 16.9 | 6.29 | 21.3 | 24.3 | 0.72 | 13.9 | 1.12 | 1.39 | 40.6 Min | 12.5 | -18.5 | 33.5 | 38.9 | -22.7 | 8.7 | 1.27 | 0.30 | 2.0 Eff Frontier | PCA | Max | 60.5 | 19.5 | 25.9 | 65.5 | -14.1 | 31.5 | 1.94 | 1.03 | 142 | Min | 22.3 | -6.12 | 27.6 | 25.8 | -9.66 | 12.0 | 1.05 | 0.66 | 44.2 AutoEnc | Max | 58.4 | 19.5 | 309 | 65.5 | -13.4 | 87.8 | 2.66 | 0.76 | 141 | Min | 15.9 | 2.11 | 35.4 | 28.1 | -10.2 | 14.3 | 1.18 | 0.78 | 11.6 Vol | Max | 57.7 | 19.5 | 25.9 | 20.4 | -13.3 | 22.0 | 2.0 | 0.91 | 142 | Min | 11.6 | -1.93 | 28.6 | 15.2 | 3.49 | 11.4 | 0.89 | 0.98 | 27.7 S&P | 14.4 | 1.29 | 13.6 | 20.7 | -5.13 | 9.0 | 0.79 | 0.82 | 30.4 Figure 3. Latent space representation from a trained Cauchy autoencoder Latent space representation from a trained Cauchy autoencoder groups sectors together. ### Dynamic Clustering We used trading data from January 1st, 2012 to January 1st, 2019 to evaluate the performance of the model. On the first day of each quarter, clusters are formed following agglomerative and affinity propagation clustering techniques using daily closing price from the previous quarter, and new clusters are created each quarter. A portfolio is created by selecting stocks with minimum Euclidean distance to cluster centers in the lower-dimensional transformed space. We also generated two portfolios using PCA with KMeans clustering, with fixed clusters and quarterly updated clusters respectively, as our benchmark portfolios. Figure 4 compares the various portfolios using monthly rebalance strategy and holding the S&P 500. Both KMeans cluster portfolios lead to comparable performance with the S&P 500. KMeans with dynamic cluster update performs better than using the same clusters throughout the time horizon. Both agglomerative clustering and affinity propagation outperformed S&P 500 index, with affinity propagation generating the highest returns throughout seven years. Since affinity propagation identifies the most suitable number of clusters for the given data set, the resulting cluster size can differ from what is used in agglomerative and KMeans clustering. The likelihood of a stock being assigned to the correct cluster is higher for affinity propagation, therefore stocks that constantly beat the S&P 500 are likely to be assigned to the same cluster and correspond to minimum distance in the spectral embedding space. Creating portfolios using such stocks steadily outgrows the performance of S&P 500. We found that both cluster update and cluster size can influence the quality of the portfolio. Involving temporal changes to the clusters invariably boost the performance. Figure 4. Rebalancing with different clustering strategies vs. S&P 500 from 2012 to 2018 Affinity propogation dynamic clusters outperforms other strategies and benchmarks in simulation test from 2012 to 2018. ## Conclusion We showed that graphical models learn useful information and correlation between stocks only based on their returns. We developed a portfolio selection using PCA and autoencoder models and rebalance strategy that selects high return, low risk portfolios. We also explored the effect of dynamic clustering on overall portfolio returns. We observed that dynamic cluster update yields higher returns than using fixed clusters. A flexible cluster size also improves the performance than using a constant cluster size. When stocks are assigned to the correct clusters throughout the time horizon, with rebalancing strategies that minimizes risk, we are able to create a portfolio with steadily increasing returns. For future work, we can include more data into the analysis and model training, such as using trading volume, expanding number of years and stock sectors. There are other experimental factors which can be varied such as dimensions in latent space, stock selection strategy, rebalance frequency and timing. There are many questions to explore, and this work shows that graphical models have interesting and useful applications in asset management. ## References * (1) * Buser (1977) Stephen A. Buser. 1977\. Mean-Variance Portfolio Selection with Either a Singular or Nonsingular Variance-Covariance Matrix. _Journal of Financial and Quantitative Analysis_ 12, 3 (1977), 347–361. * Frey and Dueck (2007) Brendan J. Frey and Delbert Dueck. 2007. Clustering by Passing Messages Between Data Points. _Science_ 315, 5814 (2007), 972–976. * Gu et al. (2019) Shihao Gu, Bryan T. Kelly, and Dacheng Xiu. 2019\. Autoencoder Asset Pricing Models. _SSRN Electronic Journal_ (2019). https://doi.org/10.2139/ssrn.3335536 * Heaton et al. (2016) J. B. Heaton, N. G. Polson, and J. H. Witte. 2016\. Deep Learning for Finance: Deep Portfolios. _Applied Stochastic Models in Business and Industry_ 33, 1 (2016), 3–12. https://doi.org/10.1002/asmb.2209 * Isogai (2014) Takashi Isogai. 2014\. Clustering of Japanese stock returns by recursive modularity optimization for efficient portfolio diversification*. _Journal of Complex Networks_ 2, 4 (07 2014), 557–584. * Isogai (2016) Takashi Isogai. 2016\. Building a dynamic correlation network for fat-tailed financial asset returns. _Applied Network Science_ 1, 7 (08 2016). * Isogai (2017) Takashi Isogai. 2017\. Dynamic correlation network analysis of financial asset returns with network clustering. _Applied Network Science_ 2, 1 (05 2017). Issue 8. * Kingma and Welling (2013) Diederik P Kingma and Max Welling. 2013. Auto-Encoding Variational Bayes. _CoRR_ (2013). arXiv:1312.6114 [stat.ML] http://arxiv.org/abs/1312.6114v10 * Ledoit and Wolf (2003) Olivier Ledoit and Michael Wolf. 2003. Improved estimation of the covariance matrix of stock returns with an application to portfolio selection. _Journal of Empirical Finance_ 10, 5 (2003), 603 – 621. * Liu et al. (2015) Han Liu, John Mulvey, and Tianqi Zhao. 2015. A Semiparametric Graphical Modelling Approach for Large-Scale Equity Selection. _Quantitative Finance_ 16, 7 (2015), 1053–1067. https://doi.org/10.1080/14697688.2015.1101149 * Polson and Tew (2000) Nicholas G. Polson and Bernard V. Tew. 2000. Bayesian Portfolio Selection: An Empirical Analysis of the S&P 500 Index 1970-1996. _Journal of Business & Economic Statistics_ 18, 2 (2000), 164–173. * Talih and Hengartner (2005) Makram Talih and Nicolas Hengartner. 2005. Structural Learning With Time-Varying Components: Tracking the Cross-Section of Financial Time Series. _Journal of the Royal Statistical Society: Series B (Statistical Methodology)_ 67, 3 (2005), 321–341. https://doi.org/10.1111/j.1467-9868.2005.00504.x
11institutetext: Laboratory for Atmospheric and Space Physics, University of Colorado at Boulder, 3665 Discovery Drive, Boulder, CO, USA 11email<EMAIL_ADDRESS>22institutetext: NOAA/National Centers for Environmental Information, 325 Broadway, Boulder, CO, USA 33institutetext: High Altitude Observatory, National Center for Atmospheric Research, P.O. Box 3000, Boulder, CO, USA 44institutetext: Naval Research Laboratory, Washington, DC, USA 55institutetext: Colorado Research Associates Division, NorthWest Research Associates, 3380 Mitchell Lane, Boulder, CO, USA 66institutetext: NASA Goddard Space Flight Center, 8800 Greenbelt Road, Greenbelt, MD, USA 77institutetext: Institute of Physics & Kanzelhöhe Observatory for Solar and Environmental Research, University of Graz, A-8010 Graz, Austria 88institutetext: Royal Observatory of Belgium, Avenue Circulaire 3, 1180 Uccle, Belgium 99institutetext: Reflective X-ray Optics LLC, New York, NY, USA # SunCET: The Sun Coronal Ejection Tracker Concept James Paul Mason 11 Phillip C. Chamberlin 11 Daniel Seaton 22 Joan Burkepile 33 Robin Colaninno 44 Karin Dissauer 55 Francis G. Eparvier 11 Yuhong Fan 33 Sarah Gibson 33 Andrew R. Jones 11 Christina Kay 66 Michael Kirk 66 Richard Kohnert 11 W. Dean Pesnell 66 Barbara J. Thompson 66 Astrid M. Veronig 77 Matthew J West 88 David Windt 99 Thomas N. Woods 11 ###### Abstract The Sun Coronal Ejection Tracker (SunCET) is an extreme ultraviolet imager and spectrograph instrument concept for tracking coronal mass ejections through the region where they experience the majority of their acceleration: the difficult-to-observe middle corona. It contains a wide field of view (0–4 $R_{\odot}$) imager and a 1 Å spectral-resolution-irradiance spectrograph spanning 170–340 Å. It leverages new detector technology to read out different areas of the detector with different integration times, resulting in what we call “simultaneous high dynamic range”, as opposed to the traditional high dynamic range camera technique of subsequent full-frame images that are then combined in post-processing. This allows us to image the bright solar disk with short integration time, the middle corona with a long integration time, and the spectra with their own, independent integration time. Thus, SunCET does not require the use of an opaque or filtered occulter. SunCET is also compact — $\sim$15 $\times$ 15 $\times$ 10 cm in volume — making it an ideal instrument for a CubeSat or a small, complementary addition to a larger mission. Indeed, SunCET is presently in a NASA-funded, competitive Phase A as a CubeSat and has also been proposed to NASA as an instrument onboard a 184 kg Mission of Opportunity. ###### keywords: EUV instrument – Coronal Mass Ejections – high dynamic range – CubeSat ## 1 Introduction and Science Drivers The primary science question that the Sun Coronal Ejection Tracker (SunCET) instrument concept is designed to address is: What are the dominant physical mechanisms for coronal mass ejection acceleration as a function of altitude and time? In the standard model configuration of a coronal mass ejection (CME; Figure 1), a CME must overcome the constraint of overlying field in order to escape. Perhaps the simplest model of this defines a 1D, horizontal background magnetic field that declines in strength with height, characterized by the “decay index” (Bateman, 1978; Kliem and Török, 2006). If the background field decays too rapidly, the so-called torus instability of the embedded flux rope occurs, meaning the flux rope erupts. The decay index has a direct impact on the CME kinematics. The acceleration curves in the bottom of Figure 2, derived from magnetohydrodynamic (MHD) simulations by Török and Kliem (2007), correspond to decay index profiles, with each increase in acceleration corresponding to an increase in in the decay index profile and the final CME speed. Thus, the acceleration profile of a CME acts as a natural probe of the surrounding magnetic field. There are many complications layered on top of this simple model in reality, described later in this introduction. Figure 1: Standard cartoon CME model. The flux rope extends through the page. Overlying fields resist the flux rope’s elevation and expansion. Magnetic reconnection releases the energy stored in the field to accelerate the flux rope, producing a CME. Adapted from Forbes et al. (2018). The bulk of the CME acceleration profile in all cases occurs either in the observational gap or in the region where existing instruments are not optimized. This gap exists between extreme ultraviolet (EUV) imagers (widest outer field of view [FOV] of 1.7 $R_{\odot}$) and coronagraphs (typical inner FOV of 2.5 $R_{\odot}$ but effectively higher due to diffraction-degraded spatial resolution). Some instruments observe only part of the low-middle corona (Solar TErrestrial RElations Observatory [STEREO; Kaiser et al. 2007] / Coronagraph-1 [COR1; Howard et al. 2008], Geostationary Operational Environmental Satellite [GOES] / Solar Ultraviolet Imager [SUVI; Martínez- Galarce et al. 2010] , Project for On-Board Autonomy [PROBA2] / Sun Watcher with Active Pixels and Image Processing [SWAP; Seaton et al. 2013]). Some have low signal to noise in the middle corona (SUVI, SWAP). Some are ground-based with duty cycles $<20$% (K-Cor). Some have limitations on cadence (COR1). SunCET, however, avoids all of these issues because it is specifically optimized for this study of CMEs. Directly observing the CME height-time profile through the whole low and middle corona allows the derivation of complete speed-time and acceleration-time profiles, and thus strong model constraints, requiring accurate modeling of the magnetic environment to obtain the observed profiles. Such constraints do not presently exist, but SunCET can provide them. Figure 2: Top: Composite of SDO/AIA 171 Å image and SOHO/LASCO/C2 white-light coronagraph image. The longstanding observational gap is shown in dark grey. Bottom: Modeled acceleration profiles of torus instability CMEs, adapted from Török and Kliem (2007) Fig. 3. The different curves result from different background magnetic field decay index profile assumptions, with each higher acceleration peak corresponding to a larger decay index profile. Most of the acceleration occurs in the observational gap that SunCET fills. The torus instability is not the only mechanism involved in CME eruptions. Complicating factors are introduced by, e.g., the 3D structure of the erupting material and the surrounding magnetic field, by potential drainage of dense plasma, and by continued magnetic reconnection freeing more energy to drive the CME. The influence of these factors also evolve with altitude and time, as the CME dynamics play out. There have been at least 26 review papers on the topic over the last two decades (Green et al. 2018, and references therein) — a testament to the sustained, intense interest in this topic. Figure 3: Simulated CME kinematic profiles. Solid lines indicate the unperturbed torus instability. Dashed lines from right to left correspond to increasing durations (6 $\tau_{A}$ up to 10$\tau_{A}$) of an upward, linearly rising velocity perturbation, resulting in fundamentally different acceleration profiles. The SunCET FOV (0–4 $R_{\odot}$; indicated in light blue) covers and extends beyond this simulation. Adapted from Schrijver et al. (2008) Fig. 7. For example, a relatively modest complication to layer into the torus instability model is to add an upward velocity perturbation with finite duration. MHD simulations by Schrijver et al. (2008) showed that simply changing the duration of this perturbation results in fundamentally different acceleration profiles (Figure 3). With brief perturbations, the profile is single-peaked and occurs at later times. Increasing the duration of the perturbation does not simply result in an earlier peak, but in two peaks. Just as in Figure 2, the heights that these acceleration profiles differentiate themselves occurs across the Heliophysics System Observatory (HSO) measurement gap. SunCET observations can discriminate between single-peak versus double- peak CME acceleration profiles, which then determines the duration of a velocity perturbation in the torus instability model. Another CME initiation mechanism arises from the magnetic field topology of the flux rope. Hood and Priest (1981) showed that if the total twist in a flux rope exceeds a critical threshold (448°), a “helical kink” instability will occur, causing the flux rope to erupt. Such contortions lead to an impulsive acceleration and a large rotation of the flux rope (Fan 2016, Figure 4). Note the substantial differences in the simulated acceleration profiles between Figures 2, 3, and 4; and that they all occur in the under observed region. Figure 4: MFE simulation containing the helical kink instability, resulting in impulsive CME acceleration. The SunCET FOV (0–4 $R_{\odot}$; indicated in light blue) captures the impulse and small jerks. Adapted from Fan (2016). The other aspect of acceleration is direction: CMEs can be deflected away from “pure” radial propagation by as much as $\sim$30$\degree$, which is again determined primarily by Bex (Figure 5). This force has a non-radial component because the field is not perfectly symmetric about the flux rope, causing a magnetic gradient on the CME’s sides as the loops drape around the rising CME. The Forecasting a Coronal Mass Ejection’s Altered Trajectory (ForeCAT) analytical model accounts for these and other forces on a CME to determine its non-radial velocity (Kay et al., 2013, 2015, 2016; Kay and Gopalswamy, 2018). Furthermore, Kay and Opher (2015) modeled 200 CMEs in ForeCAT and found that deflection occurring in the middle corona accounts for nearly all of the deflection that occurs between initiation and 1 AU. The background magnetic field and radial CME speed are two free parameters in ForeCAT that are critical to get right; SunCET observations can strictly constrain them via forward modeling. Figure 5: ForeCAT simulations of a CME propagating through background magnetic fields (PFSS) of various strengths. R is radial distance. CMEs experience greater non-radial velocity in middle corona environments with stronger magnetic fields. The SunCET FOV (0–4 $R_{\odot}$; indicated in light blue) captures the majority of CME deflection. Adapted from Kay (2016). Additionally, coronal dimming often occurs as a result of CMEs. The faster a CME departs, the steeper the decline in coronal emission. The more mass the CME takes with it, the deeper the drop in coronal emission. A large number of studies have demonstrated this link with coronal imagers (e.g., Aschwanden 2009; Aschwanden et al. 2009; Dissauer et al. 2018, 2019; Thompson et al. 2000) and with spectral irradiance data (Woods et al., 2011; Mason et al., 2014, 2016, 2019). A major advantage of dimming measurements is that they are effective measures of CME kinematics even when they occur at disk center. Coronagraphs and imagers suffer from the problem of determining halo CME speed and/or mass. Dimming is an effective measure of CME kinematics both on and off-disk (Dissauer et al., 2019; Chikunova et al., 2020). Thus, instrument suites that can capture both the dimming and direct observations of limb CMEs are ideal for CME observation. This is precisely what SunCET does. SunCET will be the first mission that allows continuous measurements of CMEs during their initial acceleration phase using only a single instrument. This is advantageous compared to currently used instruments, where, e.g. EUV imagers in the low corona are combined with white-light coronagraphs higher up to track this phase. Artifacts can be introduced in the resulting CME kinematics using this combined data due to the tracking of different structures in the different instruments, since the observed emission is generated by different physical processes. SunCET is not dependent on other instruments to observe CME initiation and acceleration but does have a sufficiently wide field of view to overlap with coronagraphs for further expanded studies. The same challenges with different CME structures in EUV versus white light will be present, but SunCET’s broader temperature response should mitigate this somewhat. ## 2 Instrument Design SunCET is an instrument with a Ritchey-Chrétien, wide-field-of-view telescope (4 $R_{\odot}$), an off-rowland-circle EUV spectrograph, and a novel, simultaneous-high-dynamic-range detector. This new detector technology allows us to image the bright solar disk and CMEs through the dim middle corona simultaneously. It also allows us to measure solar irradiance spectra on the unused portion of the same detector with an integration time independent of the telescope image. The entire design is compact, fitting in a $\sim$15 $\times$ 15 $\times$ 10 cm volume; or about 2.5 CubeSat Units. This makes it ideal as a CubeSat or as a compact instrument suite to include on larger spacecraft that requires few physical resources. SunCET observes in the EUV rather than white light because 1) CMEs have already been demonstrated to be visible in the EUV and 2) it allows for major simplifications in the technical design of the instrument. While white-light observations are independent of temperature since they rely on light Thomson scattered from free electrons, SunCET observations do have the caveat that their temperature dependence (emission from ions at particular temperatures) means that CMEs whose plasma is not at ambient coronal temperatures will not be visible. The dynamic range between on and off disk in the EUV is already large ($\sim$105 by 2 $R_{\odot}$) but this is orders of magnitude larger in white light ($\sim$108), increasing the technical challenge. Moreover, the absolute brightness are vastly different; there are far more visible light photons. This presents a major challenge with scattered light: even small imperfections in optics would result in enough of the numerous disk photons to land on the part of the detector with the exceptionally faint middle corona, swamping out CME observations. This is further exacerbated by the fact that most surfaces scatter light more efficiently in visible light than in EUV light. Therefore, SunCET observes CMEs in the EUV. ## 3 Imager Design The SunCET imager was designed to provide high-dyamic range with moderate spatial resolution while providing a large field-of-view not heard of in historical on-disk EUV imagers out to 4 $R_{\odot}$. This section describes the technical design details that were traded in order to close on the science question. ### 3.1 Dynamic Range The SunCET imager requires a dynamic range of at least $7\text{\times}{10}^{4}$, based on GOES-16/SUVI observations of CMEs and SunCET’s design optimizations. The dimmest target of interest is a CME at the outer FOV, and the brightest is the coronal loops of an active region associated with a CME. SUVI-observed radiances are used to estimate brightness in SunCET (see Section 3.7). At 3.5 $R_{\odot}$, CMEs are $6.9\text{\times}{10}^{-4}$ W/m2/sr. A few of the brightest pixels in active regions reach $\sim$70 W/m2/sr, but are typically $\sim$4.8 W/m2/sr in SunCET. Another factor of 10 is included to distinguish the loops from the background solar disk. Thus, we have a required dynamic range of (4.8 / $6.9\text{\times}{10}^{-4}$) $\times$ 10 = $7\text{\times}{10}^{4}$. We allow solar flares and a small number of the brightest pixels inside active regions to saturate because 1) they are not our target of interest, 2) our entrance filter mesh mitigates diffraction (Section 3.4), and 3) the blooming in our detector is modest: only a few percent ranging across a few pixels (verified during the 33.336 NASA sounding rocket flight and in the lab). Projected performance: CME brightness at the outer SunCET FOV of 4 $R_{\odot}$ is $2.1\text{\times}{10}^{-4}$ W/m2/sr. That implies a dynamic range of $2.3\text{\times}{10}^{5}$. From 0–1.05 $R_{\odot}$, we run exposures of 0.025 seconds and from 1.05–4 $R_{\odot}$ the exposures will be 10 seconds — a factor of 400$\times$ dynamic range. Our detector has a native dynamic range of $\sim$$5\text{\times}{10}^{3}$. 2$\times$2 pixel binning provides an additional factor of 4. Combining these, we obtain SunCET’s high dynamic range of $8\text{\times}{10}^{6}$, well above the required range of $7\text{\times}{10}^{4}$. For comparison, the SDO/AIA dynamic range is $1\text{\times}{10}^{4}$ (Lemen et al., 2012). ### 3.2 Field of View Most CMEs accelerate through the low and middle corona (Bein et al., 2011; D’Huys et al., 2014). We set our required minimum field of view (FOV) at 0.5 $R_{\odot}$, corresponding to $\pm$30$\degree$ from disk-center. Lower than this and the events tend to be halo CMEs, which are difficult to obtain height-time profiles from. The outer FOV requirement is set to 3.5 $R_{\odot}$. SunCET covers the gap between existing instruments and includes enough overlap to ensure a smooth transition in any complementary height-time profiles. SOHO/LASCO’s inner FOV is 2.4 $R_{\odot}$ and its upcoming replacement, NOAA’s GOES-U/CCOR and SWFO/CCOR, will have an inner FOV of 3 $R_{\odot}$. The aforementioned traditional CME measurements, which are from white-light coronagraphs, use occulters that are mechanically restricted to be a limited distance away; therefore these observations have significantly degraded spatial resolution in their inner FOV that is much worse than their stated plate-scale resolution, sometimes upwards of 1 arc-min in the inner FOV. These effects are primarily due to vignetting (e.g. Koutchmy 1988; Aime et al. 2019). This is not the case with SunCET as it does not require an occulter to observe the CMEs in the low- and middle-corona, so its spatial resolution is not diffraction limited and is superior even in the FOV region that overlaps with the coronagraphs. Projected performance: The FOV of SunCET is 0–4 $R_{\odot}$ (5.6 $R_{\odot}$ in image corners). ### 3.3 Temporal Resolution: Exposure and Cadence SunCET is required to observe CMEs with speeds up to at least 1000 km/s, which accounts for 98% of all CMEs (Gopalswamy et al., 2009; Barlyaeva et al., 2018). Given the cadence described below and the field of view, SunCET’s projected performance is to observe CMEs with speeds up to 3900 km/s. The fastest CME in the CDAW catalog is $\sim$3400 km/s, meaning that SunCET will be able to track CMEs with any previously observed speed. SunCET requires an exposure time $\leq$23 seconds in order to avoid motion blur of the CME. Combining the fastest required CME to observe (1000 km/s), our required spatial resolution of 30/resolution-element, and the conversion of angular to spatial resolution at 1 AU ($\sim$750 km/arcsec), we obtain 750 $\times$ 30 / 1000 $\approx$ 23 seconds/resolution-element. Projected performance - exposure: SunCET’s exposure times are 0.025 seconds from 0–1.05 $R_{\odot}$ and 10 seconds beyond that. SunCET requires a cadence $\leq$3.2 minutes. SunCET must be able to track a 1000 km/s CME from the solar limb through its FOV, a range of 2.5 $R_{\odot}$, or $1.74\text{\times}{10}^{6}$ km. Therefore, the minimum time a CME would be in the FOV is 29 minutes. We require at least 9 height-time samples to distinguish acceleration profiles (Figure 3). Thus, our cadence must be less than 29 minutes / 9 samples = 3.2 minutes. Projected performance - cadence: The SunCET mission is designed to downlink 1 minute cadence data. The designed FOV actually extends to 4 $R_{\odot}$, meaning we will capture 38 height-time points for limb-CMEs traveling at a speed of 1000 km/s and more points for CMEs that start slightly on disk and/or with slower speeds. For example, the average CME speed is 490 km/s (Webb and Howard, 2012) and if it crosses from 0.7–4 $R_{\odot}$, we will obtain 78 height-time points. ### 3.4 Bandpass: Coatings and Filters Table 1: Strong emission lines in the SunCET bandpass. Irradiance measured by SDO/EVE (Woods et al., 2012) [HTML]1A73C9 Ion | $\uplambda$ [Å] | log10(T [K]) | Quiet Sun Irradiance [$\upmu$W/m2/Å] ---|---|---|--- Fe IX | 171.1 | 5.9 | 67 Fe X | 174.5 | 6.1 | 73 Fe X | 177.2 | 6.1 | 48 Fe XI | 180.4 | 6.2 | 77 | Fe XI --- (doublet) 188.2 | 6.2 | 61 Fe XII | 193.5 | 6.2 | 45 Fe XII | 195.1 | 6.2 | 63 CMEs have been routinely identified in narrowband EUV imagers sensitive to temperatures between $\sim$0.6–1.6 MK (e.g., GOES/SUVI). Therefore, SunCET is required to observe at least one of the emission lines identified in Table 1. Projected performance: SunCET’s baseline bandpass is 170–200 Å — capturing all of the emission lines in Table 1, which boosts the signal (Section 3.7). The telescope mirrors employ reflective multilayer coatings designed to provide broad spectral response spanning the instrument bandpass. These coatings follow an aperiodic design, and comprise 15 repetitions of alternating layers of B4C, Mo, and Al, with individual layer thicknesses ranging from $\sim$5–100 Å. The aperiodic coating design provides an average reflectance of $\sim$33% from 170–200 Å, as shown in Figure 6. For reference, periodic multilayer coatings operating in this portion of the EUV are generally used for narrow- band response: for example, the periodic Si/Mo coatings used for the 195 Å channel of the GOES/SUVI instrument, also shown in Figure 6, achieve a peak reflectance of $\sim$34% with a spectral bandpass of $\sim$9.5 Å full-width- half-max (FWHM). Figure 6 also shows the periodic Al/Zr coatings used for the Hi-C rocket instrument (Kobayashi et al., 2014), which achieve a peak reflectance of $\sim$50% with a spectral bandpass of $\sim$8.5 Å FWHM. The aperiodic B4C/Mo/Al multilayer coatings are currently under development with funding from the NASA H-TIDeS program. Figure 6: Calculated reflectance near normal incidence (5°) of the broad-band, aperiodic B4C/Mo/Al multilayers used for the SunCET telescope mirrors (green), and for reference, the narrow-band, periodic Si/Mo multilayer coatings used for the GOES/SUVI instrument (red), and the Al/Zr multilayer coatings used for the Hi-C rocket instrument (blue). The C/Al/C entrance filter from Luxel Corporation prevents visible light from entering the chamber and has high heritage (24 of them in GOES/EXIS). It is supported by a 5 lines/inch mesh, which has heritage from the Hi-C sounding rocket flights and avoids the diffraction issues of the 70 lines/inch mesh used on SDO/AIA and TRACE (Lemen et al., 2012; Lin et al., 2001). A second C/Al/C filter directly in front of the detector eliminates visible light from possible pinholes in the primary filter or from stray light in the instrument. ### 3.5 Spatial Resolution SunCET requires spatial resolution better than 30. CME flux ropes often manifest observationally as a cavity which trails behind a bright front (Forsyth et al., 2006). The smallest cavities have a diameter of 0.2 $R_{\odot}$ (180) and are approximately circular, which corresponds to a circumference of $\sim$600 (Fuller and Gibson, 2009). To account for non- circularities, we require $\sim$20 points outlining the cavity, which results in our spatial resolution requirement of 600/20 = 30. Figure 7 shows a cavity observed in PROBA2/SWAP (3.16 resolution) binned down to demonstrate that cavities can be resolved at this resolution in practice. Projected performance: SunCET provides 20 resolution. Its plate scale is 4.8/pixel so 2$\times$2 binning can be applied, which meets the Nyquist sampling criterion. Figure 7: CME cavity observed in PROBA2/SWAP 174 Å binned down to SunCET required resolution of 30 (projected performance is 20). The cavity remains easily identifiable. SunCET’s SNR will be 9–30$\times$ higher off disk, making CME identification even easier. The 1.7 $R_{\odot}$ FOV shown here, the largest of any solar EUV imager to date, is SWAP’s; SunCET’s extends to 4 $R_{\odot}$. Adapted from Byrne et al. (2014). ### 3.6 Mirrors Figure 8: SunCET’s compact Ritchey-Chrétien telescope, which fits inside a 6U CubeSat with all typical bus components. SunCET contains a Ritchey-Chrétien (RC) telescope encased in a vacuum chamber with a one-time-release door (Figure 8). This type of telescope has good performance for wide fields of view (Figure 9) and has been used frequently for similar instruments (e.g., SOHO/EIT, STEREO/EUVI, GOES/SUVI). Despite its compact size, the telescope achieves nearly flat resolution across the wide FOV. The mount for the secondary mirror is designed with a coefficient of thermal expansion matching the mirror to account for focus sensitivity. Figure 9: Left: Ray trace of SunCET optics. Right: 80% encircled spot diameter over the FOV. This simple design yields excellent performance, with a mean resolution of 20 that is flat across nearly the entire FOV. ### 3.7 Signal to Noise Ratio (SNR) SunCET requires a signal to noise ratio (SNR) $\geq$10\. This is the international standard that defines digital image quality as “acceptable” (ISO 12232, 2019). The same standard defines SNR of 40 as “excellent”. These numbers are in line with the expectations of experts that have done CME image processing with coronagraph and EUV imager data. Table 2: SunCET SNRs for on-disk features and CME loops above the limb. Radiances are from GOES/SUVI 195 Å images of the 2017-09-10 CME (Seaton and Darnel, 2018) and are extrapolated beyond its FOV of 1.7 $R_{\odot}$. SNR at all heights is above the level that ISO 12232 defines as “excellent”. [HTML]1A73C9 | Quiet Sun | Active Region | Flare | 1.05 $R_{\odot}$ | 1.5 $R_{\odot}$ | 2 $R_{\odot}$ | 3 $R_{\odot}$ | 3.5 $R_{\odot}$ | 4 $R_{\odot}$ ---|---|---|---|---|---|---|---|---|--- | Radiance --- [W/m2/sr] 0.1 | 10 | 40 | 0.2 | $1.5\text{\times}{10}^{-2}$ | $3\text{\times}{10}^{-3}$ | $3\text{\times}{10}^{-4}$ | $1\text{\times}{10}^{-4}$ | $3\text{\times}{10}^{-5}$ Effective exposure [s] | 0.025 | 0.025 | 0.025 | 0.025 | 10 | 10 | 10 | 10 | 10 e-/res-element | $1.48\text{\times}{10}^{4}$ | $1.48\text{\times}{10}^{6}$ | $5.94\text{\times}{10}^{6}$ | $2.97\text{\times}{10}^{4}$ | $8.9\text{\times}{10}^{5}$ | $1.78\text{\times}{10}^{5}$ | $1.78\text{\times}{10}^{4}$ | $5.94\text{\times}{10}^{3}$ | $1.78\text{\times}{10}^{3}$ | Saturation limit --- [e-/res-element] $1.08\text{\times}{10}^{5}$ | $1.08\text{\times}{10}^{5}$ | $1.08\text{\times}{10}^{5}$ | $1.08\text{\times}{10}^{5}$ | $1.08\text{\times}{10}^{6}$ | $1.08\text{\times}{10}^{6}$ | $1.08\text{\times}{10}^{6}$ | $1.08\text{\times}{10}^{6}$ | $1.08\text{\times}{10}^{6}$ SNR | 122 | Saturated | Saturated | 172 | 944 | 422 | 133 | 77 | 42 Projected performance: Table 2 shows the SunCET SNR as a function of distance from the sun, based on the parameters shown in Table 3. Conservative radiance estimates come from GOES/SUVI 195 Å images of a CME that was tracked all the way to the edge of the SUVI 1.7 $R_{\odot}$ FOV (Seaton and Darnel, 2018). For the solar disk, the effective exposure is the median of three 0.025-second images; for 1.05–4 $R_{\odot}$, it is the median of ten 1-second exposures. This removes energetic particle tracks and, for the long exposure, increases the full-well saturation limit of the detector by a factor of 10. These conservative estimates show that SunCET CME measurements would have an excellent SNR of 42 even out at 4 $R_{\odot}$. Table 3: SunCET instrument parameters needed to calculate SNR. [HTML]1A73C9 Instrument parameter | Value | Description ---|---|--- Wavelength | 170–200 Å | Broadband response defined by mirror coating | Aperture --- size 44.9 cm2 | | 9.6 cm diameter truncated on two sides --- to a height of 7.62 cm and a 4.8 cm diameter secondary mirror obscuring its center | Weighted factor --- for broadband 6.88 | | 7 emissions in the bandpass weighted by their --- quiet-Sun intensity to the 195 Å emission line (see Table 1) Pixel size | 7 $\upmu$m $\times$ 7 $\upmu$m | e2v CIS115 datasheet and confirmed in house Pixel array | 1500 $\times$ 1500 | Full array is 1504 $\times$ 2000; $\sim$5 rows dedicated to dark FOV | 4 $R_{\odot}$ | Design FOV (requirement is 3.5 $R_{\odot}$) Plate scale | 4.8/pixel | | From pixel size, number of pixels, and FOV; --- Note that 2$\times$2 binning will be applied, resulting in 9.6/resolution-element | Optics --- throughput 0.06 | | 2 mirrors with B4C/Mo/Al coatings (0.35 each), --- entrance Al/C filter (0.6) with 5 lpi filter mesh (0.98), Al secondary/pinhole filter (0.85) Quantum yield | 18.3 e-/ph | Average over 170–200 Å bandpass Dark noise | $<$0.08 e-/pixel/sec | At -10°C, from LASP lab tests Readout noise | 5 e-/pixel | From LASP lab tests Fano noise | 1.3 e-/pixel | Fano factor of 0.1 for Si Max read rate | | 0.1 sec (full frame) --- 0.025 (up to 500 rows) In SunCET, 500 rows corresponds to 0–1.33 $R_{\odot}$ Few observations of the extended corona above $\sim$2 $R_{\odot}$ have been made in the EUV, but among these there is clear evidence that the CME signal will be detectable (Tadikonda et al. 2019, Figure 10). At about 3 $R_{\odot}$, noise in SUVI becomes comparable to solar signal. SunCET, however, is optimized for this large FOV. SunCET has a larger primary mirror geometric area (3.5$\times$), broadband wavelength response (6.88$\times$), and larger pixel solid angle (16$\times$) for a total 385$\times$ boost in signal compared to SUVI. Furthermore, the SunCET mirrors will be polished to highest degree possible, up to 3 times the smoothness of SUVI’s, to minimize scattered light. Figure 10: Composite of GOES/SUVI 195 Å off-point images that shows solar structure out to 3 $R_{\odot}$ — even without a bright CME — before straylight in the instrument becomes comparable with the coronal signal. Adapted from Tadikonda et al. (2019). ## 4 Spectrograph Design The SunCET irradiance spectrograph channel is a high-heritage off-Rowland circle design based on the SDO/EVE Multiple EUV Grating Spectrographs A2 (MEGS-A2) channel (Crotser et al., 2007). It provides the full-Sun solar irradiance from 170-340 Å at 1 Å spectral resolution. This EUV range is important for overlapping with the SunCET imager EUV bands for calibration purposes and provides additional science capability. It observes Fe IX through Fe XVI emission lines that often experience coronal dimming during CMEs (Woods et al., 2011; Mason et al., 2014, 2016, 2019). This allows for halo CME kinematics to be tracked even if SunCET is not deployed on multiple platforms with stereoscopic viewing angles. It also enables study of the energetics powering the CME as a function of time. It shares the vacuum door and detector with the SunCET imager, but has its own optical path including the entrance slit, filters, and grating. These measurements are especially pressing because EVE/MEGS-A experienced a CCD electronics anomaly in 2014 May, preventing the continued solar observations by MEGS-A. While other EVE channels and new GOES EUV Sensor (EUVS) channels are continuing solar EUV observations in the 170-340 Å range, they are only broadband measurements that are not optimized for coronal dimming irradiance observations nor for detailed calibration of solar EUV imagers. ### 4.1 Spectrograph Dynamic Range The solar irradiance values, as measured from SDO/EVE (Woods et al., 2012), from 170-340 Å range from ${10}^{-6}$–${10}^{-2}$ W/m2/nm due to variations in the peaks of the emission line in this range, the reduced irradiance values between the strong emission lines, as well as solar activity including solar minimum times and during the largest solar flares; therefore, the required dynamic range of the spectrograph is $1\text{\times}{10}^{4}$. Projected performance: The $8\text{\times}{10}^{6}$ dynamic range discussed in Section 3.1 is more than two orders of magnitude better than needed for the spectrograph. ### 4.2 Spectrograph Spectral Range and Resolution The SunCET spectrograph requires a spectral range between 170-340 Å and 1 Å spectral resolution. The entrance to the spectrograph is a 3 $\times$ 0.028 mm in order to maximize the slit image height (cross-dispersion direction) on the allotted 500 pixel height of the detector to maximize the SNR, while the width is optimized to meet the 1 Å spectral resolution requirement — it is this slit width and the grating ruling that limits the spectral resolution. The grating ruling, distance and curvature are all optimized in order to meet the spectral range and resolution as well. The optical path after being dispersed from the grating goes through the hole in the secondary imager mirror and onto the common detector. The grating is a Type-I concave imaging grating in order to image the slit onto the detector. There is an Al/C entrance filter mounted to the entrance slit in order to limit the spectral bandpass close to the required range, and an additional Al filter prior for additional bandpass rejection at the entrance to the imager optical cavity as well as to reduce any stray light or pinholes that may develop in the first filter. Given the 1500 allotted pixels in the dispersion range, this gives a plate- scale resolution of approximately 0.11 Å per pixel; therefore the spectrograph will oversample the spectral resolution by about a factor of 9$\times$, or 4.5$\times$ with the 2$\times$2 pixel binning. This allows for fits to spectral lines to be performed and allow for Doppler shift measurements of emission lines and plasma velocity flows during flares to be calculated (Chamberlin, 2016; Hudson et al., 2011) Projected performance: SunCET provides 1 Å spectral resolution across the fully observed 170-340 Å spectral range. ### 4.3 Spectrograph Signal to Noise Ratio (SNR) The SunCET spectrograph also requires a SNR of 10 or better as discussed in 3.7. This is achieved by using a long-slit and minimal optical elements, along with the high QE detector. The slit was also sized, and filter thickness optimized, to maximize the SNR without while conservatively not saturating or even go beyond the linear full well capacity of the the CMOS sensor. Even with a very large factor of 10 increase (Chamberlin et al., 2008, 2018) during flares for these lines given in Table 4, they will still be almost another factor of 2 below the full-well of this sensor. Table 4: The SunCET spectrograph SNRs for various strong emission lines. Irradiances are from SDO/EVE (Woods et al., 2011). SNR at all heights is above the level that ISO 12232 defines as “excellent”. [HTML]1A73C9 Wavelength (Å) | 171 | 193.5 | 195 | 304 | 335 ---|---|---|---|---|--- | Irradiance --- [W/m2/sr] $6.7\text{\times}{10}^{-4}$ | $4.5\text{\times}{10}^{-4}$ | $6.3\text{\times}{10}^{-4}$ | $1.0\text{\times}{10}^{-3}$ | $1.0\text{\times}{10}^{-4}$ Integration [s] | 10 | 10 | 10 | 10 | 10 Counts/Pixel | 737 | 495 | 693 | 1100 | 110 SNR | 272 | 237 | 282 | 444 | 145 Projected performance: Table 4 shows the SunCET spectrograph SNR for five strong emission lines, based on the parameters shown in Table 5. These estimates show that SunCET solar spectral irradiance measurements would have an excellent SNR of better than 100. Table 5: SunCET spectrograph instrument parameters needed to calculate SNR. [HTML]1A73C9 Instrument parameter | Value | Description ---|---|--- Wavelength | 170–340 Å | | Contains various strong emission lines, --- including some that show coronal dimming. Defined by grating equation. | Aperture --- size 0.0098 cm2 | 3.0 mm tall $\times$ 28 $\upmu$m wide | Number of Pixels --- per emission line 2000 | | 500 pixels tall $\times$ 4 pixels wide --- (defined by slit) Pixel size | 7 $\upmu$m $\times$ 7 $\upmu$m | Teledyne e2v CIS115 datasheet and confirmed in house Pixel allocation | 500 $\times$ 1500 | Full array is 1504 $\times$ 2000; $\sim$5 rows dedicated to dark FOV | Full Sun | Solar Irradiance, image the slit Plate scale | 0.011 nm | | From pixel size, number of pixels, wavelength range; --- Note: oversampling spectral resolution of 0.1nm | Optics --- throughput 0.0122 | | Grating Efficiency (0.06), Pt Grating Coating (0.4), --- Al/C entrance filter (0.6), Al secondary/pinhole filter (0.85) Quantum yield | 18.3 e-/ph | Average over 170-200 Å bandpass Dark noise | $<$0.08 e-/pixel/sec | At -10°C, from LASP lab tests Readout noise | 5 e-/pixel | From LASP lab tests Fano noise | 1.3 e-/pixel | Fano factor of 0.1 for Si ## 5 Detector Figure 11: The Teledyne e2v CIS115 detector and LASP Compact Camera and Processor (CCAP) that flew successfully on a NASA sounding rocket in 2018; CCAP is now flying on the CSIM CubeSat launched in 2018. SunCET uses a Teledyne e2v CIS115 back-illuminated, back-thinned CMOS sensor (Table 3, Figure 11). This sensor is a 1504$\times$2000 pixel array, where a square area of 1500$\times$1500 pixels will be dedicated to the image while the remaining 500$\times$1500 pixels will record the spectrally dispersed slit image from the irradiance spectrograph. Using a single detector to record data from two technically different but scientifically complementary channels significantly reduces the technical resources needed while maximizing science potential. In 2017, LASP developed custom electronics for readout of this sensor that enables independent exposure control per row. A per-pixel readout is now being developed. LASP’s “Compact Camera and Processor” (CCAP; Figure 11) system with this detector was successfully flown in 2018 on the NASA 36.336 sounding rocket (PI: T. Woods, U. of Colorado/LASP) and more recently in January 2020 on the NASA 36.356 sounding rocket (PI: S. Bailey, Virginia Tech). CCAP includes a Xilinx Kintex-7 FPGA with an embedded 32-bit processor and dedicated image compression core. ## 6 Instrument Requirements on Spacecraft The instruments described above place requirements on the performance and capabilities of whatever spacecraft hosts them. They are primarily driven by the imager. Pointing accuracy must be better than 30with stability better than 30 RMS over 23 seconds and knowledge better than 10. This ensures that the center of the sun stays in the center of portion of the detector dedicated to the imager and does not drift significantly during or between integrations. This pointing performance is achievable even on CubeSat platforms as demonstrated by the Miniature X-ray Solar Spectrometer (MinXSS), Arcsecond Space Telescope Enabling Research in Astrophysics (ASTERIA), Compact Spectral Irradiance Monitor (CSIM), and others (Mason et al., 2017; Pong, 2018). Prime science data generation is heavily dependent on CME occurrence rates, but downlink schemes can easily be designed for flexibility and the “poorest” CMEs can be ignored if there are bandwidth limitations. For CME occurrence rates at the middle of the rising phase of the solar cycle, SunCET generates $\sim$28 MB/day for the imager, and $\sim$65 MB/day of data for the spectrograph. These data are compressed using a lossless JPEG-LS scheme. ## 7 Conclusions The SunCET instrument fills a crucial, historically under observed region of the Sun — the middle corona — precisely the region where CMEs experience the majority of their acceleration. This region is inherently very difficult to observe because of the extreme intensity dynamic range between the bright solar disk and the dim corona. SunCET introduces a new technology that avoids the limitations of previous instruments. By developing a detector that can vary exposure time across its surface, we can simultaneously observe the disk without saturating and the dim middle corona; allowing us to track CMEs from their initiation all the way through their primary acceleration phase. Moreover, we can image spectra on the same detector with their own, independent integration time. Figure 12: Tracking a very fast CME in GOES/SUVI 195 Å base-difference images. The CME quickly extended beyond the FOV of SUVI. SunCET’s FOV (light blue shading) is more than twice as large. Adapted from Veronig et al. (2018). There is a large body of knowledge for tracking CMEs in coronagraphs and EUV imagers (Sarkar et al., 2019; O’Hara et al., 2019; Veronig et al., 2018; Byrne et al., 2014; Mierla et al., 2013; Bein et al., 2011; Gopalswamy et al., 2009; Vršnak et al., 2007). SunCET data processing will employ the techniques already developed for other observatories but improve the results because of its wider FOV (e.g., Veronig et al. 2018; Figure 12) and that it does not require the serendipitous alignment between instrument off-point campaigns and CME occurrence (e.g., O’Hara et al. 2019). Below we summarize: 1. 1. The majority of CME acceleration occurs in a historical observational gap: the middle corona 2. 2. Observations of full CME acceleration profiles provide tight constraints on models and thus our physical understanding of how the magnetically-dominated corona influences CME kinematics 3. 3. SunCET provides these observations, overcoming the limits of traditional technologies with a novel simultaneous high-dynamic-range detector 4. 4. SunCET is compact and thus suitable for CubeSat missions or an instrument on a larger spacecraft SunCET is presently in a NASA-funded, competitive Phase A as a 6U CubeSat and has also been proposed to NASA as an instrument onboard a 184 kg Mission of Opportunity. ## 8 Acknowledgements J.P.M thanks the numerous people who contributed to the development of the SunCET concept design and the reviewers for their commments that made this paper stronger. A.M.V. and K.D. acknowledge the Austrian Space Applications Programme of the Austrian Research Promotion Agency FFG (ASAP-11 4900217 CORDIM and ASAP-14 865972 SSCME, BMVIT). ## References * Aime et al. (2019) Aime, C., C. Theys, R. Rougeot, and H. Lantéri, 2019. Principle of Fredholm image reconstruction in the vignetting zone of an externally occulted solar coronagraph: Application to ASPIICS. _Astronomy & Astrophysics_, 622, A212. 10.1051/0004-6361/201833843, URL https://doi.org/10.1051/0004-6361/201833843. * Aschwanden (2009) Aschwanden, M. J., 2009. 4-D modeling of CME expansion and EUV dimming observed with STEREO/EUVI. _Annales Geophysicae_ , 27(8), 3275–3286. 10.5194/angeo-27-3275-2009, URL http://www.ann-geophys.net/27/3275/2009/. * Aschwanden et al. (2009) Aschwanden, M. J., N. V. Nitta, J.-P. Wuelser, J. R. Lemen, A. Sandman, A. Vourlidas, and R. C. Colaninno, 2009. First Measurements of the Mass of Coronal Mass Ejections From the EUV Dimming Observed With Stereo EUVI A + B Spacecraft. _The Astrophysical Journal_ , 706(1), 376–392. 10.1088/0004-637X/706/1/376, URL http://stacks.iop.org/0004-637X/706/i=1/a=376?key=crossref.88f60571a09db37b8197341ac713fd1a. * Barlyaeva et al. (2018) Barlyaeva, T., J. Wojak, P. Lamy, B. Boclet, and I. Toth, 2018. Periodic behaviour of coronal mass ejections, eruptive events, and solar activity proxies during solar cycles 23 and 24. _Journal of Atmospheric and Solar-Terrestrial Physics_ , 177, 12–28. 10.1016/j.jastp.2018.05.012. * Bateman (1978) Bateman, G., 1978. MHD Instabilities. MIT Press, Cambridge, Massachusetts. ISBN 9780262021319. * Bein et al. (2011) Bein, B. M., S. Berkebile-Stoiser, A. M. Veronig, M. Temmer, N. Muhr, I. Kienreich, D. Utz, and B. Vršnak, 2011. Impulsive Acceleration of Coronal Mass Ejections. I. Statistics and Coronal Mass Ejection Source Region Characteristics. _Astrophysical Journal_ , 738(2), 191. 10.1088/0004-637X/738/2/191, URL http://stacks.iop.org/0004-637X/738/i=2/a=191?key=crossref.f0398b90f91cbeb1263748f98e279bbd. * Byrne et al. (2014) Byrne, J. P., H. Morgan, D. B. Seaton, H. M. Bain, and S. R. Habbal, 2014. Bridging EUV and white-light observations to inspect the initiation phase of a “two-stage” solar eruptive event. _Solar Physics_ , 289(12), 4545–4562. 10.1007/s11207-014-0585-8. * Chamberlin (2016) Chamberlin, P. C., 2016. Measuring Solar Doppler Velocities in the He ii 30.38 nm Emission Using the EUV Variability Experiment (EVE). _Solar Physics_ , 291(6), 1665–1679. 10.1007/s11207-016-0931-0, URL https://link.springer.com/article/10.1007/s11207-016-0931-0. * Chamberlin et al. (2018) Chamberlin, P. C., T. N. Woods, L. Didkovsky, F. G. Eparvier, A. R. Jones, et al., 2018. Solar Ultraviolet Irradiance Observations of the Solar Flares During the Intense September 2017 Storm Period. _Space Weather_ , 16(10), 1470–1487. 10.1029/2018SW001866, URL https://agupubs.onlinelibrary.wiley.com/doi/full/10.1029/2018SW001866. * Chamberlin et al. (2008) Chamberlin, P. C., T. N. Woods, and F. G. Eparvier, 2008. Flare Irradiance Spectral Model (FISM): Flare component algorithms and results. _Space Weather_ , 6(5), n/a–n/a. 10.1029/2007SW000372, URL http://doi.wiley.com/10.1029/2007SW000372. * Chikunova et al. (2020) Chikunova, G., K. Dissauer, T. Podladchikova, and A. M. Veronig, 2020. Coronal dimmings associated with coronal mass ejections on the solar limb. _submitted_. URL http://arxiv.org/abs/2005.03348. * Crotser et al. (2007) Crotser, D. A., T. N. Woods, F. G. Eparvier, M. A. Triplett, and D. L. Woodraska, 2007. SDO-EVE EUV spectrograph optical design and performance. In S. Fineschi and R. A. Viereck, eds., Solar Physics and Space Weather Instrumentation II, vol. 6689, 66890M. SPIE. 10.1117/12.732592, URL http://proceedings.spiedigitallibrary.org/proceeding.aspx?doi=10.1117/12.732592. * D’Huys et al. (2014) D’Huys, E., D. B. Seaton, S. Poedts, and D. Berghmans, 2014. Observational characteristics of coronal mass ejections without low-coronal signatures. _Astrophysical Journal_ , 795(1), 49. 10.1088/0004-637X/795/1/49. * Dissauer et al. (2019) Dissauer, K., A. M. Veronig, M. Temmer, and T. Podladchikova, 2019. Statistics of coronal dimmings associated with coronal mass ejections. II. Relationship between coronal dimmings and their associated CMEs. _The Astrophysical Journal_ , 874(2), 123. 10.3847/1538-4357/ab0962, URL http://stacks.iop.org/0004-637X/874/i=2/a=123?key=crossref.833720587c5d6f444910c7dec84f30d9http://arxiv.org/abs/1810.01589. * Dissauer et al. (2018) Dissauer, K., A. M. Veronig, M. Temmer, T. Podladchikova, and K. Vanninathan, 2018\. Statistics of Coronal Dimmings Associated with Coronal Mass Ejections. I. Characteristic Dimming Properties and Flare Association. _The Astrophysical Journal_ , 863(2), 169. 10.3847/1538-4357/aad3c6, URL http://stacks.iop.org/0004-637X/863/i=2/a=169?key=crossref.04b1e3af0e5af3d849583869b3fb6f27. * Fan (2016) Fan, Y., 2016. Modeling the Initiation of the 2006 December 13 Coronal Mass Ejection in AR 10930: The Structure and Dynamics of the Erupting Flux Rope. _The Astrophysical Journal_ , 824(93), 12. 10.3847/0004-637x/824/2/93, URL http://arxiv.org/abs/1604.05687http://dx.doi.org/10.3847/0004-637X/824/2/93. * Forbes et al. (2018) Forbes, T. G., D. B. Seaton, and K. K. Reeves, 2018. Reconnection in the Post-impulsive Phase of Solar Flares. _The Astrophysical Journal_ , 858, 70. 10.3847/1538-4357/aabad4. * Forsyth et al. (2006) Forsyth, R. J., V. Bothmer, C. Cid, N. U. Crooker, T. S. Horbury, et al., 2006. ICMEs in the inner heliosphere: Origin, evolution and propagation effects: Report of working group G. _Space Science Reviews_ , 123, 383–416. 10.1007/s11214-006-9022-0. * Fuller and Gibson (2009) Fuller, J., and S. E. Gibson, 2009. A survey of coronal cavity density profiles. _The Astrophysical Journal_ , 700, 1205–1215. 10.1088/0004-637X/700/2/1205. * Gopalswamy et al. (2009) Gopalswamy, N., S. Yashiro, G. Michalek, G. Stenborg, A. Vourlidas, F. S. L, and R. A. Howard, 2009. The SOHO / LASCO CME Catalog. _Earth Moon Planet_ , 104, 295–313. 10.1007/s11038-008-9282-7. * Green et al. (2018) Green, L. M., T. Török, B. Vršnak, W. Manchester, and A. Veronig, 2018. The Origin, Early Evolution and Predictability of Solar Eruptions. _Space Science Reviews_ , 214(1), 46. 10.1007/s11214-017-0462-5. * Hood and Priest (1981) Hood, A. W., and E. R. Priest, 1981. Critical Conditions for Magnetic Instabilities in Force-Free Coronal Loops. _Geophysical & Astrophysical Fluid Dynamics_, 17(1), 297–318. 10.1080/03091928108243687. * Howard et al. (2008) Howard, R. A., J. D. Moses, A. Vourlidas, J. S. Newmark, D. G. Socker, et al., 2008\. Sun Earth Connection Coronal and Heliospheric Investigation (SECCHI). _Space Science Reviews_ , 136(1-4), 67–115. 10.1007/s11214-008-9341-4, URL http://link.springer.com/10.1007/s11214-008-9341-4. * Hudson et al. (2011) Hudson, H. S., T. N. Woods, P. C. Chamberlin, L. Fletcher, G. Del Zanna, L. Didkovsky, N. Labrosse, and D. Graham, 2011. The EVE Doppler Sensitivity and Flare Observations. _Solar Physics_ , 273(1), 69–80. 10.1007/s11207-011-9862-y, URL http://link.springer.com/10.1007/s11207-011-9862-y. * ISO 12232 (2019) ISO 12232, 2019. Photography – Digital Still Cameras – Determination of Exposure Index, ISO Speed Ratings, Standard Output Sensitivity, and Recommended Exposure Index. _Tech. rep._ , International Organization for Standardization, Geneva, CH. * Kaiser et al. (2007) Kaiser, M. L., T. A. Kucera, J. M. Davila, O. C. St. Cyr, M. Guhathakurta, and E. Christian, 2007. The STEREO Mission: An Introduction. _Space Science Reviews_ , 136(1-4), 5–16. 10.1007/s11214-007-9277-0, URL http://link.springer.com/10.1007/s11214-007-9277-0. * Kay and Gopalswamy (2018) Kay, C., and N. Gopalswamy, 2018. The Effects of Uncertainty in Initial CME Input Parameters on Deflection, Rotation, Bz , and Arrival Time Predictions. _Journal of Geophysical Research: Space Physics_ , 123, 7220–7240. 10.1029/2018JA025780. * Kay and Opher (2015) Kay, C., and M. Opher, 2015. The Heliocentric Distance Where the Deflections and Rotations of Solar Coronal Mass Ejections Occur. _The Astrophysical Journal Letters_ , 811, L36. 10.1088/2041-8205/811/2/L36. * Kay et al. (2016) Kay, C., M. Opher, R. C. Colaninno, and A. Vourlidas, 2016. Using ForeCAT Deflections and Rotations to Constrain the Early Evolution of CMEs. _The Astrophysical Journal_ , 827(1), 70. 10.3847/0004-637X/827/1/70, URL http://arxiv.org/abs/1606.03460%****␣main.bbl␣Line␣250␣****http://dx.doi.org/10.3847/0004-637X/827/1/70http://stacks.iop.org/0004-637X/827/i=1/a=70?key=crossref.fd824e09edb6a2c662122ed885296355. * Kay et al. (2013) Kay, C., M. Opher, and R. M. Evans, 2013. Forecasting a Coronal Mass Ejection’s Altered Trajectory: ForeCAT. _The Astrophysical Journal_ , 775(1), 5. 10.1088/0004-637X/775/1/5, URL http://stacks.iop.org/0004-637X/775/i=1/a=5?key=crossref.60dd88082ab7f70bf71897944c86b722. * Kay et al. (2015) Kay, C., M. Opher, and R. M. Evans, 2015. Global Trends of CME Deflections Based on CME and Solar Parameters. _The Astrophysical Journal_ , 805(2), 168. 10.1088/0004-637X/805/2/168, URL http://stacks.iop.org/0004-637X/805/i=2/a=168?key=crossref.7f7c0fc9c0ff14b3f341631e5120e261. * Kay (2016) Kay, C. D., 2016. ForeCAT - A Model for Magnetic Deflections of Coronal Mass Ejections. Ph.D. thesis, Boston University. URL https://search.proquest.com/docview/1767403214. * Kliem and Török (2006) Kliem, B., and T. Török, 2006. Torus Instability. _Physical Review Letters_ , 96(1), 4. 10.1103/PhysRevLett.96.255002. * Kobayashi et al. (2014) Kobayashi, K., J. Cirtain, A. R. Winebarger, K. Korreck, L. Golub, et al., 2014\. The High-Resolution Coronal Imager (Hi-C). _Solar Physics_ , 289(11), 4393–4412. 10.1007/s11207-014-0544-4, URL http://link.springer.com/10.1007/s11207-014-0544-4. * Koutchmy (1988) Koutchmy, S., 1988. Space-Born Coronagraphy. _Space Science Reviews_ , 47, 95–143. URL http://articles.adsabs.harvard.edu/pdf/1988SSRv...47...95K. * Lemen et al. (2012) Lemen, J. R., A. M. Title, D. J. Akin, P. F. Boerner, C. Chou, et al., 2012. The Atmospheric Imaging Assembly (AIA) on the Solar Dynamics Observatory (SDO). _Solar Physics_ , 275(1-2), 17–40. 10.1007/s11207-011-9776-8, URL http://link.springer.com/10.1007/s11207-011-9776-8. * Lin et al. (2001) Lin, A. C., R. W. Nightingale, and T. D. Tarbell, 2001. Diffraction pattern analysis of bright trace flares. _Solar Physics_ , 198(2), 385–398. 10.1023/A:1005213527766. * Martínez-Galarce et al. (2010) Martínez-Galarce, D., J. Harvey, M. Bruner, J. Lemen, E. Gullikson, R. Soufli, E. Prast, and S. Khatri, 2010. A novel forward-model technique for estimating EUV imaging performance: design and analysis of the SUVI telescope. In Space Telescopes and Instrumentation 2010: Ultraviolet to Gamma Ray, vol. 7732, 773,237–1. 10.1117/12.864577. * Mason et al. (2019) Mason, J. P., R. Attie, C. N. Arge, B. Thompson, and T. N. Woods, 2019. The SDO/EVE Solar Irradiance Coronal Dimming Index Catalog. I. Methods and Algorithms. _The Astrophysical Journal Supplement Series_ , 244(1), 13\. 10.3847/1538-4365/ab380e, URL https://iopscience.iop.org/article/10.3847/1538-4365/ab380e. * Mason et al. (2017) Mason, J. P., M. Baumgart, B. Rogler, C. Downs, M. Williams, et al., 2017. MinXSS-1 CubeSat On-Orbit Pointing and Power Performance: The First Flight of the Blue Canyon Technologies XACT 3-axis Attitude Determination and Control System. _Journal of Small Satellites_ , 6(3), 651–662. URL http://arxiv.org/abs/1706.06967https://jossonline.com/letters/minxss-1-cubesat-on-orbit-pointing-and-power-performance-the-first-flight-of-the-blue-canyon-technologies-xact-3-axis-attitude-determination-and-control-system/. * Mason et al. (2014) Mason, J. P., T. N. Woods, A. Caspi, B. J. Thompson, and R. A. Hock, 2014. Mechanisms and Observations of Coronal Dimming for the 2010 August 7 Event. _The Astrophysical Journal_ , 789(1), 61. 10.1088/0004-637X/789/1/61, URL http://adsabs.harvard.edu/abs/2014ApJ...789...61M. * Mason et al. (2016) Mason, J. P., T. N. Woods, D. F. Webb, B. J. Thompson, R. C. Colaninno, and A. Vourlidas, 2016. Relationship of EUV Irradiance Coronal Dimming Slope and Depth to Coronal Mass Ejection Speed and Mass. _The Astrophysical Journal_ , 830(20), 12. 10.3847/0004-637X/830/1/20, URL http://stacks.iop.org/0004-637X/830/i=1/a=20?key=crossref.2d956aff9237fc3069d8edd80c37186d. * Mierla et al. (2013) Mierla, M., D. B. Seaton, D. Berghmans, I. Chifu, A. De Groof, B. Inhester, L. Rodriguez, G. Stenborg, and A. N. Zhukov, 2013. Study of a Prominence Eruption using PROBA2/SWAP and STEREO/EUVI Data. _Solar Physics_ , 286(1), 241–253. 10.1007/s11207-012-9965-0, URL http://link.springer.com/10.1007/s11207-012-9965-0. * O’Hara et al. (2019) O’Hara, J. P., M. Mierla, O. Podladchikova, E. D’Huys, and M. J. West, 2019\. Exceptional Extended Field-of-view Observations by PROBA2 /SWAP on 2017 April 1 and 3 . _The Astrophysical Journal_ , 883(1), 59. 10.3847/1538-4357/ab3b08, URL http://dx.doi.org/10.3847/1538-4357/ab3b08. * Pong (2018) Pong, C., 2018. On-Orbit Performance and Operation of the Attitude and Pointing Control Subsystems on ASTERIA. In AIAA/USU Conference on Small Satellites. Logan, UT. URL https://digitalcommons.usu.edu/smallsat/2018/all2018/361. * Sarkar et al. (2019) Sarkar, R., N. Srivastava, M. Mierla, M. J. West, and E. D’Huys, 2019. Evolution of the Coronal Cavity From the Quiescent to Eruptive Phase Associated with Coronal Mass Ejection. _The Astrophysical Journal_ , 875, 101. 10.3847/1538-4357/ab11c5. * Schrijver et al. (2008) Schrijver, C. J., C. Elmore, B. Kliem, T. Torok, and A. M. Title, 2008. Observations and Modeling of the Early Acceleration Phase of Erupting Filaments Involved in Coronal Mass Ejections. _The Astrophysical Journal_ , 674(1), 586–595. 10.1086/524294. * Seaton et al. (2013) Seaton, D. B., D. Berghmans, B. Nicula, J. P. Halain, A. De Groof, et al., 2013\. The SWAP EUV Imaging Telescope Part I: Instrument Overview and Pre-Flight Testing. _Solar Physics_ , 286(1), 43–65. 10.1007/s11207-012-0114-6. * Seaton and Darnel (2018) Seaton, D. B., and J. M. Darnel, 2018. Observations of an Eruptive Solar Flare in the Extended EUV Solar Corona. _The Astrophysical Journal Letters_ , 852, L9. 10.3847/2041-8213/aaa28e. * Tadikonda et al. (2019) Tadikonda, S. K., D. C. Freesland, R. R. Minor, D. B. Seaton, G. J. Comeyne, and A. Krimchansky, 2019. Coronal Imaging with the Solar UltraViolet Imager. _Solar Physics_ , 294, 28. 10.1007/s11207-019-1411-0. * Thompson et al. (2000) Thompson, B. J., E. W. Cliver, N. V. Nitta, C. Delannée, and J. P. Delaboudiniere, 2000. Coronal Dimmings and Energetic CMEs in April-May 1998. _Geophysical Research Letters_ , 27(10), 1431–1434. * Török and Kliem (2007) Török, T., and B. Kliem, 2007. Numerical simulations of fast and slow coronal mass ejections. _Astronomische Nachrichten_ , 328(8), 743–746. 10.1002/asna.200710795. * Veronig et al. (2018) Veronig, A. M., T. Podladchikova, K. Dissauer, M. Temmer, D. B. Seaton, D. Long, J. Guo, B. Vršnak, L. Harra, and B. Kliem, 2018. Genesis and Impulsive Evolution of the 2017 September 10 Coronal Mass Ejection. _The Astrophysical Journal_ , 868, 107. 10.3847/1538-4357/aaeac5, URL https://doi.org/10.3847/1538-4357/aaeac5. * Vršnak et al. (2007) Vršnak, B., D. Maričić, A. L. Stanger, A. M. Veronig, M. Temmer, and D. Roša, 2007. Acceleration phase of coronal mass ejections: I. Temporal and spatial scales. _Solar Physics_ , 241(1), 85–98. 10.1007/s11207-006-0290-3, URL https://link.springer.com/article/10.1007/s11207-006-0290-3. * Webb and Howard (2012) Webb, D. F., and T. A. Howard, 2012. Coronal Mass Ejections: Observations. _Living Reviews in Solar Physics_ , 9, 3. 10.12942/lrsp-2012-3, URL http://link.springer.com/10.12942/lrsp-2012-3. * Woods et al. (2012) Woods, T. N., F. G. Eparvier, R. A. Hock, A. R. Jones, D. L. Woodraska, et al., 2012\. Extreme Ultraviolet Variability Experiment (EVE) on the Solar Dynamics Observatory (SDO): Overview of Science Objectives, Instrument Design, Data Products, and Model Developments. _Solar Physics_ , 275, 115–143. 10.1007/s11207-009-9487-6, URL http://link.springer.com/10.1007/s11207-009-9487-6. * Woods et al. (2011) Woods, T. N., R. A. Hock, F. G. Eparvier, A. R. Jones, P. C. Chamberlin, et al., 2011. New Solar Extreme-Ultraviolet irradiance Observations During Flares. _The Astrophysical Journal_ , 739, 59. 10.1088/0004-637X/739/2/59.